<iframe src="//www.googletagmanager.com/ns.html?id=GTM-5TSRKG" height="0" width="0" style="display: none; visibility: hidden"></iframe>
Experimental EndourologyOpen AccessOpen Access license

Support System of Cystoscopic Diagnosis for Bladder Cancer Based on Artificial Intelligence

    Published Online:https://doi.org/10.1089/end.2019.0509

    Abstract

    Introduction: Nonmuscle-invasive bladder cancer has a relatively high postoperative recurrence rate despite the implementation of conventional treatment methods. Cystoscopy is essential for diagnosing and monitoring bladder cancer, but lesions are overlooked while using white-light imaging. Using cystoscopy, tumors with a small diameter; flat tumors, such as carcinoma in situ; and the extent of flat lesions associated with the elevated lesions are difficult to identify. In addition, the accuracy of diagnosis and treatment using cystoscopy varies according to the skill and experience of physicians. Therefore, to improve the quality of bladder cancer diagnosis, we aimed to support the cystoscopic diagnosis of bladder cancer using artificial intelligence (AI).

    Materials and Methods: A total of 2102 cystoscopic images, consisting of 1671 images of normal tissue and 431 images of tumor lesions, were used to create a dataset with an 8:2 ratio of training and test images. We constructed a tumor classifier based on a convolutional neural network (CNN). The performance of the trained classifier was evaluated using test data. True-positive rate and false-positive rate were plotted when the threshold was changed as the receiver operating characteristic (ROC) curve.

    Results: In the test data (tumor image: 87, normal image: 335), 78 images were true positive, 315 true negative, 20 false positive, and 9 false negative. The area under the ROC curve was 0.98, with a maximum Youden index of 0.837, sensitivity of 89.7%, and specificity of 94.0%.

    Conclusion: By objectively evaluating the cystoscopic image with CNN, it was possible to classify the image, including tumor lesions and normality. The objective evaluation of cystoscopic images using AI is expected to contribute to improvement in the accuracy of the diagnosis and treatment of bladder cancer.

    Introduction

    Spatial and temporal development at multiple sites is a clinical characteristic of bladder cancer, and the frequencies of its presence at multiple intravesical sites at the time of diagnosis and intravesical recurrence after complete endoscopic resection of visible lesions are high.1 The standard treatment for nonmuscle-invasive bladder cancer (NMIBC) is endoscopic transurethral resection of the bladder tumor, but it has been reported that intravesical recurrence occurs within 2 years after operation in ∼50% of cases.2 Moreover, the recurrence rates at the first cystoscopy after TUR-BT vary widely among institutions, and the diagnostic accuracy also varies according to differences in the urologist's skill and experience.3,4

    Cystoscopy is essential for diagnosing bladder cancer and observing the course, but lesions are overlooked during observation under white-light imaging (WLI) in 10%–20% of cases.5,6 The reported sensitivity and specificity of diagnosis under WLI are ∼60% and 70%,7 respectively, and it is difficult to identify the expansion of flat lesions accompanying flat-type tumors and elevated lesions of tumors with a small diameter and carcinoma in situ. Endoscopic techniques, such as narrow band imaging (NBI) and photodynamic diagnosis (PDD), have been developed to improve visibility of bladder cancers, and their usefulness has been demonstrated8–12; however, WLI is still the primary method of observation.

    Imaging diagnosis utilizing artificial intelligence (AI) has recently been developed in the field of medicine. Endoscopic imaging-based diagnosis using AI has been clinically applied in the field of gastroenterology, but its application to the field of urology was initiated only recently.13–16 In this study, to improve the quality of bladder cancer diagnosis, we aimed to objectively evaluate normal and tumor images obtained by white-light cystoscopy using AI.

    Materials and Methods

    Preparation of training and test image sets

    Endoscopic images of the bladder acquired at the University of Tsukuba Hospital between February 2017 and July 2018 were used. For endoscopic images, still images acquired using a flexible endoscope (CYF-VHA; Olympus Medical System, Co., Ltd., Tokyo, Japan) at the outpatient clinic were used. For the endoscopic image of the bladder, 2102 images were obtained with the TIFF file of 1350 × 1080 pixels by white light. Images with urine turbidity and images that were out of focus were excluded. One dataset included 431 images of tumors, and one urologist marked the sites judged to be a tumor using two patterns, namely, elevated and flat lesions, regardless of whether the tumor was malignant or benign. The actual judgment was also confirmed with the pathologic results as ground truth. The other dataset contained 1671 normal images, judged by the same doctor as not showing tumor lesions based on cystoscopy performed 1637 times around the same period. Figure 1 presents the normal image, tumor image, and a sample of the annotation data.

    FIG. 1.

    FIG. 1. Sample of cystoscopic and annotation images.

    Constructing a convolutional neural network and outcome measures

    The cystoscopic images were evaluated using the convolutional neural network (CNN) model based on GoogLeNet17 with transfer learning. GoogLeNet is a CNN model that has been awarded the first prize at ImageNet large scale visual recognition challenge 2014, and developed by training through the task of 1000-class classification using ImageNet18 dataset consisting of 1.2 million natural images. In the proposed method, network parameters of this pretrained CNN model were transferred to the initial network parameters to learn cystoscopic images additionally in supervised learning (Fig. 2). Then, all network parameters of this CNN were fine-tuned using the Adam algorithm19 to identify lesions and normal images. The training dataset was augmented by adding new data that were generated by rotating and blurring the images, which were randomly chosen from the original image. Using this augmentation, the numbers of tumor and normal images in the training dataset were obtained in the ratio of 1:1. All endoscopic images of the bladder were randomly divided into two (the training and test) datasets with a ratio of 8:2. The learning rate was 1e-5, and the epoch count was 150. Whether the image is an endoscopic image of the bladder containing lesions overall or a normal image could be distinguished by performing learning by GoogLeNet as a classifier using multilayer perceptron in the discriminator with set learning data (tumor images: 344, normal images: 1336) and test data (tumor images: 87, normal images: 335).

    FIG. 2.

    FIG. 2. A method based on transfer learning with pretrained convolutional neural networks.

    Statistical analysis

    The programming language Python was used throughout the experiment. We used “Chainer,” which is a deep learning framework to model and learn CNN. When evaluating, we used the machine learning library “scikit-learn” to calculate the confusion matrix and receiver operating characteristic (ROC) curve. The classifier was trained to output 1 if the image showed a tumor and 0 if the image was normal using training data. The performance of the classifier was evaluated using test data. True-positive rate and false-positive rate were plotted when the threshold was changed from 0 to 1 as the ROC curve. The area under the curve (AUC) was calculated from the ROC curve. In addition, we calculated the maximum Youden index (YI), sensitivity, and specificity.

    Ethics

    This study was approved by the Ethics Committee of the University of Tsukuba Hospital (no. H28-235) and National Institute of Advanced Industrial Science and Technology (no. Hi2016-224). An opt-out approach was used when obtaining consent from the patients before the study participation.

    Results

    Table 1 presents the summary of the patients recruited and the tumor images. A total of 124 bladder endoscopies were performed on 109 patients (97 men and 12 women) during the study period. Among the 431 images obtained, 1 (0.2%) showed a papilloma and a benign tumor. However, the remaining images showed urothelial carcinoma. By T classification, 15 (3.5%) images were T2 muscle-invasive bladder cancer, and the rest (96.3%) were Ta, Tis, and T1 NMIBC. The level of malignancy was of low and high grades at 45.3% and 54.5%, respectively. The annotation data of the lesion image showed an elevated lesion for 265 (61.5%) images, flat lesion for 76 (17.6%), and a mixture of raised and flat lesions for 90 (20.9%). The proportion of images with a tumor size occupying <10%, 10%–50%, and >50% of the overall image was 10.2%, 56.9%, and 32.9%, respectively.

    Table 1. Patient (n = 109) and Tumor Characteristics in the Image Set (Four Hundred Thirty-One Images)

     n%
    Male9790.0
    Female1210.0
    Age, years
     Median (interquartile range)74 (64.5–77) 
    Tumor form  
     Elevated lesion26561.5
     Flat lesion7617.6
     Mixed lesion9020.9
    Pathology analysis of tumor  
     Benign, papilloma10.2
     Urothelial carcinoma43099.8
    TNM stage  
     Ta32976.5
     T1388.8
     Ta +1163.7
     Tis214.9
     Ta + is112.6
     T2153.5
    Gradea
     Low grade19545.3
     High grade23554.7
    Tumor size
     Proportion of the overall image occupied by the lesion
      >10%4410.2
      10%–50%24556.9
      >50%14232.9

    a1973 World Health Organization classification.

    CIS = carcinoma in situ; TNM = tumor–node–metastasis.

    In the ROC curve (Fig. 3), AUC was 0.98, maximum YI was 0.837, sensitivity was 89.7%, and specificity was 94.0%. In the test data (tumor image: 87, normal image: 335), 78 images were true positive, 315 true negative, 20 false positive, and 9 were false negative. Of the nine images that were false negative, six were early discovered raised lesions that included Ta tumors. In addition, two images showed flat lesions of Ta tumors, and one image was a T1 tumor for which annotation of a mixed raised and flat lesion was performed. Of the nine images, eight images showed a small lesion that occupied <10% of the overall image. Figure 4a shows the ROC curve according to the proportion of lesions in the image. In <10%, 10%–50%, and >50% of the images, AUC and YI were 0.88 and 0.62, 0.88 and 0.90, and 0.88 and 0.92, respectively. Figure 4b shows the ROC curve according to the T stage of tumor. Ta, T1, Ta +1, Tis, Ta + is, and T2, AUC and YI were 0.98 and 0.84, 0.98 and 0.86, 1.00 and 1.00, 0.98 and 0.96, 0.99 and 0.98, and 1.00 and 1.00, respectively. Figure 4c shows the ROC curve according to the form of tumor. Elevated, flat, and mixed AUC and YI were 0.98 and 0.85, 0.96 and 0.87, and 0.99 and 0.92, respectively.

    FIG. 3.

    FIG. 3. MLP ROC curve. MLP = multilayer perceptron; ROC = receiver operating characteristic.

    FIG. 4.

    FIG. 4. (a) MLP ROC curve based on the proportion of lesions in the image. (b) MLP ROC curve based on the T stage of tumor. (c) MLP ROC curve based on the form of tumor.

    Discussion

    Regarding the mechanism of intravesical recurrence of NMIBC, dissemination of tumor cells, expansion of precancerous lesions, and overlooking of micro-disseminated lesions (daughter tumors) are considered rather than the de novo development. Therefore, it is of utmost importance to ensure the resection of disseminated lesions (daughter tumors), carcinoma in situ, and expansion of precancerous lesions that are difficult to observe without overlooking during TUR-BT, in addition to the main tumor. To improve the quality of bladder cancer diagnosis, we have proposed an objective evaluation method of cystoscopic imaging using AI, which allowed us to discriminate normal and tumor images with high accuracy.

    AI has recently been applied in the medical field, such as in the diagnosis of pulmonary nodules on computed tomography,20 diagnosis of skin cancer,21 classification of retinal lesions based on fundus photographs,22 endoscopic imaging diagnosis, detection of gastric cancer,13 evaluation of Helicobacter pylori infection of the stomach,14 classification of colon polyps,15 and automatic classification of regions in the upper gastrointestinal endoscopic images.23 Diagnosis of prostate cancer lesions on prostate magnetic resonance imaging has been attempted in the field of urology,24 but only one study from the urologic endoscopy field has been reported.25

    In this study, using white-light endoscopic images, favorable results of diagnosis were achieved for tumors with high sensitivity and specificity. It was clarified that through learning of ∼1700 cystoscopic images, normal and tumor lesions can be classified by distinguishing using transfer learning based on CNN. In general, learning of AI requires millions of large amount and high-quality data, but it was possible to detect tumor lesions in cystoscopic images even when learning as few as several thousand images through transfer learning, whereby the features of cystoscopic images are extracted from a deep learning model constructed using 1.2 million general images. It is suggested that transfer learning adopting a deep learning model to extract features facilitates an adjunctive diagnosis by machine learning and increases the quality of diagnosis and treatment of bladder cancer although a large volume of imaging data for learning cannot be prepared, such as the collection of medical images at single institutions. Transfer learning used in this study is similar to the fact that humans recognize normal and abnormal states by learning, that is, we observe various general images as we grow from birth, and after becoming a medical student and then physician, we learn endoscopic images through training, thereby acquiring the skills to make a diagnosis.

    We do not consider that this achievement replaces diagnoses made by urologists. If an automatic visual detection method for cystoscopic images can be established using this method, it may support doctors to make a diagnosis based on not only still images but also videos and real-time cystoscopy, which leads to accurate identification of the range of the tumor and reduction of the diagnostic miss rate, increasing the quality of medical care.

    The accuracy of annotating images of lesions, based on learned data, is a limitation of this study. At present, one urologist prepared all annotation data, so that the diagnoses of lesions are consistent, but the possibility of overlooking a lesion cannot be ruled out. In addition, the method on how the test data are classified as tumor images or normal images by multiple doctors has not been verified. Moreover, annotation data, such as inflammation-induced changes in the bladder mucosa, were not learned because these were not prepared. The difficulty of learning may be increased by the turbidity of urine, being out of focus, and whether the image is acquired from a near or distant site.

    If a tumor is reflected in a cystoscopic image, abnormalities can be identified. When doctors recognize a tumor, they examine it closely. However, if they do not recognize the tumor, they would not examine it closely and may therefore overlook it. These issues may be solved by diagnostic support using AI. In addition, for tumor lesions that are likely to be overlooked by WLI, learning of NBI, PDD, and high-quality endoscopic images such as 4K may reduce these and lead to a higher diagnostic accuracy.

    At present, this study has resulted in a low diagnostic accuracy for small lesions that occupied <10% of the overall image. This may be because of insufficient learning by the AI system as only ∼10% of the images used contained small lesions. These small lesions are also likely to be missed by urologists. In general, the performance of an AI system is dependent on the amount and quality of data used for learning. Although the AI system was able to extract and objectively assess features in cystoscopy images, new algorithms and additional images representing a variety of lesions are needed to improve its accuracy.

    Conclusion

    This study proposed a support system for the cystoscopic diagnosis of bladder cancer based on AI. The proposed AI system, as an initial support system of cystoscopic diagnosis, was based on a pretrained CNN model with ImageNet and was fine-tuned to learn images from cystoscopy performed 1637 times additionally in supervised learning. We demonstrated that our AI system is capable of classifying tumor lesions and normality with high accuracy for 431 test images from 109 patients. For clinical application, the proposed AI system would be further verified when in clinical use, so as to further develop it as a solution for diagnosis support. The objective evaluation of cystoscopic images using AI is expected to contribute to improvement in the accuracy of the diagnosis and treatment of bladder cancer.

    Ethical Standards

    All procedures followed were in accordance with the ethical standards of the responsible committee on human experimentation (institutional and national) and with the Declaration of Helsinki of 1964 and later versions. Informed consent or substitute for it was obtained from all patients for being included in the study.

    Author Disclosure Statement

    No competing financial interests exist.

    Funding Information

    This study was supported by Japan Society for the Promotion of Science KAKENHI Grant Number from JP16H02903 and JP17K16775.

    References

    • 1. Höglund M. On the origin of syn- and metachronous urothelial carcinomas. Eur Urol 2007;51:1185–1193. Crossref, MedlineGoogle Scholar
    • 2. Herr HW, Shipley WU, Bajorin DF. Cancer of the bladder. In: Devita VTHellman SRosenberg SA, eds. Cancer: Principles and Practice of Oncology. Sixth Edition. Philadelphia, PA: Lippincott Williams and Wilkins, 2001, pp. 1005–1020. Google Scholar
    • 3. Brausi M, Collette L, Kurth K, et al. Variability in the recurrence rate at first follow-up cystoscopy after TUR in stage Ta T1 transitional cell carcinoma of the bladder: A combined analysis of seven EORTC studies. Eur Urol 2002;41:523–531. Crossref, MedlineGoogle Scholar
    • 4. Rolevich A, Minich A, Nabebina T, et al. Surgeon has a major impact on long-term recurrence risk in patients with non-muscle invasive bladder cancer. Cent Eur J Urol 2016;69:170–177. MedlineGoogle Scholar
    • 5. Kausch I, Sommerauer M, Montorsi F, et al. Photodynamic diagnosis in non-muscle-invasive bladder cancer: A systematic review and cumulative analysis of prospective studies. Eur Urol 2010;57:595–606. Crossref, MedlineGoogle Scholar
    • 6. Burger M, Grossman HB, Droller M, et al. Photodynamic diagnosis of non-muscle-invasive bladder cancer with hexaminolevulinate cystoscopy: A meta-analysis of detection and recurrence based on raw data. Eur Urol 2013;64:846–854. Crossref, MedlineGoogle Scholar
    • 7. Jocham D, Witjes F, Wagner S, et al. Improved detection and treatment of bladder cancer using hexaminolevulinate imaging: A prospective, phase III multicenter study. J Urol 2005;174:862–866. Crossref, MedlineGoogle Scholar
    • 8. Lee JY, Cho KS, Kang DH, et al. A network meta-analysis of therapeutic outcomes after new image technology-assisted transurethral resection for non-muscle invasive bladder cancer: 5-aminolaevulinic acid fluorescence vs hexylaminolevulinate fluorescence vs narrow band imaging. BMC Cancer 2015;15:566. Crossref, MedlineGoogle Scholar
    • 9. Naya Y, Oishi M, Yamada Y, et al. Initial experience of combined use of photodynamic diagnosis and narrow band imaging for detection of flat urothelial lesion. Int J Clin Oncol 2015;20:593–597. Crossref, MedlineGoogle Scholar
    • 10. Kang W, Cui Z, Chen Q, et al. Narrow band imaging-assisted transurethral resection reduces the recurrence risk of non-muscle invasive bladder cancer: A systematic review and meta-analysis. Oncotarget 2017;8:23880–23890. Crossref, MedlineGoogle Scholar
    • 11. Xiong Y, Li J, Ma S, et al. A meta-analysis of narrow band imaging for the diagnosis and therapeutic outcome of non-muscle invasive bladder cancer. PLoS One 2017;12:e0170819. Crossref, MedlineGoogle Scholar
    • 12. Drejer D, Béji S, Oezeke R, et al. Comparison of white light, photodynamic diagnosis, and narrow-band imaging in detection of carcinoma in situ or flat dysplasia at transurethral resection of the bladder: The DaBlaCa-8 study. Urology 2017;102:138–142. Crossref, MedlineGoogle Scholar
    • 13. Hirasawa T, Aoyama K, Tanimoto T, et al. Application of artificial intelligence using a convolutional neural network for detecting gastric cancer in endoscopic images. Gastric Cancer 2018;21:653–660. Crossref, MedlineGoogle Scholar
    • 14. Nosato H, Sakanashi H, Takahashi E, et al. Image retrieval method for multiscale objects from optical colonoscopy images. Int J Biomed Imaging 2017;2017:7089213. Crossref, MedlineGoogle Scholar
    • 15. Shichijo S, Nomura S, Aoyama K, et al. Application of convolutional neural networks in the diagnosis of Helicobacter pylori infection based on endoscopic images. EBioMedicine 2017;25:106–111. Crossref, MedlineGoogle Scholar
    • 16. Komeda Y, Handa H, Watanabe T, et al. Computer-aided diagnosis based on convolutional neural network system for colorectal polyp classification: Preliminary experience. Oncology 2017;93:30–34. Crossref, MedlineGoogle Scholar
    • 17. Szegedy V, Liu W, Jia Y, et al. Going deeper with convolutions. 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Boston, MA, 2015, pp.1–9. Google Scholar
    • 18. ImageNet. http://www.image-net.org (Accessed October 1, 2016). Google Scholar
    • 19. Kingma DP, Ba J. Adam: A Method for Stochastic Optimization. CoRR 2014. abs/1412.6980. Google Scholar
    • 20. Kang G, Liu K, Hou B, et al. 3D multi-view convolutional neural networks for lung nodule classification. PLoS One 2017;12:e0188290. Crossref, MedlineGoogle Scholar
    • 21. Esteva A, Kuprel B, Novoa RA, et al. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017;542:115–118. Crossref, MedlineGoogle Scholar
    • 22. Choi JY, Yoo TK, Seo JG, et al. Multi-categorical deep learning neural network to classify retinal images: A pilot study employing small database. PLoS One 2017;12:e0187336. Crossref, MedlineGoogle Scholar
    • 23. Takiyama H, Ozawa T, Ishihara S, et al. Automatic anatomical classification of esophagogastroduodenoscopy images using deep convolutional neural networks. Sci Rep 2018;8:7497. Crossref, MedlineGoogle Scholar
    • 24. Le MH, Chen J, Wang L, et al. Automated diagnosis of prostate cancer in multi-parametric MRI based on multimodal convolutional neural networks. Phys Med Biol 2017;62:6497–6514. Crossref, MedlineGoogle Scholar
    • 25. Shkolyar E, Jia X, Xing L, et al. Automated cystoscopic detection of bladder cancer using deep-learning. J Urol 2019;201 [Epub ahead of print]; DOI: 10.1097/01.JU.0000557512.76700.42. CrossrefGoogle Scholar
    Abbreviations Used
    AI

    artificial intelligence

    AUC

    area under the curve

    CIS

    carcinoma in situ

    CNN

    convolutional neural network

    DCNN

    deep convolutional neural network

    MLP

    multilayer perceptron

    NBI

    narrow band imaging

    NMIBC

    nonmuscle-invasive bladder cancer

    PDD

    photodynamic diagnosis

    ROC

    receiver operating characteristic

    TNM

    tumor–node–metastasis

    WLI

    white-light imaging

    YI

    Youden index

    Back to Top