Review |
© 2024 Jasmin Gurung, Mladen Doykov, Gancho Kostov, Bozhidar Hristov, Peter Uchikov, Krasimir Kraev, Daniel Doykov, Katya Doykova, Siyana Valova, Emiliya Nacheva-Georgieva, Eduard Tilkiyan.
This is an open access article distributed under the terms of the Creative Commons Attribution License (CC BY 4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Citation:
Gurung J, Doykov M, Kostov G, Hristov B, Uchikov P, Kraev K, Doykov D, Doykova K, Valova S, Nacheva-Georgieva E, Tilkiyan E (2024) The expanding role of artificial intelligence in the histopathological diagnosis in urological oncology: a literature review. Folia Medica 66(3): 303-311. https://doi.org/10.3897/folmed.66.e124998
|
The ongoing growth of artificial intelligence (AI) involves virtually every aspect of oncologic care in medicine. Although AI is in its infancy, it has shown great promise in the diagnosis of oncologic urological conditions. This paper aims to explore the expanding role of artificial intelligence in the histopathological diagnosis in urological oncology. We conducted a focused review of the literature on AI in urological oncology, searching PubMed and Google Scholar for recent advancements in histopathological diagnosis using AI. Various keyword combinations were used to find relevant sources published before April 2nd, 2024. We approached this article by focusing on the impact of AI on common urological malignancies by incorporating the use of different AI algorithms. We targeted the capabilities of AI’s potential in aiding urologists and pathologists in histological cancer diagnosis. Promising results suggest AI can enhance diagnosis and personalized patient care, yet further refinements are needed before widespread hospital adoption. AI is transforming urological oncology by improving histopathological diagnosis and patient care. This review highlights AI’s advancements in diagnosing prostate, renal cell, and bladder cancer. It is anticipated that as AI becomes more integrated into clinical practice, it will have a greater influence on diagnosis and improve patient outcomes.
AI, histopathological diagnosis with AI, urological oncology, urology
The ongoing growth of artificial intelligence (AI) involves virtually every aspect of oncologic care in medicine. AI is a computer’s ability to replicate the cognitive functions of the human mind at a higher level, continuously improving its performance over time. This is done through the perception of external stimuli and the determination of an effective strategy for a desired outcome.[
This paper aims to explore the expanding role of artificial intelligence in the histopathological diagnosis of urological oncology. By analyzing the complexities of AI, we will evaluate the strengths and weaknesses of different divisions of tumor oncology.
We conducted a focused, non-systematic review of published literature concerning the use of AI in urological oncology. The PubMed and Google Scholar databases were searched to gain insights into recent advancements in histopathological diagnosis using AI and their impact on urological oncology. We analyzed the existing diagnostic methods and how AI algorithms have enhanced them, facilitating more efficient diagnosis for pathologists and urologists, especially with specimens that are harder to diagnose requiring extensive training and expertise. Various combinations of keywords, such as urological oncology, AI, urology, and histopathological diagnosis with AI, were used to narrow down and identify relevant sources. Only materials published before April 2nd, 2024, were considered for inclusion. Relevant articles were selected, and information related to the keywords was extracted for analysis.
Understanding the fundamentals of AI is important to evaluate its implementation. Artificial Intelligence (AI) is a field of computer science that uses machine learning techniques to make predictions.
Machine learning (ML)-based approaches entail machines “learning” from the data provided to them to make predictions. Deep learning (DL) is a specific ML approach that has evolved through the advancement of artificial neural networks. DL involves multiple layers of artificial neural networks consisting of an input layer, an output layer, and multiple hidden layers. The algorithm is then used for different tasks, such as classification, segmentation, and detection.[
Between 1983 and 2002, the prevalence of renal cancer increased from 7.1 to 10.8 cases per 100,000 patients. The diagnosis was made incidentally when imaging studies were conducted for unrelated clinical reasons. Despite the progress in diagnosis, patient outcomes have not improved, with a reported increase in mortality from 1.5 to 6.5 deaths per 100,000 patients within the same time interval.[
Renal cell carcinoma (RCC) is the sixth most diagnosed cancer in men and the tenth in women among all cancers. It possesses extensive histopathological features, creating challenges for accurate diagnosis and prognosis.[
In some cases, interobserver variability remains a barrier to RMB interpretation.[
In the era of the “Big Data” generation, the use of electronic health records, digitalized radiology, and virtual pathology has created an abundant source of digital data that is applicable to the data-characterization algorithms of AI.[
Oncocytoma is the most frequent benign renal mass found. Pathologists encounter difficulties when differentiating oncocytoma from chromophobe RCC.[
With the help of two pre-trained convolutional neural networks (CNN), Tabibu et al. differentiated between clear cell and chromophobe RCC with a classification accuracy of 93.39% and 87.34%, respectively. Furthermore, they amplified the deep network with a directed acyclic graph support vector machine (DAG-VSM) for subtype classification. This functionality not only enhanced the model’s performance but also addressed data imbalances inherent in multi-class classification tasks involving histopathological images. High-probability cancerous regions and tissues from different origins were effectively characterized.[
RCC classification is a challenging task because of the complexity of the method. The periodic updates to the classification system additionally complicate the process. Clear cell papillary RCC is a relatively new subtype that has only been recognized as of 2016.[
Some researchers have used neural networks initially developed to diagnose different tumors to enhance diagnosis. Faust et al. utilized a CNN primarily programmed to recognize the histomorphology of brain tumors on 550 digital images of 396 clear cell RCC and 154 papillary RCC. Five hundred and twelve different features were extracted from more than 550 digital WSIs of RCC to perform image set-clustering, therefore allowing the algorithm to evaluate clinical and biological differences between the patient subgroups. By employing this methodology, researchers were able to recognize different subtypes of RCC and predict survival rates within each group. The algorithm could also identify specific patterns and features specific to kidney cancer. The results of this study demonstrate that CNNs pretrained on large histologic datasets can learn new pathologies with or without unsupervised applications.[
Bladder cancer (BCa) is the most prevalent malignancy of the urinary tract. In 2018, there were 549,393 new cases, of which 199,922 resulted in death.[
Shkolyar et al. developed CystoNet, which is a CNN-based model for automated detection of bladder tumors. The model was established upon a dataset of 2335 frames of normal bladder mucosa and 417 frames of histologically confirmed papillary urothelial carcinoma obtained from 95 patients. The effectiveness of CystoNet was assessed on 54 more patients, showing a sensitivity of 90.9% and a specificity of 98.6% per frame. The model was successful in detecting 39 out of 41 papillary and three of three flat bladder cancers, demonstrating that it could assist in training, and diagnostic decision-making while ensuring consistent performance among providers enhancing the accessibility and quality of cystoscopy.[
Urologists are trained to identify various cancers during cystoscopy; nonetheless, the diagnosis relies on their expertise. It is particularly difficult to recognize tumors with a small diameter and flat tumors like carcinoma in situ. A model based on GoogLeNet (Szegedy et al.) was utilized by Ikeda et al. to mitigate human error and increase the quality of the diagnosis. The dataset consisted of a total of 2102 cystoscopic images, 1671 images of normal tissue, and 431 images of tumor lesions. To further expand the data, the authors enlarged the training dataset by rotating and blurring the cystoscopy images. They also used transfer learning to train their model. The original GoogLeNet model, which consisted of 22 layers and over seven million parameters, could readily fit the small dataset utilized in this study. A final model evaluation of a cystoscopic image with AI assistance showed a sensitivity of 89.7% and a specificity of 94.00%. As a whole, the use of AI in cystoscopy image evaluation has a lot of potential.[
Lymph node staging is essential in the diagnosis and treatment of patients with bladder cancer. Wu et al. produced a diagnostic model based on whole-slide images of lymph node metastases. This retrospective, multicenter, and diagnostic study examined the effectiveness in different groups: five validation sets, a single lymph node test set, a multi-cancer test set, and a subset for comparing performance between the model and the pathologists. The dataset included 8177 images and 20,954 lymph nodes obtained from 1012 patients with bladder cancer and radical cystectomy, along with lymph node dissection. After the exclusion of 14 patients with non-bladder cancer and 21 low-quality images, the samples included 998 patients and 7991 images. The AUC for accuracy ranged from 0.978 to 0.998 in the five validation sets. The performance of the model came to 0.983 and 0.906 for the junior pathologist in comparison with 0.947 for the senior pathologist. The AI assistance improved the sensitivity of both the junior and senior pathologists, from 0.906 to 0.953 and from 0.947 to 0.986, respectively. In the multi-cancer test, the model maintained an AUC of 0.943 in breast cancer and 0.922 in prostate cancer images. Micro metastases missed by the pathologists, leading to negative results, were detected by the model. The receiver operating characteristic curve demonstrated that the model would allow pathologists to exclude 80% to 92% of the negative slides while maintaining 100% sensitivity. The model holds significant potential for clinical applications in detecting lymph node metastases, especially micrometastases, enhancing the precision and effectiveness of pathologists’ work.[
Due to the invasive nature of cystoscopy, patients experience discomfort during the procedure, which poses a financial burden to the healthcare system. Research into the implications of AI for efficient and prompt diagnosis through urine is underway.[
To address the absence of readily available computed screening systems for urine specimens and the issue of poor interobserver agreement, Sanghvi et al. devised a computational pipeline for analyzing digitized liquid-based urine cytology slides. One thousand six hundred and fifteen whole slide images (WSIs) were used for training, and 790 WSIs were used for testing their pipeline, encompassing multiple tiers of CNN models. This algorithm yielded an AUC of 0.88.[
The implication of artificial intelligence using the VisioCyt test in the diagnosis of bladder carcinoma was reported by Lebret et al. A large multicenter trial was executed in 14 centers on 1360 patients, divided into two groups: 1) patients with bladder carcinoma of varied histological grades and stages; and 2) a control group of patients with negative cystoscopy. The process included algorithm development and algorithm validation. This was followed by a comparison of the VisioCyt results and those obtained from experienced uropathologists. VisioCyt showed a sensitivity of 84.9% and a cytology of 43%. Regarding high-grade tumors, VisioCyt’s sensitivity was 92.6%, and that of uropathologists was 61.1%. For low-grade tumors, VisioCyt showed a sensitivity of 77%, and that of uropathologists was 26.3%. If confirmed by a validation cohort, the algorithm can be beneficial for pathologists.[
A study performed by Sokolov et al. explored the use of nano-scale-resolution images of urine cells by atomic force microscopy (AFM). The evaluation of urine samples from 43 patients without evidence of bladder cancer and 25 cancer patients with confirmed bladder cancer displayed a diagnostic accuracy of 0.94 when studying five cells per urine sample. A significant improvement in diagnostic accuracy was achieved in comparison with the standard cystoscopy, as validated by the assessment of 43 control and 25 bladder cancer patients.[
With the epidemiological range for prostate cancer patients ranging from 45 to 60, it is associated with the highest cancer-associated mortality rate in Western countries.[
Prostate cancer screening with prostate-specific antigen (PSA) has substantially reduced mortality from prostate cancer (>50%).[
Diagnosis of prostate cancer is done by microscopic assessment of tissue samples; however, limitations remain in the presence of insufficiently trained pathologists. Some researchers have developed an augmented reality microscope (ARM). The deep learning system presents real-time feedback to pathologists as they are viewing the histopathology slides. In relation to prostate cancer detection, the algorithm gained an AUC of 0.93 at the ×10 objective and an AUC of 0.99 at the ×20 objective. The integration of ARM can be seamlessly incorporated into the existing microscopy workflow, thereby eliminating the requirement for expensive IT infrastructure and whole slide image scanners. Subjective tasks like stain quantification can be overcome, especially in places with limited access to highly experienced pathologists. Early experience has been successful with promising results, indicating its utility in clinical medicine with continued development and data set incorporation.[
Due to the similar challenges encountered, Ström et al. created an AI system to assist urological pathologists in detection, localization, and Gleason grading. Digitization of 6682 slides from needle core biopsy, mainly from the STHLM3 population-based prostate cancer screening trial. The obtained images were used in the training of deep neural networks for the evaluation of the prostate biopsy samples. The system received an AUC of 0.997 for differentiating between benign (n=10) and malignant (n=721) tumors. In accordance with Gleason’s grading, the AI system displayed a mean pairwise Cohen’s kappa of 0.62, which fell within the range of corresponding values (0.60–0.73) obtained by 23 experienced urological pathologists. By using this algorithm, accurate grading of prostate cancer along with a second opinion from a pathologist can be obtained.[
The Gleason grading of the biopsy specimen holds significance within case management. Nagpal et al. constructed a deep learning system to assist in the grading of the specimen into 5 categories: nontumor, GG1, GG2, GG3, GG4, or GG4-5. Seven hundred and fifty-two deidentified digitized images from prostate needle biopsy specimens were acquired from three institutes in the United States. By allowing immunohistochemical access to the histologic sections for the biopsy samples, specialists were able to minimize levels of uncertainty in the diagnosis. The rate of agreement with subspecialists was higher at 71.7% than the 19 general pathologists (58%, p<0.001). Differentiation between tumor and non-tumor, the rate of agreement with the subspecialist was 94.3%, and the general pathologist had a similar result of 94.7%. This study demonstrates that support from AI is beneficial in enhancing expertise at Gleason grading of prostate needle biopsy specimens.[
Another computer-aided diagnostic system for automatic grading of prostate cancer by multiple pathologists can be visualized in this study demonstrated by Nir et al. The workflow was composed of multi-scale features: glandular, cellular, and image-based. The classifier was trained on 333 tissue microarray (TMA) samples, each graded by six pathologists. Each TMA slide consisted of 160 cores (10×16 grid) from approximately 80 different radical prostatectomy samples. For ease of annotation by multiple pathologists of individual core images, they developed an Android-based application known as “PathMarker”. This contributed to the establishment of an extensive and complex database with multiple labels. To evaluate the performance of the automatic grading system, a cross-validation study comprising a dataset of 231 patients was conducted. This received a kappa value of 0.51 between the system and pathologist, which falls within the spectrum of previously reported ranges of inter-pathologist agreements (0.45–0.62). This finding reflects positively on the classifier’s performance as it validates its efficacy, indicating potential employment in clinical practice.[
The use of AI is rapidly evolving; however, there are some roadblocks impacting the evolving involvement of these programs in clinical practice. Some studies have based their research on a retrospective design, resulting in occasional data loss and bias. Alongside that, small sample sizes hinder reproducibility and clinical applicability due to the algorithm being taught on small datasets, preventing its widespread use for patients from various epidemiological backgrounds. This limitation is reinforced by the use of an internal dataset, which further limits cross-validation and potentially leads to bias.[
Further validation of the images by multiple doctors was not conducted. Annotating the images of inflammatory changes in the bladder was not prepared, causing the unavailability of the data for the algorithm. Not providing the algorithm certain images indicates either overperformance in the training set or underperformance on the new, unseen data. This may reduce effectiveness in real-world application as it may struggle to generalize new data and the predictions may be inaccurate due to the lack of comprehensive data.[
This table has been designed to summarize some of the articles to clearly display the key findings of some studies
Study | Date of publication | Objective | Algorithm/Method | Study Design | Results |
Erdim et al.[ |
January 2020 | Detection of benign and malignant renal solid mass | Machine learning algorithm | 79 patients with 84 solid renal masses (21 benign; 63 malignant) | AUC of 0.91 for CT texture analysis in differentiating benign and malignant renal solid mass |
Fenstermaker et al.[ |
July 2020 | Identify and differentiate renal cell carcinoma (RCC) histology subtypes and grades based on histological specimen | Convolutional neural network | 3000 normal samples 12,168 RCC tissue samples from 42 patients |
Accuracy of 97.5% in differentiating subtypes and Fuhrman grade accuracy of 94.8% |
Zhu et al.[ |
March 2021 | Classify digitized surgical resection and biopsy slides | Deep neural network | 78 surgical resection whole slides and 79 biopsy slides 917 surgical resection slides from the Cancer Genome Atlas database |
Average AUC of 0.98 on classifying subtypes |
Tabibu et al.[ |
July 2019 | Classification of RCC subtypes | Convolutional neural networks | 1027 clear cell RCC, 303 papillary RCC, 254 chromophobe RCC tumor slide images | Accuracy of 93.39% to distinguish clear cell and 87.34% for chromophobe from normal tissue. |
Abdeltawab et al.[ |
April 2022 | Distinction between clear cell RCC and papillary RCC | Convolutional neural network based computerized-aided diagnostic system | 64 image slides of 4 categories: fat, parenchyma, clear cell RCC and clear cell papillary RCC | Accuracy of 91% using cases from institution files and 90% in diagnosis of clear cell RCC using external dataset |
Shkolyar et al.[ |
September 2019 | Automated detection of bladder tumor | Convolutional neural network augmented cystoscopy | 95 patients for algorithm training and 5 patients for testing Diagnostic performance validated on 54 patients |
Per-frame sensitivity of 90.9% Specificity of 98.6% Per-tumor sensitivity of 90.9% for detection of bladder tumors |
Ikeda et al.[ |
March 2020 | Evaluate normal and tumor images obtained by white-light cystoscopy | Convolutional neural network | 2102 cystoscopic images: 1671 images of normal tissue, 431 images of tumor lesions | Sensitivity of 89.7% and specificity of 94% for cystoscopic image evaluation |
Wu et al.[ |
March 2023 | Lymph node metastases diagnostic model on whole slide images | AI-based diagnostic model | 998 patients 7991 whole slide images |
AI assistance improved sensitivity of an AUC of 0.906 to 0.953 in junior pathologists and 0.947 to 0.986 in senior pathologists |
Sanghvi et al.[ |
August 2019 | Analyze digitized liquid-based urine cytology slides | Computational pipeline with multiple tiers of models | 1615 whole slide images Additional 790 cases for validation |
AUC of 0.88 for cytopathological diagnosis of high-grade urothelial carcinoma |
Lebret et al.[ |
March 2021 | Using VisioCyt test to improve diagnosis of bladder carcinoma using voided urine cytology | Machine and deep learning algorithm | 1360 patients divided in two groups followed by comparison of results with experienced uropathologists | VisioCyt sensitivity of 92.6% compared to 61.1% of uropathologists for high-grade tumors |
Chen et al.[ |
August 2019 | Cost-effective augmented reality microscope for diagnosis and staging of prostate cancer | Deep learning algorithm | 1360 specimen in the observer’s field of view from 34 prostate slides using ×10 and ×20 objectives | AUC of 0.93 for prostate cancer detection |
Ström et al.[ |
January 2020 | Prostate cancer detection, localization and Gleason grading | Deep neural network | Digitization of 6682 slides from needle core prostate biopsy | AUC of 0.997 for differentiating benign and malignant tumors Gleason grading of mean pairwise kappa of 0.62 |
Nagpal et al.[ |
September 2020 | Evaluate ability of deep learning system to grade diagnostic prostate biopsy specimens | Deep learning system | 752 deidentified digitized images from prostate needle biopsy specimens | Agreement with subspecialists higher with deep learning system at 71.7% than 58% with pathologists |
AI is revolutionizing urological oncology by enhancing histopathological diagnosis, optimizing patient care, and improving clinical decision-making. With the rise in technology, AI is becoming pivotal in precision medicine and is at the center of scientific research from a diagnostic and therapeutic perspective. In this review, we focused on highlighting the advancements of AI in a histopathological setting by evaluating the diagnostic role of AI for prostate, renal cell, and bladder cancer. While AI has demonstrated significant potential in urological oncology, further refinement is necessary to address the existing challenges. Despite its promise, there are hurdles that must be overcome to optimize effectiveness, such as larger patient cohort data, more refined training, and data analysis to ensure its utility in clinical practice with a wide variety of patients. As AI continues to evolve and integrate into clinical practice, its impact on diagnosis is expected to grow exponentially, ultimately leading to positive outcomes and improved quality of life for patients.
The authors have no support to report.
The authors have no funding to report.
The authors have declared that no competing interests exist.