The development and validation of pathological sections based U-Net deep learning segmentation model for the detection of esophageal mucosa and squamous cell neoplasm
Original Article

The development and validation of pathological sections based U-Net deep learning segmentation model for the detection of esophageal mucosa and squamous cell neoplasm

Feng Su1#, Wei Zhang2#, Yunzhong Liu3#, Shanglin Chen1#, Miao Lin1, Mingxiang Feng1, Jun Yin1, Lijie Tan1, Yaxing Shen1,4

1Department of Thoracic Surgery, Zhongshan Hospital, Fudan University, Shanghai, China; 2Department of Cardio and Thoracic Surgery, Hanzhong Central Hospital, Hanzhong, China; 3Department of Cardio and Thoracic Surgery, The First Affiliated Hospital of Hainan Medical University, Haikou, China; 4Department of Thoracic Surgery, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China

Contributions: (I) Conception and design: F Su, Y Shen; (II) Administrative support: Y Shen; (III) Provision of study materials or patients: S Chen, M Lin, J Yin, M Feng, L Tan, Y Shen; (IV) Collection and assembly of data: F Su, W Zhang, S Chen, M Lin, M Feng, Y Shen; (V) Data analysis and interpretation: F Su, W Zhang, Y Shen; (VI) Manuscript writing: All authors; (VII) Final approval of manuscript: All authors.

#These authors contributed equally to this work.

Correspondence to: Yaxing Shen, MD. Department of Thoracic Surgery, Zhongshan Hospital, Fudan University, No. 180 Fenglin Road, Xuhui District, Shanghai 200032, China; Department of Thoracic Surgery, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, China. Email: shen.yaxing@zs-hospital.sh.cn.

Background: Deep learning methods have demonstrated great potential for processing high-resolution images. The U-Net model, in particular, has shown proficiency in the segmentation of biomedical images. However, limited research has examined the application of deep learning to esophageal squamous cell carcinoma (ESCC) segmentation. Therefore, this study aimed to develop deep learning segmentation systems specifically for ESCC.

Methods: A Visual Geometry Group (VGG)-based U-Net neural network architecture was utilized to develop the segmentation models. A pathological image cohort of surgical specimens was used for model training and internal validation, with two additional endoscopic biopsy section cohort for external validation. Model efficacy was evaluated across several metrics including Intersection over Union (IOU), accuracy, positive predict value (PPV), true positive rate (TPR), specificity, dice similarity coefficient (DSC), area under the receiver operating characteristic curve (AUC), and F1-Score.

Results: Surgical samples from ten patients were analyzed retrospectively, with each biopsy section cohort encompassing five patients. Transfer learning models based on U-Net weights yielded optimal results. For mucosa segmentation, the in internal validation achieved 93.81% IOU, with other parameters exceeding 96% (96.96% accuracy, 96.45% PPV, 96.65% TPR, 98.41% specificity, 96.81% DSC, 96.11% AUC, and 96.55% F1-Score). The tumor segmentation model attained an IOU of 91.95%, along with other parameters surpassing 95% (95.90% accuracy, 95.62% PPV, 95.71% TPR, 97.88% specificity, 95.81% DSC, 94.92% AUC, and 95.67% F1-Score). In the external validation for tumor segmentation model, IOU was 59.86% for validation database 1 (72.74% for accuracy, 76.03% for PPV, 77.17% for TPR, 83.80% for specificity, 74.89% for DSC, 71.83% for AUC, and 76.60% for F1-Score), and 50.88% for validation cohort 2 (68.03% for accuracy, 59.02% for PPV, 66.87% for TPR, 78.48% for specificity, 67.44% for DSC, 64.68% for AUC, and 62.70% for F1-Score).

Conclusions: The models exhibited satisfactory results, paving the way for their potential deployment on standard computers and integration with other artificial intelligence models in clinical practice in the future. However, limited to the size of study, the generalizability of models is impaired in the external validation, larger pathological section cohort would be needed in future development to ensure robustness and generalization.

Keywords: Esophageal squamous cell carcinoma (ESCC); deep learning; pathological image; convolutional neural network (CNN); U-Net


Submitted Jul 16, 2023. Accepted for publication Sep 08, 2023. Published online Sep 29, 2023.

doi: 10.21037/jgo-23-587


Highlight box

Key findings

• This study developed U-Net based deep learning segmentation models for esophageal squamous cell carcinoma.

What is known and what is new?

• Innovations in digital pathology enable deep learning applications for pathological images, improving processing efficacy. The pathological image classification, identification and segmentation models are well established for cancers of breast, lung and skin.

• In this study, a Visual Geometry Group-based U-Net neural network was utilized to develop mucosa and tumor segmentation models. Meanwhile, tumor segmentation was additionally validated in two external datasets.

What is the implication, and what should change now?

• The models demonstrated promising mucosa and tumor segmentation capabilities. With integration of other artificial intelligence models, clinical deployment on common computers is expected. Further extensive researches into deep learning for esophageal cancer are warrant to assist in fast and accurate diagnosis.


Introduction

With a 5-year survival rate of 20%, esophageal carcinoma stands as a significant contributor to cancer-related fatalities worldwide (1). Notably, China accounts for half of the annual reported cases globally (2). The results of pathology examinations of gastroscopy biopsies and surgical specimens are the gold standard for diagnosing and staging this malignancy. It is imperative to identify tumor cells and determine the depth of wall penetration during the scrutiny of pathological sections to attain precise diagnoses (3).

The advent of digital pathology techniques, which involve the digitization of microscopic histologic slides including scanning, compression, storage, and digital display, makes the conversion of microscopic histologic slides into high resolution images possible (4,5). As an advanced technique, the acquisition of whole slide image (WSI) has become a routine practice in cancer diagnosis, which facilitates the storage of entire pathological slides in high definition and the consultations for challenging cases (6). Typically, a high-resolution WSI comprises thousands of cells and trillions of pixels to ensure intricate cellular structures identifiable. However, when dealing with lesions or regions of interest (ROIs) that are relatively small compared to the entire sections, manually scanning of whole slide under microscopy or with WSI is a time-consuming task, even for experienced pathologists (7). The available pool of expert pathologists remains insufficient to meet the substantial clinical demand for manual pathological examination. Fortunately, WSIs enable the application of artificial intelligence for pathological diagnosis, offering a potential solution to the aforementioned challenge through automated image processing (8,9).

Computer-assisted diagnosis techniques could potentially recognize informative patterns in images that transcend human limitations and boundaries. These techniques have already been implemented in medical imaging, including endoscopic imaging, computed tomography (CT) and positron emission tomography (PET) (10,11). Artificial intelligence could also assist pathologists in identifying distinctive features in WSI, with the goal of early detection and prognosis prediction. The digital pathological images deep learning technologies have yielded remarkable capabilities in detection, segmentation, and classification for carcinomas in locations such the breast, lung and skin (12-14).

The convolutional neural networks (CNNs) have already existed for decades, large biomedical datasets containing thousands of training images are still required for the generation of complex outputs. The U-Net network, initially proposed by Ronneberger et al. in 2015, has shown precise segmentations even with very few training images for biomedical domain (15). The network primarily consists of a contracting path and an expansive path, both of which have been validated for their effectiveness in biomedical image processing and segmentation across various diseases (16,17). For esophageal cancer, previous studies have demonstrated that CNN networks could achieve comparable accuracy as clinicians, especially with computed tomography-based or endoscopic datasets (18,19). However, few studies focused on the artificial intelligence based pathological images processing in esophageal carcinoma. As the only gold standard for diagnosing esophageal cancer, and determining the depth of invasion (20), this study aimed to develop deep U-Net architecture-based learning models for the segmentation of esophageal mucosa and esophageal squamous cell carcinoma (ESCC). We present this article in accordance with the TRIPOD reporting checklist (available at https://jgo.amegroups.com/article/view/10.21037/jgo-23-587/rc).


Methods

Datasets

Pathological images of surgical specimen from patients with ESCC who underwent esophagectomy at the Department of Thoracic Surgery, Zhongshan Hospital, Fudan University between January 2021 and June 2021 were retrospectively reviewed as the primary dataset of the study. The inclusion criteria were: (I) age between 18 to 80 years; (II) histological diagnosis of ESCC; (III) underwent radical esophagectomy with available surgical specimens. And the exclusion criteria included: (I) history of other tumors; (II) unavailable or unqualified surgical specimens. The study was conducted in accordance with the Declaration of Helsinki (as revised in 2013). The study was approved by ethics review board of Zhongshan Hospital, Fudan University (No. ZS-B2021-708) and individual consent for this retrospective analysis was waived. WSIs were acquired from hematoxylin and eosin (H&E) staining of paraffin-embedded tissues, and the slides would be re-imaged if distortion, blurs, or occlusions were detected on the quality assessment. WSIs of tumor areas were allocated into tumor cohort, while WSIs of paracancerous area and normal tissue comprised mucosa cohort. Small patches were then generated based on a sliding-window sampling strategy, with patches of less than 10% ROIs eliminated. Patches in mucosa and tumor cohort were utilized for the construction of segmentation model for esophageal mucosa and ESCC, respectively. After cohort generation, patches were allocated into the training and validation sets randomly with a 9:1 split ratio. Specifically, attention was focused on the effectiveness of the tumor identification model. And therefore, two additional WSI datasets from endoscopic biopsy slides outside the Department of Pathology, Zhongshan Hospital, Fudan University, were allocated as external validation cohorts, encompassing samples from five ESCC patients each. Patches in the external validation cohorts were acquired through random window strategy sampling from WSIs.

Basic network architecture

Input patches were resized to of 512×512 pixels with matching resolution for segmentation model outputs. The basic network was U-Net structure, a convolutional network advantageous for medical image segmentation (15). It is composed of a contracting path for feature extraction and an expansive component for segmentation prediction as well (21). The Visual Geometry Group (VGG) Network was applied as the backbone for the contracting path, which consisted of repeated application of convolutions, followed by a rectified linear unit and a max pooling operation (22). And the prediction step consisted of an upsampling followed by convolutions. Each layer of the expansive component was concatenated and fused with the corresponding layer in the contracting path via skip-connection to prevent edge information loss. Besides training from scratch, pre-trained Xception, U-Net, VGG16, VGG19 and ResNet50 models were used for the transfer learning (23-25). Training epoch weights and performance were recorded to identify contributing features. The network architecture is available in Figure 1, with main hyperparameters presented in the Table 1. The deep learning process was performed on a laptop equipped with GeForce RTX3050Ti (NVIDIA) graphics processing unit and CUDA 11.0.2 for graphics processing unit (GPU) acceleration.

Figure 1 The overview of the U-Net neural network architecture with input image matrix size of 512×512 pixels.

Table 1

List of hyperparameters submitted in the framework

Hyperparameters Values
Patch size, pixels 512×512
Batch size 2
Learning rate 1E−4
Optimizer type Adam
Momentum 0.9
Epoch number 100

ROIs

ROIs were manually annotated by two authors with Labelme under the assistance of Department of Pathology as the gold diagnoses of mucosa and cancer, and the generated JSON files were converted to mask map and trained on the Pytorch platform. Two ROI sets were labeled—mucosa/non-mucosa and tumor/non-tumor. Reproducibility was assessed in a blinded manner with disagreements resolved by consensus or by appealing to the third experienced thoracic surgeon. To assure the accurate diagnoses of ESCC and mucosa, all the ROIs would be verified by an experienced thoracic surgeon.

Statistical analysis

The performance of the models in predicting histological ROIs would be assessed with the following parameters: (I) the Intersection over Union (IOU); (II) the positive predictive value (PPV); (III) the true positive rate (TPR); (IV) the accuracy; (V) the specificity; (VI) the dice similarity coefficient (DSC); (VII) the area under the receiver operating characteristic curve (AUC); (VIII) the balanced F score (F1-Score) (26). Loss function is a parameter evaluating the relative differences between prediction and truth; IOU is a parameter assessing the consistency between the ground truth and the segmentation mask, IOU = TP/(TP + FP + FN); PPV is a measurement of precision, PPV = TP/(TP + FP); TPR could reflect the sensitivity, TPR = TP/(TP + FN); the accuracy = (TP + TN)/(TP + FP + FN + TN); specificity = TN/(TN + FP); DSC is a statistical measure of spatial overlap for segmentation accuracy, DSC = 2 TP/(2 TP + FP + FN); while AUC is a criteria to optimize the model hyperparameters; and the balanced F score is the harmonic mean of the sensitivity and specificity and is used to establish optimal system accuracy (TP, true positive; FN, false negative; FP, false positive; TN, true negative).


Results

Training dataset and validation datasets

A retrospective analysis was conducted on surgical samples from ten patients (mean age: 62.7 years; male: 90%). Clinical data for patients in the primary cohort and two additional cohorts is available in Table 2. A total of 30 pairs of carcinomas and para-carcinomas tissues were obtained from surgical specimens and subsequently underwent H&E staining. Following WSI pre-processing, 292 and 239 patches were derived for the mucosa and tumor cohort with a sliding window sampling approach, respectively. Meanwhile, the two additional datasets primarily featured tumor-dominant WSIs, from which 59 and 48 patches were generated for external validation cohort 1 and validation cohort 2 following random attention sampling, respectively. An overview of the training and validation process is depicted in the Figure 2.

Table 2

Clinical characteristics for patients with ESCC in Zhongshan Hospital, Fudan University

Characteristics Main cohort Additional cohort 1 Additional cohort 2
Patients 10 5 5
Age (years) 62.7±5.0 67.4±6.7 57.4±7.2
Male 9 [90] 5 [100] 4 [80]
Tumor location
   Middle 8 [80] 5 [100] 3 [60]
   Lower 2 [20] 0 [0] 2 [40]
Clinical stage
   II 5 [50] 5 [100] 4 [80]
   III 5 [50] 0 [0] 1 [20]
Histological grade
   2 9 [90] 3 [60] 4 [80]
   3 1 [10] 2 [40] 1 [20]

Data are shown as mean ± SD or n [%]. ESCC, esophageal squamous cell carcinoma; SD, standard deviation.

Figure 2 The illustration of the framework for the development of segmentation models for hematoxylin and eosin staining WSIs. Magnification: 10×40. WSI, whole slide image.

Performance in the deep learning for mucosa

There were 262 and 30 patches allocated in the training and internal validation cohort for the mucosa segmentation task. The performance of models exhibited variability when different pretrained models were applied, however, all models consistently demonstrated robust discriminatory capability between mucosa and non-mucosa composition after 100 epochs of training (Table 3). The ascension of IOU in the training is shown in Figure 3A. Notably, the trained U-Net model exhibited the highest potential, achieving an IOU of 93.81%, accuracy of 96.96%, PPV of 96.45%, TPR of 96.65%, specificity of 98.41%, DSC of 96.81%, AUC of 96.11% and F1-Score of 96.55% in the internal validation. Besides, the transfer learning VGG19 (IOU =82.52%) model and model trained from scratch (IOU =84.70%) also demonstrated commendable identification abilities, with most parameters basically reaching 90%.

Table 3

The performance of segmentation system for mucosa in internal validation

Models IOU (%) Accuracy (%) PPV (%) TPR (%) Specificity (%) DSC (%) AUC (%) F1-Score (%)
U-Net 93.81 96.96 96.45 96.65 98.41 96.81 96.11 96.55
VGG16 75.69 82.75 87.47 89.87 89.55 86.16 82.62 88.65
VGG19 82.52 90.52 89.35 90.33 94.77 90.42 88.65 89.84
ResNet50 68.23 75.74 84.52 87.31 84.85 81.12 76.54 85.89
Xception 58.28 67.00 78.99 81.74 78.88 73.64 68.58 80.34
No pretrained 84.70 90.75 91.79 92.70 94.86 91.72 89.78 92.24

IOU, Intersection over Union; PPV, positive predictive value; TPR, true positive rate; DSC, dice similarity coefficient; AUC, area under the receiver operating characteristic curve; F1-Score, balanced F score.

Figure 3 The IOU in the training process. (A) Mucosa segmentation models; (B) tumor segmentation models. VGG, Visual Geometry Group; ResNet, Residual Network; IOU, Intersection over Union.

Performance in the deep learning for tumor

In the tumor segmentation, the training and internal validation cohorts consisted of 215 and 24 patches respectively. The IOU curves of models are displayed in the Figure 3B and the performance of models in the internal validation is summarized in Table 4. Except the U-Net model, most models performed poorly, with IOU not exceeding 52%. The U-Net based model outperformed all other models, achieving an IOU of 91.95%, accuracy of 95.90%, PPV of 95.62%, TPR of 95.71%, specificity of 97.88%, DSC of 95.81%, AUC of 94.92% and F1-Score of 95.67%.

Table 4

The performance of segmentation system for tumor in internal validation

Models IOU (%) Accuracy (%) PPV (%) TPR (%) Specificity (%) DSC (%) AUC (%) F1-Score (%)
U-Net 91.95 95.90 95.62 95.71 97.88 95.81 94.92 95.67
VGG16 51.01 65.71 51.75 69.51 73.81 67.56 62.41 59.33
VGG19 51.05 65.72 51.73 69.58 73.79 67.59 62.42 59.34
ResNet50 51.16 65.80 51.80 69.69 73.84 67.69 62.50 59.43
Xception 51.11 65.71 51.76 69.70 73.75 67.65 62.43 59.41
No pretrained 51.15 65.89 51.69 69.57 73.93 67.68 62.54 59.31

IOU, Intersection over Union; PPV, positive predictive value; TPR, true positive rate; DSC, dice similarity coefficient; AUC, area under the receiver operating characteristic curve; F1-Score, balanced F score.

External validation for tumor segmentation

Given the clinical value of tumor segmentation, the performance of models, especially the U-Net based model, were validated on two additional datasets.

Dataset 1: the results of external validation for dataset 1 are summarized in Table 5. As expected, U-Net outperformed other models. The VGG16, VGG19, ResNet50, Xception and scratch-based models yielded IOU of 25%, while U-Net model attained 59.86% for overall IOU. Notably, the resolution capability of U-Net model for tumor segmentation was specifically investigated, yielding 79.00% for IOU, 86.00% for accuracy, 90.00% for PPV, 90.66% for TPR, 92.21% for specificity, 88.27% for DSC, 85.61% for AUC, and 90.33% for F1-Score.

Table 5

The performance of segmentation system for tumor in external validation data set 1

Models IOU (%) Accuracy (%) PPV (%) TPR (%) Specificity (%) DSC (%) AUC (%) F1-Score (%)
U-Net for tumors 79.00 86.00 90.00 90.66 92.21 88.27 85.61 90.33
U-Net 59.86 72.74 76.03 77.17 83.80 74.89 71.83 76.60
VGG16 25.09 51.53 38.36 32.84 70.73 40.12 47.91 35.39
VGG19 25.00 52.28 36.08 32.39 70.47 40.00 47.73 34.14
ResNet50 25.61 52.45 37.41 33.35 70.76 40.78 48.19 35.27
Xception 25.64 54.21 33.86 32.73 70.84 40.82 48.24 33.28
No pretrained 25.68 52.44 37.64 33.48 70.80 40.87 48.24 35.44

IOU, Intersection over Union; PPV, positive predictive value; TPR, true positive rate; DSC, dice similarity coefficient; AUC, area under the receiver operating characteristic curve; F1-Score, balanced F score.

Dataset 2: in the dataset 2, similar result of external validation as dataset 1 was achieved (Table 6). The performance of 5 models other than U-Net was far from satisfactory. Interestingly, the prediction ability of the whole model in the dataset 2 (U-Net achieved 50.88% for IOU, 68.03% for accuracy, 59.02% for PPV, 66.87% for TPR, 78.48% for specificity, 67.44% for DSC, 64.68% for AUC, and 62.70% for F1-Score) might be not as good as that in dataset 1, however, the differentiation ability of tumor tissue surpassed the dataset 1 (IOU of 89.00%, accuracy of 91.00%, PPV of 97.00%, TPR of 97.59%, specificity of 94.78%, DSC of 94.18%, AUC of 91.89% and F1-Score of 97.29%).

Table 6

The performance of segmentation system for tumor in external validation data set 2

Models IOU (%) Accuracy (%) PPV (%) TPR (%) Specificity (%) DSC (%) AUC (%) F1-Score (%)
U-Net for tumors 89.00 91.00 97.00 97.59 94.78 94.18 91.89 97.29
U-Net 50.88 68.03 59.02 66.87 78.48 67.44 64.68 62.70
VGG16 5.00 34.80 9.59 5.52 60.05 9.52 32.53 7.00
VGG19 5.55 36.31 7.59 6.15 56.23 10.52 30.89 6.79
ResNet50 6.06 36.86 9.12 6.76 58.19 11.43 32.12 7.77
Xception 10.32 44.37 14.59 11.85 64.42 18.71 37.37 13.08
No pretrained 5.96 36.34 10.08 6.65 59.49 11.25 32.72 8.02

IOU, Intersection over Union; PPV, positive predictive value; TPR, true positive rate; DSC, dice similarity coefficient; AUC, area under the receiver operating characteristic curve; F1-Score, balanced F score.


Discussion

In this article, a rigorous development and validation of novel deep learning models for segmenting ESCC pathology were presented. And the model based on the U-Net pretrained parameters has demonstrated outstanding performance for mucosa and tumor segmentation in both internal and external validation. To the best of our knowledge, this is a pioneering study exploring deep learning methods for pathological images segmentation of esophageal carcinoma.

The rise and widespread adoption of digital image system and WSI have enabled features extraction and quantitative histomorphometry analysis. The implementation of deep learning based artificial intelligence assists on various tasks including diagnosis, grading and subtyping for carcinomas located in breast, lung and digestive tract (27). Based on CNNs, deep learning can ideally deal with the complex digital pathological image analysis tasks, and it could serve as a promising frontier. Previous studies have documented the utilization of CNNs in the classification for esophageal carcinoma H&E-stained images (7,28). In addition to classification, CNNs are also implemented in tumor image segmentation (29). In order to achieve precise segmentation, various networks structures have been proposed (21,30). However, current deep learning techniques for tumor segmentation primarily focused on morphological images including ultrasound (US), X-ray, CT, magnetic resonance imaging (MRI), and PET/CT, with only a limited emphasis on pathological images of certain carcinomas. This article delves into the viability of integrating deep learning techniques with pathological WSIs, widely regarded as the benchmark for tumor diagnosis. On the other hand, the development of traditional CNN network demands a sizable image cohort to ensure efficacy, while the U-Net network has demonstrated the ability to segment biomedical images effectively with minimal training data. To ensure the effectiveness of models with limited sample size, we applied the U-Net segmentation model in our study and achieved satisfactory results.

In this study, the U-Net framework was used for the segmentation learning, with the VGG net serving as the backbone. In the absence of specific threshold values, we basically assumed a threshold of 80% for parameters. Notably, apart from learning from scratch, transfer learning, a common technique in artificial intelligence, was employed to explore compatible segmentation models. In artificial intelligence community, finely tuned pre-trained models through transfer learning are utilized to reduce training time and potentially achieve better results. The IOU curves of transfer learning models reached a performance plateau in fewer epochs compared to models learned from scratch. This might be attributed to inherent differences between digital pathological images and images in ImageNet datasets. Moreover, compared to previous network structures, the U-Net model demonstrated several advantages. The corresponding layers in the expansive path and contracting path were seamed in order to ensure the sufficient features were extracted from limited images, as a result, a large number of feature channels in the upsampling part allowed the network to propagate characteristics information to higher resolution layers (15). Besides, the adoption of a tiling strategy enables the prediction of border region in large images without consuming excessive GPU memory. Therefore, it was expected that the U-Net based transfer learning model would exhibit superior performance in both mucosa and tumor segmentation in esophageal carcinoma, which was aligned with the findings of this study.

The current study has several strengths. Firstly, the employed networks, especially the U-Net transfer learning model, appeared capable of segmenting ESCC WSIs, generating accurate ROI segmentation masks with clear borderline. This indicates its potential clinical application in tumor staging and sections screening for surgical specimen. Even with a small training cohort, models demonstrated acceptable efficacy on external validation. Secondly, this study is the first exploration of the deep learning segmentation for ESCC pathology images, establishing a foundation to guide future researches. The mucosa segmentation model could aid in the tumor stage of ESCC, and tumor segmentation models may help pathologists rapidly screen slides and identify concerning areas. Moreover, previous studies have also demonstrated systems detecting neoplasia in endoscopic images to guide biopsy, these segmentation models could facilitate biopsy diagnosis (19,31). The combination of multiple deep learning segmentation systems could assist diagnosis and clinical decision, especially the feasibility of minimal invasive surgery. Overall, these strengths highlight the significance of this study and its potential implications for improving the diagnosis and management of ESCC.

This study has several limitations as well. In the first place, despite appealing results, the cohort was relatively small. Cancer heterogeneity means that limited sample size might affect the generalizability of the model, potentially contributing to poor segmentation in the external validation. Expanding the cohort, particularly with clinically studied patients, would improve the robustness of models. Besides, even though the VGG based U-Net network structure is highly regarded for medical images segmentation, the efficacy of other network architectures warrant exploration, and in further studies, comparisons should be made between the segmentation ability of different network structures. Based on limitations above, the model has not been applied in clinical practice yet. The model could be expected to be deployed on common computers and combined with other artificial intelligence models in clinical applications, while mature models developed with more extensive, diverse data are prerequisite for the clinical practice.


Conclusions

In conclusion, a deep learning framework for surgical pathology segmentation in ESCC was successfully developed and validated using a VGG based U-Net network in this study. The model based on U-Net pretraining weights offered a promising performance in contouring accurately mucosa and tumor in particular. These results hold potential for advancing quantitative histopathology diagnosis methodologies for patients with ESCC after surgery. Importantly, these segmentation models could be deployed on common computers or even personal laptops embedded with medical image equipment, offering the capabilities to highlight the ROIs for ESCC, thereby enhancing clinical decision making. Further studies will be needed to train these models in a comprehensive clinical WSI databases to distill the complex heterogeneity encountered in real-world clinical practice.


Acknowledgments

Funding: This work was supported by the National Natural Science Foundation of China (No. 81400681), the China Postdoctoral Science Foundation Grant (No. 2018M631394), and the Science and Technology Innovation Plan of Shanghai Science and Technology Commission (No. 22Y11907200). The funding agencies had no role in study design, collection and analyses of data, decision to publish, or manuscript preparation.


Footnote

Reporting Checklist: The authors have completed the TRIPOD reporting checklist. Available at https://jgo.amegroups.com/article/view/10.21037/jgo-23-587/rc

Data Sharing Statement: Available at https://jgo.amegroups.com/article/view/10.21037/jgo-23-587/dss

Peer Review File: Available at https://jgo.amegroups.com/article/view/10.21037/jgo-23-587/prf

Conflicts of Interest: All authors have completed the ICMJE uniform disclosure form (available at https://jgo.amegroups.com/article/view/10.21037/jgo-23-587/coif). The authors have no conflicts of interest to declare.

Ethical Statement: The authors are accountable for all aspects of the work in ensuring that questions related to the accuracy or integrity of any part of the work are appropriately investigated and resolved. The study was conducted in accordance with the Declaration of Helsinki (as revised in 2013). The study was approved by ethics review board of Zhongshan Hospital, Fudan University (No. ZS-B2021-708) and individual consent for this retrospective analysis was waived.

Open Access Statement: This is an Open Access article distributed in accordance with the Creative Commons Attribution-NonCommercial-NoDerivs 4.0 International License (CC BY-NC-ND 4.0), which permits the non-commercial replication and distribution of the article with the strict proviso that no changes or edits are made and the original work is properly cited (including links to both the formal publication through the relevant DOI and the license). See: https://creativecommons.org/licenses/by-nc-nd/4.0/.


References

  1. Siegel RL, Miller KD, Fuchs HE, et al. Cancer statistics, 2022. CA Cancer J Clin 2022;72:7-33. [Crossref] [PubMed]
  2. Li J, Chen B, Wang X, et al. Prognosis of patients with esophageal squamous cell carcinoma undergoing surgery versus no surgery after neoadjuvant chemoradiotherapy: a retrospective cohort study. J Gastrointest Oncol 2022;13:903-11. [Crossref] [PubMed]
  3. Waters JK, Reznik SI. Update on Management of Squamous Cell Esophageal Cancer. Curr Oncol Rep 2022;24:375-85. [Crossref] [PubMed]
  4. Niazi MKK, Parwani AV, Gurcan MN. Digital pathology and artificial intelligence. Lancet Oncol 2019;20:e253-61. [Crossref] [PubMed]
  5. Bashshur RL, Krupinski EA, Weinstein RS, et al. The Empirical Foundations of Telepathology: Evidence of Feasibility and Intermediate Effects. Telemed J E Health 2017;23:155-91. [Crossref] [PubMed]
  6. Jahn SW, Plass M, Moinfar F. Digital Pathology: Advantages, Limitations and Emerging Perspectives. J Clin Med 2020;9:3697. [Crossref] [PubMed]
  7. Tomita N, Abdollahi B, Wei J, et al. Attention-Based Deep Neural Networks for Detection of Cancerous and Precancerous Esophagus Tissue on Histopathological Slides. JAMA Netw Open 2019;2:e1914645. [Crossref] [PubMed]
  8. Madabhushi A, Lee G. Image analysis and machine learning in digital pathology: Challenges and opportunities. Med Image Anal 2016;33:170-5. [Crossref] [PubMed]
  9. Zhu J, Liu M, Li X. Progress on deep learning in digital pathology of breast cancer: a narrative review. Gland Surg 2022;11:751-66. [Crossref] [PubMed]
  10. van der Sommen F, Curvers WL, Nagengast WB. Novel Developments in Endoscopic Mucosal Imaging. Gastroenterology 2018;154:1876-86. [Crossref] [PubMed]
  11. Avanzo M, Stancanello J, Pirrone G, et al. Radiomics and deep learning in lung cancer. Strahlenther Onkol 2020;196:879-87. [Crossref] [PubMed]
  12. Huang B, Tian S, Zhan N, et al. Accurate diagnosis and prognosis prediction of gastric cancer using deep learning on digital pathological images: A retrospective multicentre study. EBioMedicine 2021;73:103631. [Crossref] [PubMed]
  13. Kawahara D, Murakami Y, Tani S, et al. A prediction model for pathological findings after neoadjuvant chemoradiotherapy for resectable locally advanced esophageal squamous cell carcinoma based on endoscopic images using deep learning. Br J Radiol 2022;95:20210934. [Crossref] [PubMed]
  14. Li F, Yang Y, Wei Y, et al. Deep learning-based predictive biomarker of pathological complete response to neoadjuvant chemotherapy from histological images in breast cancer. J Transl Med 2021;19:348. [Crossref] [PubMed]
  15. Ronneberger O, Fischer P, Brox T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In: Navab N, Hornegger J, Wells WM, et al. editors. Medical Image Computing and Computer-Assisted Intervention – MICCAI 2015. Cham: Springer International Publishing; 2015:234-41.
  16. Swiderska-Chadaj Z, Pinckaers H, van Rijthoven M, et al. Learning to detect lymphocytes in immunohistochemistry with deep learning. Med Image Anal 2019;58:101547. [Crossref] [PubMed]
  17. Zhou T, Tan T, Pan X, et al. Fully automatic deep learning trained on limited data for carotid artery segmentation from large image volumes. Quant Imaging Med Surg 2021;11:67-83. [Crossref] [PubMed]
  18. Hu Y, Xie C, Yang H, et al. Computed tomography-based deep-learning prediction of neoadjuvant chemoradiotherapy treatment response in esophageal squamous cell carcinoma. Radiother Oncol 2021;154:6-13. [Crossref] [PubMed]
  19. de Groof AJ, Struyvenberg MR, van der Putten J, et al. Deep-Learning System Detects Neoplasia in Patients With Barrett's Esophagus With Higher Accuracy Than Endoscopists in a Multistep Training and Validation Study With Benchmarking. Gastroenterology 2020;158:915-929.e4. [Crossref] [PubMed]
  20. Mönig S, Chevallay M, Niclauss N, et al. Early esophageal cancer: the significance of surgery, endoscopy, and chemoradiation. Ann N Y Acad Sci 2018;1434:115-23. [Crossref] [PubMed]
  21. Jiang HY, Diao ZS, Yao YD. Deep learning techniques for tumor segmentation: a review. Journal of Supercomputing 2022;78:1807-51. [Crossref]
  22. Chen W, Wang WP, Liu L, et al. New ideas and trends in deep multimodal content understanding: a review. Neurocomputing 2021;426:195-215. [Crossref]
  23. Chollet F, IEEE. Xception: Deep Learning with Depthwise Separable Convolutions. 30th IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2017) 2017:1800-7.
  24. Szegedy C, Ioffe S, Vanhoucke V, et al. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning. Thirty-first AAAI Conference on Artificial Intelligence 2017:4278-84.
  25. Deng J, Dong W, Socher R, et al. ImageNet: A Large-Scale Hierarchical Image Database. CVPR: 2009 IEEE Conference on Computer Vision and Pattern Recognition 2009:248-55.
  26. Taha AA, Hanbury A. Metrics for evaluating 3D medical image segmentation: analysis, selection, and tool. BMC Med Imaging 2015;15:29. [Crossref] [PubMed]
  27. Jiang Y, Yang M, Wang S, et al. Emerging role of deep learning-based artificial intelligence in tumor pathology. Cancer Commun (Lond) 2020;40:154-66. [Crossref] [PubMed]
  28. Beuque M, Martin-Lorenzo M, Balluff B, et al. Machine learning for grading and prognosis of esophageal dysplasia using mass spectrometry and histological imaging. Comput Biol Med 2021;138:104918. [Crossref] [PubMed]
  29. Xu J, Luo X, Wang G, et al. A Deep Convolutional Neural Network for segmenting and classifying epithelial and stromal regions in histopathological images. Neurocomputing (Amst) 2016;191:214-23. [Crossref] [PubMed]
  30. Rezaeijo SM, Jafarpoor Nesheli S, Fatan Serj M, et al. Segmentation of the prostate, its zones, anterior fibromuscular stroma, and urethra on the MRIs and multimodality image fusion using U-Net model. Quant Imaging Med Surg 2022;12:4786-804. [Crossref] [PubMed]
  31. Islam MM, Poly TN, Walther BA, et al. Deep Learning for the Diagnosis of Esophageal Cancer in Endoscopic Images: A Systematic Review and Meta-Analysis. Cancers (Basel) 2022;14:5996. [Crossref] [PubMed]
Cite this article as: Su F, Zhang W, Liu Y, Chen S, Lin M, Feng M, Yin J, Tan L, Shen Y. The development and validation of pathological sections based U-Net deep learning segmentation model for the detection of esophageal mucosa and squamous cell neoplasm. J Gastrointest Oncol 2023;14(5):1982-1992. doi: 10.21037/jgo-23-587

Download Citation