Clin Endosc > Volume 56(1); 2023 > Article
Meinikheim, Messmann, and Ebigbo: Role of artificial intelligence in diagnosing Barrett’s esophagus-related neoplasia


Barrett’s esophagus is associated with an increased risk of adenocarcinoma. Thorough screening during endoscopic surveillance is crucial to improve patient prognosis. Detecting and characterizing dysplastic or neoplastic Barrett’s esophagus during routine endoscopy are challenging, even for expert endoscopists. Artificial intelligence-based clinical decision support systems have been developed to provide additional assistance to physicians performing diagnostic and therapeutic gastrointestinal endoscopy. In this article, we review the current role of artificial intelligence in the management of Barrett’s esophagus and elaborate on potential artificial intelligence in the future.


Gastroesophageal reflux disease, family history, age, or sex predisposes to Barrett’s esophagus (BE), a complication in which stratified esophageal squamous epithelium at the level of the gastroesophageal junction is replaced by metaplastic columnar epithelium.1,2 BE is a precursor to esophageal adenocarcinoma (EAC).3 The incidence of EAC has increased significantly in recent years4,5 and is often associated with poor prognosis due to delayed diagnosis.3 Risk factors for the progression of BE to dysplastic BE or EAC include the length of the BE segment, age, ethnicity, lifestyle, and medication.6,7
The gold standard for diagnosing BE and Barrett’s esophagus-related neoplasia (BERN) is endoscopic evaluation with histological confirmation. However, differentiating between non-dysplastic BE and BERN can be challenging, even for expert endoscopists. Existing biopsy strategies are suboptimal, with EAC and BE miss rates of >20% and 50%, respectively.8,9 This is partly attributable to poor compliance with existing biopsy protocols and the complexity of differentiation between non-dysplastic BE and BERN during endoscopic evaluation.8,9 Using imaging techniques such as narrow band imaging (NBI) with standardized classification systems for BE and BERN can help improve the diagnostic performance of endoscopists.10,11 Additionally, advanced imaging techniques such as chromoendoscopy with indigo carmine or acetic acid are valuable options and are recommended for high-quality assessment of BE (Fig. 1).11-14 However, implementation of advanced imaging techniques in daily practice requires extensive experience. Recently, several research groups have developed deep learning algorithms to improve the detection and characterization of BERN.


Artificial intelligence: a brief introduction

Time constraints and cost issues have led to the search for more efficient modalities for diagnosis and treatment of patients. To this end, artificial intelligence (AI) has become increasingly relevant to the field of medicine, especially for the early diagnosis of neoplasia. AI is an umbrella term for a wide range of topics, and the general idea is to solve problems using algorithms that require characteristics similar to human intelligence, such as the ability to learn. Machine learning (ML), as a subdiscipline of AI, describes algorithms employed to learn from pre-existing data. The field of AI most relevant to medicine, particularly endoscopy, is deep learning. Deep learning is a subtype of ML and describes a method that aims to solve defined problems with little to no supervision using vast amounts of data. Similar to the human brain, the applied algorithms or convolutional neural networks (CNNs) consist of numerous layers of neurons. CNNs learn to recognize certain patterns within the provided input data and produce a prediction or output.15 For example, in gastrointestinal (GI) endoscopy, the output could predict the dignity of an observed lesion and differentiate between neoplasia and non-neoplasia. This task is called computer-aided diagnosis (CADx). Meanwhile, identifying the lesion of interest is called computer-aided detection (CADe).16 One way to quantify the accuracy of an AI system during CADe is to determine the intersection over union (IOU). It compares the ground truth represented by a box with the output of the AI algorithm represented by a bounding box. IOU is the result of the division of the “area of overlap” by the “area of union”.17 Another way to quantify the accuracy of object detection is the Dice coefficient (or Sørensen–Dice coefficient), which is calculated by dividing the area of union by the total number of pixels in the individual areas.18

Relevance of AI for diagnostic purposes in Barrett’s esophagus

Detecting high-grade dysplasia (HGD) and EAC during endoscopy is difficult and challenging. Considering the consequences of false negative results or missed lesions a “second opinion” during endoscopic examination is an appealing thought.
van der Sommen et al.19 developed a pattern recognition-based AI system that could detect EAC on images of BE with a sensitivity and specificity of >80%. de Groof et al.20 developed an AI system based on a CNN that outperformed general endoscopists during an image-based trial with a sensitivity of 93% and specificity of 72%, compared to endoscopists with a sensitivity of 88% and specificity of 73%. In a follow-up study, the same group achieved a sensitivity, specificity, and accuracy of 91%, 89%, and 90%, respectively, in the differentiation between BE and BERN on high-definition white light endoscopy (HD-WLE) images.21 Similarly, several other groups have successfully differentiated BE from BERN during image-based studies. Hashimoto et al.22 managed to classify BE correctly with a sensitivity, specificity, and accuracy of 96.4%, 94.2%, and 95.4%, respectively. Furthermore, they managed to detect with an IOU of 0.3 and a mean average precision of 0.75. Iwagami et al.23 focused on an Asian population and developed an AI system that detects cancer at the esophagogastric junction with a sensitivity, specificity, and accuracy of 94%, 42%, and 66%, respectively. This study was conducted using 232 HD-WLE still images of 36 cancer and 43 non-cancer cases. The performance was compared to that of experts with a sensitivity, specificity, and accuracy of 88%, 43%, and 63%, respectively.
Ghatwary et al.24 compared the different methods used for the CNN-based development of CADe. On 100 HD-WLE still images, the single-shot multibox detector performed better (sensitivity, specificity, and F-score: 96%, 92%, and 0.94, respectively) than other region-based CNNs.
Struyvenberg et al.25 developed an AI system that could classify BE and BERN using NBI with a sensitivity, specificity, and accuracy of 88%, 78%, and 84%, respectively. Hussein et al.26 trained their CNN with BE and BERN images in HD-WLE and optical chromoendoscopy (i-scan from Pentax Hoya, Tokyo, Japan). They achieved a sensitivity, specificity, and area under the receiver operator curve of 91%, 79%, and 93%, respectively, during a classification task. In the same study, their CNN had an average Dice score of 50% during the segmentation task.26
Ebigbo et al.27 used HD-WLE images, NBI images, and texture and color enhancement imaging for training a CNN. Hence, they can offer multimodal CADe and CADx with promising results. Initial studies based on still images with HD-WLE—images from the data of the Medical Image Computing and Computer Assisted Interventions Society demonstrated a sensitivity and specificity of 92% and 100%, respectively. In a study with an independent data set from the University Hospital of Augsburg, these results could be reproduced using HD-WLE images (sensitivity/specificity of 97%/88%) and NBI images (sensitivity/specificity of 94%/80%). Subsequently, the algorithm demonstrated its potential for real-life applications. The AI system captured images randomly off an endoscopic live stream and differentiated BE from EAC with an accuracy of 89.9%.28
During the early phase of this research field, even though several AI systems were able to demonstrate promising results in preclinical and pilot-phase clinical studies, AI algorithms were mostly only able to offer a per-image evaluation of BE and BERN. Ebigbo et al.29 were one of the first research groups to differentiate BE from BERN in 3 different imaging modalities in real time (Fig. 2).
Differentiating between T1a and T1b lesions during endoscopic examination is extremely difficult. Ebigbo et al.30 developed an algorithm that was able to differentiate T1a from T1b adenocarcinoma in an image-based pilot study on 230 HD-WLE images with a sensitivity, specificity, F1-score, and accuracy of 77%, 64%, 74%, and 71%, respectively. This performance was comparable to that of expert endoscopists. A meta-analysis that included 6 studies with 561 endoscopic images of patients with BE was published by Lui et al.31 Three studies used CNN as the backbone, whereas 3 used a non-CNN backbone. Overall, pooled sensitivity and specificity were approximately 86% to 88% and 86% to 90% respectively, demonstrating the promising potential of AI systems for detecting neoplastic lesions in BE.31 Nevertheless, although a meta-analysis generally incorporates data from different trials, comparing AI studies in the preclinical phase is particularly difficult because of the heterogeneity of data samples and the algorithms used for the various trials. This meta-analysis was published in 2020, and since then, more data have emerged on this topic.
Volumetric laser endomicroscopy (VLE) is an advanced imaging method that applies the principles of optical coherence tomography.32 In an observational study conducted by Smith et al.,33 VLE-guided biopsy improved, compared to random biopsies, the detection of BERN by 700% in cases where with other imaging methods no visual cues for neoplasia had been detected. However, the application of VLE and the interpretation of acquired information require practice and experience. Therefore, several research groups have attempted to address this problem using AI. Trindade et al.34 were one of the first to develop an AI system that could detect and demarcate previously determined characteristics of dysplasia on VLE images. In a case report of a patient with long-segment BE with no visual cues on HD-WLE or NBI and negative random biopsies, VLE-guided histological acquisition demonstrated focal low-grade dysplasia. Struyvenberg et al.35 developed an AI system with a sensitivity of 91% and specificity of 82% compared to VLE experts with a pooled sensitivity and specificity of 70% and 81%, respectively.
During a pilot clinical study with a manufactured spectral endoscope, Waterhouse et al.36 tested an AI system that could differentiate the spectra of BE from those of BERN with a sensitivity of 83.7% and a specificity of 85.5%.
Beyond preclinical and pilot phase clinical trials, several AI systems have already been approved for clinical use and are now commercially available. WISE VISION (NEC Corp, Tokyo, Japan) was developed and can differentiate between BE and BERN, and offers a visual representation of the area that has been classified as HGD or neoplastic. AI system was developed (CADU, Odin Medical Ltd, London, UK) that can differentiate BE from BERN and offers a visual representation when the observed image is deemed dysplastic.
As more researchers are currently in the preclinical phase or on the verge of clinical trials, a standardized way to ensure minimum requirements during the development process and later in terms of performance is urgently needed. The American Society for Gastrointestinal Endoscopy developed a guideline, the Preservation and Incorporation of Valuable Endoscopic Innovations, for the integration of new imaging technology in the context of BE. For technologies that intend to replace random biopsies with targeted biopsies, a minimum performance of 90% sensitivity, 98% negative predictive value, and 80% specificity for HGD or EAC is recommended.37 In addition to standardized threshold performance requirements, guidelines that ensure quality standards during the developmental process of AI systems are urgently needed (Table 1).19-23,25-28,30,35,36

Computer-aided quality control of upper gastrointestinal endoscopy

AI has the potential to improve various aspects of GI endoscopy such as inter-examiner variability. For example, Pan et al.38 developed an AI system that automatically identifies the squamous–columnar junction and gastroesophageal junction on images. Ali et al.39 worked on an AI system that automatically determined the extension of the BE according to the Prague classification. With the extension of BE as a relevant factor for risk stratification, AI-assisted standardized and automated reporting has the potential to significantly improve patient care.
Additionally, a complete examination with thorough inspection is crucial to avoid missed lesions. The European Society of Gastrointestinal Endoscopy and British Society of Gastroenterology recommend photo documentation of specific landmarks during upper GI endoscopy.40,41 According to the European Society of Gastrointestinal Endoscopy, a photo documentation rate of ≥90% is recommended to meet the minimum quality requirements for upper GI endoscopy.41 Incomplete examinations during upper GI endoscopy can lead to an increased cancer miss rate. AI applications have the potential to provide immediate feedback on the quality of endoscopic examinations. Wu et al.42 developed an AI system, WISENSE, that can detect blind spots, document examination time, and automatically record images for photo documentation during the procedure. A randomized controlled trial compared upper GI endoscopy with or without the support of WISENSE and demonstrated a lower blind spot rate in the group that received support from the AI system (Table 2).38,39,42


Despite currently several research teams being on the verge of clinical studies and real-life applications, their good results from preclinical studies are often limited to data from their own centers. This is particularly relevant when AI systems are commercially available and used at different centers. Moreover, during development, a heterogeneous set of data with not only a tremendous number of frames but also a sufficient number of different cases is important for the robustness of an AI system.
Furthermore, to date, no standardized method for evaluating the performance of an AI system has been created, thus indicating an urgent need to establish standardized evaluation methods. This also includes uniform terminology when describing the methods and results of the respective studies.


Modern medicine, with its ever-growing complexity coupled with limited human and material resources, is urgently needed for more efficient workflow while maintaining a high level of patient care. AI may help to solve some of these problems. Current AI applications are not being developed to replace human physicians but to support physicians during complex diagnostic and therapeutic processes. Correct interpretation of the additional information provided by AI systems is crucial for optimal performance. Human–computer interaction should be a focus during the development of AI systems as the performance is. Creating AI systems that seamlessly integrate themselves into the daily routine of examiners is important. Furthermore, offering feedback on the confidence of an AI system in its current prediction is crucial. AI systems are only as good as the data input they receive, and low-quality data result in lower diagnostic performance. Although most AI systems for BE and BERN are still in the initial and preclinical phases, the immense potential of AI in routine clinical practice is evident. In the future, AI will optimize endoscopic practice and improve long-term patient outcomes.


Conflicts of Interest
The authors have no potential conflicts of interest.
Author Contributions
Conceptualization: MM, HM, AE; Data curation: MM, HM, AE; Formal analysis: MM, HM, AE; Investigation: MM, HM, AE; Methodology: MM, HM, AE; Project administration: MM, HM, AE; Resources: MM, HM, AE; Software: MM, HM, AE; Supervision: MM, HM, AE; Validation: MM, HM, AE; Visualization: MM, HM, AE; Writing–original draft: MM, HM, AE; Writing–review & editing: MM, HM, AE.

Fig. 1.
Images of Barrett’s esophagus-related neoplasia during endoscopy with an Olympus Evis X1 system (Olympus, Tokyo, Japan) in high-definition white light endoscopy (A), narrow band imaging (B), acetic acid chromoendoscopy (C), and chromoendoscopy with indigo carmine (D).
Fig. 2.
Detection and characterization of Barrett’s esophagus-related neoplasia during endoscopy with Olympus Evis X1 system using an AI system developed by the University Hospital of Augsburg and Ostbayerische Technische Hochschule Regensburg (OTH-Regensburg) with classification and segmentation in narrow band imaging (A), texture and color enhancement imaging (B) and high-definition white light endoscopy (C). The corresponding heatmaps are available at the top left corner of the user interface.
Table 1.
Summary of studies exploring the application of AI during the evaluation of Barrett’s esophagus
Study Aim of the study Data for the development Experimental design Performance
van der Sommen et al. (2016)19 AI stand-alone performance during detection of early neoplasia in BE 100 HD-WLE images from 44 patients Per image sensitivity and specificity of 83%
de Groof et al. (2020)20 Evaluation of AI stand-alone performance and comparing it to the performance of nonexpert endoscopists Pretraining: 494,364 images from all intestinal segments 1. Validation-only dataset: 80 HD-WLE images of 80 patients 1. CADx: sensitivity, 90%; specificity, 88%; accuracy, 89%
Training: 1,544 BE and BERN HD-WLE images of 509 patients 2. Dataset for validation and comparison to 53 nonexpert endoscopists: 80 HD-WLE images of 80 patients  CADe: optimal biopsy spot in 97%
2. CADx: sensitivity, 93%; specificity, 83%; accuracy, 88%
 CADe: optimal biopsy spot in 92%
 Endoscopists: sensitivity, 72%; specificity, 74%; accuracy, 73%
de Groof et al. (2020)21 Detection of BERN during real-life endoscopic examination Training: 1,544 BE and BERN HD-WLE images of 509 patients Evaluation of 144 HD-WLE of 10 patients with BERN and 10 with BE Sensitivity, 76%; specificity, 86%; accuracy, 84%
Hashimoto et al. (2020)22 Evaluation of AI stand-alone performance during BERN detection Total: 916 images of 65 patients with BERN, 919 images of 30 patients without dysplastic BE Evaluation of 225 BERN and 233 BE images Sensitivity, 96.4%; specificity, 94.2%; accuracy, 95.4%
Training: 691 images with BERN, 686 with BE HD-WLE and NBI images Mean average precision: 0.75
IOU: 0.3
Iwagami et al. (2021)23 AI stand-alone performance during detection of adenocarcinoma at the EGJ and comparison to 15 experts Training: 1,172 images from 166 EGJ cancer cases, 2,271 images of normal EGJ mucosa Evaluation of 232 HD-WLE images from 79 EGJ cancer and non-cancer cases AI stand-alone: sensitivity, 94%; specificity, 42%; accuracy, 66%
Comparison to 15 experts Experts: sensitivity, 88%; specificity, 43%; accuracy, 63%
Struyvenberg et al. (2021)25 AI stand-alone performance during differentiation between BE and BERN on near focus videos Pretraining: 494,364 images from all intestinal segments Internal validation: 71 BE and 112 BERN near focus NBI images Internal validation: sensitivity, 88%; specificity, 78%; accuracy: 84%
Training: 557 BE and 690 BERN HD-WLE overview images, 71 BE and 112 BERN near focus NBI images External validation: 59 BERN and 98 BE near focus NBI videos External validation: sensitivity, 85%; specificity, 83%; accuracy, 83%
Hussein et al. (2022)26 AI stand-alone performance during classification and localization of BE and BERN For classification: Classification: 264 i-scan images of 28 BERN and 16 BE patients Sensitivity, 91%; specificity, 79%; Dice score, 50% (with one expert)
Training: 148,936 frames of 31 BERN, 31 BE and 2 normal esophagus Segmentation: 86 i-scan images of 28 BERN patients
Validation: 25,161 frames of 6 BERN and 5 BE
For segmentation:
Training: 94 images of 30 BERN
Validation: 12 images of 6 BERN
Ebigbo et al. (2019)27 AI stand-alone performance during detection of BE and BERN MICCAI data: 100 (HD-WLE) images of 39 BE and BERN cases MICCAI data (HD-WLE images-only): sensitivity, 92%; specificity, 100%; Dice coefficient, 0.56
Augsburg data: 148 (HD-WLE/NBI) images of 74 BE and BERN cases Augsburg data (HD-WLE/NBI): sensitivity, 97%/94%; specificity, 88%/80%; Dice coefficient, 0.72
Comparison to expert segmentation
Ebigbo et al. (2020)28 Detection of BERN during real-life endoscopic examination Training: 129 images of 129 cases of BE and BERN Validation of the AI system under real-life examination conditions with 14 patients Sensitivity, 83.7%; specificity, 100%; accuracy, 89.9%
Real-time evaluation of 36 extracted BERN and 26 BE images
Ebigbo et al. (2021)30 Prediction of submucosal invasion of BERN with the help of AI; comparison to expert endoscopists Images of pT1a and pT1b adenocarcinoma Differentiation between pT1a and pT1b BERN AI stand-alone: sensitivity, 77%; specificity, 64%; accuracy, 71%
108 pT1a and 122 pT1b Experts: sensitivity, 63%; specificity, 78%; accuracy: 70%
HD-WLE BERN images
Comparison to 5 experts
Struyvenberg et al. (2021)35 AI-aided detection of BERN during VLE Training: 22 patients with 134 BE and 38 BERN targets Validation set: 95 BE and 51 BERN targets of 25 patients AI stand-alone: sensitivity, 91%; specificity, 82%; accuracy, 85%
Comparison to 10 VLE experts Experts: sensitivity, 70%; specificity, 81%; accuracy, 77%
Waterhouse et al. (2021)36 AI-aided differentiation of BE from BERN during spectral endoscopy Training: 572 spectra Differentiation of BE from BERN during spectral endoscopy Sensitivity, 83.7%; specificity, 85.5%; accuracy: 84.8%
Test-set: 143 spectra

AI, artificial intelligence; BE, non-dysplastic Barrett’s esophagus; HD-WLE, high-definition white light endoscopy; BERN, Barrett’s esophagus-related neoplasia; CADx, computer-aided diagnosis; CADe, computer-aided detection; IOU, intersection over union; EGJ, esophagogastric junction; NBI, narrow band imaging; MICCAI, Medical Image Computing and Computer Assisted Interventions Society; VLE, volumetric laser endomicroscopy.

Table 2.
Overview of current studies in the context of computer-aided quality control during upper gastrointestinal endoscopy
Study Application
Pan et al. (2021)38 Automatic AI-aided identification of the squamous–columnar junction and gastroesophageal junction
Ali et al. (2021)39 Automatic AI-aided determination of BE extension
Wu et al. (2019)42 WISENSE: automatic time measurement, recording of images, and detection of blind spots

AI, artificial intelligence; BE, Barrett's esophagus.


1. Spechler SJ, Souza RF. Barrett’s esophagus. N Engl J Med 2014;371:836–845.
crossref pmid
2. Qumseya BJ, Bukannan A, Gendy S, et al. Systematic review and meta-analysis of prevalence and risk factors for Barrett’s esophagus. Gastrointest Endosc 2019;90:707–717.
crossref pmid
3. Smyth EC, Lagergren J, Fitzgerald RC, et al. Oesophageal cancer. Nat Rev Dis Primers 2017;3:17048.
crossref pmid pmc pdf
4. Coleman HG, Xie SH, Lagergren J. The epidemiology of esophageal adenocarcinoma. Gastroenterology 2018;154:390–405.
crossref pmid
5. Sung H, Ferlay J, Siegel RL, et al. Global cancer statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J Clin 2021;71:209–249.
crossref pmid pdf
6. Kambhampati S, Tieu AH, Luber B, et al. Risk factors for progression of Barrett’s esophagus to high grade dysplasia and esophageal adenocarcinoma. Sci Rep 2020;10:4899.
crossref pmid pmc pdf
7. Chandrasekar VT, Hamade N, Desai M, et al. Significantly lower annual rates of neoplastic progression in short- compared to long-segment non-dysplastic Barrett’s esophagus: a systematic review and meta-analysis. Endoscopy 2019;51:665–672.
crossref pmid
8. Visrodia K, Singh S, Krishnamoorthi R, et al. Magnitude of missed esophageal adenocarcinoma after Barrett’s esophagus diagnosis: a systematic review and meta-analysis. Gastroenterology 2016;150:599–607.
crossref pmid pmc
9. Singer ME, Odze RD. High rate of missed Barrett's esophagus when screening with forceps biopsies. Esophagus 2023;20:143–149.
10. Sharma P, Bergman JJ, Goda K, et al. Development and validation of a classification system to identify high-grade dysplasia and esophageal adenocarcinoma in Barrett’s esophagus using narrow-band imaging. Gastroenterology 2016;150:591–598.
crossref pmid
11. ASGE Technology Committee, Thosani N, Abu Dayyeh BK, et al. ASGE Technology Committee systematic review and meta-analysis assessing the ASGE Preservation and Incorporation of Valuable Endoscopic Innovations thresholds for adopting real-time imaging-assisted endoscopic targeted biopsy during endoscopic surveillance of Barrett’s esophagus. Gastrointest Endosc 2016;83:684–698.
crossref pmid
12. Qumseya BJ, Wang H, Badie N, et al. Advanced imaging technologies increase detection of dysplasia and neoplasia in patients with Barrett's esophagus: a meta-analysis and systematic review. Clin Gastroenterol Hepatol 2013;11:1562–1570.
crossref pmid pmc
13. Tholoor S, Bhattacharyya R, Tsagkournis O, et al. Acetic acid chromoendoscopy in Barrett’s esophagus surveillance is superior to the standardized random biopsy protocol: results from a large cohort study (with video). Gastrointest Endosc 2014;80:417–424.
crossref pmid
14. Chedgy FJ, Subramaniam S, Kandiah K, et al. Acetic acid chromoendoscopy: improving neoplasia detection in Barrett’s esophagus. World J Gastroenterol 2016;22:5753–5760.
crossref pmid pmc
15. Topol EJ. High-performance medicine: the convergence of human and artificial intelligence. Nat Med 2019;25:44–56.
crossref pmid pdf
16. van der Sommen F, de Groof J, Struyvenberg M, et al. Machine learning in GI endoscopy: practical guidance in how to interpret a novel field. Gut 2020;69:2035–2045.
crossref pmid pmc
17. Hsiao CH, Lin PC, Chung LA, et al. A deep learning-based precision and automatic kidney segmentation system using efficient feature pyramid networks in computed tomography images. Comput Methods Programs Biomed 2022;221:106854.
crossref pmid
18. Zou KH, Warfield SK, Bharatha A, et al. Statistical validation of image segmentation quality based on a spatial overlap index. Acad Radiol 2004;11:178–189.
crossref pmid pmc
19. van der Sommen F, Zinger S, Curvers WL, et al. Computer-aided detection of early neoplastic lesions in Barrett’s esophagus. Endoscopy 2016;48:617–624.
crossref pmid
20. de Groof AJ, Struyvenberg MR, van der Putten J, et al. Deep-learning system detects neoplasia in patients with Barrett’s esophagus with higher accuracy than endoscopists in a multistep training and validation study with benchmarking. Gastroenterology 2020;158:915–929.
crossref pmid
21. de Groof AJ, Struyvenberg MR, Fockens KN, et al. Deep learning algorithm detection of Barrett’s neoplasia with high accuracy during live endoscopic procedures: a pilot study (with video). Gastrointest Endosc 2020;91:1242–1250.
crossref pmid
22. Hashimoto R, Requa J, Dao T, et al. Artificial intelligence using convolutional neural networks for real-time detection of early esophageal neoplasia in Barrett’s esophagus (with video). Gastrointest Endosc 2020;91:1264–1271.
crossref pmid
23. Iwagami H, Ishihara R, Aoyama K, et al. Artificial intelligence for the detection of esophageal and esophagogastric junctional adenocarcinoma. J Gastroenterol Hepatol 2021;36:131–136.
crossref pmid pdf
24. Ghatwary N, Zolgharni M, Ye X. Early esophageal adenocarcinoma detection using deep learning methods. Int J Comput Assist Radiol Surg 2019;14:611–621.
crossref pmid pmc pdf
25. Struyvenberg MR, de Groof AJ, van der Putten J, et al. A computer-assisted algorithm for narrow-band imaging-based tissue characterization in Barrett’s esophagus. Gastrointest Endosc 2021;93:89–98.
crossref pmid
26. Hussein M, González-Bueno Puyal J, Lines D, et al. A new artificial intelligence system successfully detects and localises early neoplasia in Barrett’s esophagus by using convolutional neural networks. United European Gastroenterol J 2022;10:528–537.
crossref pmid pmc pdf
27. Ebigbo A, Mendel R, Probst A, et al. Computer-aided diagnosis using deep learning in the evaluation of early oesophageal adenocarcinoma. Gut 2019;68:1143–1145.
crossref pmid pmc
28. Ebigbo A, Mendel R, Probst A, et al. Real-time use of artificial intelligence in the evaluation of cancer in Barrett’s oesophagus. Gut 2020;69:615–616.
crossref pmid pmc
29. Ebigbo A, Mendel R, Probst A, et al. Multimodal imaging for detection and segmentation of Barrett’s esophagus-related neoplasia using artificial intelligence. Endoscopy 2022;54:E587.
crossref pmid
30. Ebigbo A, Mendel R, Rückert T, et al. Endoscopic prediction of submucosal invasion in Barrett’s cancer with the use of artificial intelligence: a pilot study. Endoscopy 2021;53:878–883.
crossref pmid
31. Lui TK, Tsui VW, Leung WK. Accuracy of artificial intelligence-assisted detection of upper GI lesions: a systematic review and meta-analysis. Gastrointest Endosc 2020;92:821–830.
crossref pmid
32. Elsbernd BL, Dunbar KB. Volumetric laser endomicroscopy in Barrett’s esophagus. Tech Innov Gastrointest Endosc 2021;23:P69–P76.
33. Smith MS, Cash B, Konda V, et al. Volumetric laser endomicroscopy and its application to Barrett’s esophagus: results from a 1,000 patient registry. Dis Esophagus 2019;32:doz029.
crossref pmid pmc
34. Trindade AJ, McKinley MJ, Fan C, et al. Endoscopic surveillance of Barrett’s esophagus using volumetric laser endomicroscopy with artificial intelligence image enhancement. Gastroenterology 2019;157:303–305.
crossref pmid
35. Struyvenberg MR, de Groof AJ, Fonollà R, et al. Prospective development and validation of a volumetric laser endomicroscopy computer algorithm for detection of Barrett’s neoplasia. Gastrointest Endosc 2021;93:871–879.
crossref pmid
36. Waterhouse DJ, Januszewicz W, Ali S, et al. Spectral endoscopy enhances contrast for neoplasia in surveillance of Barrett’s esophagus. Cancer Res 2021;81:3415–3425.
crossref pmid pmc pdf
37. Sharma P, Savides TJ, Canto MI, et al. The American Society for Gastrointestinal Endoscopy PIVI (Preservation and Incorporation of Valuable Endoscopic Innovations) on imaging in Barrett’s Esophagus. Gastrointest Endosc 2012;76:252–254.
crossref pmid
38. Pan W, Li X, Wang W, et al. Identification of Barrett’s esophagus in endoscopic images using deep learning. BMC Gastroenterol 2021;21:479.
crossref pmid pmc pdf
39. Ali S, Bailey A, Ash S, et al. A pilot study on automatic three-dimensional quantification of Barrett’s esophagus for risk stratification and therapy monitoring. Gastroenterology 2021;161:865–878.
crossref pmid
40. Beg S, Ragunath K, Wyman A, et al. Quality standards in upper gastrointestinal endoscopy: a position statement of the British Society of Gastroenterology (BSG) and Association of Upper Gastrointestinal Surgeons of Great Britain and Ireland (AUGIS). Gut 2017;66:1886–1899.
crossref pmid pmc
41. Bisschops R, Areia M, Coron E, et al. Performance measures for upper gastrointestinal endoscopy: a European Society of Gastrointestinal Endoscopy (ESGE) Quality Improvement Initiative. Endoscopy 2016;48:843–864.
crossref pmid
42. Wu L, Zhang J, Zhou W, et al. Randomised controlled trial of WISENSE, a real-time quality improving system for monitoring blind spots during esophagogastroduodenoscopy. Gut 2019;68:2161–2169.
crossref pmid pmc
PDF Links  PDF Links
PubReader  PubReader
ePub Link  ePub Link
XML Download  XML Download
Full text via DOI  Full text via DOI
Download Citation  Download Citation
Web of Science
Related articles
As how artificial intelligence is revolutionizing endoscopy  
Application of Artificial Intelligence in Capsule Endoscopy: Where Are We Now?  2018 November;51(6)
Editorial Office
Korean Society of Gastrointestinal Endoscopy
#817, 156 Yanghwa-ro (LG Palace, Donggyo-dong), Mapo-gu, Seoul, 04050, Korea
TEL: +82-2-335-1552   FAX: +82-2-335-2690    E-mail:
Copyright © Korean Society of Gastrointestinal Endoscopy.                 Developed in M2PI
Close layer