Keywords
Colorectal Polyp, colorectal cancer, convolutional neural network, prediction model, computer aided diagnosis
This article is included in the Oncology gateway.
Colorectal Polyp, colorectal cancer, convolutional neural network, prediction model, computer aided diagnosis
Colorectal cancer (CRC) is the third most common cancer worldwide. Among the causes of cancer-related death, CRC ranks second1. Although the biological pathways that transform the normal colonic into malignant tissue are different, polyps are presumed to be the precursor lesions for malignant tumors in all cases2. Colorectal polyps (CRP) can be seen in various forms or shapes on colon endoscopy. CRP are histologically classified into two main categories based on growth pattern, namely serrated and adenomatous (adenomas) polyps2.
A new systematic review and meta-analysis combined 70 primary population-based cross-sectional studies that used colonoscopy for assessing different colorectal neoplastic lesions. This study reported the worldwide prevalence of adenoma, advanced adenoma and CRC as 25.9%, 5.2% and 0.6% in patients older than 50 years, respectively3. Brenner et al. investigated almost 850,000 colonoscopies of German national screening of CRC. They projected that 10-year risk of progression of advanced adenomas to CRC varied between 25% and 40% in men and women older than 50 years of age4.
The role of colonoscopy screening has been proven as an effective strategy for preventing CRC development. This intervention can affect CRC mortality by two pathways; first, polypectomy or removing adenomas and second, detecting early stages of CRC5. Because removal of CRP does not reduce the risk of CRC to zero, current guidelines recommend a 5–10-year rescreening interval after a negative colonoscopy result6,7. Recently, a large-scale multi-center cohort study followed 200,000 persons who underwent baseline colonoscopies for more than eight years. This large cohort study showed that high-risk and low-risk adenomas resulted in 2.6 and 1.3 fold increase, respectively, in the risk of CRC development when compared to persons without adenoma8.
During the last two decades, despite efforts made to screen CRC/CRP in target groups (e.g., population aged more than 50 years old), studies have shown that even conventional colonoscopy by expert gastroenterologists can miss adenomas or pre-cancerous lesions in up to 25% of cases9,10. Over the last few decades, several measures have been developed and recommenced to evaluate the quality of colonoscopy in the diagnosis of polyps and colon cancers11. An important quality measure is the Adenoma Detection Rate (ADR). A recent large-scale investigation on more than 300,000 colonoscopies performed by 136 colonoscopists during 13 years showed that an 1.0% increase in the ADR was associated with a 3.0% decrease in the risk of CRC12.
In recent years, several technologies have been combined with colonoscopy, leading to new endoscopic methods for improving detection rate of colorectal lesions. Image-Enhanced Endoscopy (IEE), cap-assisted colonoscopy, the Third Eye® Retroscope®, wide-angle colonoscope, Endocuff® device, and water-assisted colonoscopy are just a few of these new methods of colonoscopy that are available to augment the detection, diagnosis, and treatment of these subtle lesions13,14.
In the last few years, several adjunct techniques or devices are under investigation for improving ADR in colonoscopy settings. These are methods categorized under the generic term “Computer Aided Diagnosis” (CAD)15. The collection of models such as “Artificial Intelligence”, “Deep Neural Networks” and “Machine Learning” are under CAD. Recent investigations have shown that CAD-methods, along with colonoscopy data, have advantages; higher CRC/CRP detection, better histopathologic differentiation, decreasing overall healthcare costs, and decreasing operator independency15.
Convolutional neural networks (CNNs) are a class of deep neural networks that are highly effective at performing image and video analyses. CAD–CNN models for colonoscopy could assist endoscopists in detecting polyps and performing optical diagnosis. CNNs are trained using thousands of colonoscopy images to identify and differentiate between hyperplastic and adenomatous polyps16.
Since 2018, several systematic reviews have been published to diagnose different clinical outcomes using CNN methods. The assessed outcomes were in skin cancer17, breast cancer18, hepatocellular carcinoma or hepatic mass19, and ischemic brain strokes20.
Based on our knowledge, no systematic review or meta-analysis to evaluate the diagnostic accuracy for discrimination of CRC from CRP has been published. Unfortunately, nearly all systematic reviews assessing the diagnostic accuracy of CNNs in detecting or differentiating different health outcomes, especially different cancers that have been published so far have some limitations. First, the papers had no study protocols and no evidence was provided regarding registration of the protocols. Second, all the published systematic reviews were performed without doing a meta-analysis. Third, in the systematic review methods, risk of bias assessment or critical appraisal was not designed. Therefore, we will aim to design and conduct a systematic review and meta-analysis using a standard and high-quality method to evaluate diagnostic accuracy for discrimination of CRC/CRP by CNN.
This protocol was first designed based on the “priori” approach and then registered in Open Science Framework21.
Assessing the accuracy of CNN model (Intervention: I) for discrimination malignancy (cancer) and/or polyp (Outcome: O) from normal colorectal tissue (Comparison: C) in CRC or CRP probable patients (Participants: P) who attended colonoscopic clinics or centers, based on colonoscopic images or videos.
Type of primary studies. Primary studies will be included as follows: first, all of the study subjects have colonoscopy videos or images (at least one per subject); second, all of the study subjects have pathological or cytological diagnoses (as the gold standard or reference standard); third, the study design is cross-sectional, prospective, or retrospective approach. Therefore, all observational studies with case-control, cohort, or cross-sectional design will be included. Interventional studies (trials, experimental or quasi-experimental), reviews (secondary research), editorials, letters, and similar articles will be excluded.
Type of participants. Since CRPs are usually diagnosed in adults older than 50 years of age, all studies conducted on adult populations of either gender (i.e., patients older than 18 years) will be eligible to be included.
Reference standard. The histopathologic examination results of the colorectal lesions (CRC or CRP) will be used as the reference standard.
Index test (the output of prediction model). CNNs are a supervised learning method. They can learn and find the relationship between the input (images or videos) and the class labels. CNN layers are generally divided into two categories: hidden layers and fully connected layers. The task of the hidden layers is to extract the features. The fully connected layers are used for the classification and detection object in the input images, at the end of the CNN network. The different class labels of all the assessed images or videos should be reported.
We will search the following bibliographic databases: PubMed/MEDLINE, SCOPUS, Web of Science, IEEE (Institute of Electrical and Electronics Engineers), Inspec, ProQuest, Google Scholar, Microsoft Academic Search, ScienceOpen, arXiv, bioRxiv. Moreover, we will assess relevant conferences for content (Conference on Computer Vision and Pattern Recognition, International Conference of Computer Vision, European Conference on Computer Vision). We will search Gastroenterology, Pattern Recognition, Scientific Reports as key journals by hand-searching. The search time interval will cover the 1st of Jan 2010 up to the 31th of July 2020. Our search will not be restricted based on language or geographical area. The PubMed search syntax is provided as Extended data21.
After searching in the sources mentioned above, we will screen all the primary studies based on titles or abstracts. A screening checklist will be developed using 4 to 6 criteria. The criteria will be selected based on the most common components reported in the abstracts of primary research. We will select included or potentially included studies for further assessment. Two reviewers will independently evaluate the considered studies based on full-text papers or documents. We will resolve any disagreement between the two reviewers by consensus.
Two reviewers will independently assess the methodological quality (risk of bias) of the included studies. The risk of bias checklist will be selected based on PROBAST tool22. The checklist has two main domains: risk of bias (ROB) and applicability domains. The ROB domains contain four items: participants, predictors, outcome, analysis. The applicability domain has three items: participants, predictors, and outcome. All seven items will be assessed according to one of three options: low risk of bias, high risk of bias and unclear risk of bias. The overall quality status (overall risk of bias status) will be determined based on defined guidelines (PROBAST guideline). We will resolve any disagreement between the two reviewers by consensus.
We will design an extraction form based on the study objectives. After testing of the extraction form (at least one primary study), we will finalize the data extraction form. Two reviewers will independently extract the required data from all the primary studies (papers or documents). We will resolve any disagreement in the extracted data between the two reviewers by consensus.
The quantitative primary measures are one of two performance indicators: area under curve-receiver operating characteristic curve (AUC-ROC) (C or Concordance Statistics) or accuracy measure. As primary data synthesis (meta-analysis), we will combine the AUC and/or accuracy measures.
The secondary data are the CNN architecture model (VGG23, AlexNet24, GoogLeNet25, etc.) or the features of CNN architecture, such as the number of layers, the size of layers (kernel (filter) and stride size), kind of pooling layers (max or average) and so on, the CNN model sensitivity, specificity, and diagnostic odds ratio.
The subgroup variable data will be transfer learning (existing or absence), learning rate, the features of CNN architecture (number of layers, size of the kernel and stride, and so on).
Combining the primary and secondary data will be performed based on Debray et al guidelines26. We will use Stata version 14.2 (StataCorp. College Station, TX) and R 4.0.0 for conducting the meta-analysis method.
We will use a forest plot for presenting the accuracy measure pooling and I2 (inconsistency) measure and Q Cochrane test for assessing heterogeneity27. We will use the subgroup analysis based on the variables mentioned above. The subgroup analysis or meta-regression will be utilized for finding the potential sources of heterogeneity. The funnel plot, Begg's or Egger's tests28, and Fill and Trim methods29 will be used for evaluating publication or reporting bias.
We will apply sensitivity analysis for assessing the relation between the primary study quality and the accuracy amount, and a one-out remove method will be used too30.
The Grading of Recommendations, Assessment, Development and Evaluation (GRADE) tool to assess the quality of evidence and strength of recommendations was originally developed for interventional systematic review and meta-analysis, but the standard tool was changed later for diagnostic and predictive systematic reviews. However, we will use the modified version of GRADE31 for prediction model systematic review.
We will present the systematic review and meta-analysis findings at relevant conferences. Moreover, we will submit the manuscript to relevant scholarly peer-reviewed journal.
This systematic review is currently in the search phase for information sources.
CNN is a promising advanced method that can be useful in differentiating CRC from benign lesions, such as polyps. The results of this systematic review and meta-analysis will further our knowledge about the role of CNN in predicting colorectal lesions. As far as we know, no comprehensive systematic review has been published so far elaborating the role of CNN in predicting colorectal lesions.
We will try to conduct the study strictly adhering to the methods described. Any changes in the study conduction will be clarified in the final manuscript. Since heterogeneity is a common phenomenon in prediction model studies, we will try to address this concern by appropriate methods such as meta-regression and/or subgroup analysis.
Open Science Framework: Classification of different polyps in colorectal Endoscopy using Convolutional Neural Network; A Systematic Review Protocol Review title and timescale, https://doi.org/10.17605/OSF.IO/QJ7EU21.
This project contains the following extended data:
Repository name: PRISMA-P checklist32 ‘Classifying colorectal cancer or colorectal polyps in endoscopic setting using Convolutional Neural Network (CNN): protocol for a systematic review and meta-analysis’, https://doi.org/10.17605/OSF.IO/QJ7EU21.
Data are available under the terms of the Creative Commons Zero "No rights reserved" data waiver (CC0 1.0 Public domain dedication).
Views | Downloads | |
---|---|---|
F1000Research | - | - |
PubMed Central
Data from PMC are received and updated monthly.
|
- | - |
Is the rationale for, and objectives of, the study clearly described?
Yes
Is the study design appropriate for the research question?
Yes
Are sufficient details of the methods provided to allow replication by others?
Yes
Are the datasets clearly presented in a useable and accessible format?
Yes
Competing Interests: I have already contacted the journal and received permission for reviewing this protocol.
Reviewer Expertise: Systematic review, designing search strategies in academic databases
Alongside their report, reviewers assign a status to the article:
Invited Reviewers | |
---|---|
1 | |
Version 1 02 Sep 20 |
read |
Provide sufficient details of any financial or non-financial competing interests to enable users to assess whether your comments might lead a reasonable person to question your impartiality. Consider the following examples, but note that this is not an exhaustive list:
Sign up for content alerts and receive a weekly or monthly email with all newly published articles
Already registered? Sign in
The email address should be the one you originally registered with F1000.
You registered with F1000 via Google, so we cannot reset your password.
To sign in, please click here.
If you still need help with your Google account password, please click here.
You registered with F1000 via Facebook, so we cannot reset your password.
To sign in, please click here.
If you still need help with your Facebook account password, please click here.
If your email address is registered with us, we will email you instructions to reset your password.
If you think you should have received this email but it has not arrived, please check your spam filters and/or contact for further assistance.
Comments on this article Comments (0)