Skip to main content

Detection of the pathological exposure of pulp using an artificial intelligence tool: a multicentric study over periapical radiographs

Abstract

Background

Introducing artificial intelligence (AI) into the medical field proved beneficial in automating tasks and streamlining the practitioners’ lives. Hence, this study was conducted to design and evaluate an AI tool called Make Sure Caries Detector and Classifier (MSc) for detecting pathological exposure of pulp on digital periapical radiographs and to compare its performance with dentists.

Methods

This study was a diagnostic, multi-centric study, with 3461 digital periapical radiographs from three countries and seven centers. MSc was built using Yolov5-x model, and it was used for exposed and unexposed pulp detection. The dataset was split into a train, validate, and test dataset; the ratio was 8–1-1 to prevent overfitting. 345 images with 752 labels were randomly allocated to test MSc. The performance metrics used to test MSc performance included mean average precision (mAP), precision, F1 score, recall, and area under receiver operating characteristic curve (AUC). The metrics used to compare the performance with that of 10 certified dentists were: right diagnosis exposed (RDE), right diagnosis not exposed (RDNE), false diagnosis exposed (FDE), false diagnosis not exposed (FDNE), missed diagnosis (MD), and over diagnosis (OD).

Results

MSc achieved a performance of more than 90% in all metrics examined: an average precision of 0.928, recall of 0.918, F1-score of 0.922, and AUC of 0.956 (P<.05). The results showed a higher mean of 1.94 for all right (correct) diagnosis parameters in MSc group, while a higher mean of 0.64 for all wrong diagnosis parameters in the dentists group (P<.05).

Conclusions

The designed MSc tool proved itself reliable in the detection and differentiating between exposed and unexposed pulp in the internally validated model. It also showed a better performance for the detection of exposed and unexposed pulp when compared to the 10 dentists’ consensus.

Peer Review reports

Background

Dental caries is a multifactorial disease caused by the demineralization of the tooth's hard tissues and the product of bacterial activity [1]. Consequently, a caries lesion develops, which may lead to pulpal exposure. It is necessary to avoid pulpal exposure since the vitality of the pulp is critical to the tooth's ability to function physiologically in the mouth. In most cases, pulpal inflammation is caused by microbial attacks. As the carious lesion reaches the pulp, the intensity of the inflammatory response rises [2]. The magnitude and severity of pulp-related issues should not be underestimated because they can result in oral sepsis. Consequently, proper diagnosis and management are critical. To determine an appropriate type of restoration and treatment planning, practitioners must have a precise understanding of the status of exposure [2]. 

Furthermore, clinical and radiographic tests are the major criteria for determining the depth and risk of dental caries to prevent pulpal exposure. In addition, the thickness of residual dentine cannot be assessed clinically, which is why a radiographic examination is required to support the clinician's decision [2]. 

However, utilizing radiographs alone to detect pulp exposure is difficult, especially for dentists who lack specialized expertise or time allocated to comprehensive examination. Numerous studies have reported that the reliability and accuracy of identifying dental caries vary greatly depending on the clinician's degree of expertise [3,4,5,6]. 

This is why it is critical to develop methods for determining the pulp status of teeth with deep caries. One of these methods is to employ Artificial Intelligence (AI) [7]. 

Today, AI is beginning to emerge in the healthcare field, reducing diagnostic mistakes in regular practice. Several studies have demonstrated that AI can match or surpass human experts in image-based diagnoses from a variety of medical specialties, including pneumonia in radiology (a convolutional neural network (CNN) trained with labeled frontal chest X-ray images outperformed radiologists in detecting pneumonia), dermatology (a CNN trained with clinical images was found to accurately identify skin lesions), and pathology (a CNN diagnosed heart attack with high accuracy) [8,9,10,11]. 

In addition, AI has piqued the interest of many dental researchers, particularly in dental radiography. Numerous well-written evaluations that presented fundamental principles or radiologist's guides of AI application have been published, attracting more dental researchers to explore its use in dentistry [12,13,14,15]. The current rapid advancement of technology has also spurred the development of numerous AI applications for dental radiography [16, 17]. 

Using AI, such as CNNs, may aid in detecting tiny anomalies in radiographic images that are otherwise difficult to identify with the naked eye. This proposes the possibility of using certain algorithms as automated tools to aid in the detection of deep caries and pulpitis [7]. 

AI can also help detect proximal caries and periapical pathologies that are frequently missed by human eyes on radiographs because of picture noise and/or low contrast [18]. Several researches revealed high-performance results in diagnosing dental caries in radiographies using various image processing approaches followed by machine leaning (ML) classifiers [18,19,20,21,22,23].

Dentist’s workflow became more efficient with automated suggestions for complex cases, better treatment planning, and prediction of diseases and outcomes [2]. 

Hence, this study was conducted to design and evaluate an AI tool called Make Sure Caries Detector and Classifier (MSc) for detecting pathological exposure of pulp on digital periapical radiographs and to compare how correct is the diagnosis between MSc and Dentists. The study was testing the hypothesis to evaluate if the designed AI tool was able to detect exposed/unexposed pulp caries correctly as compared to dentists. MSc was a trial model created by Smile with Confidence (SWC) Company to test the ability of AI to detect exposed and unexposed pulp.

Methods

Study design

This study was designed as a retrospective, diagnostic, and multi-centric study to develop and assess an AI tool for detecting pathological pulp exposure on digital periapical radiographs and comparing how accurate the diagnosis was between MSc and dentists. The reporting in this study follows the checklists for STARD 2015 [24] and Artificial Intelligence in Dental Research [25].

The study protocol was approved by the Institutional Review Board (IRB) in the local committee for ethics of health and scientific research in health affairs in Medina region (IRB 25/2021), Daejeon Dental Hospital, Wonkwang University College of Dentistry, and Complutense University of Madrid, Spain. Informed consent was waived by the IRB in the local committee for ethics of health and scientific research in health affairs in Medina region (IRB 25/2021) due to retrospective nature of the study. All methods were performed in accordance with the Declaration of Helsinki.

Study population

Three thousand four hundred sixty-one anonymized labeled digital periapical radiographs with 3106 exposed pulp caries and 4612 unexposed pulp caries were selected between April 2021 and November 2021 from seven centers in three countries, including Saudi Arabia (Specialized Dental Center, Aohd Dental Center, Alhijra Dental Center, and Faculty of Dentistry, Taibah University), Spain (Faculty of Dentistry, Complutense University of Madrid), and Korea (Faculty of Dentistry Daejeon Dental Hospital). The periapical radiographs included in this research were retrospectively selected by two certified dentists from 18,000 collected periapical radiographs.

The research included all sizes (size 0, 1, and 2 and as JPG and TXT format) of digital periapical radiographs. All carious teeth, including those with periapical or periodontal issues were included. Additionally, radiographs with noticeable caries by human eyes were included (after interrater agreement between two collaborated dentists), whether permanent or deciduous, anterior or posterior teeth, and upper or lower. Radiographs with any number of exposures or carious lesions were included.

Digital periapical radiographs with root caries, restorations (intra-coronal, crowns, and bridges), orthodontic brackets, and wires affecting the interpretation of the carious teeth were excluded. Also, radiographs with more than half of the film missing or unclear or that are difficult to discern due to extreme distortion, artificial noise, blur, and poor image quality were omitted.

Data cleaning and labeling

The data from each hospital or dental center was cleaned and labeled internally by two collaborated, qualified endodontists with more than two years of experience before being sent to the principal investigator (PI) (Specialist of Restorative Dentistry) through an electronic cloud (Google Drive). All radiographs from restorative, endodontic, and pedodontic categories were collected. The PI then revised the cleaning and labeling processes by randomly distributing all collected labeled data to another two dental practitioners with more than two years of experience, who checked and confirmed that all radiographs included dental caries and met our eligibility criteria and then excluded any data that did not. All data was labeled and confirmed when the diagnosis was confirmed by the two dentists; any radiograph on which the interrater disagreed was excluded. At last, the qualified information was sent to the electronic cloud (Google Drive) of the PI.

All confirmed radiograph images were sent to 10 clinicians (ages 25–32, endodontists and general practitioners, two females and eight males, and from the organizations from which data was collected), who manually classified the digital periapical radiographs based on whether the pulp was exposed or not using "Labellmg (Windows_v1.8.0, tzutalin)". This was used to establish the gold standard. A different digital file was sent to each clinician. The annotation was made using boxes.

The PI received the clinicians’ responses as radiographs images in JPG format and TXT files as YOLO format, without knowing their names or contacts (only their titles were known). Clinicians in various hospitals were not given access to each other's data, so they were unfamiliar with one another. All radiographs were coded by sequential numbers. Later, during data processing, the labeled dataset was randomly divided into train (2755), validate (345) (internal validation), and test (345) datasets using Python's random package. The two programmers had full access to all labeled data.

QuestionPro testing platform was used for randomization. All data was randomized by sample randomization method and was masked.

Data processing

Pre-processing procedures were applied using CLAHE to create more contrasted black-and-white images with clipLimit = 5.0 (Fig. 1). The dataset was divided using an 8–1-1 ratio. The test dataset was used to test the MSc and compare its results to those of the other 10 dentists. The validation set, on the other hand, was solely used for validation.

Fig. 1
figure 1

Before and After CLAHE

The MSc model makes use of a variety of CNN algorithm optimization techniques, including auto-learning bounding box anchors, mosaic data augmentation, and the cross-stage partial network. Yolo (You Only Look Once) is an object detection method that it employs. It divided images into cells. Each cell is in charge of detecting objects within it. Yolo processes the entire image with a single neural network, then divides it into sections and predicts the bounding boxes for each section. This algorithm looks at the image once, making predictions after a single forward pass through the neural network. The identified objects are then delivered. Its architecture is primarily composed of three components (Fig. 2):

  • Backbone (CSPDarknet): It is used to extract key features (rich in useful characteristics) from the input image

  • Neck (PANet): A series of layers to mix and combine image features to pass them forward to prediction

  • Head (Output): It is responsible for the final detection step

Fig. 2
figure 2

Yolov5 network architecture

The data is initially sent into the backbone for feature extraction, then into PANet for feature fusion, and lastly into the Yolo layer for output.

The following parameters and hyperparameters were used in our model:

  • Image size: 640

  • Initial weights: yolov5s.pt

  • epoch: 50

  • Learning Rate: 0.001

  • Momentum: 0.999

  • Optimizer: SGD

  • Batch Size: 8

  • IoU threshold: 0.5

  • Device: Running on GPU with Cuda from AWS (p3.2xlarge)

Study groups

The randomized test dataset included 345 images with 752 labels and was used to assess MSc's performance. The data was also utilized to compare the diagnostic disparities between 10 qualified dentists with more than two years of experience and the MSc tool. The data for the dentist groups was randomly divided into 10 blocks, with each block including 34 to 35 radiographs. Data was labeled at various intervals since it was gathered at different times. The reference standards were provided to the performers when using MSc.

Reference test

MSc metrics

Primary

Mean Average Precision 0.5 (mAP@0.5): mAP, calculated by taking the mean AP (accuracy of our AI tool) over all exposed and unexposed pulp caries and/or overall 0.5 (IoU) thresholds

Secondary

  • Precision (Specificity): The ratio of correctly predicted positive exposed/unexposed pulp caries to the total predicted exposed/unexposed pulp caries: Prec. = TP/ TP + FP

  • Recall (Sensitivity): Calculates how many actual exposed/unexposed pulp caries true positives the model has captured, labeling them as positives. Recall = TP/TP + FN

  • F1 Score: Defined as the function of precision and recall. It is calculated when a balance between precision and recall is needed. F1 = 2 × Precision × Recall/ Precision + Recall

  • AUC: Area under the receiver operating characteristic (ROC) curve (AUC). AUC integrated from (0, 0) to (1, 1) gave the aggregate measure of all possible exposed and unexposed pulp caries detection and classification thresholds

Diagnostics metrics

Primary

  • Correct Diagnosis: Overall diagnostic accuracy (true negatives + true positives)

Secondary

  • Wrong diagnosis: The sum of False Diagnosis Exposed (FDE), which is all exposed pulp caries that were diagnosed as unexposed pulp caries based on the gold standard, False Diagnosis Not Exposed (FDNE), which is all unexposed pulp caries that were diagnosed as exposed pulp caries based on the gold standard, Over Diagnosis (OD, Exposed\Unexposed), which is all non-carious objects that were diagnosed as exposed pulp caries or unexposed pulp caries based on the gold standard, and Missed Diagnosis (MD, Exposed\Unexposed), which is all exposed\unexposed pulp caries that were undiagnosed based on the gold standard.

Sample size calculation

According to the results of Lee et al., study [3],  which aimed to estimate optimal deep CNN algorithm weight factors for training and validation dataset of both carious and non-carious molars and premolars teeth, at diagnostic accuracy 82.0%, sensitivity 81.0%, specificity 83.0%, PPV 82.7%, and NPV 81.4%, and with an alpha error of 5% and a confidence interval of 95%, a sample size of 3,000 periapical radiographs in total were chosen. To achieve higher diagnostic performance metrics, the teeth were not classified based on tooth position, and 3,445 digital periapical radiographs were selected in this study.

Statistical analysis

The descriptive metrics of MSc model performance presented as the percentage of mAP @0.5, Precision, Recall, F1 score, and AUC of the test dataset were calculated using the Keras library on top of TensorFlow "Yolo v5" in Python. On the other hand, the mean, frequency distribution, median, and range of the diagnostics measures (RDE, FDE, RDNE, FDNE, OD, MD, correct diagnosis, and wrong diagnosis) were calculated to compare both the performance of the MSc model and 10 certified dentists.

The Wilcoxon Signed Ranks Test was used to determine the significant difference between the values (RDE, FDE, RDNE, FDNE, OV, MD, correct diagnosis, and wrong diagnosis) for both the dentists over test dataset and MSc over the test dataset. The statistical significance was set at p < 0.05.

Results

The total data assessed for eligibility consisted of 18,000 images; however, 14,539 images were excluded for not meeting the inclusion criteria and for technical reasons. Most images were excluded because they were caries free; therefore, only 3461 images were randomized. In addition, further 16 images were excluded because of inter-rater disagreement; hence, 3445 images were labeled and annotated (7718 labels in total). The data were split in a ratio of (8:1:1) for train 2491 (exposed), test (315 exposed), and validate (300 exposed), respectively. The train set consisted of 2,755 images (6,171 labels), the validation set consisted of 345 images (795 labels), and the test set comprised of 345 images (752 labels). Lastly, the test dataset was used for analyzing MSc group and dentists group. (Tables 1, 2) (Fig. 3).

Table 1 Data and labelled distribution
Table 2 Data Evaluation
Fig. 3
figure 3

Flow Diagram

MSc performance metrics

The number of true positive (TP) exposed and unexposed pulp caries that were detected by MSc over the test dataset is 691 labels. The number of false positive (FP) exposed and unexposed pulp caries that were detected by MSc over the test dataset is 56 labels, and the number of false negative (FN) exposed and unexposed pulp caries that were detected by MSc over the test dataset is 61 labels. Therefore, the obtained results of our model showed an mAP < 0.5 of 95.6%, precision of 92.8%, a recall of 91.8%, and an F1 score of 92.2% (P > 0.05) (Table 3). The AUC value was 0.956. (P > 0.05) (Figs. 4 and 5).

Table 3 MSc performance metrics
Fig. 4
figure 4

AUC for Precision-Recall curve

Fig. 5
figure 5

Confusion Matrix for the test dataset

Diagnostics measures

The mean value of RDE, RDNE, OD, and RD for the MSc group was higher than that of the dentists group. Meanwhile, the mean value for the FDE, FDNE, MD, and wrong diagnosis for the MSc group was lower than that of the dentists group. The mean and median values are listed in (Table 4).

Table 4 Dentists versus Msc’s diagnosis

Discussion

The current study designed and analyzed an AI tool named MSc to detect pathological pulp exposure on digital periapical radiographs, indicating that MSc diagnoses exposed/unexposed pulp caries more accurately than dentists. This is considered the first study to test the use of AI in detecting exposed and unexposed pulps. It's important to note that the tool isn't commercially accessible; it's an experimental model designed to evaluate AI's ability to recognize exposed and unexposed pulp.

It is challenging to diagnose dental caries using radiographs alone, especially for the dentists who do not have enough specialized training or time devoted for detailed diagnosis; the reasons are the settings of various parameters, including brightness, shadow, and contrast [3].

Nowadays, AI is becoming essential in radiology due to its ability to detect abnormalities in radiographic images unnoticed by the naked human eye [26]. However, many people, including clinicians and scientists, are not yet familiar with the concepts and true potential of AI and the impact that it can have on our personal and professional lives [27]. 

It can help in standardizing dental caries according to caries depth [7]. Caries detection in dental radiographs might be achieved by evaluating the radiodensity of images. The enamel and dentin of normal teeth are radiopaque. Caries cause these structures to lose their mineralization and hence become radiolucent [18]. Using these characteristics, AI prediction capability enables AI tools to "learn" to analyze dental radiographs and predict pulp status.

This is why we developed our MSc tool: to compare its accuracy in detecting exposed and unexposed pulps to that of dentists. Yolov5x, an object detection method, was used in this tool. We used 10 examiners from various hospitals and locations of the world in our investigation. The dentists who participated in the comparison were all qualified specialists, and their results served as the "gold standard." The inclusion of dentists from various hospitals and places of the world assured that diverse experiences in defining the gold standard would result in a broader picture of the many MSc outcomes [28].

However, this was not in line with a systematic review [29] that showed that trials with only one examiner yielded the best results, indicating that the same caries detection criteria were always utilized [18]. Also, another study used four experts to analyze the photos, which was considered the second best [3].  Finally, the trial with two examiners produced the worst accuracy result [30, 31].

The number of images (3445) used in our study was also comparable with that of Cantu et al., who used 3292 bitewing radiographs to detect carious exposure. When these data were compared to the technique and findings of the current investigation, the number of images employed and the verified diagnostic performance were nearly similar [32].

In the present study, MSc showed a high mAP (0.956) at 50% Confidence level in detecting exposed and unexposed pulp, and this is considered outstanding. In addition, with an AUC of 0.956 for all classes in our trial, MSc’s performance is considered exceptional since an AUC value of 0.8 to 0.9 is excellent, and more is exceptional [33].

MSc also showed statistically significant outcomes in detecting exposed and unexposed pulp caries accurately and with less wrong detections in comparison with the dentists group. This means that the model yielded better scores than the dentists. Moreover, the model seemed to be more effective and reliable than the dentists, since the 10 experienced dentists did not show good consistency and stability, and the model was much faster and more accurate in lesion classification. However, further research needs to be conducted with a larger dataset and different experienced dentists for more reliable results.

This was in line with a study [3] that used deep CNN-based computer-vision for dental caries detection, and the results were 89.0%, 88.0%, and 82.0% for diagnostic accuracies of premolar, molar, and both premolar and molar models, respectively. The deep CNN algorithm achieved an AUC of 0.917 on premolar, an AUC of 0.890 on molar, and an AUC of 0.845 on both premolar and molar models. In another study conducted by Kühnisch et al., the CNN accurately detected cavities in 92.5% of instances (SE, 89.6; SP, 94.3; AUC, 0.964). This was also similar to the finding in other studies [28, 32].

Another study in India evaluated the diagnostic performance of an algorithm where a neural network is developed to detect dental caries in digital radiographs. The system gave an accuracy of 97.1%, FP rate of 2.8%, ROC area of 0.987 and precision recall curve (PRC) area of 0.987. This is considered high accuracy compared to our study, but the authors only employed 105 photos and performed tenfold cross-validation; their dataset wasn’t enough [18]. 

Other accuracy comparisons with studies that used different AI methods were as follows: Singh et al. suggested a caries detection method based on Radon Transformation and DCT employing dental X-ray images [23].  Selected characteristics are retrieved using the PCA approach and used to the Random Forest classifier, yielding an accuracy of 86%.

During another study, [21] a caries detection system utilized SVM and got 86.15% accuracy for the training dataset and 77.34%accuracy for the test dataset.

A study [34] developed CNN-based categorization of major dental illnesses, which achieved an accuracy of 87.5% in detecting dental caries.

Additionally, a similar study [35] assessed the detection of carious exposure using AI on bitewings, and the neural network had an accuracy of 80%, while dentists had a considerably lower mean accuracy of 71%. The neural network demonstrated significant sensitivities of 70% or higher for both initial and advanced lesions. Dentists' sensitivities for initial lesions were generally low while those for advanced lesions ranged between 40 and 75%. Authors suggested that this was due to the dentists being prudent in their decisions, unlike the AI tool. When these data were compared to the technique and findings of the current investigation with MSc, the number of images employed and the verified diagnostic performance were nearly comparable. Nevertheless, our MSc model had higher mean in the over diagnosis variable, which means higher probability for detecting non carious objects as exposed or unexposed pulp caries. That can be viewed as limitation for AI, as it can consider some radiographic artificial changes as carious lesions. This is why it is important to note that the final diagnosis needs to be confirmed by the clinician as our tool is only a decision support system.

We also had an overfitting problem during our study. Overfitting is a crucial aspect of AI. "The essence of overfitting," according to Burnham and Anderson, "is to have unwittingly retrieved part of the residual variation as if that variation reflected underlying model structure" [36]. When a model is overfitted, it is so particular to the original data that applying it to data obtained in the future might result in problematic or incorrect outcomes, and so less-than-optimal judgments [37]. However, we tried to resolve this by increasing the number of training dataset.

There were also more limitations in our study. Clinical parameters were not included, which is an aspect that should be taken into account to have a more accurate diagnosis. Also, neural networks, in general, including our tool MSc, are black boxes that cannot explain machine learning characteristics and the grounds for making decisions based on that learning. The limitations of the digital periapical radiographs, such as image magnification and distortion and the lack of three-dimensional information, may lower the MSc tool's diagnostic accuracy. This is because it is challenging to diagnose critical cases, such as nearly exposed pulp with a small layer of dentine. Therefore, it is necessary to take clinical assessment into consideration. Also, the dataset was not divergent regarding the age and sex because it was collected without prior knowledge of the patients’ details. We also did not categorize the teeth according to their types, which may have affected the accuracy of the results according to the teeth type. Finally, further research needs to be conducted with a larger dataset and different experienced dentists for more reliable results.

Future initiatives for improving AI-based caries diagnosis on intraoral pictures should involve image segmentation as an alternate option, which should be carried out by well-trained and calibrated dental practitioners under the supervision of senior specialists. To accomplish this, caries lesions must be marked pixel by pixel on each accessible image and the diagnosis accuracy must be reassessed. In comparison to the currently utilized classification methodology, this more precise but otherwise time- and resource-intensive approach provides thorough caries localization.

Conclusions

From the above results and discussion, it is concluded that:

  1. 1.

    The designed AI model proved itself reliable in the detection of pathological pulp exposure, and in the differentiation between exposed and unexposed pulp caries on digital periapical radiographs.

  2. 2.

    The designed AI model detected pathological pulp exposure on digital periapical radiographs more correctly and effectively than the10 dentists.

Availability of data and materials

Part of the test dataset is included in the supplementary files to allow the replication of the results.

MSc was a trial model created by Smile with Confidence (SWC) Company to test the ability of AI to detect exposed and unexposed pulp. We are showing the results of a tool that is not available commercially to show that AI could be used to support the work of dentists. As a company specialized in AI, we were testing the power of this technology.

References

  1. Gil AMC, Gispert AEÁ. General remarks about minimal intervention dentistry. Rev Cubana Estomatol. 2016;53(2):37–44.

  2. Leite AF, Vasconcelos KF, Willems H, Jacobs R. Radiomics and machine learning in oral healthcare. Proteomics Clin Appl. 2020;14(3):e1900040.

  3. Lee JH, Kim DH, Jeong SN, Choi SH. Detection and diagnosis of dental caries using a deep learning-based convolutional neural network algorithm. J Dent. 2018;77:106–11.

  4. Estay J, Bersezio C, Arias R, Fernández E, Oliveira Junior OB, Ferrarezi de Andrade M. Effect of Clinical Experience on Accuracy and Reliability of Radiographic Caries Detection. Int J Odontostomatol. 2017;3:347–52.

    Article  Google Scholar 

  5. Diniz MB, Rodrigues JA, Neuhaus KW, Cordeiro RC, Lussi A. Influence of examiner's clinical experience on the reproducibility and accuracy of radiographic examination in detecting occlusal caries. Clin Oral Investig. 2010;14(5):515–23.

  6. Shams N, Panahandeh N, Aghababa H, Shams B, Hemati E. Effects of education and experience on detection of proximal caries on digital radiographs. Oral Radiol. 2015;32(3):154–9.

  7. Zheng L, Wang H, Mei L, Chen Q, Zhang Y, Zhang H. Artificial intelligence in digital cariology: a new tool for the diagnosis of deep caries and pulpitis using convolutional neural networks. Ann Transl Med. 2021;9:763.

    Article  PubMed  PubMed Central  Google Scholar 

  8. Esteva A, Robicquet A, Ramsundar B, Kuleshov V, DePristo M, Chou K, et al. A guide to deep learning in healthcare. Nat Med. 2019;1:24–9.

    Article  Google Scholar 

  9. Bejnordi B Ehteshami, Veta M, van P Johannes Diest, van B Ginneken, Karssemeijer N, Litjens G, et al. Diagnostic Assessment of Deep Learning Algorithms for Detection of Lymph Node Metastases in Women With Breast Cancer. JAMA. 2017;318(22):2199–210.

    Article  Google Scholar 

  10. Wang X, Peng Y, Lu L, Lu Z, Bagheri M, Summers RM. Chestx-ray8: hospital-scale chest x-ray database and benchmarks on weakly-supervised classification and localization of common thorax diseases. Honolulu: IEEE Conference on Computer Vision and Pattern Recognition (CVPR); 2017. p. 3462–71.

  11. Strodthoff N, Strodthoff C. Detecting and interpreting myocardial infarction using fully convolutional neural networks. Physiol Meas. 2019;1:15001.

    Article  Google Scholar 

  12. Thrall JH, Li X, Li Q, Cruz C, Do S, Dreyer K. Artificial Intelligence and Machine Learning in Radiology: Opportunities, Challenges, Pitfalls, and Criteria for Success. J Am Coll Radiol. 2018;15(3 Pt B):504–8.

    Article  PubMed  Google Scholar 

  13. Bluemke DA, Moy L, Bredella MA, Ertl-Wagner BB, Fowler KJ, Goh VJ, Halpern EF, Hess CP, Schiebler ML, Weiss CR. Assessing radiology research on artificial intelligence: a brief guide for authors, reviewers, and readers-from the radiology editorial board. Radiology. 2020;294(3):487–9.

  14. Do S, Song KD, Chung JW. Basics of Deep Learning: A Radiologist’s Guide to Understanding Published Radiology Articles on Deep Learning. Korean J Radiol. 2020;1:33–41.

    Article  Google Scholar 

  15. Soffer S, Ben-Cohen A, Shimon O, Amitai MM, Greenspan H, Klang E. Convolutional Neural Networks for Radiologic Images: A Radiologist’s Guide. Radiology. 2019;3:590–606.

    Article  Google Scholar 

  16. Hung K, Montalvao C, Tanaka R, Kawai T, Bornstein MM. The use and performance of artificial intelligence applications in dental and maxillofacial radiology: A systematic review. Dentomaxillofac Radiol. 2020;1:20190107.

    Article  Google Scholar 

  17. Nagi R, Aravinda K, Rakesh N, Gupta R, Pal A, Mann AK. Clinical applications and performance of intelligent systems in dental and maxillofacial radiology: A review. Imaging Sci Dent. 2020;2:81–92.

    Article  Google Scholar 

  18. Geetha V, Aprameya KS, Hinduja DM. Dental caries diagnosis in digital radiographs using back-propagation neural network. Heal Inf Sci Syst. 2020;1:8.

    Article  Google Scholar 

  19. Devito KL, de Souza BF, Felippe Filho WN. An artificial multilayer perceptron neural network for diagnosis of proximal dental caries. Oral Surg Oral Med Oral Pathol Oral Radiol Endod. 2008;6:879–84.

    Article  Google Scholar 

  20. Yu Y, Li Y, Li Y-J, Wang JM, Lin DH, Ye WP. Tooth decay diagnosis using back propagation neural network. Dalian: In proceedings of the 2006 international conference on machine learning and cybernetics; 2006. p. 3956–9.

  21. Li W, Kuang W, Li Y, Li YJ, Ye WP: Clinical X-ray image based tooth decay diagnosis using SVM. Proceedings of international conference on machine learning and cybernetics. 2007:1616–19.

  22. Ali RB, Ejbali R, Zaied M. Detection and classification of dental caries in x-ray images using deep neural networks. International conference on software engineering advances. 2016:223–7.

  23. Singh P, Sehgal P. Automated caries detection based on radon transformation and DCT. 2017 8th international conference on computing, communication and networking technologies (ICCCNT). 2017:1–6.

  24. Bossuyt PM, Reitsma JB, Bruns DE, Bruns DE, Glasziou PP, Irwig L, et al. STARD 2015: An updated list of essential items for reporting diagnostic accuracy studies1. Radiology. 2015;3:826–32.

    Article  Google Scholar 

  25. Schwendicke F, Singh T, Lee JH, Gaudin R, Chaurasia A, Wiegand T, Uribe S, Krois J. IADR e-oral health network and the ITU WHO focus group AI for Health. Artificial intelligence in dental research: Checklist for authors, reviewers, readers. J Dent. 2021;107:103610.

  26. Kim JE, Nam NE, Shim JS, Jung YH, Cho BH, Hwang JJ. Transfer learning via deep neural networks for implant fixture system classification using periapical radiographs. J Clin Med. 2020;9(4):1117.

  27. Sur J, Bose S, Khan F, Dewangan D, Sawriya E, Roul A. Knowledge, attitudes, and perceptions regarding the future of artificial intelligence in oral radiology in India: A survey. Imaging Sci Dent. 2020;3:193.

    Article  Google Scholar 

  28. Lian L, Zhu T, Zhu F, Zhu H. Deep learning for caries detection and classification. Diagnostics (Basel). 2021;11(9):1672.

  29. Prados-Privado M, Villalón JG, Martínez-Martínez CH, Ivorra C, Prados-Frutos JC. Dental caries diagnosis and detection using neural networks: A systematic review. J Clin Med. 2020;11:1–13.

    Google Scholar 

  30. Schwendicke F, Samek W, Krois J. Artificial Intelligence in Dentistry: Chances and Challenges. J Dent Res. 2020;7:769–74.

    Article  Google Scholar 

  31. Berdouses ED, Koutsouri GD, Tripoliti EE, Matsopoulos GK, Oulis CJ, Fotiadis DI. A computer-aided automated methodology for the detection and classification of occlusal caries from photographic color images. Comput Biol Med. 2015;119–35.

  32. Kühnisch J, Meyer O, Hesenius M, Hickel R, Gruhn V. Caries detection on intraoral images using artificial intelligence. J Dent Res. 2022;101(2):158–65.

  33. Mandrekar JN. Receiver operating characteristic curve in diagnostic test assessment. J Thorac Oncol Off Publ Int Assoc Study Lung Cancer. 2010;9:1315–6.

    Google Scholar 

  34. Prajapati SA, Nagaraj R, Mitra S. Classification of dental diseases using CNN and transfer learning. Proceedings of the 2017 5th International Symposium on Computational and Business Intelligence (ISCBI), Dubai, United Arab Emirates. 2017:70–4.

  35. Cantu AG, Gehrung S, Krois J, Chaurasia A, Rossi JG, Gaudin R, Elhennawy K, Schwendicke F. Detecting caries lesions of different radiographic extension on bitewings using deep learning. J Dent. 2020;100:103425.

  36. Burnham KP, Anderson DR. Model Selection and Multimodel Inference. New York, NY, USA: Springer; 2004. (978-0-387-95364-9.

    Book  Google Scholar 

  37. Mutasa S, Sun S, Ha R. Understanding artificial intelligence based radiology studies: What is overfitting? Clin Imaging. 2020;65:96-9.

Download references

Acknowledgements

Not applicable

Funding

This study was funded by SWC Company.

Author information

Authors and Affiliations

Authors

Contributions

All authors contributed to the conception, data acquisition and interpretation, performed the statistical analysis and drafted and critically revised the manuscript.

Corresponding author

Correspondence to A. Altukroni.

Ethics declarations

Ethics approval and consent to participate

The study protocol was approved by the IRB in the local committee for ethics of health and scientific research in health affairs in Medina region (IRB 25/2021), Daejeon Dental Hospital, Wonkwang University College of Dentistry, and Complutense University of Madrid Informed consent was waived by the IRB in the local committee for ethics of health and scientific research in health affairs in Medina region (IRB 25/2021) due to retrospective nature of the study. All methods were performed in accordance with the Declaration of Helsinki.

Consent for publication

Not applicable.

Competing interests

The authors declare no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Additional file 1:

Test Dataset 1: Digital radiograph of upper posterior teeth. Test Dataset 2: Digital radiograph of upper posterior teeth, Test Dataset 3: Digital radiograph of upper posterior teeth, Test Dataset 4: Digital radiograph of upper posterior teeth, Test Dataset 5: Digital radiograph of upper anterior teeth, Test Dataset 6: Digital radiograph of upper anterior teeth, Test Dataset 7: Digital radiograph of lower posterior teeth, Test Dataset 8: Digital radiograph of upper posterior teeth, Test Dataset 9: Digital radiograph of lower anterior teeth, Test Dataset 10: Digital radiograph of lower anterior teeth, Test Dataset 11: Digital radiograph of lower posterior teeth, Test Dataset 12: Digital radiograph of lower anterior teeth, Test Dataset 13: Digital radiograph of upper posterior teeth, Test Dataset 14: Digital radiograph of lower teeth, Test Dataset 15: Digital radiograph of lower deciduous teeth, Test Dataset 16: Digital radiograph of lower deciduous teeth, Test Dataset 17: Digital radiograph of lower posterior teeth, Test Dataset 18: Digital radiograph of lower deciduous posterior teeth, Test Dataset 19: Digital radiograph of upper posterior teeth, Test Dataset 20: Digital radiograph of lower posterior teeth, Test Dataset 21: Digital radiograph of lower posterior teeth, Test Dataset 22: Digital radiograph of upper posterior teeth, Test Dataset 23: Digital radiograph of upper posterior teeth, Test Dataset 24: Digital radiograph of lower posterior teeth, Test Dataset 25: Digital radiograph of upper posterior teeth, Test Dataset 26: Digital radiograph of lower deciduous posterior teeth, Test Dataset 27: Digital radiograph of lower deciduous posterior teeth, Test Dataset 28: Digital radiograph of lower posterior teeth, Test Dataset 29: Digital radiograph of lower posterior teeth, Test Dataset 30: Digital radiograph of upper deciduous posterior teeth, Test Dataset 31: Digital radiograph of upper anterior teeth, Test Dataset 32: Digital radiograph of lower deciduous posterior teeth, Test Dataset 33: Digital radiograph of upper anterior teeth, Test Dataset 34: Digital radiograph of lower deciduous posterior teeth, Test Dataset 35: Digital radiograph of lower posterior teeth, Test Dataset 36: Digital radiograph of lower deciduous posterior teeth, Test Dataset 37: Digital radiograph of lower deciduous posterior teeth, Test Dataset 38: Digital radiograph of lower posterior teeth, Test Dataset 39: Digital radiograph of upper posterior teeth, Test Dataset 40: Digital radiograph of lower anterior teeth, Test Dataset 41: Digital radiograph of upper anterior teeth, Test Dataset 42: Digital radiograph of lower posterior teeth, Test Dataset 43: Digital radiograph of lower posterior teeth, Test Dataset 44: Digital radiograph of upper posterior teeth, Test Dataset 45: Digital radiograph of upper anterior teeth, Test Dataset 46: Digital radiograph of upper posterior teeth, Test Dataset 47: Digital radiograph of lower deciduous posterior teeth, Test Dataset 48: Digital radiograph of upper deciduous posterior teeth, Test Dataset 49: Digital radiograph of lower posterior teeth, Test Dataset 50: Digital radiograph of lower deciduous posterior teeth. 

Additional file 2:

STARD 2015 Guidelines.

Additional file 3:

Artificial intelligence in dental research: Checklist for authors, reviewers, readers.

Additional file 4.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Altukroni, A., Alsaeedi, A., Gonzalez-Losada, C. et al. Detection of the pathological exposure of pulp using an artificial intelligence tool: a multicentric study over periapical radiographs. BMC Oral Health 23, 553 (2023). https://doi.org/10.1186/s12903-023-03251-0

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12903-023-03251-0

Keywords