Skip to main content

Caries detection with tooth surface segmentation on intraoral photographic images using deep learning

Abstract

Background

Intraoral photographic images are helpful in the clinical diagnosis of caries. Moreover, the application of artificial intelligence to these images has been attempted consistently. This study aimed to evaluate a deep learning algorithm for caries detection through the segmentation of the tooth surface using these images.

Methods

In this prospective study, 2348 in-house intraoral photographic images were collected from 445 participants using a professional intraoral camera at a dental clinic in a university medical centre from October 2020 to December 2021. Images were randomly assigned to training (1638), validation (410), and test (300) datasets. For image segmentation of the tooth surface, classification, and localisation of caries, convolutional neural networks (CNN), namely U-Net, ResNet-18, and Faster R-CNN, were applied.

Results

For the classification algorithm for caries images, the accuracy and area under the receiver operating characteristic curve were improved to 0.813 and 0.837 from 0.758 to 0.731, respectively, through segmentation of the tooth surface using CNN. Localisation algorithm for carious lesions after segmentation of the tooth area also showed improved performance. For example, sensitivity and average precision improved from 0.890 to 0.889 to 0.865 and 0.868, respectively.

Conclusion

The deep learning model with segmentation of the tooth surface is promising for caries detection on photographic images from an intraoral camera. This may be an aided diagnostic method for caries with the advantages of being time and cost-saving.

Peer Review reports

Background

Dental caries is one of the most common infectious diseases globally that can cause oral pain, infection, and even tooth loss without proper treatment [1,2,3]. Untreated caries can be a biological, social, and financial burden both for the individual and society as a whole [4]. However, it is usually not life-threatening; therefore, many patients visit the dental clinic for treatment when the caries are at an advanced stage and serious complications, which are expensive or difficult to treat, have already occurred.

Early detection of dental caries reduces the treatment costs and time. It also conforms to the concept of minimally invasive dentistry by avoiding aggressive treatments, such as root canal treatment or tooth extraction [5]. Thus, the early detection of dental caries is required, and this warrants a method that could accurately, easily, and rapidly diagnose dental caries at an early stage.

Conventional examination for caries detection is primarily performed by visual inspection, tactile sensation, and radiography [6], which are clinical evaluations. However, oral health factors are unequal worldwide, and there are still many people in poor condition with limited availability and accessibility to dental professionals [4]. Therefore, there is a need to develop a more convenient and cost-effective technique for early detection of carious lesions. Moreover, there is variability in the ability of dentists to detect caries; therefore, it would be good if a standardised diagnostic method could assist the practitioners [7].

In the field of dentistry, novel techniques beneficial for patients and clinicians are being developed. One such device is an intraoral camera (IOC), which is inexpensive, easy to operate with digital storage, and capable of capturing high-quality images [8, 9]. It has also been shown that caries can be identified simply and reliably using the IOC [10, 11]. Recently, some studies have suggested the application of deep learning algorithms in the diagnosis of caries using oral photographic images [12, 13].

Deep learning (DL) is a subset of machine learning in artificial intelligence (AI) and is based on deep neural networks that consist of multiple hidden layers to progressively learn representations from raw data. In DL, a convolutional neural network (CNN or ConvNet) has demonstrated excellent performance in computer vision and has been most commonly applied to analyse visual imagery [14, 15]. Recently, CNNs have been rapidly emerging in the medical field and have demonstrated excellent performance in computer vision, including object, facial, and activity recognition, tracking, and three-dimensional mapping and localisation [16]. In the dental field, several studies have applied CNNs to detect carious lesions on periapical radiographs [17], radiovisiography [18], and oral photographs [19, 20].

The detection of dental caries from intraoral photographs using DL is inexpensive and accessible, allowing advanced oral healthcare conditions. Furthermore, this computer-aided diagnostic (CAD) technique is a reliable and standardised assistant. However, studies on CNN models for the detection of dental caries in oral photographs are still limited. Zhang et al. showed that DL using oral photographs captured with smartphones was useful for screening dental caries. However, some oral areas are difficult to capture owing to their location; therefore, smartphones are not appropriate for clinical use in professional dental diagnosis [19]. Kühnisch et al. demonstrated the high accuracy of a CNN model, using oral photographs including only one tooth captured by a digital single-lens reflex (DSLR) camera, in caries detection. Also, the photographs were strictly selected according to methodological requirements, and the images similar to caries, such as tooth discoloration or abrasion, were excluded from the datasets; therefore, limitations in daily use in dental clinics may be present [20]. However, relatively large area including several teeth can be captured in one intraoral photograph frequently, which may affect accuracy of CNN model. In case of dental panoramic images, there were some trials to detect carious lesion better by focusing tooth surfaces through image preprocessing with segmentation [21]. Segmentation task by CNN is a kind of classification at every pixel of the input image and can discriminate different anatomical structures in medical images [22]. However, in case of intraoral photograph, there was little trials to separate tooth area from background by segmentation for caries detection. Therefore, it is necessary to develop CNN algorithms to detect caries among photographs including various clinical conditions such as developmental tooth defect, discoloration, or dental restorations which can be obtained in dental clinics generally along with applying segmentation for tooth surfaces to improve accuracy.

Finally we evaluated the performance of the CNN algorithms in the detection of dental caries using photograph, which can generally be taken by IOC in dental clinics. Moreover, to improve the performance in the detection of caries, we pre-processed the photographic images by segmenting the tooth surface using the CNN algorithm.

Methods

Study protocol and dataset

The study procedure was approved by the Institutional Review Board of the Kyungpook National University (KNU-2021-0097) and was performed in accordance with the Declaration of Helsinki. This study followed the guidelines of the Standards for Reporting of Diagnostic Accuracy Studies (STARD). Participants were recruited from patients who visited the dental clinic of the university hospital, after explaining the purpose and process of the study according to the approved study protocol, from October 2020 to December 2021. Among those, patients who have at least one permanent tooth, and could cooperate with taking photographs using an intraoral camera were selected, and finally 445 patients who provided written informed consents were enrolled. A total of 2348 RGB intraoral photographic images were taken by an experienced dentist using a professional intraoral camera (Qraypen, AIOBIO, Seoul, Republic of Korea) after oral examination. For oral examination, dental mirror and explorer were used carefully under dental light in a unit chair with an air syringe by three dentists. Photographs which have poor image quality for dentist to diagnose and make ground truth, such as blurred, unintended, or duplicated images, were not included in dataset. However, to include general photographic images that can be normally acquired at general dental clinics, there was no criteria for the number of teeth or amount of saliva in the photographic image if the diagnosis of caries was possible. Similarly, various non-carious defects, such as tooth stain, hypomineralisation, tooth wear, or dental restoration, were included in the dataset. All images with a resolution of 1280 × 720 pixels were used for analysis without additional processing.

Annotation for tooth surface and caries

Each photographic image was examined to detect carious lesion on a personal computer monitor by a board-certified dentist, who is an expert in the epidemiological oral examination conducted by the government agency in Korea. First, all tooth surfaces in each image were labelled pixel-wise by the dentist using a tagging tool (https://supervise.ly) as a reference standard for tooth surface segmentation. Second, ground truth for carious lesion was made on tooth surface using bounding box according to International Caries Detection and Assessment System (ICDAS) considering each clinical chart record [23]. If the carious lesion was separated by a sound tooth surface on one tooth, each bounding box was created independently using the tagging tool mentioned above. Therefore, according to the carious lesion, each photographic image had none or more than one bounding box. To minimise ambiguousness of caries diagnosis using photographic images only, distinct caries classified as codes 4, 5, or 6 according to the ICDAS (ICDAS Code 4: An underlying dark shadow from dentin with or without localised enamel breakdown, Code 5: Distinct cavity with visible dentin, Code 6: Extensive distinct cavity with visible dentin) were annotated as caries cases, which may be closely related to the necessity of dental treatment clinically [23, 24].

Deep learning algorithm application for tooth surface segmentation caries detection

Among the 2348 intraoral photographic images obtained from 445 patients during actual dental treatment, 998 images showing caries symptoms and 1350 healthy teeth without caries symptoms were divided. The number of local carious lesions found in the 998 images where caries was observed was 1999. This dataset is divided into three subsets: training, validation, and testing. 150 carious images corresponding to 15% of about 1000 carious images were first assigned to the test set, and the remainder were divided into a training set and an evaluation set at a ratio of 4:1. Among the collected data, since the number of normal images is larger than that of caries images, the number of test sets was set to 150, which is the same as that of caries images, to balance the classification model evaluation. For the remaining data, the training set and the evaluation set were divided at a ratio of 4:1. In the end, the dataset contained 1638 (69.8%), 410 (17.5%), and 300 (12.8%) images for training, validation, and testing (Table 1). Both the caries image and the normal image are used in the dataset for training or evaluation of the classification model and the segmentation model, and only the caries-containing image is used for the caries detection model.

Table 1 Dataset composition of intraoral photographic images

Images assigned to the testing subset of the CNN model were never used in the training or validation step for the independent evaluation of the deep learning model. To predict the tooth surface, U-Net was used, which is a CNN for biomedical image segmentation and can be used to extract objects from images [25]. Segmentation of the tooth surface improves the overall caries detection performance by darkening areas not classified as tooth surfaces in each image. ResNet-18 and Faster R-CNN were used for classification and localization of carious lesions, respectively. Figure 1 shows the caries detection structure using U-net and Faster R-CNN in IOC images. Through the U-Net segmentation model, a single tooth is targeted and the output of the RPN of Faster R-CNN is concentrated, so that false alarms are reduced compared to when the background is not removed through segmentation. ResNet is a CNN network that won the ImageNet Large-Scale Visual Recognition Challenge (ISVRC) in 2015. They proposed residual learning using skip connections and showed that the performance improves as the layer deepens compared to previous plain networks, such as AlexNet and VGGNet [26]. To check the usefulness of the deep learning technique in IOC data, we checked the general performance of the latest deep learning technique in caries classification using the ResNet-18 structure, which has the simplest structure among the ResNets.

Fig. 1
figure 1

The architecture of U-net and Faster R-CNN for caries detection in IOC images. a U-net finds the tooth area in the IOC images and removes the background. U-net can divide the area through the process of progressively compressing and reconstructing the input information, and the skip connection helps to maintain the morphological characteristics of the previous information. Faster R-CNN creates a feature map by compressing input information with a backbone network such as VGG, and can recommend a caries region with a predefined anchor size through RPN (Region Proposal Network). Finally, we detect caries through a fully connected layer to the recommended location. b Images that have not been segmented increase false alarms in the RPN stage due to background noise. c After segmentation, it can be seen that the caries detection performance in the IOC has changed much more reliably

Faster R-CNN is a deep learning algorithm that dramatically reduces the cost of the entire process by sharing the convolutional features of the classification stage with the region proposal network (RPN), which predicts the scores for the boundaries and locations of objects. We used Faster R-CNN to apply a state-of-the-art technique, which is commonly used in the field of object detection, for the location detection of caries [27].

Training strategy of deep learning model

ResNet and Faster R-CNN were fine-tuned using pre-trained parameters with ImageNet and COCO train 2017 datasets, respectively. Besides, image augmentation, such as data shifting, image symmetry, and blurring, was applied to improve the generalisation performance of the model, but data generated owing to medical image compliance were not saved. For training, k-fold cross-validation was applied, and learning was performed by sequentially replacing separate validation sets of a certain ratio. The ratio of the training set to the validation set was 4:1, and 5-fold validation was performed on an average of 15–20 epochs; whether the validation loss increased by applying early stopping was used as the stopping criterion.

The selection of a loss function in deep learning model training can vary depending on which problem is to be solved. Ren, Shaoqing, et al. [27] used a multi-task loss function synthesizing classification loss and regression loss to simultaneously learn information on object presence and location in a Faster R-CNN model, and Mehdi Khoshboresh-Masouleh, et al. [28] optimizes the model more than before using weighted cross entropy to improve the result in the case of imbalance between the number of target pixels and non-target pixels in the segmentation problem. We used binary cross entropy and soft L1 loss by referring to the general method of Faster R-CNN. Binary cross-entropy loss was used for classification task between caries and non-caries in the Faster R-CNN model and each of predicted probabilities was compared with the ground truth either 0 or 1. Smooth L1 loss was used for box regression to estimate the location of caries. The loss function is defined as

$$L\left({p}_{i}, {t}_{i}\right) = \frac{1}{{N}_{cls}}{\sum }_{i}^{ }{L}_{cls}\left(p, {p}_{i}^{*}\right) + \lambda \frac{1}{{N}_{reg}}{\sum }_{i}^{{P}_{i}^{*}}{L}_{reg}\left({t}_{i}, {t}_{i}^{*}\right)$$
(1)

where \({N}_{cls}\) and \({N}_{reg}\) are the mini-batch size and the number of anchor locations, respectively. The anchor is a location information predefined as a candidate of the bounding box at each location of the image. \(i\)is the index of an anchor in a mini-batch and \({p}_{i}\) is the predicted probability of anchor \(i\)being an object, \({p}_{i}^{*}\)is the ground truth, \(t\)is the vector representation of the bounding box, \({t}_{i}\) is the predicted box coordinate, and \({t}_{i}^{*}\) is the ground truth. The classification loss Lcls and regression loss Lreg are balanced by λ balancing parameter to prevent imbalances caused by different normalization factors between Ncls and Nreg. λ is normally and empirically set by 10 as default value. The classification loss Lcls is the binary cross entropy loss and the regression loss Lreg uses the smooth L1 loss function. The classification loss function is defined as

$${L}_{cls}\left({p}_{i}, {p}_{i}^{*}\right) = -{p}_{i}^{*}\text{log}{p}_{i} - \left(1-{p}_{i}^{*}\right)\text{log}\left(1-{p}_{i}\right)$$
(2)

The IOU (Intersection over Union) threshold was set to 0.2, and NMS (non-maximum suppression) was used to select the most suitable candidate area. The threshold value was set lower than for general object detection tasks; however, owing to the characteristics of caries, the boundary of the shape was not clear, and when compared with the ground truth, a false positive case occurred that was determined retrospectively to be correct. Thus, position errors were allowed to some extent. The SGD optimiser was used to update the parameters, and the learning rate was set to 0.001. In the case of ResNet, a dropout of 0.5 was applied, and the batch size was set as large as possible as memory allowed. For model training, two NVIDIA RTX2080 Ti GPUs and Itel Xeon E5-1650 CPU were used, and the deep learning frameworks, PyTorch 1.11, Cuda 11.4, and cuDNN 8.2, were used. Our code and benchmarks are available at github.com/hyeonrae/ioai.

Statistical analysis

To evaluate the overall performance for caries classification, the accuracy, sensitivity, specificity, negative predictive value (NPV), precision, and area under the receiver operating characteristic curve (AUC) were calculated according to the segmentation of the tooth surface.

In the case of localization of caries, it is difficult to accurately identify a carious lesion in a photographic image in many cases even with a dental examination, and it is difficult to define a clear feature on the boundary between a carious area and a normal tooth area even on image data. It becomes difficult to precisely define the threshold of the boundary. Therefore, the allowable IOU threshold between the ground truth (GT) and the predicted bounding box was defined as 20%, and it was set more generously than other object detection models. If the IOU value was greater than 40%, the predicted box was regarded as the true detection case. True positive (TP), false positive (FP), and false negative (FN) cases were classified on this basis. To evaluate the overall performance of caries localisation, sensitivity, specificity, precision, NPV, precision, and average precision were calculated according to segmentation of the tooth surface.

Results

The evaluation metrics of the classification algorithms for caries among 300 test images (150 images with carious teeth, 150 images of healthy teeth, with 163 carious lesions included) are shown in Table 2.

Table 2 Evaluation metrics of classification for caries images using test set

When image classification was performed with segmentation of the tooth surface, all the evaluation indexes improved better than those without segmentation. For example, AUC improved to 0.831 from 0.731. The accuracy improved to 0.813 from 0.756, which means that our CNN algorithms correctly classified 244 of the 300 test images of caries existence. The sensitivity and precision improved from 0.740 to 0.867 to 0.722 and 0.779, respectively. Pre-processing of each photographic image for tooth surface segmentation also improved the performance of CNN algorithms for the localisation of carious lesions. The sensitivity, precision, and average precision improved from 0.890, 0.874, and 0.889 to 0.865, 0.750, and 0.868, respectively, through segmentation of the tooth surface (Table 3).

Table 3 Evaluation metrics of localisation for caries before and after tooth surface segmentation using test set

From these results, we found that training CNN algorithms to predict the tooth surface in each photographic image can improve its performance in terms of both tooth classification and localisation of carious lesions. The detection results are shown in Fig. 2 for qualitative verification, in which the green box, GT annotated by the dentist and yellow box is predicted by the CNN model. In the TP cases, caries was appropriately detected, although the size of the bounding box may be slightly different from that of the GT (Fig. 2a). In many cases of false detection, they were classified as negative because there was a difference between the GT and prediction box. However, in such cases, clinical effectiveness could be expected because it properly included a significant proportion of the caries.

Fig. 2
figure 2

Results images. The green box is the ground truth cases and the yellow box is the predicted results by Faster R-CNN. a True positive (TP) cases when the IoU threshold is 0.2 b The results of false positive (FP) and false negative (FN)

Discussion

AI has recently proven promising in the detection and diagnosis of dental caries. Meanwhile, diagnostic imaging with AI in dentistry could operate much faster and more accurately than humans do, which lowers costs, eliminates subjective individual examinations, and increases the effectiveness of care [29]. However, most studies have applied deep neural networks to the analysis of dental X-ray images, such as panorama, bitewing, or periapical radiographs to detect dental caries [17, 30,31,32]. In clinical practice, visual-tactile examination is usually performed to detect caries as the first method. The caries detection method using intraoral photographs was more accurate than the inspection method, probably due to the ability to magnify the image on the computer monitor and thus easy to find [33]. Intraoral photographs can be beneficial for conducting dental public health surveys, especially in suburban areas. It could be utilised as an effective medium of communication between the patient and dentist [8] and in advancing tele-dentistry [34]. Intraoral photographs may be captured by a smartphone, professional single-reflex lens camera with flash, or intraoral camera for clinical use. Because smartphones have limitations in the magnification of the tooth surface, they are not recommended for identification of initial caries. In contrast, a professional single-reflex lens camera with flash is excellent in magnification and finding initial caries, but it is heavy, inconvenient to operate, and requires an intraoral mirror for the molar teeth. Therefore, intraoral cameras are generally used in dental clinics.

Therefore, in this study, we evaluated the performance of caries detection using a deep learning algorithm with photographs captured by dentists with intra oral camera for the classification of caries images and localisation of carious lesions. To improve the performance, we pre-processed the original images with a deep learning algorithm for segmentation of the tooth surface, which can extract the tooth image by eliminating the background from the original image automatically. As a result, all the performance indices, such as accuracy and AUC of the classification algorithm for caries images, were improved through segmentation of the tooth area. In the case of the localisation algorithm for carious lesions, improvement of the performance index, such as mean precision, was observed through segmentation of the tooth area. The detection system using Faster R-CNN is processed at an average speed of 270 ms per input image of 1280 × 720 size on the GPU (RTX2080), and the results can be utilized in real time by shooting with a camera in clinical practice.

In case of carious lesion detection, our model showed 89.0% of sensitivity and 87.4% of precision, which are better than 64.6% of sensitivity of previous study using oral photographs captured by smartphone [19]. However, in case of carious image classification, our model showed 83.7% of AUC, which is similar to 85.65% of the previous study [19]. According to another previous study on classification of caries image captured by DSLR, a high accuracy of 92.5% was achieved, which is better than 81.3% of our model [20]. However, as discussed in their paper, strict methodological requirements such as only single-tooth included image that did not have non-carious cases, for example, developmental defects, fillings, or discoloration were applied at their dataset, which cause potential problems in its application in general dental clinics [20].Therefore direct comparison of their performance with those of our CNN model is unsuitable considering that our dataset including various dental restorations, multiple teeth in one image, discoloration on occlusal tooth surface, which may have bad influence on prediction of CNN model. There are several drawbacks to intraoral photographs taken in dental clinics, such as presence of saliva, multiple teeth in one photographic image, images obliquely captured to the tooth surface, presence of dental restorations (such as amalgam and resin) or various lesions similar to caries. Therefore, this study attempted to improve the applicability of the DL model by reducing the limitations of the photography dataset and including various caries-like lesions for training.

However, most CNN models may show reduced performance when applied to real-world images with blur and noise because the neural networks are susceptible to these distortions. Therefore, most DL models may need to be trained using augmented images processed for various distortion [35]. Nevertheless, since augmented images are highly correlated with the original versions, these augmented versions do not provide many new features for the CNN to learn compared with independent original images. Therefore, we attempted to include various forms of oral photographic images as a dataset [36]. Besides, we pre-processed the original images through the segmentation of the tooth surface.

After our CNN model was trained for segmentation of the tooth surface, the resulting images included the tooth surface only by removing the background from the original images without manual intervention. Training the DL model for caries detection using these pre-processed images showed improved model performance compared to training using the original photographic images without pre-processing. Similar results have been observed in previous studies [21, 37,38,39,40]. In a study by Lian et al. a DL model showed good performance in detection and classification by localising the carious lesion after segmenting only the tooth outline within the entire image of the dental panorama [21].

Automated segmentation attempts using DL models have mainly been performed on radiography in dentistry, and this study is the first to be performed on intraoral photographic images. It is reasonable that segmentation of the tooth surface which removes the background from the original image and leaves only the tooth area, would contribute to improved accuracy by allowing the AI model to focus only on the tooth surface.

However, this study had several limitations. First, compared with an X-ray image, intraoral photographic images cannot express the inside of the tooth and the interproximal tooth surface. Therefore, there is a limit to fundamentally finding all carious lesions without X-ray images or tactile examination. In future studies, other diagnostic tools to complement the intraoral photographic images should be used as a dataset to evaluate the performance of the AI model. Second, the relatively low sensitivity and negative predictive value need to be improved along with the general improvement of all the evaluation indexes.

Nevertheless, this study is significant as it is the first report to achieve relatively acceptable performance of deep learning models in both classification of caries images and localisation of carious lesions using photographic images by intraoral camera. Diagnosis may vary depending on the clinician’s experience and skills therefore, CAD can be an effective diagnostic support tool to overcome this [41]. This study is expected to initiate the evaluation of the DL model performance using intraoral photographs and active investigations to diagnose caries easily and accurately.

Conclusion

In conclusion, this study indicate that caries can be detected to some extent, even on intraoral photographic images, and this study can be developed to support the judgment of the medical staff.

Availability of data and materials

The datasets generated and/or analysed during the current study are not publicly available due to privacy of participants but are available from the corresponding author on reasonable request.

Abbreviations

CNN:

Convolutional neural networks

IOC:

Intraoral camera

DL:

Deep learning

AI:

Artificial intelligence

CAD:

Computer-aided diagnostic

DSLR:

Digital single-lens reflex

STARD:

Standards for reporting of diagnostic accuracy studies

ICDAS:

International caries detection and assessment system

ISVRC:

ImageNet large-scale visual recognition challenge

RPN:

Region proposal network

IOU:

Intersection over union

NMS:

Non-maximum suppression

NPV:

Negative predictive value

AUC:

Area under the receiver operating characteristic curve

TP:

True positive

FP:

False positive

FN:

False negative

GT:

Ground truth

References

  1. Selwitz RH, Ismail AI, Pitts NB. Dental caries. Lancet. 2007;369:51–9.

    Article  Google Scholar 

  2. Fejerskov O, Nyvad B, Kidd E, editors. Dental caries: the disease and its clinical management. John Wiley & Sons; 2015.

  3. Broadbent JM, Thomson WM, Poulton R. Progression of dental caries and tooth loss between the third and fourth decades of life: a birth cohort study. Caries Res. 2006;40:459–65.

    Article  Google Scholar 

  4. Petersen PE, Bourgeois D, Ogawa H, Estupinan-Day S, Ndiaye C. The global burden of oral diseases and risks to oral health. Bull World Health Organ. 2005;83:661–9.

    Google Scholar 

  5. Murdoch-Kinch CA, McLean ME. Minimally invasive dentistry. J Am Dent Assoc. 2003;134:87–95.

    Article  Google Scholar 

  6. Yılmaz H, Keleş S. Recent methods for diagnosis of dental caries in dentistry. Meandros Med Dent J. 2018;19:1–8.

    Article  Google Scholar 

  7. Dabiri D, Eckert GJ, Li Y, Seow K, Schroth RJ, Warren J, et al. Diagnosing developmental defects of enamel: pilot study of online training and accuracy. Pediatr Dent. 2018;15(40):105–9.

    Google Scholar 

  8. Pentapati KC, Siddiq H. Clinical applications of intraoral camera to increase patient compliance—current perspectives. Clin Cosmet Investig Dent. 2019;11:267–78.

    Article  Google Scholar 

  9. Snyder TL. The intraoral camera: a popular computerized tool. J Am Dent Assoc. 1995;126:177–8.

    Article  Google Scholar 

  10. Forgie AH, Pine CM, Pitts NB. The assessment of an intra-oral video camera as an aid to occlusal caries detection. Int Dent J. 2003;53:3–6.

    Article  Google Scholar 

  11. Erten H, Uçtasli MB, Akarslan ZZ, Uzun O, Baspinar E. The assessment of unaided visual examination, intraoral camera and operating microscope for the detection of occlusal caries lesions. Oper Dent. 2005;30:190–4.

    Google Scholar 

  12. Moutselos K, Berdouses E, Oulis C, Maglogiannis I. Recognizing occlusal caries in dental intraoral images using deep learning. In: Annu Int Conf IEEE Eng Med Biol Soc 41st annual international conference of the IEEE engineering in medicine and biology society (EMBC). IEEE Publications; 2019. p. 1617–20.

  13. Askar H, Krois J, Rohrer C, Mertens S, Elhennawy K, Ottolenghi L, et al. Detecting white spot lesions on dental photography using deep learning: a pilot study. J Dent. 2021;107:103615.

    Article  Google Scholar 

  14. Xception CF. Deep learning with depthwise separable convolutions. In: Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR). IEEE Publications; 2017. p. 1251–8.

  15. Esteva A, Kuprel B, Novoa RA, Ko J, Swetter SM, Blau HM, et al. Dermatologist-level classification of skin cancer with deep neural networks. Nature. 2017;542:115–8.

    Article  Google Scholar 

  16. Sklan JE, Plassard AJ, Fabbri D, Landman BA. Toward content based image retrieval with deep convolutional neural networks. Proc SPIE Int Soc Opt Eng. 2015;9417:94172C.

    Google Scholar 

  17. Lee JH, Kim DH, Jeong SN, Choi SH. Detection and diagnosis of dental caries using a deep learning-based convolutional neural network algorithm. J Dent. 2018;77:106–11.

    Article  Google Scholar 

  18. Prajapati SA, Nagaraj R, Mitra S. Classification of dental diseases using CNN and transfer learning. In: 5th international symposium on computational and business intelligence (ISCBI), 2017. IEEE Publications; 2017. p. 70–4.

  19. Zhang X, Liang Y, Li W, Liu C, Gu D, Sun W, et al. Development and evaluation of deep learning for screening dental caries from oral photographs. Oral Dis. 2022;28:173–81.

    Article  Google Scholar 

  20. Kühnisch J, Meyer O, Hesenius M, Hickel R, Gruhn V. Caries detection on intraoral images using artificial intelligence. J Dent Res. 2022;101:158–65.

    Article  Google Scholar 

  21. Lian L, Zhu T, Zhu F, Zhu H. Deep learning for caries detection and classification. Diagnostics. 2021;13(11):1672.

    Article  Google Scholar 

  22. Kayalibay B, Jensen G, van der Smagt P. CNN-based segmentation of medical imaging data. arXiv preprint arXiv:1701.03056.

  23. Ismail AI, Sohn W, Tellez M, Amaya A, Sen A, Hasson H, Pitts NB. The international caries detection and assessment system (ICDAS): an integrated system for measuring dental caries. Community Dent Oral Epidemiol. 2007;35(3):170–8.

    Article  Google Scholar 

  24. Cheng L, Zhang L, Yue L, Ling J, Fan M, Yang D, Huang Z, Niu Y, Liu J, Zhao J, Li Y, Guo B, Chen Z, Zhou X. Expert consensus on dental caries management. Int J Oral Sci. 2022;14(1):17.

    Article  Google Scholar 

  25. Ronneberger O, Fischer P, Brox T. U-net: Convolutional networks for biomedical image segmentation. In: Lecture notes in computer science international conference on medical image computing and computer-assisted intervention (MICCAI). Cham: Springer; 2015. p. 234–41.

  26. He K, Zhang X, Ren X, Sun J. Deep residual learning for image recognition. In: Proceedings of the 2016 IEEE conference on computer vision and pattern recognition (CVPR); 2016. p. 770–8.

  27. Ren S, He K, Girshick R, Sun J, Faster R-CNN. Faster R-CNN Towards real-time object detection with region proposal networks. IEEE Trans Pattern Anal Mach Intell. 2017;39:1137–49.

    Article  Google Scholar 

  28. Khoshboresh-Masouleh M, Shah-Hosseini R. A deep learning method for near-real-time cloud and cloud shadow segmentation from gaofen-1 images. Comput Intell Neurosci. 2020. https://doi.org/10.1155/2020/8811630.

    Article  Google Scholar 

  29. Schwendicke F, Samek W, Krois J. Artificial intelligence in dentistry: chances and challenges. J Dent Res. 2020;99:769–74.

    Article  Google Scholar 

  30. Devito KL, de Souza BF, Felippe Filho WN. An artificial multilayer perceptron neural network for diagnosis of proximal dental caries. Oral Surg Oral Med Oral Pathol Oral Radiol Endod. 2008;106:879–84.

    Article  Google Scholar 

  31. Srivastava MM, Kumar P, Pradhan L, Varadarajan S. Detection of tooth caries in bitewing radiographs using deep learning. arXiv preprint arXiv:1711.07312, 2017.

  32. Ali RB, Ejbali R, Zaied M. Detection and classification of dental caries in X-ray images using deep neural networks. In: International conference on software engineering advances (ICSEA) Rome, Italy. 2016. p. 236.

  33. Lee JT, Lee KH, Seo JH, Chun JA, Park JH. The evaluation for oral examination by using of intra-oral camera. Int J Clin Prev Dent. 2014;10:113–20.

    Article  Google Scholar 

  34. Ghai S. Teledentistry during COVID-19 pandemic. Diabetes Metab Syndr. 2020;14:933–5.

    Article  Google Scholar 

  35. Dodge S, Karam L. Understanding how image quality affects deep neural networks. In: 2016 eighth international conference on quality of multimedia experience (QoMEX). IEEE. 2016 June. p. 1–6.

  36. Chan HP, Samala RK, Hadjiiski LM, Zhou C. Deep learning in medical image analysis. Adv Exp Med Biol. 2020;1213:3–21.

    Article  Google Scholar 

  37. Cantu AG, Gehrung S, Krois J, Chaurasia A, Rossi JG, Gaudin R, et al. Detecting caries lesions of different radiographic extension on bitewings using deep learning. J Dent. 2020;100:103425.

    Article  Google Scholar 

  38. Bayraktar Y, Ayan E. Diagnosis of interproximal caries lesions with deep convolutional neural network in digital bitewing radiographs. Clin Oral Investig. 2022;26:623–32.

    Article  Google Scholar 

  39. Lee S, Oh SI, Jo J, Kang S, Shin Y, Park JW. Deep learning for early dental caries detection in bitewing radiographs. Sci Rep. 2021. https://doi.org/10.1038/s41598-021-96368-7.

    Article  Google Scholar 

  40. Lin XJ, Zhang D, Huang MY, Cheng H, Yu H. Evaluation of computer-aided diagnosis system for detecting dental approximal caries lesions on periapical radiographs. Zhonghua Kou Qiang Yi Xue Za Zhi. 2020;9(55):654–60.

    Google Scholar 

  41. Shibata T, Teramoto A, Yamada H, Ohmiya N, Saito K, Fujita H. Automated detection and segmentation of early gastric cancer from endoscopic images using mask R-CNN. Appl Sci. 2020;10:3842.

    Article  Google Scholar 

Download references

Acknowledgements

This work was supported by the 2021 Yeungnam University Research Grant.

Funding

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. 2020R1F1A1070070).

Author information

Authors and Affiliations

Authors

Contributions

EYP contributed to drafting the article and acquisition of data, and HRC contributed to analysis and interpretation of data. SHK contributed to acquisition of data and SMJ contributed to revising it critically for important intellectual content. EKK contributed to the conception and design of the study. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Eun-Kyong Kim.

Ethics declarations

Ethics approval and consent to participate

The study procedure was approved by the Institutional Review Board of the Kyungpook National University (KNU-2021-0097) and was performed in accordance with the Declaration of Helsinki. This study followed the guidelines of the Standards for Reporting of Diagnostic Accuracy Studies. Only participants who provided informed consents were enrolled.

Consent for publication

Not applicable.

Competing interests

The authors certify that no competing interests were involved in this paper.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Park, E.Y., Cho, H., Kang, S. et al. Caries detection with tooth surface segmentation on intraoral photographic images using deep learning. BMC Oral Health 22, 573 (2022). https://doi.org/10.1186/s12903-022-02589-1

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12903-022-02589-1

Keywords