Skip to main content

Detection of periodontal bone loss patterns and furcation defects from panoramic radiographs using deep learning algorithm: a retrospective study

Abstract

Background

This retrospective study aimed to develop a deep learning algorithm for the interpretation of panoramic radiographs and to examine the performance of this algorithm in the detection of periodontal bone losses and bone loss patterns.

Methods

A total of 1121 panoramic radiographs were used in this study. Bone losses in the maxilla and mandibula (total alveolar bone loss) (n = 2251), interdental bone losses (n = 25303), and furcation defects (n = 2815) were labeled using the segmentation method. In addition, interdental bone losses were divided into horizontal (n = 21839) and vertical (n = 3464) bone losses according to the defect patterns. A Convolutional Neural Network (CNN)-based artificial intelligence (AI) system was developed using U-Net architecture. The performance of the deep learning algorithm was statistically evaluated by the confusion matrix and ROC curve analysis.

Results

The system showed the highest diagnostic performance in the detection of total alveolar bone losses (AUC = 0.951) and the lowest in the detection of vertical bone losses (AUC = 0.733). The sensitivity, precision, F1 score, accuracy, and AUC values were found as 1, 0.995, 0.997, 0.994, 0.951 for total alveolar bone loss; found as 0.947, 0.939, 0.943, 0.892, 0.910 for horizontal bone losses; found as 0.558, 0.846, 0.673, 0.506, 0.733 for vertical bone losses and found as 0.892, 0.933, 0.912, 0.837, 0.868 for furcation defects (respectively).

Conclusions

AI systems offer promising results in determining periodontal bone loss patterns and furcation defects from dental radiographs. This suggests that CNN algorithms can also be used to provide more detailed information such as automatic determination of periodontal disease severity and treatment planning in various dental radiographs.

Peer Review reports

Background

Periodontitis is a chronic inflammatory disease that is characterized by damage to the supporting tissues of the teeth and can result in tooth loss if not controlled [1]. This common disease in society is associated with many systemic diseases in the body as well as dental problems such as chewing loss [2, 3]. In this respect, early diagnosis and treatment planning are very important in periodontitis cases [2, 3].

Radiographic evaluations play an important role in the diagnosis of periodontitis in addition to clinical periodontal evaluations such as probing pocket depth, attachment loss, and gingival recession [4]. Intraoral radiographs such as periapical radiography and bitewing radiography are commonly used to determine periodontal status [4,5,6]. Panoramic radiographs, which is one of the extraoral dental radiographic techniques, have advantages such as working with low-dose radiation providing quick and easy radiological imaging, and also helping in determining general dental problems and periodontal conditions [6,7,8]. In addition, these radiographs allow the observation of bone loss patterns (such as horizontal bone loss, vertical bone loss, and furcation defects) in periodontitis cases [9]. In periodontitis cases, it is very important to determine the bone loss patterns and bone morphology to make an appropriate treatment plan and achieve successful results [10, 11]. Horizontal bone loss occurs when the bone supporting the tooth melts at the same levels on all tooth-related surfaces (mesial, distal, vestibular/buccal, lingual/palatal). In this type of bone defect, the alveolar crest levels around the teeth are resorbed parallel to the line that should be in the healthy periodontium and are positioned more apically [12,13,14]. Vertical bone losses are bone losses that occur obliquely and angularly in the interdental region [12,13,14]. In multi-rooted teeth, furcation defects occur when the alveolar bone between the tooth roots is also affected by periodontal disease. In furcation defects, both diagnosis and treatment planning become much more complex and difficult for dentists [15]. The correct estimation of bone loss patterns enables the correct treatment selection and facilitates the work of both patients and dentists. For example, flap surgery and/or resective surgical periodontal therapy may be preferred rather than regenerative treatment in horizontal bone losses, while regenerative treatment may be an alternative method in vertical bone losses [14]. According to the current classification of periodontal and peri-implant diseases and conditions developed in 2018, the detection of periodontal bone loss patterns is very important for determining disease stages [16]. In this classification system, Stage I and II periodontitis are associated with the presence of horizontal bone loss, while the presence of vertical bone losses and furcation defects in Stage III and IV periodontitis is noted [16, 17]. Current academic studies using image processing and machine learning technologies and aiming at automatic periodontal disease classification have pointed out the determination of periodontal bone loss patterns as one of the main classification criteria [17, 18]. Therefore, the determination of periodontal bone loss pattern is of clinical importance in making the current periodontal disease classification complete and accurate [17, 18].

Various studies about artificial intelligence (AI) have been carried out in dentistry, following the use of AI systems in the medical field for disease diagnosis and treatment planning [19, 20]. In these studies, it is seen that AI systems are used for the determination of many pathologies such as dental caries, apical lesions, tooth numbering, and root fractures on two-dimensional (2D) radiographs [21,22,23,24]. The main purpose of using AI in dentistry practice is to automatically detect pathologies, diseases, or anatomical structures and determine disease severity. While achieving these goals, it also provides additional benefits such as preventing situations that may be overlooked due to the physician’s inexperience, intensity, and fatigue, ensuring early diagnosis of diseases, and recording patient data more regularly in the digital environment [6, 20].

There are several studies for evaluating bone losses and periodontal problems using AI systems [3, 6, 18, 25,26,27,28,29,30]. However, as far as we know, there is no AI-based study aiming to determine the periodontal bone loss patterns using the segmentation method. This study aimed to examine the performance of a convolutional neural network (CNN)-based AI system in the detection of periodontal problems such as horizontal bone losses, vertical bone losses, and furcation defects on panoramic radiographs by segmentation method.

Methods

Study design

All of the images were obtained from the radiology archive of the Eskişehir Osmangazi University Faculty of Dentistry for this retrospective study. Ethics board approval was obtained from the Non-interventional Clinical Research Ethics Committee of Eskişehir Osmangazi University before starting the study (decision number: 2019 − 227). The guidelines of the Declaration of Helsinki were followed throughout all phases of the study.

Data selection

All images used in the study were obtained with the same radiography device (Planmeca Promax 2D, Planmeca, Helsinki, Finland) with the following parameters: 68 kVp, 16 mA, and 13 s. All images were selected from images of individuals over the age of 18, without paying attention to age and gender differences. In radiographs, 2 mm apical of the imaginary line passing through the cemento-enamel junction of the teeth was accepted as the healthy periodontium level [31,32,33]. Cases whose alveolar bone crest was coronal to this imaginary line were considered to have healthy periodontium and were excluded from the study. Images of periodontitis cases of different ages and genders were re-evaluated in terms of the criteria stated below: ‘1. The images with dense artifacts, 2. The images with low quality due to patient positioning error or patient movement during acquisition, 3. The images of patients who received orthognathic treatment, 4. The images of patients with bone metabolism disorders, 5. The images of patients with unusual alveolar bone morphology (bone pathologies due to conditions such as cysts and tumors), 6. The images of patients with cleft lip and palate, 7. The images of poor quality and blurred images in the alveolar bone region due to conditions such as dental crowding, 8. The images of patients with metal restorations that cause artifacts to complicate periodontal diagnosis’. Images that meet these criteria, which would make diagnosis difficult and disrupt the standardization of the study, were also excluded from the study.

Ground truth

All data that were compatible with the inclusion criteria were uploaded into the CranioCatch labeling module (CranioCatch, Eskisehir, Turkey) in an anonymized manner. Initially, intra-observer and inter-observer compatibility were analyzed for the 4 observers (3 periodontists: SKB, MBY, NS, and 1 oral maxillofacial radiologist: İŞB) who planned to carry out the labeling to ensure labeling standardization. For intra-observer evaluation, the observers labeled all the parameters to be evaluated in this study in line with the criteria below on 10 radiographs with the segmentation method. Each observer repeated this procedure 1 week later. A computer program (Python version 3.6.1, Python Software Foundation) and NumPy library were used and the splicing of the segmented fields was done and Intersection over Union (IoU) values were calculated. When it was decided that there was intra-observer and inter-observer aggregation for labeling, all data were evaluated by the same observers according to the same criteria, and the labeling of the relevant parameters was completed.

The criteria taken into consideration during the labeling phase were:

For total alveolar bone loss: The unit of this parameter was the jaw (maxilla and mandible). Two separate segmentation processes were performed for the maxilla and mandible in all panoramic radiographs. While performing the segmentation, the root parts of the teeth without bone support and the interdental regions following these areas were included in the label area. Initially, a segmentation line was created from the cemento-enamel junction of all teeth. After that, the segmentation line was combined to follow the line on the distal surfaces of the most distal teeth and the border of the bone crests in the relevant jaw. In Fig. 1-a, an example image was presented to better understand the path followed when labeling this parameter. (Fig. 1-a)

Fig. 1
figure 1

Images of the labeling module. (a) Labeling of a patient’s total alveolar bone loss; (b) Labeling of a patient’s interdental bone loss (red: vertical, blue: horizontal, purple: furcation); (c) Labeled version of all data of a patient

In the next stage, the interdental regions were evaluated one by one. A separate labeling process was made for these areas, taking into account the bone loss patterns (in the form of horizontal and vertical bone loss) (Fig. 1-b). The units of these parameters were the interdental region.

For horizontal bone losses: Interdental regions where the imaginary lines passing through the cemento-enamel junctions of two adjacent teeth and the alveolar bone crest were parallel to each other and where the alveolar bone level was in a more apical position than it should be in the healthy periodontium were labeled with the horizontal bone loss label [12,13,14]. During the segmentation process, the cemento-enamel junctions of two adjacent teeth, the most distal border of the mesial tooth, the alveolar crest line in the relevant interdental region, and the most mesial border of the distal tooth were followed, and the specified reference areas were combined in the form of a closed curve. When the relevant labels were completed, all of the segmentation areas resembled geometric shapes such as squares, rectangles, and trapezoids (Fig. 1-b).

For vertical bone losses: Two imaginary lines were created from the cemento-enamel junctions of two adjacent teeth and the alveolar bone crest line. The interdental regions where these two lines are at an angle to each other were labeled with the vertical bone destruction label [12,13,14]. When completed, the labels showed a triangular geometry. Minimal ridge angles due to changes in the tooth axis are not included in this label. The criteria specified for horizontal bone loss were followed when creating labeling boundaries (Fig. 1-b).

For furcation defects: The unit of this parameter was the tooth. Furcation areas of multi-rooted teeth were examined and radiolucent images compatible with bone resorption were labeled with the furcation defect label. While labeling was carried out, the root boundaries of the relevant tooth and the boundaries of the radiolucent region formed by the lesion were followed and completed to form a closed curve (Fig. 1-b).

After labeling was completed, the data set was rechecked by the observers. At this stage, instead of reviewing the boundaries of segmentation, the bone resorption form and geometry of the labeled region were quickly reviewed to avoid standardization deficiencies that may arise in horizontal-vertical bone loss decisions. In addition, consensus was reached in cases where dentists were conflicted about labeling some initial furcation defects. Also, the final evaluation made by 4 dentists together allowed the forgotten and overlooked labels to be noticed and completed before AI training. Labels with no consensus on diagnosis were removed and not used in algorithm training and result metrics.

Model training

1. Pre-processing steps: All images were resized as 1024 × 512 pixels. Panoramic radiographs to be used in the data sets created for developing the furcation defect model were cropped to 4 (left mandible, right mandible, left maxilla, and right maxilla). Since furcation defects were only seen in multi-rooted molars, it was aimed to focus more on the relevant areas in this way. The radiographs to be used in other parameter models were not cropped but used as a whole. Images that do not have the related parameter label were excluded from the main data sets of that parameter (Figs. 2 and 3). In line with these criteria, 1121 panoramic radiographs for total alveolar bone losses, 1120 panoramic radiographs for horizontal bone losses, 828 panoramic radiographs for vertical bone losses, and 1941 cropped panoramic radiographs (890 panoramic radiographs) for furcation defects were included in the main datasets (Table 1; Fig. 3).

Fig. 2
figure 2

Workflow demonstrating the application of inclusion and exclusion criteria

Table 1 Numerical data on the development of deep learning algorithms
Fig. 3
figure 3

The stages of creating the datasets to be used in the development of AI models for all parameters. (a) Total alveolar bone loss; (b) Horizontal bone loss; (c) Vertical bone loss; (d) Furcation defect

2. Training, Validation, and Testing Data: The main datasets were created by combining all panoramic radiographs containing the relevant parameter labels for each parameter. Then these main datasets were split as a training set (80%), validation set (10%), and testing set (10%) randomly for the development of AI systems (Table 1).

Training set: 935 images (1878 labels) for total alveolar bone loss, 934 images (18232 labels) for horizontal bone loss, 690 images (2869 labels) for vertical bone loss, and 1619 cropped images (2358 labels) for furcation defect were used as training data-set.

The validation set: 93 images (186 labels) for total alveolar bone loss, 93 images (1786 labels) for horizontal bone loss, 69 images (287 labels) for vertical bone loss, and 161 cropped images (227 labels) for furcation defect were used as validation data-set.

Test set: 93 images (187 labels) for total alveolar bone loss, 93 images (1821 labels) for horizontal bone loss, 69 images (308 labels) for vertical bone loss, and 161 cropped images (230 labels) for furcation defect were used as testing data-set.

The labels in the training set were used to create the algorithm of the AI model. This model was tested with the data in the validation set for adjusting the algorithm and questioning the need for further training. The final version of the algorithm, which was developed using the training set and successful trials were carried out with the labels in the validation set, was used in the detection of periodontal bone destruction and pattern on the radiographs in the testing set. The diagnostic results of the system in the testing set were compared with the observer labels in this dataset by a computer command, and the success metrics of the system were presented. In short, while determining the final success of the model, the observer diagnoses in the test dataset were accepted as the gold standard and these labels were used as the ground truth dataset. (Fig. 3). These steps were performed separately for each parameter.

3. Description of CNN architecture: The U-Net architecture was used to perform deep learning (Table 1) [24, 34]. U-Net is a type of CNN that can perform semantic segmentation assignments on various images such as radiography images. U-Net contains four block levels with 32, 64, 128, and 256 convolution filters in each block. The working mechanism of this architecture is to convert images to vectors for pixel classifications and then convert these vectors back to images for segmentation. The encoder and decoder paths are used in order and perform model training. There is a maximum pool layer in the paths of encoders and up-convolution layers in the paths of decoders. For the current study, the U-Net architecture steps were followed and the working mechanism was presented in Fig. 4 in detail [24, 34].

Fig. 4
figure 4

U-Net architecture and deep learning stages. Red tiles were used to show the multichannel feature map and light tiles were used for the copied feature map. Different colored arrows represented different processes in the architecture [34]

4. Method for training: PyTorch library (v. 3.6.1; Python Software Foundation, Wilmington, DE, USA) and Adaptive Moment Estimation (ADAM) optimizer (for updating the learning rate for each parameter) were used in the model development stages. All training was made with a computer equipped with 16 GB RAM and NVIDIA GeForce GTX 1660 TI graphics card. U-Net architecture was used in the training of all parameters and all of them were trained with 800 epochs. Training of total alveolar bone loss and horizontal bone loss parameters was carried out with a learning rate of 0.0001, and training of vertical bone loss and furcation defect parameters was carried out with a learning rate of 0.00001.

Assessment metrics

The statistical phase of the study was carried out with the confusion matrix method and ROC analysis. A script was created to perform these evaluations automatically with the help of a computer. A computer program (Python version 3.6.1, Python Software Foundation) and OpenCV and NumPy libraries were used to develop this script. The script used at this stage accepted the labels made by experienced dentists in the testing set as the gold standard. It automatically presented a success metric by comparing the diagnostics of the algorithm in the relevant dataset with the observers’ diagnoses. While the script was running, it presented results using some calculations below.

Sensitivity, precision, and F1 scores were calculated using the confusion matrix method [24]. Initially, three different calculations were for the confusion matrix method: numbers of true positives (TP, there was a periodontal problem and it was segmented correctly), numbers of false positives (FP, there wasn’t a periodontal problem but it was detected wrongly), and numbers of false negatives (FN, there was a periodontal problem but it wasn’t detected). Then, sensitivity, precision, and F1 scores were determined with the following calculations:

Sensitivity (recall): TP/ (TP + FN).

Precision: TP/ (TP + FP).

F1 score: 2TP/ (2TP + FP + FN).

In addition to the confusion matrix method, receiver operating characteristic (ROC) curve analysis and precision-recall evaluation were also performed to provide more detailed data. The area under the ROC curve (AUC) values are calculated by this analysis. ROC is a probability curve and the area under it, AUC, represents the degree or measure of separability. As the area under the curve increases, the discrimination performance between classes increases. It is known that as the success of the system increases, the AUC grows and approaches the value of 1. In line with this information, interpretations were made and the results were presented. In addition, the labels and system results of the radiographs in the test data set in which the evaluations were carried out were reviewed by experienced dentists for each parameter.

Results

Study data

A total of 2949 radiographs were examined in the study. 1804 of them were excluded from the study because they were not periodontitis cases. In addition, 24 of the remaining images were excluded from the study due to radiography (n = 21) and case-based reasons (n = 3). Observers agreed on the periodontitis diagnosis of 1121 panoramic radiographs. These radiographs contained a total of 30369 labels as 2251 total alveolar bone loss, 25303 interdental bone losses (21839 horizontal bone losses, 3464 vertical bone losses), and 2815 furcation defects in total (Fig. 5). The total number of labels presented for each parameter indicated the presence of pathology related to the relevant parameter (bone losses or furcation defects). If the relevant pathology was not present, a label was not available in the relevant area.

Fig. 5
figure 5

Standards for Reporting Diagnostic Accuracy (STARD) flow diagram

Intra-observer and inter-observer agreements

It was found that each investigator showed high consistency in the diagnosis and segmentation of these parameters at different times (IoU > 0.91). In addition, kappa values were calculated to evaluate the consistency of the observers’ decision for horizontal and vertical bone losses in the interdental regions. Kappa values showed excellent agreement when the two times were compared in the decision of bone loss characteristics in each investigator (K = 0.81–0.99). Also, the labeling areas of different observers were examined for evaluation of inter-observer agreement using this data set. The IoU values obtained were found to be between 0.8126 and 0.8737. While the results were 0.85 and above for horizontal bone loss, total alveolar bone loss, and furcation defect, they were lower for vertical bone loss (> 0.81).

Outcomes

While the confusion matrix evaluates a given classifier with a fixed threshold, the AUC evaluates that classifier over all possible thresholds. Therefore, AUC was considered as the primary outcome. AUC values were 0.951 for total alveolar bone loss, 0.910 for horizontal bone losses, 0.733 for vertical bone losses, and 0.868 for furcation defects (Table 2). The developed model showed the most successful results when determining total alveolar bone loss (AUC = 0.951). The model’s lowest success rate was seen when detecting vertical bone loss (AUC = 0.733).

Table 2 The results obtained with the confusion matrix method and ROC analysis

The results were calculated with the confusion matrix method, which demonstrates the success of the developed AI model, were presented in Table 2. When the diagnostic performance metrics were evaluated, sensitivity, precision, F1 score, and accuracy values, respectively, were 1, 0.995, 0.997, 0. 994 for total alveolar bone loss; 0.947, 0.939, 0.943, 0.892 for horizontal bone losses; 0.558, 0.846, 0.673, 0.506 for vertical bone losses and 0.892, 0.933, 0.912, 0.837 for furcation defects at 50% IoU. In Fig. 6, The input and output images of the radiographs of some patients in the AI system were shown as an example.

Fig. 6
figure 6

Input and output data of some radiographs of patients. a-b. Total alveolar bone loss; c-d. Horizontal bone loss; e. Vertical bone loss, f: Furcation defect

Also, ROC curve analysis and precision-recall results for all periodontal parameters were presented in Fig. 7.

Fig. 7
figure 7

ROC curve analysis and precision-recall results of all parameters. (a) Total alveolar bone loss; (b) Horizontal bone loss; (c) Vertical bone loss; (d) Furcation defect

Observations

In addition, the input and output images of each case in the testing set were reviewed by the observers who labeled them, and it was analyzed in which situations the system made more mistakes. When the output images of total alveolar bone loss were examined, it was observed that the success of the system was quite good, and there were few errors in the detection of only some deep angular bone loss areas, and furcation areas. When the output images of horizontal bone loss were examined, it was analyzed that the system could not diagnose in some radiographs in large interdental regions caused by tooth deficiency. It was observed that some vertical bone losses were determined as horizontal bone loss in regions with partial superimposition of anatomical structures. The most common error made by the system in detecting vertical bone losses was to perceive very minimal bone angulations as vertical bone losses. The system identified some of the bone losses in the interdental regions of roots close to each other as furcation defects. Another mistake was that the AI model overlooked and failed to detect some of the initial level furcation defects.

Discussion

In recent years, the use of AI systems in medicine and dentistry in image interpretation has attracted great interest and CNN systems change these areas rapidly [24]. However, the number of studies conducted in the field of periodontology is still limited. The current study aims to automatically evaluate panoramic radiographs by an AI system and to determine the success of the related system in detecting periodontal disease findings on radiographs. The results of this study showed that AI systems can be a decision-support mechanism for dentists in the diagnosis of periodontal disease, one of the most common diseases in the world.

Different architectures can be used in AI-based studies. U-Net architecture, a CNN-based AI architecture, is one of the image segmentation techniques that enable the evaluation of images in the medical field with AI systems and provides a more precise and successful evaluation with fewer training sets [35, 36]. Therefore, in the current study, image processing was performed using the U-Net architecture.

In the literature, there were many studies in which AI algorithms were used in the processing and interpretation of patient data, which played a role in diagnosis and treatment planning, such as dental radiographs [20, 37,38,39], intraoral photographs [40,41,42,43] and pathology images [44]. In addition, these systems were also used in situations such as pre-examination evaluation and risk estimation [45]. Also, it was seen that some of the studies aiming to determine periodontal disease from 2D dental radiographs using AI systems were performed on periapical radiographs [21, 28, 46] and some on panoramic radiographs [3, 6, 25, 29, 30].

In one of the studies in the literature, Lee et al. (2018) used 1740 periapical radiographs and evaluated the prediction success of the CNN algorithm they developed. In the results of this study, they reported that the accuracy of determining premolar teeth with periodontal damage was 81% and 76.7% for molars [27]. Similar to this study, Lee et al. (2021) performed another study on 693 periapical radiographs [18]. In this study, it was aimed to find the radiographic bone loss by grouping it according to the severity of destruction (stages I-II-III) [18] High success rates were found in the detection of periodontal disease in the results of these two studies.

On the other hand, Khan et al. (2021) tried to find dental problems such as caries, bone losses, and furcation defects on 206 periapical images using different AI architectures (U-Net, X Net, and Seg Net) in their study [28]. Although this study was similar to our study in terms of using the segmentation method in the detection of periodontal bone loss, no evaluation was made in terms of bone destruction characteristics (horizontal, vertical) [28]. Another difference was Khan et al. (2021) performed their studies on periapical radiographs and with fewer data [28]. Although the evaluations were made on different types of radiographs, it could be said that the success rates of our study were higher. We think that the reason for the higher success rates of the AI system developed in our study may be due to the use of a larger data set. Because as the number of data used for training in AI studies increases, the success rates of the model also increase [47].

In another study, Kurt Bayrakdar et al. (2021) used Google Net Inception v3 architecture on a large data set (2276 panoramic radiographs). They evaluated the success of the AI systems they developed in the determination of radiographs of periodontitis cases [6]. In this study, 1137 radiographs of patients with bone loss and panoramic radiographs of 1139 periodontal healthy individuals were used for the development of the AI model. F1 score and accuracy were found as high as 91% in this study [6]. However, no labeling was made, only the classification method was used to train the radiographs in the form of patient/healthy and to find them by the system. In this respect, they are quite different from our study in terms of this study planning. Based on this information, it can be said that the segmentation technique we used in our study is the most advantageous method and provides the physician with more detailed information for diagnosis and treatment planning.

There were also studies in the literature comparing the diagnosis of dentists with different experiences and AI model predictions. The results obtained in this way are undoubtedly more interpretable and reveal the success of the system more clearly. For example, Krois et al. (2019) compared the evaluation of 6 dentists and the results of AI in a panoramic radiography study with the CNN technique. They reported the accuracy, specificity, and sensitivity rates for the system as 81% in the results of their study [25]. Since this study deals with the evaluation of many dentists, it was a more comprehensive and superior study in terms of planning. However, in the study of Krois et al. (2019), periodontal bone loss patterns were not classified as horizontal or vertical.

In another similarly planned study, Kim et al. (2020) compared the evaluations of 5 different clinicians with the performance of AI in their study for the determination of bone resorption sites using 12179 panoramic radiography [29]. They reported that while the average F1 scores of the clinicians’ results in determining bone loss were 69%, AI showed higher success and the F1 score was 75%. This study reveals the success of AI systems in radiography interpretation and shows promise for the future use of these systems [29]. Also, Chang et al. (2020) tried to perform the determination of bone loss from panoramic radiographs without any evaluation of bone destruction angulation and defect type [3]. In this study, the staging was tried to be made according to the 2017 periodontitis classification by calculating the amount of bone destruction and destruction [3].

Finally, Jiang et al. (2022) tried to detect periodontal bone destruction in their study using the CNN model using 640 panoramic radiographs. This study is the most similar to our study in the literature [30] because Jiang et al. (2022) also determined the bone loss patterns of periodontal disease in the form of vertical/horizontal/furcation defects in their study [30]. Although it was similar to our study in this respect, the object detection method was used in this study for labeling. The segmentation method we use is a much more advantageous method since it determines the defective area with its borders. Because it provides more detailed information to the dentist to determine the severity of the disease and to plan the treatment in the next process. In addition, in this method, the defect area is processed like a detailed map and provides more advanced diagnostic support visually. On the other hand, when two studies were compared in detail, it was seen that they used multiple observers, and two different AI architectures (U-Net and Yolo-v4), and included the severity of the disease in their assessment.

One of the limitations of the study was that the decision of multiple observers was not separately compared with the predictions of AI. Another limitation was that no measurements were made to determine disease severity in the study and it was aimed to detect only bone losses. More extensive research could have been done using calibrated panoramic radiographs, based on direct measurement, or to determine the percentage of root affected. However, It should be noted that in all 2D radiographic imaging techniques, the evaluation of bone craters, lamina dura, and periodontal bone level is limited to the projection geometry and superposition of adjacent anatomical structures [48]. In other words, these radiographs do not provide clear and reliable information for measurement and treatment planning. For this reason, performing AI-based studies with three-dimensional (3D) radiographic imaging techniques such as cone-beam computed tomography systems (CBCT) will prevent this limitation. Undoubtedly, as the number of studies in this field increases, AI algorithms with stronger decision-support capability and providing much more detailed information will be developed. Despite the limitations, our study is promising for such studies to be carried out in the future. In addition, the performance of the AI model in determining vertical bone losses in our study was lower than other periodontal parameters. It is a known fact that vertical bone losses are less common than horizontal bone loss, and therefore, the number of labels for vertical bone losses was more limited in our study [14]. We think that the number of labels used in the detection of vertical bone losses with AI in our study was therefore less, and this situation caused the performance of the model to be found to be lower. Because the most important thing in the success of AI studies is to work with large data sets. This limitation could be overcome, albeit to a limited extent, by using different AI architectures and making different technical plans. For example, in our study, cross-validation techniques could be used during model development and multi-class training could be done. On the other hand, the more important reason for the low success rates for the vertical bone loss parameter also may be the incomplete understanding of the outline of the alveolar bone crest in 1, 2, and 3-walled defects. We think that interobserver agreements were also lower for this parameter due to this limitation, which may be due to the limited view of panoramic radiography. This diagnostic difficulty can be eliminated by using some radiography techniques that provide more detailed images in future studies.

Conclusions

When the literature was examined, it was seen that the use of AI systems in dental radiographs for periodontal status determination has very successful and promising results. To the best of our knowledge, the present study is the first study aimed at detecting bone resorption and its patterns using deep learning algorithms and segmentation methods. These bone loss patterns play an important role in periodontal treatment planning and the prognosis of teeth. With this aspect, these systems are candidates to be a decision-support mechanism for dentists in radiographic interpretation. The development of these systems with more data sets will increase success rates. There is a need for more comprehensive studies on 2D and 3D radiographs in this regard.

Data availability

The datasets generated and/or analyzed during the current study are not publicly available due to the conditions specified when receiving approval from the ethics committee but are available from the corresponding author upon reasonable request.

Abbreviations

CNN:

Convolutional Neural Network

AI:

Artificial intelligence

2D:

Two-dimensional

IoU:

Intersection over Union

ADAM:

Adaptive Moment Estimation

TP:

Numbers of true positives

FP:

Numbers of false positives

FN:

Numbers of false negatives

ROC:

Receiver operating characteristic

AUC:

Area under the ROC curve

3D:

Three-dimensional

CBCT:

Cone-beam computed tomography systems

STARD:

Standards for Reporting Diagnostic

References

  1. Dentino A, Lee S, Mailhot J, Hefti AF. Principles of periodontology. Periodontol 2000. 2013;61:16–53.

    Article  PubMed  Google Scholar 

  2. Bourgeois D, Inquimbert C, Ottolenghi L, Carrouel F. Periodontal pathogens as Risk factors of Cardiovascular diseases, Diabetes, Rheumatoid Arthritis, Cancer, and Chronic Obstructive Pulmonary Disease-Is there cause for Consideration? Microorganisms 2019;7.

  3. Chang H-J, Lee S-J, Yong T-H, Shin N-Y, Jang B-G, Kim J-E, et al. Deep learning hybrid method to automatically diagnose periodontal bone loss and stage periodontitis. Sci Rep. 2020;10:1–8.

    Google Scholar 

  4. Mol A. Imaging methods in periodontology. Periodontol 2000. 2004;34:34–48.

    Article  PubMed  Google Scholar 

  5. Scarfe WC, Azevedo B, Pinheiro LR, Priaminiarti M, Sales MAO. The emerging role of maxillofacial radiology in the diagnosis and management of patients with complex periodontitis. Periodontol 2000. 2017;74:116–39.

    Article  PubMed  Google Scholar 

  6. Kurt Bayrakdar S, Özer Ç, Bayrakdar IS, Orhan K, Bilgir E, Odabaş A, et al. Success of artificial intelligence system in determining alveolar bone loss from dental panoramic radiography images. Cumhuriyet Dent J. 2020;23:318–24.

    Article  Google Scholar 

  7. Rushton VE, Horner K. The use of panoramic radiology in dental practice. J Dent. 1996;24:185–201.

    Article  PubMed  Google Scholar 

  8. Clerehugh V, Tugnait A. Diagnosis and management of periodontal diseases in children and adolescents. Periodontol 2000. 2001;26:146–68.

    Article  PubMed  Google Scholar 

  9. Ivanauskaite D, Lindh C, Rangne K, Rohlin M. Comparison between Scanora panoramic radiography and bitewing radiography in the assessment of marginal bone tissue. Stomatologija. 2006;8(1):9–15. PMID: 16687909.

    PubMed  Google Scholar 

  10. de Faria Vasconcelos K, Evangelista KM, Rodrigues CD, Estrela C, de Sousa TO, Silva MA. Detection of periodontal bone loss using cone beam CT and intraoral radiography. Dentomaxillofac Radiol. 2012;41:64–9.

    Article  PubMed  PubMed Central  Google Scholar 

  11. Haas LF, Zimmermann GS, De Luca Canto G, Flores-Mir C, Corrêa M. Precision of cone beam CT to assess periodontal bone defects: a systematic review and meta-analysis. Dentomaxillofac Radiol. 2018;47:20170084.

    Article  PubMed  Google Scholar 

  12. Persson RE, Hollender LG, Laurell L, Persson GR. Horizontal alveolar bone loss and vertical bone defects in an adult patient population. J Periodontol. 1998;69:348–56.

    Article  PubMed  Google Scholar 

  13. Gomes-Filho IS, Sarmento VA, de Castro MS, da Costa NP, da Cruz SS, Trindade SC, et al. Radiographic features of periodontal bone defects: evaluation of digitized images. Dentomaxillofac Radiol. 2007;36:256–62.

    Article  PubMed  Google Scholar 

  14. Jayakumar A, Rohini S, Naveen A, Haritha A, Reddy K. Horizontal alveolar bone loss: a periodontal orphan. J Indian Soc Periodontol. 2010;14:181–5.

    Article  PubMed  PubMed Central  Google Scholar 

  15. Braun X, Ritter L, Jervøe-Storm PM, Frentzen M. Diagnostic accuracy of CBCT for periodontal lesions. Clin Oral Investig. 2014;18:1229–36.

    Article  PubMed  Google Scholar 

  16. Papapanou PN, Sanz M, Buduneli N, Dietrich T, Feres M, Fine DH, et al. Periodontitis: Consensus report of workgroup 2 of the 2017 World workshop on the classification of Periodontal and Peri-implant diseases and conditions. J Periodontol. 2018;89(Suppl 1):173–s82.

    Google Scholar 

  17. Ertaş K, Pence I, Cesmeli MS, Ay ZY. Determination of the stage and grade of periodontitis according to the current classification of periodontal and peri-implant diseases and conditions (2018) using machine learning algorithms. J Periodontal Implant Sci. 2023;53(1):38–53. https://doi.org/10.5051/jpis.2201060053.

    Article  PubMed  Google Scholar 

  18. Lee CT, Kabir T, Nelson J, Sheng S, Meng HW, Van Dyke TE, et al. Use of the deep learning approach to measure alveolar bone level. J Clin Periodontol. 2022;49:260–9.

    Article  PubMed  Google Scholar 

  19. Sen D, Chakrabarti R, Chatterjee S, Grewal DS, Manrai K. Artificial intelligence and the radiologist: the future in the Armed Forces Medical Services. BMJ Mil Health. 2020;166:254–6.

    Article  PubMed  Google Scholar 

  20. Kurt Bayrakdar S, Orhan K, Bayrakdar IS, Bilgir E, Ezhov M, Gusarev M, et al. A deep learning approach for dental implant planning in cone-beam computed tomography images. BMC Med Imaging. 2021;21:86.

    Article  PubMed  Google Scholar 

  21. Lee JH, Kim DH, Jeong SN, Choi SH. Detection and diagnosis of dental caries using a deep learning-based convolutional neural network algorithm. J Dent. 2018;77:106–11.

    Article  PubMed  Google Scholar 

  22. Fukuda M, Inamoto K, Shibata N, Ariji Y, Yanashita Y, Kutsuna S, et al. Evaluation of an artificial intelligence system for detecting vertical root fracture on panoramic radiography. Oral Radiol. 2020;36:337–43.

    Article  PubMed  Google Scholar 

  23. Bilgir E, Bayrakdar İ, Çelik Ö, Orhan K, Akkoca F, Sağlam H, et al. An artifıcial ıntelligence approach to automatic tooth detection and numbering in panoramic radiographs. BMC Med Imaging. 2021;21:124.

    Article  PubMed  PubMed Central  Google Scholar 

  24. Bayrakdar IS, Orhan K, Çelik Ö, Bilgir E, Sağlam H, Kaplan FA, et al. A U-Net Approach to apical lesion segmentation on panoramic radiographs. Biomed Res Int. 2022;2022:7035367.

    Article  PubMed  PubMed Central  Google Scholar 

  25. Krois J, Ekert T, Meinhold L, Golla T, Kharbot B, Wittemeier A, et al. Deep learning for the Radiographic detection of Periodontal Bone loss. Sci Rep. 2019;9:8495.

    Article  PubMed  PubMed Central  Google Scholar 

  26. Thanathornwong B, Suebnukarn S. Automatic detection of periodontal compromised teeth in digital panoramic radiographs using faster regional convolutional neural networks. Imaging Sci Dent. 2020;50:169–74.

    Article  PubMed  PubMed Central  Google Scholar 

  27. Lee JH, Kim DH, Jeong SN, Choi SH. Diagnosis and prediction of periodontally compromised teeth using a deep learning-based convolutional neural network algorithm. J Periodontal Implant Sci. 2018;48:114–23.

    Article  PubMed  PubMed Central  Google Scholar 

  28. Khan HA, Haider MA, Ansari HA, Ishaq H, Kiyani A, Sohail K, et al. Automated feature detection in dental periapical radiographs by using deep learning. Oral Surg Oral Med Oral Pathol Oral Radiol. 2021;131:711–20.

    Article  PubMed  Google Scholar 

  29. Kim J, Lee HS, Song IS, Jung KH. DeNTNet: deep neural transfer network for the detection of periodontal bone loss using panoramic dental radiographs. Sci Rep. 2019;9:17615.

    Article  PubMed  PubMed Central  Google Scholar 

  30. Jiang L, Chen D, Cao Z, Wu F, Zhu H, Zhu F. A two-stage deep learning architecture for radiographic staging of periodontal bone loss. BMC Oral Health. 2022;22:106.

    Article  PubMed  PubMed Central  Google Scholar 

  31. Helmi MF, Huang H, Goodson JM, Hasturk H, Tavares M, Natto ZS. Prevalence of periodontitis and alveolar bone loss in a patient population at Harvard School of Dental Medicine. BMC Oral Health. 2019;19:254. https://doi.org/10.1186/s12903-019-0925-z.

    Article  PubMed  PubMed Central  Google Scholar 

  32. Wylleman A, Van Der Veken D, Teughels W, Quirynen M, Laleman I. Alveolar bone level at deciduous molars in flemish children: a retrospective, radiographic study. J Clin Periodontol. 2020;47:660–7. https://doi.org/10.1111/jcpe.13280.

    Article  PubMed  Google Scholar 

  33. Castro LO, Castro IO, de Alencar AHG, Valladares-Neto J, Estrela C. Cone beam computed tomography evaluation of distance from cementoenamel junction to alveolar crest before and after nonextraction orthodontic treatment. Angle Orthod. 2015;86:543–9. https://doi.org/10.2319/040815-235.1.

    Article  PubMed  PubMed Central  Google Scholar 

  34. Ronneberger O, Fischer P, Brox T. U-net: Convolutional networks for biomedical image segmentation. International Conference on Medical image computing and computer-assisted intervention: Springer; 2015. p. 234 – 41.

  35. Zhou Z, Siddiquee MMR, Tajbakhsh N, Liang J. UNet++: A nested U-Net Architecture for Medical Image Segmentation. Deep Learn Med Image Anal Multimodal Learn Clin Decis Support (2018). 2018;11045:3–11.

    Article  PubMed  Google Scholar 

  36. Du G, Cao X, Liang J, Chen X, Zhan Y. Medical image segmentation based on u-net: a review. J Imaging Sci Technol. 2020;64:1–12.

    Article  Google Scholar 

  37. Lee S, Kim D, Jeong HG. Detecting 17 fine-grained dental anomalies from panoramic dental radiography using artificial intelligence. Sci Rep. 2022;12:5172.

    Article  PubMed  PubMed Central  Google Scholar 

  38. Kılıc MC, Bayrakdar IS, Çelik Ö, Bilgir E, Orhan K, Aydın OB, et al. Artificial intelligence system for automatic deciduous tooth detection and numbering in panoramic radiographs. Dentomaxillofac Radiol. 2021;50:20200172.

    Article  PubMed  PubMed Central  Google Scholar 

  39. Pauwels R, Brasil DM, Yamasaki MC, Jacobs R, Bosmans H, Freitas DQ, et al. Artificial intelligence for detection of periapical lesions on intraoral radiographs: comparison between convolutional neural networks and human observers. Oral Surg Oral Med Oral Pathol Oral Radiol. 2021;131:610–6.

    Article  PubMed  Google Scholar 

  40. Zhang X, Liang Y, Li W, Liu C, Gu D, Sun W, et al. Development and evaluation of deep learning for screening dental caries from oral photographs. Oral Dis. 2022;28:173–81.

    Article  PubMed  Google Scholar 

  41. Takahashi T, Nozaki K, Gonda T, Mameno T, Ikebe K. Deep learning-based detection of dental prostheses and restorations. Sci Rep. 2021;11:1960.

    Article  PubMed  PubMed Central  Google Scholar 

  42. Fu Q, Chen Y, Li Z, Jing Q, Hu C, Liu H, et al. A deep learning algorithm for detection of oral cavity squamous cell carcinoma from photographic images: a retrospective study. EClinicalMedicine. 2020;27:100558.

    Article  PubMed  PubMed Central  Google Scholar 

  43. Li W, Liang Y, Zhang X, Liu C, He L, Miao L, et al. A deep learning approach to automatic gingivitis screening based on classification and localization in RGB photos. Sci Rep. 2021;11:16831.

    Article  PubMed  PubMed Central  Google Scholar 

  44. Zehra T, Shaikh A, Shams M. Dawn of artificial intelligence - enable digital Pathology in Pakistan: a paradigm shift. J Pak Med Assoc. 2021;71:2683–4.

    PubMed  Google Scholar 

  45. Alhazmi A, Alhazmi Y, Makrami A, Masmali A, Salawi N, Masmali K, et al. Application of artificial intelligence and machine learning for prediction of oral cancer risk. J Oral Pathol Med. 2021;50:444–50.

    Article  PubMed  Google Scholar 

  46. Alotaibi G, Awawdeh M, Farook FF, Aljohani M, Aldhafiri RM, Aldhoayan M. Artificial intelligence (AI) diagnostic tools: utilizing a convolutional neural network (CNN) to assess periodontal bone level radiographically-a retrospective study. BMC Oral Health. 2022;22:399.

    Article  PubMed  PubMed Central  Google Scholar 

  47. Surya L. An exploratory study of AI and Big Data, and it’s future in the United States. Int J Creative Res Thoughts (IJCRT), ISSN 2015;2320 – 882.

  48. Acar B, Kamburoğlu K. Use of cone beam computed tomography in periodontology. World J Radiol. 2014;28(5):6.

    Google Scholar 

Download references

Acknowledgements

This study has been supported by Eskisehir Osmangazi University Scientific Research Projects Coordination Unit under project number 202045E06.

Funding

There is no funding. 

Author information

Authors and Affiliations

Authors

Contributions

All authors have made substantial contributions to the conception and design of the study. SKB, ISB, MBY, and NS have been involved in data collection and ÖÇ has been involved in data analysis. SKB, ISB, MBY, NS, OK, BCUS, and RJ have been involved in the data interpretation and drafting of the manuscript. SKB, ISB, BK, and KO have been involved in revising it critically and have given final approval for the version to be published. * The all participants in this study gave informed consent (both oral and written).

Corresponding author

Correspondence to Sevda Kurt-Bayrakdar.

Ethics declarations

Ethics approval and consent to participate

The research protocol was approved by the Eskisehir Osmangazi University Ethics Committee (decision number: 2019 − 227). All participants gave informed consent. We confirmed that the study was carried out according to the relevant guidelines and regulations of the Helsinki Declaration.

Consent for publication

Additional informed consent was obtained from all individual participants included in the study.

Competing interests

The authors declare that there is no conflict of interest.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Kurt-Bayrakdar, S., Bayrakdar, İ.Ş., Yavuz, M.B. et al. Detection of periodontal bone loss patterns and furcation defects from panoramic radiographs using deep learning algorithm: a retrospective study. BMC Oral Health 24, 155 (2024). https://doi.org/10.1186/s12903-024-03896-5

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12903-024-03896-5

Keywords