Skip to main content

Automatic classification of 3D positional relationship between mandibular third molar and inferior alveolar canal using a distance-aware network

Abstract

The purpose of this study was to automatically classify the three-dimensional (3D) positional relationship between an impacted mandibular third molar (M3) and the inferior alveolar canal (MC) using a distance-aware network in cone-beam CT (CBCT) images. We developed a network consisting of cascaded stages of segmentation and classification for the buccal-lingual relationship between the M3 and the MC. The M3 and the MC were simultaneously segmented using Dense121 U-Net in the segmentation stage, and their buccal-lingual relationship was automatically classified using a 3D distance-aware network with the multichannel inputs of the original CBCT image and the signed distance map (SDM) generated from the segmentation in the classification stage. The Dense121 U-Net achieved the highest average precision of 0.87, 0.96, and 0.94 in the segmentation of the M3, the MC, and both together, respectively. The 3D distance-aware classification network of the Dense121 U-Net with the input of both the CBCT image and the SDM showed the highest performance of accuracy, sensitivity, specificity, and area under the receiver operating characteristic curve, each of which had a value of 1.00. The SDM generated from the segmentation mask significantly contributed to increasing the accuracy of the classification network. The proposed distance-aware network demonstrated high accuracy in the automatic classification of the 3D positional relationship between the M3 and the MC by learning anatomical and geometrical information from the CBCT images.

Peer Review reports

Introduction

Extraction of the mandibular third molar (M3) is one of the most common surgeries in the oral and maxillofacial field [1,2,3]. Inferior alveolar nerve injury, an important surgical complication, occurs in about 0.35–8.4% of impacted M3 extractions [4]. The positional relationship between an impacted M3 and the inferior alveolar canal (MC) is the main factor that determines the risk of inferior alveolar nerve injury [5]. Panoramic radiographs have been used for preoperative imaging to predict and minimize such complications [2, 6, 7]. To predict the relationship between the M3 and the MC in panoramic radiographs, clinicians have to infer specific radiological signs (e.g., darkening or narrowing of the root, bifid apex, and interruption or diversion of the cortical outline of the MC) [8]. However, the anatomical position of the impacted M3 in relation to the MC cannot be determined easily because of superimposition and distortion of the surrounding anatomical structures in the two-dimensional (2D) panoramic radiographs [5, 9].

Cone-beam CT (CBCT) has been widely used to overcome the limitations of 2D panoramic radiographs in the oral and maxillofacial field [10]. CBCT has the advantages of lower radiation dose and cost compared with multi-detector CT, and shows three-dimensional (3D) information of anatomical structures including the teeth, jaw bone, and inferior alveolar nerve [9, 11, 12]. The 3D positional relationship between the MC and the M3 in the buccolingual direction can be determined using the cross-sectional images of CBCT. In this respect, six types of relationships were established based on the distance between the M3 and the MC, the level of contact, and the 3D positional relationship in the CBCT images [7]. The relationships were classified quantitatively based on the presence of the contact, periarticular, interradicular, buccal, and inferior positions [13]. In a previous analysis of risk factors for nerve damage with paresthesia after extraction of the M3, Wang et al. stated that the direct contact relationships between the inferior alveolar nerve and the root of the M3 as well as the buccal or lingual positions observed in preoperative CBCT images were important factors [13, 14]. Additionally, a previous study reported that the possibility of damage to the inferior alveolar nerve was higher if it was located lingually [12]. The rate of MC passage to the lingual side of the M3 root was high when the MC and the M3 were in contact [15]. Therefore, confirming the relative buccal or lingual relationship of the MC with the M3 is an essential procedure for accurate risk assessment and treatment planning to avoid or reduce inferior alveolar nerve damage during an M3 extraction.

Anatomic segmentation of the MC and M3 structures is essential when making an appropriate surgical plan based on their positional relationship to avoid or reduce nerve damage. However, segmentation of the MC and M3 and determination of the 3D positional relationship between them in relevant multiple slices of CBCT images is time-consuming and labor-intensive. Automatic segmentation methods have been proposed including level-set methods [16,17,18,19,20], template-based fitting methods [21], and statistical shape models [22, 23]. However, these methods have some limitations, such as initialization problems, transformation vulnerability, and additional manual annotation, which need improvement for fully automatic segmentation. Recently, many studies based on deep learning have been performed on the segmentation and classification of anatomical structures or lesions in medical or dental images [24,25,26], and they have shown impressive performance improvement in terms of overcoming limitations [27,28,29,30,31,32,33]. Several studies for detecting and segmenting the MC in CBCT images have also been performed using deep learning [27, 29]. The proximity and contact relationship between the M3 and the MC were classified in CBCT images using a ResNet-based deep learning model [34]. To the best of our knowledge, no previous studies have attempted to apply deep learning to the classification of the relative buccal or lingual relationships between the M3 and the MC. Thus, in this study, we focused on developing an end-to-end automatic method based on deep learning to replace the time- and labor-consuming process of determining the relative buccal or lingual relationship between the M3 and the MC.

We hypothesized that a deep learning model could automatically determine the relative buccal or lingual relationship between the M3 and the MC using distance information in CBCT images. Therefore, the purpose of this study was to automatically classify the 3D positional relationship between an impacted M3 and the MC using a 3D distance-aware network that consisted of cascaded stages of segmentation and classification of CBCT images. Our main contributions were that we proposed a distance-aware network for automatic and accurate classification of the 3D positional relationship between the M3 and the MC by learning anatomical and geometrical information. In addition, we applied the signed distance map (SDM) generated from the segmentation mask as a multichannel volumetric input in the 3D distance-aware classification network to guide the position and distance relationships between the M3 and the MC.

Materials and methods

Data acquisition and preparation

This study was performed with approval from the Institutional Review Board (IRB) of Seoul National University Dental Hospital (ERI18001). The ethics committee approved the waiver for informed consent because this was a retrospective study. The study was performed in accordance with the Declaration of Helsinki. CBCT images were obtained from 50 patients (27 females and 23 males; mean age 25.56 ± 6.73 years) who underwent dental implant surgery or extraction of the M3 at Seoul National University Dental Hospital in 2019–2020. The images had dimensions of 841 × 841 × 289 pixels, voxel sizes of 0.2 × 0.2 × 0.2 mm3, and 16-bit depth and were obtained at 80 kVp and 8 mA using a CBCT (CS9300; Carestream Health, New York, USA).

The anatomical structures of the M3 and the MC in the CBCT images were manually annotated by an oral and maxillofacial radiologist using the 3D Slicer for Windows 10 (Version 4.10.2; MIT, Massachusetts, USA) [35]. The ground truth of the MC was established by annotation of both the inferior alveolar nerve and the cortical bone. For multiclass segmentation of the M3 and the MC by deep learning, we prepared 64 volumes (32 patients) for the training dataset and 36 volumes (18 patients) for the test dataset from all CBCT images (50 patients), where the right mandible volume was horizontally flipped to match the left (Table 1). The training and test datasets used for segmentation comprised 3546 and 1804 axial images, respectively (Table 1). The 2D axial images were automatically cropped into images of 512 × 512 pixels centered at the region of the mandible as an input of the segmentation network.

Table 1 The dataset configuration of CBCT images used for deep learning

The buccolingual relationship of the M3 and the MC was determined by analysis of successive slices in multiplanar CBCT images by an oral and maxillofacial radiologist. The passing direction and path of the MC were evaluated based on the lamina dura of the M3. If the MC directly contacted or passed in close proximity to the inner surface of the M3 root, it was considered a lingual class, but if it directly contacted or passed in close proximity to the outer surface of the M3 root, it was classified as a buccal class (Fig. 1). We excluded the CBCT images in which it was difficult to determine the positional relationship between the MC and M3 for classification dataset. In our study, the radiologist annotating the images was unaware of critical information that could bias their assessments. This information included the patient's dental history, the patient's clinical symptoms, or demographic information. When the interpreter was not aware of the patient's clinical information, they were less likely to make assumptions about the patient's condition based on that information. This could lead to a more accurate interpretation of the image.

Fig. 1
figure 1

Visualization of mandibular canals running along the buccal side of the impacted mandibular third molar (a, b) and of mandibular canals running along the lingual side of the impacted mandibular third molar (c, d) in 3D and 2D axial slices of a CBCT image and the ground truth of segmentation

For classification by deep learning, we prepared 43 buccal and 8 lingual cases for the training dataset and 20 buccal and 11 lingual cases for the test dataset from the whole dataset of 63 buccal and 19 lingual cases. The CBCT images for the input of the classification network were cropped from the segmentation results into a volume of 256 × 256 × 32 pixels. We estimated the minimum required sample size to detect significant differences in the accuracy between the distance-aware network and the other networks when both assessed the same subjects. Based on an effect size of 0.50, a significance level of 0.05, and a statistical power of 0.80, we obtained a sample size of N = 128 (G* Power for Windows 10, Version 3.1.9.7; Universität Düsseldorf, Germany). Therefore, we split all CBCT images into 3546 and 1804 axial images for the training and test datasets, respectively.

Overall architecture of the distance-aware network

We designed a distance-aware network consisting of cascaded stages of segmentation and classification (Fig. 2). In the segmentation stage, a U-Net of DenseNet121 [36] backbone with the input of 2D axial images (S-Net) was used for multiclass segmentation of the M3 and the MC, simultaneously. In the classification stage, a 3D distance-aware classification network with input of multiple volumes (C-Net) was designed for classifying the buccal-lingual relationship of the M3 and the MC by learning their 3D anatomical and geometrical information.

Fig. 2
figure 2

The 3D distance-aware network consisting of segmentation and classification stages for classifying the positional relationship between the third molar and the mandibular canal. In the segmentation stage, the third molar and the mandibular canal were simultaneously segmented using Dense121 U-Net. In the classification stage, the 3D distance-aware network with inputs of CBCT volume and the distance map classified the buccal-lingual positional relationship between the third molar and the mandibular canal

For multiclass segmentation of the M3 and the MC, we used a pre-trained DenseNet121 backbone as the encoder of the U-Net, which consisted of multiple densely connected layers and transition layers to improve feature propagation and alleviate the vanishing gradient problem (Fig. 2). The decoder was composed of a five-level structure where each level consisted of a 2 × 2 up-sampling layer, a skip connection, and two convolutional blocks. Each convolutional block consisted of a 3 × 3 convolutional filter, a batch-normalization layer, and a rectified linear unit (ReLU) activation function. The SoftMax activation function was applied to the last activation layer for outputting multiclass segmentation of both the M3 and the MC. Several U-shaped networks such as SegNet [37], simple U-Net [38], and Attention U-Net [39] were also used as segmentation networks for performance comparison.

For classification of the positional relationship between the M3 and the MC, we designed a 3D distance-aware network (C-Net) consisting of a five-level structure where each level consisted of a 3 × 3 × 3 convolutional layer, a batch-normalization layer, the ReLU activation function, and a 2 × 2 × 2 max-pooling layer (Fig. 2). The original CBCT volume, binary segmentation mask from predictions, and corresponding SDM were used as multichannel inputs for the C-Net. The input volumes were centered at the point where the M3 and the MC were closest. In the output layer, the class probability for the relative buccal-lingual position of the MC was calculated using the SoftMax activation function following the global average pooling layer and the dense layer. The predictions by the segmentation networks of SegNet, simple U-Net, Attention U-Net, and Dense U-Net were also used as the input of the classification network for performance comparison.

The segmentation networks were trained using the Dice similarity coefficient loss and the Adam optimizer with a learning rate of 0.00025 that was reduced on a plateau by a factor of 0.5 every 25 epochs over 300 epochs with a batch size of 8. The classification network was trained using binary cross-entropy and the Adam optimizer with a learning rate of 0.001 that was reduced on a plateau by a factor of 0.5 every 25 epochs over 100 epochs with a batch size of 1. Analyses were implemented with Python3 based on Keras with a Tensorflow backend using a single NVIDIA Titan RTX GPU 24 GB.

Signed distance map (SDM) for positional information

In the classification process, the network learned the 3D anatomical and geometrical information from the multichannel volume inputs simultaneously. The SDM calculated from the mask result of the segmentation prediction was used as an input for learning the geometrical information, while the original CBCT image was used for learning the 3D anatomical information (Fig. 3). The geometric SDM between the M3 and the MC from the segmentation prediction was calculated as the signed distance transform (SDT) [40]. The SDT was defined as the Euclidean distance from the nearest background point:

$$SDT\left(x\right)= \left\{\begin{array}{c}d\left(x,\partial M\right) if x\in M\\ -d\left(x,\partial M\right) if x\in {M}^{c}\end{array}\right.$$
(1)

where \(x\) is the metric space, \(M\) is the metric space of the M3 and the MC, and \(\partial M\) is the boundary of \(M\) [41, 42]. For any \(\in X,\) \(\left(x,\partial M\right):=\underset{y\in \partial M}{\mathrm{inf}}d(x,y)\), where inf denotes the infimum.

Fig. 3
figure 3

Examples of axial image (left), segmentation mask (middle), and SDM (right) of the mandibular third molar and the canal. The demarcation line was observed radially in the signed distance map, and the difference in direction was observed when the mandibular canals traveled along the lingual side or buccal side of the M3

The SDM was derived from the 3D SDT considering the internal shape of the object and the external relationship (Eq. 2), where B denoted the binary segmentation mask of the M3 and the MC. When calculating the SDM, the inside and outside of a boundary of an object had a negative and a positive value, respectively. In the SDM from the predicted M3 and MC segmentations, the values between the boundaries of the two objects formed a line that had a constant value. This information helped the networks to learn the geometrical relationship between the M3 and the MC.

$$SDM=\left(1-B\right)*SDT\left(1-B\right)-B*(SDT\left(B\right)-1)$$
(2)

where \(B\) denotes the binary segmentation mask of the M3 and the MC.

Performance evaluation for segmentation and classification

The decision-making process of the networks for buccolingual classification was visualized and verified in the attentional area of the classification networks with multichannel inputs using gradient-weighted class activation mappings (Grad-CAMs). Grad-CAM is a powerful tool for interpreting the decision-making process of a deep learning network by visualizing the learning results of the network in terms of heat maps [43]. Therefore, a qualitative evaluation was performed by confirming the differences in the attentional areas of the Grad-CAM among the classification methods based on the segmentation network and multichannel inputs.

The segmentation performance of the Dense121 U-Net was compared with those of other networks from SegNet [37], simple U-Net [38], and Attention U-Net [39]. The evaluation matrix included precision, recall, Dice similarity coefficient (DSC), intersection over union (IoU), 3D volumetric overlap error (VOE), and relative volume difference (RVD). All matrices were calculated as volume levels. Precision \((\frac{TP}{TP+FP}\)) was the rate of correctly predicted positive predictions, recall \((\frac{TP}{TP+FN}\)) was the rate of correctly predicted ground truths, and DSC \((\frac{2TP}{2TP+FN+FP})\) was a harmonic mean of precision and recall, where TP, FP, and FN denoted true positive, false positive, and false negative, respectively. VOE \((1- \frac{{V}_{gt}\cap {V}_{pred}}{{V}_{gt}\cup {V}_{pred}}\)) is the rate between intersection and union of two sets of segmentations, and RVD (\(\frac{|{V}_{gt}-{V}_{pred}|}{{V}_{gt}}\)) is the absolute volumetric size difference of the regions, where \({V}_{gt}\) and \({V}_{pred}\) represent the number of voxels for the ground truth and the predicted volumes, respectively. Higher values of DSC, precision, recall, and IoU and lower values of VOE and RVD indicate better segmentation performance. The precision-recall curve was also computed from the network’s segmentation output by varying the IoU threshold. Average precision (AP) was calculated as that across all recall values.

The classification performance of the Dense121 U-Net with a variety of configurations of volume inputs was compared with other networks with the same volume inputs. The classification performance was evaluated using sensitivity, specificity, accuracy, and the area under the receiver operating characteristic curve (AUC). Sensitivity \((\frac{TP}{TP+FN}\)) correctly identified the positive result for the actual class, specificity \((\frac{TN}{TN+FP}\)) correctly identified the negative result for the actual class, and accuracy \((\frac{TP+TN}{TP+TN+FP+FN}\)) provided the proportion of correct predictions for all classes, where TP, FP, TN, and FN denoted true positive, false positive, true negative, and false negative, respectively. The receiver operating characteristic curve (ROC) was also computed from the network classification output by varying the class probability for each network.

Results

Table 2 shows the quantitative results of the segmentation performance of IoU, DSC, precision, recall, RVD, and VOE by U-Net models. The performance of Dense121 U-Net, Attention U-Net, Simple U-Net, and SegNet was compared over the 36 volumes of the test dataset. The Dense121 U-Net achieved segmentation performances of IoU, DSC of precision, recall, RVD, and VOE of 0.872, 0.920, 0.946, 0.918, 0.038, and 0.088, respectively, for the M3, and of 0.766, 0.861, 0.911, 0.830, 0.135, and 0.248, respectively, for the MC. The Attention U-Net showed similar or better performance in some parameters. Generally, the Dense121 U-Net showed the highest scores in the segmentation performances for the M3 and the MC among the U-Net models.

Table 2 Segmentation performances (Mean (SD)) of intersection over union (IoU), Dice similarity coefficient score (DSC), precision (PR), recall (RC), relative volume difference (RVD), and volume of error (VOE) by Dense121 U-Net, Attention U-Net, simple U-Net, and SegNet for the mandibular third molar and the mandibular canal

Figure 4 presents the precision-recall curves of the segmentation performance for the M3 and the MC by SegNet, simple U-Net, Attention U-Net, and Dense121 U-Net. Of the tested networks, the Dense121 U-Net achieved the highest AP of 0.87, 0.96, and 0.94 for the M3, the MC, and both combined, respectively (Fig. 4). Figure 5 shows the line plots for the means of DSC and Hausdorff distance values calculated from the inferior axial slice to the superior one for the M3 and the MC volumes by the U-Net models. The Dense121 U-Net usually predicted the M3 and the MC volumes more accurately with smaller fluctuations in performance.

Fig. 4
figure 4

Precision-recall curves from segmentation results for M3 (a), MC (b), and both together (c)

Fig. 5
figure 5

The line plots of Dice similarity coefficient (a, c) and Hausdorff distance (b, d) by Dense121 U-Net, Attention U-Net, simple U-Net, and SegNet for the M3 (a, b) and MC (c, d). The metric values were calculated from the inferior axial slice to the superior slice in the volume

Figures 6 and 7 present the segmentation results for the M3 and the MC in 2D and 3D by the U-Net models. The predicted segmentations for the M3 and the MC by the networks and the ground truth are superimposed on the axial CBCT image (Fig. 6). The 2D segmentation results from the Dense121 U-Net show more true positives (yellow) and fewer false negatives (green) and false positives (red) in the MC and M3 areas compared to the other networks (Fig. 6). Particularly, the Dense121 U-Net successfully segmented the areas where the cortical bone of the MC was not clearly visible due to compression or obstruction by the root of the M3, while the other networks failed to segment these areas in the CBCT images (Fig. 6c–e). As a result, 3D segmentation by Dense121 U-Net exhibited better prediction results with improved structural continuity and boundary details for the MC and M3 volumes compared with the other networks (Fig. 7).

Fig. 6
figure 6

Segmentation results for the M3 and the MC by SegNet, simple U-Net, Attention U-Net, and Dense121 U-Net. The predicted segmentation masks of the M3 and the MC were superimposed on the ground truth in CBCT images. The yellow, green, and red regions represent true positive, false negative, and false positive, respectively

Fig. 7
figure 7

The 3D visualization of the ground truth and segmentation results from SegNet, simple U-Net, Attention U-Net, and Dense121 U-Net from left to right with buccal relationships between M3 and MC (a-c) and with lingual relationships (d-f). The red line passing through the M3, the main axis of the M3, shows the buccal-lingual relationship between the M3 and the MC on the ground truth

The results in Table 3 show the classification performances for the buccal-lingual relationship between the M3 and the MC by the 3D distance-aware networks with different configurations of input volumes from the original CBCT volume and the binary segmentation mask and SDM from Dense121 U-Net, Attention U-Net, simple U-Net, and SegNet. The classification performances by the 3D classification networks with the input of the original CBCT volume and/or the segmentation mask showed lower accuracy, sensitivity, specificity, and AUC. In contrast, higher classification performances were achieved by the 3D classification networks with the input of the SDM volume. The 3D classification network by Dense121 U-Net with the input of both the CBCT image and the SDM showed the highest performance, with accuracy, sensitivity, specificity, and AUC all having values of 1.00, whereas Attention U-Net had accuracy, sensitivity, specificity, and AUC values of 0.90, 0.90, 0.91, and 0.98, respectively. The ROC curve of classification performance by Dense121 U-Net and Attention U-Net using the CBCT image and the SDM is closest to the upper left, with AUC values of 1.00 and 0.98, respectively (Fig. 8). A lower segmentation accuracy of the segmentation networks represented with SDM caused by discontinuous, fragmented, and scattered structures led to a lower classification accuracy of the 3D distance aware networks. This showed difficulty in learning the proper relationships between the M3 and the MC. The classification performance was influenced by the performance of the segmentation network as the SDM input of the classification network was directly derived from the segmentation mask.

Table 3 Classification performances of accuracy, sensitivity, specificity, and area under the receiver operating characteristic curve (AUC) by the 3D distance-aware networks with different configurations of input volumes of the original CBCT image, segmentation mask, and SDM by Dense121 U-Net, Attention U-Net, simple U-Net, and SegNet. SDM: signed distance map
Fig. 8
figure 8

The classification performance of the 3D distance-aware networks based on the receiver operating characteristic (ROC) curves. The ROC curves and AUCs from the networks with segmentation networks of SegNet, simple U-Net, Attention U-Net, and Dense121 U-Net

Figure 9 shows activation heatmaps of the Grad-CAM superimposed on the CBCT images according to the segmentation networks and the input configuration of the classification network. The classification networks using the inputs of the original CBCT image and segmentation mask did not emphasize properly the anatomical regions related to the classification because the networks identified the entire mandible and M3 or external regions that were not relevant to the classification. However, the classification networks using the inputs of the CBCT image and SDM returned reasonable areas for classification. The areas of the M3 root and the medullary space surrounding the cortical layer of the MC were identified well in cases where the network predicted the lingual relationship, and the marginal area of the M3 and the medullary space near the cortical layer of the MC was identified well in cases with a buccal relationship.

Fig. 9
figure 9

Visualization of Grad-CAM results from the 3D distance-aware networks by SegNet, simple U-Net, Attention U-Net, and Dense121 U-Net with inputs of the CBCT image and segmentation mask (left), the SDM (middle), and the CBCT image and SDM (right). Ground truth (on CBCT image) and predicted classes are shown on each image, where B and L indicate buccal and lingual relationships, respectively

Discussion

Accurate segmentation of the inferior alveolar nerve is an essential step in oral and maxillofacial surgery, such as implant placement in the mandible, extraction of an impacted M3, and orthognathic surgery [1,2,3, 44]. The incidence of inferior alveolar nerve injuries in M3 extraction surgery increases if the MC and the root of the M3 are closely located [4]. In addition, the positional relationship of the MC with the M3 in the buccolingual aspect is a risk factor as important as proximity and is identified routinely in the clinical field [14, 45]. Therefore, diagnosis based on preoperative imaging is essential for accurate prediction of and sufficient preparation for this risk. CBCT images, which overcome the limitations of distortion and superimposition of structures in 2D panoramic images [5, 9], have been widely used in dental clinics [10]. However, accurate identification of the path of the MC and its 3D relationship with the M3 on cross-sections of the CBCT image is labor- and time-consuming due to the low contrast and high noise [46].

The application of deep learning for segmentation or classification of the M3 or the MC has been performed in 2D panoramic radiographs, 3D CT, and CBCT images [27, 29, 31, 47,48,49,50]. Another study has segmented the M3 and the MC in CBCT images and classified them into three types according to proximity and contact between the two structures [34]. However, no previous studies have been performed on the automatic classification of the buccal-lingual positional relationship between the M3 and the MC using deep learning. In this study, a distance-aware CNN was proposed to segment the M3 and the MC and to classify the positional relationship between them in CBCT images, automatically. In the segmentation stage, pre-trained Dense121 U-Net was used for multiclass segmentation of the M3 and the MC. In the classification stage, a 3D classification network simultaneously learned anatomical and geometrical information from the CBCT image and the distance map generated from the segmentation prediction of the M3 and the MC.

The M3 and MC segmentation by Dense121 U-Net showed the highest performance among all segmentation networks, indicating that the network had balanced precision and recall. The results of the Dense121 U-Net showed a significant reduction in the number of discontinuous areas where the cortical bone of the MC was not clearly visible and was compressed or obstructed by the root of the M3 in CBCT images. Those of other networks showed a larger number of areas of discontinuity for the MC. The plots representing the segmentation performance of DSC and Hausdorff distances for the M3 and the MC by Dense121 U-Net demonstrated this quantitatively with higher values and smaller fluctuations compared to those of the other networks. These results showed that the Dense121 U-Net is the most effective network for segmenting small target objects with consistent accuracy. Therefore, 3D segmentation by Dense121 U-Net demonstrated the best performance with improved structural continuity and boundary details for segmentation of the MC and the M3.

The 3D distance-aware network was designed to automatically classify the buccal-lingual relationship between the M3 and the MC, and it incorporated the SDM as an input to learn additional 3D geometrical information between them. Using both the 3D CBCT image and the SDM as multichannel inputs of the network, the network achieved better classification performance compared to other input combinations. The performance of the classification network was highest, reaching 0.90 or more when using an SDM generated from the segmentation outcomes of Attention U-Net or Dense121 U-Net with higher segmentation performance. These results demonstrated that the classification network using both the original CBCT image and the SDM as multichannel inputs learned the 3D positional information more effectively to analyze the relationship between the M3 and the MC.

We utilized Grad-CAM visualization to create visual interpretations for the decision-making process in the buccolingual classification of the M3 and the MC by the classification network. The heat maps generated by the networks using inputs of the CBCT image and SDM focused specifically on the root of the M3 and the periphery of the MC, while those using other inputs emphasized regions that were not relevant to classification. As a result, the SDM provided geometrical guidance that provided precise information on the position and distance relationships between the M3 and the MC during classification. In addition to the information on the anatomical regions for the M3 and the MC, the geometrical information for the position and shape of adjacent anatomical structures was important for better classification performance.

The purpose of including a control group in a study is to provide a baseline for comparison. In studies like ours, the control group can be used to compare the results of the group which uses the developed network in classification for the buccal-lingual relationship in order to see if there is a statistically significant difference. As there was no control group in this study, it was difficult to say whether the group using the developed network demonstrated better performance in the classification than the control group not using the network. However, deep learning models in medical and dental image interpretation can be advantageous in the following ways [24, 30, 51, 52]. Deep learning models can interpret medical images much faster than human interpreters, which can lead to a faster diagnosis for patients [30]. Deep learning models can automate the process of medical image interpretation, which can free up human interpreters to focus on other tasks [30, 51]. Deep learning models can help to improve the efficiency of medical image interpretation by reducing the time and labor required [24, 52].

Our study did have some limitations. First, due to the insufficient availability of the CBCT images, only the relative buccal or lingual relationship between the M3 and the MC was considered as a criterion for classification, although the MC could appear in a greater variety of positions and orientations with respect to the M3, such as through, beneath, or between the roots. Future studies should include more diverse criteria for detailed classification. Second, our study had a potential limitation of generalization ability due to the use of CBCT images from a single organization. Overfitting of training of a deep learning model, which results in the model learning statistical regularity specific to the training dataset, could negatively impact the ability of the model to generalize to a new dataset [53]. Therefore, in future studies, it is important to improve and evaluate the performance of the network on CBCT images of a variety of relationships of the M3 and the MC and those obtained from different devices.

Conclusions

In this study, the proposed distance-aware network demonstrated high accuracy in automatic classification of the 3D positional relationship between the M3 and the MC by learning anatomical and geometrical information from CBCT images. The distance map generated using the segmentation mask as one of the multichannel inputs in the classification network significantly contributed to increasing the accuracy in the classification of the 3D buccal-lingual relationship between the M3 and the MC. The network could contribute to the automatic, accurate, and efficient classification of the 3D positional relationship between the M3 and the MC for preoperative planning of M3 extraction surgery to avoid surgical complications. This research could be a cornerstone for the automatic and accurate classification of more diverse relationships between the M3 and the MC.

Availability of data and materials

The datasets generated and/or analyzed during the current study are not publicly available due to the restriction by the IRB of Seoul National University Dental Hospital to protect patient privacy but are available from the corresponding author on reasonable request. Please contact the corresponding author for any commercial implementation of our research.

References

  1. Sayed N, Bakathir A, Pasha M, Al-Sudairy S. Complications of third molar extraction: a retrospective study from a tertiary healthcare centre in Oman. Sultan Qaboos Univ Med J. 2019;19(3):e230.

    PubMed  PubMed Central  Google Scholar 

  2. Leung Y, Cheung L. Risk factors of neurosensory deficits in lower third molar surgery: a literature review of prospective studies. Int J Oral Maxillofac Surg. 2011;40(1):1–10.

    PubMed  Google Scholar 

  3. Bui CH, Seldin EB, Dodson TB. Types, frequencies, and risk factors for complications after third molar extraction. J Oral Maxillofac Surg. 2003;61(12):1379–89.

    PubMed  Google Scholar 

  4. Sarikov R, Juodzbalys G. Inferior alveolar nerve injury after mandibular third molar extraction: a literature review. J Oral Maxillofac Res. 2014;5(4):e1.

    PubMed  PubMed Central  Google Scholar 

  5. Hasegawa T, Ri S, Shigeta T, Akashi M, Imai Y, Kakei Y, et al. Risk factors associated with inferior alveolar nerve injury after extraction of the mandibular third molar—a comparative study of preoperative images by panoramic radiography and computed tomography. Int J Oral Maxillofac Surg. 2013;42(7):843–51.

    PubMed  Google Scholar 

  6. Jerjes W, Upile T, Shah P, Nhembe F, Gudka D, Kafas P, et al. Risk factors associated with injury to the inferior alveolar and lingual nerves following third molar surgery—revisited. Oral Surg Oral Med Oral Pathol Oral Radiol Endod. 2010;109(3):335–45.

    PubMed  Google Scholar 

  7. Maglione M, Costantinides F, Bazzocchi G. Classification of impacted mandibular third molars on cone-beam CT images. J Clin Exp Dent. 2015;7(2):e224.

    PubMed  PubMed Central  Google Scholar 

  8. Blaeser BF, August MA, Donoff RB, Kaban LB, Dodson TB. Panoramic radiographic risk factors for inferior alveolar nerve injury after third molar extraction. J Oral Maxillofac Surg. 2003;61(4):417–21.

    PubMed  Google Scholar 

  9. Maegawa H, Sano K, Kitagawa Y, Ogasawara T, Miyauchi K, Sekine J, et al. Preoperative assessment of the relationship between the mandibular third molar and the mandibular canal by axial computed tomography with coronal and sagittal reconstruction. Oral Surg Oral Med Oral Pathol Oral Radiol Endod. 2003;96(5):639–46.

    PubMed  Google Scholar 

  10. Matzen L, Wenzel A. Efficacy of CBCT for assessment of impacted mandibular third molars: a review–based on a hierarchical model of evidence. Dentomaxillofacial Radiol. 2015;44(1):20140189.

    Google Scholar 

  11. Ueda M, Nakamori K, Shiratori K, Igarashi T, Sasaki T, Anbo N, et al. Clinical significance of computed tomographic assessment and anatomic features of the inferior alveolar canal as risk factors for injury of the inferior alveolar nerve at third molar surgery. J Oral Maxillofac Surg. 2012;70(3):514–20.

    PubMed  Google Scholar 

  12. Ghaeminia H, Meijer G, Soehardi A, Borstlap W, Mulder J, Bergé S. Position of the impacted third molar in relation to the mandibular canal Diagnostic accuracy of cone beam computed tomography compared with panoramic radiography. Int J Oral Maxillofac Surg. 2009;38(9):964–71.

    PubMed  Google Scholar 

  13. Wang W-Q, Chen MY, Huang H-L, Fuh L-J, Tsai M-T, Hsu J-T. New quantitative classification of the anatomical relationship between impacted third molars and the inferior alveolar nerve. BMC Med Imaging. 2015;15(1):1–6.

    Google Scholar 

  14. Wang D, Lin T, Wang Y, Sun C, Yang L, Jiang H, et al. Radiographic features of anatomic relationship between impacted third molar and inferior alveolar canal on coronal CBCT images: risk factors for nerve injury after tooth extraction. Arch Med Sci. 2018;14(3):532.

    PubMed  Google Scholar 

  15. Gu L, Zhu C, Chen K, Liu X, Tang Z. Anatomic study of the position of the mandibular canal and corresponding mandibular third molar on cone-beam computed tomography images. Surg Radiol Anat. 2018;40(6):609–14.

    PubMed  Google Scholar 

  16. Gan Y, Xia Z, Xiong J, Li G, Zhao Q. Tooth and alveolar bone segmentation from dental computed tomography images. IEEE J Biomed Health Inform. 2017;22(1):196–204.

    PubMed  Google Scholar 

  17. Hosntalab M, Aghaeizadeh Zoroofi R, Abbaspour Tehrani-Fard A, Shirani G. Segmentation of teeth in CT volumetric dataset by panoramic projection and variational level set. Int J Comput Assist Radiol Surg. 2008;3(3):257–65.

    Google Scholar 

  18. Gao H, Chae O. Individual tooth segmentation from CT images using level set method with shape and intensity prior. Pattern Recogn. 2010;43(7):2406–17.

    Google Scholar 

  19. Ji DX, Ong SH, Foong KWC. A level-set based approach for anterior teeth segmentation in cone beam computed tomography images. Comput Biol Med. 2014;50:116–28.

    PubMed  Google Scholar 

  20. Moris B, Claesen L, Sun Y, Politis C. Automated tracking of the mandibular canal in cbct images using matching and multiple hypotheses methods. 2012 Fourth International Conference on Communications and Electronics (ICCE). Hue: IEEE; 2012.

  21. Barone S, Paoli A, Razionale AV. CT segmentation of dental shapes by anatomy-driven reformation imaging and B-spline modelling. Int J Numeri Method Biomed Eng. 2016;32(6):e02747.

    Google Scholar 

  22. Abdolali F, Zoroofi RA, Abdolali M, Yokota F, Otake Y, Sato Y. Automatic segmentation of mandibular canal in cone beam CT images using conditional statistical shape model and fast marching. Int J Comput Assist Radiol Surg. 2017;12(4):581–93.

    PubMed  Google Scholar 

  23. Kainmueller D, Lamecker H, Seim H, Zinser M, Zachow S. Automatic extraction of mandibular nerve and bone from cone-beam CT data. International Conference on Medical Image Computing and Computer-Assisted Intervention–MICCAI 2009: 12th International Conference, London, UK, September 20-24, 2009, Proceedings, Part II 12. Springer Berlin Heidelberg; 2009.

  24. Hwang J-J, Jung Y-H, Cho B-H, Heo M-S. An overview of deep learning in the field of dentistry. Imaging Sci Dent. 2019;49(1):1–7.

    PubMed  PubMed Central  Google Scholar 

  25. Jeoun B-S, Yang S, Lee S-J, Kim T-I, Kim J-M, Kim J-E, et al. Canal-Net for automatic and robust 3D segmentation of mandibular canals in CBCT images using a continuity-aware contextual network. Sci Rep. 2022;12(1):13460.

    PubMed  PubMed Central  Google Scholar 

  26. Kim S-H, Kim J, Yang S, Oh S-H, Lee S-P, Yang HJ, et al. Automatic and quantitative measurement of alveolar bone level in OCT images using deep learning. Biomed Opt Express. 2022;13(10):5468–82.

    PubMed  PubMed Central  Google Scholar 

  27. Kwak GH, Kwak E-J, Song JM, Park HR, Jung Y-H, Cho B-H, et al. Automatic mandibular canal detection using a deep convolutional neural network. Sci Rep. 2020;10(1):1–8.

    Google Scholar 

  28. Cui Z, Li C, Wang W. ToothNet: automatic tooth instance segmentation and identification from cone beam CT images. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; 2019.

  29. Jaskari J, Sahlsten J, Järnstedt J, Mehtonen H, Karhu K, Sundqvist O, et al. Deep learning method for mandibular canal segmentation in dental cone beam computed tomography volumes. Sci Rep. 2020;10(1):1–8.

    Google Scholar 

  30. Litjens G, Kooi T, Bejnordi BE, Setio AAA, Ciompi F, Ghafoorian M, et al. A survey on deep learning in medical image analysis. Med Image Anal. 2017;42:60–88.

    PubMed  Google Scholar 

  31. Vinayahalingam S, Xi T, Bergé S, Maal T, de Jong G. Automated detection of third molars and mandibular nerve by deep learning. Sci Rep. 2019;9(1):1–7.

    Google Scholar 

  32. Yong T-H, Yang S, Lee S-J, Park C, Kim J-E, Huh K-H, et al. QCBCT-NET for direct measurement of bone mineral density from quantitative cone-beam CT: A human skull phantom study. Sci Rep. 2021;11(1):1–13.

    Google Scholar 

  33. Kang S-R, Shin W, Yang S, Kim J-E, Huh K-H, Lee S-S, et al. Structure-preserving quality improvement of cone beam CT images using contrastive learning. Comput Biol Med. 2023;158:106803.

    PubMed  Google Scholar 

  34. Liu M-Q, Xu Z-N, Mao W-Y, Li Y, Zhang X-H, Bai H-L, et al. Deep learning-based evaluation of the relationship between mandibular third molar and mandibular canal on CBCT. Clin Oral Invest. 2022;26(1):981–91.

    Google Scholar 

  35. Fedorov A, Beichel R, Kalpathy-Cramer J, Finet J, Fillion-Robin J-C, Pujol S, et al. 3D Slicer as an image computing platform for the Quantitative Imaging Network. Magn Reson Imaging. 2012;30(9):1323–41.

    PubMed  PubMed Central  Google Scholar 

  36. Huang G, Liu Z, Van Der Maaten L, Weinberger KQ. Densely connected convolutional networks. Proceedings of the IEEE conference on computer vision and pattern recognition; 2017.

  37. Badrinarayanan V, Kendall A, Cipolla R. SegNet: a deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans Pattern Anal Mach Intell. 2017;39(12):2481–95.

    PubMed  Google Scholar 

  38. Ronneberger O, Fischer P, Brox T. U-net: convolutional networks for biomedical image segmentation. International Conference on Medical image computing and computer-assisted intervention–MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18. Springer International Publishing; 2015.

  39. Oktay O, Schlemper J, Folgoc LL, Lee M, Heinrich M, Misawa K, et al. Attention u-net: Learning where to look for the pancreas. arXiv preprint arXiv:180403999. 2018.

  40. Ye Q-Z. The signed Euclidean distance transform and its applications. 9th International conference on pattern recognition. Rome: IEEE Computer Society; 1988.

  41. Malladi R, Sethian JA, Vemuri BC. Shape modeling with front propagation: A level set approach. IEEE Trans Pattern Anal Mach Intell. 1995;17(2):158–75.

    Google Scholar 

  42. Chan T, Zhu W. Level set based shape prior segmentation. 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. San Diego: IEEE; 2005.

  43. Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, Batra D. Grad-cam: visual explanations from deep networks via gradient-based localization. Proceedings of the IEEE international conference on computer vision; 2017.

  44. Juodzbalys G, Wang HL, Sabalys G, Sidlauskas A, Galindo-Moreno P. Inferior alveolar nerve injury associated with implant surgery. Clin Oral Implant Res. 2013;24(2):183–90.

    Google Scholar 

  45. Tachinami H, Tomihara K, Fujiwara K, Nakamori K, Noguchi M. Combined preoperative measurement of three inferior alveolar canal factors using computed tomography predicts the risk of inferior alveolar nerve injury during lower third molar extraction. Int J Oral Maxillofac Surg. 2017;46(11):1479–83.

    PubMed  Google Scholar 

  46. Pauwels R, Jacobs R, Singer SR, Mupparapu M. CBCT-based bone quality assessment: are Hounsfield units applicable? Dentomaxillofac Radiol. 2015;44(1):20140238.

    PubMed  Google Scholar 

  47. Dhar MK, Yu Z. Automatic tracing of mandibular canal pathways using deep learning. arXiv preprint arXiv:211115111. 2021.

  48. Lahoud P, Diels S, Niclaes L, Van Aelst S, Willems H, Van Gerven A, et al. Development and validation of a novel artificial intelligence driven tool for accurate mandibular canal segmentation on CBCT. J Dent. 2022;116:103891.

    PubMed  Google Scholar 

  49. Lim H-K, Jung S-K, Kim S-H, Cho Y, Song I-S. Deep semi-supervised learning for automatic segmentation of inferior alveolar nerve using a convolutional neural network. BMC Oral Health. 2021;21(1):1–9.

    Google Scholar 

  50. Ariji Y, Mori M, Fukuda M, Katsumata A, Ariji E. Automatic visualization of the mandibular canal in relation to an impacted mandibular third molar on panoramic radiographs using deep learning segmentation and transfer learning techniques. Oral Surg Oral Med Oral Pathol Oral Radiol. 2022;134:749–57.

    PubMed  Google Scholar 

  51. Varoquaux G, Cheplygina V. Machine learning for medical imaging: methodological failures and recommendations for the future. NPJ Digit Med. 2022;5(1):48.

    PubMed  PubMed Central  Google Scholar 

  52. Heo M-S, Kim J-E, Hwang J-J, Han S-S, Kim J-S, Yi W-J, et al. Artificial intelligence in oral and maxillofacial radiology: what is currently possible? Dentomaxillofac Radiol. 2021;50(3):20200375.

    PubMed  Google Scholar 

  53. Kwon O, Yong T-H, Kang S-R, Kim J-E, Huh K-H, Heo M-S, et al. Automatic diagnosis for cysts and tumors of both jaws on panoramic radiographs using a deep convolution neural network. Dentomaxillofac Radiol. 2020;49(8):20200185.

    PubMed  PubMed Central  Google Scholar 

Download references

Acknowledgements

This work was supported by a National Research Foundation of Korea (NRF) grant funded by the Korean government (MSIT) (No. 2023R1A2C200532611). This work also was supported by a Korea Medical Device Development Fund grant funded by the Korean government (Ministry of Science and ICT; Ministry of Trade, Industry, and Energy; Ministry of Health & Welfare; Ministry of Food and Drug Safety) (Project Number: 1711194231, KMDF_PR_20200901_0011, 1711174552, KMDF_PR_20200901_0147).

Funding

This work was supported by a National Research Foundation of Korea (NRF) grant funded by the Korean government (MSIT) (No. 2023R1A2C200532611). This work also was supported by a Korea Medical Device Development Fund grant funded by the Korean government (Ministry of Science and ICT; Ministry of Trade, Industry, and Energy; Ministry of Health & Welfare; Ministry of Food and Drug Safety) (Project Number: 1711194231, KMDF_PR_20200901_0011, 1711174552, KMDF_PR_20200901_0147).

Author information

Authors and Affiliations

Authors

Contributions

So-Young Chun, Yun-Hui Kang, and Su Yang contributed to study conception and design; data acquisition, analysis, and interpretation; and drafted the manuscript. Se-Ryong Kang, Jun-Min Kim, Jo-Eun Kim, Kyung-Hoe Huh, Sam-Sun Lee, Min-Suk Heo, and Sang-Jeong Lee contributed to study conception and design, data interpretation, and drafted the manuscript. Won-Jin Yi contributed to study conception and design; data acquisition, analysis, and interpretation; and drafted and critically revised the manuscript. All authors provided final approval and agreed to be accountable for all aspects of the work.

Corresponding author

Correspondence to Won-Jin Yi.

Ethics declarations

Ethics approval and consent to participate

This study was performed with approval from the Institutional Review Board (IRB) of Seoul National University Dental Hospital (ERI18001). The ethics committee approved the waiver for informed consent because this was a retrospective study. The study was performed in accordance with the Declaration of Helsinki.

Consent for publication

Not applicable.

Competing interests

The authors declare no competing interests.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Chun, SY., Kang, YH., Yang, S. et al. Automatic classification of 3D positional relationship between mandibular third molar and inferior alveolar canal using a distance-aware network. BMC Oral Health 23, 794 (2023). https://doi.org/10.1186/s12903-023-03496-9

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12903-023-03496-9

Keywords