Skip to main content

Evaluation of transfer ensemble learning-based convolutional neural network models for the identification of chronic gingivitis from oral photographs

Abstract

Background

To evaluate the performances of several advanced deep convolutional neural network models (AlexNet, VGG, GoogLeNet, ResNet) based on ensemble learning for recognizing chronic gingivitis from screening oral images.

Methods

A total of 683 intraoral clinical images acquired from 134 volunteers were used to construct the database and evaluate the models. Four deep ConvNet models were developed using ensemble learning and outperformed a single model. The performances of the different models were evaluated by comparing the accuracy and sensitivity for recognizing the existence of gingivitis from intraoral images.

Results

The ResNet model achieved an area under the curve (AUC) value of 97%, while the AUC values for the GoogLeNet, AlexNet, and VGG models were 94%, 92%, and 89%, respectively. Although the ResNet and GoogLeNet models performed best in classifying gingivitis from images, the sensitivity outcomes were not significantly different among the ResNet, GoogLeNet, and Alexnet models (p>0.05). However, the sensitivity of the VGGNet model differed significantly from those of the other models (pā€‰<ā€‰0.001).

Conclusion

The ResNet and GoogLeNet models show promise for identifying chronic gingivitis from images. These models can help doctors diagnose periodontal diseases efficiently or based on self-examination of the oral cavity by patients.

Peer Review reports

Background

Gingivitis is a chronic disease primarily caused by bacterial infection. Gingivitis affects public health worldwide, and its main clinical symptoms are bleeding, redness, and bad breath. Without prompt treatment, the continuous progression of gingivitis can lead to the resorption of the alveolar bone and loss of the periodontal ligament. Meanwhile, the latest epidemiological survey on oral health reported that 88% of adults suffer from gingivitis each year [1]. As the first stage of periodontal disease, gingivitis is closely related to other serious diseases such as cardiovascular disease, leukemia, and tumors [2, 3]. Protecting gum health is a key factor in preventing periodontal disease [4]. Gingivitis is clinically diagnosed through a series of conventional oral examinations, which include measurements of the degree of redness and swelling and bleeding on probing [5]. Although chairside oral examination is the most reliable method for gingivitis detection, other methods should be considered to improve the efficiency of diagnosis and reduce the clinical burden on physicians.

In recent years, deep learning algorithms (e.g., deep convolutional neural networks [DCNNs]) have shown high efficiency and accuracy for classifying and analyzing medical characteristics or features [6, 7]. These algorithms can enable the automatic screening of several diseases using imagery captured with cameras or smartphones. For example, Li et al. classified tooth types in dental images using contrast-limited adaptive histogram equalization, gray-level co-occurrence matrix, and extreme learning machine approaches [8]. Krois et al. applied CNNs to detect periodontal bone loss in panoramic dental radiographs [9]. In a previous study, we screened for gingivitis and its irritants (dental calculus and soft deposits) in oral photos using a novel multi-task learning CNN model and obtained good accuracy and sensitivity for both classification and localization [10]. The above examples demonstrate that image analysis can play an important role in monitoring the gingival condition of patients, and that DCNNs show promise for supplementing clinical visits to detect health issues. However, using deep learning models to screen for gum conditions remains under-explored.

In this study, we selected several advanced deep ConvNet models (Alexnet [11], VGG [12], GoogLeNet [13], and ResNet [14]) for the classification of anatomical gingival soft tissue structures from oral screening images and compared their performance for recognizing chronic gingivitis in terms of both accuracy and efficiency. AlexNet is renowned for its breakthrough performance in the ImageNet challenge, which reignited the interest in neural networks for computer vision tasks. GoogLeNet, with its inception modules, demonstrates the power of network-in-network architectures to increase depth and width without a significant increase in computational cost. ResNet introduced residual learning, enabling the training of much deeper networks by addressing the vanishing gradient problem, leading to remarkable improvements in accuracy. Lastly, VGGNet is celebrated for its simplicity and depth, utilizing very small convolutional filters to build deeper networks, which has shown to be effective in capturing fine details in images. Together, these models encompass a range of approaches that have significantly advanced the field of computer vision. By comparing the accuracy of the four advanced models in identifying gingivitis, we aim to initiate a discourse on the integration of deep learning into oral self-examination tools for enhancing public dental health.

Materials and methods

In this section, Firstly, we present the data collection protocol employed in this study along with the generated dataset, followed by the annotation of the collected data. Next, we demonstrate the limitations associated with gingivitis diagnosis and propose four deep learning model architectures. The implementation and training of the model are subsequently elaborated upon. Finally, we elucidate the metrics and statistical analysis methodologies employed to validate multiple models for discerning gingivitis.

Ethical approval

The project was approved by the Ethical Review Board at local University (approval number NJSH-2022NL-069). Any reports related to the results of this study will be subject to confidentiality and compliance with data protection regulations

Subjects and dataset

We built an in-house dataset of oral photos collected in the Department of Periodontics, Orthodontics and Endodontics, Stomatological Hospital, from January 2020 to December 2022. The dataset contained 683 images captured by postgraduate dentists from 134 gingivitis patients and the healthy population. The images cover a wide age range from 14 to 60 years old. Images of teeth with severe cervical caries and periodontitis with severe gingival recession were excluded. To approximate the image quality in practical scenarios, a diverse range of equipment was utilized for photo collection, including iPhone, Samsung Galaxy, Canon 6D and so on. The methods were conducted in accordance with the relevant guidelines and regulations, written informed consent was obtained from each participant. The diagnosis of chronic gingivitis requires that two criteria be met: (1) clinical symptoms including bleeding with tooth brushing, blood in the saliva, and gingival swelling and redness (Fig.Ā 1); and (2) no attachment loss in periodontal probe examination and no loss of supporting structures in radiographic analysis.

Fig. 1
figure 1

Clinical symptoms of gingivitis

In our preprocessing step, images of different sizes are first scaled, and all images are uniformly scaled to a size of 224ā€‰Ć—ā€‰224 pixels. This process uses bilinear interpolation to maintain the aspect ratio and detail information of the image. Subsequently, we normalized the scaled image to normalize the pixel values to between 0 and 1.

Given the constrained size of our dataset, it was imperative to explore strategies that could artificially expand our dataā€™s diversity without compromising its integrity. To this end, we integrated a comprehensive suite of data augmentation techniques directly into our training process. This dynamic augmentation occurs in real-time during the training phase, presenting the model with an enriched dataset. Our augmentation strategy includes a range of transformations, such as random rotations zooming, horizontal flipping, and shear transformations.

Implicit sorting relationships may exist in the initially gathered data, which may negatively affect the DCNN modelā€™s accuracy. Therefore, during training, we initially use a shuffling procedure to disrupt the order of the data.We divided the dataset into training, validation, and testing subsets by randomly splitting the photos into three groups. We used the training set to update the model. We changed the hyperparameters in the validation set, and the test set was used to evaluate the modelā€™s performance. To increase the use of data and train the model with a large number of parameters, we improved the efficiency of data utilization through cross validation. FigureĀ 2 provides details about the distribution of the dataset.

Fig. 2
figure 2

Numbers of images with positive and negative findings assigned to the training, validation, and testing subsets

1) Training dataset

The identification of gingivitis is a complex process that involves considering the color of the gums, the level of swelling, and the bleeding condition. Accordingly, we selected models with large numbers of structural layers to extract the complex features of gingivitis. To update the modelā€™s parameters, the model was trained many times with the training data.

2) Validation dataset

The modelā€™s variables, including the number of layers and neurons, can affect the ultimate recognition accuracy. In this study, the model that performed best in the validation set was chosen as the final version after we tested numerous iterations of each model. As the number of training sessions increases, the model accumulates considerable useless knowledge, including the brightness of the image, the placement of the teeth, the size of the teeth, and other features of oral diseases such as black stains and dental calculus. Additionally, we used the validation set to halt model updates early.

3) Test dataset

We used the test set to test the final performance of the model based on accuracy and analysis of the receiver operating characteristic (ROC) curves.

4) Cross-validation method

Due to the complexity of the task and the small amount of data, we needed to increase the use of the collected data. We used a cross-validation method to train/test multiple groups of models with different training/test sets, as shown in Fig.Ā 3.

Fig. 3
figure 3

Cross validation was used to average the data and take several samples to observe the results of a variety of collected data

This rigorous validation framework was designed to enhance the reliability of our modelā€™s performance metrics by carefully segregating the dataset into non-overlapping training, validation, and testing segments. The nested cross-validation process involved an outer 5-fold cross-validation for delineating training and testing data, complemented by an inner 5-fold cross-validation within the training dataset for hyperparameter tuning.

ConvNet models

We selected several models and compared their performances by training to identify the model with the best performance. Finally, we outperformed a single model by using ensemble learning. FigureĀ 4 shows the overall framework of the model.

Fig. 4
figure 4

Architecture of AlexNet, Note ā€œConvā€ is short for convolutional, and ā€œFCā€ stands for the fully connected layer

1) AlexNet

AlexNet uses ReLU as the activation function. Only the most obvious elements of a region are kept. As an illustration, consider the color of the gingiva. Only the components whose color depth exceed the threshold are kept, and the rest of the features are eliminated (Fig.Ā 5).

Fig. 5
figure 5

Architecture of VGG. Note ā€œConvā€ is short for convolutional, and ā€œFCā€ stands for the fully connected layer

To increase the diversity of the data, we randomly eliminated some of the intermediate results using the dropout function.

2) VGG

For the model to quickly collect the general information of the image, a lot of distant information needs to be swiftly aggregated. We used the 3ā€‰Ć—ā€‰3 convolution kernel to quickly gather tooth and gingival information over large distances. Therefore, we obtained a general assessment without a precise division of tooth position (Fig.Ā 6).

Fig. 6
figure 6

Architecture of GoogleNet. Note ā€œConvā€ is short for convolutional. Inception is a structure composed of 5 convolutional networks, where the dimensions of each network are as shown in the figure

3) GoogLeNet

Each model layerā€™s dimension may correlate to the toothā€™s higher-dimensional attributes. To minimize the featureā€™s dimension, we used a 1ā€‰Ć—ā€‰1 convolution kernel (Fig.Ā 7).

Fig. 7
figure 7

Architecture of ResNet. Note: ā€œConvā€ is short for convolutional, ā€œResā€ is short for Residual Network, and ā€œFCā€ stands for the fully connected layer

4) ResNet

We used a 34-layer version of the ResNet model (Fig.Ā 8). We discovered that the starting information and position information in the image were helpful in the recognition process. Basic knowledge of the structure and color can be learned in the first few layers, and this knowledge remains helpful in the subsequent layers. ResNet adds the results of previous layers to subsequent layers by using residual structures and skipping intermediate steps. In this way, the knowledge learned in the previous layer can be directly transferred to the next layer.

Fig. 8
figure 8

We used ensemble learning to identify images that are difficult for a single model to identify

5) Ensemble learning

The AlexNet, GoogLeNet, and ResNet models had good performance based on their accuracy and area under the ROC curve (AUC). In contrast, the VGGNet model performed poorly. We know that for locating particular data, some models perform poorly, while others show good performance. Consequently, we combined the findings from the three models with good performance using ensemble learning. Three modelsā€”Alexnet, Googlenet, and Resnetā€”were used to make predictions for one image. The output of ensemble learning was then determined by the outcomes of the three models with the highest number. The ensemble learning process is shown in Fig.Ā 4.

Training strategies

To train the model quickly, we employed transfer learning hyperparameters. The optimization method included a few hyperparameters. Transfer learning [15] was first utilized for initialization. The model parameters developed using the open dataset were utilized by all four models. Compared with the loss value of the model, which will decrease later, the loss value of the model in the early stage of training is larger. The early loss value of training was greater because the hyperparameters used in transfer learning were different from those employed in our jobs. The GoogLeNet and ResNet models both showed substantial loss values in the first few epochs, whereas the VGG and AlexNet models displayed smaller loss values. The loss values of the four prediction algorithms converged to lower values as the number of epochs increased [16].

In our endeavor to develop a model capable of generalizing well beyond the confines of our training data, we implemented a comprehensive strategy to prevent overfitting. Apart from leveraging early stopping as a safeguard against overtraining, we incorporated dropout with a rate of 0.5 within our modelā€™s architecture, which served to randomly deactivate a portion of the neurons during each training pass. This randomness introduced by dropout helps in reducing the modelā€™s sensitivity to specific features of the training data. In parallel, L2 regularization with a value of 1ā€‰Ć—ā€‰10āˆ’ā€‰4 was applied across the networkā€™s parameters, imposing a constraint that penalizes the magnitude of the weights, thereby discouraging complex models that could overfit the training data.

To evaluate the effectiveness of the proposed strategy, we ran numerous experiments. Inference accuracy and training loss were measured. Finally, we examined the performances the AlexNet, VGG, GoogLeNet, and ResNet deep learning models.

Table 1 The precision, recall, F1 score, cross entropy loss and accuracy of models

In our study, we evaluated several deep learning modelsā€™ performance metrics, including precision, recall, F1 score, cross-entropy loss, and accuracy, as summarized in TableĀ 1. AlexNet and GoogLeNet demonstrated superior performance, particularly in precision and F1 score, indicating their effectiveness in accurately identifying positive cases.

The convergence performance of a model is reflected by the loss value. In most cases, the loss value will stabilize after a certain time and eventually decrease to some amount. The training is then considered to be finished. The modelā€™s training outcome can be measured by the accuracy. The initial prediction accuracy of the AlexNet, VGG, and GoogleNet models was roughly 85%. The VGG model began to exhibit accuracy fluctuations at approximately 20 epochs, after which the accuracy increased steadily before flattening out. The same phenomenon was observed for the AlexNet model at approximately 23 epochs. At approximately 29 epochs, the accuracy of the AlexNet model peaked before starting to decrease and returning to the initial accuracy.

A recognition model is prone overfitting if there are too many training iterations, which increase the accuracy on the training dataset. However, this will result in substandard generalization performance, leading to poor performance on unseen data. Among the four models, the ResNet and GoogLeNet models performed well with high recognition accuracy. We also constructed the ROC curves of the four models and calculated the corresponding AUC values. Among the models, VGGNet had the smallest AUC (89%), while the AUCs of GoogLeNet and ResNet were larger (94% and 97%, respectively). It is worth noting that GoogLeNet showed better generalization performance than ResNet.

It is important to note that after sufficient training, several models experienced overfitting. As a result, we kept the results once the training reached the highest level of accuracy.

We examined the p values between different models and correct result after extracting the data at random. ResNet and GoogLeNet and label(The label is the real classification of the data. During the process of data collection, the researchers who collect the data will mark the data classification as the label.)had the highest p values out of all the tested models, and no substantial differences were observed in the projected outcomes of these models. The p values between the VGGNet model and the other models were all less than 0.001, indicating significant differences.

Table 2 Calculated p values between different models and labels after extracting the data at random

Discussion

Chronic gingivitis is a common oral disease that influences human health and quality of life [17]. In this study, we evaluated four precise ConvNet models for the recognition of chronic gingivitis from screening oral photographs taken by either mobile phones or digital cameras. This approach differs from the commonly used deep learning systems based on clinical imaging for the computer-assisted diagnosis of gingivitis. Among the four tested models, the ResNet and GoogLeNet models presented the best performance. These models can be applied to help people maintain oral hygiene through self-screening, thereby reducing the treatment burden on physicians and the financial burden on patients.

Previous studies have verified the advantages of using photographs for the detection of chronic gingivitis by dentists. Eke et al. reported self-screening measures for periodontal diseases [18]. Alalharith et al. evaluated state-of-the-art object detection and recognition techniques based on deep learning for the automatic detection of gingivitis in orthodontic patients based on intraoral images [19]. Lang et al. further assessed the ability to diagnose plaque-induced gingivitis from intraoral photographs based on the symptoms of gingivitis [20]. These studies all suggest that clinical photographs are sufficient for dentists to draw conclusions that are statistically similar to those based on visual inspection.

Unlike the above-described studies, we compared several advanced transfer learning-based ConvNet models (Alexnet, VGG, GoogLeNet, and ResNet) for the automated detection of gingivitis from oral photographs. The most closely related study in the literature is Xianwei et al. (2019), who assessed a deep learning method (a mask DCNN model with a multichannel gray-level co-occurrence matrix associated with a particle swarm optimization neural network) for the detection of gingivitis based on dental images. However, compared to their work, our study provides four innovations: (1) we identified gingivitis using oral images from both consumer mobile phones and digital cameras; (2) we employed transfer learning to effectively train the models and compensate for the low data volume; (3) we used a cross-validation method to process the training, validation, and testing datasets, and each group was trained once to make efficient use of the datasets; and (4) we use ensemble learning, a supervised learning algorithms, to select three models that performed better than VGGNet.

At present, transfer learning has three main aspects [21]: (1) the new data are used for training after loading all the model parameters; (2) only the last few layers of parameters are trained and loaded after the weights are determined; and (3) after the weight is loaded, a complete connection adds one layer to the original network. Because our task is not the same as the publicly available dataset, we adopted transfer learning and trained the model faster than random initialization.

Other studies have used transfer learning methods to diagnose oral diseases. Jin et al [22]. proposed a deep transfer learning method for the diagnosis of various diseases based on computer-aided facial recognition. Rahman et al. [23]. proposed a transfer learning model based on AlexNet to extract rank features from oral squamous cell carcinoma biopsy images and achieved high classification accuracies. Chang et al. [24]. presented a transfer learning-based method for the automatic diagnosis of parotid gland tumors from multimodal magnetic resonance images. However, none of these studies used ensemble learning to train for low-quality spatial annotations of disease.

In this work, we developed four trained CNN models to automatically identify and localize gingivitis from photographs. We created a high-specificity point (blue) and a high-sensitivity point (orange) for the model to show the accuracy; as a result, the prediction results at various operating points could be represented visually. To reduce false positives, we established a high-specificity point (blue) with a high discrimination threshold. To maintain a low missing rate, we selected a high-sensitivity point (orange) with a low discriminating threshold (Fig.Ā 9). We also provide the ROC curve or accuracy of the ensemble model in Fig.Ā 10. It is important to note that the two points of ResNet can be obtained at the same site. These points were not simultaneously established for VGGNet since its prediction result was suboptimal. Regarding their ability to recognize gingivitis, the ResNet model achieved a larger AUC of 97% than the GoogLeNet (AUCā€‰=ā€‰94%), Alexnet (AUCā€‰=ā€‰92%), and VGG (AUCā€‰=ā€‰89%) models (Fig.Ā 9). The ResNet and GoogLeNet models and label (correct category result) had the highest sensitivity of all models tested, and there was not a substantial difference in the projected outcomes between the ResNet and GoogLeNet groups (p>0.05). Meanwhile, the sensitivity of VGGNet was significantly different from those of the other groups (pā€‰<ā€‰0.001; TableĀ 2; Fig.Ā 11).

Fig. 9
figure 9

ROC curves of AlexNet, VGG, GoogLeNet, and ResNet during training. To reduce false positives, we established a high-specificity point (blue) with a high discrimination threshold. To maintain a low missing rate, we selected a high-sensitivity point (orange) with a low discriminating threshold. These points were not simultaneously established for the VGGNet model since its prediction result was suboptimal

Fig. 10
figure 10

ROC curves with ensemble Model

Fig. 11
figure 11

Sensitivity of the AlexNet, VGG, GoogLeNet, and ResNet models during training and testing. The AUC value increased during training and decreased during testing. Among the models, the prediction result of ResNet changed the most

The task addressed in this study was a dichotomous classification task, and the amount of data was less than required for model training datasets. To make use of this small amount of data and ensure the accuracy of the model training process, we employed transfer learning during training. Well-learned networks retain some common abstractions of gingivitis, and the new neural network can adopt the high-latitude features of the previous underlying general-purpose network. Transfer learning can hasten the convergence of the model and enhance recognition performance. The loss of GoogleNet and ResNet models continued to fluctuate until 25 epochs, after which they stabilized (Figs.Ā 12 and 13). After a few variations, Alexnet and VGG quickly stabilized after roughly the same number of epochs. It is important to note that the modelā€™s prediction performance is shown by the accuracy on the right. We chose the best result as the final model since accuracy decreases as the number of training epoch increases.

Our study has several shortcomings. First, it is worth mentioning that several models have the problem of overfitting gradient descent after sufficient training; in other words, the modelā€™s prediction is highly consistent with the training set, making it difficult to accurately detect the new data. Therefore, we retained the results when the training reached the highest accuracy; later updates will only consider the original parameters after the original ideal recognition performance is achieved. Second, although we employed transfer learning to improve the accuracy of the ResNet model in identifying gingivitis from oral photographs, our dataset collected from patients was limited; thus, we plan to expand the amount of data from patients of different ages and for a wider range of ethnic groups. Third, some issues during the training process may lead to the failure of gingivitis detection; for instance, improper lighting can cause light spots on gingiva, which may be misidentified as gingivitis (see Fig.Ā 8). Thus, we need composite images with fewer light spots, and pre-processing steps should be designed to reduce noise without eliminating real features. Finally, to improve the model accuracy and reliability, some text datasets such as clinically diagnosed cases or traditional questionnaires [25] can be used to complement image datasets [26].

Fig. 12
figure 12

Loss and accuracy of the AlexNet, VGG, GoogLeNet, and ResNet models during training

Fig. 13
figure 13

Loss and accuracy of the AlexNet, VGG, GoogLeNet, and ResNet models of Validation

We show Grad-CAM based on the last convolutional layer of each model on several test data samples. These heatmaps can help understand which areas of the image the model focuses on (Fig.Ā 14). Simultaneously, the incorrect predictions of the picture by the model is likely attributed to issues with reflective light during the shooting process. For instance, the subsequent heat map failed to accurately detect gingivitis. FigureĀ 15 is a heat map illustrating a selection of the images.

Fig. 14
figure 14

Grad-CAM of correct predictions of gingivitis

Fig. 15
figure 15

Grad-CAM of incorrect predictions of gingivitis

Conclusion

Deep learning plays an essential role in dental disease recognition. We extracted data multiple times for complex tasks with limited data and used ensemble learning to improve model performance. Among the tested models, the ResNet and GoogleNet models performed best, and transfer learning increased the accuracy of the models to recognize gingivitis from oral images.

Data availability

The data used in current study were collected from Medical School of Nanjing University and is available only for the granted research. However, the data can be made available if requested within data protection and regulation guideline. https://github.com/Eating-GET/Tooth_Data. The datasets used during the study will be available from the corresponding author upon request.

References

  1. Kamath DG, Nadimpalli H, Nayak SU, Rajendran V, Natarajan S. Comparison of antiplaque and anti-gingivitis effects of aloe vera mouthwash with chlorhexidine in fixed orthodontic patients-A randomized controlled trial. Int J Dental Hygiene. 2023;21(1):211ā€“8.

    ArticleĀ  Google ScholarĀ 

  2. Shayani A, Aravena PC, RodrĆ­guez-Salinas C, Escobar-Silva P, Diocares-MonsĆ”lvez Y, Angulo-GutiĆ©rrez C, Rivera C. Chemotherapy as a risk factor for caries and gingivitis in children with acute lymphoblastic leukemia: a retrospective cohort study. Int J Paediatr Dent. 2022;32(4):538ā€“45.

    ArticleĀ  PubMedĀ  Google ScholarĀ 

  3. Nazir MA. Prevalence of periodontal disease, its association with systemic diseases and prevention. Int J Health Sci (Qassim). 2017;11(2):72ā€“80.

    PubMedĀ  Google ScholarĀ 

  4. Murakami S, Mealey BL, Mariotti A, Chapple ILC. Dental plaque-induced gingival conditions. J Periodontol. 2018;89(1):S17ā€“27.

    PubMedĀ  Google ScholarĀ 

  5. Ohira H, Yamamoto I, Sadamori T, Miyauchi H, Nakagawa K, Hukushige H, Suzuki E, Yamada Y, Nakagawa K, Sakurai T, Harada H, Ishizu K, Teramura N, M.Takagaki Y, Takeuchi R. Transmission of dental data on amicrodotted paper-based dental chart from a digital pen - trial operations at disaster exercise sites and at a landslide disaster at Hiroshima City. Can Soc Forensic Sci J. 2016;49:189ā€“202.

    ArticleĀ  Google ScholarĀ 

  6. Litjens G, Kooi T, Bejnordi BE, Setio AAA, Ciompi F, Ghafoorian M, van der Laak JAWM, van Ginneken B, SĆ”nchez CI. A survey on deep learning in medical image analysis. Med Image Anal. 2017;42:60ā€“88.

    ArticleĀ  PubMedĀ  Google ScholarĀ 

  7. Wang G, Li W, Zuluaga MA, Pratt R, Patel PA, Aertsen M, Doel T, David AL, Deprest J, Ourselin S, Vercauteren T. Interactive medical image Segmentation using Deep Learning with Image-Specific Fine tuning. IEEE Trans Med Imaging. 2018;37(7):1562ā€“73.

    ArticleĀ  PubMedĀ  Google ScholarĀ 

  8. Li W, Chen. YY, Sun. WB, Mackenzie. B, Zhang. X, Wang, SH. Miao. LY. A gingivitis identification method based on contrast-limited adaptive histogram equalization, gray-level co-occurrence matrix, and extreme learning machine. Int J Imaging Syst Technol. 2019;29:77ā€“82.

    ArticleĀ  CASĀ  Google ScholarĀ 

  9. Ekert T, Krois J, Meinhold L, Elhennawy K, Emara R, Golla T, Schwendicke F. Deep learning for the Radiographic Detection of Apical Lesions. J Endod. 2019;45(7):917ā€“e9225.

    ArticleĀ  PubMedĀ  Google ScholarĀ 

  10. Li W, Liang Y, Zhang X, et al. A deep learning approach to automatic gingivitis screening based on classification and localization in RGB photos. Sci Rep. 2021;11(1):16831.

    ArticleĀ  CASĀ  PubMedĀ  PubMed CentralĀ  Google ScholarĀ 

  11. Iandola FN, Han S, Moskewicz MW, Ashraf K, Dally WJ, Keutzer K. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <ā€‰0.5Ā MB model size, arXiv preprint arXiv:1602.07360. 2016.

  12. Xu Y, Jia Z, Wang LB, Ai Y, Zhang F, Lai M, Chang EI. Large scale tissue histopathology image classification, segmentation, and visualization via deep convolutional activation features. BMC Bioinformatics. 2017;18(1):281.

    ArticleĀ  PubMedĀ  PubMed CentralĀ  Google ScholarĀ 

  13. Lee H, Kwon H. Going deeper with contextual CNN for Hyperspectral Image classification. IEEE Trans Image Process. 2017;26(10):4843ā€“55.

    ArticleĀ  PubMedĀ  Google ScholarĀ 

  14. Ioffe SC, Alemi SVV. AA. Inception-v4, inception-resnet and the impact of residual connections on learning. Thirty-first AAAI conference on artificial intelligence. 2017; 4278ā€“4283.

  15. Cheplygina V, de Bruijne M, Pluim JPW. Not-so-supervised: a survey of semi-supervised, multi-instance, and transfer learning in medical image analysis. Med Image Anal. 2019;54:280ā€“96.

    ArticleĀ  PubMedĀ  Google ScholarĀ 

  16. Zhang ZJ. Improved adam optimizer for deep neural networks, 2018 IEEE/ACM 26th International Symposium on Quality of Service (IWQoS). IEEE. 2018.

  17. Mummolo S, Cirillo E, CiribĆØ M, Manenti RJ, Galeotti A. Periodontology. Part 1: Gingivitis in adolescence. Review of the literature and case reports. Eur J Paediatr Dent. 2022;23(1):79ā€“82.

    CASĀ  PubMedĀ  Google ScholarĀ 

  18. Eke PI, Dye BA, Wei L, et al. Self-reported measures for surveillance of periodontitis. J Dent Res. 2013;92(11):1041ā€“7.

    ArticleĀ  CASĀ  PubMedĀ  Google ScholarĀ 

  19. Alalharith DM, Alharthi HM, Alghamdi WM, et al. A deep learning-based Approach for the Detection of Early Signs of Gingivitis in Orthodontic patients using faster region-based convolutional neural networks. Int J Environ Res Public Health. 2020;17(22):8447.

    ArticleĀ  PubMedĀ  PubMed CentralĀ  Google ScholarĀ 

  20. Lang NP, Bartold PM. Periodontal health. J Periodontol. 2018;89(Suppl 1):S9ā€“16.

    PubMedĀ  Google ScholarĀ 

  21. Greener JG, Kandathil SM, Moffat L, Jones DT. A guide to machine learning for biologists. Nat Rev Mol Cell Biol. 2022;23(1):40ā€“55.

    ArticleĀ  CASĀ  PubMedĀ  Google ScholarĀ 

  22. Jin B, Cruz L, GonƧalves N. Deep Facial Diagnosis: Deep Transfer Learning From Face Recognition to Facial Diagnosis, IEEE ACCESS, 2020.

  23. Rahman AU, Alqahtani A, Aldhafferi N, et al. Histopathologic oral Cancer prediction using oral squamous cell Carcinoma Biopsy empowered with transfer learning. Sens (Basel). 2022;22(10):3833.

    ArticleĀ  Google ScholarĀ 

  24. Chang YJ, Huang TY, Liu YJ, Chung HW, Juan CJ. Classification of parotid gland tumors by using multimodal MRI and deep learning. NMR Biomed. 2021;34(1):e4408.

    ArticleĀ  PubMedĀ  Google ScholarĀ 

  25. Liang Y, Fan HW, Fang ZJ et al. OralCam: enabling self-examination and awareness of oral health using a smartphone camera. CHI 2020, Honolulu, HI, USA. 2020;25ā€“30.

  26. Choi E, Xiao C, Stewart WF, Sun J. Mime: Multilevel medical embedding of electronic health records for predictive healthcare. Adv Neural Inf Process Syst. 2018;4:456ā€“61.

    Google ScholarĀ 

Download references

Acknowledgements

The authors are very grateful for all respondents who contributed the data used in this paper. Hong Zhao and Enting Guo gratefully acknowledge the financial support by the China Scholarship Council, China (No. 202208050053 and No. 202208050055).

Funding

This work was supported by the Natural Science Foundation of Jiangsu under Grant (grant no. BK20221177), ā€œ2015ā€ Cultivation Program for Reserve Talents for Academic Leaders of Nanjing Stomatological School Medical School of Nanjing Univeristy (grant no. 0223A203 and no. 0223A210), the Key Project supported by Medical Science and Technology Development Foundation, Nanjing Department of Healthunder Grant (grant no. ZKX22055), Postgraduate Research & Practice Innovation Program of Jiangsu Province (grant no. KYCX23_0192), the General Project supported by Medical Science and Technology Development Foundation, Nanjing Department of Health under (grant no.YKK23183).The figures and tables was drawn by Enting Guo. The authors declare no potential conflicts of interest with respect to the authorship and/or publication of this article.

Author information

Authors and Affiliations

Authors

Contributions

Wen Li: Study conception, Design, Drafting, Critical review of the manuscript. EntingGuo: Data analysis, Interpretation, Drafting of the manuscript. Hong Zhao: Investigation, Data acquisition, Editing, Validation. Yuyang Li: Formal analysis, Investigation, Data curation. Leiying Miao: Data curation, Writing ā€“ review & editing, Funding acquisition. Chao Liu: Study conception, Design as well as critical review of the manuscript, Funding acquisition. Weibin Sun: Visualization and resources, Review and editing, Validation and original draft.All authors reviewed the manuscript.

Corresponding authors

Correspondence to Chao Liu or Weibin Sun.

Ethics declarations

Ethics approval and consent to participate

The study was performed in accordance with the Declaration of Helsinki, and approved by the Ethical Review Board at Nanjing University. Ethics committee approval number: NJSH-2022NL-069. All respondents were informed comprehensively and provided written informed consent. For children under the age of 16, the written informed consent was obtained from their parents/guardians.

Consent for publication

Not applicable.

Competing interests

The authors declare no competing interests.

Additional information

Publisherā€™s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the articleā€™s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the articleā€™s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Li, W., Guo, E., Zhao, H. et al. Evaluation of transfer ensemble learning-based convolutional neural network models for the identification of chronic gingivitis from oral photographs. BMC Oral Health 24, 814 (2024). https://doi.org/10.1186/s12903-024-04460-x

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12903-024-04460-x

Keywords