Skip to main content
Advertisement
  • Loading metrics

Deep learning-based assessment of pulp involvement in primary molars using YOLO v8

  • Aydin Sohrabi,

    Roles Conceptualization, Data curation, Funding acquisition, Investigation, Methodology, Writing – original draft, Writing – review & editing

    Affiliations Department of Orthodontics, Faculty of Dentistry, Tabriz University of Medical Sciences, Tabriz, Iran, Faculty of Medicine & Dentistry, Mike Petryk School of Dentistry, University of Alberta, Edmonton, Canada

  • Nazila Ameli,

    Roles Formal analysis, Methodology, Writing – original draft, Writing – review & editing

    Affiliation Faculty of Medicine & Dentistry, Mike Petryk School of Dentistry, University of Alberta, Edmonton, Canada

  • Masoud Mirimoghaddam,

    Roles Conceptualization, Data curation, Writing – review & editing

    Affiliation Faculty of Medicine & Dentistry, Mike Petryk School of Dentistry, University of Alberta, Edmonton, Canada

  • Yuli Berlin-Broner,

    Roles Validation, Writing – review & editing

    Affiliation Faculty of Medicine & Dentistry, Mike Petryk School of Dentistry, University of Alberta, Edmonton, Canada

  • Hollis Lai,

    Roles Validation, Writing – review & editing

    Affiliation Faculty of Medicine & Dentistry, Mike Petryk School of Dentistry, University of Alberta, Edmonton, Canada

  • Maryam Amin

    Roles Conceptualization, Methodology, Writing – review & editing

    maryam.amin@ualberta.ca

    Affiliation Faculty of Medicine & Dentistry, Mike Petryk School of Dentistry, University of Alberta, Edmonton, Canada

Abstract

Dental caries is a major global public health problem, especially among young children. Rapid decay progression often necessitates pulp treatment, making accurate pulp condition assessment crucial. Despite advances in pulp management techniques, diagnostic methods for assessing pulp involvement have not significantly improved. This study aimed to develop a machine learning (ML) model to diagnose pulp involvement using radiographs of carious primary molars. Clinical charts and bitewing radiographs of 900 children treated from 2018-2022 at the University of Alberta dental clinic were reviewed, yielding a sample of 482 teeth. images were preprocessed, standardized, and labeled based on clinical diagnoses. Data were split into training, validation, and test sets, with data augmentation applied to classify 2 categories of outcomes. The YOLOv8m-cls model architecture included convolutional and classification layers, and performance was evaluated using top-1 and top-5 accuracy metrics. The YOLOv8m-cls model achieved a top-1 accuracy of 78.7% for upper primary molars and 87.8% for lower primary molars. Validation datasets showed higher accuracy for lower primary teeth. Performance on new test images demonstrated precision, recall, accuracy, and F1-scores, highlighting the model’s effectiveness in diagnosing pulp involvement, with lower primary molars showing superior results. This study developed a promising CNN model for diagnosing pulp involvement in primary teeth using bitewing radiographs, showing promise for clinical application in pediatric dentistry. Future research should explore whole bitewing images, include clinical variables, and integrate heat maps to enhance the model. This tool could streamline clinical practice, improve informed consent, and assist in dental student training.

Author summary

We developed an AI-based diagnostic tool for identifying dental pulp involvement in primary teeth, which is a common complication of severe cavities requiring careful assessment to determine optimal treatment plans. Our system analyzes dental X-rays of primary teeth to assist in diagnosis of pulp involvement in deep caries. Results demonstrate high diagnostic accuracy, offering several benefits: faster, more reliable diagnoses improving treatment efficiency; enhanced communication between dental practitioners and parents; and educational value as a training resource for dental students to supplement their limited clinical experience and improve their learning. By using modern technology in children’s dental care, we hope to make diagnoses more accurate and ultimately enhance children’s oral health outcomes.

Introduction

Dental caries is a serious global public health problem, especially among young children. The severity ranges from the white spot lesions in enamel to extensive cavitation [1]. According to the National Health and Nutrition Examination Survey, 51% of children aged 6–8 in the United States have dental caries in their primary teeth [2].

Caries in the primary teeth tend to progress rapidly and extend to involve the pulp that necessitates pulp treatment [3]. Therefore, assessing the pulp condition is a crucial diagnostic procedure in pediatric dentistry for reaching a proper treatment decision. Conventionally, dentists rely on patients’ history of pain, clinical examinations, and radiographic images of the teeth and the periodontal tissues, as well as pulpal and periodontal diagnostic tests performed [4]. The three techniques commonly used to manage pulpitis in primary teeth include direct pulp capping, pulpotomy, and pulpectomy; each using various evolving and improving materials. These techniques have been successfully developed over the past few decades. However, there has not been a significant advancement in the diagnostic methods for determining whether dental caries have reached the pulpal tissue.

One of the most challenging decisions in the treatment of carious primary molars is assessing the severity and progression of caries and accurately determining if the decay has reached the pulp, as well as the extent of pulp inflammation or infection. The reliability of the dental history provided by the child patient complicates clinical decisions about the probability of pulp involvement and the need for pulp treatment. Clinicians, therefore, commonly make these decisions relying on bitewing images to assess the location and extent of caries, as well as any evidence of radicular pathology in x-rays [5]. They also consider clinical findings such as tenderness and pain history. However, correlating symptoms with pulpal status is often challenging in pediatric patients [6]. In some uncertain cases, caries removal method is employed in that removal of the carious lesion is done layer-by-layer and in small amounts preserves the pulp vitality and intactness in the case if the caries has not reached the pulp yet [7]. Making this decision requires substantial knowledge and experience.

Furthermore, in borderline cases where it is uncertain whether pulp treatment is necessary, obtaining informed consent from parents becomes challenging. Clinicians must consider parents’ preferences, the family’s ability to manage dental care costs, and the uncertainty surrounding the need for pulp treatment, which add to the complexity of treatment planning. A reliable diagnostic tool would streamline the treatment process, making it easier to obtain informed consent and provide a more accurate cost estimate to parents. This would benefit not only students in dental school clinics but also early-career dentists.

Artificial intelligence (AI), particularly deep learning (DL) with convolutional neural networks (CNN), is rapidly advancing in dental research. Especially in imaging, the use of DL with CNNs to process various types of images has been actively researched and has shown promising performance [8,9]. Recently successful AI models have emerged for caries detection [10], dental age estimation [1113], sex estimation from teeth [10], working length determination, and morphology detection [14], and assessing endodontic treatment difficulty [15].

The process of classification in a CNN model involves: 1) Convolutional layers (the first step) to extract features such as gradients or edges from the input image using the mathematical transformations, 2) Non-linear activation functions, which are placed between any two layers and guide the input signals into output signals required for the NN to act, 3) Pooling layer, which reduces the number of parameters to learn and the amount of computation to summarize the features generated by the convolution layer, 4) Fully connected layers that are responsible for the interpretation of the feature representations learned by preceding layers [16]. YOLO is an algorithm employing CNN principles above, designed to precisely identify objects in real-time [17].

The aims of this study were: (1) to develop an ML model to diagnose the risk of pulp involvement in bitewing images of carious primary molars, and (2) to test the diagnostic performance of the developed model in a new set of samples.

Methods and sample

Ethics statement

This study was conducted in accordance with the ethical standards of the Declaration of Helsinki. Ethical approval was obtained from the Research Ethics Board of the University of Alberta (Approval Number: Pro00121109).

Study population

This study was approved by the Health Research Ethics Board University of Alberta (HREB-Pro00121109). Clinical charts of 900 pediatric patients treated by Year 3 and Year 4 dental students between January 2018 and December 2022 at the University of Alberta Kaye Edmonton Clinic were screened. A sample of patients with carious primary molars was identified. Bitewing radiographs of the treated carious primary molars were examined, along with clinical data including post-treatment diagnoses of pulp involvement condition. Data was obtained from the axiUm database of the pediatric dentistry clinic. Samples with missing radiographic or clinical data, and low-quality images, such as those with cone-cuts, overlaps, or other artifacts, were excluded.

Data preprocessing

Preprocessing steps involved image standardization to a consistent resolution, size (280X280 pixels), and format, which was automatically adjusted to 288 to ensure compatibility with the model’s stride. The images were then normalized to compensate for under- and overexposure. In the preprocessing stage, the bitewing radiographs were cropped to display only one tooth per image in the optimal position. If multiple teeth in the same patient were treated during a session, the bitewing images were cropped to create individual images for each treated tooth.

Sample labeling

Clinical data from patient charts were used as ground truth to label bitewing radiographs for training the proposed AI model. The images were labeled as “pulp involved”, for caries with definitive pulp involvement, and “pulp not involved”, for caries without definitive pulp involvement (teeth with caries only), according to the real final clinical situation of pulp involvement as registered in the patients’ dental chart records and received treatment.

The two datasets, which comprised radiographs of lower and upper primary molars were labeled. Data extraction was repeated twice on 30 images to ensure accuracy. The datasets were then split into training, validation, and test sets with a ratio of approximately 70:20:10, Data augmentation was done before splitting ensuring no overlap between the sets. This was done using techniques such as rotation, scaling, horizontal flipping, and translation to increase data variability. The final cohort of radiographs included 688 images for training and 253 images for validation of upper primary molars, and 607 images for training and 238 images for validation of lower primary molars.

Model architecture

The YOLOv8m-cls model architecture included multiple convolutional layers followed by classification layers. The training utilized the AdamW optimizer with an auto-determined learning rate of 0.0001 and momentum of 0.9, aiming to minimize the classification loss function. The training process consisted of 40 epochs for both datasets, with early stopping based on validation loss to prevent overfitting. A batch size of 16 was used, and Automatic Mixed Precision (AMP) was employed to accelerate training by utilizing mixed precision arithmetic while maintaining numerical stability. Pretrained weights were utilized, with 228 out of 230 items successfully transferred to facilitate better starting performance and improved convergence. The entire training process, including the results and model architecture, was logged using TensorBoard, enabling detailed visualization and analysis of the training and validation phases.

Model performance was evaluated using top-1 accuracy metrics on the validation sets after each epoch. The final trained models’ performances were assessed on the set of new unseen test images (24 lower primary molars with pulp involvement and 32 lower primary molars without pulp involvement for the lower teeth model, and 31 upper primary molars with pulp involvement and 31 upper primary molars without pulp involvement for the upper teeth model) to evaluate its generalization capability in classifying primary tooth radiographs based on pulp involvement.

Visualization of model diagnosis

Gradient-weighted Class Activation Mapping (Grad-CAM) was used to visualize the regions of an image that were most focused on by the model during its decision-making process and to improves model transparency. Grad-CAM is a visualization technique that generates heatmaps to highlight the most influential areas in a neural network’s predictions. By computing the gradients of the target class with respect to the feature maps, Grad-CAM produces an activation map that overlays onto the original image, providing insight into the model’s focus areas [18].

Statistical analysis

Classification accuracy measures were used to evaluate outcomes from the test image sets. In ML different evaluation metrics are applied according to the type of problem. Accuracy, precision, recall, and the F1-score are used for classification tasks. As this study was based on a classification task, the evaluation criteria of accuracy, precision, recall, and the F1-score were used to evaluate the classification performance of the proposed model. A confusion matrix was used to calculate these values [19]. The confusion matrix has true positive (TP), true negative (TN), false positive (FP), and false negative (FN) values. The equations for accuracy, recall, precision, and the F1-score, which are performance evaluation metrics, are provided below:

Results

A sample of patients with carious primary molar teeth was selected, ranging in age from 3.8 to 11.7 years, with a median age of 7.8 years. Their clinical charts included 482 carious primary molars: 239 lower and 243 upper primary teeth. Of the 239 lower teeth, 124 received pulp treatment, while 115 were restored without pulp treatment. Among the 243 upper teeth, 120 had pulp involvement, and 123 received only restorations. The model performance for the upper teeth showed a top-1 accuracy of 78.7% and a top-5 accuracy of 1, while the results for the lower teeth achieved a top-1 accuracy of 87.8% and a top-5 accuracy of 1 (Fig 1-A and B).

thumbnail
Fig 1. Yolo-v8 classification model results on upper (A) and lower (B) primary molar.

https://doi.org/10.1371/journal.pdig.0000816.g001

Table 1 represents the model performance on the classification of primary molars in the validation datasets for both upper and lower primary molars. The model showed higher accuracy in classifying the lower primary teeth compared to the upper primary molars.

thumbnail
Table 1. Performance of the model on validation datasets for classifying the primary molars based on the pulp involvement.

https://doi.org/10.1371/journal.pdig.0000816.t001

Table 2 depicts the performance of the model on classifying two new sets of test images for upper and lower teeth. The F1-Score value, which is usually used to judge the overall performance of the model and is defined as the average of precision and recall values, showed that our proposed models performed highly accurately in classifying the involvement of pulp in primary teeth using bitewing radiographs. Similarly, the accuracy of the model in the correct classification of teeth with and without pulp involvement was greater in lower primary molars.

thumbnail
Table 2. Performance of the model on a new set of images for classifying the primary molars based on the pulp involvement.

https://doi.org/10.1371/journal.pdig.0000816.t002

Grad-CAM visualizations are presented in Fig 2, these heat maps were generated to illustrate the model’s focus areas during classification for both positive (pulp involved) and negative (pulp not involved) samples.

thumbnail
Fig 2. Grad-CAM visualizations illustrate the model focus areas during classification for positive (A) (pulp involved) and negative (B) (pulp not involved) samples.

https://doi.org/10.1371/journal.pdig.0000816.g002

Discussion

In this study, we developed a CNN model using YOLOv8m-cls for predicting pulp involvement on bitewing radiographs in carious primary molars. Our model demonstrated high classification accuracy of over 75% for the upper and over 80% for the lower teeth and showed the potential to significantly assist clinicians in diagnosing pulp involvement in clinical settings.

In the present study, we used bitewing images of primary teeth and trained a YOLO-v8 classification model to classify teeth into two distinct groups based on the involvement of the pulp. The annotating step was skipped in the proposed model, which resulted in a more time-efficient image pre-processing. In the literature, older versions of YOLO have been widely used in the field of dental radiology to detect mandibular fractures in panoramic radiographs [20], primary and permanent tooth detection on pediatric dental radiographs [21], detection of cysts and tumors of the jaw [22], and detection of impacted mandibular third molar teeth [23]. However, new versions of YOLO-v8, released in January 2023, demonstrate superior performance for throughput and computational load requirements [24], and provide a network architecture that requires lower computing and training requirements, hence providing a more effective feature integration method, more accurate object detection performance, a more robust loss function, and an increased label assignment and model training efficiency [25,26].

The final phase of most ML studies is diagnostic performance or accuracy testing. In this phase, the findings of ML model are validated against a gold standard defined as an error-free reference standard [27]. In the present study, the post-treatment diagnosis was used as the ground truth while most ML studies in dental and medical domains rely on diagnoses from expert panels, which can exhibit errors due to the subjective nature of the task, individual expertise, biases towards certain diagnoses, among other factors. This introduces limitations regarding the certainty of the ground truth labels. It is crucial to recognize that while expert panel labeling offers necessary reference points for model training and evaluation, it may not always reflect the true ground truth [28]. Instead, we validated the model against the actual final diagnoses recorded in the patients’ clinical charts to determine pulp involvement. The images were labeled based on this information rather than relying solely on the diagnosis of dentists. This approach allowed the developed model to exceed the accuracy of experienced dentists, approaching maximum accuracy.

In the present study, if validated against the diagnoses of one or more experts, the accuracy of the developed model would likely match the expert level. It is acknowledged that for any individual specialist, accuracy is rarely 100%, and there can be instances where a dentist unexpectedly encounters pulp involvement when removing a carious lesion.

In a recent study, Ma et. al. developed a model to classify deep carious primary molars of patients 3–8 years old into two groups: 1) The failure group defined as teeth that required non-vital pulp therapy or extraction during the next 18-month observation period, and 2) The success group defined as teeth that remained clinically and radiographically free of pathology during the observation period. They labeled their samples based on the actual clinical outcomes rather than relying solely on the expert panel diagnosis. They utilized 348 samples of periapical radiographs of primary teeth, achieving 90% accuracy in predicting the prognosis. The outcomes included whether the teeth received non-vital pulp therapy, were extracted, or remained pathology-free during observation [29]. The methodologies of the two studies are comparable, with both demonstrating that utilizing ML models is promising for assessing pulp conditions in primary molars. However, there are differences as well. Ma et al. labeled samples as either needing pulp treatment or extraction within the same group, while our study excluded extracted samples, aiming to develop a model specifically for diagnosing teeth in need of pulp treatment. Their study included 65 teeth in the “failure” group and 357 in the “success” group. Another difference is related to the time interval between the x-ray taken and the treatment performed. They selected a period of up to 18 months, whereas we limited it to 6 months, excluding any x-rays taken more than 6 months before the treatment session.

In a recent study, Wang et al. evaluated ML-based predictions of pulp exposure in permanent teeth using periapical radiographic images. Their methodology closely resembled ours, particularly in the selection of reference standard and case-labeling processes. However, their study focused on permanent teeth and periapical images, whereas ours concentrated on primary teeth and bitewing radiographs. Additionally, while Wang et al. employed three ML models (VGG, ResNet, and DenseNet), we adapted the YOLOv8 architecture for classification. They trained the models on 126 samples (56 with no exposure and 70 with exposure in posterior teeth). The reference standard in this study was clinical pulp exposure observed after cavity preparation. They tested the developed models on a separate set of 40 samples (24 without exposure and 16 with exposure). DenseNet achieved the best performance in their study (AUC = 0.97), followed by ResNet (AUC = 0.89), VGG (AUC = 0.78), and senior dentist (AUC = 0.87). Their reported accuracy levels for the three trained models ranged between 75% and 85%, which are comparable to ours (78.7%–87%) [30]. Ramezanzade et al. also compared the effect of providing AI-based radiographic information versus standard radiographic and clinical information on dental students’ ability to predict pulp exposure. While their study used bitewing radiographs similar to ours, the primary goal was to assess the impact of AI assistance through the application of convolutional neural networks (CNN) based on ResNet-50 architecture on students’ diagnostic capabilitieson students’ diagnostic capabilities. Their findings showed that AI outperformed the students, the participants only benefited “slightly” when provided with AI predictions. Notably, the authors emphasized the need for more explainable AI systems to address real-world clinical complexities, including factors such as provider experience and patient expectations [31].

We used a comparable number of “positive” and “negative” cases in test sets. It is known that diagnostic accuracy is affected by disease prevalence [32]. Disease prevalence in a population impacts positive and negative predictive values. With high positive cases in the test set, a test will be reported as more effective at ‘ruling in’ the positive condition and less effective at ‘ruling it out [33]. Consequently, having a representative and comparable prevalence of “positive” and “negative” conditions will assure researchers about the validity and the generalizability of the results. This point is reflected in the comparable levels of accuracy in the preditction of the “pulp involved” and “pulp not involved” categories in our study (0.74 vs 0.79 for upper molars, and 0.88 vs 0.83 for lower molars).

In the present study, a small difference was observed in the precision achieved for upper molars compared to lower molars. While the clinical significance of this difference is uncertain, it can likely be attributed to the more complicated anatomy of pulp chambers in the upper molars, due to the complex anatomy of roots and canals compared to lower molars, as well as the fact that upper molars typically have three roots versus the two roots in lower molars.

This study has some limitations that need to be acknowledged. First, cases retrieved from the University pediatric clinic were treated by undergraduate dental students under the supervision of experienced instructors. This may have introduced the possibility of over-cavity preparation and dentine removal in some carious teeth, leading to a higher number of exposed pulps compared to ideal conditions. This could result in a bias towards predicting more pulp involvement in borderline cases. In addition, we cropped the included tooth from the bitewing images for training the ML model. Consequently, this ML model is designed to accurately analyse single tooth images rather than entire bitewing images. For future studies, a better approach would be to use whole bitewing images to train new models, enabling AI to be flexibly applied to entire images. However, this approach requires additional training steps to differentiate between intact teeth, buds of unerupted permanent teeth, and erupting permanent teeth visible in the images. Third, we used images of the carious teeth to develop the model and integrated clinical examination data along with patient-related factors such as age, eruption stage of underlying permanent teeth, estimated time to eruption of successors, and other relevant (influential) treatment planning variables. This comprehensive approach has the potential to significantly enhance the model’s accuracy. However, the ease of use in clinical practice should also be considered, as an increase in the number of input factors might make the model more complex and time-consuming to use. Finally, we did not create heat-map images to identify the focus areas of the trained model in detecting pulp involvement, which means that the principles by which this AI model diagnoses the possibility of pulp involvement are still unknown. As this is the first AI study in the field, our primary objective was to evaluate the feasibility of this approach. Therefore, future projects may focus on creating heat maps and uncovering potential hidden features in this context.

Refined models using this method could enhance the accuracy of diagnoses and treatment plans in clinical practice for pediatric dentists. This improvement would aid in providing more reliable cost estimates for treatments and obtaining precise informed consent. Additionally, it could streamline patient scheduling by better estimating the need for pulp treatment and help general dentists distinguish between manageable and complex cases, prompting appropriate referrals. Moreover, the model serves as a valuable training tool for dental students, allowing them to compare their diagnostic assessments with AI predictions and enhancing their clinical skills. The rapid development of smartphones presents an opportunity to streamline dental practice through mobile apps. Developing an app based on this model would offer convenient access, akin to commercially available apps like WebCeph for orthodontic diagnosis.

Conclusion

This study successfully developed a convolutional neural network (CNN) model, specifically using YOLOv8m-cls, for diagnosing pulp involvement in primary teeth from bitewing radiographs, achieving high classification accuracy. The model demonstrated over 75% accuracy for both upper and lower primary molars, with superior results for lower molars. Future research should focus on training models with whole bitewing images and incorporating additional clinical variables to further improve accuracy. Despite some limitations, this study lays a solid foundation for AI applications in pediatric dentistry, empowering clinicians to make informed decisions, enhance patient care processes, and advance dental education. The potential development of a mobile app underscores the transformative impact of AI on dental diagnostics and treatment planning, particularly in pediatric practice.

References

  1. 1. Featherstone JD. The science and practice of caries prevention. J Am Dent Assoc. 2000;131(7):887–99. pmid:10916327
  2. 2. Lin M, Griffin S, Gooch B, Espinoza L, Wei L, Li C-H. Oral health surveillance report: trends in dental caries and sealants, tooth retention, and edentulism, United States: 1999–2004 to 2011–2016. National Center for Chronic Disease Prevention and Health Promotion; 2019.
  3. 3. Nadin G, Goel B, Yeung C, Glenny A. Pulp treatment for extensive decay in primary teeth. Cochrane Database of Systematic Reviews. 2003;2003(1):CD003220.
  4. 4. Gluskin AH, Peters CI, Peters OA. Minimally invasive endodontics: challenging prevailing paradigms. Br Dent J. 2014;216(6):347–53. pmid:24651341
  5. 5. American Academy of Pediatric Dentistry. Pulp therapy for primary and immature permanent teeth. The Reference Manual of Pediatric Dentistry. 2021:399–407.
  6. 6. Igna A, Mircioagă D, Boariu M, Stratul Ș., . A diagnostic insight of dental pulp testing methods in pediatric dentistry. Medicina (Kaunas). 2022;58(5):Page range needed.
  7. 7. Fuks AB. Vital pulp therapy with new materials for primary teeth: new directions and Treatment perspectives. Pediatr Dent. 2008;30(3):211–9. pmid:18615986
  8. 8. Ghaffari M, Zhu Y, Shrestha A. A review of advancements of artificial intelligence in dentistry. Dentistry Review. 2024;4(2):100081.
  9. 9. Hartoonian S, Hosseini M, Yousefi I, Mahdian M, Ghazizadeh Ahsaie M. Applications of artificial intelligence in dentomaxillofacial imaging: a systematic review. Oral Surg Oral Med Oral Pathol Oral Radiol. 2024;138(5):641–55. pmid:38637235
  10. 10. Albano D, Galiano V, Basile M, Di Luca F, Gitto S, Messina C, et al. Artificial intelligence for radiographic imaging detection of caries lesions: a systematic review. BMC Oral Health. 2024;24(1):274. pmid:38402191
  11. 11. Dogan OB, Boyacioglu H, Goksuluk D. Machine learning assessment of dental age classification based on cone-beam CT images: a different approach. Dentomaxillofac Radiol. 2024;53(1):67–73. pmid:38214945
  12. 12. Kokomoto K, Kariya R, Muranaka A, Okawa R, Nakano K, Nozaki K. Automatic dental age calculation from panoramic radiographs using deep learning: a two-stage approach with object detection and image classification. BMC Oral Health. 2024;24(1):143. pmid:38291396
  13. 13. Merdietio Boedi R, Shepherd S, Oscandar F, Franco AJ, Mânica S. Machine learning assisted 5-part tooth segmentation method for CBCT-based dental age estimation in adults. J Forensic Odontostomatol. 2024;42(1):22–9. pmid:38742569
  14. 14. Aminoshariae A, Kulild J, Nagendrababu V. Artificial Intelligence in Endodontics: Current Applications and Future Directions. J Endod. 2021;47(9):1352–7. pmid:34119562
  15. 15. Karkehabadi H, Khoshbin E, Ghasemi N, Mahavi A, Mohammad-Rahimi H, Sadr S. Deep learning for determining the difficulty of endodontic treatment: a pilot study. BMC Oral Health. 2024;24(1):574. pmid:38760686
  16. 16. Sarvamangala DR, Kulkarni RV. Convolutional neural networks in medical image understanding: a survey. Evol Intell. 2022;15(1):1–22. pmid:33425040
  17. 17. Astuti E, Putra R, Putri D, Ramadhani N, Ahmad T, Putra B. The sensitivity and specificity of YOLO v4 for tooth detection on panoramic radiographs. J Int Dental Med Res. 2023;16(1):442–6.
  18. 18. Selvaraju R, Cogswell M, Das A, Vedantam R, Parikh D, Batra D. Grad-CAM: Visual explanations from deep networks via gradient-based localization. International Journal of Computer Vision. 2020;128(2):336–59.
  19. 19. Ertaş K, Pence I, Cesmeli MS, Ay ZY. Determination of the stage and grade of periodontitis according to the current classification of periodontal and peri-implant diseases and conditions (2018) using machine learning algorithms. J Periodontal Implant Sci. 2023;53(1):38–53. pmid:36468476
  20. 20. Son D-M, Yoon Y-A, Kwon H-J, An C-H, Lee S-H. Automatic detection of mandibular fractures in panoramic radiographs using deep learning. Diagnostics (Basel). 2021;11(6):933. pmid:34067462
  21. 21. Kaya E, Gunec HG, Gokyay SS, Kutal S, Gulum S, Ates HF. Proposing a CNN Method for primary and permanent tooth detection and enumeration on pediatric dental radiographs. J Clin Pediatr Dent. 2022;46(4):293–8. pmid:36099226
  22. 22. Yang H, Jo E, Kim H, Cha I, Jung Y, Nam W. Deep learning for automated detection of cyst and tumors of the jaw in panoramic radiographs. J Clin Med. 2020;9(6).
  23. 23. Celik ME. Deep learning based detection tool for impacted mandibular third molar teeth. Diagnostics (Basel). 2022;12(4):942. pmid:35453990
  24. 24. Hussain M. YOLO-v1 to YOLO-v8, the rise of YOLO and its complementary nature toward digital manufacturing and industrial defect detection. Machines. 2023;11(7):677.
  25. 25. Ju R-Y, Cai W. Fracture detection in pediatric wrist trauma X-ray images using YOLOv8 algorithm. Sci Rep. 2023;13(1):20077. pmid:37973984
  26. 26. Wang C-Y, Bochkovskiy A, Liao H-YM, editors. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition; 2023.
  27. 27. Walsh T. Fuzzy gold standards: Approaches to handling an imperfect reference standard. J Dent. 2018;74(Suppl 1):S47–9.
  28. 28. Casalegno F, Newton T, Daher R, Abdelaziz M, Lodi-Rizzini A, Schürmann F, et al. Caries detection with near-infrared transillumination using deep learning. J Dent Res. 2019;98(11):1227–33. pmid:31449759
  29. 29. Ma T, Zhu J, Wang D, Xu Z, Bai H, Ding P, et al. Deep learning-based detection of irreversible pulpitis in primary molars. Int J Paediatr Dent. 2025;35(1):57–67. pmid:38725105
  30. 30. Wang C-Y, Bochkovskiy A, Liao H-YM, editors. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition; 2023.
  31. 31. Ramezanzade S, Dascalu TL, Ibragimov B, Bakhshandeh A, Bjørndal L. Prediction of pulp exposure before caries excavation using artificial intelligence: Deep learning-based image data versus standard dental radiographs. J Dent. 2023;138:104732. pmid:37778496
  32. 32. Šimundić A-M. Measures of diagnostic accuracy: basic definitions. EJIFCC. 2009;19(4):203–11. pmid:27683318
  33. 33. Shreffler J, Huecker MR. Diagnostic testing accuracy: Sensitivity, specificity, predictive values and likelihood ratios. 2020.