This is an open access journal, and articles are distributed under the terms of the Creative Commons Attribution-NonCommercial-ShareAlike 4.0 License, which allows others to remix, tweak, and build upon the work non-commercially, as long as appropriate credit is given and the new creations are licensed under the identical terms.
This is an open access journal, and articles are distributed under the terms of the Creative Commons Attribution-NonCommercial-ShareAlike 4.0 License, which allows others to remix, tweak, and build upon the work non-commercially, as long as appropriate credit is given and the new creations are licensed under the identical terms.
Dentists begin the diagnosis by identifying and enumerating teeth. Panoramic radiographs are widely used for tooth identification due to their large field of view and low exposure dose. The automatic numbering of teeth in panoramic radiographs can assist clinicians in avoiding errors. Deep learning has emerged as a promising tool for automating tasks. Our goal is to evaluate the accuracy of a two-step deep learning method for tooth identification and enumeration in panoramic radiographs.
In this retrospective observational study, 1007 panoramic radiographs were labeled by three experienced dentists. It involved drawing bounding boxes in two distinct ways: one for teeth and one for quadrants. All images were preprocessed using the contrast-limited adaptive histogram equalization method. First, panoramic images were allocated to a quadrant detection model, and the outputs of this model were provided to the tooth numbering models. A faster region-based convolutional neural network model was used in each step.
Average precision (AP) was calculated in different intersection-over-union thresholds. The AP50 of quadrant detection and tooth enumeration was 100% and 95%, respectively.
We have obtained promising results with a high level of AP using our two-step deep learning framework for automatic tooth enumeration on panoramic radiographs. Further research should be conducted on diverse datasets and real-life situations.
Artificial intelligence (AI) refers to the use of a machine to simulate human intelligence and perform specific tasks, such as recognizing objects, making decisions, and solving problems. Machine learning is a subcategory of AI that uses algorithms to learn data patterns and predict outcomes.
Panoramic radiographs are widely used in dental practice due to their advantages, such as low radiation dose for the patient and ease and speed of production.
Study design
This retrospective observational study was conducted and reported according to the Checklist for AI in Medical Imaging guideline.
Patient selection
One thousand and seven panoramic radiographs were obtained from various sources in Iran and Brazil, which were as follows:
The Department of Oral and Maxillofacial Radiology at Shahid Beheshti University of Medical Sciences, Tehran, Iran: Samples were taken from the Iranian population. The panoramic device was Promax Dimax 3 Digital Pan/Ceph device (Planmeca, Helsinki, Finland). The images were exported to.jpg format with the size 3252 × 1536. A total of 83 images were annotated from this resource. The device setting was 64–66 kilovoltage peak, 4–7 milliampere, and 15–18 s exposure time A private oral and maxillofacial radiology center, Tehran, Iran: Samples were taken from the Iranian population. The panoramic device was Promax Planmeca ProMax (Planmeca, Helsinki, Finland). The device setting was 64–72 kVp, 6.3–12.5 mA, and 13.8–16 s exposure time. The images were exported to.jpg format with the size of 2949 × 1435. A total of 535 images were annotated from this resource UFBA_UESC_DENTAL_IMAGES_DEEP dataset:
All radiographs were anonymized before including in the study. Radiographs with tilted teeth, implants, retained roots, and crowns or bridges were included. Radiographs with low quality, motion artifacts, deciduous teeth, supernumerary teeth, impacted teeth, and edentulous patients were excluded from the study.
Reference data set
Three independent dentists with at least 3 years of clinical experience provide the ground truth by drawing bounding boxes. They held a calibration session and labeled the first 20 images together. The images were labeled in two separate ways: (1) R. R. and S. S. annotated teeth using LabelImg.
Preprocessing
First, we applied contrast-limited adaptive histogram equalization (CLAHE). CLAHE is a histogram-based image enhancement method that limits amplification based on the clipping done in the histogram to restrict it to a predefined level. Moreover, all the images were resized to 224 × 224 before feeding the model.
Data partitions
First, 1007 panoramic images were allocated to a quadrant detection model, and the outputs of this model were provided to the tooth numbering models. Afterward, the tooth numbering models were provided with the 4028 quadrant images that had been generated. In each step (quadrant detection and tooth enumeration), 60% of the data were used for training, 20% for validation, and 20% for testing. To avoid data leakage, all four quadrants of a patient were included in the same set.
Model
Briefly, the first model, named “Quadrant detection model,” splits the image into four separate quadrants. The second model takes a single quadrant, an output of the first model, as an input and finds the tooth enumeration using object detection. We also trained a one-stage tooth enumeration model as a baseline for the comparison.
Quadrant detection model
The first object detector refines the class score of a region to be a quadrant for the object detection task and generates the final bounding box coordinates using faster RCNN using pretrained weights. The ResNet-50 was used as a base CNN for the quadrant detection model. We have tried to evaluate several approaches for quadrant detection task:
The 4-class method: Our first method divided quadrants into four classes, upper lefts, upper rights, lower lefts, and lower rights. We trained end-to-end faster RCNN The 2-class method: In the second method, quadrants are divided into two classes, including the upper and lower quadrants. From quadrant bounding boxes, the right and left quadrants of each upper and lower jaw are determined by a rule-based postprocessing method based on bounding box coordinates The 1-class method: In the third method, all quadrants were named by one class, which means that each of the quadrants was labeled as a single class. Similar to the previous approach, based on bounding box coordinates, each quadrant was determined by a rule-based postprocessing method.
The original images were cropped with an additional margin to ensure that all related teeth are within the cropped images. In the enumeration model, these cropped images become the input. We select the best approach based on the performance of each of the three mentioned methods for the following step.
Enumerating detection model
The enumeration model performs an object detection task, where it identifies and labels individual tooth numbers based on the FDI notation. Each quadrant has four different alignments. To minimize the variance, we used two approaches. First, we flipped the right quadrants from right to left. Then, we divided the tooth enumerating models into two distinct models: one for upper quadrants and the other for lower quadrants. We trained two separate Faster RNNs for each model. Following the enumeration task, quadrants were flipped back into their original alignment.
Training
All model architecture and optimization processes were developed using Python programming language through the Detectron version 2 library.
Evaluation
Various metrics have been used in this study, such as intersection-over-union (IoU), precision, and recall. The overlap between two boundary boxes is measured by the IoU, widely used in object detection. We evaluated our model's performance on unseen data (test set). We computed IoU scores for all pairs of objects using [INSIDE:1] where A is the ground truth and B is the predicted bounding boxes. The IoU threshold t was selected to identify correctly detected objects. True positive (TP) is defined when IoU is more than t, false positive (FP) is defined when IoU is <t, and false negative (FN) is defined when ground truth is present in the image, but the model fails to detect the object.
We computed precision (P), and recall (R), using a fixed IoU threshold t as follows:
[INLINE:1]
The following formula calculates AP:
[INLINE:2]
where Recall iis from 11 points interpolated from the precision-recall curve, which is done by segmenting the recalls evenly into 11 parts: (0.5, 0.55, 0.6,…, 0.95, 1) and Precision (Recall i) is precision in each recall value. Average recall (AR) is formulated as follows:
[INLINE:3]
where t is IoU, which is ∈ {50,100}, and recall (t) is the recall at that point.
Dataset description
The descriptive distribution of each tooth is presented in
Tooth distribution in the total image database according to FDI tooth numbering system. FDI tooth numbering system: 11–18 = Upper right 1–8, 21–28 = Upper left 1–8, 31–38 = Lower left 1–8, 41–48 = Lower right 1–8; 1. Central incisor, 2. Lateral incisor, 3. Canine, 4. First premolar, 5. Second premolar, 6. First molar, 7. Second molar, 8. Third molar.
Quadrant detection model
The quadrant detection model produces promising results with fewer classes in each mouth quadrant. The results show that both the 2-class and 1-class methods can produce 100% AP of bounding boxes with an IoU of more than 50% (AP50)
Outputs of the quadrant detection model without applying any postprocessing steps.
Tooth enumerating model
(a) The output of the lower quadrant detection model, which contains the tooth number and prediction confidence. (b) The output of the upper quadrant detection model, which contains the tooth number and prediction confidence.
In every dental procedure, tooth enumeration and charting are the first steps, which is instrumental in giving the clinician the most accurate treatment plan. The charting process is critical in diagnosing, managing, referrals, and treatment. As dental diseases are either directly associated with teeth or closely located to them, the initial charting serves as the foundation for all subsequent dental procedures.
Several studies have examined the automatic numbering of teeth in panoramic images. However, the results must be interpreted with caution. The AP and predefined threshold have not been reported in Prados-Privado et al.,
Our end-to-end single-step approach produced unsatisfactory results. Therefore, we incorporate a two-step approach, similar to Yüksel et al.'s study.
A significant limitation of our study is the exclusion of radiographs from children with deciduous teeth. Since dentition is rapidly changing, we require large quantities of annotated radiographs at every stage of dental development. Future studies must be conducted in databases of panoramic radiographs in children with mixed dentition, as accurate numbering and charting are vital in mixed dentition, particularly in extraction cases. Furthermore, retained roots, supernumerary teeth, impacted teeth, and implants must be considered in future studies to enhance diversity and generalizability.
We proposed a two-step deep learning-based framework for automatic tooth enumeration on panoramic radiographs. We have obtained promising results with a high level of AP and recall. There is a need for further research on diverse datasets and real-life settings.
Financial support and sponsorship
Nil.
Conflicts of interest
The authors of this manuscript declare that they have no conflicts of interest, real or perceived, financial or non-financial in this article.