?
DRJ Dent Res J Dental Research Journal Dental Research Journal

This is an open access journal, and articles are distributed under the terms of the Creative Commons Attribution-NonCommercial-ShareAlike 4.0 License, which allows others to remix, tweak, and build upon the work non-commercially, as long as appropriate credit is given and the new creations are licensed under the identical terms.

1735-3327 2008-0255 Wolters Kluwer India Pvt. Ltd. India
DRJ-20-116 Original Article Deep learning for tooth identification and enumeration in panoramic radiographs Sadr Soroush Mohammad-Rahimi Hossein Ghorbanimehr Mohammad S Rokhshad Rata Abbasi Zahra Soltani Parisa Moaddabi Amirhossein Shahab Shahriar Rohban Mohammad H Department of Endodontics, School of Dentistry, Hamadan University of Medical Sciences, Hamadan, Iran Topic Group Dental Diagnostics and Digital Dentistry, ITU/WHO Focus Group AI on Health, Berlin, Germany; Department of Computer Engineering, Sharif University of Technology, Tehran, Iran, Department of Computer Science and Software Engineering, Concordia University, Montreal, Canada Topic Group Dental Diagnostics and Digital Dentistry, ITU/WHO Focus Group AI on Health, Berlin, Germany; Department of Medicine, Section of Endocrinology, Nutrition, and Diabetes, Boston University Medical Center, Boston, MA, USA, Department of Oral Health Sciences, Faculty of Dentistry, University of British Columbia, Vancouver, Canada Department of Oral and Maxillofacial Radiology, Dental Implants Research Center, School of Dentistry, Dental Research Institute, Isfahan University of Medical Sciences, Isfahan, Iran Department of Oral and Maxillofacial Surgery, Dental Research Center, School of Dentistry, Mazandaran University of Medical Sciences, Sari, Iran Department of Oral and Maxillofacial Radiology, School of Dentistry, Shahed University of Medical Sciences, Tehran, Iran Department of Computer Engineering, Sharif University of Technology, Tehran, Iran Address for correspondence:Hossein Mohammad-Rahimi, Topic Group Dental Diagnostics and Digital Dentistry, ITU/WHO Focus Group AI on Health, Berlin; Department of Computer Engineering, Sharif University of Technology, Tehran, ramtin.rhm@gmail.com Jan–Dec 2023 20 1 116 116 20 5 2023 Copyright: © Dental Research Journal 2023

This is an open access journal, and articles are distributed under the terms of the Creative Commons Attribution-NonCommercial-ShareAlike 4.0 License, which allows others to remix, tweak, and build upon the work non-commercially, as long as appropriate credit is given and the new creations are licensed under the identical terms.

Background:

Dentists begin the diagnosis by identifying and enumerating teeth. Panoramic radiographs are widely used for tooth identification due to their large field of view and low exposure dose. The automatic numbering of teeth in panoramic radiographs can assist clinicians in avoiding errors. Deep learning has emerged as a promising tool for automating tasks. Our goal is to evaluate the accuracy of a two-step deep learning method for tooth identification and enumeration in panoramic radiographs.

Materials and Methods:

In this retrospective observational study, 1007 panoramic radiographs were labeled by three experienced dentists. It involved drawing bounding boxes in two distinct ways: one for teeth and one for quadrants. All images were preprocessed using the contrast-limited adaptive histogram equalization method. First, panoramic images were allocated to a quadrant detection model, and the outputs of this model were provided to the tooth numbering models. A faster region-based convolutional neural network model was used in each step.

Results:

Average precision (AP) was calculated in different intersection-over-union thresholds. The AP50 of quadrant detection and tooth enumeration was 100% and 95%, respectively.

Conclusion:

We have obtained promising results with a high level of AP using our two-step deep learning framework for automatic tooth enumeration on panoramic radiographs. Further research should be conducted on diverse datasets and real-life situations.

Deep learning panoramic radiography tooth identification tooth numbering
</sec> <sec> <title>Introduction

Artificial intelligence (AI) refers to the use of a machine to simulate human intelligence and perform specific tasks, such as recognizing objects, making decisions, and solving problems. Machine learning is a subcategory of AI that uses algorithms to learn data patterns and predict outcomes. 1 Deep learning, a category of machine learning models, has recently gained interest due to increasing data, computing power availability, and superior performance compared with conventional machine learning models. 2 Deep learning refers to deep (multilayered) neural networks. A neural network consists of several artificial neurons and connections with mathematical operations inspired by human neurons. It can automatically learn data patterns without explicit direction when given a large amount of data. Convolutional neural networks (CNNs) are introduced for processing complex images. They use mathematical convolution functions, which allow them to detect local connectivity patterns such as edges and corners in images. 3 CNNs have been studied in maxillofacial imaging for automated diagnosing and treatment planning. They are used primarily to perform tasks such as semantic segmentation (e.g. segmenting of all teeth as tooth), 4 instance segmentation (e.g. segmentation of all teeth as each individual tooth), and object detection. 3 , 5 Object detection identifies the locations of objects in an image with rectangular bounding boxes and classifies them into defined groups. 6 Several CNN architectures have been proposed for this task, including region-based CNNs (R-CNNs) and You Only Look Once.

Panoramic radiographs are widely used in dental practice due to their advantages, such as low radiation dose for the patient and ease and speed of production. 7 They provide a two-dimensional (2D) image that typically includes all the present teeth of both jaws and their supporting structures. They provide valuable information about the patient's dental state that can be used for charting, screening, treatment planning, and forensic investigation. 8 , 9 To analyze panoramic radiographs, several steps need to be followed, starting with tooth identification. 6 Automated tooth identification and enumeration is the first step toward achieving a fully automatic diagnosis and treatment plan. This task can be carried out using CNN architectures. Object detection was applied for automatic tooth enumeration in periapical, 10 , 11 bitewing, 12 and panoramic radiographs. 6 , 9 , 13 , 14 , 15 , 16 However, there are two main challenges in training deep learning models on panoramic radiographs: (1) there are up to 32 tooth classes in a single image which makes it difficult for the model to be trained on them and (2) the image contains many structures other than teeth such as cervical vertebrae, nasal spine, maxillary sinuses, and mandibular condyles which results in too much unnecessary data for deep learning models. We propose a two-step method to overcome these issues: first, we trained the model to detect the quadrants, and then, we trained the model to detect the teeth within each quadrant. We hypothesize that this method will enhance the average precision (AP) and recall of the model compared to previous studies.

Materials and Methods

Study design

This retrospective observational study was conducted and reported according to the Checklist for AI in Medical Imaging guideline. 17 First, we trained a model for detecting quadrants automatically. Then, we trained models to detect and classify teeth from one to eight in the upper and lower quadrants.

Patient selection

One thousand and seven panoramic radiographs were obtained from various sources in Iran and Brazil, which were as follows:

The Department of Oral and Maxillofacial Radiology at Shahid Beheshti University of Medical Sciences, Tehran, Iran: Samples were taken from the Iranian population. The panoramic device was Promax Dimax 3 Digital Pan/Ceph device (Planmeca, Helsinki, Finland). The images were exported to.jpg format with the size 3252 × 1536. A total of 83 images were annotated from this resource. The device setting was 64–66 kilovoltage peak, 4–7 milliampere, and 15–18 s exposure time

A private oral and maxillofacial radiology center, Tehran, Iran: Samples were taken from the Iranian population. The panoramic device was Promax Planmeca ProMax (Planmeca, Helsinki, Finland). The device setting was 64–72 kVp, 6.3–12.5 mA, and 13.8–16 s exposure time. The images were exported to.jpg format with the size of 2949 × 1435. A total of 535 images were annotated from this resource

UFBA_UESC_DENTAL_IMAGES_DEEP dataset: 18 Data were acquired from a GitHub repository ( https://github.com/IvisionLab/deep-dental-image). Samples were taken from the Brazilian population. The images were in.jpg format with the size of 1991 × 1127. A total of 389 images were annotated from this resource.

All radiographs were anonymized before including in the study. Radiographs with tilted teeth, implants, retained roots, and crowns or bridges were included. Radiographs with low quality, motion artifacts, deciduous teeth, supernumerary teeth, impacted teeth, and edentulous patients were excluded from the study.

Reference data set

Three independent dentists with at least 3 years of clinical experience provide the ground truth by drawing bounding boxes. They held a calibration session and labeled the first 20 images together. The images were labeled in two separate ways: (1) R. R. and S. S. annotated teeth using LabelImg. 19 (2) R. R and Z. A. labeled the quadrants. Labeling was double-checked by H. M. R. In case of any conflict or ambiguity in tooth numbers, the sample was excluded. The teeth labeling process involved drawing a rectangular bounding box around the outer edges of each tooth and classifying it according to the two-digit FDI tooth numbering system. For quadrant labeling, in the maxillary and mandibular edentulous quadrants, the midline was taken as the reference line. For the maxilla, the coronoid processes of the mandible are the superior-posterior reference point. At the same time, in the inferior-posterior area, the upper one-third of the retromolar pad serves as the reference point. On the mandible, the superior-posterior reference point was the upper two-thirds of the retromolar pad, and the lower-posterior reference point was the line passing under the inferior mandibular canal. In quadrants with teeth, all teeth with their roots were covered. On the maxilla, the line connecting the palatal roots of the molars and the canine roots was considered the upper reference line. The inferior line was the line under the incisal edges of the central and canine. On the mandible, the superior reference line is the line above the molar's cusp tips, and the inferior reference line is the line under the lowest apex of mandibular roots.

Preprocessing

First, we applied contrast-limited adaptive histogram equalization (CLAHE). CLAHE is a histogram-based image enhancement method that limits amplification based on the clipping done in the histogram to restrict it to a predefined level. Moreover, all the images were resized to 224 × 224 before feeding the model.

Data partitions

First, 1007 panoramic images were allocated to a quadrant detection model, and the outputs of this model were provided to the tooth numbering models. Afterward, the tooth numbering models were provided with the 4028 quadrant images that had been generated. In each step (quadrant detection and tooth enumeration), 60% of the data were used for training, 20% for validation, and 20% for testing. To avoid data leakage, all four quadrants of a patient were included in the same set.

Model

Briefly, the first model, named “Quadrant detection model,” splits the image into four separate quadrants. The second model takes a single quadrant, an output of the first model, as an input and finds the tooth enumeration using object detection. We also trained a one-stage tooth enumeration model as a baseline for the comparison.

Quadrant detection model

The first object detector refines the class score of a region to be a quadrant for the object detection task and generates the final bounding box coordinates using faster RCNN using pretrained weights. The ResNet-50 was used as a base CNN for the quadrant detection model. We have tried to evaluate several approaches for quadrant detection task:

The 4-class method: Our first method divided quadrants into four classes, upper lefts, upper rights, lower lefts, and lower rights. We trained end-to-end faster RCNN

The 2-class method: In the second method, quadrants are divided into two classes, including the upper and lower quadrants. From quadrant bounding boxes, the right and left quadrants of each upper and lower jaw are determined by a rule-based postprocessing method based on bounding box coordinates

The 1-class method: In the third method, all quadrants were named by one class, which means that each of the quadrants was labeled as a single class. Similar to the previous approach, based on bounding box coordinates, each quadrant was determined by a rule-based postprocessing method.

The original images were cropped with an additional margin to ensure that all related teeth are within the cropped images. In the enumeration model, these cropped images become the input. We select the best approach based on the performance of each of the three mentioned methods for the following step.

Enumerating detection model

The enumeration model performs an object detection task, where it identifies and labels individual tooth numbers based on the FDI notation. Each quadrant has four different alignments. To minimize the variance, we used two approaches. First, we flipped the right quadrants from right to left. Then, we divided the tooth enumerating models into two distinct models: one for upper quadrants and the other for lower quadrants. We trained two separate Faster RNNs for each model. Following the enumeration task, quadrants were flipped back into their original alignment.

Training

All model architecture and optimization processes were developed using Python programming language through the Detectron version 2 library. 20 The training procedure was performed on an NVIDIA Tesla K80 with 12GB of GDDR5 VRAM, Intel Xeon Processor with two cores at 2.20 GHz, and 13 GB RAM. The initial learning rate for this model was chosen as 0.001 with further exponential decay. A batch size of 128 was used to train two enumerating networks. Several models with different methods were trained, and both quadrant and tooth detection models were trained for 700 iterations. The Grid search was used for hyperparameter tuning batch size, learning rate, and optimizer. Early stopping allowed the best model weights to be saved based on their performance of AP on a validation dataset, thereby reducing the possibility of overfitting.

Evaluation

Various metrics have been used in this study, such as intersection-over-union (IoU), precision, and recall. The overlap between two boundary boxes is measured by the IoU, widely used in object detection. We evaluated our model's performance on unseen data (test set). We computed IoU scores for all pairs of objects using [INSIDE:1] where A is the ground truth and B is the predicted bounding boxes. The IoU threshold t was selected to identify correctly detected objects. True positive (TP) is defined when IoU is more than t, false positive (FP) is defined when IoU is <t, and false negative (FN) is defined when ground truth is present in the image, but the model fails to detect the object.

We computed precision (P), and recall (R), using a fixed IoU threshold t as follows:

[INLINE:1]

The following formula calculates AP:

[INLINE:2]

where Recall iis from 11 points interpolated from the precision-recall curve, which is done by segmenting the recalls evenly into 11 parts: (0.5, 0.55, 0.6,…, 0.95, 1) and Precision (Recall i) is precision in each recall value. Average recall (AR) is formulated as follows:

[INLINE:3]

where t is IoU, which is ∈ {50,100}, and recall (t) is the recall at that point.

Results

Dataset description

The descriptive distribution of each tooth is presented in Figure 1. Mandibular incisors and canines are the most prevalent teeth in images while maxillary third molars were the least frequently observed.

Tooth distribution in the total image database according to FDI tooth numbering system. FDI tooth numbering system: 11–18 = Upper right 1–8, 21–28 = Upper left 1–8, 31–38 = Lower left 1–8, 41–48 = Lower right 1–8; 1. Central incisor, 2. Lateral incisor, 3. Canine, 4. First premolar, 5. Second premolar, 6. First molar, 7. Second molar, 8. Third molar.

Figure 1

Quadrant detection model

The quadrant detection model produces promising results with fewer classes in each mouth quadrant. The results show that both the 2-class and 1-class methods can produce 100% AP of bounding boxes with an IoU of more than 50% (AP50) Table 1. Therefore, the quadrant detection model can be successfully used to generate the subsequent models' training data. An example of an output of the quadrant detection model without applying any postprocessing steps can be seen in Figure 2.

Outputs of the quadrant detection model without applying any postprocessing steps.

Figure 2 {Table 1}

Tooth enumerating model

Figure 3a and Figure 3b shows examples of the outputs of the enumeration model. The output label format contains the tooth number and prediction confidence. The AP of upper and lower enumerations was 95.93 and 95.05, respectively. The AP (50:95), AR (50:95), and AP75 for the enumeration of teeth in quadrants are presented in Table 2. Table 3shows the AP for each tooth class. The lower first molars and lower lateral incisors have the highest and lowest APs, respectively. The results for the end-to-end approach are presented in Table 4for comparison.

(a) The output of the lower quadrant detection model, which contains the tooth number and prediction confidence. (b) The output of the upper quadrant detection model, which contains the tooth number and prediction confidence.

Figure 3 {Table 2}{Table 3}{Table 4}

Discussion

In every dental procedure, tooth enumeration and charting are the first steps, which is instrumental in giving the clinician the most accurate treatment plan. The charting process is critical in diagnosing, managing, referrals, and treatment. As dental diseases are either directly associated with teeth or closely located to them, the initial charting serves as the foundation for all subsequent dental procedures. 5 Consequently, tooth identification and numbering hold significant importance as they form the basis for more intricate AI-based tasks in dental radiographic images. Using digital images to identify teeth automatically is a crucial component of intelligent health care. 9 , 10 , 16 , 21 Studies have been conducted on AI for charting purposes on cone-beam computed tomographies, bitewing radiographs, and periapical radiographs. 11 , 12 , 22 , 23 Nevertheless, a panoramic radiograph is the most suitable technique for charting. It provides an overview of the entire dentition in a single image with minimal radiation dose. 7

Several studies have examined the automatic numbering of teeth in panoramic images. However, the results must be interpreted with caution. The AP and predefined threshold have not been reported in Prados-Privado et al., 14 Estai et al., 24 Bilgir et al., 16 Tuzoff et al., 9 and Muramatsu et al. 13 studies. AP is a metric commonly used in object detection literature derived from precision and recall. As the data are in 2D space, a predefined threshold of IoUs of bounding boxes is used to define the model's true and false predictions. Therefore, AP is reported in various thresholds, mostly 0.5 and 0.75. Tuzoff et al. relied on expert opinion about the judgment of the model's prediction, which is subject to individual bias. 9 Moreover, some studies simplified tooth identification into 4-category (incisor, canine, premolar, and molar) and 3-category (incisor, canine, molar) classifications. 6 , 13

Our end-to-end single-step approach produced unsatisfactory results. Therefore, we incorporate a two-step approach, similar to Yüksel et al.'s study. 15 The difference is they used segmentation for quadrant detection and we used object detection task in both steps. Yüksel et al. reported an AP of 89.4% for the enumeration task. 15 Chung et al. used one-step point-wise localization and distant regularization, instead of the anchor-based method. 25 The 32 tooth boxes were annotated regardless of whether the tooth was present in the image. The detection was achieved with a multitasking, class-agnostic identification neural network that incorporated parallel training for center offsets. They reached an AP of 91%. We obtained a 95% AP score in tooth enumeration, which is an improvement over previous studies.

A significant limitation of our study is the exclusion of radiographs from children with deciduous teeth. Since dentition is rapidly changing, we require large quantities of annotated radiographs at every stage of dental development. Future studies must be conducted in databases of panoramic radiographs in children with mixed dentition, as accurate numbering and charting are vital in mixed dentition, particularly in extraction cases. Furthermore, retained roots, supernumerary teeth, impacted teeth, and implants must be considered in future studies to enhance diversity and generalizability.

Conclusion

We proposed a two-step deep learning-based framework for automatic tooth enumeration on panoramic radiographs. We have obtained promising results with a high level of AP and recall. There is a need for further research on diverse datasets and real-life settings.

Financial support and sponsorship

Nil.

Conflicts of interest

The authors of this manuscript declare that they have no conflicts of interest, real or perceived, financial or non-financial in this article.

Schwendicke F Samek W Krois J Artificial intelligence in dentistry: Chances and challenges J Dent Res 769 74 LeCun Y Bengio Y Hinton G Deep learning Nature 436 44 Yamashita R Nishio M Do RK Togashi K Convolutional neural networks: An overview and application in radiology Insights Imaging 611 29 Cantu AG Gehrung S Krois J Chaurasia A Rossi JG Gaudin R Detecting caries lesions of different radiographic extension on bitewings using deep learning J Dent 103425 Umer F Habib S Adnan N Application of deep learning in teeth identification tasks on panoramic radiographs Dentomaxillofac Radiol 20210504 Kim C Kim D Jeong H Yoon SJ Youm S Automatic tooth detection and numbering using a combination of a CNN and heuristic algorithm Appl Sci 5624 Terlemez A Tassoker M Kizilcakaya M Gulec M Comparison of cone-beam computed tomography and panoramic radiography in the evaluation of maxillary sinus pathology related to maxillary posterior teeth: Do apical lesions increase the risk of maxillary sinus pathology? Imaging Sci Dent 115 22 Heinrich A Güttler F Wendt S Schenkl S Hubig M Wagner R Forensic Odontology: Automatic identification of persons comparing antemortem and postmortem panoramic radiographs using computer vision Rofo 1152 8 Tuzoff DV Tuzova LN Bornstein MM Krasnov AS Kharchenko MA Nikolenko SI Tooth detection and numbering in panoramic radiographs using convolutional neural networks Dentomaxillofac Radiol 20180051 Chen H Zhang K Lyu P Li H Zhang L Wu J A deep learning approach to automatic teeth detection and numbering based on object detection in dental periapical films Sci Rep 3840 Görürgöz C Orhan K Bayrakdar IS Çelik Ö Bilgir E Odabaş A Performance of a convolutional neural network algorithm for tooth detection and numbering on periapical radiographs Dentomaxillofac Radiol 20210246 Yasa Y Çelik Ö Bayrakdar IS Pekince A Orhan K Akarsu S An artificial intelligence proposal to automatic teeth detection and numbering in dental bite-wing radiographs Acta Odontol Scand 275 81 Muramatsu C Morishita T Takahashi R Hayashi T Nishiyama W Ariji Y Tooth detection and classification on panoramic radiographs for automatic dental chart filing: Improved classification by multi-sized input data Oral Radiol 13 9 Prados-Privado M García Villalón J Blázquez Torres A Martínez-Martínez CH Ivorra C A convolutional neural network for automatic tooth numbering in panoramic images Biomed Res Int 3625386 Yüksel AE Gültekin S Simsar E Özdemir ŞD Gündoğar M Tokgöz SB Dental enumeration and multiple treatment detection on panoramic X-rays using deep learning Sci Rep 12342 Bilgir E Bayrakdar İŞ Çelik Ö Orhan K Akkoca F Sağlam H An artifıcial ıntelligence approach to automatic tooth detection and numbering in panoramic radiographs BMC Med Imaging 124 Mongan J Moy L Kahn CE Jr Checklist for artificial intelligence in medical imaging (CLAIM): A guide for authors and reviewers Radiol Artif Intell e200029 Jader G Fontineli J Ruiz M Abdalla K Pithon M Oliveira L editors Deep Instance Segmentation of Teeth in Panoramic X-Ray Images.Parana, Brazil: SIBGRAPI; 2018 p Deep Instance Segmentation of Teeth in Panoramic X-Ray Images Parana, Brazil: SIBGRAPI; 2018 p 400-7 Tzutalin: LabelImg com/tzutalin/labelImg Available from 2022 10] Wu Y Kirillov A Massa F Lo WY Girshick R Detectron 2; 2019 [Last accessed on Jan ]. Detectron 2; 2019 Available from: https://githubcom/facebookresearch/detectron2 [Last accessed on 2023 Jan 10] Yuniarti A Nugroho A Amaliah B Arifin AZ Classification and numbering of dental radiographs for an automated human identification system TELKOMNIKA 137 46 Hosntalab M Aghaeizadeh Zoroofi R Abbaspour Tehrani-Fard A Shirani G Classification and numbering of teeth in multi-slice CT images using wavelet-Fourier descriptor Int J Comput Assist Radiol Surg 237 49 Yaren Tekin B Ozcan C Pekince A Yasa Y An enhanced tooth segmentation and numbering according to FDI notation in bitewing radiographs Comput Biol Med 105547 Estai M Tennant M Gebauer D Brostek A Vignarajan J Mehdizadeh M Deep learning for automated detection and numbering of permanent teeth on panoramic images Dentomaxillofac Radiol 20210296 Chung M Lee J Park S Lee M Lee CE Lee J Individual tooth detection and identification from dental panoramic X-ray images via point-wise localization and distance regularization Artif Intell Med 101996

Refbacks

  • There are currently no refbacks.