Research Article
BibTex RIS Cite

Comparison of VT-based and CNN-based Models on Teeth Segmentation

Year 2025, Volume: 13 Issue: 2, 148 - 156
https://doi.org/10.17694/bajece.1457754

Abstract

Semantic segmentation is a crucial task in computer vision with a wide array of applications across various fields, especially in medical imaging. One of the most important applications of semantic segmentation is in the field of dentistry, where teeth segmentation plays a significant role in diagnosing and treating oral health issues. Accurate segmentation of teeth in dental images is vital for detecting abnormalities, planning treatments, and monitoring the progress of dental procedures. In this paper, a comprehensive comparative analysis is presented, focusing on the use of Convolutional Neural Network (CNN)-based and Vision Transformer (VT)-based models for image segmentation within the context of dentistry. The paper presents a comparison of eight different models, contributing to the literature on dental image segmentation and showcasing practical applications in clinical dental settings. The research presented in this study uses several state-of-the-art segmentation models, namely U-Net, LinkNet, and Swin U-Net, along with different backbones to perform teeth segmentation on publicly available two datasets: one representing adults and the other children. The experiments were conducted to determine which models and backbones provided the best segmentation performance for each dataset. The study also emphasizes that the segmentation modeling process should be handled separately since the alignment of child and adult teeth is different. The U-Net model with the ResNet101 backbone achieved the best performance on the adults dataset, while for the children dataset, the U-Net model with the same ResNet101 backbone also demonstrated superior results. The highest Dice scores obtained were 0.9543 for the adults dataset and 0.9019 for the children dataset, indicating the effectiveness of these models in accurately segmenting teeth. The findings from this research demonstrate the potential of deep learning techniques in improving the accuracy and efficiency of dental diagnosis and treatment planning. Codes used throughout the study will be publicly available at https://github.com/FidanVural/Teeth-Segmentation-in- Panoramic-Radiography/tree/main

Thanks

TÜBİTAK-BİLGEM

References

  • [1] Y. Zhao et al., “TSASNet: Tooth segmentation on dental panoramic X-ray images by Two-Stage Attention Segmentation Network,” Knowl Based Syst, vol. 206, p. 106338, 2020.
  • [2] A. Haghanifar, M. M. Majdabadi, S. Haghanifar, Y. Choi, and S.-B. Ko, “PaXNet: Tooth segmentation and dental caries detection in panoramic X-ray using ensemble transfer learning and capsule classifier,” Multimed Tools Appl, vol. 82, no. 18, pp. 27659–27679, 2023.
  • [3] Y. Ariji et al., “Automatic detection and classification of radiolucent lesions in the mandible on panoramic radiographs using a deep learning object detection technique,” Oral Surg Oral Med Oral Pathol Oral Radiol, vol. 128, no. 4, pp. 424–430, 2019.
  • [4] M. Fukuda et al., “Evaluation of an artificial intelligence system for detecting vertical root fracture on panoramic radiography,” Oral Radiol, vol. 36, pp. 337–343, 2020.
  • [5] B. C. Uzun Saylan et al., “Assessing the effectiveness of artificial intelligence models for detecting alveolar bone loss in periodontal disease: a panoramic radiograph study,” Diagnostics, vol. 13, no. 10, p. 1800, 2023.
  • [6] L. Schneider et al., “Federated vs local vs central deep learning of tooth segmentation on panoramic radiographs,” J Dent, vol. 135, p. 104556, 2023.
  • [7] S. Park et al., “Deep learning-based automatic segmentation of mandible and maxilla in multi-center ct images,” Applied Sciences, vol. 12, no. 3, p. 1358, 2022.
  • [8] N. Kumbasar, M. T. Güller, Ö. Miloğlu, E. A. Oral, and I. Y. Ozbek, “Deep-learning based fusion of spatial relationship classification between mandibular third molar and inferior alveolar nerve using panoramic radiograph images,” Biomed Signal Process Control, vol. 100, p. 107059, 2025.
  • [9] M. T. Güller, N. Kumbasar, and Ö. Miloğlu, “Evaluation of the effectiveness of panoramic radiography in impacted mandibular third molars on deep learning models developed with findings obtained with cone beam computed tomography,” Oral Radiol, pp. 1–16, 2024.
  • [10] A. B. Oktay, Z. Akhtar, and A. Gurses, “Dental biometric systems: a comparative study of conventional descriptors and deep learning-based features,” Multimed Tools Appl, vol. 81, no. 20, pp. 28183–28206, 2022.
  • [11] Ö. Miloğlu, N. Kumbasar, Z. T. Tosun, M. T. Güller, and \.Ibrahim Yücel Özbek, “Gender Classification With Hand-Wrist Radiographs Using the Deep Learning Method,” Current Research in Dental Sciences, vol. 35, no. 1, pp. 2–7, 2025.
  • [12] C.-W. Wang et al., “A benchmark for comparison of dental radiography analysis algorithms,” Med Image Anal, vol. 31, pp. 63–76, 2016.
  • [13] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in Medical image computing and computer-assisted intervention–MICCAI 2015: 18th international conference, Munich, Germany, October 5-9, 2015, proceedings, part III 18, 2015, pp. 234–241.
  • [14] J. Long, E. Shelhamer, and T. Darrell, “Fully convolutional networks for semantic segmentation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 3431–3440.
  • [15] A. Chaurasia and E. Culurciello, “Linknet: Exploiting encoder representations for efficient semantic segmentation,” in 2017 IEEE visual communications and image processing (VCIP), 2017, pp. 1–4.
  • [16] A. Vaswani, “Attention is all you need,” Adv Neural Inf Process Syst, 2017.
  • [17] A. Dosovitskiy, “An image is worth 16x16 words: Transformers for image recognition at scale,” arXiv preprint arXiv:2010.11929, 2020.
  • [18] Z. Liu et al., “Swin transformer: Hierarchical vision transformer using shifted windows,” in Proceedings of the IEEE/CVF international conference on computer vision, 2021, pp. 10012–10022.
  • [19] J. Chen et al., “Transunet: Transformers make strong encoders for medical image segmentation,” arXiv preprint arXiv:2102.04306, 2021.
  • [20] J. Wu, “Promptunet: Toward interactive medical image segmentation,” arXiv preprint arXiv:2305.10300, vol. 2, 2023.
  • [21] G. Silva, L. Oliveira, and M. Pithon, “Automatic segmenting teeth in X-ray images: Trends, a novel data set, benchmarking and future perspectives,” Expert Syst Appl, vol. 107, pp. 15–31, 2018.
  • [22] T. L. Koch, M. Perslev, C. Igel, and S. S. Brandt, “Accurate segmentation of dental panoramic radiographs with U-Nets,” in 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019), 2019, pp. 15–19.
  • [23] S. Sivagami, P. Chitra, G. S. R. Kailash, and S. R. Muralidharan, “Unet architecture based dental panoramic image segmentation,” in 2020 International Conference on Wireless Communications Signal Processing and Networking (WiSPNET), 2020, pp. 187–191.
  • [24] G. Jader, J. Fontineli, M. Ruiz, K. Abdalla, M. Pithon, and L. Oliveira, “Deep instance segmentation of teeth in panoramic X-ray images,” in 2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI), 2018, pp. 400–407.
  • [25] A. Wirtz, S. G. Mirashi, and S. Wesarg, “Automatic teeth segmentation in panoramic X-ray images using a coupled shape model in combination with a neural network,” in Medical Image Computing and Computer Assisted Intervention–MICCAI 2018: 21st International Conference, Granada, Spain, September 16-20, 2018, Proceedings, Part IV 11, 2018, pp. 712–719.
  • [26] J.-H. Lee, S.-S. Han, Y. H. Kim, C. Lee, and I. Kim, “Application of a fully deep convolutional neural network to the automation of tooth segmentation on panoramic radiographs,” Oral Surg Oral Med Oral Pathol Oral Radiol, vol. 129, no. 6, pp. 635–642, 2020.
  • [27] S. Zhao, Q. Luo, and C. Liu, “Automatic tooth segmentation and classification in dental panoramic X-ray images,” 2020.
  • [28] B. Silva, L. Pinheiro, L. Oliveira, and M. Pithon, “A study on tooth segmentation and numbering using end-to-end deep neural networks,” in 2020 33rd SIBGRAPI conference on graphics, patterns and images (SIBGRAPI), 2020, pp. 164–171.
  • [29] C. Sheng et al., “Transformer-based deep learning network for tooth segmentation on panoramic radiographs,” J Syst Sci Complex, vol. 36, no. 1, pp. 257–272, 2023.
  • [30] S. Arora, S. K. Tripathy, R. Gupta, and R. Srivastava, “Exploiting multimodal CNN architecture for automated teeth segmentation on dental panoramic X-ray images,” Proc Inst Mech Eng H, vol. 237, no. 3, pp. 395–405, 2023.
  • [31] M. Kanwal, M. M. Ur Rehman, M. U. Farooq, and D.-K. Chae, “Mask-transformer-based networks for teeth segmentation in panoramic radiographs,” Bioengineering, vol. 10, no. 7, p. 843, 2023.
  • [32] M. K. Dhar, M. Deb, D. Madhab, and Z. Yu, “A Deep Learning Approach to Teeth Segmentation and Orientation from Panoramic X-rays,” arXiv preprint arXiv:2310.17176, 2023.
  • [33] A. Ghafoor, S.-Y. Moon, and B. Lee, “Multiclass Segmentation Using Teeth Attention Modules for Dental X-Ray Images,” IEEE Access, vol. 11, pp. 123891–123903, 2023.
  • [34] Y. Zhang et al., “Children’s dental panoramic radiographs dataset for caries segmentation and dental disease detection,” Sci Data, vol. 10, no. 1, p. 380, 2023.
  • [35] W. Brahmi and I. Jdey, “Automatic tooth instance segmentation and identification from panoramic X-Ray images using deep CNN,” Multimed Tools Appl, vol. 83, no. 18, pp. 55565–55585, 2024.
  • [36] E. Asci et al., “A Deep Learning Approach to Automatic Tooth Caries Segmentation in Panoramic Radiographs of Children in Primary Dentition, Mixed Dentition, and Permanent Dentition,” Children, vol. 11, no. 6, p. 690, 2024.
  • [37] S. Wathore and S. Gorthi, “Bilateral symmetry-based augmentation method for improved tooth segmentation in panoramic X-rays,” Pattern Recognit Lett, vol. 188, pp. 1–7, 2025.
  • [38] G. Altan and A. Al Samar, “Tooth segmentation on dental panoramic X-rays using Mask R-CNN,” in Mining Biomedical Text, Images and Visual Features for Information Retrieval, Elsevier, 2025, pp. 481–498.
  • [39] S. Hou, T. Zhou, Y. Liu, P. Dang, H. Lu, and H. Shi, “Teeth U-Net: A segmentation model of dental panoramic X-ray images for context semantics and contrast enhancement,” Comput Biol Med, vol. 152, p. 106296, 2023.
Year 2025, Volume: 13 Issue: 2, 148 - 156
https://doi.org/10.17694/bajece.1457754

Abstract

References

  • [1] Y. Zhao et al., “TSASNet: Tooth segmentation on dental panoramic X-ray images by Two-Stage Attention Segmentation Network,” Knowl Based Syst, vol. 206, p. 106338, 2020.
  • [2] A. Haghanifar, M. M. Majdabadi, S. Haghanifar, Y. Choi, and S.-B. Ko, “PaXNet: Tooth segmentation and dental caries detection in panoramic X-ray using ensemble transfer learning and capsule classifier,” Multimed Tools Appl, vol. 82, no. 18, pp. 27659–27679, 2023.
  • [3] Y. Ariji et al., “Automatic detection and classification of radiolucent lesions in the mandible on panoramic radiographs using a deep learning object detection technique,” Oral Surg Oral Med Oral Pathol Oral Radiol, vol. 128, no. 4, pp. 424–430, 2019.
  • [4] M. Fukuda et al., “Evaluation of an artificial intelligence system for detecting vertical root fracture on panoramic radiography,” Oral Radiol, vol. 36, pp. 337–343, 2020.
  • [5] B. C. Uzun Saylan et al., “Assessing the effectiveness of artificial intelligence models for detecting alveolar bone loss in periodontal disease: a panoramic radiograph study,” Diagnostics, vol. 13, no. 10, p. 1800, 2023.
  • [6] L. Schneider et al., “Federated vs local vs central deep learning of tooth segmentation on panoramic radiographs,” J Dent, vol. 135, p. 104556, 2023.
  • [7] S. Park et al., “Deep learning-based automatic segmentation of mandible and maxilla in multi-center ct images,” Applied Sciences, vol. 12, no. 3, p. 1358, 2022.
  • [8] N. Kumbasar, M. T. Güller, Ö. Miloğlu, E. A. Oral, and I. Y. Ozbek, “Deep-learning based fusion of spatial relationship classification between mandibular third molar and inferior alveolar nerve using panoramic radiograph images,” Biomed Signal Process Control, vol. 100, p. 107059, 2025.
  • [9] M. T. Güller, N. Kumbasar, and Ö. Miloğlu, “Evaluation of the effectiveness of panoramic radiography in impacted mandibular third molars on deep learning models developed with findings obtained with cone beam computed tomography,” Oral Radiol, pp. 1–16, 2024.
  • [10] A. B. Oktay, Z. Akhtar, and A. Gurses, “Dental biometric systems: a comparative study of conventional descriptors and deep learning-based features,” Multimed Tools Appl, vol. 81, no. 20, pp. 28183–28206, 2022.
  • [11] Ö. Miloğlu, N. Kumbasar, Z. T. Tosun, M. T. Güller, and \.Ibrahim Yücel Özbek, “Gender Classification With Hand-Wrist Radiographs Using the Deep Learning Method,” Current Research in Dental Sciences, vol. 35, no. 1, pp. 2–7, 2025.
  • [12] C.-W. Wang et al., “A benchmark for comparison of dental radiography analysis algorithms,” Med Image Anal, vol. 31, pp. 63–76, 2016.
  • [13] O. Ronneberger, P. Fischer, and T. Brox, “U-net: Convolutional networks for biomedical image segmentation,” in Medical image computing and computer-assisted intervention–MICCAI 2015: 18th international conference, Munich, Germany, October 5-9, 2015, proceedings, part III 18, 2015, pp. 234–241.
  • [14] J. Long, E. Shelhamer, and T. Darrell, “Fully convolutional networks for semantic segmentation,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2015, pp. 3431–3440.
  • [15] A. Chaurasia and E. Culurciello, “Linknet: Exploiting encoder representations for efficient semantic segmentation,” in 2017 IEEE visual communications and image processing (VCIP), 2017, pp. 1–4.
  • [16] A. Vaswani, “Attention is all you need,” Adv Neural Inf Process Syst, 2017.
  • [17] A. Dosovitskiy, “An image is worth 16x16 words: Transformers for image recognition at scale,” arXiv preprint arXiv:2010.11929, 2020.
  • [18] Z. Liu et al., “Swin transformer: Hierarchical vision transformer using shifted windows,” in Proceedings of the IEEE/CVF international conference on computer vision, 2021, pp. 10012–10022.
  • [19] J. Chen et al., “Transunet: Transformers make strong encoders for medical image segmentation,” arXiv preprint arXiv:2102.04306, 2021.
  • [20] J. Wu, “Promptunet: Toward interactive medical image segmentation,” arXiv preprint arXiv:2305.10300, vol. 2, 2023.
  • [21] G. Silva, L. Oliveira, and M. Pithon, “Automatic segmenting teeth in X-ray images: Trends, a novel data set, benchmarking and future perspectives,” Expert Syst Appl, vol. 107, pp. 15–31, 2018.
  • [22] T. L. Koch, M. Perslev, C. Igel, and S. S. Brandt, “Accurate segmentation of dental panoramic radiographs with U-Nets,” in 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019), 2019, pp. 15–19.
  • [23] S. Sivagami, P. Chitra, G. S. R. Kailash, and S. R. Muralidharan, “Unet architecture based dental panoramic image segmentation,” in 2020 International Conference on Wireless Communications Signal Processing and Networking (WiSPNET), 2020, pp. 187–191.
  • [24] G. Jader, J. Fontineli, M. Ruiz, K. Abdalla, M. Pithon, and L. Oliveira, “Deep instance segmentation of teeth in panoramic X-ray images,” in 2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI), 2018, pp. 400–407.
  • [25] A. Wirtz, S. G. Mirashi, and S. Wesarg, “Automatic teeth segmentation in panoramic X-ray images using a coupled shape model in combination with a neural network,” in Medical Image Computing and Computer Assisted Intervention–MICCAI 2018: 21st International Conference, Granada, Spain, September 16-20, 2018, Proceedings, Part IV 11, 2018, pp. 712–719.
  • [26] J.-H. Lee, S.-S. Han, Y. H. Kim, C. Lee, and I. Kim, “Application of a fully deep convolutional neural network to the automation of tooth segmentation on panoramic radiographs,” Oral Surg Oral Med Oral Pathol Oral Radiol, vol. 129, no. 6, pp. 635–642, 2020.
  • [27] S. Zhao, Q. Luo, and C. Liu, “Automatic tooth segmentation and classification in dental panoramic X-ray images,” 2020.
  • [28] B. Silva, L. Pinheiro, L. Oliveira, and M. Pithon, “A study on tooth segmentation and numbering using end-to-end deep neural networks,” in 2020 33rd SIBGRAPI conference on graphics, patterns and images (SIBGRAPI), 2020, pp. 164–171.
  • [29] C. Sheng et al., “Transformer-based deep learning network for tooth segmentation on panoramic radiographs,” J Syst Sci Complex, vol. 36, no. 1, pp. 257–272, 2023.
  • [30] S. Arora, S. K. Tripathy, R. Gupta, and R. Srivastava, “Exploiting multimodal CNN architecture for automated teeth segmentation on dental panoramic X-ray images,” Proc Inst Mech Eng H, vol. 237, no. 3, pp. 395–405, 2023.
  • [31] M. Kanwal, M. M. Ur Rehman, M. U. Farooq, and D.-K. Chae, “Mask-transformer-based networks for teeth segmentation in panoramic radiographs,” Bioengineering, vol. 10, no. 7, p. 843, 2023.
  • [32] M. K. Dhar, M. Deb, D. Madhab, and Z. Yu, “A Deep Learning Approach to Teeth Segmentation and Orientation from Panoramic X-rays,” arXiv preprint arXiv:2310.17176, 2023.
  • [33] A. Ghafoor, S.-Y. Moon, and B. Lee, “Multiclass Segmentation Using Teeth Attention Modules for Dental X-Ray Images,” IEEE Access, vol. 11, pp. 123891–123903, 2023.
  • [34] Y. Zhang et al., “Children’s dental panoramic radiographs dataset for caries segmentation and dental disease detection,” Sci Data, vol. 10, no. 1, p. 380, 2023.
  • [35] W. Brahmi and I. Jdey, “Automatic tooth instance segmentation and identification from panoramic X-Ray images using deep CNN,” Multimed Tools Appl, vol. 83, no. 18, pp. 55565–55585, 2024.
  • [36] E. Asci et al., “A Deep Learning Approach to Automatic Tooth Caries Segmentation in Panoramic Radiographs of Children in Primary Dentition, Mixed Dentition, and Permanent Dentition,” Children, vol. 11, no. 6, p. 690, 2024.
  • [37] S. Wathore and S. Gorthi, “Bilateral symmetry-based augmentation method for improved tooth segmentation in panoramic X-rays,” Pattern Recognit Lett, vol. 188, pp. 1–7, 2025.
  • [38] G. Altan and A. Al Samar, “Tooth segmentation on dental panoramic X-rays using Mask R-CNN,” in Mining Biomedical Text, Images and Visual Features for Information Retrieval, Elsevier, 2025, pp. 481–498.
  • [39] S. Hou, T. Zhou, Y. Liu, P. Dang, H. Lu, and H. Shi, “Teeth U-Net: A segmentation model of dental panoramic X-ray images for context semantics and contrast enhancement,” Comput Biol Med, vol. 152, p. 106296, 2023.
There are 39 citations in total.

Details

Primary Language English
Subjects Bioengineering (Other)
Journal Section Araştırma Articlessi
Authors

Şilan Fidan Vural 0009-0000-9488-3809

Nida Kumbasar 0000-0001-5497-4618

Early Pub Date July 11, 2025
Publication Date
Submission Date March 23, 2024
Acceptance Date February 24, 2025
Published in Issue Year 2025 Volume: 13 Issue: 2

Cite

APA Vural, Ş. F., & Kumbasar, N. (2025). Comparison of VT-based and CNN-based Models on Teeth Segmentation. Balkan Journal of Electrical and Computer Engineering, 13(2), 148-156. https://doi.org/10.17694/bajece.1457754

All articles published by BAJECE are licensed under the Creative Commons Attribution 4.0 International License. This permits anyone to copy, redistribute, remix, transmit and adapt the work provided the original work and source is appropriately cited.Creative Commons Lisansı