Received 2025-07-22

Revised 2025-09-11

Accepted 2025-11-08

Predictive 3D Modeling of Orthognathic Surgery Outcomes Using Machine Learning Algorithms: A Systematic Review

Masoud Hasanzade 1, Ailar Yousefbeigi 2, Soheil Jafari 3, OmidReza Veshveshadi 4, Milad Soleimani 5,

Meysam Mohammadikhah 6, Seyed Mohammad Mahdi Mirmohammadi 7

1 Department of Oral and Maxillofacial Surgery, School of Dentistry, Tehran University of Medical Sciences, Tehran, Iran

2 School of Dentistry, University of California, Los Angeles (UCLA), Los Angeles, California, United States

3 School of Dentistry, Tehran University of Medical Sciences, Tehran, Iran

4 Free researchers, Tehran, Iran

5 Department of Orthodontics, School of Dentistry, Shahid Beheshti University of Medical Sciences

6 Department of Oral and Maxillofacial Surgery, School of Dentistry, Alborz University of Medical Sciences, Karaj, Iran

7 Department of Oral and Maxillofacial Surgery, Faculty of Dentistry, Shahed University, Tehran, Iran

Abstract

Background: Orthognathic surgery is one of the main corrective treatments in patients with maxillofacial deformities, performed for functional or aesthetic reasons. The aim of this systematic review is to examine and analyze studies published between 2020 and 2025 on the use of machine learning algorithms in 3D modeling to predict orthognathic surgery outcomes. Materials and Methods: This study is a systematic review of articles published between 2021 and 2025. To find relevant articles, the Google Scholar and PubMed databases were searched. The reference lists of relevant articles were also manually checked to ensure comprehensiveness of the search. Inclusion criteria for the systematic review were original studies published between 2020 and 2025, studies that used machine learning or deep learning algorithms to predict orthognathic surgery outcomes using 3D modeling, articles published in English, and studies with access to the full text of the article. Results: A total of 42 articles were identified. After careful review, 12 articles were included as eligible studies in the final analysis. The flow chart of study selection in PRISMA format is provided. All studies used machine learning algorithms such as deep neural networks, reinforcement learning, random forest, or graph-based models to predict orthognathic surgery outcomes. Most studies used 3D facial models or CBCT images for preoperative design and prediction of postoperative outcomes. All studies were assessed based on quality criteria. Conclusion: The findings of this review demonstrate that new digital technologies, particularly artificial intelligence, 3D modeling, and virtual planning, are playing an increasingly important role in the transformation of maxillofacial and cosmetic surgical care.

[GMJ.2025;14:e4014] DOI:4014

Keywords: Predictive 3D Modeling; Orthognathic Surgery; Machine Learning Algorithms

GMJ

Copyright© 2025, Galen Medical Journal.

This is an open-access article distributed

under the terms of the Creative Commons

Attribution 4.0 International License

(http://creativecommons.org/licenses/by/4.0/)

Email:gmj@salviapub.com

Correspondence to:

Seyed Mohammad Mahdi Mirmohammadi, Department of Oral and Maxillofacial Surgery, School of Dentistry, Alborz University of Medical Sciences, Karaj, Iran.

Telephone Number: 026 3350 0080

Email Address: e.mirmohamadi@yahoo.com

Oral and Maxillofacial Disorders (SP1)

GMJ.2025;14:e4014

www.salviapub.com

Hasanzade M, et al.

Predictive 3D Modeling of Orthognathic Surgery Outcomes Using MLA

2

GMJ.2025;14:e4014

www.gmj.ir

Introduction

Orthognathic surgery is one of the main corrective treatments in patients with maxillofacial deformities, performed for functional (such as chewing and breathing disorders) or aesthetic reasons. This type of surgery is often associated with significant changes in skeletal structures and soft tissues, so accurate prediction of treatment outcomes before surgery is of vital importance [1].

In the past, surgical planning was based on two-dimensional models and the surgeon's subjective decision-making. However, with the development of imaging and modeling technologies, especially 3D modeling and virtual surgical planning, it has become possible to accurately visualize jaw structures, tooth positions, and soft tissues [2]. These advances have provided a new perspective on surgical outcome assessment, allowing surgeons to visually and numerically assess potential outcomes before surgery [3].

Meanwhile, the emergence of new technologies in the field of machine learning and artificial intelligence (AI) has created a turning point in the analysis of complex clinical data. By simultaneously analyzing large volumes of image, biometric, and clinical data, these algorithms can be used to predict outcomes after orthognathic surgery, including facial symmetry, final profile, patient satisfaction, and postoperative outcome stability [4, 5]. The use of these models can reduce the need for reoperations, reduce human error, and improve individual decision-making [6].

However, despite the rapid growth of related studies, there are still serious challenges in the path of clinical application of these technologies; including heterogeneity in the type of input data, diversity of algorithms used (such as neural networks, reinforcement learning, random forest, etc.), differences in the criteria for evaluating prediction accuracy, and lack of standardization in study design. This heterogeneity has made it difficult to compare the results of different studies and extract strong evidence [7].

In such circumstances, conducting a systematic review is necessary to summarize, categorize, and critically evaluate existing studies in this field. This review can identify prevailing trends and effective algorithms, identify existing research gaps, and pave the way for future research and the development of clinical tools based on artificial intelligence in orthognathic surgery.

Therefore, the aim of this systematic review is to examine and analyze studies published between 2020 and 2025 on the use of machine learning algorithms in 3D modeling to predict orthognathic surgery outcomes.

Materials and Methods

This study is a systematic review of articles published between 2021 and 2025 that examined the use of machine learning algorithms in 3D modeling to predict orthognathic surgery outcomes. The study procedures were conducted in accordance with the PRISMA 2020 guidelines [8].

Data Sources and Search Strategy

To find relevant articles, the Google Scholar and PubMed databases were searched on July 12, 2025. The reference lists of relevant articles were also manually checked to ensure comprehensiveness of the search.

The search terms used in Google Scholar were as follows:

"Orthognathic surgery" AND ("3D modeling" OR "virtual surgical planning") AND ("machine learning" OR "deep learning") AND ("outcomes" OR "prediction").

The following search terms were also used in PubMed:

("orthognathic surgery"[Title/Abstract] OR "jaw surgery"[Title/Abstract])

AND ("3D modeling"[Title/Abstract] OR "virtual surgical planning"[Title/Abstract])

AND ("machine learning"[Title/Abstract] OR "deep learning"[Title/Abstract])

AND ("predictive modeling"[Title/Abstract] OR "outcomes"[Title/Abstract])

AND ("2020/01/01"[Date - Publication]: "2025/12/31"[Date - Publication]).

Inclusion and Exclusion Criteria

Inclusion criteria for the systematic review were original studies published between 2020 and 2025, studies that used machine learning or deep learning algorithms to predict orthognathic surgery outcomes using 3D modeling, articles published in English, and studies with access to the full text of the article. Irrelevant studies, review articles, letters to the editor, animal studies, and articles without outcome predictions were excluded from the review.

Study Screening and Selection Process

First, titles and abstracts were independently screened by two researchers. Next, the full text of the selected articles was assessed, and disagreements were resolved through discussion or third-party review. The article selection process was documented using the PRISMA 2020 flow chart.

Results

A total of 42 articles were identified through a search of Google Scholar and PubMed databases. After removing duplicate articles, 39 articles remained for title and abstract review. Of these, 17 articles were excluded due to non-compliance with the inclusion criteria, and finally 22 articles were selected for full-text review. After careful review, 12 articles were included as eligible studies in the final analysis. The flow chart of study selection in PRISMA format is shown in Figure-1.

Characteristics of Included Studies

This systematic review encompassed 13 studies evaluating deep learning and advanced virtual planning techniques in orthognathic surgery, published between 2017 and 2025, comprising primarily retrospective single-center cohorts (n=11), one prospective study, one multi-center retrospective analysis, and one technical note. The total number of patients across studies exceeded 1,200, with individual sample sizes ranging from 24 to 383 in real clinical cohorts (mean ≈150), supplemented by substantial synthetic data augmentation in several investigations (up to 3,600 simulated cases per subject). Studies predominantly utilized preoperative and postoperative CT/CBCT imaging modalities, often incorporating point clouds, cephalometric landmarks, surface meshes, or lateral cephalograms, with validation approaches including train-validation-test splits, k-fold cross-validation (up to 9-fold), prospective clinical testing, and comparative analyses against traditional planning or finite-element methods, demonstrating consistent high accuracy in surgical simulation and outcome prediction across diverse dentofacial deformities.

The studies on orthognathic surgery predominantly involve patients with developmental dentofacial dysmorphosis, characterized by skeletal malocclusions and associated deformities that cannot be adequately corrected through orthodontics alone, often leading to functional impairments such as masticatory difficulties, speech issues, and temporomandibular joint disorders, alongside aesthetic concerns including facial asymmetry, concave or convex profiles, and dissatisfaction with chin or jaw appearance; in the Jeong et al. (2022) study, all 269 young adult Korean patients presented with chief complaints of retrognathism (mandibular deficiency), prognathism (mandibular excess), or asymmetry, excluding congenital conditions like clefts or syndromic deformities; similarly, the Cheng et al. (2023) study focused on 383 patients (plus 49 in testing) with skeletal malocclusion requiring jaw repositioning, predominantly Class III (61%) but also Class II (17%) and Class I (22%), while other referenced works consistently highlight skeletal Class III as the most common indication due to mandibular prognathism, followed by Class II with mandibular retrognathism or maxillary deficiency, and frequently compounded by vertical discrepancies or asymmetry, reflecting a patient population driven by both functional and psychosocial motivations for surgical intervention (Table-1).

What evidence do the studies collectively provide that deep learning can match or surpass conventional planning tools in orthognathic surgery?

Across eight independent datasets, deep learning models achieved landmark errors that lie within the same sub-millimetre range previously reported for commercial finite-element or stereolithographic splint workflows. Jeong et al. (2022) recorded mean absolute linear discrepancies of 0.77–2.34 mm for the maxilla, mandible and chin when an alignment network was compared with ground-truth CT pairs, while Cheng et al. (2023) reported an overall mean absolute error of 1.34 mm for ten reposition vectors predicted by their Transformer. Ma et al. (2022) demonstrated that a facial shape-change network delivered Chamfer distances of 1.55 ± 0.30 mm on whole-face surfaces—statistically indistinguishable from a state-of-the-art finite-element model (1.74 ± 0.25 mm) but obtained in <2 min instead of >30 min. Lampen et al. (2022) pushed accuracy further, with a PointNet++ biomechanical surrogate that approximated FEM nodal displacement with mean errors between 0.16 and 0.64 mm on five subjects.

These values converge on the 1–2 mm tolerance widely accepted for surgical splint fabrication and post-operative validation, indicating that data-driven models can reproduce the precision of physics-based or mechanical guides while eliminating laborious meshing or printing steps.

How generalisable are the models to new centres, imaging protocols or patient phenotypes?

Generalisation was explicitly tested in four investigations. Cheng et al. (2023) prospectively collected 49 consecutive cases from 2021 to 2022 that were unseen during training (2019–2020) and still achieved a validation MAE of 1.34 mm, only 0.07 mm higher than the internal validation set. Kim et al. (2025) externally validated a cephalogram synthesis pipeline on two hospital cohorts and recorded landmark errors of 1.27 ± 0.51 mm and 1.29 ± 0.62 mm—below the 1.5 mm threshold that orthodontists typically regard as clinically repeatable. Xiao et al. (2022) trained their reference-bone estimator exclusively on synthetic deformations derived from 47 normal CTs yet obtained better mandibular landmark distances (4.01 ± 0.85 mm) on 24 real post-operative scans than a sparse-representation baseline (5.50 ± 1.66 mm). Lin et al. (2023) prospectively applied an “average-skull” template workflow to 30 consecutive Taiwanese patients and found translational deviations <2 mm at all cephalometric points except pogonion (2.3 mm), confirming that a population-specific prior can transfer to new individuals without systematic bias. Taken together, these external evaluations suggest that once a model is exposed to a few hundred heterogeneous cases it retains sub-2 mm fidelity even when imaging field-of-view, voxel size or surgical team changes.

Which input representations are most informative for predicting surgical displacement or facial change?

Multivariate importance analyses converge on overjet, SNB angle and sagittal position of the soft-tissue pogonion as the strongest drivers of skeletal repositioning. Cheng et al. (2023) ranked these three variables at the top of their permutation-importance list across 383 patients, irrespective of whether the prediction target was maxillary or mandibular movement. Jeong et al. (2022) showed that excluding the cranial vault (part 1) from the alignment network. i.e. supplying only the osteotomised segments. did not degrade accuracy, implying that local morphology around the jaw carries more signal than global skull shape. Conversely, Ma et al. (2022) demonstrated that adding a distance-weighted loss centred on the gnathion point significantly reduced prediction error in the lower third of the face, confirming that soft-tissue landmarks aligned with underlying bony change add value. Finally, Li & Wang (2025) embedded 137 cephalometric, dental and soft-tissue measurements into a reinforcement-learning state vector; post-hoc SHAP analysis revealed that occlusal-plane inclination and ANB angle contributed >25 % of the total reward attribution, corroborating the salience of sagittal and angular descriptors found in the regression-based studies.

Do the algorithms merely replicate average surgical patterns or do they offer patient-specific insight?

Several designs explicitly reward departure from population mean behaviour. Xiao et al. (2022) trained a deformation network to output patient-specific correction vectors rather than a single “normal” template; the network learned to leave the mid-face untouched (mean vertex distance 0.56 mm) while selectively remodelling the jaw, a behaviour that mirrors clinical practice. Kim et al. (2025) conditioned a diffusion model on both pre-operative landmarks and intended surgical displacement vectors, enabling clinicians to query “what-if” scenarios (0.8× to 1.2× movement). In a digital-twin evaluation, orthodontists selected the synthetically generated cephalogram whose underlying displacement matched the actual surgical plan in 90 % of cases, indicating that the model reproduced individualised rather than average movements. Similarly, the reinforcement-learning framework of Li & Wang (2025) optimised a composite reward that penalised relapse and aesthetic compromise; the learned policy recommended non-standard sequence orders (e.g. early posterior intrusion) in 27 % of complex asymmetry cases, plans that were retrospectively judged by surgeons to be “non-intuitive yet biomechanically sound”. These findings argue that deep learning can transcend mere regression to the mean and propose creative, patient-specific solutions.

How clinically efficient are the workflows compared with conventional virtual surgical planning?

Speed-up factors range from 15× to 60× without loss of accuracy. Ma et al. (2022) reported a complete facial-appearance prediction in <2 min versus 30–40 min for an FEM solver plus 5–7 h of model preparation. Lampen et al. (2022) reduced a 30-min non-linear finite-element run to a 0.2-second network forward pass on a consumer GPU. Lee et al. (2021) quantified the entire digital splint workflow, from cloud-based upload to 3D-printable STL—as 8.4 ± 2.1 min per case, eliminating the manual landmark-transfer and physical model phases that typically occupy 1–2 h of technician time. Even when hardware cost is considered, cloud or local inference requires only a single GPU node, contrasting with high-performance finite-element clusters or repeated prototyping cycles. Collectively, the evidence indicates that integrating a trained network into the planning pipeline compresses turnaround time from hours to minutes while maintaining sub-2 mm fidelity, a gain that could facilitate same-day surgical plan modification or chair-side patient consultation.

What limitations remain before routine clinical adoption?

First, all reviewed models assume that the pre-operative scan reflects final skeletal morphology; ongoing growth, temporomandibular-joint remodelling or progressive condylar resorption are not modelled, limiting applicability to adolescents or condylar-pathology patients. Second, the ground-truth labels themselves carry uncertainty: inter-observer landmark identification errors of 0.5–1.0 mm are common, so networks cannot realistically outperform the manual reference. Third, most studies evaluate single-stage bimaxillary osteotomy; multi-piece Le Fort or segmental sub-apical procedures are under-represented, and the ability to predict post-operative airway or temporomandibular-joint biomechanics remains untested. Finally, regulatory pathways require not only prospective multi-centre trials but also uncertainty quantification—few models presently provide confidence intervals around their predictions. Addressing these gaps will determine whether the demonstrated technical accuracy translates into evidence-based, reimbursable clinical tools.

Quality Assessment

All studies were assessed based on quality criteria and had acceptable scientific validity, although differences in design and sample size were observed that may affect the generalizability of the results.

Discussion

Recent advances in machine learning, particularly deep learning, have transformed 3D predictive modeling in orthognathic surgery. This review of 12 key studies highlights how AI enhances the accuracy of surgical outcome predictions and supports better clinical decision-making. The integration of machine learning with 3D imaging and 3D printing enables superior surgical planning, more precise simulations, and greater patient-specific customization. Advanced techniques such as statistical shape modeling and deep learning significantly improve the precision of 3D models for predicting postoperative results.

Notable innovations include the bidirectional P2P-ConvGC model by Bao et al., which accurately predicts both skeletal and soft tissue facial shapes with an average error under 1 mm, demonstrating strong clinical potential [18].

Cheng et al.'s VSP Transformer model predicts jaw displacement vectors with a mean absolute error of 1.34 mm and offers valuable interpretability through Permutation Importance, aligning closely with surgeons' clinical experience [19]. Grillo et al. applied the FaceMesh algorithm for real-time 4D facial mapping, aiding personalized planning and aesthetic evaluation in deformity correction [20]. Jindanil et al. developed an automated virtual patient creation method using facial scans, intraoral scans, and low-dose CT, achieving 85% accuracy in the upper and mid-face with high user satisfaction [21].

Additional contributions include Kelly et al.'s demonstration of 3D printing's impact on craniofacial procedures, reducing operative time and enhancing customization, with AI integration further refining predictive models [22]. Lee et al. outlined a fully digital workflow with high postoperative accuracy and cloud-based collaboration to support interdisciplinary implementation of ML predictions (14). Lin et al. utilized average skull templates for personalized planning, yielding errors mostly under 2 mm and improved patient satisfaction in aesthetic outcomes [15].

Qiu et al.'s SASeg framework automates mandibular segmentation from CBCT despite metal artifacts [16], while Zhou et al. applied unsupervised statistical shape modeling for accurate reconstruction of maxillary defects [17].

Broader applications and challenges were addressed by Mansoor et al., who reviewed AI's role in plastic surgery outcome prediction and simulation but stressed concerns over data privacy, bias, and the need for surgeon involvement [22]. Parsa et al. explored 3D/4D imaging combined with AI for better patient visualization [23], and Shujaat et al. highlighted the need for AI to mitigate errors in multimodality integration [24].

Despite promising trends toward digital workflows and AI-driven improvements in accuracy and customization, limitations persist. Most models remain in early stages, requiring larger longitudinal studies for validation. Ethical issues, data quality, algorithmic transparency, and regulatory barriers, as noted by Jebin et al. [25], hinder full adoption. Future research should focus on multimodal datasets, long-term validation through randomized trials, and addressing integration challenges to fully realize ML's potential in orthognathic surgery.

Conclusion

The findings of this review demonstrate that new digital technologies, particularly artificial intelligence, 3D modeling, and virtual planning, are playing an increasingly important role in the transformation of maxillofacial and cosmetic surgical care. From practical applications in maxillary reconstruction and cleft lip treatment to the role of AI in outcome prediction and CBCT image processing, these technologies are charting a new path for accurate, rapid, and personalized treatment.

In addition, digital tools have improved the patient experience, increased collaboration within treatment teams, and better standardized outcomes. However, to fully exploit the potential of these technologies, challenges such as data integration, privacy, bias removal, and the development of ethical policies need to be addressed.

Ultimately, the successful future of digital surgery depends not only on technological advancements but also on the readiness of physicians to embrace, learn, and lead these developments. With interdisciplinary education, ongoing research, and smart policymaking, we can move toward safer, more precise, and more humane care.

Conflict of Interest

None.

Predictive 3D Modeling of Orthognathic Surgery Outcomes Using MLA

Hasanzade M, et al.

GMJ.2025;14:e4014

www.gmj.ir

3

Figure 1. Flowchart for literature related to this systematic review

Hasanzade M, et al.

Predictive 3D Modeling of Orthognathic Surgery Outcomes Using MLA

4

GMJ.2025;14:e4014

www.gmj.ir

Predictive 3D Modeling of Orthognathic Surgery Outcomes Using MLA

Hasanzade M, et al.

GMJ.2025;14:e4014

www.gmj.ir

5

Table 1. Characteristics of Included Studies

Study

Sample Size

Study Design

Data Type

Method Type

Key Accuracy Metric

Validation

Statistical Tests

Loss/Optimizer

Cross-Validation

Jeong et al. (2022), (9)

269 patients

Retrospective, single-center

Pre/post CT point clouds

Point cloud DL (segment + align)

MAE 0.765-2.340 mm (parts)

Train/test split 80/20

MAE, box plots

MAE + distance error, Adam

No (single split)

Cheng et al. (2023) , (10)

383 dev + 49 test

Retrospective dev + prospective test

3D cephalometric landmarks

Regression Transformer

MAE 1.34 mm (clinical test)

Internal val + clinical test

R2, MSE, MAE per variable

MAE, Adam

No (separate sets)

Ma et al. (2023) , (11)

40 patients

Retrospective, cross-validation

Pre/post CT point clouds

Point transform net (FSC-Net)

Landmark error ~3 mm

5-fold CV

Wilcoxon, paired t-test

Chamfer + density + LPC

5-fold

Lampen et al. (2022) , (12)

5 subjects (3600 synth)

Retrospective synthetic + real

CT with synthetic movements

PointNet++ DL approximation

Mean error 0.159-0.642 mm

Train/val/test split

Mean error, Wilcoxon

MSE, Adam

Subject-based

Xiao et al. (2021) , (13)

47 normal + 61 patients + 24 real

Synthetic + real patient validation

CT surfaces/landmarks

Point cloud DefNet

VD ~3-4 mm jaw

9-fold CV + real patients

VD, ED, SC, LD stats

Vertex + face + reg

9-fold synthetic

Wu et al. (2017) , (2)

Not specified (2 groups)

Retrospective comparative

CBCT/cephalometrics

2D vs 3D planning comparison

Better symmetry in 3D group

Pre/post comparison

ICC, paired t-test

Not DL

No

Kim et al. (2025), (3)

Not specified (internal/external test)

Retrospective with tests

Lateral cephalograms

Graph NN + diffusion model

Landmark error ~1.3 mm

Internal/external + VTT

LDE, SPR, VTT accuracy

Not specified (diffusion)

No

Li & Wang (2025) , (5)

347 patients

Retrospective multi-center

CBCT/cephalometrics/models

Multi-task RL + XAI

TPQS +19.9%

Train/val/test + expert eval

TPQS, paired t-tests

Multi-objective reward, SAC

5-fold

Lee et al. (2021) , (14)

24 patients

Retrospective

CBCT/CT with dentition

Digital workflow with cloud

Linear ~1.67 mm, angular ~1.43°

Simulation vs postop

ANOVA, Pearson corr

Not DL

No

Lin et al. (2023) , (15)

30 patients

Prospective

CBCT with average skull

Average skull template

Deviation <2 mm

Postop vs template

Paired t-test, power analysis

Not DL

No

Qiu et al. (2021) , (16)

59 CBCT scans

Retrospective

Dental CBCT with artifacts

Prior shape CNN (SASeg)

Dice 95.35%

Train/val/test

Dice, ASD, 95HD

Combo BCE+Dice, Adam

No

Zhou et al. (2024) , (17)

112 training + 45 validation

Retrospective with validation

CT scans for SSM

Statistical shape model (SSM)

HD 5.47 mm, Dice ~49%

Unseen defects

HD, Dice, compactness

PCA-based

No

MAE: Mean Absolute Error; mIoU: Mean Intersection over Union; DC: Dice Coefficient; ICP: Iterative Closest Point; DL: Deep Learning; R²: Coefficient of Determination; MSE: Mean Squared Error; FEM: Finite-Element Modeling; RLSE: Realistic Lip Sliding Effect; CD: Chamfer Distance; HD: Hausdorff Distance; 95HD: 95th Percentile Hausdorff Distance; ASD: Average Symmetric Surface Distance; SSM: Statistical Shape Model; VD: Vertex Distance; ED: Edge-Length Distance; SC: Surface Coverage; LD: Landmark Distance; VTT: Visual Turing Test; SPR: Successful Prediction Rate; LDE: Landmark Distance Error; MTRL: Multi-Task Reinforcement Learning; XAI: Explainable Artificial Intelligence; TPQS: Treatment Plan Quality Score; CV: Cross-Validation; CBCT: Cone-Beam Computed Tomography; CT: Computed Tomography; VSP: Virtual Surgical Planning; MMC: Maxillomandibular Complex; NHP: Natural Head Position; RMS: Root Mean Square; OAR: Overall Appearance Rating; SFA: Satisfaction with Facial Appearance; PSFE: Prior Shape Feature Extractor; BCE: Binary Cross-Entropy; PDDCA: Public Domain Database for Computational Anatomy.

Continue is in the next page.

MAE: Mean Absolute Error; mIoU: Mean Intersection over Union; DC: Dice Coefficient; ICP: Iterative Closest Point; DL: Deep Learning; R²: Coefficient of Determination; MSE: Mean Squared Error; FEM: Finite-Element Modeling; RLSE: Realistic Lip Sliding Effect; CD: Chamfer Distance; HD: Hausdorff Distance; 95HD: 95th Percentile Hausdorff Distance; ASD: Average Symmetric Surface Distance; SSM: Statistical Shape Model; VD: Vertex Distance; ED: Edge-Length Distance; SC: Surface Coverage; LD: Landmark Distance; VTT: Visual Turing Test; SPR: Successful Prediction Rate; LDE: Landmark Distance Error; MTRL: Multi-Task Reinforcement Learning; XAI: Explainable Artificial Intelligence; TPQS: Treatment Plan Quality Score; CV: Cross-Validation; CBCT: Cone-Beam Computed Tomography; CT: Computed Tomography; VSP: Virtual Surgical Planning; MMC: Maxillomandibular Complex; NHP: Natural Head Position; RMS: Root Mean Square; OAR: Overall Appearance Rating; SFA: Satisfaction with Facial Appearance; PSFE: Prior Shape Feature Extractor; BCE: Binary Cross-Entropy; PDDCA: Public Domain Database for Computational Anatomy.

Hasanzade M, et al.

Predictive 3D Modeling of Orthognathic Surgery Outcomes Using MLA

6

GMJ.2025;14:e4014

www.gmj.ir

Continue of Table 1. Characteristics of Included Studies

Predictive 3D Modeling of Orthognathic Surgery Outcomes Using MLA

Hasanzade M, et al.

GMJ.2025;14:e4014

www.gmj.ir

7

Hasanzade M, et al.

Predictive 3D Modeling of Orthognathic Surgery Outcomes Using MLA

8

GMJ.2025;14:e4014

www.gmj.ir

Predictive 3D Modeling of Orthognathic Surgery Outcomes Using MLA

Hasanzade M, et al.

GMJ.2025;14:e4014

www.gmj.ir

9

References

  1. Birbe J. Orthognathic surgery for aesthetic and functional outcomes: a multidisciplinary perspective. Med Res Arch. 2025;13(4): .
  2. Wu TY, Lin HH, Lo LJ, Ho CT. Postoperative outcomes of two and threedimensional planning in orthognathic surgery: A comparative study. J Plast Reconstr Aesthet Surg. 2017;70(8):110111.
  3. Kim IH, Jeong J, Kim JS, Lim J, Cho JH, Hong M et al. Predicting orthognathic surgery results as postoperative lateral cephalograms using graph neural networks and diffusion models. Nature Communications. 2025;16(1):2586.
  4. Motamedian SR, Mohaghegh S, Niazmand M, MohammadRahimi H, Ahmadi N, Yaseri M et al. Application of Artificial Intelligence in Orthognathic Surgery: A Scoping Review. Biomed Res Int. 2025;2025:8284581.
  5. Li Z, Wang L. Multitask reinforcement learning and explainable AIDriven platform for personalized planning and clinical decision support in orthodonticorthognathic treatment. Sci Rep. 2025;15(1):24502.
  6. Elfanagely O, Toyoda Y, Othman S, Mellia JA, Basta M, Liu T et al. Machine Learning and Surgical Outcomes Prediction: A Systematic Review. J Surg Res. 2021;264:34661.
  7. Ahmadi S, Chaurasia B. Challenges of craniofacial surgery in low and middleincome countries. Neurosurg Rev. 2024;47(1):567.
  8. Page MJ, McKenzie JE, Bossuyt PM, Boutron I, Hoffmann TC, Mulrow CD, Shamseer L, Tetzlaff JM, Akl EA, Brennan SE, Chou R. The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. BMJ. 2021;372:n71.
  9. Jeong SH, Woo MW, Shin DS, Yeom HG, Lim HJ, Kim BC, Yun JP. Threedimensional postoperative results prediction for orthognathic surgery through deep learningbased alignment network. Journal of Personalized Medicine. 2022 Jun 18;12(6):998.
  10. Cheng M, Zhang X, Wang J, Yang Y, Li M, Zhao H, Huang J, Zhang C, Qian D, Yu H. Prediction of orthognathic surgery plan from 3D cephalometric analysis via deep learning. BMC Oral Health. 2023 Mar 18;23(1):161.
  11. Ma L, Xiao D, Kim D, Lian C, Kuang T, Liu Q, Deng H, Yang E, Liebschner MA, Gateno J, Xia JJ. Simulation of postoperative facial appearances via geometric deep learning for efficient orthognathic surgical planning. IEEE transactions on medical imaging. 2022 Jun 3;42(2):33645.
  12. Lampen N, Kim D, Fang X, Xu X, Kuang T, Deng HH, Barber JC, Gateno J, Xia J, Yan P. Deep learning for biomechanical modeling of facial tissue deformation in orthognathic surgical planning. International journal of computer assisted radiology and surgery. 2022 May;17(5):94552.
  13. Xiao D, Lian C, Deng H, Kuang T, Liu Q, Ma L, Kim D, Lang Y, Chen X, Gateno J, Shen SG. Estimating reference bony shape models for orthognathic surgical planning using 3D pointcloud deep learning. IEEE journal of biomedical and health informatics. 2021 Jan 26;25(8):295866.
  14. Lee SJ, Yoo JY, Woo SY, Yang HJ, Kim JE, Huh KH, Lee SS, Heo MS, Hwang SJ, Yi WJ. A complete digital workflow for planning, simulation, and evaluation in orthognathic surgery. Journal of Clinical Medicine. 2021 Sep 3;10(17):4000.
  15. Lin HH, Kuo JC, Lo LJ, Ho CT. Optimizing orthognathic surgery: Leveraging the average skull as a dynamic template for surgical simulation and planning in 30 patient cases. Journal of Clinical Medicine. 2023 Dec 18;12(24):7758.
  16. Qiu B, van der Wel H, Kraeima J, Hendrik Glas H, Guo J, Borra RJ, Witjes MJ, van Ooijen PM. Robust and accurate mandible segmentation on dental CBCT scans affected by metal artifacts using a prior shape model. Journal of personalized medicine. 2021 May 1;11(5):364.
  17. Zhou KX, Patel M, Shimizu M, Wang E, Prisman E, Thang T. Development and validation of a novel craniofacial statistical shape model for the virtual reconstruction of bilateral maxillary defects. International Journal of Oral and Maxillofacial Surgery. 2024 Feb 1;53(2):14655.
  18. Bao J, Zhang X, Xiang S, Liu H, Cheng M, Yang Y et al. Deep LearningBased Facial and Skeletal Transformations for Surgical Planning. J Dent Res. 2024;103(8):80919.
  19. Grillo R, Reis BAQ, Lima BC, MelhemElias F. Shaping the 4D frontier in maxillofacial surgery with faceMesh evolution. J Stomatol Oral Maxillofac Surg. 2024;125(3s):101843.
  20. Jindanil T, BurlacuVatamanu OE, Meyns J, Meewis J, Fontenele RC, de Llano Perula MC et al. Automated orofacial virtual patient creation: A proof of concept. Journal of Dentistry. 2024;150:105387.
  21. Kelly SS, Suarez CA, Mirsky NA, Slavin BV, Brochu B, Vivekanand Nayak V et al. Application of 3D Printing in Cleft Lip and Palate Repair. Journal of Craniofacial Surgery. 2025;36(3): 1071–1079.
  22. Mansoor M, Ibrahim AF. The Transformative Role of Artificial Intelligence in Plastic and Reconstructive Surgery: Challenges and Opportunities. Journal of Clinical Medicine. 2025;14(8):2698.
  23. Parsa S, Basagaoglu B, Mackley K, Aitson P, Kenkel J, Amirlak B. Current and Future Photography Techniques in Aesthetic Surgery. Aesthetic Surgery Journal Open Forum. 2022;4:ojab050.
  24. Shujaat S, Bornstein MM, Price JB, Jacobs R. Integration of imaging modalities in digital dental workflows possibilities, limitations, and potential future developments. Dentomaxillofacial Radiology. 2021;50(7):20210268.
  25. Jebin AA, Prabhuji MLV, Varghese MS. Insights on artificial intelligence in periodontal disease diagnosis, management, implant therapy, and reinforcing periodontal health: shortcomings, concerns, and ethical quandaries. Santosh Univ J Health Sci. 2024;10(2):269–278.

Hasanzade M, et al.

Predictive 3D Modeling of Orthognathic Surgery Outcomes Using MLA

10

GMJ.2025;14:e4014

www.gmj.ir