Skip to main content

ORIGINAL RESEARCH article

Front. Cell Dev. Biol., 30 October 2024
Sec. Molecular and Cellular Pathology
This article is part of the Research Topic Artificial Intelligence Applications in Chronic Ocular Diseases, Volume II View all 13 articles

PtosisDiffusion: a training-free workflow for precisely predicting post-operative appearance in blepharoptosis patients based on diffusion models

Shenyu Huang&#x;Shenyu Huang1Jiajun Xie&#x;Jiajun Xie1Boyuan YangBoyuan Yang2Qi Gao
Qi Gao1*Juan Ye
Juan Ye1*
  • 1Eye Center, The Second Affiliated Hospital, School of Medicine, Zhejiang University, Zhejiang Provincial Key Laboratory of Ophthalmology, Zhejiang Provincial Clinical Research Center for Eye Diseases, Zhejiang Provincial Engineering Institute on Eye Diseases, Hangzhou, Zhejiang, China
  • 2Department of Mechanical Science and Engineering, University of Illinois at Urbana-Champaign, Urbana, IL, United States

Purpose: This study aims to develop a diffusion-based workflow to precisely predict postoperative appearance in blepharoptosis patients.

Methods: We developed PtosisDiffusion, a training-free workflow that combines face mesh with ControlNet for accurate post-operative predictions, and evaluated it using 39 preoperative photos from blepharoptosis patients. The performance of PtosisDiffusion was compared against three other diffusion-based methods: Conditional Diffusion, Repaint, and Dragon Diffusion.

Results: PtosisDiffusion demonstrated superior performance in subjective evaluations, including overall rating, correction, and double eyelid formation. Statistical analyses confirmed that PtosisDiffusion achieved the highest overlap ratio (0.87 ± 0.07) and an MPLPD ratio close to 1 (1.01 ± 0.10). The model also showed robustness in extreme cases, and ablation studies confirmed the necessity of each model component.

Conclusion: PtosisDiffusion generates accurate postoperative appearance predictions for ptosis patients using only preoperative photographs. Among the four models tested, PtosisDiffusion consistently outperformed the others in both subjective and statistical evaluation.

1 Introduction

Blepharoptosis (Ptosis) is characterized by an abnormally low upper eyelid margin in the primary gaze, which narrows the eye opening and partially covers the eye (Finsterer, 2003). Unilateral or bilateral ptosis can impair appearance and visual function, significantly affecting quality of patient life by causing amblyopia and increased anxiety and depression related to appearance, thereby impacting overall patient wellbeing. The primary treatment for managing ptosis is surgery (Bacharach et al., 2021). Surgical correction of ptosis is recommended not only for cosmetic improvement but also to prevent visual impairments (Salamah et al., 2022). However, ptosis surgery is highly personalized, with the type of procedure determined by the underlying cause of ptosis, its severity, and the function of the levator muscle. This personalized approach, combined with the potential for unexpected surgical outcomes, can increase patient anxiety and depression, and may even reduce patient confidence in decision-making. Therefore, accurate prediction the of postoperative appearance is essential for the success of ptosis surgery (Koka and Patel, 2019).

Accurate predictions provide surgeons with visual feedback on the anticipated changes, aiding in the optimization of surgical plans and the precise adjustment of eyelid positioning to correct any residual deformities. This approach not only improves the precision and effectiveness of the surgery but also enhances patient understanding and confidence in the outcomes. There have been attempts to predict postoperative results in ptosis surgery. Mawatari et al. utilize Adobe Photoshop for predicting levator resection images (Mawatari and Fukushima, 2016) and employ mirror image processing software for ptosis surgery (Mawatari et al., 2021). However, this method is limited by the subjective nature of manual image manipulation and operator variability. Sun el al employed a Generative Adversarial Network (Goodfellow et al., 2014) (GAN) trained on paired pre- and post-surgery data to perform image translation tasks for ptosis prediction (Sun et al., 2022). The GAN approach have been used in image translation tasks widely (Song et al., 2023; Armanious et al., 2020). Though more automated, it faces several constraints. It requires paired data of pre- and post-operative images, which are difficult to acquire. Additionally, GANs are known for training instability and artifacts in generated images. Some results exhibit unrealistic artifacts in the center of the eyebrow.

Recently, diffusion-based methods (Ho et al., 2020; Nichol and Dhariwal, 2021; Rombach et al., 2021) have demonstrated significant success in generating realistic images. Leveraging these advancements, we developed a workflow named PtosisDiffusion, which can accurately predict postoperative outcomes in blepharoptosis without requiring additional model training. We compared PtosisDiffusion with three other diffusion-based methods using human evaluation and abstract statistical analysis. Our results indicate that PtosisDiffusion outperforms the other three methods.

Our primary contributions are as follows.

Unpaired Data Requirement: In contrast to other image translation methods, our model operates without the need for paired data.

Training Efficiency: Our model functions effectively without necessitating additional training.

State-of-the-Art Performance: Our model achieves state-of-the-art results.

New Evaluation Parameters Proposed: We propose two evaluation parameters that are agnostic to real-world measurements and are solely based on image analysis.

2 Methods

We began with the quantification of facial attributes, followed by a brief introduction to diffusion methods and a detailed explanation of each individual method.

2.1 Ptosis attributes measurements

To advance the study of ptosis, it is crucial to develop efficient methods for quantifying various clinical measurements. Traditionally, this process requires the use of a reference object with known dimensions, such as a ruler or a sticker of predetermined size, placed adjacent to the face. While this approach provides accurate measurements, the availability of such reference objects is limited to specific circumstances, thereby restricting the broader application of these clinical measurements. Leveraging artificial intelligence (AI) offers a promising avenue for achieving this goal. Specifically, the use of face mesh, a machine learning model that accurately maps facial landmarks, presents a robust starting point. Face mesh and iris detection (Kartynnik et al., 2019; Ablavatski et al., 2020) detects 468 face landmark 3D points and iris locations, see Figure 1A. Comparing with traditional face landmarks detection with typically 68 points, it achieves better accuracy and offers more flexible face landmarks choices.

Figure 1
www.frontiersin.org

Figure 1. Face attributes illustration. (A) Example of face mesh (human face generated by AI). (B) MRR=r0/r1 where r0 indicates pixel distance from iris center point to upper lid point and r1 indicates pixel distance from iris center point to iris top point. (C) MPLPD indicates pixel distance from iris center point to lid point for angle θ.

Inspired by the marginal reflex distance-1 (MRD1) (Bodnar et al., 2016) and mid-pupil lid distances (MPLDs) (Morris et al., 2011), we define analogous quantities for measuring ptosis attributes. First, we introduce the marginal reflex ratio (MRR), see Figure 1B. This ratio is determined by two vertical distances: the first is the distance along the y-axis from the center of the pupil to the top edge of the pupil’s bounding box, and the second is the distance from the center of the pupil to the top edge of the upper eyelid. Next, we define mid-pupil lid pixel distances (MPLPDs), see Figure 1C, which are similar to MPLDs but measured in pixels. These two criteria are utilized for guiding the models and for evaluating their performance metrics.

2.2 Diffusion methods

Diffusion models are a class of generative models that have gained significant attention for their ability to produce high-quality synthetic samples, including images, audio, and text. In the realm of medical imaging, diffusion methods have been applied to tasks such as classification, segmentation, denoising and reconstruction (Wolleb et al., 2022; Sanchez et al., 2022; Chung and Ye, 2022; Friedrich et al., 2023). Compared to other generative methods, diffusion models excel in generation quality and training stability. Numerous studies have expanded their generative capabilities, including control over the structure of generated results and the manipulation of color schemes (Ruiz et al., 2023; Zhang et al., 2023; Hu et al., 2021).

Recent stable diffusion models trained on billions of real images have achieved remarkable results (Rombach et al., 2021). We leverage this capability for post-operative prediction. We developed a workflow called PtosisDiffusion based on face mesh (Kartynnik et al., 2019; Ablavatski et al., 2020) and ControlNet (Zhang et al., 2023) for accurate post-operative prediction. Additionally, we compared this workflow with three other diffusion-based methods, namely, Conditional Diffusion (Song et al., 2020), Repaint (Lugmayr et al., 2022) and Dragon Diffusion (Mou et al., 2023).

2.2.1 Denoising diffusion probabilistic model (DDPM)

As the foundational diffusion model, Denoising Diffusion Probabilistic Models (Ho et al., 2020) (DDPM) serve as the basis for all subsequent diffusion-based methods. DDPM consists of two processes: the forward process, which repeatedly adds small amounts of Gaussian noise to the sample x0 until it becomes random noise xT, and the reverse process, which gradually denoises xT back to the original sample x0 based on the model estimated noise, see Figure 2. Assuming the data distribution of xt is q(xt). The forward process can be characterized by Equation 1:

qxtxt1=Nxt;1βtxt1,βtI,(1)

where βt(0,1) is a predefined hyper-parameter. Moreover, a sample at any given time t can be characterized as Equation 2:

xt=ᾱtx0+1ᾱtϵt,ϵtN0,I(2)

where αt=1βt,ᾱt=i=1Tαi=i=1T(1βi). The reverse process is learnable and characterized by Equation 3:

pθxt1xt=Nxt1;μθxt,t,Σθxt,t(3)

and further reduced to Equation 4:

xt1=1αtxt1αt1ᾱtϵθxt,t+σtz,(4)

where zN(0,I), σt is a constant derived from βt and ϵθ is an estimation of ϵt.

Figure 2
www.frontiersin.org

Figure 2. The workflow for DDPM. Red arrow indicates forward process which gradually adds random Gaussian noise to the image. Blue arrow indicates reverse process which gradually denoises the image.

During training, a U-Net model is trained to estimate ϵθ. The loss function can be derived as Equation 5:

EtU1,T,x0qx0,ϵN0,Iλtϵϵθxt,t2(5)

where λ(t) is a positive weighting function, xt is computed by Equation 2, tU[[1,T]] is a uniform distribution over the set {1,2,,T}.

During inference, we sample xT from Gaussian distribution. Following reverse process, the noise ϵθ at each step t is estimated and subtracted from xt. At final step, we obtain a clean image x0.

2.2.2 Conditional Diffusion

Conditional Diffusion leverages a probabilistic model to generate images guided by predefined conditions, such as postoperative targets. By incorporating prior information, the diffusion process is directed to produce images that match desired outcomes. This technique is often used for tasks like image synthesis and inpainting, ensuring that the generated images meet specific criteria. Conditional diffusion model generates result under condition label y. Using Bayesian rule, it can be shown (Song et al., 2020) that conditional diffusion can generate images using an unconditional diffusion model, augmented by an additional term xtlogpt(y|xt), where xt is the image sample at step t, to account for the conditional information. We use the target MRR as condition label and create a Gaussian distribution for pt(y|xt). Then the conditional term becomes xt(yy0)2 where y is the measured MRR for image xt and y0 is the target MRR. This approach enables the generation of images with an MRR value close to the desired target, ensuring accurate prediction of postoperative outcomes.

2.2.3 Repaint

Repaint (Lugmayr et al., 2022) is a variation of the diffusion inpainting technique that focuses on editing image content within a specified mask region. During the denoising process, it systematically adds noise to the masked areas, ensuring that their latent distribution gradually aligns with that of the adjacent unmasked sections. Repaint iteratively refines targeted regions, blending them seamlessly with the surrounding image, achieving the desired modifications while maintaining coherence with the unedited sections. This approach is frequently used for tasks requiring selective changes, such as image correction or restoration. In this study, it ensures that edited areas of the eyelid are integrated smoothly into the overall postoperative prediction.

2.2.4 Dragon Diffusion

Dragon Diffusion (Mou et al., 2023) enables drag-style manipulation through gradient guidance and cross-attention guidance, everaging advanced mathematical principles such as attention mechanisms and gradient descent. This approach allows for dynamic adjustments in image generation, facilitating tasks such as object movement, resizing, appearance replacement, and object pasting on diffusion models. The ability to manipulate image elements in real-time makes Dragon Diffusion particularly valuable in applications such as augmented reality and image editing, where precise control over visual content is essential. In this study, we utilize a face mesh to detect key points on the upper eyelid, and manually determine the targeted dragging points. This allows for fine-tuned adjustments to the eyelid’s position, which is crucial for generating accurate predictions of postoperative appearance.

2.2.5 PtosisDiffusion

We designed a PtosisDiffusion workflow specifically designed for ptosis post-operative prediction, leveraging ControlNet to provide additional control over the image generation process. ControlNet introduces external conditioning information (such as the edge map) into the diffusion process via zero-convolution, utilizing convolutional neural networks (CNNs) to influence the latent space of the diffusion model, enabling more targeted and constrained generation. A common implementation of ControlNet involves integrating Canny edge detection (Canny, 1986) as a conditioning method. Canny edge detection is a widely-used algorithm that identifies boundaries within an image by detecting areas of rapid intensity change, producing a binary image where edges are highlighted. This representation offers a clear and concise depiction of the structural outlines within the original image. The pre-trained ControlNet model guides the diffusion model to generate images that have similar Canny edge detection results.

By employing face mesh technology, we developed a robust workflow for accurate ptosis postoperative prediction. Initially, the face mesh is used to detect the patient’s eye contour and iris location, from which a Canny edge image and a binary mask are generated. The Canny edge image reflects the appropriate MRR value and serves as input for ControlNet diffusion inpainting, while the binary mask is utilized in the inpainting process. This methodology ensures that the generated images retain the desired structural characteristics, which are crucial for accurate and reliable predictions of postoperative outcomes. Furthermore, this approach has applications in other fields requiring precise image manipulation, such as cosmetic surgery and facial reconstruction. A schematic representation of this workflow is illustrated in Figure 3.

Figure 3
www.frontiersin.org

Figure 3. The workflow of PtosisDiffusion.

3 Experiment results

We collected pre-operative photos from a total of 39 patients. Among them, 13 had bilateral ptosis, 26 had unilateral ptosis, 26 were male, and 13 were female. Figure 4 displays a selection of postoperative predictions generated by different methods. A comprehensive set of all 39 predicted images is provided in the Supplementary Information (SI) for complete transparency. We conducted both human evaluation and statistical analysis to assess the performance of each method.

Figure 4
www.frontiersin.org

Figure 4. Typical examples of predicted postoperative images obtained with different models.

3.1 Subjective evaluation

We engaged two doctors—a junior ophthalmologist and a senior ophthalmologist—to assess the prediction results of four models across three different aspects: overall rating, correction amount, and double eyelid formation (DEF). In selecting the metrics for better evaluating prediction quality, we focused on both functional and aesthetic outcomes. The correction amount is vital for ensuring proper eyelid functionality post-surgery, which directly impacts the patient’s visual comfort and overall satisfaction. Additionally, the creation of a well-defined and persistent double-eyelid crease is of paramount importance, especially for East Asian patients (Lee et al., 2013), where aesthetic outcomes are closely tied to patient satisfaction. Therefore, double eyelid formation (DEF) was chosen as a key metric to assess the accuracy of aesthetic predictions. By incorporating both correction amount and DEF, our evaluation framework provides a comprehensive assessment of the surgical outcomes that are most relevant to patients and clinicians alike. For the overall rating, we adopted a three-point scale, evaluating and scoring in three distinct categories: 0 (poor), 1 (average), and 2 (excellent), larger value indicates better result. For correction amount, under-correction was scored as −1, over-correction as 1, and correct correction as 0, value closer to zero indicates better result. For DEF, we used a similar rating standard to the overall rating.

The results, as shown in Table 1 and Figure 5, indicate that our PtosisDiffusion model outperformed the other three models in all aspects.

Table 1
www.frontiersin.org

Table 1. Mean and standard deviation of evaluation metrics. For overall rating and double eyelid formation (DEF), larger value indicates better result. For correction amount, value closer to zero indicates better result.

Figure 5
www.frontiersin.org

Figure 5. Subjective evaluation of predicted postoperative images obtained with different models: three-point scale of overall rating, correction and double eyelid formation (DEF).

3.2 Statistical evaluation

Since we only collected pre-operative images of patients, we do not have post-operative images for direct comparison. Instead, due to the symmetry of the eyes, we selected patients with unilateral ptosis and used the unaffected eye as a standard for comparison. There are a total of 26 patients, with 15 having left eye ptosis and 11 having right eye ptosis. Among them, 20 are male and 6 are female. For comparison, we first aligned the two eyes and then horizontally flipped one eye so that it overlaps with the other. We then detected the contours for both eyes, as shown in Figure 6. We use two stats for evaluation. One is overlap ratio, which is defined as the intersection area of the two eyes over the union area of the two eyes. Another metric is the MPLPD ratio, defined as the ratio of the MPLPDs of the two eyes at a given angle, θ. We select θ at increments of 15° from 0° to 360°. We also introduce a control group, which involves calculating statistics for the original, unmodified images.

Figure 6
www.frontiersin.org

Figure 6. Objective assessment of the predicted postoperative images obtained with different models. (A) The clinical characteristics for patients with unilateral ptosis. (B) Illustration for eye contour and overlap ratio: the ratio of the intersection area over the union area of the predicted eye region (green) and the original eye region (yellow). (C, D) The plots for the overlap ratio and MPLPD ratio across the four models and the control group.

As shown in Table 2 and Figures 6C, D, all four models perform better than the control group. Our model outperforms the other three models in terms of overlap ratio and performs comparably in MPLPD ratio to Dragon Diffusion and Repaint.

Table 2
www.frontiersin.org

Table 2. Mean and standard deviation of evaluation metrics. For overlap ratio, larger values indicate better result. For MPLPD ratio, values closer to one are better.

3.3 Robustness evaluation

We investigate post-operative predictions in extreme cases with severe ptosis, as shown in Figure 7. The other three models fail to output meaningful post-operative images. Yet face meshes are still detected for these cases, and thus output reasonable measurements based on face mesh. This elucidates why the subjective evaluation of our model demonstrates significantly superior performance compared to other models, whereas the statistical evaluation shows relatively less pronounced improvements.

Figure 7
www.frontiersin.org

Figure 7. An example of predicted postoperative images obtained with different models. (A) Model outputs under extreme case. (B) Face mesh detections for test models.

3.4 Ablation study

In our ablation study, we conducted a series of comparisons to evaluate the individual contributions of different components within our workflow. First, we examined the results after removing ControlNet from the complete workflow. Second, we assessed the impact of excluding the inpainting component. Finally, we compared these modified workflows against the results produced by the full, unaltered workflow. The results in Figure 8 clearly demonstrate that both components are essential to our workflow.

Figure 8
www.frontiersin.org

Figure 8. An example of input preoperative images and predicted postoperative images obtained with three different methods (w/o ControlNet, the model w/o inpainting, the full model).

4 Discussion

Our study employs a diffusion model to predict the postoperative appearance of ptosis patients using preoperative photographs. We use both statistics and subjective evaluations to assess the predicted images. To the best of our knowledge, this is the first time a diffusion model has been used for such predictions in ptosis patients.

Subjective evaluation results show that PtosisDiffusion model significantly outperformed the other three models in overall rating, correction, and double eyelid formation. This highlights the superior quality and accuracy of the images generated by our diffusion model, underscoring its potential for clinical use in predicting postoperative outcomes. Statistical evaluation shows that PtosisDiffusion model achieved the highest overlap ratio and second best MPLPD ratio. Compared to the unmodified control, there is a significant improvement, indicating that our model produces better postoperative symmetry in patients with unilateral ptosis. These findings are consistent with the results from the subjective evaluation.

The robustness evaluation reveals that our model maintains consistent performance even in extreme cases, such as fully closed eyes, where other models fail to produce reasonable reconstructions. This robustness underscores the effectiveness of our diffusion model in handling a variety of challenging scenarios, further validating its potential for reliable clinical application.

The results of our ablation studies demonstrate that each component of our model is essential. The removal of the ControlNet component results in the model losing the ability to control the eye contour of the image. Similarly, omitting the inpainting component results in a loss of identity preservation for the original patients.

Compared to studies using GAN networks, our diffusion model does not require paired data and extra training. This is a significant advantage as it reduces the complexity and resource intensity associated with data collection and preparation. The ability to generate the prediction without the need for paired preoperative and postoperative images can streamline the process and make it more feasible to implement in clinical settings. Notably, the model takes approximately 12 s to generate predictions on a Nvidia T4 GPU, which underscores its computational efficiency and feasibility for implementation in a clinical environment. Furthermore, the images generated by the diffusion model are more realistic, enhancing the visual accuracy and reliability of the predicted postoperative appearances.

An important consideration in studies involving postoperative image prediction is the quality of the preoperative images. In previous studies, image-to-image translation methods were employed, where the quality of the input preoperative image directly influenced the quality of the generated postoperative prediction. However, in our study, we utilized the inpainting approach, which is a generative process rather than a translation. This method synthesizes the postoperative appearance based on the input image’s features rather than simply altering the original image. As a result, the quality of the preoperative photograph does not significantly impact the quality of the predicted image, provided that the preoperative image is not extremely blurry—a condition that is typically avoided in clinical practice. The inpainting method we used in this study mitigates the dependency on preoperative image quality, thus ensuring the reliability and applicability of the findings. While our inpainting method offers improved robustness, certain factors may still influence the accuracy and consistency of the model’s predictions. The accuracy of the facial mesh used in the workflow is critical; if key facial landmarks are incorrectly identified, it can lead to distorted outputs. Additionally, variability in facial expressions in the preoperative images can impact the predictions, as different expressions may alter the appearance of features and affect synthesis outcomes. Furthermore, the parameters and settings applied during the inpainting process can also affect the quality of predictions. Careful optimization of these settings is essential to accommodate different input images and ensure high-quality postoperative predictions.

One limitation of our study is the lack of actual postoperative photographs for comparison, as the diffusion model does not require paired data. This absence of real postoperative images as a benchmark may impact the thorough evaluation of the model’s predictive accuracy. However, the use of unpaired data is also an advantage, as it allows for greater flexibility and applicability of the model, particularly in scenarios where obtaining paired data is challenging. Despite this limitation, we have employed statistical methods based on symmetry and overlap rates to evaluate the model’s performance. These methods, while valuable, may not fully capture the nuances of individual patient outcomes. Therefore, future work should consider integrating methods to validate the predicted images against actual postoperative results to further substantiate the model’s effectiveness. Expanding the dataset to include postoperative images will enable more comprehensive and direct comparisons, thereby enhancing the overall evaluation of the model’s predictive capabilities.

Another limitation is the lack of precise control over the generated images. While we generate canny edge maps based on the MRR value for ControlNet reference, the output results often exhibit different MRR values. We attribute this discrepancy to the inherent characteristics of the ControlNet and the diffusion model. The accuracy of ControlNet is contingent upon the precision of the canny edge detection, which may not be ideal for our specific case. Additionally, as a probabilistic model, the diffusion model may produce slightly varying results with each iteration. This variability in generated images introduces a degree of uncertainty that could be a limitation when applying the model to clinical decision-making. Future research should aim to develop methods that enhance the control over image generation, possibly by refining edge detection techniques or incorporating additional constraints to reduce output variability. If further techniques become available for precise control of the diffusion model, our model could potentially achieve even greater performance enhancements.

Despite the limitations, our study opens several promising avenues for future research and clinical applications. PtosisDiffusion could be particularly useful in preoperative planning, allowing surgeons to simulate and visualize potential outcomes based on individual patient characteristics. This capability could enhance patient counseling, helping patients set realistic expectations and making informed decisions about their surgery. Additionally, PtosisDiffusion could serve as a valuable tool in surgical training, enabling trainees to explore different surgical outcomes without actual patient involvement. Future studies could also explore the application of diffusion models in predicting outcomes for other types of ocular and facial surgeries, broadening the scope of its clinical utility. Additionally, integrating longitudinal data that captures different stages of postoperative recovery could offer more comprehensive predictions, aiding in both surgical planning and patient counseling. Moreover, the integration of longitudinal data capturing different stages of postoperative recovery could provide more dynamic and comprehensive predictions, aiding surgeons in tailoring postoperative care and optimizing recovery outcomes.

5 Conclusion

This study introduced a diffusion model approach to predict postoperative appearance for ptosis patients using preoperative photographs, without the need for paired data and extra training. We tested four different diffusion models, with PtosisDiffusion, producing the best results. Subjective and statistical evaluations demonstrated that PtosisDiffusion outperformed others in overall rating, correction, and double eyelid formation, achieving better postoperative symmetry.

Despite limitations such as lack of real postoperative photographs for comparison, our model showed robustness in generating satisfactory images under extreme conditions. Overall, our diffusion model, particularly PtosisDiffusion, represents a significant advancement in predicting postoperative outcomes for ptosis patients, with the potential to enhance clinical practice.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving humans were approved by The Ethics Committee of the Second Affiliated Hospital of Zhejiang University, College of Medicine. The studies were conducted in accordance with the local legislation and institutional requirements. Written informed consent for participation in this study was provided by the participants and apos; legal guardians/next of kin. Written informed consent was obtained from the individual(s), and minor(s) and apos; legal guardian/next of kin, for the publication of any potentially identifiable images or data included in this article.

Author contributions

SH: Conceptualization, Data curation, Writing–original draft, Writing–review and editing, Investigation. JX: Data curation, Formal Analysis, Writing–review and editing, Supervision. BY: Conceptualization, Formal Analysis, Investigation, Methodology, Writing–review and editing. QG: Conceptualization, Formal Analysis, Project administration, Supervision, Writing–review and editing. JY: Conceptualization, Funding acquisition, Project administration, Supervision, Writing–review and editing.

Funding

The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This work has been financially supported by Key Research and Development Program of Zhejiang Province (2024C03204), Key Program of the National Natural Science Foundation of China (82330032), National Natural Science Foundation Regional Innovation and Development Joint Fund (U20A20386).

Acknowledgments

We acknowledge the use of stable-diffusion-v1-5 (Huggingface) for generating images in this study. This model is available at https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fcell.2024.1459336/full#supplementary-material

References

Ablavatski, A., Vakunov, A., Grishchenko, I., Raveendran, K., and Zhdanovich, M. (2020). Real-time pupil tracking from monocular video for digital puppetry, 11341. arXiv preprint arXiv:2006.

Google Scholar

Armanious, K., Jiang, C., Fischer, M., Küstner, T., Hepp, T., Nikolaou, K., et al. (2020). Medgan: medical image translation using gans. Comput. Med. imaging Graph. 79, 101684. doi:10.1016/j.compmedimag.2019.101684

PubMed Abstract | CrossRef Full Text | Google Scholar

Bacharach, J., Lee, W. W., Harrison, A. R., and Freddo, T. F. (2021). A review of acquired blepharoptosis: prevalence, diagnosis, and current treatment options. Eye 35, 2468–2481. doi:10.1038/s41433-021-01547-5

PubMed Abstract | CrossRef Full Text | Google Scholar

Bodnar, Z. M., Neimkin, M., and Holds, J. B. (2016). Automated ptosis measurements from facial photographs. JAMA Ophthalmol. 134, 146–150. doi:10.1001/jamaophthalmol.2015.4614

PubMed Abstract | CrossRef Full Text | Google Scholar

Canny, J. (1986). A computational approach to edge detection. IEEE Trans. pattern analysis Mach. Intell. 8, 679–698. doi:10.1109/tpami.1986.4767851

PubMed Abstract | CrossRef Full Text | Google Scholar

Chung, H., and Ye, J. C. (2022). Score-based diffusion models for accelerated mri. Med. image Anal. 80, 102479. doi:10.1016/j.media.2022.102479

PubMed Abstract | CrossRef Full Text | Google Scholar

Finsterer, J. (2003). Ptosis: causes, presentation, and management. Aesthetic Plast. Surg. 27, 193–204. doi:10.1007/s00266-003-0127-5

PubMed Abstract | CrossRef Full Text | Google Scholar

Friedrich, P., Wolleb, J., Bieder, F., Thieringer, F. M., and Cattin, P. C. (2023). “Point cloud diffusion models for automatic implant generation,” in International conference on medical image computing and computer-Assisted Intervention (Springer), 112–122.

CrossRef Full Text | Google Scholar

Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., et al. (2014). Generative adversarial nets. Adv. neural Inf. Process. Syst. 27, 2672–2680.

Google Scholar

Ho, J., Jain, A., and Abbeel, P. (2020). Denoising diffusion probabilistic models. Adv. neural Inf. Process. Syst. 33, 6840–6851.

Google Scholar

Hu, E. J., Shen, Y., Wallis, P., Allen-Zhu, Z., Li, Y., Wang, S., et al. (2021). Lora: low-rank adaptation of large language models. arXiv Prepr. arXiv:2106.09685.

Google Scholar

Kartynnik, Y., Ablavatski, A., Grishchenko, I., and Grundmann, M. (2019). Real-time facial surface geometry from monocular video on mobile gpus. arXiv Prepr. arXiv:1907.06724.

Google Scholar

Koka, K., and Patel, B. C. (2019). Ptosis correction

Google Scholar

Lee, C. K., Ahn, S. T., and Kim, N. (2013). Asian upper lid blepharoplasty surgery. Clin. plastic Surg. 40, 167–178. doi:10.1016/j.cps.2012.07.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Lugmayr, A., Danelljan, M., Romero, A., Yu, F., Timofte, R., and Van Gool, L. (2022). “Repaint: inpainting using denoising diffusion probabilistic models,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 11461–11471.

CrossRef Full Text | Google Scholar

Mawatari, Y., and Fukushima, M. (2016). Predictive images of postoperative levator resection outcome using image processing software. Clin. Ophthalmol. 10, 1877–1881. doi:10.2147/OPTH.S116891

PubMed Abstract | CrossRef Full Text | Google Scholar

Mawatari, Y., Kawaji, T., Kakizaki, H., Vaidya, A., and Takahashi, Y. (2021). Usefulness of mirror image processing software for creating images of expected appearance after blepharoptosis surgery. Int. Ophthalmol. 41, 1151–1156. doi:10.1007/s10792-020-01671-3

PubMed Abstract | CrossRef Full Text | Google Scholar

Morris, C. L., Morris, W. R., and Fleming, J. C. (2011). A histological analysis of the müllerectomy: redefining its mechanism in ptosis repair. Plastic Reconstr. Surg. 127, 2333–2341. doi:10.1097/PRS.0b013e318213a0cc

PubMed Abstract | CrossRef Full Text | Google Scholar

Mou, C., Wang, X., Song, J., Shan, Y., and Zhang, J. (2023). Dragondiffusion: enabling drag-style manipulation on diffusion models. arXiv Prepr. arXiv:2307.02421.

Google Scholar

Nichol, A. Q., and Dhariwal, P. (2021). “Improved denoising diffusion probabilistic models,” in International conference on machine learning (PMLR) 139, 8162–8171.

Google Scholar

Rombach, R., Blattmann, A., Lorenz, D., Esser, P., and Ommer, B. (2021). High-resolution image synthesis with latent diffusion models

Google Scholar

Ruiz, N., Li, Y., Jampani, V., Pritch, Y., Rubinstein, M., and Aberman, K. (2023). “Dreambooth: fine tuning text-to-image diffusion models for subject-driven generation,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 22500–22510.

CrossRef Full Text | Google Scholar

Salamah, M., Eid, A. M., Albialy, H., and Sharaf El Deen, S. (2022). Anterior approach levator plication for congenital ptosis, absorpable versus non absorpable sutures. Eur. J. Ophthalmol. 32, 134–139. doi:10.1177/11206721211005320

PubMed Abstract | CrossRef Full Text | Google Scholar

Sanchez, P., Kascenas, A., Liu, X., O’Neil, A. Q., and Tsaftaris, S. A. (2022). “What is healthy? generative counterfactual diffusion for lesion localization,” in MICCAI Workshop on Deep generative models (Springer), 34–44.

CrossRef Full Text | Google Scholar

Song, F., Zhang, W., Zheng, Y., Shi, D., and He, M. (2023). A deep learning model for generating fundus autofluorescence images from color fundus photography. Adv. Ophthalmol. Pract. Res. 3, 192–198. doi:10.1016/j.aopr.2023.11.001

PubMed Abstract | CrossRef Full Text | Google Scholar

Song, Y., Sohl-Dickstein, J., Kingma, D. P., Kumar, A., Ermon, S., and Poole, B. (2020). Score-based generative modeling through stochastic differential equations, 13456. arXiv preprint arXiv:2011.

Google Scholar

Sun, Y., Huang, X., Zhang, Q., Lee, S. Y., Wang, Y., Jin, K., et al. (2022). A fully automatic postoperative appearance prediction system for blepharoptosis surgery with image-based deep learning. Ophthalmol. Sci. 2, 100169. doi:10.1016/j.xops.2022.100169

PubMed Abstract | CrossRef Full Text | Google Scholar

Wolleb, J., Sandkühler, R., Bieder, F., Valmaggia, P., and Cattin, P. C. (2022). “Diffusion models for implicit image segmentation ensembles,” in International conference on medical imaging with Deep learning (PMLR) 172, 1336–1348.

Google Scholar

Zhang, L., Rao, A., and Agrawala, M. (2023). “Adding conditional control to text-to-image diffusion models,” in Proceedings of the IEEE/CVF International conference on computer vision, 3836–3847.

CrossRef Full Text | Google Scholar

Keywords: deep learning, diffusion model, blepharoptosis, post-operative appearance, prediction

Citation: Huang S, Xie J, Yang B, Gao Q and Ye J (2024) PtosisDiffusion: a training-free workflow for precisely predicting post-operative appearance in blepharoptosis patients based on diffusion models. Front. Cell Dev. Biol. 12:1459336. doi: 10.3389/fcell.2024.1459336

Received: 04 July 2024; Accepted: 17 October 2024;
Published: 30 October 2024.

Edited by:

Weihua Yang, Shenzhen Eye Institute, Shenzhen Eye Hospital, Jinan University, China

Reviewed by:

Shikha Joon, National Cancer Institute at Frederick (NIH), United States
Gangyong Jia, Hangzhou Dianzi University, China
Yongxin Zhang, Shenzhen Eye Hospital, China, in collaboration with reviewer GJ
Wei Wang, Sun Yat-sen University, China

Copyright © 2024 Huang, Xie, Yang, Gao and Ye. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Qi Gao, Z2FvcWk4OTc3QHpqdS5lZHUuY24=; Juan Ye, eWVqdWFuQHpqdS5lZHUuY24=

These authors have contributed equally to this work and share first authorship

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.