A conceptual illustration of theoretical analysis. Theorem 1 (Corollary 1 for diffusion model) translates the perturbation in the parameter space into the set of perturbed distributions. Theorem 2 (Corollary 2 for diffusion model) shows that flat minima lead to robustness against the distribution gap…
A conceptual illustration of theoretical analysis. Theorem 1 (Corollary 1 for diffusion model) translates the perturbation in the parameter space into the set of perturbed distributions. Theorem 2 (Corollary 2 for diffusion model) shows that flat minima lead to robustness against the distribution gap. Credit: arXiv (2025). DOI: 10.48550/arxiv.2503.11078
When users ask ChatGPT to generate an image in a Ghibli style, the actual image is created by DALL-E, a tool powered by diffusion models. Although these models produce stunning images—such as transforming photos into artistic styles, creating personalized characters, or rendering realistic landscapes—they also face certain limitations. These include occasional errors, like three-fingered hands or distorted faces, and challenges in running on devices with limited computational resources, like smartphones, due to their massive number of parameters.
A research team, jointly led by Professors Jaejun Yoo and Sung Whan Yoon of the UNIST Graduate School of Artificial Intelligence at UNIST, has proposed a new design principle for generative AI that addresses these issues. They have shown, through both theoretical analysis and extensive experiments, that training diffusion models to reach “flat minima”—a specific type of optimal point on the loss surface—can simultaneously improve both the robustness and the generalization ability of these models.
Their study was presented at the International Conference on Computer Vision (ICCV 2025), and the findings are posted on the arXiv preprint server.
Diffusion models are widely used in popular AI applications, including tools like DALL-E and Stable Diffusion, enabling a range of tasks from style transfer and cartoon creation to realistic scene rendering. However, deploying these models often leads to challenges, such as error accumulation during short generation cycles, performance degradation after model compression techniques like quantization, and vulnerability to adversarial attacks—small, malicious input perturbations designed to deceive the models.
The research team identified that these issues stem from fundamental limitations in the models’ ability to generalize—meaning their capacity to perform reliably on new, unseen data or in unfamiliar environments.
To address this, the research team proposed guiding the training process toward “flat minima”—regions in the model’s loss landscape characterized by broad, gentle surfaces. Such minima help the model maintain stable and reliable performance despite small disturbances or noise. Conversely, “sharp minima”—narrow, steep valleys—tend to cause performance to deteriorate when faced with variations or attacks.
Among various algorithms designed to find flat minima, the team identified Sharpness-Aware Minimization (SAM) as the most effective. Models trained with SAM demonstrated reduced error accumulation during rapid generation tasks, maintained higher quality outputs after compression, and exhibited a sevenfold increase in resistance to adversarial attacks, significantly boosting their robustness.
While previous research addressed issues like error accumulation, quantization errors, and adversarial vulnerabilities separately, this study shows that focusing on flat minima offers a unified and fundamental solution to all these challenges.
The researchers highlight that their findings go beyond simply improving image quality. They provide a fundamental framework for designing trustworthy, versatile generative AI systems that can be effectively applied across various industries and real-world scenarios. Additionally, this approach could pave the way for training large-scale models like ChatGPT more efficiently, even with limited data.
More information: Taehwan Lee et al, Understanding Flatness in Generative Models: Its Role and Benefits, arXiv (2025). DOI: 10.48550/arxiv.2503.11078
Journal information: arXiv
Citation: Design principles for more reliable and trustworthy AI artists (2025, November 6) retrieved 6 November 2025 from https://techxplore.com/news/2025-11-principles-reliable-trustworthy-ai-artists.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.