Diffusion Models with Implicit Conditions Driven by Latent Shifts
Citations

WEB OF SCIENCE

0
Citations

SCOPUS

0

초록

This paper introduces a diffusion model that replaces the target prior distribution from a standard Gaussian to non-zero-mean Gaussian priors, with shifted latent trajectories determined by non-zero-mean Gaussian noise, thereby maintaining the closed-form of conventional diffusion models. Unlike conventional conditional models such as conditional denoising diffusion probabilistic models or classifier guidance, the proposed model implicitly aligns each class with a unique component in the Gaussian priors. Carefully devised positioning strategies uniformly distribute Gaussian components without introducing additional learnable parameters, which is essential to implicit-only conditioning on our model. Qualitative and quantitative experiments demonstrate that, even without additional conditioning layers or a classifier, the proposed framework achieves performance comparable to explicit methods in terms of Fr & eacute;chet Inception Distance, precision, and recall. The framework is further validated under unsupervised settings by replacing class labels with pseudo-labels generated from K-means clustering on feature representations.

키워드

Clustering algorithmsComputer visionDeep learningDiffusion modelsGaussian distributionGaussian mixture modelGenerative AINeural networksRepresentation learningUnsupervised learning
제목
Diffusion Models with Implicit Conditions Driven by Latent Shifts
저자
Lee, Da EunNakamura, KensukeHong, Byung-Woo
DOI
10.1109/ACCESS.2025.3603215
발행일
2025-09
유형
Article
저널명
IEEE Access
13
페이지
152651 ~ 152668