ExplorerData ScienceMachine Learning
Research PaperResearchia:202603.16051

Fractals made Practical: Denoising Diffusion as Partitioned Iterated Function Systems

Ann Dooms

Abstract

What is a diffusion model actually doing when it turns noise into a photograph? We show that the deterministic DDIM reverse chain operates as a Partitioned Iterated Function System (PIFS) and that this framework serves as a unified design language for denoising diffusion model schedules, architectures, and training objectives. From the PIFS structure we derive three computable geometric quantities: a per-step contraction threshold $L^_t$, a diagonal expansion function $f_t(λ)$ and a global exp...

Submitted: March 16, 2026Subjects: Machine Learning; Data Science

Description / Details

What is a diffusion model actually doing when it turns noise into a photograph? We show that the deterministic DDIM reverse chain operates as a Partitioned Iterated Function System (PIFS) and that this framework serves as a unified design language for denoising diffusion model schedules, architectures, and training objectives. From the PIFS structure we derive three computable geometric quantities: a per-step contraction threshold LtL^*_t, a diagonal expansion function ft(λ)f_t(λ) and a global expansion threshold λλ^{**}. These quantities require no model evaluation and fully characterize the denoising dynamics. They structurally explain the two-regime behavior of diffusion models: global context assembly at high noise via diffuse cross-patch attention and fine-detail synthesis at low noise via patch-by-patch suppression release in strict variance order. Self-attention emerges as the natural primitive for PIFS contraction. The Kaplan-Yorke dimension of the PIFS attractor is determined analytically through a discrete Moran equation on the Lyapunov spectrum. Through the study of the fractal geometry of the PIFS, we derive three optimal design criteria and show that four prominent empirical design choices (the cosine schedule offset, resolution-dependent logSNR shift, Min-SNR loss weighting, and Align Your Steps sampling) each arise as approximate solutions to our explicit geometric optimization problems tuning theory into practice.


Source: arXiv:2603.13069v1 - http://arxiv.org/abs/2603.13069v1 PDF: https://arxiv.org/pdf/2603.13069v1 Original Link: http://arxiv.org/abs/2603.13069v1

Please sign in to join the discussion.

No comments yet. Be the first to share your thoughts!

Access Paper
View Source PDF
Submission Info
Date:
Mar 16, 2026
Topic:
Data Science
Area:
Machine Learning
Comments:
0
Bookmark
Fractals made Practical: Denoising Diffusion as Partitioned Iterated Function Systems | Researchia