Denoising as Adaptation: Noise-Space Domain Adaptation for Image Restoration
Kang Liao, Zongsheng Yue, Zhouxia Wang, Chen Change Loy
TL;DR
The paper tackles the domain gap between synthetic and real degraded images in image restoration by introducing noise-space domain adaptation. It leverages a conditional diffusion model as a training proxy, jointly optimizing a restoration network with diffusion-based guidance to align outputs with a target clean distribution, while discarding the diffusion model after training. To prevent shortcut learning, it adds a channel-shuffling layer and a residual-swapping contrastive loss, ensuring robust, domain-level alignment without relying on easily distinguishable features. Across denoising, deraining, and deblurring, the approach outperforms both feature-space and pixel-space DA methods and scales across architectures, offering a stable, generalizable, and inference-efficient solution to real-world restoration tasks.
Abstract
Although learning-based image restoration methods have made significant progress, they still struggle with limited generalization to real-world scenarios due to the substantial domain gap caused by training on synthetic data. Existing methods address this issue by improving data synthesis pipelines, estimating degradation kernels, employing deep internal learning, and performing domain adaptation and regularization. Previous domain adaptation methods have sought to bridge the domain gap by learning domain-invariant knowledge in either feature or pixel space. However, these techniques often struggle to extend to low-level vision tasks within a stable and compact framework. In this paper, we show that it is possible to perform domain adaptation via the noise space using diffusion models. In particular, by leveraging the unique property of how auxiliary conditional inputs influence the multi-step denoising process, we derive a meaningful diffusion loss that guides the restoration model in progressively aligning both restored synthetic and real-world outputs with a target clean distribution. We refer to this method as denoising as adaptation. To prevent shortcuts during joint training, we present crucial strategies such as channel-shuffling layer and residual-swapping contrastive learning in the diffusion model. They implicitly blur the boundaries between conditioned synthetic and real data and prevent the reliance of the model on easily distinguishable features. Experimental results on three classical image restoration tasks, namely denoising, deblurring, and deraining, demonstrate the effectiveness of the proposed method.
