One Noise to Rule Them All:
Learning a Unified Model of Spatially-Varying Noise Patterns

Transactions on Graphics (Proceedings of SIGGRAPH 2024)

Arman Maesumi
Brown University
Dylan
Hu

Brown University
Krishi
Saripalli

Brown University
Vladimir G. Kim
Adobe Research
Matthew Fisher
Adobe Research
Sören Pirk
CAU
Daniel Ritchie
Brown University

[Paper, 34mb] [Supp., 12mb] [arXiv] [Bibtex]

Code release coming soon!

Abstract

Procedural noise is a fundamental component of computer graphics pipelines, offering a flexible way to generate textures that exhibit "natural" random variation. Many different types of noise exist, each produced by a separate algorithm. In this paper, we present a single generative model which can learn to generate multiple types of noise as well as blend between them. In addition, it is capable of producing spatially-varying noise blends despite not having access to such data for training. These features are enabled by training a denoising diffusion model using a novel combination of data augmentation and network conditioning techniques. Like procedural noise generators, the model's behavior is controllable via interpretable parameters plus a source of randomness. We use our model to produce a variety of visually compelling noise textures. We also present an application of our model to improving inverse procedural material design; using our model in place of fixed-type noise nodes in a procedural material graph results in higher-fidelity material reconstructions without needing to know the type of noise in advance.



High-level Summary

Goal: Learn a denoising diffusion model that can produce multiple types of parametrized noise functions and blend between them spatially.

Problem: Training a denoising U-Net without spatially-varying noise data results in poor adherence to the conditioning signal.

Insights:

  1. The U-Net's bottleneck architecture causes a smearing effect on the network's learned features.

  2. Spatially-varying synthesis requires the U-Net to denoise along multiple denoising trajectories simultaneously, which is difficult in the absence of spatially-varying noise data.

Solution: We propose a novel variant of CutMix data augmentation, which creates synthetic training samples that are collage-like amalgamations of multiple noise types. This forces the U-Net to respect the spatially-varying conditioning signal, and improves its ability to denoise along multiple trajectories.


Figure 3. During training, we first transform the current data sample (highlighted in blue) by cutting and patching together a set of other random samples from the dataset, resulting in a training image x0. The noise parameters for each image patch are passed to an MLP, which projects the parameter sets into an embedding space that encodes both the noise type (class) and the noise parameters. The resulting feature vectors are tiled to form a feature grid, which is used as a conditioning signal in the U-Net's SPADE blocks.


Figure 4. At inference time, we query our network using artificially constructed feature grids, enabling a flexible way to synthesize spatially-varying noise patterns. Here we embed four sets of noise parameters, pictorially shown as one of four colors. We blend the feature vectors using bilinear interpolation, creating a smoothly-varying feature grid, which our U-Net is able to transform into a Voronoi noise pattern with non-uniform scale and distortion characteristics.


Results














Inverse Procedural Material Design

We demonstrate that our model can be used as a differentiable proxy for noise generator nodes in a procedural material graph. Using our model in place of fixed-type noise nodes results in higher-fidelity material reconstructions.

Figure 7. Given a procedural material graph and a target photograph, our noise generative model serves as a useful prior over the space of noise functions, facilitating the recovery of non-trivial patterns that are present in the target photos, and improving the baseline result given by MATch. Two similarity scores are reported for each result with respect to the target image: MATch's feature-based texture similarity metric (left) and LPIPS (right). Lower is better for both scores.

Figure 10. After optimization, we can easily modify the resulting procedural material graphs. Here a marble material graph is optimized (first row), and edited in various ways (second row). We show the results of editing the noise colorization, resampling our diffusion noise, and modifying our model's conditioning inputs.



Paper and Supplementary Material

A. Maesumi, D. Hu, K. Saripalli, V. Kim, M. Fisher, S. Pirk, D. Ritchie
One Noise to Rule Them All:
Learning a Unified Model of Spatially-Varying Noise Patterns.

In ACM Transactions on Graphics (Proceedings of SIGGRAPH 2024).
(hosted on arXiv)

Bibtex:
@article{maesumi2024noise, author = {Maesumi, Arman and Hu, Dylan and Saripalli, Krishi and Kim, Vladimir G. and Fisher, Matthew and Pirk, Sören and Ritchie, Daniel}, title = {One Noise to Rule Them All: Learning a Unified Model of Spatially-Varying Noise Patterns}, year = {2024}, booktitle = {ACM Transactions on Graphics (Proceedings of SIGGRAPH 2024)}, publisher = {Association for Computing Machinery} }

Acknowledgements

This material is based upon work that was supported by the National Science Foundation Graduate Research Fellowship under Grant No. 2040433. Part of this work was done while Arman Maesumi was an intern at Adobe Research. This project page template was originally made by Phillip Isola and Richard Zhang.