Photorealistic Material Editing Through Direct Image Manipulation
Name:
Preprintfile1.pdf
Size:
8.962Mb
Format:
PDF
Description:
Pre-print
Embargo End Date:
2021-07-20
Type
ArticleKAUST Department
Computer Science ProgramVisual Computing Center (VCC)
Computer, Electrical and Mathematical Sciences and Engineering (CEMSE) Division
Date
2020-07-20Preprint Posting Date
2019-09-12Online Publication Date
2020-07-20Print Publication Date
2020-07Embargo End Date
2021-07-20Permanent link to this record
http://hdl.handle.net/10754/660691
Metadata
Show full item recordAbstract
Creating photorealistic materials for light transport algorithms requires carefully fine-tuning a set of material properties to achieve a desired artistic effect. This is typically a lengthy process that involves a trained artist with specialized knowledge. In this work, we present a technique that aims to empower novice and intermediate-level users to synthesize high-quality photorealistic materials by only requiring basic image processing knowledge. In the proposed workflow, the user starts with an input image and applies a few intuitive transforms (e.g., colorization, image inpainting) within a 2D image editor of their choice, and in the next step, our technique produces a photorealistic result that approximates this target image. Our method combines the advantages of a neural network-augmented optimizer and an encoder neural network to produce high-quality output results within 30 seconds. We also demonstrate that it is resilient against poorly-edited target images and propose a simple extension to predict image sequences with a strict time budget of 1–2 seconds per image.Citation
Zsolnai-Fehér, K., Wonka, P., & Wimmer, M. (2020). Photorealistic Material Editing Through Direct Image Manipulation. Computer Graphics Forum, 39(4), 107–120. doi:10.1111/cgf.14057Sponsors
We would like to thank Reynante Martinez for providing us the geometry and some of the materials for the Paradigm (Fig. 1) and Genesis scenes (Fig. 3), ianofshields for the Liquify scene that served as a basis for Fig. 9, Robin Marin for the material test scene, Andrew Price and Gábor Mészáros for their help with geometry modeling, Felícia Zsolnai-Fehér for her help improving our figures, Christian Freude, David Ha, Philipp Erler and Adam Celarek for their useful comments. We also thank the anonymous reviewers for their help improving this manuscript and NVIDIA for providing the hardware to train our neural networks. This work was partially funded by Austrian Science Fund (FWF), project number P27974.Publisher
WileyJournal
Computer Graphics ForumarXiv
1909.11622Additional Links
https://onlinelibrary.wiley.com/doi/abs/10.1111/cgf.14057ae974a485f413a2113503eed53cd6c53
10.1111/cgf.14057