TY - GEN
T1 - SEAN: Image Synthesis With Semantic Region-Adaptive Normalization
AU - Zhu, Peihao
AU - Abdal, Rameen
AU - Qin, Yipeng
AU - Wonka, Peter
N1 - KAUST Repository Item: Exported on 2020-10-01
PY - 2020
Y1 - 2020
N2 - We propose semantic region-adaptive normalization (SEAN), a simple but effective building block for Generative Adversarial Networks conditioned on segmentation masks that describe the semantic regions in the desired output image. Using SEAN normalization, we can build a network architecture that can control the style of each semantic region individually, e.g., we can specify one style reference image per region. SEAN is better suited to encode, transfer, and synthesize style than the best previous method in terms of reconstruction quality, variability, and visual quality. We evaluate SEAN on multiple datasets and report better quantitative metrics (e.g. FID, PSNR) than the current state of the art. SEAN also pushes the frontier of interactive image editing. We can interactively edit images by changing segmentation masks or the style for any given region. We can also interpolate styles from two reference images per region.
AB - We propose semantic region-adaptive normalization (SEAN), a simple but effective building block for Generative Adversarial Networks conditioned on segmentation masks that describe the semantic regions in the desired output image. Using SEAN normalization, we can build a network architecture that can control the style of each semantic region individually, e.g., we can specify one style reference image per region. SEAN is better suited to encode, transfer, and synthesize style than the best previous method in terms of reconstruction quality, variability, and visual quality. We evaluate SEAN on multiple datasets and report better quantitative metrics (e.g. FID, PSNR) than the current state of the art. SEAN also pushes the frontier of interactive image editing. We can interactively edit images by changing segmentation masks or the style for any given region. We can also interpolate styles from two reference images per region.
UR - http://hdl.handle.net/10754/664682
UR - https://ieeexplore.ieee.org/document/9156510/
U2 - 10.1109/CVPR42600.2020.00515
DO - 10.1109/CVPR42600.2020.00515
M3 - Conference contribution
SN - 978-1-7281-7169-2
BT - 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)
PB - IEEE
ER -