VISTEC
Rayong, Thailand
Phranakhon Rajabhat University
Bangkok, Thailand
CVPR 2023
StyleGAN Salon Our method can transfer the hairstyle from any reference hair image in the top row to Tom Holland, in the second row.
Our paper seeks to transfer the hairstyle of a reference image to an input photo for virtual hair try-on. We target a variety of challenges scenarios, such as transforming a long hairstyle with bangs to a pixie cut, which requires removing the existing hair and inferring how the forehead would look, or transferring partially visible hair from a hat-wearing person in a different pose. Past solutions leverage StyleGAN for hallucinating any missing parts and producing a seamless face-hair composite through so-called GAN inversion or projection. However, there remains a challenge in controlling the hallucinations to accurately transfer hairstyle and preserve the face shape and identity of the input. To overcome this, we propose a multi-view optimization framework that uses "two different views" of reference composites to semantically guide occluded or ambiguous regions. Our optimization shares information between two poses, which allows us to produce high fidelity and realistic results from incomplete references. Our framework produces high-quality results and outperforms prior work in a user study that consists of significantly more challenging hair transfer scenarios than previously studied.
We first align the input face If and reference hair Ih and use them to construct guide images Iguide, in two different viewpoints, which specifies the target appearance for each output region.
Real images
FFHQ
User study results on hairstyle transfer (percentage of user preferring each method). Our method outperforms state-of-the-art hairstyle transfer methods on FFHQ datasets in all challenging scenarios. A total of 450 pairs are used in this study, 150 pairs in FFHQ-P and 300 in FFHQ-S. For each pair, we asked 3 unique participants to select the best result for hairstyle transfer.
User study on pose-invariant hairstyle transfer. Our method outperforms others on all pose difference ranges.
BibTex
@inproceedings{Khwanmuang2023StyleGANSalon, author = {Khwanmuang, Sasikarn and Phongthawee, Pakkapon and Sangkloy, Patsorn and Suwajanakorn, Supasorn}, title = {StyleGAN Salon: Multi-View Latent Optimization for Pose-Invariant Hairstyle Transfer}, booktitle = {IEEE Conference on Computer Vision and Pattern Recognition (CVPR)}, year = {2023}, }