L2M-GAN: Learning to Manipulate Latent Space Semantics for Facial Attribute Editing

Abstract

A deep facial attribute editing model strives to meet two requirements: (1) attribute correctness -- the target attribute should correctly appear on the edited face image; (2) irrelevance preservation -- any irrelevant information (e.g., identity) should not be changed after editing. Meeting both requirements challenges the state-of-the-art works which resort to either spatial attention or latent space factorization. Specifically, the former assume that each attribute has well-defined local support regions; they are often more effective for editing a local attribute than a global one. The latter factorize the latent space of a fixed pretrained GAN into different attribute-relevant parts, but they cannot be trained end-to-end with the GAN, leading to sub-optimal solutions. To overcome these limitations, we propose a novel latent space factorization model, called L2M-GAN, which is learned end-to-end and effective for editing both local and global attributes. The key novel components are: (1) A latent space vector of the GAN is factorized into an attribute-relevant and irrelevant codes with an orthogonality constraint imposed to ensure disentanglement. (2) An attribute-relevant code transformer is learned to manipulate the attribute value; crucially, the transformed code are subject to the same orthogonality constraint. By forcing both the original attribute-relevant latent code and the edited code to be disentangled from any attribute-irrelevant code, our model strikes the perfect balance between attribute correctness and irrelevance preservation. Extensive experiments on CelebA-HQ show that our L2M-GAN achieves significant improvements over the state-of-the-arts.

Cite

Text

Yang et al. "L2M-GAN: Learning to Manipulate Latent Space Semantics for Facial Attribute Editing." Conference on Computer Vision and Pattern Recognition, 2021. doi:10.1109/CVPR46437.2021.00297

Markdown

[Yang et al. "L2M-GAN: Learning to Manipulate Latent Space Semantics for Facial Attribute Editing." Conference on Computer Vision and Pattern Recognition, 2021.](https://mlanthology.org/cvpr/2021/yang2021cvpr-l2mgan/) doi:10.1109/CVPR46437.2021.00297

BibTeX

@inproceedings{yang2021cvpr-l2mgan,
  title     = {{L2M-GAN: Learning to Manipulate Latent Space Semantics for Facial Attribute Editing}},
  author    = {Yang, Guoxing and Fei, Nanyi and Ding, Mingyu and Liu, Guangzhen and Lu, Zhiwu and Xiang, Tao},
  booktitle = {Conference on Computer Vision and Pattern Recognition},
  year      = {2021},
  pages     = {2951-2960},
  doi       = {10.1109/CVPR46437.2021.00297},
  url       = {https://mlanthology.org/cvpr/2021/yang2021cvpr-l2mgan/}
}