We found a match
Your institution may have access to this item. Find your institution then sign in to continue.
- Title
Adversarial image generation by combining content and style.
- Authors
Liu, Songyan; Zhao, Chaoyang; Gao, Yunze; Wang, Jinqiao; Tang, Ming
- Abstract
Images can be considered as the combination of two parts: the content and the style. The authors' approach can leverage this property by extracting a certain unique style from the reference images and combining it to generate images with new contents. With a well‐defined style feature extraction module, they propose a novel framework to generate images with various styles and the same content. To train the style specific image generation model efficiently, a double‐cycle training strategy is proposed: they input two natural‐content pairs simultaneously, extract their style features, and exchange them twice to obtain the reconstruction of the input natural images. What is more, they apply the triplet margin loss to the style feature extracted from the images before and after style exchange and an adversarial discriminator to force the style‐exchanged images to be real. They perform experiments on licence‐plate image, Chinese characters, and shoes or handbags images generating, obtain photo‐realistic results and remarkably improve the corresponding supervised recognition task.
- Publication
IET Image Processing (Wiley-Blackwell), 2019, Vol 13, Issue 14, p2716
- ISSN
1751-9659
- Publication type
Article
- DOI
10.1049/iet-ipr.2019.0103