Text2FaceGAN: Face Generation from Fine Grained Textual Descriptions
Autor: | Yi Yu, Rajiv Ratn Shah, Ajit Kumar, Shailesh Kumar Jha, Manraj Singh Grover, Osaid Rehman Nasir |
---|---|
Rok vydání: | 2019 |
Předmět: |
FOS: Computer and information sciences
Computer Science - Machine Learning Computer science Computer Vision and Pattern Recognition (cs.CV) Computer Science - Computer Vision and Pattern Recognition Machine Learning (stat.ML) 02 engineering and technology 010501 environmental sciences Space (commercial competition) computer.software_genre 01 natural sciences Machine Learning (cs.LG) Domain (software engineering) Set (abstract data type) Statistics - Machine Learning FOS: Electrical engineering electronic engineering information engineering 0202 electrical engineering electronic engineering information engineering 0105 earth and related environmental sciences business.industry Image and Video Processing (eess.IV) Electrical Engineering and Systems Science - Image and Video Processing Multimedia (cs.MM) Task (computing) Face (geometry) Metric (mathematics) Task analysis 020201 artificial intelligence & image processing Noise (video) Artificial intelligence business computer Computer Science - Multimedia Natural language processing |
Zdroj: | BigMM |
DOI: | 10.1109/bigmm.2019.00-42 |
Popis: | In recent years, powerful generative adversarial networks (GAN) have been developed to automatically synthesize realistic images from text. However, most existing tasks are limited to generating simple images such as flowers from captions. In this paper, we extend this problem to the less addressed domain of face generation from fine-grained textual descriptions of face, e.g., "A person has curly hair, oval face, and mustache". We are motivated by the potential of automated face generation to impact and assist critical tasks such as criminal face reconstruction. Since current datasets for the task are either very small or do not contain captions, we generate captions for images in the CelebA dataset by creating an algorithm to automatically convert a list of attributes to a set of captions. The generated captions are meaningful, versatile and consistent with the general semantics of a face. We then model the highly multi-modal problem of text to face generation as learning the conditional distribution of faces (conditioned on text) in same latent space. We utilize the current state-of-the-art GAN (DC-GAN with GAN-CLS loss) for learning conditional multi-modality. The presence of more fine-grained details and variable length of the captions makes the problem easier for a user but more difficult to handle compared to the other text-to-image tasks. We flipped the labels for real and fake images and added noise in discriminator. Generated images for diverse textual descriptions show promising results. In the end, we show how the widely used inceptions score is not a good metric to evaluate the performance of generative models used for synthesizing faces from text. |
Databáze: | OpenAIRE |
Externí odkaz: |