Generating human faces by generative adversarial networks

Over the years, computer vision improves significantly. From recognising and understanding what lies underneath an image, we can now generate images by modelling training distribution using generative adversarial network(GAN). Since then, researchers come out with various variants of GAN and ways to...

Full description

Bibliographic Details
Main Author: Quek, Chin Wei
Other Authors: Chen Change Loy
Format: Final Year Project (FYP)
Language:English
Published: Nanyang Technological University 2020
Subjects:
Online Access:https://hdl.handle.net/10356/139259
Description
Summary:Over the years, computer vision improves significantly. From recognising and understanding what lies underneath an image, we can now generate images by modelling training distribution using generative adversarial network(GAN). Since then, researchers come out with various variants of GAN and ways to stabalize GAN training. This results in improved quality of generated image. The application of GAN has sparked the interest of many people. In this project, we first analyse the use of StarGAN, a unified generative adversarial network for multi-domain image-to-image translation task to generate human facial expressions. We also explore the possible use of StarGAN in cartoon character facial expression generation and video generation.