Loading...
|
Please use this identifier to cite or link to this item:
https://tkuir.lib.tku.edu.tw/dspace/handle/987654321/122970
|
Title: | Sketch-guided Deep Portrait Generation |
Authors: | Ho, Trang-Thi;Virtusio, John Jethro;Chen, Yung-Yao;Hsu, Chih-Ming;Hua, and Kai-Lung |
Date: | 2020-07-05 |
Issue Date: | 2023-04-28 16:33:06 (UTC+8) |
Publisher: | ACM New York, NY, USA |
Abstract: | Generating a realistic human class image from a sketch is a unique and challenging problem considering that the human body has a complex structure that must be preserved. Additionally, input sketches often lack important details that are crucial in the generation process, hence making the problem more complicated. In this article, we present an effective method for synthesizing realistic images from human sketches. Our framework incorporates human poses corresponding to locations of key semantic components (e.g., arm, eyes, nose), seeing that its a strong prior for generating human class images. Our sketch-image synthesis framework consists of three stages: semantic keypoint extraction, coarse image generation, and image refinement. First, we extract the semantic keypoints using Part Affinity Fields (PAFs) and a convolutional autoencoder. Then, we integrate the sketch with semantic keypoints to generate a coarse image of a human. Finally, in the image refinement stage, the coarse image is enhanced by a Generative Adversarial Network (GAN) that adopts an architecture carefully designed to avoid checkerboard artifacts and to generate photo-realistic results. We evaluate our method on 6,300 sketch-image pairs and show that our proposed method generates realistic images and compares favorably against state-of-the-art image synthesis methods. |
Relation: | ACM Transactions on Multimedia Computing, Communications, and Applications (TOMM) 16.3, p.1-18 |
DOI: | 10.1145/3396237 |
Appears in Collections: | [資訊工程學系暨研究所] 期刊論文
|
Files in This Item:
File |
Description |
Size | Format | |
index.html | | 0Kb | HTML | 28 | View/Open |
|
All items in 機構典藏 are protected by copyright, with all rights reserved.
|