Abstract
We expand the scope of image-to-image translation to include more distinct image domains, where the image sets have analogous structures, but may not share object types between them. Semantic-Driven Unsupervised Image-to-Image Translation for Distinct Image Domains (SUNIT) is built to more successfully translate images in this setting, where content from one domain is not found in the other. Our method trains an image translation model by learning encodings for semantic segmentations of images. These segmentations are translated between image domains to learn meaningful mappings between the structures in the two domains. The translated segmentations are then used as the basis for image generation. Beginning image generation with encoded segmentation information helps maintain the original structure of the image. We qualitatively and quantitatively show that SUNIT improves image translation outcomes, especially for image translation tasks where the image domains are very distinct.
Degree
MS
College and Department
Physical and Mathematical Sciences; Computer Science
Rights
https://lib.byu.edu/about/copyright/
BYU ScholarsArchive Citation
Ackerman, Wesley, "Semantic-Driven Unsupervised Image-to-Image Translation for Distinct Image Domains" (2020). Theses and Dissertations. 8684.
https://scholarsarchive.byu.edu/etd/8684
Date Submitted
2020-09-15
Document Type
Thesis
Handle
http://hdl.lib.byu.edu/1877/etd11429
Keywords
computer science, machine learning, image-to-image translation, generative adversarial network, deep learning, unsupervised learning, convolutional neural network
Language
english