Unsupervised Object-Level Image-to-Image Translation Using Positional Attention Bi-Flow Generative Network

Recent work in unsupervised image-to-image translation by adversarially learning mapping between different domains, which cannot distinguish the foreground and background. The existing methods of image-to-image translation mainly transfer the global image across the source and target domains. Howeve...

Full description

Bibliographic Details
Main Authors: Liuchun Yuan, Dihu Chen, Haifeng Hu
Format: Article
Language:English
Published: IEEE 2019-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/8662557/
id doaj-bd5500f2bece43758d5431d74e68fcbc
record_format Article
spelling doaj-bd5500f2bece43758d5431d74e68fcbc2021-03-29T22:18:28ZengIEEEIEEE Access2169-35362019-01-017306373064710.1109/ACCESS.2019.29035438662557Unsupervised Object-Level Image-to-Image Translation Using Positional Attention Bi-Flow Generative NetworkLiuchun Yuan0Dihu Chen1https://orcid.org/0000-0001-5432-8149Haifeng Hu2https://orcid.org/0000-0002-4884-323XSchool of Electronic and Information Technology, Sun Yat-sen University, Guangzhou, ChinaSchool of Electronic and Information Technology, Sun Yat-sen University, Guangzhou, ChinaSchool of Electronic and Information Technology, Sun Yat-sen University, Guangzhou, ChinaRecent work in unsupervised image-to-image translation by adversarially learning mapping between different domains, which cannot distinguish the foreground and background. The existing methods of image-to-image translation mainly transfer the global image across the source and target domains. However, it is evident that not all regions of images should be transferred because forcefully transferring the unnecessary part leads to some unrealistic translations. In this paper, we present a positional attention bi-flow generative network, focusing our translation model on an interesting region or object in the image. We assume that the image representation can be decomposed into three parts: image-content, image-style, and image-position features. We apply an encoder to extract these features and bi-flow generator with attention module to achieve the translation task in an end-to-end manner. To realize the object-level translation, we adopt the image-position features to label the common interesting region between the source and target domains. We analyze the proposed framework and provide qualitative and quantitative comparisons. The extensive experiments validate that our proposed model is qualified to accomplish the object-level translation and obtain compelling results with other state-of-the-art approaches.https://ieeexplore.ieee.org/document/8662557/Image-to-image translationattention mechanismGANs
collection DOAJ
language English
format Article
sources DOAJ
author Liuchun Yuan
Dihu Chen
Haifeng Hu
spellingShingle Liuchun Yuan
Dihu Chen
Haifeng Hu
Unsupervised Object-Level Image-to-Image Translation Using Positional Attention Bi-Flow Generative Network
IEEE Access
Image-to-image translation
attention mechanism
GANs
author_facet Liuchun Yuan
Dihu Chen
Haifeng Hu
author_sort Liuchun Yuan
title Unsupervised Object-Level Image-to-Image Translation Using Positional Attention Bi-Flow Generative Network
title_short Unsupervised Object-Level Image-to-Image Translation Using Positional Attention Bi-Flow Generative Network
title_full Unsupervised Object-Level Image-to-Image Translation Using Positional Attention Bi-Flow Generative Network
title_fullStr Unsupervised Object-Level Image-to-Image Translation Using Positional Attention Bi-Flow Generative Network
title_full_unstemmed Unsupervised Object-Level Image-to-Image Translation Using Positional Attention Bi-Flow Generative Network
title_sort unsupervised object-level image-to-image translation using positional attention bi-flow generative network
publisher IEEE
series IEEE Access
issn 2169-3536
publishDate 2019-01-01
description Recent work in unsupervised image-to-image translation by adversarially learning mapping between different domains, which cannot distinguish the foreground and background. The existing methods of image-to-image translation mainly transfer the global image across the source and target domains. However, it is evident that not all regions of images should be transferred because forcefully transferring the unnecessary part leads to some unrealistic translations. In this paper, we present a positional attention bi-flow generative network, focusing our translation model on an interesting region or object in the image. We assume that the image representation can be decomposed into three parts: image-content, image-style, and image-position features. We apply an encoder to extract these features and bi-flow generator with attention module to achieve the translation task in an end-to-end manner. To realize the object-level translation, we adopt the image-position features to label the common interesting region between the source and target domains. We analyze the proposed framework and provide qualitative and quantitative comparisons. The extensive experiments validate that our proposed model is qualified to accomplish the object-level translation and obtain compelling results with other state-of-the-art approaches.
topic Image-to-image translation
attention mechanism
GANs
url https://ieeexplore.ieee.org/document/8662557/
work_keys_str_mv AT liuchunyuan unsupervisedobjectlevelimagetoimagetranslationusingpositionalattentionbiflowgenerativenetwork
AT dihuchen unsupervisedobjectlevelimagetoimagetranslationusingpositionalattentionbiflowgenerativenetwork
AT haifenghu unsupervisedobjectlevelimagetoimagetranslationusingpositionalattentionbiflowgenerativenetwork
_version_ 1724191891595984896