Hier finden Sie wissenschaftliche Publikationen aus den Fraunhofer-Instituten.

Benign examples: Imperceptible changes can enhance image translation performance

: Srinivasan, V.; Müller, K.-R.; Samek, W.; Nakajima, S.


Association for the Advancement of Artificial Intelligence -AAAI-:
Thirty-Fourth AAAI Conference on Artificial Intelligence 2020. Proceedings : Thirty-Second Conference on Innovative Applications of Artificial Intelligence, the Tenth Symposium on Educational Advances in Artificial Intelligence, February 7-12, 2020, New York, New York, USA
Menlo Park: AAAI Press, 2020 (AAAI Technical Tracks 34.2020)
ISBN: 978-1-57735-835-0
Conference on Artificial Intelligence (AAAI) <34, 2020, New York/NY>
Conference on Innovative Applications of Artificial Intelligence (IAAI) <32, 2020, New York/NY>
Symposium on Educational Advances in Artificial Intelligence (EAAI) <10, 2020, New York/NY>
Fraunhofer HHI ()

Unpaired image-to-image domain translation involves the task of transferring an image in one domain to another domain without having pairs of data for supervision. Several methods have been proposed to address this task using Generative Adversarial Networks (GANs) and cycle consistency constraint enforcing the translated image to be mapped back to the original domain. This way, a Deep Neural Network (DNN) learns mapping such that the input training distribution transferred to the target domain matches the target training distribution. However, not all test images are expected to fall inside the data manifold in the input space where the DNN has learned to perform the mapping very well. Such images can have a poor mapping to the target domain. In this paper, we propose to perform Langevin dynamics, which makes a subtle change in the input space bringing them close to the data manifold, producing benign examples. The effect is significant improvement of the mapped image o n the target domain. We also show that the score function estimation by denoising autoencoder (DAE), can practically be replaced with any autoencoding structure, which most image-to-image translation methods contain intrinsically due to the cycle consistency constraint. Thus, no additional training is required. We show advantages of our approach for several state-of-the-art image-to-image domain translation models. Quantitative evaluation shows that our proposed method leads to a substantial increase in the accuracy to the target label on multiple state-of-the-art image classifiers, while qualitative user study proves that our method better represents the target domain, achieving better human preference scores.