Fast photographic style transfer based on convolutional neural networks

This data was imported from Scopus:

Authors: Wang, L., Xiang, N., Yang, X. and Zhang, J.

http://eprints.bournemouth.ac.uk/32095/

Journal: ACM International Conference Proceeding Series

Pages: 67-76

ISBN: 9781450364010

DOI: 10.1145/3208159.3208165

© 2018 ACM. The techniques for photographic style transfer have been researched for a long time, which explores effective ways to transfer the style features of a reference photo onto another content photograph. Recent works based on convolutional neural networks present an effective solution for style transfer, especially for paintings. The artistic style transformation results are visually appealing, however, the photorealism is lost because of content-mismatching and distortions even when both input images are photographic. To tackle this challenge, this paper introduces a similarity loss function and a refinement method into the style transfer network. The similarity loss function can solve the content-mismatching problem, however, the distortion and noise artefacts may still exist in the stylized results due to the content-style trade-off. Hence, we add a post-processing refinement step to reduce the artefacts. The robustness and effectiveness of our approach has been evaluated through extensive experiments which show that our method can obtain finer content details and less artefacts than state-of-the-art methods, and transfer style faithfully. In addition, our approach is capable of processing photographic style transfer in almost real-time, which makes it a potential solution for video style transfer.

The data on this page was last updated at 05:24 on October 24, 2020.