This document summarizes a research paper on content-aware style transfer using deep learning. It proposes two improvements: 1) Using semantic masks to specify where in an image to transfer style (e.g. transferring a dog's appearance without changing the background). 2) Defining style features using high-order statistics to better preserve content coherence when transferring style. An example is shown where the facial features of two different dog breeds are successfully exchanged while retaining the background. The full paper presents more results and derives the gradient formulation for the proposed generalized style transfer method.
This document summarizes a research paper on content-aware style transfer using deep learning. It proposes two improvements: 1) Using semantic masks to specify where in an image to transfer style (e.g. transferring a dog's appearance without changing the background). 2) Defining style features using high-order statistics to better preserve content coherence when transferring style. An example is shown where the facial features of two different dog breeds are successfully exchanged while retaining the background. The full paper presents more results and derives the gradient formulation for the proposed generalized style transfer method.
This document summarizes a research paper on content-aware style transfer using deep learning. It proposes two improvements: 1) Using semantic masks to specify where in an image to transfer style (e.g. transferring a dog's appearance without changing the background). 2) Defining style features using high-order statistics to better preserve content coherence when transferring style. An example is shown where the facial features of two different dog breeds are successfully exchanged while retaining the background. The full paper presents more results and derives the gradient formulation for the proposed generalized style transfer method.
http://www.cs.nthu.edu.tw/~cthsu/candy.html National Tsing Hua University Hsinchu, Taiwan
Motivation feature aggregation, to specify the spatial
correspondence. To constrain what to transfer, Recently, it has been shown that one can invert we propose a new feature statistics & = T a deep convolutional neural network originally ' ' , by introducing a high- trained for classification tasks to transfer image order convolutional matrix ' , to better match style. There is, however, a dearth of research on the style representation. Finally, we propose to content-aware style transfer. In this paper, we embed both two constraints into the style loss of generalize the original neural algorithm [1] for Equation (1) and derive the layer-wise gradient style transfer from two perspectives: where to in a general form: transfer and what to transfer. To specify where ∇ = ∑4*23 ∑1,23 ' * + , + - , ' * .- & , −& , /0 , ., ) to transfer, we propose a simple yet effective , , * , * masking out strategy to constrain the transfer where & = ' + ' . (2) layout. To illustrate what to transfer, we define a new style feature by high-order statistics to better characterize content coherency. Results
Methodology We show an example for real-life photo
transfer in Fig. 1. Using the semantic masks estimated by image matting, we successfully Given a source image (or content image) transfer the dogs’ appearance without either and a target image (or style image) , [1] aims changing background or producing noticeable to synthesize an image which simultaneously artifacts. Please refer to our paper for more style shares the visual content of and the style transfer results representation of . Specifically, the image rendering was modelled as an optimization problem by minimizing the difference between and and the difference between and in terms of content and style features, respectively. The authors characterize both features by the deep convolutional neural network (CNN). The desired image was obtained by = arg min ∑ ∈ − +∑ ∈ − + Γ , (1) where the content feature is the layer-wise T response and the style feature = Figure 1: An example of content-aware style encodes cross-feature dependencies globally. transfer. The face appearance of two different We formulate the generalized style transfer breeds of dogs, Maltese and Yorkshire terrier, based on Equation (1) under two additional are exchanged by our method. constraints: where to transfer and what to [1] L. A. Gatys, A. S. Ecker, and M. Bethge. Image transfer. To constrain where to transfer, we style transfer using convolutional neural networks. introduce a diagonal matrix , whose , In Proc. CVPR, 2016. entry ! 0 ≤ ! ≤ 1 ) is a soft indictor of