Artistic Style-Transfer
Artistic Style-Transfer
Output
Description:
It has been shown that a convolutional neural net algorithm can distinguish between the content
of a piece of art and the style that it was painted in ([1] Leon Gatys). For our final project, we would like to explore this idea in the context of image processing. Specifically, we want to create an image filter that transfers the style of one painting to the content of another image. A clear motivation for finding a solution is for applications that handle many images at a time or for applications that cannot tolerate the training/learning time overhead that comes with the CNN([2] Michael Elad).
Proposed Methods:
0. Define style transfer from style of image 2 to content of image 1
1. Extract the “content” from image 1
a. This may be achieved by feature detection (edges, corners, lines, circles, uniform
b. Rank which features are more important to the scene overall - some metrics that
may be considered are nominal size of the feature, centricity, brightness, color,
contrast, etc
2. Extract the “content” from image 2
3. Extract the “style” from image 2
a. This may be accomplished by segmenting the image and trying to find key similarities between different patches i.e. themes throughout the image
b. Map styles to features
4. Find sufficiently matched areas of “content” between image 1 and 2 and apply the appropriate “style” that fits the content given : its ranked importance in the scene, how similar it is to the “content” of image 2,
5. Fill in empty spaces and stitch / blend the various areas of the image
References:
[1] Leon A. Gatys, Alexander S. Ecker,Matthias Bethge. “A Neural Algorithm of Artistic Style.”
https://arxiv.org/pdf/1508.06576.pdf
[2] Michael Elad and Peyman Milanfar “Style-Transfer via Texture-Synthesis” arXiv:1609.03057
[3] Wei Zhang, Chen Cao, Shifeng Chen, Jianzhuang Liu, Senior Member, IEEE, and Xiaoou
Tang, Fellow, IEEE “Style Transfer Via Image Component Analysis"
ieeexplore.ieee.org/iel7/6046/6630084/06522845.pdf
[4] V. Kwatra, I. Essa, A. Bobick, and N. Kwatra, “Texture Optimization for Example-Based
Synthesis”, ACM ToG, Vol. 24, No. 3, pp. 795-802, 2005.
FOR BASE PAPER PLEASE MAIL US
DOWNLOAD SOURCE CODE CLICK HERE
Output
Description:
It has been shown that a convolutional neural net algorithm can distinguish between the content
of a piece of art and the style that it was painted in ([1] Leon Gatys). For our final project, we would like to explore this idea in the context of image processing. Specifically, we want to create an image filter that transfers the style of one painting to the content of another image. A clear motivation for finding a solution is for applications that handle many images at a time or for applications that cannot tolerate the training/learning time overhead that comes with the CNN([2] Michael Elad).
Proposed Methods:
0. Define style transfer from style of image 2 to content of image 1
1. Extract the “content” from image 1
a. This may be achieved by feature detection (edges, corners, lines, circles, uniform
b. Rank which features are more important to the scene overall - some metrics that
may be considered are nominal size of the feature, centricity, brightness, color,
contrast, etc
2. Extract the “content” from image 2
3. Extract the “style” from image 2
a. This may be accomplished by segmenting the image and trying to find key similarities between different patches i.e. themes throughout the image
b. Map styles to features
4. Find sufficiently matched areas of “content” between image 1 and 2 and apply the appropriate “style” that fits the content given : its ranked importance in the scene, how similar it is to the “content” of image 2,
5. Fill in empty spaces and stitch / blend the various areas of the image
References:
[1] Leon A. Gatys, Alexander S. Ecker,Matthias Bethge. “A Neural Algorithm of Artistic Style.”
https://arxiv.org/pdf/1508.06576.pdf
[2] Michael Elad and Peyman Milanfar “Style-Transfer via Texture-Synthesis” arXiv:1609.03057
[3] Wei Zhang, Chen Cao, Shifeng Chen, Jianzhuang Liu, Senior Member, IEEE, and Xiaoou
Tang, Fellow, IEEE “Style Transfer Via Image Component Analysis"
ieeexplore.ieee.org/iel7/6046/6630084/06522845.pdf
[4] V. Kwatra, I. Essa, A. Bobick, and N. Kwatra, “Texture Optimization for Example-Based
Synthesis”, ACM ToG, Vol. 24, No. 3, pp. 795-802, 2005.
FOR BASE PAPER PLEASE MAIL US
DOWNLOAD SOURCE CODE CLICK HERE
Comments
Post a Comment