Skip to main content
Browse Subject Areas

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Guided neural style transfer for shape stylization


Designing logos, typefaces, and other decorated shapes can require professional skills. In this paper, we aim to produce new and unique decorated shapes by stylizing ordinary shapes with machine learning. Specifically, we combined parametric and non-parametric neural style transfer algorithms to transfer both local and global features. Furthermore, we introduced a distance-based guiding to the neural style transfer process, so that only the foreground shape will be decorated. Lastly, qualitative evaluation and ablation studies are provided to demonstrate the usefulness of the proposed method.


Designing decorated shapes (e.g. logos and typefaces) can require professional skills and can be time-consuming. Fig 1 shows examples of manually decorated shapes, such as a monogram, logomark, and logotype. Designers have to follow many complicated steps of designing processes, such as sketching and digital vectoring. Therefore, a method for automatically generating decorated shapes would be an important tool for both professional designers and non-professionals.

For logos, there exist applications, such as online tools [1, 2], that can be used for aided design. These tools generate logos by letting users choose from heuristic choices. However, due to the limited choice of fonts, shapes, and design patterns, there is the possibility for duplicate logos when two or more users unintentionally choose the same heuristic choices. Azadi et al. [3] and Yang et al. [4, 5] tried to generate stylized fonts and texts. Yet, these methods require prior training of the model for specific styles and fonts.

In recent years, many studies have been conducted in the field of style transfer, which is the process of transferring styles from one image to another in order to generate a new stylized image. Recently, Gatys et al. [6] introduced the Neural Style Transfer (NST) algorithm for image style transfer using Convolutional Neural Networks (CNN) [7]. As shown in Fig 2, a new stylized image is created by synthesizing a content image and a style image. Specifically, an image is iteratively optimized from an initial image to that of an image with mixed qualities from the content image and the style image. The initial image can be anything; it could be a random image or the content image. The NST has been used in many domains, such as for fashion [8], portraits [9], photorealistic scenes [10, 11] and video style transfer [12].

Fig 2. An example of NST.

Features of the style image are blended into the structure of the content image in the results image.

The original NST method iteratively optimizes the initial image with a summary statistic, a Gram matrix. In this approach, global features are summarized into Gram matrices on specified layers of the CNN. However, the Gram matrix can be unstable because different sets of global features can be encoded into the same Gram matrix [13]. Furthermore, the Gram matrix ignores local features and their coherence. For instance, in Fig 2, the global features of the style image are densely populated multiple colors. With NST methods, the colors and density of the style image are transferred, and some local features of the style image (e.g. curved borders between the colors) are ignored. On the other hand, Li and Wand [14] proposed to capture local features by matching the patches from the feature maps in their Convolutional Neural Network Markov Random Fields (CNNMRF). In this approach, patches are extracted from feature maps, and directly matched to ensure that local features will not be ignored. The biggest disadvantage of the CNNMRF method is that it tends to transfer exact pixels of the style image. Because of that, the CNNMRF method produces nice and well-formed results only when global features of the content and style images are similar.

In this paper, we exploited the advantages of both parametric and non-parametric neural style transfer methods for stylizing images automatically. By combining these methods, we were able to transfer both correlations of global features and local features of the style image onto the content image simultaneously. However, the direct application of this method results in images that have styles in the foreground and background alike. Therefore, we constrained the style transfer process only to the areas in or near the foreground object of the content image with [15]. Some comparisons using the NST method are shown in Fig 3.

Fig 3. Examples of shape stylization.

Compared to the NST, the proposed method was able to transfer the styles only to the foreground shape of the content image.

Furthermore, the proposed method does not require any prior training for specific styles in contrast to [3, 5]. Therefore, even people without any professional designing skills can create their unique results using any image as content or style image. Content image is restricted to clip arts and binary silhouette images as required in [15], while style image can be arbitrary. Due to that, the number of designs that can be created using the proposed method is limitless in contrast to the heuristic methods.

The main contributions of this paper are as follows:

  • The first development of a neural style transfer-based shape design method.
  • Suggested the combination of parametric and non-parametric neural style transfer methods into one model.
  • Demonstrated a method of generating new and unique shapes (e.g., logos) easily.

Related work

Decorated shape generation

A classical approach to logo generation is the use of genetic algorithms. For example, Mark Maker [16] continuously generates logos based on other logos selected by the user. Each user-selected logo contributes to the seed for future generations. This genetic selection process continues until users find their desired logo. The problem with this approach is that no new logos are created; it combines user-defined text with existing shapes and attempts to guess the user’s preferred combination. Another problem is the possibility of users finding the same logo due to the limited choices. A more recent approach in logo generation is RepulsionPak by Saputra et al. [17]. In RepulsionPak, a container shape is filled with smaller shape elements iteratively. However, the smaller shape elements must be prepared for manipulation beforehand.

By using a CNN, Sage et al. [18] used Generative Adversarial Networks (GAN) [19] to generate logos. They gathered various logos from the Internet to create the Large Logo Dataset (LLD) and trained a GAN with the LLD to synthesize new logos. However, the logo synthesizing process was difficult to control. Therefore, Mino et al. [20] proposed to control the color of the resulting logos. Also, Atarsaikhan et al. [15] generated new logos from any shapes by constraining the original NST [6].

Regarding the creation of font images, Tsuchiya et al. [21] and Suveeranont et al. [22] proposed methods to use example fonts to determine the predictive features. Campbell et al. [23] and Uchida et al. [24] generated fonts using the interpolation of different fonts. More recently, fonts have been generated using NST [25] and GANs [2628]. Furthermore, fonts have been stylized by statistical methods [4, 5] and by using GANs [3, 29, 30]. However, most of these methods require prior training for the style, making them difficult to use.

Neural style transfer

Style transfer itself is not a new field. Hertzmann et al. [31] first introduced the example-based style transfer method. More recently, Gatys et al. opened a new field called, “Neural Style Transfer” [6]. Since then, many improvements and inspired methods introduced. According to Jing et al. [32], neural style transfer field is divided into image optimization-based methods that optimize images online, and model optimization-based methods that trains the neural network for stylization. Model optimization-based methods train the network for a specific style or stylization technique. This subfield is opened by Johnson et al. [33] and Ulyanov et al. [34], that an encoder-decoder network is trained for a specific style. Huang et al. [35] introduced a method that does not require training for a specific style by using instance normalization, and improved by Li et al. [36] in their universal style transfer. More recently, Jing et al. [37] achieved great results by using dynamic instance normalization layer instead of instance normalization of [35]. Image optimization-based neural style transfer is also divided into parametric methods and non-parametric methods. Our proposed method lies in both parametric and non-parametric neural style transfer.

Parametric neural style transfer.

Gatys et al. used Gram matrices to capture the correlations of high-level features in texture synthesis [38] and the NST [6]. Li et al. [39] proved that Gram matrix-based methods are equivalent to minimizing the Maximum Mean Discrepancy [40] using a quadratic polynomial kernel. A histogram loss was proposed by Risser et al. [13] to make NST more stable and faster. Li et al. [41] introduced a Laplacian loss between the content image and the generated image to preserve the low-level features. A video style transfer was also introduced [12, 42, 43] and expanded into [44]. Luan et al. [10] and Mechrez et al. [11] achieved photorealistic neural style transfer. Additionally, semantic-aware style transfer [45] and style transfer with controlled content features [46, 47] were introduced.

One drawback of the Gram matrix-based methods is that the slow speed of the style transfer process. To address the speed, Ulyanov et al. [34] trained a generator network and for a specific style image. Furthermore, Johnson et al. [33] trained a ConvDeconv neural network for fast style transfer as well as super-resolution. The training time and quality of results of these methods are improved in [35, 48].

Non-parametric neural style transfer.

In contrast to the Gram matrix-based style transfer that captures features as global correlations, patch-based style transfer can capture local features. Li and Wand [14] introduced CNNMRF, a patch-based style transfer. They replaced the Gram matrix calculation with a patch-matching process performed on the outputs of a Visual Geometry Group network (VGGNet) [49]. This work has inspired research on semantic patch-based style transfer [5052]. Furthermore, Liao et al. [53] proposed attribute style transfer. The processing time is decreased significantly by training fast semantic style transfer [54] and feed-forward networks [55, 56]. More recently, GLStyleNet [57] proposed transferring both the local features and global correlation of features.

Guided neural style transfer for decorated shape generation

The proposed method is illustrated in Fig 4. In addition to capturing global features with content loss, and correlation of global features with style loss from NST [6], we used patch matching loss [14] to capture local features, and distance transfer loss [15] to constrain the style transfer process to the foreground shape. With these losses, local and global features, and correlation of global features are simultaneously transferred onto the foreground shape of the content image. In this section, we will explain the algorithm and the losses used in the proposed method.

Fig 4. The process flow of the proposed method.

Black arrows show forward operations and red arrows show back-propagation. Details of the “Distance transform module” is shown in Fig 7 and the “Patch matching module” is in Fig 8.

Neural style transfer

The process flow of the NST method is illustrated in Fig 5. The NST method extracts content and style representations of input images using a CNN and mixes them to optimize the generated image. A pre-trained VGGNet is used as the backbone of the method. When an image is input into the pre-trained VGGNet, it produces feature maps on every layer. The feature maps produced on the lower layers tend to represent fine details from the input image. On the other hand, the feature maps from the higher layers show arrangements of the objects inside input image. The content and style losses are computed from these feature maps for the optimization goal of the generated image.

Fig 5. The process flow of the NST.

Gram matrices are calculated from style feature maps captured by inputting style and generated images. The difference between the Gram matrices becomes style loss. Also, the difference between content features is content loss. Style and content loss are accumulated into the total loss. Using this total loss the generated image is optimized via back-propagation.

The stylization process is as follows. Before commencing the optimization process of the generated image, content representation of the content image and style representation of the style image are calculated. First, the content image C is input into the VGGNet. Its feature maps of all layers are taken as the content representation. Then, the style image S is input into the VGGNet to determine style feature maps of all layers. Using the style feature maps , Gram matrices are computed and stored as the style representation of the style image. Also, the generated image X is initialized. It can be initialized as anything, e.g. as the content image or as a random noise image.

At the beginning of the optimization process, the generated image X is input into the VGGNet. Content representation and style representation are extracted in the same manner as with the content and style images, respectively. Content loss is determined as a sum of the squared difference between content representations of content and generated images. Style loss is determined as a sum of the squared difference between style representations of the style and generated images. These losses are accumulated into the total loss . Namely, the total loss is calculated by: (1) where α and β are weighting hyperparameters. With the total loss , gradients of the generated image are calculated by using back-propagation through the used layers. Then the optimizer optimizes the generated image X using its updated gradients.

As mentioned before, the original NST uses only content loss and style loss to optimize the generated image. Below is a detailed explanation for content and style losses.

Content loss

Using the feature maps of the content image and the feature maps of the generated image as the content representations, content loss for all selected set of layers Lc is defined as: (2) where l refers to a layer, Nl is the number of feature maps, Ml is the number of elements in one feature map of the layer, and is a pre-defined weight for the layer. Also, each element of and is written as and , respectively. Content loss preserves the general shape (content) of the content image by penalizing the change in the content representation of the generated image. Due to the higher layers learning high-level features such as the structures of the content image, it is preferable to calculate the content loss from higher layers.

Style loss

For the calculation of the style representation, the Gram matrix is used. The Gram matrix Gl of layer l is defined as: (3) where Sl is a matrix that consists of flattened feature maps from layer l. With Nl as the number of feature maps on layers l and as an individual flattened feature map, Sl can be written as follows: (4)

The texture information (style) is incorporated into the Gram matrix by correlating the feature maps of a given layer. To capture consistent and scalable style representation , Gram matrices are computed generally on layers conv1_1, conv2_1, conv3_1, conv4_1, conv5_1 of the VGGNet.

With style representation of the style image as , style representation of the generated image as and wl as a pre-defined weight for a style layer l, style loss on all selected set of layers Ls is defined as, (5)

In Eq 5, and refers to the values at indices (i, j) of the Gram matrices on layer l. Style loss ensures that the correlation of global features (style) are the same on both the style and generated images. On every block of layers, style loss tends to capture different features; therefore, the layers from each block are used as the style layers.

Distance transform loss

Unlike natural images, shapes are typically isolated structures with plain backgrounds, as shown in the content image of Fig 6. Consequently, standard NST methods will place unwanted style features in the background area. To address this problem, Atarsaikhan et al. [15] introduced Distance Transform Loss that is calculated from the distance transform of the content image. The distance transform assigns a value for each pixel of a binary image based on the distance to the nearest foreground pixel. Therefore, the binary image has a pixel value of 0 on the foreground shape and pixel value gradually increases based on the distance from the shape. An example of distance transform is show in Fig 6. Furthermore, the distance transform D of content image C can be tuned to emphasize the distance between the foreground and the background using the power z.

Fig 6. An example of an image with its respective distance transform.

Using generated image X, content image C, and emphasized distance transform Dz of the content image, distance transform loss is defined as, (6) where ∘ stands for pixel-wise multiplication. This process is shown in distance transform module in Fig 7. The biggest advantage of utilizing distance transform loss is that it constrains stylization to the foreground shape while permitting soft transformations near the edge of the foreground. Distance transform loss ensures that the continuity of the style around the foreground edges is preserved without any abrupt cuts. In other words, it performs a soft masking operation, instead of the regular hard masking.

Fig 7. Distance transform module.

Distance transform loss is calculated as, the difference between pixel-wise multiplication of the content image C and D, and pixel-wise multiplication of the generated image X and D.

Patch matching loss

Visually plausible results can be achieved using content loss to preserve the shape of the content image, style loss to transfer the style of the style image, and distance transform loss to constrain the transfer to the content image. However, due to the use of a Gram matrix between feature maps, the results are unstable and hard to control. The Gram matrix statistically abbreviates feature maps of the style image and tries to match with the Gram matrix of the generated image. However, there can be more than one set of feature maps producing the same Gram matrix. Thus, the Gram matrices from the style image and the generated image can be the same, even if their styles are not the same. Moreover, the Gram matrix is designed to capture correlations of global features and ignore the local features.

Li et al. [14] proposed a patch matching technique to extract styles that include local features. They divided the feature maps from the VGGNet into small patches, and the patches from style image and generated image are matched to achieve style transfer. This technique ensures that the feature maps of the style image and the generated images are similar. However, the problem with this method is that it only produces attractive results if the content and style images are already quite similar. Intuitively, patch-based style transfer methods lack the abstract abilities of the standard NST using the Gram matrices.

We combined the advantages of these two methods by using both style loss and a patch matching loss which is inspired from [14]. When an image is input into a VGGNet, patches are extracted from feature maps on specified patch layers. The size of one patch is [C × k × k] with stride = 1, where k is height and width and C as the number of nodes in the layer. We extract patches from conv3_1 and conv4_1, because these layers capture both local and global features in contrast to lower layers that capture colors and borders, or higher layers that capture global features only. Given style image S and generated image X, we extracted style patches as and generated patches as , respectively.

The process of calculating patch matching loss is shown in Fig 8. For every generated patch Pgenerated of generated patches , the best matching patch can be found from by using: (7) where p is index and P is all number of patches in . Also, p is index for patches in and pmin is the closest patch index from . Then, patch matching loss is defined as a sum of the difference between generated patches and corresponding best matching patches: (8)

Fig 8. Patch matching module.

Patches are extracted as a set of generated patches and a set of style patches from layers conv3_1 and conv4_1 of and , respectively. Because we extract patches from the same layers for generated and style images, the patch number is the same, P. Then, for every generated patch with p as index, closest style patch with index pmin is found. Patch matching loss is sum of losses on every pair of and .

Style transfer process

The entire style transfer process of the proposed method is shown in Fig 4. We calculate and store content representation of content image C and style representation of style image S as previously explained. Also, style patches of style image are extracted as from patch layers. Next, generated image X, which is initialized as either content image C or a random noise image, is input to the network. Content and style representations are calculated, as well as generated patches are extracted. By using content loss , style loss Lstyle, distance transform loss and patch matching loss , the total loss can be determined as, (9) where α, β, γ and δ are weighting factors for each loss. Once the total loss is determined, the generated image X is optimized by back-propagation with the goal of minimizing the total loss . By repeatedly running this process, the content of the content image and the style of the style image are gradually synthesized into the generated image X.

The processing time of the style transfer process depends on image size and the number of patches. The larger image size simply increases the number of nodes that had to be processed. Although larger images take longer time to stylize, the results do not necessarily become more plausible. Table 1 shows the average computation time of individual loss functions in one iteration. The patch matching loss requires the longest time to compute. Using larger patch size is more time consuming than using a smaller patch size because the number of nodes is increased with the patch size. Moreover, using stride higher than 1 significantly decreases the computation time because of less number of patches. Other losses require a small amount of time to compute however, it increased with image size. Furthermore, the initialization of the generated image affects the overall processing time. More iterations are needed to achieve the same results with random noise initialization than initializing as the content image.

Table 1. The average computation time of each loss in one iteration.

Results and discussion

We used silhouette icons, single letter images, and texts as content images. Examples of the stylized shapes using the proposed method are shown in Fig 9. In Fig 9A, a monogram is generated from a single “R” image and style image. Fig 9B shows a logomark that is created from silhouette icons and Fig 9C has a text as logotype.

Fig 9. Examples of logo generation using the proposed method.

Qualitative evaluation

To analyze the capability and quality of the proposed method, we performed the following comparison experiments:

  • NST [6]. The standard NST method which uses only style loss and content loss . NST is run for 5,000 iterations.
  • Constrained NST [15]. NST using style loss and content loss plus the additional distance transform loss . Constrained NST is also run for 5,000 iterations.
  • CNNMRF [14]. This comparison is the model proposed by Li et al. which introduced patch matching based NST. Only patch matching and the patch matching-based loss is used. This network is run for 2,000 iterations.
  • NST + Patch matching. This evaluation is to demonstrate the effects of using NST with only style loss , content loss , and the patch matching loss . Using NST + Patch matching is run for 2,000 iterations.
  • Arbitrary Style Transfer in Real-time with Adaptive Instance Normalization (AdaIN) [35]. Huang et al. stylize contents by using instance normalization. It is possible to perform spatial control by masking feature maps on each layer.
  • Universal Style Transfer via Feature Transforms (UST) [36]. Although similar to AdaIN, UST uses whitening and coloring transformation instead of instance normalization. As for spatial control, it can be achieved by masking feature maps on each layer.
  • Proposed method. This is the proposed method of using all four losses and is run for 2,000 iterations.

In the experiments, all image sizes are 256 × 256. The pre-defined weights for content layers are zero except for , and weights for style layers are zero except for , , , , . For the methods that use patch matching, patch sizes are C × 5 × 5 and are taken from the layers conv3_1, and conv4_1. The weighting factors for each loss are α = 0.1, β = 0.1, γ = 10−6 and δ = 100. All of the experiments are performed on single “GeForce GTX 1080 Ti” GPU.

In Fig 10, we compare the results of the above methods. Due to the lack of constraint from a distance transform loss, NST, CNNMRF, and NST + Patch matching have unnecessary background noise. Although, in the NST results, the contents of the content image are more visible than CNNMRF. This is because, in NST, feature maps are abstracted with a Gram matrix and do not require to look exactly like the style image. In CNNMRF, patches from feature maps are directly matched, thus, the method neurally copies and pastes style features. NST + Patch matching method transfers local features as well as correlation of global features from the style image. Thus, the results are visually comparable to the style image. However, these methods do not have constrain or control to ignore the background of the content image. So, there are styles on all of the generated images, foreground and background indiscriminately.

Fig 10. The results comparison with NST, CNNMRF, Constrained NST, NST + Patch matching, AdaIN with spatial control, and UST with spatial control methods.

The first column shows various content and style combinations, and each column shows results from each method using the content and style combination from the left.

Conversely, using the NST with distance transform loss in the Constrained NST, the style transfer is contained in the shape correctly. However, because the content images and the style images are not similar enough, the results of using just a Gram matrix for style transfer are not sufficiently impressive. There are areas that the Constrained NST could not transform and corrupted colors in the logotype results.

The spatial controls in AdaIN and UST worked fairly well. The results are appealing, and there is no style on the background. However, details (e.g. round shapes, patterns) of the style image are completely ignored. Instead, new details with color similar to the style image emerged. The biggest advantage of these methods is the feed-forward stylization. Because of that, they are generally much faster than the proposed method.

On the other hand, results from the proposed method are not only constrained to the shape of the content image but also included the details from style image. Distance transform loss constrains the style transfer and the patch matching loss ensures that there is no corrupted colors or different details. The results in Fig 10 demonstrate that the proposed method combines the benefits of patch matching and the distance constraint. Similar to the CNNMRF and NST + Patch matching, the proposed method has detailed local features, and similar to Constrained NST, the proposed method does not have an extra background style.

On simple masking

One concern would be that simply masking the style image would produce better results. However, Fig 11 shows that a simple masking operation does a poor job. The results from the masking operation have unnatural cuts of the style, whereas the results from the proposed method preserved the style as natural as possible. Specifically, as shown in the first row, the style is black areas, and thin white stripes. By masking this style with the content letter of “A”, white stripes are cut in the middle. However, the result from the proposed method has continual white stripes. In rows 2 and 3, the masking operation resulted in incoherent contents. This shows the results of the proposed method is better than a simple masking operation.

Fig 11. Comparison of the proposed method against masking the style image.

Ablation study

In this section, we discuss the effects of some of the hyperparameters of the proposed method.

Effects of weighting factors.

In Eq (9), the different losses are weighted respectively. Different weighting factors have different effects on the results. Hence, it is important to understand the effect of weighting factors. Fig 12 shows the effects of each weighting factor while other weighting factors are fixed. The magnitude of the weighting factor is attributable to the magnitude of the respective loss, in comparison to the other losses. In Fig 12A, there are results with various α for content loss. If α is larger the style transfer process is hindered, and the results are more similar to the content image. However, if α is smaller, content shapes could not be recognized from the results. Fig 12B shows various beta for style loss. With smaller β local features are dominant, while with bigger β style loss neglected other losses, especially distance transform loss. Distance transform loss could be overwhelmed by patch matching loss with large γ as shown in Fig 12C. Smaller γ means that the global feature is confined into the shape of the content image. Moreover, distance transform loss with small δ could not be able to constrain the style transfer process. On the other hand, if the δ is too large, it hindered the style transfer process. There is a trade-off between these weighting factors, and we found that α = 0.1, β = 0.1, γ = 10−6 and δ = 100 work best.

Fig 12. Style transferring results with various parameters.

Patch size.

The population of the style patches in the generated image can be controlled by changing patch size. Fig 13 shows experimental results on the patch sizes of patch matching module. With smaller patches, such as 5 × 5, more densely populated styles can fit into the shape as shown in Fig 13A and 13C. On the other hand, as shown in Fig 13B and 13D, with larger patches, like 11 × 11 the styles are scarce. The reason for that is when the patch size is small, it includes the part of feature maps that represents either of the style or the background, separately. Then, in the patch matching process, the patches that include only background features are matched to the background of the content image and patches of the foreground are matched properly. With large patches like 11 × 11, the windows are so large that they include both the foreground and background. When applied to the generated image, excess background from the style image appears.

Fig 13. The effects of patch size that used in the proposed method.


While the proposed method produces generally good results, depending on the choice of the content and style image combinations the quality of the generated images can vary. In Fig 14, style images have no specific background. Although some features of the style images were transferred onto the shape, the stylized results are not as plausible as other results. That is because the proposed method tried to generate all of the style image onto the shape of content image, which is too small for the entire style image. Fig 15 compares style images with white background and purple background. The user has to specify the background color of the input images. In this paper, we specified the background color as white because most logo images have a white background. So, when a style image with a white background is used, the white background was rejected in the patch matching process. Thus, the results have only flowers and green leaves. However, when a purple background style image is used, the proposed method cannot discriminate between the background. Hence, the generated image has all features from the style image; flowers, green leaves, and purple background.

Fig 14. Photos and images without plain background as style images.

Fig 15. White background style image vs colored background image.

Moreover, Fig 16 shows results that are generated from content images that have too narrow shapes compared with the style. The first image shows that when the content shape is too narrow, there is not enough room for the style. Thus, the results change very little, or in extreme cases does not change at all, as shown in the second image. Furthermore, the third image shows that there is no change in the narrow part of the content image whereas the wider parts are stylized.

Alternatively, Fig 17 shows various combinations of content images and style images that produce good results. With the same size objects, the style transfers much more easily because of the patch matching module and Gram matrix style transfer. We also found that style images with many different but somewhat repetitive features work well. This is due to the global correlation of features in the Gram matrix. In addition, it is better to have many instances of similar but slightly different features. Finally, the objects inside the content images should be large enough to be able to contain multiple features from the style image. This also enables them to be redrawn by the style from the style image.

Fig 17. Examples of good combinations of content and style images.

Content and style combinations that produce good results. The upper smaller images are the content and style images, respectively. The larger images are the generated results.

Table 2 shows the overall computation time of each method with various parameters. The proposed method is much slower than other methods. Due to the high number of nodes in the extracted patches, the backward calculation of gradients with regard to the total loss consumes a large amount of time. One solution is to use a larger stride. With a larger stride, the number of nodes can be decreased significantly and make the proposed method faster. NST and Constrained NST have similar computation time because distance transform loss requires a negligible amount of computation time. CNNMRF is also fast because it resizes the image to a smaller size and gradually increases the image size while stylizing. AdaIN and UST require one forward-pass for stylization, however, prior training is also needed.


In this paper, we integrated a patch matching process into the Gram matrix-based style transfer method for stylizing shapes by transferring local and global correlation of features from a style image. Using a patch matching loss, we were able to transfer the local features that lack in the regular NST. Then, we constrained the style transfer process only to the contents of the content image using a distance transform loss. Compared to hard cutting, such as masking, the proposed method ensures continuity of the transferred styles in the generated image. Through experimental results, we demonstrated that the proposed method outperformed multiple style transfer methods such as NST, CNNMRF, NST + Patch matching, Constrained NST, AdaIn, UST, and simple masking. Furthermore, we discussed the effects of different hyperparameters and different content and style images on the generated results. We demonstrated that the proposed method can generate attractive shapes easily without requiring prior training and manual instructions for the style.


  1. 1. ai L. Press Kit: Everything about;. Available from:
  2. 2. Logojoy. How it works: Design a logo you’ll love with our online logo maker;. Available from:
  3. 3. Azadi S, Fisher M, Kim VG, Wang Z, Shechtman E, Darrell T. Multi-Content GAN for Few-Shot Font Style Transfer. In: IEEE Conf. Comput. Vision and Pattern Recognition; 2018. p. 7564–7573.
  4. 4. Yang S, Liu J, Lian Z, Guo Z. Awesome Typography: Statistics-Based Text Effects Transfer. In: IEEE Conf. Comput. Vision and Pattern Recognition; 2017. p. 2886–2895.
  5. 5. Yang S, Liu J, Lian Z, Guo Z. Text effects transfer via distribution-aware texture synthesis. Computer Vision and Image Understanding. 2018;174:43—56.
  6. 6. Gatys LA, Ecker AS, Bethge M. Image style transfer using convolutional neural networks. In: IEEE Conf. Comput. Vision and Pattern Recognition; 2016. p. 2414–2423.
  7. 7. LeCun Y, Bottou L, Bengio Y, Haffner P. Gradient-based learning applied to document recognition. Proc IEEE. 1998;86(11):2278–2324.
  8. 8. Jiang S, Fu Y. Fashion Style Generator. In: Int. Joint Conf. Artificial Intell.; 2017. p. 3721–3727.
  9. 9. Selim A, Elgharib M, Doyle L. Painting style transfer for head portraits using convolutional neural networks. ACM Trans Graphics. 2016;35(4):129.
  10. 10. Luan F, Paris S, Shechtman E, Bala K. Deep Photo Style Transfer. In: IEEE Conf. Comput. Vision and Pattern Recognition; 2017. p. 4990–4998.
  11. 11. Mechrez SE Roey, Zelnik-Manor L. Photorealistic Style Transfer with Screened Poisson Equation. British Mach Vision Conf. 2017;.
  12. 12. Ruder M, Dosovitskiy A, Brox T. Artistic style transfer for videos. In: German Conf. on Pattern Recognition; 2016. p. 26–36.
  13. 13. Risser E, Wilmot P, Barnes C. Stable and controllable neural texture synthesis and style transfer using histogram losses. CoRR. 2017;.
  14. 14. Li C, Wand M. Combining markov random fields and convolutional neural networks for image synthesis. In: IEEE Conf. Comput. Vision and Pattern Recognition; 2016. p. 2479–2486.
  15. 15. Atarsaikhan G, Iwana BK, Uchida S. Contained Neural Style Transfer for Decorated Logo Generation. In: Int. Workshop Document Anal. Syst.; 2018. p. 317–322.
  16. 16. Emplemmatic. Mark Maker;. Available from:
  17. 17. Saputra RA, Kaplan C, Asente P. Improved Deformation-Driven Element Packing with RepulsionPak. IEEE transactions on visualization and computer graphics. 2019;.
  18. 18. Sage A, Agustsson E, Timofte R, Van Gool L. Logo synthesis and manipulation with clustered generative adversarial networks. In: IEEE Conf. Comput. Vision and Pattern Recognition; 2018. p. 5879–5888.
  19. 19. Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, et al. Generative adversarial nets. In: Advances in Neural Inform. Process. Syst.; 2014. p. 2672–2680.
  20. 20. Mino A, Spanakis G. LoGAN: Generating Logos with a Generative Adversarial Neural Network Conditioned on Color. In: IEEE Int. Conf. Machine Learning and Appli.; 2018. p. 965–970.
  21. 21. Tsuchiya T, Miyazaki T, Sugaya Y, Omachi S. Automatic Generation of Kanji Fonts from Sample Designs. In: Tohoku-Section Joint Conv. Inst. Elect. and Inform. Engineers; 2014. p. 36.
  22. 22. Suveeranont R, Igarashi T. Example-based automatic font generation. In: Int. Symp. Smart Graphics; 2010. p. 127–138.
  23. 23. Campbell NDF, Kautz J. Learning a manifold of fonts. ACM Trans Graphics. 2014;33(4):1–11.
  24. 24. Uchida S, Egashira Y, Sato K. Exploring the World of Fonts for Discovering the Most Standard Fonts and the Missing Fonts. In: Int. Conf. Document Anal. and Recognition; 2015. p. 441–445.
  25. 25. Atarsaikhan G, Iwana BK, Narusawa A, Yanai K, Uchida S. Neural Font Style Transfer. In: Int. Conf. Document Anal. and Recognition; 2017. p. 51–56.
  26. 26. Abe K, Iwana BK, Holmér VG, Uchida S. Font Creation Using Class Discriminative Deep Convolutional Generative Adversarial Networks. In: Asian Conf. Pattern Recognition; 2017. p. 232–237.
  27. 27. Lyu P, Bai X, Yao C, Zhu Z, Huang T, Liu W. Auto-Encoder Guided GAN for Chinese Calligraphy Synthesis. In: Int. Conf. Document Anal. and Recognition; 2017. p. 1095–1100.
  28. 28. Sun D, Zhang Q, Yang J. Pyramid Embedded Generative Adversarial Network for Automated Font Generation. CoRR. 2018;.
  29. 29. Zhao N, Cao Y, Lau RWH. Modeling Fonts in Context: Font Prediction on Web Designs. Comput Graphics Forum. 2018;37(7):385–395.
  30. 30. Yang S, Liu J, Wang W, Guo Z. TET-GAN: Text Effects Transfer via Stylization and Destylization. CoRR. 2018;.
  31. 31. Hertzmann A, Jacobs CE, Oliver N, Curless B, Salesin DH. Image analogies. In: Proceedings of the 28th annual conference on Computer graphics and interactive techniques. ACM; 2001. p. 327–340.
  32. 32. Jing Y, Yang Y, Feng Z, Ye J, Yu Y, Song M. Neural style transfer: A review. IEEE transactions on visualization and computer graphics. 2019;.
  33. 33. Johnson J, Alahi A, Fei-Fei L. Perceptual losses for real-time style transfer and super-resolution. In: European Conf. Comput. Vision; 2016. p. 694–711.
  34. 34. Ulyanov D, Lebedev V, Vedaldi A, Lempitsky V. Texture Networks: Feed-forward Synthesis of Textures and Stylized Images. In: Int. Conf. Mach. Learning; 2016. p. 1349–1357.
  35. 35. Huang X, Belongie S. Arbitrary style transfer in real-time with adaptive instance normalization. CoRR. 2017;.
  36. 36. Li Y, Fang C, Yang J, Wang Z, Lu X, Yang MH. Universal style transfer via feature transforms. In: Advances in neural information processing systems; 2017. p. 386–396.
  37. 37. Jing Y, Liu X, Ding Y, Wang X, Ding E, Song M, et al. Dynamic Instance Normalization for Arbitrary Style Transfer. In: AAAI; 2020.
  38. 38. Gatys LA, Ecker AS, Bethge M. Texture synthesis and the controlled generation of natural stimuli using convolutional neural networks. CoRR. 2015;.
  39. 39. Li Y, Wang N, Liu J, Hou X. Demystifying neural style transfer. In: Int. Joint Conf. Artificial Intell.; 2017. p. 2230–2236.
  40. 40. Gretton A, Borgwardt KM, Rasch MJ, Schölkopf B, Smola A. A kernel two-sample test. J Mach Learning Research. 2012;13(Mar):723–773.
  41. 41. Li S, Xu X, Nie L, Chua TS. Laplacian-Steered Neural Style Transfer. In: ACM Int. Conf. Multimedia; 2017. p. 1716–1724.
  42. 42. Anderson AG, Berg CP, Mossing DP, Olshausen BA. DeepMovie: Using Optical Flow and Deep Neural Networks to Stylize Movies. CoRR. 2016;.
  43. 43. Joshi B, Stewart K, Shapiro D. Bringing impressionism to life with neural style transfer in come swim. In: ACM SIGGRAPH Digital Production Symp.; 2017. p. 5.
  44. 44. Ruder M, Dosovitskiy A, Brox T. Artistic style transfer for videos and spherical images. Int J Comput Vision. 2018; p. 1–21.
  45. 45. Yin R. Content Aware Neural Style Transfer. CoRR. 2016;.
  46. 46. Gatys LA, Bethge M, Hertzmann A, Shechtman E. Preserving color in neural artistic style transfer. CoRR. 2016;.
  47. 47. Gatys LA, Ecker AS, Bethge M, Hertzmann A, Shechtman E. Controlling perceptual factors in neural style transfer. In: IEEE Conf. Comput. Vision and Pattern Recognition; 2017. p. 3730–3738.
  48. 48. Ulyanov D, Vedaldi A, Lempitsky V. Improved texture networks: Maximizing quality and diversity in feed-forward stylization and texture synthesis. In: IEEE Conf. Comput. Vision and Pattern Recognition; 2017. p. 6.
  49. 49. Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition. In: Int. Conf. Learning Representations; 2015.
  50. 50. Champandard AJ. Semantic style transfer and turning two-bit doodles into fine artworks. CoRR. 2016;.
  51. 51. Chen YL, Hsu CT. Towards Deep Style Transfer: A Content-Aware Perspective. In: British Mach. Vision Conf.; 2016. p. 8.1–8.11.
  52. 52. Mechrez R, Talmi I, Zelnik-Manor L. The contextual loss for image transformation with non-aligned data. CoRR. 2018;.
  53. 53. Liao J, Yao Y, Yuan L, Hua G, Kang SB. Visual Attribute Transfer Through Deep Image Analogy. ACM Trans Graph. 2017;36(4):120:1–120:15.
  54. 54. Lu M, Zhao H, Yao A, Xu F, Chen Y, Zhang L. Decoder network over lightweight reconstructed feature for fast semantic style transfer. In: IEEE Int. Conf. Comput. Vision; 2017. p. 2488–2496.
  55. 55. Li C, Wand M. Precomputed real-time texture synthesis with markovian generative adversarial networks. In: European Conf. Comput. Vision; 2016. p. 702–716.
  56. 56. Chen TQ, Schmidt M. Fast Patch-based Style Transfer of Arbitrary Style. CoRR. 2016;.
  57. 57. Wang Z, Zhao L, Xing W, Lu D. GLStyleNet: Higher Quality Style Transfer Combining Global and Local Pyramid Features. CoRR. 2018;.