ArticlePDF Available

An Improved Style Transfer Algorithm Using Feedforward Neural Network for Real-Time Image Conversion

Authors:
  • Qinghai University Lanzhou University

Abstract and Figures

Creation of art is a complex process for its abstraction and novelty. In order to create those art with less cost, style transfer using advanced machine learning technology becomes a popular method in computer vision field. However, traditional transferred image still troubles with color anamorphosis, content losing, and time-consuming problems. In this paper, we propose an improved style transfer algorithm using the feedforward neural network. The whole network is composed of two parts, a style transfer network and a loss network. The style transfer network owns the ability of directly mapping the content image into the stylized image after training. Content loss, style loss, and Total Variation (TV) loss are calculated by the loss network to update the weight of the style transfer network. Additionally, a cross training strategy is proposed to better preserve the details of the content image. Plenty of experiments are conducted to show the superior performance of our presented algorithm compared to the classic neural style transfer algorithm.
Content may be subject to copyright.
Sustainability 2019, 11, 5673; doi:10.3390/su11205673 www.mdpi.com/journal/sustainability
Article
An Improved Style Transfer Algorithm Using
Feedforward Neural Network for Real-Time
Image Conversion
Chang Zhou
1
, Zhenghong Gu
1
, Yu Gao
1
and Jin Wang
2,3,
*
1
College of Information Engineering, Yangzhou University, Yangzhou 225000, China;
zhouchangyz@163.com (C.Z); guzhenghong@yzu.edu.cn (Z.G.); mx120170403@yzu.edu.cn (Y.G)
2
Hunan Provincial Key Laboratory of Intelligent Processing of Big Data on Transportation, School of
Computer & Communication Engineering, Changsha University of Science & Technology,
Changsha 410000, China; jinwang@csust.edu.cn
3
School of Information Science and Engineering, Fujian University of Technology, Fuzhou 350000, China
* Correspondence: jinwang@csust.edu.cn; Tel.: +86-180-1484-9250
Received: 27 September 2019; Accepted: 11 October 2019; Published: 14 October 2019
Abstract: Creation of art is a complex process for its abstraction and novelty. In order to create those
art with less cost, style transfer using advanced machine learning technology becomes a popular
method in computer vision field. However, traditional transferred image still troubles with color
anamorphosis, content losing, and time-consuming problems. In this paper, we propose an
improved style transfer algorithm using the feedforward neural network. The whole network is
composed of two parts, a style transfer network and a loss network. The style transfer network owns
the ability of directly mapping the content image into the stylized image after training. Content loss,
style loss, and Total Variation (TV) loss are calculated by the loss network to update the weight of
the style transfer network. Additionally, a cross training strategy is proposed to better preserve the
details of the content image. Plenty of experiments are conducted to show the superior performance
of our presented algorithm compared to the classic neural style transfer algorithm.
Keywords: style transfer; convolution neural network; cross training; machine learning
1. Introduction
Advanced machine learning technology makes the automatically style transfer possible because
of its powerful fitting ability [1–5]. Style transfer as a popular method applicated in artistic creation
has attracted much attention. It commonly combines the style information from a style image with
the original content image [6–8]. The fused picture preserves the features of the content image and
style image simultaneously. The strong ability of features extraction using convolution neural
network improves the quality of the synthetic image by style transfer. By adopting style transfer,
people can create works of art easily and don’t need to care how to professionally draw a picture
[9,10]. Additionally, much repetitive work can be omitted and the business costs can be reduced. The
improved quality and the automated process make style transfer popular in artistic creation [11–13],
font style transformation [14–16], movie effects rendering [17,18] and some engineering fields [19–
22].
Traditional style transfer mainly adopts the following methods.
(1) Stroke-Based Rendering: Stroke-based rendering refers to the method of adding a virtual stroke
to a digital canvas to render a picture with a particular style [23–25]. The obvious disadvantage
is that its application scenarios are only limited to oil paintings, watercolors, and sketches and
it’s not flexible enough.
Sustainability 2019, 11, 5673 2 of 15
(2) Image Analogy: Image analogy is used to learn the mapping relationship between a pair of
source images and target images. The source images are transferred by a supervised way. The
training set includes a pair of uncorrected source images and corresponding stylized images
with a particular pattern [26,27]. The analogy method owns an effective performance and its
shortage is that the paired training data is difficult to obtain.
(3) Image Filtering: Image filtering method adopts a combination of different filters (such as
bilateral and Gaussian filters) to render a given image [28,29].
(4) Texture Synthesis: The texture denotes the repetitive visual pattern in an image. In texture
synthesis, similar textures are added to the source image [30,31]. However, those texture
synthesis-based algorithms only use low-level features and their performance is limited.
Recent years, Gatys et al. [6] present a new solution for style transfer combined with the
convolution neural network. It regards the style transfer as the optimization problem and adopts
iterations to optimize each pix in the stylized picture. The pretrained Visual Geometry Group 19
(VGG19) network is introduced to extract the content feature and style feature from the content image
and style image, respectively. Owing the greatly improved performance to the neural network, the
method Gatys et al. proposed is also called neural style transfer. Though neural style transfer
performs much better than some traditional methods, some drawbacks still trouble the researchers.
Firstly, neural style transfer needs to iterate to optimize each pix of the stylized image and it’s not
applicable to some delay-sensitive applications especially those needing real-time processing.
Secondly, though the style features can be well integrated into the stylized image, the content
information is inevitable lost. For example, the color in the content image will be mixed with the color
in the style image and the lines in the stylized image will show varying degrees of distortion.
In order to make up for the lack of the classic neural style transfer, an improved style transfer
algorithm adopting a deep neural network structure is presented. The whole network is composed
of two parts, style transfer network and loss network. The style transfer network conducts a direct
mapping between the content image and the stylized image. The loss network computes the content
loss, style loss, and TV loss between the content image, style image, and stylized image generated by
the style transfer network. Then the weight of the style transfer network can be updated according to
the calculated loss. The style transfer network needs to be trained while the loss network adopts the
first few layers of the pretrained VGG19. A cross training strategy is presented to make the style
transfer network to preserve more detailed information. Finally, numerous experiments are
conducted and the performances are compared between our presented algorithm and the classic
neural style transfer algorithm.
We outline the paper as follows. Section 1 introduces the background of style transfer. Some
parallel works are summarized in Section 2. Section 3 demonstrates the effects of feature extraction
from different layers of VGG19. Section 4 has a specific illustration of our proposed algorithm. Section
5 conducts the experiments and analyzes the experiment results. Merits and demerits are discussed
in Section 6. Section 7 makes a conclusion for the whole paper.
2. Related Work
A deep network structure with multiple convolutional layers is proposed for image classification
[32]. Small filters are introduced for detailed features extraction and less parameters need to be
trained simultaneously. Due to the favorable expansibility of the well trained VGG network, many
other researchers adopt it as a pretrained model for further training.
In order to solve the content losing problem, a deep convolution neural network with dual
streams is introduced for feature extraction and an edge capture filter is introduced for synthetic
image quality improving [33]. The convolution network contains two parts, detail recognizing
network and style transfer network. A detail filter and a style filter are respectively applied to process
the synthesized images from the detail recognizing network and style transfer network for detail
extraction and color extraction. Finally, a style fuse model is used to integrate the detailed image and
the color image into a high-quality style transfer image.
Sustainability 2019, 11, 5673 3 of 15
A style transfer method for color sketch synthesis is proposed by adopting dilated residual
blocks to fuse the semantic content with the style image and it works without dropping the spatial
resolution [34]. Additionally, a filtering process is conducted after the liner color converts.
A novel method combined with the local and global style losses is presented to improve the
quality of stylized images [35]. The local style preserves the details of style image while the global
style captures more global structural information. The fused architecture can well preserve the
structure and color of the content image and it reduces the artifacts.
An end-to-end learning schema is created to optimize both the encoder and the decoder for
better features extraction [36]. The original pretrained VGG is fine-tuned to adequately extract
features from style or content image.
In order to preserve the conspicuous regions in style and content images, the authors adopt a
localization network to calculate the region loss from the SqueezeNet network [37]. The stylized
image can preserve the conspicuous semantics regions and simple texture.
Advanced Generative Adversarial Networks (GAN) technology is introduced to style transfer
for cartoon images [38]. Network training using the unpaired images makes the training set easier to
build. To simulate the sharp edges of cartoon images, the edge loss is added to the loss function. The
Gaussian smoothing method is first used to blur the content image, and then the discriminator
determines the blurred image as a negative sample. A pre-trained process is executed in the previous
several epochs for the generator to make the GAN network converge more quickly.
A multiple style transfer method based on GAN is proposed in [39]. The generator is composed
of an encoder, a gated transformer, and a decoder. The gated transformer contains different branches
and different styles can be adopted by passing different branches.
3. Features Extraction from VGG19
In a pretrained convolution neural network, the convolution kernels own the ability to extract
the features from a picture. Therefore, similarly as the classic neural style transfer, we adopted
VGG19 to extract the features from content and style images. VGG19 is a very deep convolution
neural network trained with ImageNet dataset and has excellent performance in image classification,
object positioning, etc. VGG19 owns good versatility for features extractions and many works adopt
it as the pretrained model. Different from the classic neural style transfer, we first analyze the
extracted features by the VGG19 to select the suitable layers for feature extraction.
Since the features extracted by each layer of the VGG19 have multiple channels and cannot be
directly visualized, we adopt the gradient descent algorithm to reconstruct the original image
according to the features extracted by the different layer of the VGG19. The reconstructed images are
initialized with Gaussian noise and then we put the initialized images into the VGG19. Then the
extracted features are compared in the same layer and their 𝐿 loss are calculated. Next, the 𝐿 loss
is back propagated to the reconstructed image and the reconstructed image is updated according to
the gradient. When reconstructing the content image, we directly use the extracted features to
calculate the 𝐿 loss, as shown in Formula 1.
(y,y)= 1
𝐻𝑊𝐶𝜑(𝑦)−𝜑
(𝑦)
(1)
where y and y denote the original image and the reconstructed image, respectively. 𝜑 denotes
the output value of the j-th layer of VGG19 network. 𝐻,𝑊
,𝐶
denote the width, height, and number
of channels of the j-th layer of VGG19 network. 𝑥 denotes the Euclidean norm of the vector 𝑥.
When reconstruct the style image, the Gram matrix needs to be firstly calculated by Formula 2.
G(y) =
𝑓
𝑓
(2)
where 𝑓denotes the reshaped matrix using the extracted features of the j-th layer of VGG19. Then
the style loss can be defined as Formula 3.
Sustainability 2019, 11, 5673 4 of 15
=1
𝐶𝐺(𝑦)−𝐺(𝑦)
(3)
We exchange the layers used for features extraction and conduct much experiments. The
experiment results are shown as follows.
As we can clearly see from Figure 1, the lo wer laye rs of VGG19 c an prese rve mu ch mo re detai led
information of content images. While, the deeper layers of VGG19 are more interested in the regular
texture which represents the style of a picture. Therefore, the lower layers of VGG19 are more suitable
for content features extraction and the deeper layers are more suitable for style features extraction.
Figure 1. Feature extraction effects of different of the VGG19.
4. Proposed Method
4.1. Data Processing
We firstly process the input data of the network. In order to avoid the problem of color mixing
in the classic neural style transfer, a gray conversion is conducted for the content image. We use the
classic physiology formula to transform each RGB pixel of content image into grey pixel as Formula
4.
𝐺𝑟𝑎𝑦(𝑥)= 𝑅(𝑥) ∙0.299+𝐺(𝑥) ∙0.587+𝐵(𝑥) ∙ 0.144 (4)
where R(x), G(x), and B(x) denote the value of the Red, Green, and Blue (RBG) channels of the pixel
x in content image respectively. After the gray conversion, the original content image with three RGB
channels is transformed into a gray image with one grayed channel. Whereas, the style image is an
RGB picture with three channels, the grayed content image still needs to be converted to the form of
three channels to match the format of the style image. Thus, we just simply stack three identical
grayed channels as the content image with RGB format. The gray conversion is shown in Figure 2.
Figure 2. Gray conversion.
Sustainability 2019, 11, 5673 5 of 15
Another problem we need to solve is that the style image we used is not limited to the fully
texture image. Some regular texture may only concentrate in a centralized area, and we need to use
those local features to render the whole content image. Therefore, it’s necessary to conduct the data
augmentation for the style image to enhance the local features. Following operations are taken for
data augmentation.
(1) Zoom in on the original image and then crop the image of the same size.
(2) Randomly rotate the image at a certain angle and change the orientation of the image content.
(3) Flip the image horizontally or vertically.
(4) Randomly occlude part of the image.
(5) Randomly perturb RGB value of each pixel of the image by adding salt and pepper noise or
Gaussian noise.
The data augmentation is illustrated as Figure 3.
Figure 3. Data augmentation.
4.2. Network Model
The whole network contains two components and they are style transfer network and loss
network. The style transfer network realizes a direct mapping between the content image and the
stylized image. Then the stylized image is inputted to the loss network to calculate the content and
style losses with inputted content and the style image. Next, the weight of the style transfer network
will be updated according to the losses calculated in the loss network using gradient descent
algorithm. The style transfer network is a deep neural network composed of multiple convolution
layers and residual blocks. The weight of each layer in the style transfer network is randomly
initialized while the loss network adopts the first few convolution layers of the pretrained VGG19
network. During the training process, only the weight of the style transfer network will be updated.
The size of the images must be the same during the training phase, while in the test phase, we can
input different sizes of images. The whole network structure and the operation flow are shown in
Figure 4.
Sustainability 2019, 11, 5673 6 of 15
Figure 4. Network model and workflow.
4.3. Style Transfer Network
The style transfer network is stacked by multiple convolution layers, residual blocks, and
deconvolution layers. The convolution layers and deconvolution layers adopt short stride for down
sampling and up sampling, respectively. Specifically, the style transfer network is composed of four
convolution layers, five residual blocks, and two deconvolution layers. Besides the output layer, each
convolution or deconvolution layer is followed by a Relu activation layer. The residua block is firstly
represented by He et al. [40] in 2016. After two liner transformations, the input data and its initial
value is added through a “shortcut” and then the added value is inputted to the Relu activation
layers. The whole structure of the style transfer network is shown as Figure 5.
Figure 5. Style transfer network.
Sustainability 2019, 11, 5673 7 of 15
4.4. Loss Network and Loss Function
The input of the loss network contains three parts, the style image, the stylized image generated
by the style transform network, and the content image. The loss network adopts the first few layers
of the VGG19 to extract the features of images and its structure and workflow is shown as Figure 6.
Figure 6. Loss network.
In previous sections, we analyze the ability of feature extraction for different layers in VGG19.
The shallower convolutional layers extract the lower features of the image, thus preserving a large
amount of detailed information. While the deeper convolution layers can extract higher features in
the image, thereby preserving the style information of the image. According to the above rules, we
finally adopt “Conv3_1” in VGG19 to extract the content features. Similarly, we adopt “Conv2_1”,
“Conv3_1”, “Conv4_1”, and “Conv5_1” in VGG19 to extract the style features.
Content loss describes the difference of features between stylized image and content image. It
can be calculated using Formula 5.
(y,y)= 1
𝐵𝑆𝑛
𝑛
𝑛
(𝑐,,
−𝑐̂,,
)



(5)
where 𝐵𝑆 denotes the batch size of the input data. 𝑛
, 𝑛
, and 𝑛
denotes the height, width, and
number of channels of the l-th layer, respectively. 𝑐,,
and 𝑐̂,,
represent the 𝑖×𝑗-th value of the
k-th channel after the content image and stylized image are activated by the l-th layer of VGG19.
Gram matrix can be seen as an eccentric covariance matrix between features of an image. It can
reflect the correlation between the two features. Additionally, the diagonal elements of the Gram
matrix also reflect the trend of each feature that appears in the image. Gram matrix can measure the
features of each dimension and the relationship between different dimensions. Therefore, it can
reflect the general style of the entire image. We only need to compare the Gram matrix between
different images to represent the difference of their styles. The gram matrix can be calculated using
Formula 6.
𝐺,
(𝑦)=𝑐,,
∙𝑐̂,,


(6)
where 𝑘 and 𝑘 both denote the number of channels in the 𝑙-th layer.
Style loss means the difference between the Gram matrix of the stylized image and the Gram
matrix of the style image. It can be calculated using Formula 7.

(y,y)= 1
𝑛
𝑛
𝑛
(𝐺
,
(𝑦)−𝐺
,
(𝑦))


(7)
Sustainability 2019, 11, 5673 8 of 15
where 𝐺(𝑦) and 𝐺(𝑦) denote the Gram matrix of extracted features in 𝑙-th layer for the style
image and stylized image.
Then we define the total style loss as the weight sum of all layers and it can be defined as Formula
8.
(y,y)=𝜆∙ℒ
(y,y) (8)
where 𝜆 denotes the weight of 𝑙-th layer.
In order to make the generated stylized image smoother, TV loss is introduced to be a regularizer
to increase the smoothness of the generated image. TV loss calculates the square of the difference
between each pixel and the next pixel in the horizontal and vertical directions. TV loss can be
calculated using Formula 9.
 =(𝑐
,, −𝑐,,)


 +  (𝑐,, −𝑐,,)



(9)
Finally, we can define the total loss as the weight sum of , , and . The total loss
can be represented as Formula 10.
 =𝛼
 +𝛽 +𝛾 (10)
where 𝛼, 𝛽 and 𝛾 are three adjustment factors and their values can be adjusted according to the
actual demand. We will have a discussion on their values in Section 5.3.
The final target in the training phase is to minimize . The weight of the style transfer
network will be updated according to the total loss by gradient descent algorithm.
4.5. Cross Training Strategy
In order to preserve as much content information as possible, we use a cross training method by
rotationally adopting different loss function. When the num of iteration is even, we adopt the original
total loss as loss function, otherwise, the content loss will be chosen as the loss function. The loss
function can be defined as Formula 11.
 =
𝛼ℒ +𝛽 +𝛾 𝑖
𝑓
𝑖𝑡𝑒𝑟𝑎𝑡𝑖𝑜𝑛%2 == 0
 𝑜𝑡ℎ𝑒𝑟𝑤𝑖𝑠𝑒 (11)
The workflow of the training process is shown as Figure 7.
Figure 7. Loss network.5. Experiments and Analysis.
5.1. Experiment Environment and Parameters
In order to have an evaluation of our presented algorithm, we compare it with the classic neural
style transfer algorithm proposed by Gatys et al. The experiment environment is a workstation using
Sustainability 2019, 11, 5673 9 of 15
the Ubuntu operation system. The workstation is also equipped with a GTX 1080 Ti (10G Memory)
graphics card to accelerate the training process of the style transfer network. The relevant software
versions are shown in the table below.
Table 1. Relevant software versions.
Software Name Versions
Python 3.7
TensorFlow-GPU 1.13.1
NumPy 1.14.6
SciPy 1.1.0
Matplotlib 3.02
Os 3.7
The specific information of convolution kernels in the style transfer network are illustrated in
Table 2.
Table 2. Parameters of convolution kernels.
Convolution Name Kernels Size, Stride, Number
Conv1 9×9, 1, 32
Conv2 3×3, 2, 64
Conv3 3×3, 2, 128
Res_block_Conv 3×3, 1, 128
Deconv1 3×3, 2, 64
Deconv2 3×3, 2, 32
Conv4 9×9, 1, 3
The relevant parameters of the network are listed in Table 3.
Table 3. Parameters for training.
Software Name Versions
Batch_Size 4
Training Data Size [256,512,1024]
Number of Training Data 5000
Content_layer Conv3_1
Style_layer Conv2_1, Conv3_1, Conv4_1, Conv5_1
Epoch 100
α, β, γ 1, [1,5,10], 1
Optimizer Adam
Learning Rate 0.001
5.2. Activation Function in Output Layer
In our proposed algorithm, the output layer of the style transfer network can adopt tanh or
sigmoid as its activation function. When adopting tanh as the activation function, the final output
will adopt Formula 12.
y=tanh
(x)122.5+122.5 (12)
where x is the output value of the previous layer. When adopting sigmoid as the activation function,
the final output will adopt Formula 13. In both ways, the output value of the output layer can be
between 0 and 255.
y=sigmoid
(x)∗255 (13)
In order to have an evaluation of two different activation functions in the output layer, we test
the performance of the network using the same content and style image. The experiment result is
shown as Figure 8. We can clearly see from Figure 8 that when adopting tanh as the activation
function, the performance is poor and only part of the image is stylized. Whereas, the style image can
Sustainability 2019, 11, 5673 10 of 15
be well integrated into the content image by the method of adopting sigmoid as the activation
function in the output layer.
Figure 8. Different activation functions in the output layer.
5.3. Loss Control Factors Adjustment
As we have discussed in the loss network, α,β,γ are three parameters to adjust the proportion
of content loss, style loss, and TV loss. The proportion of content loss and style loss has a significant
influence on the levels of stylization. When the content loss accounts for a larger proportion, the
stylized image will preserve more information of the content image. On the contrary, the stylized
image will be better rendered with the value of β increasing. TV loss only affects the smoothness of
the stylized image and it owns a small impact on the overall rendering effect. The rendering can be
strengthened by decreasing the value of α or increasing the value of β. Different applications can
retrofit the values of α and β based on their requirements. As Figure 9 illustrates, when β is 1, the
stylization is shallow and when β is increased to 10, the stylization is obvious.
Sustainability 2019, 11, 5673 11 of 15
Figure 9. Stylized image under different values of β.
5.4. Comparison of Details Preserving
In a stylized image, we expect the objects in it are still recognizable while the background is well
rendered. Classic neural style transfer achieves a great performance in image rendering, however,
it’s weak to preserve the details in the content image. In order to evaluate the performance of the
presented algorithm, we compare it with the classic neural style transfer in terms of details preserving.
Both two algorithms adopt the same content and style image with 1024 × 1024 pixels. Classic
neural style transfer iterates 1000 times for fully rendering while our presented algorithm iterates 100
epochs for fully training. The experiment result is shown as Figure 10. As Figure 10 illustrates, the
classic neural style transfer destroyed partial details from the content image. As you can clearly see
in the enlarged picture, that the pillars and the roof of the pavilion have different degrees of missing.
While, in our improved algorithm, those details are preserved and the background is well rendered.
Figure 10. Comparison of details preserving.
5.5. Comparison of Characters Rendering
Sometimes, characters are contained in the content image and commonly, we expect those
characters can preserve their original features rather than be rendered. In order to have an evaluation
of the presented algorithm in terms of characters rendering, we compare it with the classic neural
style transfer. Both two algorithms use the same content and style image with 1024 × 1024 pixels.
Classic neural style transfer iterates only 500 times to preserve more features of characters and our
presented algorithm still iterates 100 epochs for fully training. The experiment result is shown as
Figure 11. As we can clearly see from Figure 11, that both two algorithms achieve a good performance
in stylization. However, the classic neural style transfer algorithm stylizes the characters the same as
the background which results in the facial features, contours, etc., of the characters become blurred
and distorted. This can be explained as that classic neural style transfer adopts the optimization
method to convert the original image to the stylized image. Therefore, the network will treat each
pixel in the picture indiscriminately, making the content image close to the style image. While in our
presented algorithm, the trained deep neural network will recognize the characters in the image and
separate them from the background. Thus, the characters still keep complete features and clear
outlines.
Sustainability 2019, 11, 5673 12 of 15
Figure 11. Comparison of characters rendering.
5.6. Comparison of Time Consuming
Finally, we have an evaluation of our presented algorithm compared with the classic neural style
transfer in terms of time consuming. Images with different pixels are tested respectively. Both in the
classic neural style transfer and our present algorithm, the network needs to be adjusted to fit
different sizes of input data. Graphics Processing Unit (GPU) is only used when execute the two
different algorithms. Images with high resolution will have a better visual effect and meanwhile, it
takes more time for the network to render. Experiment result is shown in Table 4. As we can clearly
see from Table 4, the time consuming of both two algorithms increases with image pixel increasing.
For the image with the same pixel, our proposed algorithm achieves an enhancement of three orders
of magnitude compared with the classic neural style transfer. However, in our presented algorithm,
a long time is needed to train the style transfer network.
Table 4. Comparison of time consuming of different algorithms.
Algorithm Classic Neural Style Transfer Algorithm Ours
Image Size 100 Iterations 500 Iterations 1000
Iterations
Training
Time 1 Iteration
256 × 256 5.4 s 26.1 s 51.3 s 5 h 32 m 0.05 s
512 × 512 15.1 s 69.6 s 122.7 s 8 h 47 m 0.1 s
1024 × 1024 30.5 s 138.6 s 240.1 s 12 h 17 m 0.2 s
5.7. Other Examples Using Proposed Algorithm
Some other examples using the proposed algorithm are shown as Figure 12.
Sustainability 2019, 11, 5673 13 of 15
Figure 12. Examples using proposed algorithm.
6. Discussion
The convolution neural network owns an excellent ability for features extraction and it provides
an alternative way for the feature comparison between different images. The classic neural style
transfer algorithm regards the stylization task as an image-optimization-based online processing
problem. While, our presented algorithm regards it as a model-optimization-based offline processing
problem. The most prominent advantage of the presented algorithm is the short running time for
stylization. Since the network model can be trained in advance, it’s suitable for those delay sensitive
application especially real-time style transfer. Another advantage is that it can separate important
targets from the background to avoid the content loss. Contrary to image-optimization, model-
optimization aims to train a model to directly map the content image to the stylized image. The
training process makes the model capable to recognize different objects such as characters and
buildings, therefore, it can better preserve the details of those objects and separate them from the
background.
Meanwhile, there are also some demerits of the proposed algorithm. It’s inflexible to switch the
style. Once we want to change the style, we need to train a brand-new model which may take a lot of
time. However, the image-optimization-based method only needs to change the style image and then
iterates to the final solution. Additionally, our presented algorithm needs to run on the device with
better performance such as computation and memory which increases the cost.
7. Conclusions
Classic neural style transfer has the demerits of time consuming and details losing. In order to
accelerate the speed of stylization and improve the quality of the stylized image, in this paper, we
present an improved style transfer algorithm based on a deep feedforward neural network. A style
transfer network stacked by multiple convolution layers and a loss network work based on VGG19
are respectively constructed. Three different losses which represent the content, style, and
smoothness are defined in the loss network. Then, the style transfer network is trained in advance,
adopting the training set, and the loss is calculated by the loss network to update the weight of the
style transfer network. Meanwhile, a cross training strategy is adopted during the training process.
Our feature work will mainly focus on single model based multi-style transfer and special style
transfer combined with Generative Adversarial Networks (GAN).
Author Contributions: Z.G. conceived and designed the experiments; C.Z. and Y.G. performed the experiments
and analyzed the data. J.W. wrote this paper.
Sustainability 2019, 11, 5673 14 of 15
Acknowledgments: The authors are thankful to the editor and reviewers for their hard work which largely
improve the quality of this paper.
Conflicts of Interest: The authors declare no conflict of interest.
Data Availability: The data that support the findings of this study are available from the corresponding author
upon reasonable request.
References
1. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. Proc. Ieee Conf. Comput. Vis.
Pattern Recognition 2015, arXiv:1512.03385.
2. Zou, W.; Li, X.; Li, S. Chinese painting rendering by adaptive style transfer. Chin. Conf. Pattern Recognit.
Comput. Vision 2018, 3–14, doi:10.1007/978-3-030-03338-5_1.
3. Zheng, C.; Zhang, Y. Two-stage color ink painting style transfer via convolution neural network. In 2018
15th International Symposium on Pervasive Systems, Algorithms and Networks. 2018, doi:10.1109/i-
span.2018.00039.
4. Liu, S.; Guo, C.; Sheridan, J.T. A review of optical image encryption techniques. Opt. Laser Technol. 2014, 57,
327–342.
5. Wu, C.; Ko, J.; Davis, C.C. Imaging through strong turbulence with a light field approach. Opt. Express.
2016, 24, 11975–11986.
6. Gatys, L.A.; Ecker, A.S.; Bethge, M.A. Neural algorithm of artistic style. Proc. Ieee Conf. Comput. Vis. Pattern
Recognition 2015, arXiv:1508.06576
7. Karen, S.; Andrew, Z. Very deep convolutional networks for large-scale image recognition. Proc. Int. Conf.
Learn. Representations 2015, arXiv:1409.1556
8. Wang, J.; Gao, Y.; Liu, W.; Sangaiah, A.K.; Kim, H.J. An intelligent data gathering schema with data fusion
supported for mobile sink in wireless sensor networks. Int. J. Distrib. Sens. Networks 2019, 15,
doi:10.1177/1550147719839581.
9. Qiu, H.; Huang, X. An Improved image transformation network for neural style transfer. Proc. Int. Conf.
Inf. Syst. 2017, doi:10.1007/978-3-319-68121-4_28.
10. Wang, J.; Gu, X.; Liu, W.; Sangaiah, A.K.; Kim, H. An empower hamilton loop based data collection
algorithm with mobile agent for WSNs. Human-centric Computing and Information Sciences. 2019, 9, 18.
11. Zeng, H.; Liu, Y.; Li, S.; Che, J.; Wang, X. Convolutional neural network based multi-feature fusion for non-
rigid 3D model retrieval. J. Inf. Process. Systems 2018, 14, 176–190.
12. Daru, P.; Gada, S.; Chheda, M.; Raut, P. Neural style transfer to design drapes. Proc. Ieee Conf. Comput. Intell.
Comput. Res. 2017, 1–6, arXiv:1707.09899.
13. Pan, J.S.; Kong, L.P.; Sung, T.W.; Tsai, P.W.; Snasel, V. Alpha-fraction first strategy for hierarchical wireless
sensor networks. J. Internet Technol. 2018, 19, 1717–1726.
14. Johnson, J.; Alahi, A.; Li, F.-F. Perceptual losses for real-time style transfer and super-resolution. Proceedings
of European Conference on Computer Vision. 2016; pp. 694–711.
15. Qiu, X.; Jia, W.; Li, H. A font style learning and transferring method based on strokes and structure of
Chinese characters. Proc. Int. Conf. Comput. Sci. Serv. System 2012, 1836–1839, doi:10.1109/CSSS.2012.457.
16. Pan, J.S.; Lee, C.Y.; Sghaier, A.; Zeghid, M.; Xie, J.F. Novel systolization of subquadratic space complexity
multipliers based on toeplitz matrix–vector product approach. IEEE Trans. Very Large Scale Integr. 2019, 27,
1614–1622.
17. Azadi, S.; Fisher, M.; Kim, V.G.; Wang, Z.; Shechtman, E.; Darrell, T. Multi-content gan for few-shot font
style transfer. Proceedings of the IEEE conference on computer vision and pattern recognition. 2018, 7564–7573,
arXiv:1712.00516.
18. Wang, J.; Gao, Y.; Liu, W.; Sangaiah, A.K.; Kim, H.J. Energy efficient routing algorithm with mobile sink
support for wireless sensor networks. Sensors 2019, 19, 1494.
19. Nguyen, T.T.; Pan, J.S.; Dao, T.K. An improved flower pollination algorithm for optimizing layouts of
nodes in wireless sensor network. Ieee Access 2019, 7, 75985–75998.
20. Meng, Z.Y.; Pan, J.S.; Tseng, K.K. PaDE: An enhanced differential evolution algorithm with novel control
parameter adaptstion schemes for numerical optimization. Knowl.-Based Systems 2019, 168, 80–99.
21. Pan, J.S.; Kong, L.P.; Sung, T.W.; Tsai, P.W.; Snasel, V. A clustering scheme for wireless sensor networks
based on genetic algorithm and dominating Set. J. Internet Technol. 2018, 19, 1111–1118.
Sustainability 2019, 11, 5673 15 of 15
22. Wu, T.Y.; Chen, C.M.; Wang, K.H.; Meng, C.; Wang, E.K. A provably secure certificateless public key
encryption with keyword search. J. Chin. Inst. Eng. 2019, 42, 20–28.
23. Liu, J.; Yang, W.; Sun, X.; Zeng, W. Photo stylistic brush: Robust style transfer via superpixel-based bipartite
graph. Ieee Trans. Multimed. 2017, 20, 1724–1737.
24. Wang, J.; Gao, Y.; Wang, K.; Sangaiah, A.K.; Lim, S.J. An affinity propagation-based self-adaptive clustering
method for wireless sensor networks. Sensors 2019, 19, 2579.
25. Wang, J.; Gao, Y.; Yin, X.; Li, F.; Kim, H.J. An enhanced PEGASIS algorithm with mobile sink support for
wireless sensor networks. Wirel. Commun. Mob. Computing 2018, 2018, 9472075.
26. Ghrabat, M.J.J.; Ma, G.; Maolood, I.Y.; Alresheedi, S.S.; Abduljabbar, Z.A. An effective image retrieval based
on optimized genetic algorithm utilized a novel SVM-based convolutional neural network classifier. Hum.
-Cent. Comput. Inf. Sci. 2019, 9, 31.
27. Zeng, D.; Dai, Y.; Li, F.; Wang, J.; Sangaiah, A.K. Aspect based sentiment analysis by a linguistically
regularized CNN with gated mechanism. J. Intell. Fuzzy Systems 2019, 36, 3971–3980.
28. Zhang, L.; Wang, Y. Stable and refned style transfer using zigzag learning algorithm. Neural Process. Lett.
2019, doi:10.1007/s11063-019-10024-w.
29. Tu, Y.; Lin, Y.; Wang, J.; Kim, J.U. Semi-supervised learning with generative adversarial networks on digital
signal modulation classification. Comput. Mater. Continua. 2018, 55, 243–254.
30. Li, C.; Liang, M.; Song, W.; Xiao, K. A multi-scale parallel convolutional neural network based intelligent
human identification using face information. J. Inf. Process. Systems 2018, 14, 1494–1507.
31. Liu, D.; Yu, W.; Yao, H. Style transfer with content preservation from multiple images. Proc. Pac. Rim Conf.
Multimed. 2017, doi:10.1007/978-3-319-77380-3_75.
32. Hu, J.; He, K.; Hopcroft, J.E.; Zhang, Y. Deep compression on convolutional neural network for artistic style
transfer. Proc. Natl. Conf. Theor. Comput. Sci. 2017, doi:10.1007/978-981-10-6893-5_12.
33. Wang, L.; Wang, Z.; Yang, X.; Hu, S.; Zhang, J. Photographic style transfer. Vis. Computer. 2018,
doi:10.1007/s00371-018-1609-4.
34. Zhang, W.; Li, G.; Ma, H.; Yu, Y. Automatic color sketch generation using deep style transfer. Ieee Comput.
Graph. Applicat. 2019, 39, 26–37.
35. Zhao, H. H.; Rosin, P.L.; Lai, Y.K.; Lin, M.G.; Liu, Q.Y. Image neural style transfer with global and local
optimization fusion. Ieee Access 2019, 7, 85573–85580.
36. Yoon, Y.B.; Kim, M.S.; Choi, H.C. End-to-end learning for arbitrary image style transfer. Electron. Lett. 2018,
54, 1276–1278.
37. Liu, Y.; Xu, Z.; Ye, W.; Zhang, Z.; Weng, S.; Chang, C.C.; Tang, H. Image neural style transfer with
preserving the salient regions. Ieee Access 2019, 7, 40027–40037.
38. Chen, Y.; Lai, Y.; Liu, Y. CartoonGAN: Generative adversarial networks for photo cartoonization. Proc. Ieee
Conf. Comput. Vis. Pattern Recognition 2018, doi: 10.1109/CVPR.2018.00986
39. Chen, X.; Xu, C.; Yang, X.; Song, L.; Tao, D. Gated-gan: Adversarial gated networks for multi-collection
style transfer. Ieee Trans. Image Process. 2018, 28, 546–560.
40. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. Proc. Ieee Conf. Comput. Vis.
Pattern Recognition 2016, 770–778, arXiv:1512.03385
© 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access
article distributed under the terms and conditions of the Creative Commons Attribution
(CC BY) license (http://creativecommons.org/licenses/by/4.0/).
... Neural Style Transfer was initially proposed by Gatys et al. [10]. Similarly, research on Neural Style Transfer was conducted by Zhou et al. [11] using the VGG19 pre-trained model architecture on the ImageNet dataset, requiring 5 hours 32 minutes with an image size of 256 x 256 for training the model. Further, another study conducted by Irawan and Widjaja [12] has generated Batik patterns by transferring natural styles through Neural Transfer Style using VGG-16 architecture. ...
... The loss function will force the generator to form repeated batik patterns. Therefore, the batik results from the generator will be cropped into 16 images divided into two blocks, including block_y with index {1, 3,6,8,9,11,14, 16} and blok_z with index {2, 4, 5, 7, 10, 12, 13, 15} [24]. The loss calculation is also acknowledged as a reconstruction error [28], illustrated in the following equation (1). ...
Article
Every craftsman who draws Batik can also not necessarily draw various types of Batik. On the other hand, it takes a long time ranging from weeks to months, to make Batik. Image generation is regarded as an essential part of the field of computer vision. One of the popular methods includes the Generative Adversarial Network, commonly implemented to generate a new data set from an existing one. One model of the Generative Adversarial Network is BatikGAN SL generating batik images by inserting the two Batik patterns to produce a new Batik image. Currently, the generated Batik image does not maintain the input of the Batik pattern. Therefore, this study proposes a GAN model of BatikGAN SL, with the addition of a content loss function using hyperparameters to weight the content loss function. The content loss function is added from the Neural Transfer Style method. Previously, the style loss function in this method has been implemented in BatikGAN SL, and the dataset consists of Batik patches (326 images) and real Batik (163 images). This paper compares the BatikGAN SL model from previous studies with the BatikGAN SL model by implementing hyperparameters on the content loss function. The evaluation is conducted with FID, containing FID Local and FID Global. The results obtained in this study include a collection of Batik images, test evaluation value of 42 on FID Global and 16 on FID Local. These results are obtained by implementing the content loss function with a weight value of 1.
... 4. L Total [22] function considers per-pixel reconstruction accuracy and targets how fluently predicted hole values move to adjacent surroundings. L Total is calculated by adding L Perceptual loss [16], L styleOut [16], L styleComp [16], L tv [51] (Total variation loss), L valid [44] and L hole [44]. The weights given to these loss terms are shown as follows: ...
Article
Full-text available
In computer vision, image inpainting is a famous problem to automatically reconstruct the damaged part of the image according to the undamaged portion of an image. Inpainting irregular damaged areas in the image is still challenging. Deep learning-based techniques have given us a fantastic performance over the last few years. In this paper, we propose VResNet, a deep-learning approach for image inpainting, inspired by U-Net architecture and the residual framework. Since deeper neural networks are extra hard to train, the superficial convolution block in U-Net architecture is replaced by the residual learning block in the proposed approach to simplify the training of deeper neural networks. To develop an effective and adaptable model, an extensive series of experiments was conducted using the Paris-Street-View dataset. Our proposed method achieved notable results, including a PSNR of 20.65, an SSIM of 0.65, an L1 Loss of 6.90, and a total loss (LTotal\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$_{\hbox {Total}}$$\end{document}) of 0.30 on the Paris-Street-View dataset. These outcomes clearly demonstrate the superior performance of our model when compared to other techniques. The paper presents both qualitative and quantitative comparisons to provide a comprehensive assessment of our approach.
... [47] uses Vision Transformers for image style transfer. There can be many more such fascinating use cases for NST shortly based on the user requirements discussed in [48][49][50][51][52][53][54][55]. Most of the current work uses traditional neural networks with iterative training mechanisms in solving the Video Style Transfer problem, which may lead to the time-consuming training of the whole model. ...
Article
Full-text available
Neural Style Transfer (NST) is a class of software algorithms that allows us to transform scenes, change/edit the environment of a media with the help of a Neural Network. NST finds use in image and video editing software allowing image stylization based on a general model, unlike traditional methods. This made NST a trending topic in the entertainment industry as professional editors/media producers create media faster and offer the general public recreational use. In this paper, the current progress in Neural Style Transfer with all related aspects such as still images and videos is presented critically. The authors looked at the different architectures used and compared their advantages and limitations. Multiple literature reviews focus on the Neural Style Transfer of images and cover Generative Adversarial Networks (GANs) that generate video. As per the authors’ knowledge, this is the only research article that looks at image and video style transfer, particularly mobile devices with high potential usage. This article also reviewed the challenges faced in applying for video neural style transfer in real-time on mobile devices and presents research gaps with future research directions. NST, a fascinating deep learning application, has considerable research and application potential in the coming years.
... Therefore, a sustainable system-the one that maintains its performance against a change in the object's nature-is required for the correct recognition of complex objects [3]. Object classification is the key to a sustainable visual surveillance system [4]. Besides the latter, object classification finds its application in numerous domains, including intelligent robotics, face and action recognition, video watermarking, pedestrian tracking, autonomous vehicles, semantic scene analysis, content-based image retrieval, and many more. ...
Article
Full-text available
With an overwhelming increase in the demand of autonomous systems, especially in the applications related to intelligent robotics and visual surveillance, come stringent accuracy requirements for complex object recognition. A system that maintains its performance against a change in the object’s nature is said to be sustainable; and it has become a major area of research for the computer vision research community in the past few years. In this work, we present a sustainable deep learning architecture, which utilizes multi-layer deep feature fusion and selection, for accurate object classification. The proposed approach comprises three steps: 1) By utilizing two deep learning architectures, Very Deep Convolutional Networks for Large-Scale Image Recognition and Inception V3, it extracts features based on transfer learning, 2) Fusion of all the extracted feature vectors is performed by means of a parallel maximum covariance approach, and 3) The best features are selected using Multi Logistic Regression controlled Entropy-Variances method. For verification of the robust selected features, the Ensemble Learning method named Subspace Discriminant Analysis is utilized as a fitness function. The experimental process is conducted using three publicly available datasets, including Caltech-101, Birds database, Butterflies database, and CIFAR-100, and a ten-fold validation process yields the best accuracies of 95.5%, 100%, 98%, and 68.80% for the datasets respectively. Based on the detailed statistical analysis and comparison with the existing methods, the proposed selection method gives significantly more accuracy. Moreover, the computational time of the proposed selection method is better for real-time implementation.
Chapter
Neural Style Transfer (NST) is an emerging and known technique for convolutional neural networks by generating images of distinct artistic styles. Humans have championed the skill of creating extraordinary visual experiences by blending content and image style. It might be challenging to render an image’s semantic content in many visual styles. Perhaps a key constraint for earlier methods has been the absence of explicit representations of semantic information in images that make it possible to distinguish between the content and style of a picture. Image Steganography is the practice of concealing sensitive information by embedding the information in an audio, video, image or text file. This paper explores the use of Image Steganography on abstract images produced using Neural Style Transfer to conceal sensitive information.KeywordsNeural style transferImage style and contentMethods PerformanceDeep neural network
Article
Abstract Purpose Based on the highlights of The Metropolitan Museum of Art's collection, the purpose of this paper is to examine the similarities and differences between the subject keywords tags assigned by the museum and those produced by three computer vision systems. Design/methodology/approach This paper uses computer vision tools to generate the data and the Getty Research Institute's Art and Architecture Thesaurus (AAT) to compare the subject keyword tags. Findings This paper finds that there are clear opportunities to use computer vision technologies to automatically generate tags that expand the terms used by the museum. This brings a new perspective to the collection that is different from the traditional art historical one. However, the study also surfaces challenges about the accuracy and lack of context within the computer vision results. Practical implications This finding has important implications on how these machine-generated tags complement the current taxonomies and vocabularies inputted in the collection database. In consequence, the museum needs to consider the selection process for choosing which computer vision system to apply to their collection. Furthermore, they also need to think critically about the kind of tags they wish to use, such as colors, materials or objects. Originality/value The study results add to the rapidly evolving field of computer vision within the art information context and provide recommendations of aspects to consider before selecting and implementing these technologies.
Article
Full-text available
In the poultry evisceration, the inconsistent position of visible carcass and invisible viscera may be one of the most neglected problems, which increase the level of visceral damage by gripping manipulator. In order to detect the position of the chicken carcass and viscera, a computer vision‐based automation system is developed to extract the region of interest from each chicken in the image acquisition system. The segmentation method of carcass is proposed by applying the color space transformation and the threshold segmentation. This method combines with several operations of morphology and reconstruction to remove the other regions such as wings and legs. After a midline abdominal incision of the chicken, the viscera is segmented using the active contour algorithm. Following data analysis, the change trend of relative position is presented between carcass and viscera. The average longitudinal deviation value of weighting 1,000–1,500 g and 1,500–2000 g chickens were 22.9 and 32.4 mm, separately. These results indicate that the relative position is significantly changed between carcass and viscera with chicken size in the longitudinal direction. In this study, machine vision technology can be satisfactorily applied to predict the position of chicken viscera, which can provide technical support for poultry processing. Practical applications For a long time, avian influenza virus and various pathogenic microorganism have been threatening the breeding and slaughtering of poultry, which can easily damage the health of people and cause widespread transmission of the virus. Therefore, automated evisceration technology by intelligent robot will become the development trend of the poultry slaughtering. However, due to the invisibility of the poultry viscera, it will easily cause damage of the internal organs during the manipulator evisceration. In this study, in order to avoid repetitive and boring work and manual operation errors, machine vision technology was introduced to automatically obtain the relative position between the visible carcass and the invisible viscera of individual chicken based on image segmentation algorithms. Thus, the position of the viscera can be predicted based on the data analysis and the carcass location, which provide useful information to guide the robot for evisceration, and it is also helpful to conduct further research in the slaughtering and processing of other poultry.
Article
Full-text available
A wireless sensor network (WSN) is an essential component of the Internet of Things (IoTs) for information exchange and communication between ubiquitous smart objects. Clustering techniques are widely applied to improve network performance during the routing phase for WSN. However, existing clustering methods still have some drawbacks such as uneven distribution of cluster heads (CH) and unbalanced energy consumption. Recently, much attention has been paid to intelligent clustering methods based on machine learning to solve the above issues. In this paper, an affinity propagation-based self-adaptive (APSA) clustering method is presented. The advantage of K-medoids, which is a traditional machine learning algorithm, is combined with the affinity propagation (AP) method to achieve more reasonable clustering performance. AP is firstly utilized to determine the number of CHs and to search for the optimal initial cluster centers for K-medoids. Then the modified K-medoids is utilized to form the topology of the network by iteration. The presented method effectively avoids the weakness of the traditional K-medoids in aspects of the homogeneous clustering and convergence rate. Simulation results show that the proposed algorithm outperforms some latest work such as the unequal cluster-based routing scheme for multi-level heterogeneous WSN (UCR-H), the low-energy adaptive clustering hierarchy using affinity propagation (LEACH-AP) algorithm, and the energy degree distance unequal clustering (EDDUCA) algorithm.
Article
Full-text available
The arrangement of nodes impacts the quality of connectivity and energy consumption in wireless sensor network (WSN) for prolonging the lifetime. This paper presents an improved flower pollination algorithm based on a hybrid of the parallel and compact techniques for global optimizations and a layout of nodes in WSN. The parallel enhances diversity pollinations for exploring in space search and sharing computation load. The compact can save storing variables for computation in the optimization process. In the experimental section, the selected test functions and the network topology issue WSN are used to test the performance of the proposed approach. Compared results with the other methods in the literature show that the proposed algorithm achieves the practical way of reducing the number of its stored memory variables and running times.
Article
Full-text available
Abstract Image retrieval is the process of retrieving images from a database. Certain algorithms have been used for traditional image retrieval. However, such retrieval involves certain limitations, such as manual image annotation, ineffective feature extraction, inability capability to handle complex queries, increased time required, and production of less accurate results. To overcome these issues, an effective image retrieval method is proposed in this study. This work intends to effectively retrieve images using a best feature extraction process. In the preprocessing of this study, a Gaussian filtering technique is used to remove the unwanted data present in the dataset. After preprocessing, feature extraction is applied to extract features, such as texture and color. Here, the texture feature is categorized as a gray level cooccurrence matrix, whereas the novel statistical and color features are considered image intensity-based color features. These features are clustered by k-means clustering for label formation. A modified genetic algorithm is used to optimize the features, and these features are classified using a novel SVM-based convolutional neural network (NSVMBCNN). Then, the performance is evaluated in terms of sensitivity, specificity, precision, recall, retrieval and recognition rate. The proposed feature extraction and modified genetic algorithm-based optimization technique outperforms existing techniques in experiments, with four different datasets used to test the proposed model. The performance of the proposed method is also better than those of the existing (RVM) regression vector machine, DSCOP, as well as the local directional order pattern (LDOP) and color co-occurrence feature + bit pattern feature (CCF + BPF) methods, in terms of the precision, recall, accuracy, sensitivity and specificity of the NSVMBCNN.
Article
Full-text available
This paper presents a new image synthesis method for image style transfer. For some common methods, the textures and colors in the style image are sometimes applied inappropriately to the content image, which generates artifacts. In order to improve the results, we propose a novel method based on a new strategy that combines both local and global style losses. On the one hand, a style loss function based on a local approach is used to keep the style details. On the other hand, another style loss function based on global measures is used to capture the more global structural information. Results on various images show that the proposed method reduces artifacts while faithfully transferring the style image’s characteristics and preserving the structure and color of the content image.
Article
Full-text available
In wireless sensor networks (WSNs), sensor devices must be equipped with the capabilities of sensing, computation and communication. These devices work continuously through non-rechargeable batteries under harsh conditions, the batter span of nodes determines the whole network lifetime. Network clustering adopts an energy neutral approach to extend the network life. The clustering methods can be divided into even and uneven clustering. If even clustering is adopted, it will cause the cluster head nodes (CHs) in vicinity of the base station to relay more data and cause energy hole phenomenon. Therefore, we adopt a non-uniform clustering method to alleviate the problem of energy hole. Furthermore, to further balance and remit resource overhead of the entire network, we combined the PEGASIS algorithm and the Hamilton loop algorithm, through a mixture of single-hop and multiple hops mechanisms, inserting a mobile agent node (MA) and designing an optimal empower Hamilton loop is obtained by the local optimization algorithm. MA is responsible for receiving and fusing packet from the CHs on the path. Network performance results show that the proposed routing algorithm can effectively prolong network lifetime, equalize resource expenditure and decrease the propagation delay.
Article
Full-text available
Recently, wireless sensor network (WSN) has drawn wide attention. It can be viewed as a network with lots of sensors that are autonomously organized and cooperate with each other to collect, process, and transmit data around targets to some remote administrative center. As such, sensors may be deployed in harsh environments where it is impossible for battery replacement. Therefore, energy efficient routing is crucial for applications that introduce WSNs. In this paper, we present an energy efficient routing schema combined with clustering and sink mobility technology. We first divide the whole sensor field into sectors and each sector elects a Cluster Head (CH) by calculating its members’ weight. Member nodes calculate energy consumption of different routing paths to choose the optimal scenario. Then CHs are connected into a chain using the greedy algorithm for intercluster communication. Simulation results prove the presented schema outperforms some similar work such as Cluster-Chain Mobile Agent Routing (CCMAR) and Energy-efficient Cluster-based Dynamic Routing Algorithm (ECDRA). Additionally, we explore the influence of different network parameters on the performance of the network and further enhance its performance.
Article
Full-text available
Recently, style transfer based on the convolutional neural network has achieved remarkable results. In this paper, we extend the original neural style transfer algorithm to ameliorate the instability in the reconstruction of certain structural information, and improve the ghosting artefacts in the background of image which with low texture and homogeneous areas. For that end, we adopt zigzag learning strategy: The model parameters are optimized to an intermediate target firstly, then let the model converge to the final goal. We show the zigzag learning to multi-sample model which is fabricated from resampling the style input and to loss function that is split into two sections. And also, we demonstrate experimentally the effectiveness of the proposed algorithm and provide its theoretical analysis. Finally we show how to integrate the zigzag learning strategy in fast neural style transfer framework.
Article
Full-text available
Numerous tiny sensors are restricted with energy for the wireless sensor networks since most of them are deployed in harsh environments, and thus it is impossible for battery re-change. Therefore, energy efficiency becomes a significant requirement for routing protocol design. Recent research introduces data fusion to conserve energy; however, many of them do not present a concrete scheme for the fusion process. Emerging machine learning technology provides a novel direction for data fusion and makes it more available and intelligent. In this article, we present an intelligent data gathering schema with data fusion called IDGS-DF. In IDGS-DF, we adopt a neural network to conduct data fusion to improve network performance. First, we partition the whole sensor fields into several subdomains by virtual grids. Then cluster heads are selected according to the score of nodes and data fusion is conducted in CHs using a pretrained neural network. Finally, a mobile agent is adopted to gather information along a predefined path. Plenty of experiments are conducted to demonstrate that our schema can efficiently conserve energy and enhance the lifetime of the network.
Article
Recently, sentiment analysis has become a focus domain in artificial intelligence owing to the massive text reviews of modern networks. The fast increase of the domain has led to the spring up of assorted sub-Areas, researchers are also focusing on subareas at various levels. This paper focuses on the key subtask in sentiment analysis: Aspect-based sentiment analysis. Unlike feature-based traditional approaches and long short-Term memory network based models, ourwork combines the strengths of linguistic resources and gating mechanism to propose an effective convolutional neural network based model for aspect-based sentiment analysis. First, the proposed regularizers from the real world linguistic resources can be of benefit to identify the aspect sentiment polarity. Second, under the guidance of the given aspect, the gating mechanism can better control the sentiment features. Last, the basic structure of model is convolutional neural network, which can perform parallel operations well in the training process. Experimental results on SemEval 2014 Restaurant Datasets demonstrate our approach can achieve excellent results on aspect-based sentiment analysis.
Article
Systolic finite field multiplier over $GF(2^{m})$ , because of its superior features such as high throughput and regularity, is highly desirable for many demanding cryptosystems. On the other side, however, obtaining high-performance systolic multiplier with relatively low hardware cost is still a challenging task due to the fact that the systolic structure usually involves large area complexity. Based on this consideration, in this paper, we propose to carry out two novel coherent interdependent efforts. First, a new digit-serial multiplication algorithm based on polynomial basis over binary field $(GF(2^{m}))$ is proposed. Novel Toeplitz matrix–vector product (TMVP)-based decomposition strategy is employed to derive an efficient subquadratic space complexity. Second, The proposed algorithm is then innovatively mapped into a low-complexity systolic multiplier, which involves less area-time complexities than the existing ones. A series of resource optimization techniques also has been applied on the multiplier which optimizes further the proposed design (it is the first report on digit-serial systolic multiplier based on TMVP approach covering all irreducible polynomials, to the best of our knowledge). The following complexity analysis and comparison confirm the efficiency of the proposed multiplier, that is, it has lower area-delay product (ADP) than the existing ones. The extension of the proposed multiplier for bit-parallel implementation is also considered in this paper.