Real-Time Neural Style Transfer for Videos
Title | Real-Time Neural Style Transfer for Videos |
Publication Type | Conference Paper |
Year of Publication | 2017 |
Authors | Huang, H., Wang, H., Luo, W., Ma, L., Jiang, W., Zhu, X., Li, Z., Liu, W. |
Conference Name | 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) |
Date Published | July 2017 |
Publisher | IEEE |
ISBN Number | 978-1-5386-0457-1 |
Keywords | content information, fast style transfer, feed-forward convolutional neural networks, feed-forward network, feedforward neural nets, hybrid loss, image colour analysis, image sequences, image style, image texture, input frames, learning (artificial intelligence), Metrics, neural nets, neural style transfer, Optical imaging, optical losses, Optimization, pubcrawl, real-time neural style transfer, Real-time Systems, resilience, Resiliency, Scalability, style information, stylized video frames, temporal consistency, temporal information, temporal loss, temporally consistent stylized videos, trained network, Training, training stage, two-frame synergic training mechanism, video signal processing, video style transfer method, Videos |
Abstract | Recent research endeavors have shown the potential of using feed-forward convolutional neural networks to accomplish fast style transfer for images. In this work, we take one step further to explore the possibility of exploiting a feed-forward network to perform style transfer for videos and simultaneously maintain temporal consistency among stylized video frames. Our feed-forward network is trained by enforcing the outputs of consecutive frames to be both well stylized and temporally consistent. More specifically, a hybrid loss is proposed to capitalize on the content information of input frames, the style information of a given style image, and the temporal information of consecutive frames. To calculate the temporal loss during the training stage, a novel two-frame synergic training mechanism is proposed. Compared with directly applying an existing image style transfer method to videos, our proposed method employs the trained network to yield temporally consistent stylized videos which are much more visually pleasant. In contrast to the prior video style transfer method which relies on time-consuming optimization on the fly, our method runs in real time while generating competitive visual results. |
URL | https://ieeexplore.ieee.org/document/8100228 |
DOI | 10.1109/CVPR.2017.745 |
Citation Key | huang_real-time_2017 |
- temporal loss
- real-time neural style transfer
- real-time systems
- resilience
- Resiliency
- Scalability
- style information
- stylized video frames
- temporal consistency
- temporal information
- pubcrawl
- temporally consistent stylized videos
- trained network
- Training
- training stage
- two-frame synergic training mechanism
- video signal processing
- video style transfer method
- Videos
- image texture
- fast style transfer
- feed-forward convolutional neural networks
- feed-forward network
- feedforward neural nets
- hybrid loss
- image colour analysis
- image sequences
- image style
- content information
- input frames
- learning (artificial intelligence)
- Metrics
- neural nets
- neural style transfer
- Optical imaging
- optical losses
- optimization