A Neural Net Creates Movies In The Style Of Any Artist
Written by David Conrad   
Saturday, 14 May 2016

Deep Neural Networks have not only been impressing people with their AI ability, but also with their artistic abilities. The latest work can take the style of any artist and convert existing video into that style. You have to see it to believe it.

Neural networks are good at generalizing. For example, you can use a trained network to morph an input image into something that has the style of a reference painting. The result is that the photo looks as if it had been painted by the artist.

 

artystar

Star Wars in the Style of Kandinsky.

The process is a little more complicated than this because neural networks learn features which are about the content of the image and not style information. It is reasonable to suppose that the style information is in the correlations between the features. The stylised image is input to the network and its pixels are adjusted to minimise a loss function that includes its features and a measure of how far the feature correlations are from the image being used as the style example. Notice that this isn't a straightforward use of the neural network in the sense that the network doesn't learn the style of the artist - it just provides the features that are used to characterise the style.

So far so good, but what about creating a movie in the style of an artist using the same technique. If you try the obvious approach and apply the network and optimization to each frame in turn then the result isn't very good because things jump about - there is no temporal consistency. To make a a good movie in the style of an artist you have to convert each frame with temporal constraints on how things change.

This is the method investigated by the team at the University of Freiburg. They used a deep neural network approach to estimating the optical flow from one frame to the next  Optical flow gives you a measure of how things move and change between frames. The idea is to compare the optical flow between the original and the restyled image and include the difference in the loss function used in the optimization. So now the optimization tries to keep the changes between frames the same, as well as the original features and the correlations in the style image. 

The exact details of how to run such an optimization is complicated and you can read about them in the paper. However, it is worth saying that it was implemented on a Nvidia Titan X GPU and took about eight to ten minutes for each 1024x436 frame. 

Now you know how it works, take a look at the video that explains the different constraints applied. Notice the unconstrained example at the start where the applied styles jumps about all over the place. 

 

There is also a less technical video which is worth seeing to appreciate how well it works.

 

 

So it's fun but is it useful? 

You might not want to re-style an existing movie in the say the style of Picasso or "The Scream" but it opens up the possibly of creating new movies in the style of...

 artystar2

More Information

Artistic style transfer for videos

Manuel Ruder, Alexey Dosovitskiy, Thomas Brox

Related Articles

A Neural Net Colorizes Photos

Automatic Colorizing Photos With A Neural Net

TensorFlow - Googles Open Source AI And Computation Engine

Are Your Pictures Memorable?

Microsoft Wins ImageNet Using Extremely Deep Neural Networks 

Baidu AI Team Caught Cheating - Banned For A Year From ImageNet Competition 

The Allen Institute's Semantic Scholar

Removing Reflections And Obstructions From Photos

See Invisible Motion, Hear Silent Sounds Cool? Creepy?

Computational Camouflage Hides Things In Plain Sight

Google Has Software To Make Cameras Worse

 

To be informed about new articles on I Programmer, sign up for our weekly newsletter, subscribe to the RSS feed and follow us on, Twitter, Facebook or Linkedin.

 

Banner


Java 15 Reaches General Availability
17/09/2020

Oracle has announced that Java 15 is now generally available. The announcement was made in the opening keynote of Oracle Developer Live, an online version of the usual CodeOne and OpenWorld conference [ ... ]



Learn About Serverless with The Linux Foundation on edX
11/09/2020

A new free training course that explains serverless computing and provides first-hand experience in building and deploying code directly to a Kubernetes cluster has just launched on the edX platf [ ... ]


More News

 

graphics

 



 

Comments




or email your comment to: comments@i-programmer.info

 

 

 

 

Last Updated ( Wednesday, 08 April 2020 )