Hacker News new | past | comments | ask | show | jobs | submit login
Pictures combined using Neural networks (imgur.com)
362 points by olalonde on Feb 29, 2016 | hide | past | favorite | 52 comments




This actually seems to come from a new paper: http://arxiv.org/pdf/1601.04589v1.pdf


There's a fascinating animated movie "Loving Vicent" coming up that is made in the style of Vincent van Gogh. Apparently the individual frames are actual oil paintings made by hand.

http://www.thisiscolossal.com/2016/02/loving-vincent-trailer...

When I first saw the trailer my first thought was the neural network approach shown here. I wonder how close you could get to "Loving Vincent" style by simply applying this algorithm to individual video camera frames.


There's a not very well know software that's written by one guy what's been around for 10 year so which can do this for moving images with an amazing amount of flexibility. It's a lot of fun: http://synthetik.com/


What in the world... this is like stop motion animation taken to the extreme.


> Apparently the individual frames are actual oil paintings made by hand

I don't believe it


Isn't this thread enough to proof the movie wouldn't have to be hand-painted? I'm not going to find the Van Gogh themed NN posted before.


that's incredible


There's something too oddly natural in NN based images. The jpeg compression artifacts of the visual cortex.


I did this album:

http://imgur.com/a/Gliwg


Wow, this is art, seriously. Do you have any documentation on how you made these? What were the original images you used?


Is this the same code/algorithm as: https://github.com/jcjohnson/neural-style?


The authors of the paper have their own website providing similar service http://deepart.io/page/about/ (which claims a patent over this, unfortunately).


Now we just need another neural network to remove the watermark from the free version :)


Seems like it, the source for their website is here https://github.com/SergeyMorugin/ostagram and the bottom links to jcjohnson's repo


That is a Billion dollar photo app waiting to be published ;)



"Estimated waiting time: 9145.0 minutes"

This is not how you are going to earn billions ;)


It's "a bit" faster if you pay them. They wouldn't earn a penny doing it for free.


Well, yeah, but if they can't even make one single sample image for me, I'd rather not give them money. It's been more than a day and the image is still not done. It'd be faster for me to just install caffe and one of the open source solutions and do it myself. For free, as many as I want.


A Billion second photo app.


What about the sardine rolls? Surely there's a big business waiting to happen there!



Whelp, it looks like painting and graphic art is going the way of the dodo...


...the same way we don't need actors for movies anymore since we have 3D rendered special effects. /s

More likely it's going to become another tool to use (and heavily overuse for some time). The arts will be different but people won't stop being creative. The art of making cave paintings didn't really go away - now they are called graffity and are illegal in most modern caves.


Those pictures were definitely curated. In the near future, this means the neural net is still a tool for the artist.


Who would have thought that one of the first professions to be automated by neural nets would be that of artists?


Someone still has to come up with new art styles for the machine to mix.


Before that becomes the main problem, that other problem must be solved where you give it your picture and the Mona Lisa whose style is to be copied and the way it copies the style is it paints her eyebrow atop your mouth. 'Cause it's not like this thing can really distinguish between "style" and "content". What it does is it pastes patches from the "style" image onto the "content" image. Sometimes the result is interesting, much of the time it's between boring and terrifying. It's cool and all, but it's not much more intelligent than your favorite Photoshop filter.

"Going the way of the dodo..."


> Cause it's not like this thing can really distinguish between "style" and "content

So it's going to be perfect for modern art? :D


No, it can't distinguish because it doesn't know either.


If the rating is more important than the creation, perhaps they should create a neural network with an image as input, and a single signal as output, that indicates on a scale of 0 to 1 how interesting that image is, from an artistic point of view.

Alternatively, equip a human with an EEG headset, let them look at images, and take the rating from the output of the scanner.


Interesting is not a value on a one dimensional scale, because the art isn't one dimensional.


Build a database, randomly assemble.

Recurse.

Use whatever language generator deemed appropriate* to assign proper titles for the newly-evolved art movement.

*this may be delegated to a meat-computer, in pity.


Democratize art!


Found this while browsing /r/MachineLearning: https://www.reddit.com/r/MachineLearning/comments/47zxox/pic...


Reminds me about this music video: https://www.youtube.com/watch?v=DgPaCWJL7XI I guess they've used similar tech.


Looks like standard DeepDream to me so not really the same thing. In OP they combine pictures while this just apply DeepDream over and over until it looks like that.


Do you know how they get stable videos from DeepDream? I've seen it elsewhere too, but my initial guess would have been that DeepDream would give ridiculous results if you ran it on each frame individually. I expect it's sensitive to tiny differences between frames.


Awesome, would be even cooler if it could pick up specific features like faces and replicate the style. Now it kind of brushes over most faces unless it dominates the picture.


If I wanted to get into generating this kind of image, as an art project, where would I start, assuming I'm a component programmer?


Checkout http://deepart.io/. You'll have to pay a little bit but they provide a very convenient interface for doing this.


I'm interested in the question too and, if he's of similar mind to me, that's not what he meant.

I'm curious what it takes to make a piece of software like this. People say "Using NN" and stuff. But if I make an app with a way to load a picture and then add NN to it, this won't come out. I'm wondering what the specific function of the networks are here.


Definitely. I want to understand what's going on here and be able to apply some creativity to the process, not just use existing tools without understanding.



Is the combination symmetrical? I.e. would we get the same result combining A with B than combining B with A?


It would appear that in each of the examples, the result uses the top source image for "content" and the bottom one for "style". So no, it's not commutative.



Does this actually use the same technique, or is it just a clever filter?


is this different from neural-style? it seems to be. here is neural-style with an api: http://www.somatic.io/models/somatic/neural-style-demo


It it based on it


Beautiful :)




Consider applying for YC's Spring batch! Applications are open till Feb 11.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: