Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Style Model Training Independent of Optical Flow? #13

Open
JeffMusgrave opened this issue Oct 13, 2018 · 13 comments
Open

Style Model Training Independent of Optical Flow? #13

JeffMusgrave opened this issue Oct 13, 2018 · 13 comments
Labels
question Further information is requested

Comments

@JeffMusgrave
Copy link

What is the fastest way to train a new style? Is it necessary to train an optical flow model on the hollywood dataset, for every new style?

@manuelruder
Copy link
Owner

You just need to create the hollywood dataset once, and you can use it for every new style you want to train. You can use a pretrained optical flow model, as described in the FlowNet2 repository.

Since you are asking for the "fastest way": you could in theory only use the COCO dataset and then only use the "shift", "zoom_out" and "single_image" as the datasource parameter. Results may be inferior, since the model only learned from camera motion, but maybe sufficient for your use case.

@JeffMusgrave
Copy link
Author

By the way, how long did it take you to train the hollywood dataset?

@noufpy
Copy link

noufpy commented Oct 18, 2018

@jargonfilter I computed the optical flow with the full Hollywood dataset and it took me about 1.5-2 weeks. Produced really high quality results.
Machine Specs: Linux (Ubuntu 16.04), GPU 24 GB, RAM 30 GB, CPUS 8

@manuelruder
Copy link
Owner

manuelruder commented Oct 18, 2018

I computed the optical flow on our university cluster with multiple jobs in parallel (there is no "training" involved by the way), but I don't remember the exact number of GPU days. Thanks for some reference numbers, noufali.

This is quite a long time. As pointed out in the description, the amount of data can be reduced to one fifth with a simple parameter switch. Concurrent work on video style transfer (Gupta et al., Chen et al.) use smaller datasets, too, thus I don't expect the quality to drop significantly.

@manuelruder manuelruder added the question Further information is requested label Oct 18, 2018
@bafonso
Copy link

bafonso commented Dec 9, 2018

@manuelruder do you mean the num_tuples_per_scene parameter that should be set lower than 5 ?
@noufali Any chance that optical flow dataset could be made available for download? I only have a normal 1070 and a i5 8400, god knows how long that will take :(

@manuelruder
Copy link
Owner

@bafonso yes, exactly. Can be as low as 1. The script first separates all the video clips by scene, then it ranks every possible tuple in each scene by amount of motion. Then it will take the top num_tuples_per_scene tuples per scene. Reducing this is the most natural way to reduce the dataset size, since this will mostly remove similar looking tuples. The size can be further reduced by deleting random files in "AVIClipsScenes" before executing the scripts. The scripts will dynamically detect the number of video files available.

@noufpy
Copy link

noufpy commented Dec 11, 2018

@bafonso Hey there! yea I can share mine with you. It's definitely a tedious process. Shoot me your email?

@AIaesthetic
Copy link

I'm wondering if you could share that with me also @noufali. my email is
[email protected]

That would be a lifesaving

@ryanqiutu
Copy link

@noufali Excuse me! I've calculated flow files using deepflow and reliable file using consistencyChecker of the part dataset of Hollywood. But I'm not sure the results are right. I tested the calculated reliable results with the occlusions results of MPI-Sintel datset(in my thought, they should be the same), the results are different. Could you show me some calculated results of flow file and reliable file?

@pizboese
Copy link

pizboese commented Apr 8, 2020

@noufali Sorry to bother and revive this old thread. I just started to work and explore video style transfers on some consumer hardware. It would take ages for me to compute the optical flow of the hollywood dataset. Would it be possible for you to share yours, in case you still have it.
Thanks in advance! ([email protected])

@ryletko
Copy link

ryletko commented May 13, 2020

@noufali Please could you share it with me too? I'm a newbie in the topic and don't have proper hardware and skills for the computations, but I'm very curious about this. My email is [email protected] Thank you!

@StoneCypher
Copy link

@noufali - I would also like the optical flow. Would you consider putting it in a repo, so that people don't have to ask you directly?

@chen-jimmy
Copy link

@noufali Hi, sorry to bother you but I'm currently working on a style transfer project and this step is taking ages to complete on my hardware. If you could share your optical flow results with me, I would deeply appreciate it. My email is [email protected]

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

10 participants