-
-
Notifications
You must be signed in to change notification settings - Fork 192
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Cannot get better Result like README #54
Comments
Did you train on the full dataset? I suggest put more results here and give more information about yout training, this may be userful for others to help you. I also met the problem for not able to get good result. Model trained on full dev set for ~4 epoch with 3 GTX 2080ti, with K=8 and the result is blurry, and I haven't run inference yet. |
@Jarvisss
Yes,I downloaded full dataset of VoxCeleb2 ,contains 18588 data, It seems that your batch size was very larger than me.( my batchsize was only 2...) Thank you. Nekomo |
For me, I actually got 5994 speakers and 145,569 videos in VoxCeleb2 dev
I use batchsize=6, as I have 3 gpus each 2 batch, The results I showed maybe misleading where 'batch' should be 'step' |
@Jarvisss
18588 means length of the dataLoader which is already preprocessed.
Actually,my dataset contains
and maybe my VoxCeleb2 dataset is broken...
Is 'step' means iteration during a epoch? And,if you do not mind , could you give me your pretrained model? Thanks, |
Yes, it is
ofcourse, how can I give you my model? |
Thanks a lot. Nekomo |
@nekomo Hi, sorry for late reply, I trained my model for another 10 epochs, and get comparable results to @vincent-thevenin, the result can be seen here. but when I ran the embedder_inference.py and finetuning_training.py, I also got ugly results. and if I do not finetune the model and feed forward directly, I will get result like this: There must be something wrong, I'm still debugging now. And I wonder what your result looks like without finetuning, could you please share the result? Jarvisss |
Hi Jarvisss, |
@mingkaihu
You may first skip step 4 to see if the result is resonable. If true, then do the fine_tuning and step 5 again to see if the result is better. Good luck, |
@Jarvisss
I ran embedeer_inference.py,and finetuning_training.py,so running model without finetuning was never tried.I also try that again.
I got it. I 'll modify my local code too. Thank you, |
Thanks a lot for your feedback, Jarvisss. |
@Jarvisss hi, I am so appreciated with your gread job. What is you newest code? it is "https://github.com/Jarvisss/Realistic-Neural-Talking-Head-Models", right? I knew you have chage emberdder_inference.py from 256 to 224: I figure it out, the network is trained on 224 * 224, but the code in embedder_inference.py and video_inference.py crops the input to 256 * 256, which in my case causes the ugly result.but when I read the code in "https://github.com/Jarvisss/Realistic-Neural-Talking-Head-Models", it is still 256. |
@tengshaofeng
The network is actually trained on 256 * 256, where the real input is 224 * 224 and zero-padding to 256 * 256. So the network still takes 256 * 256 as input, but I changed the cropped image to 224 * 224 and padding to 256 * 256 just like in training, instead of crop images to 256 * 256 and without padding during testing |
@Jarvisss , sorry, I am confused now. Should I change it from 256 to 224 in embedder_infernce.py, finetuning_training.py and webcam_inference.py? what is the different betweent code in the master branch with yours? |
@Jarvisss I really appliciate your support . thanks, |
@tengshaofeng Sorry for late reply, the code of my forked version (Jarvisss@da30930) was created for purpose of PR, and the code of crop was not added to that commit. By the way, what you should do is to crop the images to (224, 224), in
to make it consistent with the training data. yours, |
@Nenoko |
@Jarvisss Hi, can you upload your trained checkpoints to google drive to share with us? Thanks! |
@Jarvisss thanks for your reply.
is it right? I got the result like following: I do not thing the result is good. Can u give me some advice? |
What's the problem with your result? |
@Jarvisss Can you see my shared images? Do u think there exist mistakes in my steps? |
i see your result, but i dont understand whats the problem from the images you provide. The steps are first embed image to code ,fine tuning, and then inference, as the author suggest in readme, can you share the landmarks for inference |
|
@Jarvisss hi, your result is so cool, but I want to known, if do not finetune , what the result like? and, could you share your pre-trained model weight ? |
Hi .
I really appreciate for your Implementation.
I try to execute your program,but result is too ugly.
Here is my result.
I tried to run
train.py
for 28 epoch ,thenembedeer_inference.py
,and finallyfinetuning_training.py
for 150 epoch.I skipped some important training phase, did I?
If you have any suggestion about this result,please give me some advice.
(I'm not English speaker,so maybe this issue's grammer is wrong.I apologize for my poor English.)
Thanks,
Nekomo
The text was updated successfully, but these errors were encountered: