Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Some confusion on the training result. #18

Open
yyk-wew opened this issue Mar 3, 2021 · 3 comments
Open

Some confusion on the training result. #18

yyk-wew opened this issue Mar 3, 2021 · 3 comments

Comments

@yyk-wew
Copy link

yyk-wew commented Mar 3, 2021

Hi, I've tried to train my own xception model on LQ data of FF++ dataset and find that the model performs so unbalanced.
The accuracy is about 50% on real images while near 100% on fake images.
I've consider the ratio of the number of fake and real images (4:1), so I only choose a quarter of fake images to get a balance. But it doesn't work. The accuracy on real images is still below 60%.
Here's some training details:

  1. Use only c40 data. 720 for train, 140 for validate, 140 for test.
  2. Capture 1 frame per 10 frames each video.
  3. Train on fake data generated by all four methods (DF/F2F/FS/NT)

The result is so confused and I hope that you could give me some advices.
Thanks a lot.

@AssassinXY
Copy link

I have same porblem.After train DF/F2F/FS/NT as fakeimages, the train and val acc is always about 50%.but i have good result on DeepfakeTIMIT 、 UADFV and DeeperForensics-1.0 . I also want to know why the FF++dataset have the bad result .

@HongguLiu
Copy link
Owner

The result is about 75% with LQ images in the original paper.
image
image

  1. Increasing the number of images when you construct training datasets. (i.e. 270 images per video).
  2. Increasing the epochs of training processing.
  3. Try to balance pristine and fake images.

@yyk-wew
Copy link
Author

yyk-wew commented Mar 3, 2021

Thanks for the responses!

@AssassionXY
I tried training ResNet50 on real data and fake data generated by 3 method (DF/F2F/FS). Without NT, the real acc on validate dataset comes to 0.73. Maybe this helps.

@HongguLiu
For the third advice,

Try to balance pristine and fake images.

I've tried to get a balance by undersampling fake images. It does not work effectively.

For the second advice,

Increasing the epochs of training processing.

The model performance of detecting real faces continues to fall from the middle of the training. So training for more epochs may not work either.

For the first advice, I'll try and see.

Anyway, thanks for the advice, to both of you.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants