Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Questions about experiment results. #10

Closed
ghost opened this issue Jun 16, 2022 · 1 comment
Closed

Questions about experiment results. #10

ghost opened this issue Jun 16, 2022 · 1 comment

Comments

@ghost
Copy link

ghost commented Jun 16, 2022

I trained about 50000 iters using 2080Ti with batch size as 2, and I found that the evaluation results are quite unstable. The AP50 fluctuated around 41 and reach a maximum of 43.5. I wanna ask you how to judge the convergence of the model and select the results to report.
Thanks a lot.

@wymanCV
Copy link
Contributor

wymanCV commented Jun 16, 2022

Hi, thanks for your attention.

First of all, the instability is common due to the adversarial training, which is the same as almost all existing DAOD frameworks, e.g., EPM (our baseline), DAF, and the latest work VISGA. As far as I know, currently, no framework can guarantee stable training and completed convergence with GRL-based adversarial learning (without using Cyclegan to generate interpolated data). So, you may notice that almost all current works rely on the model selection for the benchmark comparison. If you can address the GRL-based instability, it will be a big step for this field.

Actually, longer training, two-stage training, and reducing the learning rate will make the training more stable but won't break the maximum results a lot. So, the common and easy practice is to train longer, e.g., AT uses 360000 iterations w. bs=16. Moreover, you may find that our uploaded model should have higher AP (not AP50) than your reproduced one because we report our results (bs=4) with longer training and two-stage training when the model has a tolerable fluctuation around 43. You can load it as the initialization and train further to have a look.

Besides, some interesting discussion (in Chinese) about the convergence of GRL-based DA may help you understand this common phenomenon better.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

1 participant