-
Notifications
You must be signed in to change notification settings - Fork 8
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fine Tuning The Model #1
Comments
I'm sorry I haven't seen the issue for some reason, did you solve the problem now?
|
I also encountered the situation where the loss is NaN. It happened because I only provided supervised data to the model without any unsupervised data, so when calculating, division by zero occurred, resulting in an error. In your case of fine-tuning based on the POT dataset, there should be no unsupervised data available, which might be the reason for this issue. Moreover, during training, these NaN losses do not have any practical impact. |
Do you have any experience with training from scratch or training based on the improved model provided by the author? I have tried training from scratch and training based on pre-trained models using COCO14 and GOT-10K data, but I couldn't achieve the same accuracy as the improved model mentioned by the author. Do you have any tips or experiences to share? |
I am trying to fine-tune the HDN model on the POT dataset. I have preprocessed the POT dataset as instructed. However, the training loss seems jumpy as it fluctuates from 1 to 600. I tried to debug the training flow and found that it happens when we feed the supervised training samples to the model. Also, the loss for the HMNET model comes out to be nan. Even after training the model for 1 epoch, the model performance decreases too much. Is there a bug in the training pipeline? Please assist.
The text was updated successfully, but these errors were encountered: