Skip to content
This repository has been archived by the owner on Oct 31, 2023. It is now read-only.

About the learning rate for resnet-50 #34

Open
cswaynecool opened this issue Oct 8, 2022 · 1 comment
Open

About the learning rate for resnet-50 #34

cswaynecool opened this issue Oct 8, 2022 · 1 comment

Comments

@cswaynecool
Copy link

I met an issue training resnet-50 with moco-v3. Under the distributed training setting with 16 V100 GPUs (each process only has one gpu, batch size 4096), I can get the training loss at about 27.2 in the 100-th epoch. When I lower the learning to 1.5e-4 (the default one is 0.6), the loss decreases more resonably and it reaches 27.0 in the 100-th epoch. Could you please verify if this is reasonable.

@cswaynecool
Copy link
Author

It seems that the training process is hardly convergent under the default learning rate.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant