-
Notifications
You must be signed in to change notification settings - Fork 21
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Train on your own 512 x 512 size image #32
Comments
I think the CityScale setup defaults to 512x512 patch size. Can you try that? Batch size depends on your GPU memory, I think you can start with the largest batch size you can get away with - then tune LR properly to make sure it converges. It may need some trail-and-error. |
I think our released model takes 1.0m/pixel images. Can you try resizing your images to that resolution? Also, have you fine-tuned on your own dataset? How large was your dataset? |
Also did you correctly load the pre-trained SAM ckpts? |
Thank you again for your suggestion, I used the 0.5 meter resolution image to crop it to 512*512 image size, and also loaded the ckpt of the pre-trained SAM, and also adjusted the learning rate, and re-trained, but the effect is not justified by the good effect of the two datasets in the original paper, the dataset has a total of 3065 pieces, 2453 for training, 459 for testing, 153 for verification, according to the scale of Spacenet for the data division Each image is 512 and contains the corresponding required graph data. So if I change the image to 1 meter resolution, will the end result be improved? The results of the test are like this, and it is not clear whether the thresholds of key points and roads should also be modified? |
Hi, I think if you are fine-tuning from the original SAM ckpt (not the ones I released), resolution is less crucial. How does the images look like in general? The numbers you shown seems to suggest the model did not converge at all. The size of the dataset sounds reasonable, can you try the following options:
Good luck with your experiments! |
Sorry for the late reply, thank you for your suggestion, I will follow your suggestion to carry out the experiment, I am a graduate student in a university, and the current direction of study is to use high-resolution remote sensing images for road extraction, thank you for communicating with you, can you add WeChat, my WeChat account is 18837621961, I will be honored. |
I trained SAM on the DeepGlobe dataset and the results were convincing, so I believe SAM is robust. Please carefully check your code. |
Thank you for your work sharing, but I also used DeepGlobe for training and testing, I cropped it to a 512*512 image, and trained and tested, but the result is not very ideal, but the clear road that can be extracted is incomplete The effect is not very good, can I consult your config settings and the division rules of the dataset? Or is there some other modification work and configuration work that I haven't noticed? Thanks for your answer.
|
For 512*512 images, do I need to modify the config settings for training on my own dataset? (This is the SpaceNet configuration). Also, do you have any advice on the training batch, the results of 30 sessions are not ideal? Thank you very much for your work on the SAM-Road project, and look forward to your answers in your busy schedule, thank you.
The text was updated successfully, but these errors were encountered: