We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
请问训练的显存需要多少呢? 我在单张H100 80G上也会爆显存,使用了CogVideoX-5b-I2V为初始权重,加入is_train_lora参数也会爆显存
The text was updated successfully, but these errors were encountered:
1、DeepSpeed Zero2模式,全参微调,单卡80G可能无法运行(可以尝试打开low_vram,vae.enable_tiling参数) 2、DeepSpeed Zero2模式,全参微调,双卡80G可以正常运行,甚至不需要打开low_vram,vae.enable_tiling参数(因为Zero2会将部分显存均摊到每张卡上) 3、换成lora微调,印象中只需要50G显存
Sorry, something went wrong.
好的,谢谢您。那请问batchsize设置成1大概要train多少个step呢
感谢关注,这个得具体看实验结果来判断了,我还没尝试过bs1训练。
No branches or pull requests
请问训练的显存需要多少呢?
我在单张H100 80G上也会爆显存,使用了CogVideoX-5b-I2V为初始权重,加入is_train_lora参数也会爆显存
The text was updated successfully, but these errors were encountered: