-
Notifications
You must be signed in to change notification settings - Fork 54
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug] 好像没有把internevo的MoE权重转换成huggingface版本的脚本? #262
Comments
@Cerberous 我们大概下周会把转换脚本开源 |
大佬还有一个问题咨询一下,现在的MoE都是基于internlm1的版本,很多参数都不能设置,有基于internlm2的版本的MoE版本嘛 |
还有就是internevo也没有huggingface版本的MoE modeling.py,这个也会提供嘛? |
是的,internlm2-moe之后也会开源。huggingface版本的MoE modeling.py会跟着转换脚本一起提供的。 |
好的,非常期待!!! |
转换脚本目前在 #271 实现了,可以用 |
可以啦,非常感谢 |
还有一个问题是我发现提供的hf的modeling_internlm.py里面对shared_experts是直接加到y上的
但是在internevo中实际实现的时候有一个coefficient
这里多了一个与结果的加权,如果我没理解错的话这里是没实现的? |
Describe the bug
我好像没有找到用internevo训练然后转换成对应的hf的脚本?请问有提供嘛?
Environment
官方代码
Other information
No response
The text was updated successfully, but these errors were encountered: