You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Issue:
想请教一下是否在13B以上的MoE模型上实验过? 我使用了ZeRO-2,EP_SIZE=8;
在初始化optimizer状态时会报cuda: out of memory.
而ZeRO-3则不支持MoE, 由于设备限制,我也无法采用offload加载;
是不是还是得考虑megatron-deepspeed?
Describe the issue
Issue:
想请教一下是否在13B以上的MoE模型上实验过? 我使用了ZeRO-2,EP_SIZE=8;
在初始化optimizer状态时会报cuda: out of memory.
而ZeRO-3则不支持MoE, 由于设备限制,我也无法采用offload加载;
是不是还是得考虑megatron-deepspeed?
Environment:
Command:
Log:
Screenshots:
You may attach screenshots if it better explains the issue.
The text was updated successfully, but these errors were encountered: