You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Megatron-DeepSpeed supports MoE as seen from examples_deepspeed/MoE, there is some support for PP and TP as well introduced recently: microsoft#373 However I could not get this easily running, maybe I was missing some recent deepspeed updates that required this one.
Megatron-LM has some MoE support but the older version that can easily be ported to any accelerator is lacking drop token support. Maybe that can be ported easily still?
The text was updated successfully, but these errors were encountered:
Megatron-DeepSpeed supports MoE as seen from
examples_deepspeed/MoE
, there is some support for PP and TP as well introduced recently: microsoft#373 However I could not get this easily running, maybe I was missing some recent deepspeed updates that required this one.Megatron-LM has some MoE support but the older version that can easily be ported to any accelerator is lacking drop token support. Maybe that can be ported easily still?
The text was updated successfully, but these errors were encountered: