Megatron-LM
Megatron-LM copied to clipboard
[QUESTION] Does Megatron-Core supports LLAMA models?
Does Megatron-Core supports LLAMA models?
yes
@ethanhe42 When transformer-impl
is local
, it reports the following error:
AssertionError: (RMSNorm) is not supported in FusedLayerNorm when instantiating FusedLayerNorm when instantiating TransformerLayer
When transformer-impl
is transformer_engine
, the following code does not seem to define RMSNorm?
So do I need to make any changes when I want to use llama?
You need to use mcore models. local is deprecating
@ethanhe42 When transformer-impl
is set to transformer_engine
, the following code does not seem to define RMSNorm?
It's handled by TEnorm
Marking as stale. No activity in 60 days.