Issue loading finetuned mamba model into MambaLMHeadModel
I am getting the following error trying to load a Mamba model:
TypeError: MambaConfig.__init__() got an unexpected keyword argument '_name_or_path'
This is due to the config.json having this as its first line:
"_name_or_path": "state-spaces/mamba-130m-hf"
I get the error running:
model = MambaLMHeadModel.from_pretrained(model_name, device='cuda')
When loading a model trained with Lora, I get:
TypeError: expected str, bytes or os.PathLike object, not NoneType
I am guessing this is because there is no config.json file. Am I missing something?
I guess you are mixing state-spaces/mamba-130m with state-spaces/mamba-130m-hf. Use state-spaces/mamba-130m for MambaLMHeadModel
I am encountering the same issue, maybe choice the wrong model 0.0
I guess you are mixing state-spaces/mamba-130m with state-spaces/mamba-130m-hf. Use state-spaces/mamba-130m for MambaLMHeadModel
I choice the mamba-130m, but also have this problem