FATE-LLM icon indicating copy to clipboard operation
FATE-LLM copied to clipboard

GPT2 Example job issues when supplying target_modules param in LoraConfig

Open cooleo-dude opened this issue 1 year ago • 3 comments

When following the GPT2 example, the following errors occur: image After debugging, this error is caused by the json/encoder.py file when trying to reference the target_modules = ['c_attn'] component of the LoraConfig object. Changing this value to a string 'c_attn' resolves the issue, however limits the ability to fine-tune multiple categories. However, after doing so the following error occurs. image This occurs because the t.nn.CustModel object created in the pipeline job does not reference any of the layers from the GPT2 model, and there does not appear a way to do so. Is there a workaround for this, or would this be an environment issue?

cooleo-dude avatar Nov 16 '23 15:11 cooleo-dude