Roberto
Roberto
I was trying to test my pruned model with the [generate.py](https://github.com/horseee/LLM-Pruner/blob/main/generate.py) script but I get this error: **AttributeError: 'GenerationConfig' object has no attribute 'prefill_chunk_size'**. I suppose this problem is related...
Pruning LLMs using [prune_llm.py](https://github.com/VainF/Torch-Pruning/blob/master/examples/LLMs/prune_llm.py) will impact performance severely because it prunes all the layers (I think...). Instead, in LLMPruner there's the option to "flag" the layers you're going to prune...
ignore env folders, pycharm, vscode