SwissArmyTransformer icon indicating copy to clipboard operation
SwissArmyTransformer copied to clipboard

SwissArmyTransformer is a flexible and powerful library to develop your own Transformer variants.

Results 42 SwissArmyTransformer issues
Sort by recently updated
recently updated
newest added

尝试安装时候出现以下错误 ``` PS D:\deeplearning\VisualGLM-6B> pip install SwissArmyTransformer>=0.4.4 Collecting SwissArmyTransformer>=0.4.4 Obtaining dependency information for SwissArmyTransformer>=0.4.4 from https://files.pythonhosted.org/packages/91/b9/44a3e9cc0116a3ebf8b60f0ef67340ca22c195e37b9d8f951a37e114f300/SwissArmyTransformer-0.4.4-py3-none-any.whl.metadata Using cached SwissArmyTransformer-0.4.4-py3-none-any.whl.metadata (10 kB) Requirement already satisfied: torch in c:\python311\lib\site-packages (from SwissArmyTransformer>=0.4.4) (2.0.1+cu118)...

logits, *mems = model(inputs_ids, position_ids, attention_mask) # print(logits.shape) loss_func = CrossEntropyLoss(ignore_index=-100) loss = loss_func(logits.view(-1, logits.size(-1)).float(), labels.view(-1))`` 我是这样写的loss计算方式,会出现一个/opt/conda/conda-bld/pytorch_1670525539683/work/aten/src/ATen/native/cuda/Loss.cu:242: nll_loss_forward_reduce_cuda_kernel_2d: block: [0,0,0], thread: [15,0,0] Assertion `t >= 0 && t < n_classes`...

我设置会出现mp_size=2以后进行lora训练,会出现维度不匹配的问题

加载方式 ``` model = AutoModel.from_pretrained(visualchatglm_model_path,trust_remote_code=True).to(torch.cuda.current_device()) ``` 运行环境如下: ``` SwissArmyTransformer 0.3.7 transformers 4.28.1 deepspeed 0.9.1 torch 1.11.0+cu113 torchaudio 0.11.0+rocm4.5.2 torchvision 0.12.0+cu113 cpm-kernels 1.0.11 einops 0.6.1 ``` 报错日志如下: ``` Epoch_0: 0%| |...

I did not find such a cached method using past_key_values in the SAT. Is it possible to add this? Thanks.

ModuleNotFoundError: No module named 'localAttention' 请问这是为什么呢 谢谢

直接跑源码的qlora.py,报错 ![image](https://github.com/THUDM/SwissArmyTransformer/assets/64970397/1a231ac7-d43e-414c-a1d0-73a5c1c6180d) 给model.child = LoraLinear(100, 200, 10)改为model.child = LoraLinear(100, 200, 10,10,2)后,又报错 ![image](https://github.com/THUDM/SwissArmyTransformer/assets/64970397/a2d16427-e2a4-4178-807a-55af76c48589)

## What I did I pass the "-h" option to print the help message. However, my script complains `TypeError: %o format: an integer is required, not dict` in `/usr/lib/python3.8/argparse.py:Line633`. The...

[2023-07-06 16:43:31,720] [INFO] [RANK 0] Try to load tokenizer from Huggingface transformers... Explicitly passing a `revision` is encouraged when loading a model with custom code to ensure no malicious code...

Collecting SwissArmyTransformer Using cached SwissArmyTransformer-0.3.7-py3-none-any.whl (2.4 MB) Requirement already satisfied: torch in /usr/local/lib/python3.8/site-packages (from SwissArmyTransformer) (1.13.1) Requirement already satisfied: transformers in /usr/local/lib/python3.8/site-packages (from SwissArmyTransformer) (4.27.1) Requirement already satisfied: sentencepiece in...