paulpaulzhang
                                            paulpaulzhang
                                        
                                    @ver217 this is my code ``` python def trainer(train_dataloader, args, val_dataloader=None): start_epoch = 0 shard_strategy = TensorShardStrategy() with ZeroInitContext(target_device=torch.cuda.current_device(), shard_strategy=shard_strategy, shard_param=True): config = BertConfig.from_pretrained(args.model_name_or_path, num_labels=200) model = BertForSequenceClassification.from_pretrained(args.model_name_or_path, config=config) optimizer...
> This usually occurs because of CUDA out-of-memory. Yes, after open the zero, seems to have happened the memory overflow, memory growth until oom. and after turning on Zero, colossalai...
已解决,自己操作问题😂 Mcoy Jiang 于2018年7月2日周一 上午10:17写道: > 报错log贴出来 看一下 > > — > You are receiving this because you authored the thread. > Reply to this email directly, view it on...
I replaced control_hint by control_cond, new error occurred, maybe unet not have `control` param