LeoWood

Results 9 issues of LeoWood

请问有没有试过在一台机器上开启多个服务(设置不同的端口),有一台机器最多开启5个服务,后面再开就一直停留在load pb file最后一步,一直没有出现ready and listening。查看GPU显存还有机器内存都还有很多空闲,不知道为什么。

您好,这边用了最新的代码之后,使用多卡进行预训练就会报错,主要是出现在mp.spawn那一步,错误信息如下: Traceback (most recent call last): File "pretrain.py", line 133, in main() File "pretrain.py", line 129, in main trainer.train_and_validate(args) File "/data/leo/Projects/uer-py-1/uer/trainer.py", line 56, in train_and_validate mp.spawn(worker, nprocs=args.ranks_num, args=(args.gpu_ranks, args, model),...

最近看到仓库在陆续丰富下游任务的相关代码,但一直没有看到Relation Extraction这个任务。关系抽取也是NLP里面比较重要的一个任务之一,想请问下项目组,是否有计划增加关系抽取任务?

enhancement

在hugging face下载的模型:https://cdn.huggingface.co/bert-base-chinese-pytorch_model.bin ,使用scripts中convert_bert_from_huggingface_to_uer.py,报错: Traceback (most recent call last): File "convert_bert_from_huggingface_to_uer.py", line 22, in output_model["embedding.layer_norm.gamma"] = input_model["bert.embeddings.LayerNorm.weight"] KeyError: 'bert.embeddings.LayerNorm.weight' 脚本: python convert_bert_from_huggingface_to_uer.py \ --input_model_path ../models/bert-base-chinese-pytorch_model.bin \ --output_model_path ../models/google_zh_model.bin

请问多个节点多GPU训练的具体操作是怎样的? ReadME中的例子: ``` Node-0 : python3 pretrain.py --dataset_path dataset.pt --vocab_path models/google_zh_vocab.txt \ --pretrained_model_path models/google_model.bin --output_model_path models/output_model.bin \ --encoder bert --target bert --world_size 16 --gpu_ranks 0 1 2 3 4 5...

请问eHealth-base模型文件里面是否少了saved_weights.pdparams? 还是说这个文件是跟Ernie-base一致的?

Hi, i have a question that the number of PMC articles is huge and the pre-process procedure requires sentences segmentation for paragraphs, so how do you finish your sentence segmentation...

when i input Chinese, like: **lm.do("你好")** it returns ' ⁇ , ⁇ , ⁇ ' It seems that the current model does not support Chinese Is there any recommended Chinese...

question

看到huggingface上的tf模型是.h5格式的,请问是否有.ckpt格式的模型? 或者如何自己生成.ckpt格式的模型呢?