DongYongFei
DongYongFei
 这是我生成的yolov5s.xml  这是作者的yolov5s.xml 显然,拓扑结构是不一样的 我运行我产生的yolov5s.xml结果  很奇怪,我也是v3.1
Can you provide a simple demo?thank you
okay, thank you for your outstanding work
> > python3 predict_system.py > > --kie_algorithm=LayoutXLM > > --re_model_dir=./inference/re_vi_layoutxlm_xfund_infer > > --ser_model_dir=./inference/ser_vi_layoutxlm_xfund_infer > > --image_dir=./docs/kie/input/zh_val_42.jpg > > --ser_dict_path=../ppocr/utils/dict/kie_dict/xfund_class_list.txt > > --vis_font_path=../doc/fonts/simfang.ttf > > --ocr_order_method="tb-yx" > > --mode=kie > >...
> @DYF-AI Yes, during the fine-tuning phase, the vision encoder is trained as well. Thank you for your answer, and one more question, is there any code that provides training...
If relative_attention is closed, are the highlights mentioned in the paper meaningless?