FunPony
FunPony
我分别跑了ernie1.0,ernie3.0和ernie-gram,结果分别如下: ernie1.0: Sentence-Level Performance: Detection metric: F1=0.7262449528936744, Recall=0.6642048252092565, Precision=0.8010688836104513 Correction metric: F1=0.687742651136994, Recall=0.6105366814377154, Precision=0.7873015873015873 ernie3.0: Sentence-Level Performance: Detection metric: F1=0.6717277486910995, Recall=0.6317085179714427, Precision=0.7171604248183343 Correction metric: F1=0.6490947816826411, Recall=0.6001969473165928, Precision=0.7066666666666667 ernie-gram还没跑完,但是我看到原文好像是sighan13,14,15分开测试,而这个代码好像是把三个整理成一个test数据集了?所以不能简单的对比(但基本上跟原文mlm_phenetic有差距),毕竟我们用的不是他的pre_train。 
> 我分别跑了ernie1.0,ernie3.0和ernie-gram,结果分别如下: ernie1.0: Sentence-Level Performance: Detection metric: F1=0.7262449528936744, Recall=0.6642048252092565, Precision=0.8010688836104513 Correction metric: F1=0.687742651136994, Recall=0.6105366814377154, Precision=0.7873015873015873 > > ernie3.0: Sentence-Level Performance: Detection metric: F1=0.6717277486910995, Recall=0.6317085179714427, Precision=0.7171604248183343 Correction metric: F1=0.6490947816826411, Recall=0.6001969473165928, Precision=0.7066666666666667...