BertTokenizers
BertTokenizers copied to clipboard
The tokenization for Korean text seems not correct.
Is the tokenizer implementation the same as python version? It seems that it can not cut the Korean text into correct wordpiece.