keras-nlp
keras-nlp copied to clipboard
The BytePairTokenizer class is extremely, extremely slow at tokenizing
vocabulary size 6400
text = "Are you OK? "
start = time.time()
for i in range(10):
tokenizer.tokenize(text + str(i))
end = time.time()
print(end - start)
3.8366940021514893 seconds