gpt-2 icon indicating copy to clipboard operation
gpt-2 copied to clipboard

BPE using sequence of bytes. HOW ?

Open bayethiernodiop opened this issue 5 years ago • 1 comments

Hello, I read the paper about GPT 2 : it says that they used BPE on a sequence of bytes and that they only needed a vocab size of 256. I researched the internet but didn't find any explanation on how BPE on sequence of bytes work and why the 256 vocab size. I am confusing since I don't know how this works compared to applying BPE on normal characters and what are the clear motivations since they also say that character/byte level LMs don't work great. How this is different. THANKS.

bayethiernodiop avatar Jan 25 '20 23:01 bayethiernodiop

whats a BPE????????????????????????????????????????????????????????????????????????????????????????

ghost avatar Oct 22 '20 13:10 ghost