cappy
cappy copied to clipboard
Is it multilnigual or "en" only?
Interesting paper. Regarding pretrained model I'm wondering - are they Roberta based or R-XLM? Did you evaluate performance wrt mDeberta as base model?
And finally - how would one use such/this model to score contexts beyond 512 tokens?