DeBERTa
DeBERTa copied to clipboard
The implementation of DeBERTa
DeBERTa is a very interesting language model, however XLNet is a vastly superior baseline over BERT, XLNet achieves much greater acuracy on key NLP tasks. Unfortunately, for contingent reasons, the...
Hi, it would be very interesting to also see a comparison of pre-training times for DeBERTa v2 versus the recently released v3, that is using RTD. The v2 paper mentioned...
Am trying to create a tokenizer from Model **'microsoft/deberta-v2-xlarge'** Initially I got no-offset-mapping error while setting return_offset_mapping = True. Later, I created tokenizer using pretokenizerfast : ```model = 'microsoft/deberta-v2-xlarge' tokenizer...
Are there weights for the masked language model? Could you provide example of how to use it, because on homepage it is only shown how to tokenize it and i...
Hi, I was wondering if it would be possible to release the source distribution for pip through PyPI or GitHub releases? Reason is that I would like to add this...
Hey For some reason I want to try doing some futher pretraining on another dataset. Where can I access the RTD Model's pretrained version (Sine it's not included in the...
Import ABC from `collections` was deprecated and removed in Python 3.10. Use `collections.abc` .
Impressed by the performance debertaV3, and I'm trying to re-produce debertaV3. I have a question regarding the model architecture in DebertaV3. Is EMD used in generator when doing MLM task?...
Hello, @BigBird01 is deberta v3 built with machine translation capabilities
use this command can get checkpoint?and in where? if i want to save checkpoint,can i add which argument? python3 -m DeBERTa.apps.run --task_name $task --do_train \ --data_dir $cache_dir/glue_tasks/$task \ --eval_batch_size 128...