spacy-transformers
spacy-transformers copied to clipboard
🛸 Use pretrained transformers like BERT, XLNet and GPT-2 in spaCy
Switch to offset mapping-based alignment for fast tokenizers. With this change, slow vs. fast tokenizers will not give identical results with `spacy-transformers`. Additional modifications: * Update package setup for cython...
**Draft:** still needs docs, but I first wanted to discuss this proposal. The output of a transformer is passed through in two different ways: - Prediction: the data is passed...
Preserve `tokenizer.is_fast` as `use_fast` in `HFShim` in order to be able to restore the same slow/fast tokenizer settings. This will be necessary when fast tokenizers switch to offset alignments from...
The output of a transformer is passed through in two different ways: - Prediction: the data is passed through the `Doc._.trf_data` attribute. - Training: the data is broadcast directly to...
## Description Extend support to transformers v4.39. ### Types of change ? ## Checklist - [x] I confirm that I have the right to submit this contribution under the project's...
## Description Relax the upper bound a little. ### Types of change Maintenance ## Checklist - [x] I confirm that I have the right to submit this contribution under the...