spacy-transformers icon indicating copy to clipboard operation
spacy-transformers copied to clipboard

🛸 Use pretrained transformers like BERT, XLNet and GPT-2 in spaCy

Results 6 spacy-transformers issues
Sort by recently updated
recently updated
newest added

Switch to offset mapping-based alignment for fast tokenizers. With this change, slow vs. fast tokenizers will not give identical results with `spacy-transformers`. Additional modifications: * Update package setup for cython...

feat / alignment

**Draft:** still needs docs, but I first wanted to discuss this proposal. The output of a transformer is passed through in two different ways: - Prediction: the data is passed...

bug
feat / pipeline

Preserve `tokenizer.is_fast` as `use_fast` in `HFShim` in order to be able to restore the same slow/fast tokenizer settings. This will be necessary when fast tokenizers switch to offset alignments from...

feat / serialize

The output of a transformer is passed through in two different ways: - Prediction: the data is passed through the `Doc._.trf_data` attribute. - Training: the data is broadcast directly to...

bug
feat / pipeline

## Description Extend support to transformers v4.39. ### Types of change ? ## Checklist - [x] I confirm that I have the right to submit this contribution under the project's...

## Description Relax the upper bound a little. ### Types of change Maintenance ## Checklist - [x] I confirm that I have the right to submit this contribution under the...