keras-nlp icon indicating copy to clipboard operation
keras-nlp copied to clipboard

Modular Natural Language Processing workflows with Keras

Results 360 keras-nlp issues
Sort by recently updated
recently updated
newest added

This could be similar to very similar in structure to the [lstm seq2seq guide](https://keras.io/examples/nlp/lstm_seq2seq/) on keras.io, but show using either the ByteTokenizer or UnicodeCharacterTokenizer (or both). We should demo training...

Currently TransformerDecoder must have both `encoder_inputs` and `decoder_inputs`, while for models like GPT2, only `decoder_inputs` is required. We should make a change to mark `encoder_inputs` as optional.

We should write a example very similar to the [Spanish English translation example](https://keras.io/examples/nlp/neural_machine_translation_with_transformer/) already on keras.io. We can use the same dataset, and same basic model structure, but we should...

We should update https://github.com/keras-team/keras-nlp/tree/master/examples/bert to use `keras_nlp.layers.TransformerEncoder`.

type:feature

KerasNLP is always looking for new examples on [keras.io](https://keras.io/keras_nlp) the demonstrate how to use the library. This issue will stay open on a "contributions welcome" list forever. If you have...

type:docs
good first issue
stat:contributions welcome
team-created

Currently we are using the BackupAndRestore callback to resume training on our examples after a failure. We also need to make sure that we reset the dataset iterator to the...

type:feature

Fixes #166 Hey @chenmoneygithub Following our discussion I think the PR is ready for review!

We should add a `vocabulary_size` argument to the WordPieceTokenizer layer that forces the vocabulary size by truncating the passed in vocabulary if necessary. Potential docstring: ``` vocabulary_size: Force the vocabulary...

type:feature
good first issue
stat:contributions welcome

Currently the BERT example writes [custom code](https://github.com/keras-team/keras-nlp/blob/master/examples/bert/create_pretraining_data.py#L392) to generate MLM mask, which is slow. We should replace it with the [MLMMaskGenerator](https://github.com/keras-team/keras-nlp/blob/master/keras_nlp/layers/mlm_mask_generator.py).

good first issue
contributions welcome

This is an issue for tracking the progress of training BERT example. The model has different sizes: tiny, small, base and large. Only tiny and small fit in a common...

type:feature