keras-nlp icon indicating copy to clipboard operation
keras-nlp copied to clipboard

Modular Natural Language Processing workflows with Keras

Results 360 keras-nlp issues
Sort by recently updated
recently updated
newest added

[MLMMaskGenerator](https://github.com/keras-team/keras-nlp/blob/master/keras_nlp/layers/preprocessing/mlm_mask_generator.py) is a helpful tool to generate masks for input sequences. The current docstring only has two basic examples, and it would be better for users if we can provide...

good first issue
contributions welcome

We need to generate symbol documentation and host them keras.io (and tensorflow.org?).

documentation

Branching off from the issue which @aflah02 opened a few weeks ago, https://github.com/keras-team/keras-nlp/issues/39: Is the KerasNLP team interested in implementing adversarial attacks? We could start off with simple attacks on...

Are there any plans to add a BIO Tagging Layer to KerasNLP? This layer could take sentences and relevant spans as input and output their BIO representations. BIO Tags are...

type:feature

We should add support for ragged inputs to the `SinePositionEncoding` layer. Our `PositionEmbedding` layer shows a rough example of how this can be done on [this line](https://github.com/keras-team/keras-nlp/blob/master/keras_nlp/layers/position_embedding.py#L93)

enhancement
good first issue
contributions welcome

We would like to use type annotations in KerasNLP. We should add them for the encoder/decoder blocks... https://github.com/keras-team/keras-nlp/blob/master/keras_nlp/layers/transformer_decoder.py https://github.com/keras-team/keras-nlp/blob/master/keras_nlp/layers/transformer_encoder.py

The tokenization structure is different from languages like English or other languages that start from left to right, and those language like Arabic and langues that start from right to...

type:feature

Not sure if this is worth implementing, but basically, we can have functions/classes, which users can call in order to get decoded text. As far as I know, we have...

Instead of implementing language specific tokenizers, we can try to look at more robust way of dealing with such problems. We can maybe look at a more robust way, like...

type:feature

The gMLP model is from the paper "[Pay Attention to MLPs](https://arxiv.org/abs/2105.08050)". It has a decent number of citations - around 40. Every Encoder Block merely consists of linear layers, a...

type:feature