Alan Akbik
Alan Akbik
@Dashulkin the deadline has already passed, sorry! But more positions will be posted next year!
@dchaplinsky I tried today with the most recent lightning release and DDP strategy, using the code from the example that comes with the PR: ```python trainer = LanguageModelTrainer(accelerator="gpu", devices="auto", strategy='ddp')...
Hello @aniketmaurya sure, I'll build a minimal example for you.
Here is a minimal training data example for my training script. Unpack this where you like and point the script to the root folder: [penn_lm.zip](https://github.com/flairNLP/flair/files/9899426/penn_lm.zip). Note that we usually split...
Quick update: it trains successfully if the `train` folder only contains a single split, so maybe the data loader messes up the execution.
@dchaplinsky form what I can tell, it does a mini-batch on each GPU and then gathers the gradients for a single update. So in the same amount of time, you...
Hello @quantarb we've had such issues before. In this case, I first use a regular tokenizer, and then additionally split all tokens on the offset positions to get the final...
@stefan-it that would be great!
Could you try again? We just merged @stefan-it 's new version of the model.
@himkt thanks for improving this! And thanks @helpmefindaname for testing!