Nan Jiang
Nan Jiang
1. This is likely due to a problem in the vocabulary part. Are you using the pre-trained GPT model I shared when you train your own GPT-CoNuT model? 2. It...
The point of the first possible cause is that, when you train your own GPT-CoNuT model, did you only change the `train_file` and `valid_file` in `src/trainer/gpt_conut_trainer.py` and keep the `vocab_file`...
@msintaha Looks like you only run the `prepare_cure_input` function. there are two remaining steps: 1. run subword-nmt to tokenize these lines into subwords. 2. run `clean_testing_bpe` to finalize the input...
> First i generated the vocab using `subword-nmt learn-joint-bpe-and-vocab --input training_tokenize.txt -s 50000 -o subword.txt --write-vocabulary vocabulary.txt` > > Then i ran: > > ``` > subword-nmt apply-bpe -c subword.txt...
That looks reasonable. Could you enclose the call of `generate_gpt_conut` with try-catch and see if it crashes for every input or just some? Another possibility I can imagine is the...
I think this issue is caused by that lines 38-41 are not correctly executed. You may want to check if you load the model correctly, is the model an instance...
These two files are prepared from JDK. For keywords, I think you can search online for JavaScript keywords. For Java class, we crawled the JDK8 to extract this. You can...
@faysalhossain2007 please check `data/data/prepare_testing_data.py`, which is the script to prepare test input for new data. If your ground-truth is inserting a new statement, you could use the line after the...
@msintaha @nashid I used things like: `n_positions=1024, n_ctx=1024, n_embd=384, n_layer=8, n_head=6` You can also try other reasonable settings, as this is just empirically set. Actually, the checkpoint contains the config...