kgTransformer icon indicating copy to clipboard operation
kgTransformer copied to clipboard

Query structure for specific tasks

Open Jasonqte opened this issue 1 year ago • 3 comments

Nice work! One question, I want to run this model on my own dataset, but the query logic of my downstream tasks mainly focus on is 1p, 2p, 3p, 4p, 5p. Can I only use the above five structures during training process? Does this have an impact on model performance?

Jasonqte avatar Aug 02 '23 03:08 Jasonqte

Thank you! Yes, the model supports tuning toward specific query structures. This can be set in the configuration file. Here is an example, which is also how we boost the model performance.

sshockwave avatar Aug 02 '23 09:08 sshockwave

Thank you for your reply! Is that mean I should change the pretrain task from '1p 2p 3p 2i 3i' to '1p, 2p, 3p, 4p, 5p' or keep the original pretrain task and set a new finetune task like reasoning_5p which has reasoning_train_modes of '1p, 2p, 3p, 4p, 5p' and reasoning_test_modes of '5p'. Which one seems better?

Jasonqte avatar Aug 03 '23 02:08 Jasonqte

Adding 5p sampling to the pre-train task might or might not improve the performance, which really depends on your dataset. Either way, I believe that a separate finetune task is best for datasets that are unseen during pre-training.

sshockwave avatar Aug 07 '23 02:08 sshockwave