Convolutional-Complex-Knowledge-Graph-Embeddings icon indicating copy to clipboard operation
Convolutional-Complex-Knowledge-Graph-Embeddings copied to clipboard

Is the formula right ?

Open SsnAgo opened this issue 4 years ago • 9 comments

Hi, I read the paper and I find that (3) and (4) are not equivalent to (1) and (5)? Also I think (3) even not equivalent to (4) ? image image image And I notice that it is calculated using formula (4) in the code. If it is my misunderstanding, please kindly advise me. thanks~

SsnAgo avatar Sep 26 '21 07:09 SsnAgo

Dear @SsnAgo,

Thank you for your question. A score of a triple is computed via Eq. 4 as you mentioned. Through Eq. 1 and Eq 5, we previously aimed to higlight the connection of conv. with Hermitian inner product. However, there is an indeed typo. Hence, it is not your misunderstanding. You may have a look at our updated manuscript where we fixed this issue and please let me know if something is still unclear.

Cheers, CD

Demirrr avatar Sep 27 '21 09:09 Demirrr

Thanks for your reply~ @Demirrr Ok, thanks ~ finally,may I ask, when I used the parameters given in the paper to run the code, I found that the experiment result did not reach the performance in the paper, is there a more suitable parameter?

SsnAgo avatar Sep 27 '21 09:09 SsnAgo

Firstly, you may want to use pretrained models provided.

If you would like to reproduce experiments, I would suggest you the followings

  1. Fix the seed of random number generator to 1 and
  2. Carry out grid search whose ranges are defined in Section 5.3.

If you have already followed these steps, could you elaborate on results and generated log files?

Demirrr avatar Sep 27 '21 09:09 Demirrr

Thanks~ @Demirrr I noticed that the ablation experiment in the paper, the Conv(') operation was eliminated. Does this mean the model complete degeneration into ComplEx?

SsnAgo avatar Nov 04 '21 07:11 SsnAgo

Yes, it does indeed.

At evaluation/inference time, we removed the Conv(.) function to obtain scores of input triples as shown in here. Thereofre, scores are generated solely via the Hermitian inner product.

Demirrr avatar Nov 04 '21 10:11 Demirrr

Thanks a lot ! @Demirrr

SsnAgo avatar Nov 04 '21 10:11 SsnAgo

Lemme know if you face any difficulties

Demirrr avatar Nov 04 '21 10:11 Demirrr

Hi, I add ConEx in my own model to predict the score as decoder. While Encoder is based on GCN and attention network. But ConEx get lower MRR than InteractE/ConvE in this situation. So I remove my Encoder and I find ConEx is still far from the performance in the paper. It get 0.33 Valid MRR about 200 epoch. I use the the parameters given in the paper: embedding_dim = 100 ; kernel_size = 3; num_of_output_channels = 16; input_dropout = 0.3; feature_map_dropout = 0.4; batch_size = 512; learning_rate = 0.01; label_smoothing = 0.1 and seed = 1. Is there something else I haven't noticed? The code all I used is the 'class ConEx' in complex_models.py. I used BCE loss and KvsAll scoring technique.

Jim-Hou avatar Nov 19 '21 12:11 Jim-Hou

Hello @Jim-Hou

You may want to take a look at #7. Therein, I elucidated possible solutions.

Demirrr avatar Nov 20 '21 12:11 Demirrr