tensor2tensor icon indicating copy to clipboard operation
tensor2tensor copied to clipboard

[Question] Inadequate translation

Open lkluo opened this issue 5 years ago • 3 comments

NMT is better than SMT in fluency, while it suffers from inadequate translation for long sentences. I have come across research where coverage is modelled for NMT. Does Transformer have better solutions for this (optimal setting, etc)?

lkluo avatar Oct 11 '18 09:10 lkluo

In this paper https://arxiv.org/pdf/1609.08144.pdf, based on beam search algorithm, they includes a coverage penalty to favor translations that fully cover the source sentence according to the attention module.The scoring function s(Y, X) that we employ to rank candidate translations is defined asfollows: (Equation 14 in paper page 12)

    s(Y, X) = log(P(Y|X))/lp(Y ) + cp(X; Y )

The first part of s(Y, X) is length normalization, found it here in transformer, but for the second part cp(X; Y ) which means coverage penalty, I didn't find which piece of code is implementing this function in transformer decoder. Does Transformer have better solutions for this?

crystal0913 avatar Oct 18 '18 03:10 crystal0913

@crystal0913: I think inadequate translation of NMT is a common problem in the community. Adding penalty can work but not a complete solution. I am thinking of hybrid translation with statistical MT.

lkluo avatar Nov 13 '18 04:11 lkluo

Is there any schedule on adding beta for coverage penalty?

ccmehk avatar Jul 30 '20 04:07 ccmehk