multi-task-learning-example
multi-task-learning-example copied to clipboard
MergeLoss with regular item is \log_{sigma} in paper but \log_{sigma}^2 in code
I see that as well, should it not be ( precision**2 )/ 2 instead of just precision?
So, what is the correct? Did you try them?
if ‘precision = K.exp(-log_var[0])’,then the network learning $\log{\sigma}^2$,'precision * (y_true - y_pred)**2. + log_var[0]' is $\frac{1}{\sigma ^ 2} L(w) + 2 * \log{\sigma}$;
if ‘precision = K.exp(-log_var[0]) ** 2 / 2’,then the network learning $\log{\sigma}$,'precision * (y_true - y_pred)**2. + log_var[0]' is $\frac{1}{2 * \sigma ^ 2} L(w) + \log{\sigma}$;
The difference between the two is the coefficient 2,for network training ,they are the same.