attentional_factorization_machine icon indicating copy to clipboard operation
attentional_factorization_machine copied to clipboard

AFM训练中出现nan

Open Atomu2014 opened this issue 6 years ago • 3 comments

AFM训练起来很容易nan,请问您遇到过这种情况吗?对于调参数有什么建议?哪些参数比较敏感?

Atomu2014 avatar Feb 10 '18 15:02 Atomu2014

nan应该是因为我们自己用exp函数实现了softmax, 导致容易overflow。换用tensorflow实现的softmax应该就不会nan了。

On Sat, Feb 10, 2018 at 11:48 PM, KevinKune [email protected] wrote:

AFM训练起来很容易nan,请问您遇到过这种情况吗?对于调参数有什么建议?哪些参数比较敏感?

— You are receiving this because you are subscribed to this thread. Reply to this email directly, view it on GitHub https://github.com/hexiangnan/attentional_factorization_machine/issues/9, or mute the thread https://github.com/notifications/unsubscribe-auth/ABGxjuXIl51KSyJ121ucnDpYetMQ64Wiks5tTboxgaJpZM4SA-Kv .

-- Best Regards, Xiangnan He

hexiangnan avatar Feb 10 '18 23:02 hexiangnan

感谢回答,最近发现调节softmax的温度会比较容易训练,在我的数据集上调节温度比attention_lambda更有效,能显著控制过拟合,建议代码中增加调节温度的参数

Atomu2014 avatar Feb 15 '18 06:02 Atomu2014

谢谢告知!

On Thu, Feb 15, 2018 at 2:50 PM, KevinKune [email protected] wrote:

感谢回答,最近发现调节softmax的温度会比较容易训练,在我的数据集上调节温度比attention_lambda更有效,能显著控制过拟合, 建议代码中增加调节温度的参数

— You are receiving this because you commented. Reply to this email directly, view it on GitHub https://github.com/hexiangnan/attentional_factorization_machine/issues/9#issuecomment-365840062, or mute the thread https://github.com/notifications/unsubscribe-auth/ABGxjgrFgN4ZzcHQz8khApezHDNIjFkmks5tU9PIgaJpZM4SA-Kv .

-- Best Regards, Xiangnan He

hexiangnan avatar Feb 15 '18 11:02 hexiangnan