Kaggle_Lux_AI_2021
Kaggle_Lux_AI_2021 copied to clipboard
In calculate baseline_loss why use Td-lambada rather than vtrace advance
what the different between use TD-lamda and baseline_value to calculate baseline_loss than use original vtrace return and baseline_value ???
In alphastar paper they said this make a lot of differnt I run your code agian and again ,I cannot understand why and what make this differents