tanlong Du
tanlong Du
我也遇到相同的错误,尝试使用+actor_rollout_ref.actor.fsdp_config.mixed_precision.param_dtype=${dtype} \ 但是发现训练很不稳定
fsdp上的混合精度是设置是这样的: if mixed_precision_config is not None: param_dtype = PrecisionType.to_dtype(mixed_precision_config.get("param_dtype", "bf16")) reduce_dtype = PrecisionType.to_dtype(mixed_precision_config.get("reduce_dtype", "fp32")) buffer_dtype = PrecisionType.to_dtype(mixed_precision_config.get("buffer_dtype", "fp32")) else: param_dtype = torch.bfloat16 reduce_dtype = torch.float32 buffer_dtype = torch.float32 mixed_precision =...
Thank you for your response. However, I didn't understand why Verl moved this part of the code to this specific path. Was the code in the "recipe" path no longer...
ok,thanks
@vermouth1992 Hello, the saved HF model lacks the weight file. How can this be resolved?