ilyasoulk
Results
2
issues of
ilyasoulk
Hello, I am currently training a model using DPO, and I'm adapting the dataset dynamically during training. My current approach looks like this: ```python trainer = DPOTrainer( model, None, args=training_args,...
❓ question
🏋 DPO
Hello, I am trying to reproduce the benchmark results mentioned in the Qwen2.5-Coder technical report. However, I couldn’t find detailed information about the generation parameters (e.g., temperature, top-k, top-p, num...