ALAE icon indicating copy to clipboard operation
ALAE copied to clipboard

Fine-tuning trained model on new dataset

Open 5agado opened this issue 4 years ago • 3 comments

A big advantage of StyleGAN is the seamless fine-tuning process, where a previous checkpoint can be used as a starting point for training on a new dataset (say for example fine-tune the FFHQ model on paintings).

Is this possible for ALAE too? Do you have any pointers or feedback on how to approach it?

5agado avatar May 31 '20 11:05 5agado

Go to StyleGAN2 page, use the script to convert a dataset to tfrecords. Then use the training script from this repo.

smthomas-sci avatar Jul 01 '20 21:07 smthomas-sci

@smthomas-sci thanks, but my question was more about the fine-tuning process, then simple dataset conversion.

However I just tried myself, and simply pointing to already present ALAE checkpoints and passing a new dataset to the training config works. Two caveats:

  • as we reached already top resolution, only one checkpoint is created, and overwritten every time, so you can't roll back to previous fine-tuned versions. The solution is simply to edit the code to create unique checkpoints more often.
  • process doesn't seem as stable as for StyleGAN. As you can see from my results, it diverges quickly and abruptly

sample_182_0 sample_186_0

5agado avatar Jul 06 '20 11:07 5agado

I see what you mean now. I can’t offer much advice here, beyond that it looks like mode collapse. Decrease your learning rate and increase your batch size? You could try adding cumulative gradients to the training script to increase your batch size if memory is an issue.

smthomas-sci avatar Jul 06 '20 11:07 smthomas-sci