Janus
Janus copied to clipboard
wrong number of image tokens
How do you solve the problem of the wrong number of image tokens generated by llm? For example, if the desired number of image tokens is 1024, will the model necessarily generate 1024 tokens? Should there be more or less sometimes? How to relieve?