Classic

Results 3 issues of Classic

This is a very powerful tool, it would be nice to see it implemented using the optimized code: https://github.com/jquesnelle/txt2imghd

The ability to write the start of a message in chat mode and then send it off to the model for it to complete the rest would be very useful....

enhancement

When using llama.cpp, the entire prompt gets processed at each generation making things like chat mode unbearably slow. The problem compounds as the context gets larger and larger as well....