Carlos Echevarria

Results 2 comments of Carlos Echevarria

> For now, there is a hard limitation of 2048 input tokens and a maximum of 1920 output tokens. We might consider increasing these numbers in the future. Is there...

> Hey - could you elaborate a bit? static context length is kind of a intrinsic thing to transformer based LLMs. Sorry. Dependent on the models, the context can be...