Carlos Echevarria
Results
2
comments of
Carlos Echevarria
> For now, there is a hard limitation of 2048 input tokens and a maximum of 1920 output tokens. We might consider increasing these numbers in the future. Is there...
> Hey - could you elaborate a bit? static context length is kind of a intrinsic thing to transformer based LLMs. Sorry. Dependent on the models, the context can be...