j
Model Maximum Context Length Error
Yes, but when I do that, specifically setting to 8192, I get a separate error saying that I have exceeded the maximum context length. But in general, even if I manage to set it a little higher, won't I run into the same problem then?
20 replies
Model Maximum Context Length Error
But am now running into a new error:
I didn't see this when using the serverless endpoints. So my question:
- Is there something I can be setting on vLLM to automatically manage the context length for me? I.e. to delete tokens from the
prompt
or messages
automatically for me? Or do I need to manage this myself?
Thanks!20 replies