You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When sending text that is larger than the context lenght of the model, vLLM throws an error with:
This model\'s maximum context length is 256 tokens. However, you requested 258 tokens in the input for embedding generation. Please reduce the length of the input.
Is there an option to enable some sort of auto truncation?
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
-
When sending text that is larger than the context lenght of the model, vLLM throws an error with:
Is there an option to enable some sort of auto truncation?
Beta Was this translation helpful? Give feedback.
All reactions