VScode continue plugin in agent mode fails after a while #9132
Unanswered
YanickNoblanc
asked this question in
Help
Replies: 2 comments
-
|
Copying VSCode console logs when problem happens: |
Beta Was this translation helpful? Give feedback.
0 replies
-
|
I tried : restarting plugin, uninstalling/re-installing plugin, switching between Pre-release and Release plugin versions, Problem persists whatever I do. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi all, I am using VSCode Continue plugin (version 1.3.24) on a Windows computer. It can successfully communicate with my vLLM offline server (server has not access to internet). The vLLM server can serve several language models. (Llama-3.1-8B-Instruct, Mistral-7B-Instruct-v0.2, ...). At VSCode side, I have set a local Continue config file at our project root folder. All of this is working fine and we are able to ask questions in a chat prompt, when used in agent mode, Continue plugin is able to locate our source code files, suggest modifications, etc ... But systematicaly, and whatever the language model used is, after a while using agent, Continue plugin fails communicating with our server. vllm logs at server side show that max token have been reached, log:
ValueError: 'max_tokens' or 'max_completion_tokens' is too large: 4096. This model's maximum context length is 32768 tokens and your request has 28739 input tokens (4096 > 32768 - 28739).
The problem only seems to be ate VSCode/Continue side because my model still properly answers questions when going through something else than VSCode (curl requests). Hence my question: Is there a way in agent mode to avoid reaching max_tokens? When this happens, how can we reset or dump history at VSCode side to recover?
Beta Was this translation helpful? Give feedback.
All reactions