continue
continue copied to clipboard
Long edits are getting cut off / finish early
Before submitting your bug report
- [X] I believe this is a bug. I'll try to join the Continue Discord for questions
- [X] I'm not able to find an open issue that reports the same bug
- [X] I've seen the troubleshooting guide on the Continue Docs
Relevant environment info
- OS: Windows 11 23H2 (Build 22361.4169)
- Continue: v0.8.52
- IDE: VSCode 1.92.1
- Model: deepseek-coder-v2:latest
- config.json:
{
"models": [
{
"title": "Ollama",
"provider": "ollama",
"model": "AUTODETECT"
}
],
"customCommands": [
{
"name": "test",
"prompt": "{{{ input }}}\n\nWrite a comprehensive set of unit tests for the selected code. It should setup, run tests that check for correctness including important edge cases, and teardown. Ensure that the tests are complete and sophisticated. Give the tests just as chat output, don't edit any file.",
"description": "Write unit tests for highlighted code"
},
{
"name": "desc",
"prompt": "{{{ input }}}\n\nWrite a comprehensive comment for each block of the selected code. It should describe what it does and show possible caveats.",
"description": "Comment the highlighted code"
}
],
"tabAutocompleteModel": {
"title": "Starcoder 3b",
"provider": "ollama",
"model": "starcoder2"
},
"allowAnonymousTelemetry": false,
"embeddingsProvider": {
"provider": "transformers.js"
},
"contextProviders": [
{
"name": "open",
"params": {
"onlyPinned": false
}
}
],
"experimental": {
"readResponseTTS": true
},
"ui": {
"showChatScrollbar": true
},
"docs": []
}
Description
While trying to convert a big Python2 file to Python3, output always stops after ~103 ± 3 lines. It sometimes cuts off in the middle of a variable or function name. The source-script is almost 600 lines. The Ollama-log doesn't show any errors. It behaves as if the answer is complete after 2-3 Minutes:
[GIN] 2024/09/19 - 15:00:08 | 200 | 2m40s | 10.8.0.21 | POST "/api/generate"
[GIN] 2024/09/19 - 15:03:41 | 200 | 2m18s | 10.8.0.21 | POST "/api/generate"
I'm not sure if this is a timeout-issue or token exhaustion or something else entirely. The Model has a context length of 163840 and an embedding length of 2048
When using starcoder2 (16384 context length and 3072 embedding length), I get ~150 lines.
To reproduce
- Find a big Python 2 script and open it in VS-Code
- Select All, press Ctrl+I
- Enter "Convert this Python2 program to Python3"
Log output
No response