made settings prettier and also broke whatever settings were currently set. Had to do it at some point
This commit is contained in:
parent
259730806f
commit
e4ee5d6dd3
|
@ -47,22 +47,22 @@
|
|||
"default": "openhermes2.5-mistral:7b-q4_K_M",
|
||||
"description": "The model to use for generating completions"
|
||||
},
|
||||
"ollama-autocoder.raw-input": {
|
||||
"ollama-autocoder.raw input": {
|
||||
"type": "boolean",
|
||||
"default": false,
|
||||
"description": "Prompt the model without formatting. Disables system message. Turn this on if you are having trouble with a model falling out of coding mode."
|
||||
},
|
||||
"ollama-autocoder.system-message": {
|
||||
"ollama-autocoder.system message": {
|
||||
"type": "string",
|
||||
"default": "You are a code autocompletion engine. Respond with a continuation of the code provided and nothing else. Code should not be in a code block. Anything that is not code should be written as a code comment.",
|
||||
"description": "The system message to use for code completions. Type DEFAULT for Makefile."
|
||||
},
|
||||
"ollama-autocoder.max-tokens-predicted": {
|
||||
"ollama-autocoder.max tokens predicted": {
|
||||
"type": "integer",
|
||||
"default": 500,
|
||||
"description": "The maximum number of tokens generated by the model."
|
||||
},
|
||||
"ollama-autocoder.prompt-window-size": {
|
||||
"ollama-autocoder.prompt window size": {
|
||||
"type": "integer",
|
||||
"default": 2000,
|
||||
"description": "The size of the prompt in characters. NOT tokens, so can be set about 1.5-2x the max tokens of the model (varies)."
|
||||
|
|
|
@ -15,10 +15,10 @@ function updateVSConfig() {
|
|||
VSConfig = vscode.workspace.getConfiguration("ollama-autocoder");
|
||||
apiEndpoint = VSConfig.get("apiEndpoint") || "http://localhost:11434/api/generate";
|
||||
apiModel = VSConfig.get("model") || "openhermes2.5-mistral:7b-q4_K_M"; // The model I tested with
|
||||
apiSystemMessage = VSConfig.get("system-message");
|
||||
numPredict = VSConfig.get("max-tokens-predicted") || 500;
|
||||
promptWindowSize = VSConfig.get("prompt-window-size") || 2000;
|
||||
rawInput = VSConfig.get("raw-input");
|
||||
apiSystemMessage = VSConfig.get("system message");
|
||||
numPredict = VSConfig.get("max tokens predicted") || 500;
|
||||
promptWindowSize = VSConfig.get("prompt window size") || 2000;
|
||||
rawInput = VSConfig.get("raw input");
|
||||
|
||||
if (apiSystemMessage == "DEFAULT" || rawInput) apiSystemMessage = undefined;
|
||||
}
|
||||
|
|
Loading…
Reference in New Issue