Fabelous-Autocoder/package.json

105 lines
2.9 KiB
JSON

{
"name": "ollama-autocoder",
"displayName": "Ollama Autocoder",
"description": "A simple to use Ollama autocompletion engine with options exposed and streaming functionality",
"version": "0.0.4",
"icon": "icon.png",
"publisher": "10nates",
"license": "MIT",
"bugs": {
"url": "https://github.com/10Nates/ollama-autocoder/issues"
},
"sponsor": {
"url": "https://ko-fi.com/natehedge"
},
"repository": {
"type": "git",
"url": "https://github.com/10Nates/ollama-autocoder"
},
"engines": {
"vscode": "^1.73.0"
},
"categories": [
"Machine Learning",
"Snippets",
"Programming Languages"
],
"keywords": [
"llama",
"ollama",
"gpt",
"coding",
"autocomplete",
"open source",
"assistant",
"ai",
"llm"
],
"galleryBanner": {
"color": "#f5ead1"
},
"activationEvents": [
"onStartupFinished"
],
"main": "./out/extension.js",
"contributes": {
"configuration": {
"title": "Ollama Autocoder",
"properties": {
"ollama-autocoder.endpoint": {
"type": "string",
"default": "http://localhost:11434/api/generate",
"description": "The endpoint of the ollama REST API"
},
"ollama-autocoder.model": {
"type": "string",
"default": "openhermes2.5-mistral:7b-q4_K_M",
"description": "The model to use for generating completions"
},
"ollama-autocoder.raw input": {
"type": "boolean",
"default": false,
"description": "Prompt the model without formatting. Disables system message. Turn this on if you are having trouble with a model falling out of coding mode."
},
"ollama-autocoder.system message": {
"type": "string",
"default": "You are a code autocompletion engine. Respond with a continuation of the code provided and nothing else. Code should not be in a code block. Anything that is not code should be written as a code comment.",
"description": "The system message to use for code completions. Type DEFAULT for Makefile."
},
"ollama-autocoder.max tokens predicted": {
"type": "integer",
"default": 500,
"description": "The maximum number of tokens generated by the model."
},
"ollama-autocoder.prompt window size": {
"type": "integer",
"default": 2000,
"description": "The size of the prompt in characters. NOT tokens, so can be set about 1.5-2x the max tokens of the model (varies)."
}
}
},
"commands": [
{
"command": "ollama-autocoder.autocomplete",
"title": "Autocomplete with Ollama"
}
]
},
"scripts": {
"vscode:prepublish": "npm run compile",
"compile": "tsc -p ./",
"lint": "eslint \"src/**/*.ts\"",
"watch": "tsc -watch -p ./"
},
"devDependencies": {
"@types/node": "^16.18.34",
"@types/vscode": "^1.73.0",
"@typescript-eslint/eslint-plugin": "^6.7.0",
"@typescript-eslint/parser": "^6.7.0",
"eslint": "^8.26.0",
"typescript": "^5.3.2"
},
"dependencies": {
"axios": "^1.6.2"
}
}