1.2 KiB
1.2 KiB
Ollama Autocoder
A simple to use Ollama autocompletion engine with options exposed and streaming functionality
Requirements
- Ollama must be serving on the API endpoint applied in settings
- For installation of Ollama, visit ollama.ai
- Ollama must have the model applied in settings installed.
How to Use
- In a text document, press space. The option
Autocomplete with Ollama
or a preview of the first line of autocompletion will appear. Pressenter
to start generation.- Alternatively, you can run the
Autocomplete with Ollama
command from the command pallete (or set a keybind).
- Alternatively, you can run the
- After startup, the tokens will be streamed to your cursor.
- To stop the generation early, press the "Cancel" button on the "Ollama Autocoder" notification or type something.
- Once generation stops, the notification will disappear.
Notes
- For fastest results, an Nvidia GPU or Apple Silicon is recommended. CPU still works on small models.
- The prompt only sees behind the cursor. The model is unaware of text in front of its position.
- For CPU-only, low end, or battery powered devices, it is highly recommended to disable the
response preview
option, as it automatically triggers the model.