The extension now supports local AI models via Ollama. Useful for those who prefer not to pay for Copilot or OpenAI.
Here's how to set it up:
- Download and install Ollama for your operating system.
- Set the environment variables to use less memory:
OLLAMA_KEEP_ALIVE = 1h
OLLAMA_FLASH_ATTENTION = 1
- Restart Ollama.
- Download the AI model Gemma 2:
ollama pull gemma2:2b
- Change the Proofread settings:
proofread.ai.vendor = ollama
proofread.ai.model = gemma2:2b
That's it!
Gemma 2 is a lightweight model that uses about 1GB of memory and works quickly without a GPU. For good results, send only a few paragraphs at a time for proofreading or translation.
For larger documents or improved results, try models like mistral
or mistral-nemo
.