Beyond Chat: How to Set Up Local LLM Code Completion in VS Code with Ollama