From Demo to Production: Self-Hosting LLMs with Ollama and Docker