Right now this tool requires OpenAI to use. Instead the tool should use a locally running olllama server that it accesses via a local socket connection. Specifically:
- Add ollama as an LLM backend
- Add a command line option to select OpenAI or ollama
- Add a menu toggle to switch between the two modes