You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Thanks for coding this little wonder of extension. Kudos! I've been using it for a bit, and I have noticed that every autocompletion generates an extra request to the /api/tags endpoint in Ollama:
I suspect it comes from the call to ollamaCheckModel() in provideInlineCompletionItems():
In my view it should not be necessary to send a request to the /api/tags endpoint every time. I am aware the latency it introduces is orders of magnitude lower than the /api/generate cat, but still ... it's extra job for the extension that (in my view) does not need to do.
I'd suggest to go for a different strategy 🤔 Perhaps do the check once and save the list of available models to check locally. Then check again whenever the configuration changes, or every now and then.
Thanks!
The text was updated successfully, but these errors were encountered:
Hi!
Thanks for coding this little wonder of extension. Kudos! I've been using it for a bit, and I have noticed that every autocompletion generates an extra request to the
/api/tags
endpoint in Ollama:I suspect it comes from the call to
ollamaCheckModel()
inprovideInlineCompletionItems()
:llama-coder/src/prompts/provider.ts
Line 89 in 996ac71
In my view it should not be necessary to send a request to the
/api/tags
endpoint every time. I am aware the latency it introduces is orders of magnitude lower than the/api/generate
cat, but still ... it's extra job for the extension that (in my view) does not need to do.I'd suggest to go for a different strategy 🤔 Perhaps do the check once and save the list of available models to check locally. Then check again whenever the configuration changes, or every now and then.
Thanks!
The text was updated successfully, but these errors were encountered: