Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Need clarification: Ollama and codellama-70b running. Will Llama Coder use this? #30

Open
ewebgh33 opened this issue Jan 31, 2024 · 2 comments

Comments

@ewebgh33
Copy link

ewebgh33 commented Jan 31, 2024

As title says
If I've already pulled the new (as of 2024-01-30) codellama-70b from meta (or python variant)
Will Llama Coder use this?
Or does it download the 34b and run that?

Does it just run whatever I'm running in Ollama and the list of models you provide are more "recommendations"?
Instructions seem to contradict or are not clear.
One one hand it simply says:

Local Installation
Install Ollama on local machine and then launch the extension in VSCode, everything should work as it is.

But then below that, the list of models doesn't go to 70b and probably doesn't include the new meta ones 70b, 70b python and 70b instruct?

Since my machine is capable of running it, I would prefer to.
Successfully running (and quite fast!) ollama run codellama:70b from here: https://ollama.ai/library/codellama:70b

Only reason I haven't simply installed and launched is because I don't want to end up with a 34b download in an unspecified location despite already running 70b :)

Thanks

@Kevsnz
Copy link
Contributor

Kevsnz commented Feb 5, 2024

For this extension you gonna need FIM (Fill-In-Middle) version of CodeLlama-70b, i.e. https://ollama.ai/library/codellama:70b-code or it's derivatives.

To use it, in extension settings select 'Custom' in 'Inference: Model' dropdown and put model name in 'Inference > Custom: Model' textbox, codellama:70b-code for example. Also make sure 'codellama' option is selected in 'Inference > Custom: Format' dropdown list.

You could try to use non-FIM model 'codellama:70b' (configured the same way as described above) but I doubt it will produce output the extension expects.

@ex3ndr
Copy link
Owner

ex3ndr commented Feb 16, 2024

Latest version has 70b model

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants