Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Any way to use local models? #2

Open
Neltherion opened this issue Oct 12, 2024 · 2 comments
Open

Any way to use local models? #2

Neltherion opened this issue Oct 12, 2024 · 2 comments

Comments

@Neltherion
Copy link

I'm real tired of OpenAI/Groq/... API keys and having to call their APIs over the net. Is there any way to spin up a local LLM/Vision Model and direct the queries to that?

@simularlyon
Copy link
Collaborator

Yes, you can use Ollama for local models.

@saa1605
Copy link
Collaborator

saa1605 commented Oct 14, 2024

The easiest way in the current setup is to use vLLM (https://docs.vllm.ai/en/latest/serving/openai_compatible_server.html) to locally load and host a model on your machine. After this you can use the LMMEnginevLLM by passing the required engine_params and setting the vLLM_ENDPOINT_URL to the locally hosted model.

  1. Export the vLLM_ENDPOINT_URL:
export vLLM_ENDPOINT_URL=YOUR_URL_ADDRESS
  1. Specify engine params to use vllm before instantiating the GraphSearchAgent
engine_params = {'engine_type': 'vllm', 'model': 'YOUR_HOSTED_MODEL_NAME'}

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants