Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat: support a local custom cortex engine #3764

Closed
RyanMetcalfeInt8 opened this issue Oct 3, 2024 · 6 comments
Closed

feat: support a local custom cortex engine #3764

RyanMetcalfeInt8 opened this issue Oct 3, 2024 · 6 comments
Labels
category: providers Local & remote inference providers duplicate This issue or pull request already exists move to Cortex

Comments

@RyanMetcalfeInt8
Copy link

Hello, I have a custom cortex engine (compiled into an engine.dll). Right now I am able to make use of it in a super hacky way, by replacing the default cortex.llamacpp engine.dll with my custom version. But obviously I don't want to have to do that..

So, I am wondering -- What the correct way to make use of my custom engine? I see in model.json that there is an 'engine' field, which is typically set to 'nitro'. My confusion here is that 'nitro' seems equivalent to 'cortex' (which I also use), but I don't see any field that can specify which specific engine to use (cortex.llamacpp vs. mine).

It's very possible that I missed / overlooked something..

Perhaps there is also some feature needed in cortex.cpp layer as well?

Thanks!
Ryan

@github-project-automation github-project-automation bot moved this to Investigating in Menlo Oct 3, 2024
@louis-menlo
Copy link
Contributor

louis-menlo commented Oct 3, 2024

Hi @RyanMetcalfeInt8, that's awesome! You're right, existing models have defined nitro as the engine, but the extension will actually route to the cortex.cpp server underneath, it would run on the llama.cpp engine (version 0.5.5). We're working on multiple engines support, which is tracked here. #3690

@RyanMetcalfeInt8
Copy link
Author

Hey @louis-jan -- Thanks for getting back to me so quickly! Makes sense.. I'll subscribe to that issue.. and happy to help test new features as they become available. Cheers!

@github-project-automation github-project-automation bot moved this from Investigating to Review + QA in Menlo Oct 3, 2024
@louis-menlo
Copy link
Contributor

Hey @RyanMetcalfeInt8, I think this is a valid issue to track for improving the extension dev experience. I encountered the same issue while working on MLX support. I will reopen it for tracking.

@louis-menlo louis-menlo reopened this Oct 3, 2024
@github-project-automation github-project-automation bot moved this from Review + QA to In Progress in Menlo Oct 3, 2024
@freelerobot freelerobot changed the title [Question] How can I make use of my custom cortex engine? feat: support a local custom cortex engine Oct 14, 2024
@freelerobot
Copy link
Contributor

related #3773

@freelerobot freelerobot added category: providers Local & remote inference providers and removed category: local providers labels Oct 17, 2024
@imtuyethan
Copy link
Contributor

@imtuyethan
Copy link
Contributor

imtuyethan commented Dec 10, 2024

@github-project-automation github-project-automation bot moved this from In Progress to QA in Menlo Dec 10, 2024
@imtuyethan imtuyethan moved this from QA to Completed in Menlo Dec 10, 2024
@imtuyethan imtuyethan added the duplicate This issue or pull request already exists label Dec 10, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
category: providers Local & remote inference providers duplicate This issue or pull request already exists move to Cortex
Projects
Archived in project
Development

No branches or pull requests

4 participants