Replies: 2 comments
-
Hey @jagrut14, great to see you diving into another challenge with LangChain! 🚀 Based on the provided context, it seems that the "ms-marco-MultiBERT-L-12" model is being used in the To specify a local path for a model, you would typically need to adjust the model initialization logic to check for a local model path before downloading it. This often involves passing a path to the model's constructor or initialization method, or setting a configuration option to use a local model path. Unfortunately, without the specific implementation details of the model loading mechanism in LangChain, the precise code changes cannot be provided. You might need to explore the implementation of As for the I hope this information helps. If you need further assistance, please provide more details about the model loading mechanism in LangChain or the specific part of the code where the "ms-marco-MultiBERT-L-12" model is being loaded. Sources
|
Beta Was this translation helpful? Give feedback.
-
hi @jagrut14 , I found from FlashRerank code at https://github.com/PrithivirajDamodaran/FlashRank/blob/main/flashrank/Config.py, it show the default cache folder is "/tmp". If it is running on Windows, most likely it is at C:/tmp/ |
Beta Was this translation helpful? Give feedback.
-
Checked other resources
Commit to Help
Example Code
Description
I want to use the Reranker function with my following code. Without it downloading the mode "ms-marco-MultiBERT-L-12" every time. Is there a way where I can locate this model locally and provide the path of the model to this function?
System Info
langchain==0.1.9
Beta Was this translation helpful? Give feedback.
All reactions