Replies: 5 comments 3 replies
-
2025-05-07T16:20:28.966533_a_390634270895595536_llm_request.json |
Beta Was this translation helpful? Give feedback.
-
qwen3:14b with 16K context window worked on a taskv2: ![]() |
Beta Was this translation helpful? Give feedback.
-
Qwen3:14b with 40960 size context window did work to get Top 3. ![]() |
Beta Was this translation helpful? Give feedback.
-
It may not be on Skyvern's roadmap but optimizing an open source model and the prompts for that file may be a winning path. |
Beta Was this translation helpful? Give feedback.
-
qwen2.5vl:32b with an updated context window of 125k has my workflows working. It fails with the default context window. 7b failed with 125k and default. It seemed to have the right responses so I am wondering if there is some other issue. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Are you getting the "AttributeError: 'str' object has no attribute 'get'" error? You're probably losing context.
I have been doing some experimenting with the default "Top 3 results on Hacker News" topic to figure out what is going on. My Top 3 results were always the bottom results from the HN list, which didn't make sense unless the top of the prompt was being lost.
This led me to discover that Ollama has a default context length of 2048. So if you just use a model without updating that context length, you'll get weird results because it only gets the end of the context. Skyvern passes the DOM/html, which can mean very large context, much larger than 2048. For example, many of the instructions, like outputting valid JSON, are at the beginning of the message but the beginning of the message gets lost to the context window.
The solution is to add a context size model parameter called num_ctx. The larger the number, the more memory your system needs to have. I could not get the full context processed even at num_ctx 8192. It worked with num_ctx 16384 but I did not try anything less as I have 128gb of RAM.
The simplest way to set this is by updating the Ollama model file. You can read how to do so here: https://help.nurgo-software.com/article/202-optimizing-ollama-models-for-brainsoup
I named my updated model as qwen3:30b-a3b-max_context. You then need to add this model to your docker-compose file.
So, I am now getting valid JSON. However, it is not giving me the top 3 results. Almost like it is taking the request as a request on how to get the Top 3 rather than a request to provide the Top 3. I will try to figure out why next. It may also be context related.
I am attaching my docker-compose.yml, request.json, and response.json.
Beta Was this translation helpful? Give feedback.
All reactions