Replies: 1 comment 2 replies
-
Thank you for the detailed question. From how you describe the issue, I can see at least the following possible reason: As you are only training on pairs, meaning not a multi-turn conversation but only single question + single answer, the inference your are performing in the chat tab with multi-turn conversations is completely out of distribution. You might want to chain multiple questions and answers together if you want the model to also work well with this setup. Your dataset would look like this:
of course can chain more Q/A pairs together or mix up the number of the chains a bit to add diversity to your train dataset. Regarding your second question: |
Beta Was this translation helpful? Give feedback.
-
Hi,
i 'm not very deep into finetuning LLMs (two Days, to be precise) and i have a rather "simple" Question.
After feeding a .csv with QA-Pairs about a certain programming language, using h2oai/h2o-danube3-500m-base (can´t use more because iḿ limited to two GPUS with 6 GB each), Metric BLEU, 512 Tokens, LORA, Token Averaged Cross Entropy, Batch Size 2, up to 30 Epochs, the Model (running in "chat") initially gives a correct answer after the first question, but wrong answers for the following questions and mixes up the first correct answer with the new answer, means its hallucinating. I can only prevent this by closing the chat entirely and open a new one.
Can anybody explain this behavior and a possible cause (wrong Training parameters, "works as intended")?
Also, could it be possible with LLM Studio to train - not - on GPU, but on CPU (two Xeons, 128 GB RAM) or automatically swap the load off to the CPU(s)? Performing slower wouldn´t be the issue for me, but it would open the possibility to use larger (better) models for finetuning.
Best regards & thank you !
Beta Was this translation helpful? Give feedback.
All reactions