-
Notifications
You must be signed in to change notification settings - Fork 206
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
how to set the eval part in generation tasks #72
Comments
Hi, we have a generation example with all the details: Let me know if something is not clear |
Thank you for your instant response. I have already referred to this part of the content before. However, my main point of confusion is whether I should set the eval function name as “ranking” in this generation scenario, as well as any subsequent settings that may be required. Or should I still leave the eval function name to 'accuracy'? |
The short answer is that you should not modify the eval function, because this is done in the More details on the process:
|
and I still got an error like others 'Keyerror' related to samples ### Here is my default_config: use_wandb: False annotator: predictor: meta_prompts: eval: llm: stop_criteria: ### Following is the logging: Starting step 0 ...... |
the command is: run_generation_pipeline.py |
It seems like the model was not able to generate samples. There are log files that are generated in the dump folder that might help to understand the root cause, feel free to reach out in the discord channel if you need help with the debugging process. |
log file content: 2024-07-17 16:22:22,175 - DEBUG - load_ssl_context verify=True cert=None trust_env=True http2=False
Please classify the text accordingly, ensuring that the quality of the generated sample is evaluated based on how well it adheres to the given user prompt. |
It seems from the log that you are trying to use GPT-3.5.
|
I have changed to mdoel to gpt-4-1106-preview, and error remains the same. Can you please help me check with the prompt files I use? predictor: meta_prompts: |
perhaps it's related with settings: annotator: begging for a generating example T.T |
This is fine since there is no annotator in the generation optimization (only the evaluator uses the ranker). |
well I restore all the default parameters, use GPT-4, and run classifying task. I still get 'Keyerror' of 'samples'. So I revised the 'optimization_pipeline.py' scripts and added some parse functions to extract 'Sample #' from the 'text'. And finally I successfully run the classifying sample. However, generation sample remains unsolved. I get the error: |
It seems that the issue was sensitive to lower/upper cases (you used 'azure' and the system expected 'Azure') |
In face of generation situations, I guess we should set function name to ranking. And as so we should set function_params. That's not so obvious.
the following not works for me.
Can we have a concrete demo for generation tasks? Thanks in advance
eval:
function_name: 'ranking'
function_params:
prompt: 'prompts/predictor_completion/prediction_generation.prompt'
mini_batch_size: 1
llm:
type: '...'
num_workers: 10
mode: 'verbose'
label_schema: ["1","2","3","4","5"]
The text was updated successfully, but these errors were encountered: