-
Notifications
You must be signed in to change notification settings - Fork 30
Open
Labels
enhancementNew feature or requestNew feature or request
Description
Problem
When code is generated and submitted to PDD Cloud, we do not capture:
- Which LLM was used
- Temperature setting
- Thinking token budget
- Reasoning type (effort/budget)
From Dec 13 Benchmarking Meeting:
"Right now I do not put the LM that was used. I do not put down a temperature that was used... I do not put in thinking tokens"
Current Firestore Schema (inferred)
prompt, generated_code, generated_example, generated_test, embedding
Proposed Addition
llm_model, temperature, thinking_tokens, reasoning_type, generation_cost, timestamp
Benefits
- Understand which models produce best results for which prompts
- Enable model routing based on historical success
- Track cost trends over time
- Support future cloud model routing - automatically pick best model for problem type
Context from Meeting
"Different models excel at different tasks (jagged intelligence). Users should not have to manually figure out which model works best."
This metadata capture is a prerequisite for intelligent model routing in PDD Cloud.
Metadata
Metadata
Assignees
Labels
enhancementNew feature or requestNew feature or request