Skip to content

Toloka/podvodka

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Podvodka: Training a Language Model to Write Prompts for Stable Diffusion

Step 1:

Train a base model on prompts from Stable Diffusion Discord and image descriptions extracted from them using GPT-3. See behavior_cloning.

Fine-tuned GPT-2 Large: https://huggingface.co/toloka/gpt2-large-supervised-prompt-writing.

Use the following template: image description</s>

W&B report of the final model: https://wandb.ai/toloka-research/gpt-2-self-supervised-prompt-writing/runs/byxqftxf

Step 2:

Collect the annotation using pairwise comparisons of images generated from prompts written by the base model using preference_collection

Train a reward model using reward_model

The final model: https://huggingface.co/toloka/prompts_reward_model

W&B report: https://wandb.ai/toloka-research/prompts_reward_model/runs/bbq7xxbh.

Step 3:

RL fine-tuned base model: https://huggingface.co/toloka/gpt2-large-rl-prompt-writing

W&B Run: https://wandb.ai/pavlichenko/trlx/runs/2v7wfl36

Fine-tuning was done through Carper's trlx.

About

No description, website, or topics provided.

Resources

License

Security policy

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published