Skip to content

Latest commit

 

History

History
32 lines (15 loc) · 1.02 KB

README.md

File metadata and controls

32 lines (15 loc) · 1.02 KB

Podvodka: Training a Language Model to Write Prompts for Stable Diffusion

Step 1:

Train a base model on prompts from Stable Diffusion Discord and image descriptions extracted from them using GPT-3. See behavior_cloning.

Fine-tuned GPT-2 Large: https://huggingface.co/toloka/gpt2-large-supervised-prompt-writing.

Use the following template: image description</s>

W&B report of the final model: https://wandb.ai/toloka-research/gpt-2-self-supervised-prompt-writing/runs/byxqftxf

Step 2:

Collect the annotation using pairwise comparisons of images generated from prompts written by the base model using preference_collection

Train a reward model using reward_model

The final model: https://huggingface.co/toloka/prompts_reward_model

W&B report: https://wandb.ai/toloka-research/prompts_reward_model/runs/bbq7xxbh.

Step 3:

RL fine-tuned base model: https://huggingface.co/toloka/gpt2-large-rl-prompt-writing

W&B Run: https://wandb.ai/pavlichenko/trlx/runs/2v7wfl36

Fine-tuning was done through Carper's trlx.