From 2135beaa082f8a9fa4059b1d034dcd793aff620c Mon Sep 17 00:00:00 2001 From: randaller Date: Sun, 19 Mar 2023 18:40:03 +0300 Subject: [PATCH] Update README.md --- README.md | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/README.md b/README.md index 690e3ed..c43bd1b 100644 --- a/README.md +++ b/README.md @@ -217,7 +217,7 @@ DATA_FILE_PATH = 'datasets/stable_diffusion_prompts.csv' OUTPUT_DIR = './trained' ``` -*Note: You may also prepare your own dataset, for example, with Positive: and Negative: and even Sampler etc lines interleaving in csv.* +*Note: You may also prepare your own dataset, for example, with Prompt: and Negative: and even Sampler etc lines interleaving in csv.* Then run the training, then after a long-long time, use something like this as a prompt for LLaMA to generate SD prompts: @@ -225,10 +225,10 @@ Then run the training, then after a long-long time, use something like this as a batch = tokenizer("A portrait of a beautiful girl, ", return_tensors="pt") ``` -*Note: If you have prepared and used own dataset with Positive: Negative: lines, the initial LLaMA prompt may look like:* +*Note: If you have prepared and used own dataset with Prompt: Negative prompt: lines, the initial LLaMA prompt may look like:* ``` -batch = tokenizer("Positive: A warship flying thru the Wormhole, ", return_tensors="pt") +batch = tokenizer("Prompt: A warship flying thru the Wormhole, ", return_tensors="pt") ``` Run inference, this should return continued prompt for SD.