hollowstrawberry
commited on
Commit
Β·
7669666
1
Parent(s):
073df5e
Update README.md
Browse files
README.md
CHANGED
@@ -434,30 +434,28 @@ With those way smarter resources out of the way, I'll try to produce a simple gu
|
|
434 |
|
435 |
1. Click the play button on *π¦ Mount your google drive* and give it access when it asks you to. Do the same for *π§ Installation*. Proceed to the next step while it installs on Google's servers.
|
436 |
|
437 |
-
1. Scroll down to *π¨ Settings* but don't run it yet. Here in **Setup** you may give any name you want to your project. You may also change the base model for training, but we'll be using AnythingV3_fp16 as it is the foundation of
|
438 |
|
439 |
1. **Creating a dataset** <a name="datasets"></a>[β²](#index)
|
440 |
|
441 |
This is the largest part of Lora training. You will need to create a "dataset" of images to train with, along with corresponding text files containing descriptions for those images (tags in the case of anime).
|
442 |
|
443 |
-
1. Find some images online representing the character/artstyle/concept you want to convey, possibly on sites such as [
|
444 |
|
445 |
-
1.
|
446 |
|
447 |
-
1.
|
448 |
-
|
449 |
-
1. Once your images and their tags are ready, put them all in a folder following this structure: A folder with your project name, containing at least 1 folder in the format `repetitions_somename`, which each contain your images and matching tags. Like this:
|
450 |
|
451 |
![Folder structure for training](images/trainfolder.png)
|
452 |
|
453 |
-
1. At this point when naming your folders you get to choose the number of repetitions for your dataset.
|
454 |
|
455 |
1. Upload the entire parent folder (the one with your project's name) into your Google Drive's `lora_training/datasets` folder.
|
456 |
|
457 |
1. **Training Parameters** <a name="trainparams"></a>[β²](#index)
|
458 |
|
459 |
* Under **Files**, you don't need to change anything this time.
|
460 |
-
* Under **Steps**, you can change your number of epochs. More epochs means more control over your Lora's progress
|
461 |
* Under **Training**, the `unet_lr` or "learning rate" is the most important parameter. 1e-3 is the default and good for small amounts of images, but it can go as far as 1e-5. You will also find the dim, which is the size of your Lora. I'd recommend dim 16 for most cases, which produces 18 MB Loras. More dim means more capacity for learning but also more capacity to learn garbage.
|
462 |
|
463 |
1. You can now press play on *π¨ Settings*, wait for the model to download, and finally start the training with *π© Cook the Lora*. It should take 20 to 60 minutes if you set a reasonable amount of repetitions and epochs. If it starts counting the steps that means you're in the clear. If you instead encounter an error you can seek troubleshooting or contact me.
|
|
|
434 |
|
435 |
1. Click the play button on *π¦ Mount your google drive* and give it access when it asks you to. Do the same for *π§ Installation*. Proceed to the next step while it installs on Google's servers.
|
436 |
|
437 |
+
1. Scroll down to *π¨ Settings* but don't run it yet. Here in **Setup** you may give any name you want to your project. You may also change the base model for training, but we'll be using AnythingV3_fp16 as it is the foundation of most anime models. I also recommend `animefull-final-pruned` if you can find it. If you want to train with photographs you may copy the link to the [base SD 1.5 model](https://huggingface.co/runwayml/stable-diffusion-v1-5/resolve/main/v1-5-pruned-emaonly.safetensors) or the realistic model you wish to use in the end (such as [Deliberate](https://civitai.com/api/download/models/15236)). In that case remember to change the `model_type` to safetensors.
|
438 |
|
439 |
1. **Creating a dataset** <a name="datasets"></a>[β²](#index)
|
440 |
|
441 |
This is the largest part of Lora training. You will need to create a "dataset" of images to train with, along with corresponding text files containing descriptions for those images (tags in the case of anime).
|
442 |
|
443 |
+
1. Find some images online representing the character/artstyle/concept you want to convey, possibly on sites such as [gelbooru](https://gelbooru.com/). You will need at least 10 images, I'd recommend at least 20. Optionally, you can get hundreds of them using [Grabber](https://github.com/Bionus/imgbrd-grabber/releases) if you want better results.
|
444 |
|
445 |
+
1. If you only have a few images you may tag them yourself, but it may be slow and inaccurate. Optionally, add the [Tagger extension](https://github.com/toriato/stable-diffusion-webui-wd14-tagger) to your webui, through which you can automatically analyze all your training images and generate accurate tags for them. In which case I also recommend you get the [Tag Editor extension](https://github.com/toshiaki1729/stable-diffusion-webui-dataset-tag-editor).
|
446 |
|
447 |
+
1. Once your images and their tags are ready, put them all in a folder following this structure: A folder with your project name, containing at least 1 folder in the format `repetitions_name`, which each contain some images and their tags. Like this:
|
|
|
|
|
448 |
|
449 |
![Folder structure for training](images/trainfolder.png)
|
450 |
|
451 |
+
1. At this point when naming your folders you get to choose the number of repetitions for your dataset. I recommend that your amount of images multiplied by their repetitions does not exceed a total of 400. So, if you have 20 images, I'd recommend 10 to 20 repetitions. More files should mean less repetitions. Then, your inner folder should be called `10_mynewlora` or something similar.
|
452 |
|
453 |
1. Upload the entire parent folder (the one with your project's name) into your Google Drive's `lora_training/datasets` folder.
|
454 |
|
455 |
1. **Training Parameters** <a name="trainparams"></a>[β²](#index)
|
456 |
|
457 |
* Under **Files**, you don't need to change anything this time.
|
458 |
+
* Under **Steps**, you can change your number of epochs. More epochs means more control over your Lora's progress and more time to learn, but don't go overboard. You can also increase the batch size if you have lots of images.
|
459 |
* Under **Training**, the `unet_lr` or "learning rate" is the most important parameter. 1e-3 is the default and good for small amounts of images, but it can go as far as 1e-5. You will also find the dim, which is the size of your Lora. I'd recommend dim 16 for most cases, which produces 18 MB Loras. More dim means more capacity for learning but also more capacity to learn garbage.
|
460 |
|
461 |
1. You can now press play on *π¨ Settings*, wait for the model to download, and finally start the training with *π© Cook the Lora*. It should take 20 to 60 minutes if you set a reasonable amount of repetitions and epochs. If it starts counting the steps that means you're in the clear. If you instead encounter an error you can seek troubleshooting or contact me.
|