This repository primarily provides a Gradio GUI for Kohya's Stable Diffusion trainers. However, support for Linux OS is also offered through community contributions. macOS support is not optimal at the moment but might work if the conditions are favorable.
The GUI allows you to set the training parameters and generate and run the required CLI commands to train the model.
- Kohya's GUI
This Colab notebook was not created or maintained by me; however, it appears to function effectively. The source can be found at: https://github.com/camenduru/kohya_ss-colab.
I would like to express my gratitude to camendutu for their valuable contribution. If you encounter any issues with the Colab notebook, please report them on their repository.
Colab | Info |
---|---|
kohya_ss_gui_colab |
To install the necessary dependencies on a Windows system, follow these steps:
-
Install Python 3.10.11.
- During the installation process, ensure that you select the option to add Python to the 'PATH' environment variable.
-
Install CUDA 11.8 toolkit.
-
Install Git.
-
Install the Visual Studio 2015, 2017, 2019, and 2022 redistributable.
To set up the project, follow these steps:
-
Open a terminal and navigate to the desired installation directory.
-
Clone the repository by running the following command:
git clone https://github.com/bmaltais/kohya_ss.git
-
Change into the
kohya_ss
directory:cd kohya_ss
-
Run the setup script by executing the following command:
.\setup.bat
During the accelerate config step, use the default values as proposed during the configuration unless you know your hardware demands otherwise. The amount of VRAM on your GPU does not impact the values used.
The following steps are optional but will improve the learning speed for owners of NVIDIA 30X0/40X0 GPUs. These steps enable larger training batch sizes and faster training speeds.
- Run
.\setup.bat
and select2. (Optional) Install cudnn files (if you want to use the latest supported cudnn version)
.
To install the necessary dependencies on a Linux system, ensure that you fulfill the following requirements:
-
Ensure that
venv
support is pre-installed. You can install it on Ubuntu 22.04 using the command:apt install python3.10-venv
-
Install the CUDA 11.8 Toolkit by following the instructions provided in this link.
-
Make sure you have Python version 3.10.9 or higher (but lower than 3.11.0) installed on your system.
To set up the project on Linux or macOS, perform the following steps:
-
Open a terminal and navigate to the desired installation directory.
-
Clone the repository by running the following command:
git clone https://github.com/bmaltais/kohya_ss.git
-
Change into the
kohya_ss
directory:cd kohya_ss
-
If you encounter permission issues, make the
setup.sh
script executable by running the following command:chmod +x ./setup.sh
-
Run the setup script by executing the following command:
./setup.sh
Note: If you need additional options or information about the runpod environment, you can use
setup.sh -h
orsetup.sh --help
to display the help message.
The default installation location on Linux is the directory where the script is located. If a previous installation is detected in that location, the setup will proceed there. Otherwise, the installation will fall back to /opt/kohya_ss
. If /opt
is not writable, the fallback location will be $HOME/kohya_ss
. Finally, if none of the previous options are viable, the installation will be performed in the current directory.
For macOS and other non-Linux systems, the installation process will attempt to detect the previous installation directory based on where the script is run. If a previous installation is not found, the default location will be $HOME/kohya_ss
. You can override this behavior by specifying a custom installation directory using the -d
or --dir
option when running the setup script.
If you choose to use the interactive mode, the default values for the accelerate configuration screen will be "This machine," "None," and "No" for the remaining questions. These default answers are the same as the Windows installation.
To install the necessary components for Runpod and run kohya_ss, follow these steps:
-
Select the Runpod pytorch 2.0.1 template. This is important. Other templates may not work.
-
SSH into the Runpod.
-
Clone the repository by running the following command:
cd /workspace git clone https://github.com/bmaltais/kohya_ss.git
-
Run the setup script:
cd kohya_ss ./setup-runpod.sh
-
Run the GUI with:
./gui.sh --share --headless
or with this if you expose 7860 directly via the runpod configuration:
./gui.sh --listen=0.0.0.0 --headless
-
Connect to the public URL displayed after the installation process is completed.
To run from a pre-built Runpod template, you can:
-
Open the Runpod template by clicking on https://runpod.io/gsc?template=ya6013lj5a&ref=w18gds2n.
-
Deploy the template on the desired host.
-
Once deployed, connect to the Runpod on HTTP 3010 to access the kohya_ss GUI. You can also connect to auto1111 on HTTP 3000.
If you prefer to use Docker, follow the instructions below:
-
Ensure that you have Git and Docker installed on your Windows or Linux system.
-
Open your OS shell (Command Prompt or Terminal) and run the following commands:
git clone --recursive https://github.com/bmaltais/kohya_ss.git cd kohya_ss docker compose up -d --build
Note: The initial run may take up to 20 minutes to complete.
Please be aware of the following limitations when using Docker:
- All training data must be placed in the
dataset
subdirectory, as the Docker container cannot access files from other directories. - The file picker feature is not functional. You need to manually set the folder path and config file path.
- Dialogs may not work as expected, and it is recommended to use unique file names to avoid conflicts.
- This Dockerfile has been designed to be easily disposable. You can discard the container at any time and docker build it with a new version of the code. To update the system, run update scripts outside of Docker and rebuild using
docker compose down && docker compose up -d --build
.
If you are running Linux, an alternative Docker container port with fewer limitations is available here.
- All training data must be placed in the
You may want to use the following Dockerfile repositories to build the images:
- Standalone Kohya_ss template: https://github.com/ashleykleynhans/kohya-docker
- Auto1111 + Kohya_ss GUI template: https://github.com/ashleykleynhans/stable-diffusion-docker
To upgrade your installation to a new version, follow the instructions below.
If a new release becomes available, you can upgrade your repository by running the following commands from the root directory of the project:
-
Pull the latest changes from the repository:
git pull
-
Run the setup script:
.\setup.bat
To upgrade your installation on Linux or macOS, follow these steps:
-
Open a terminal and navigate to the root directory of the project.
-
Pull the latest changes from the repository:
git pull
-
Refresh and update everything:
./setup.sh
To launch the GUI service, you can use the provided scripts or run the kohya_gui.py
script directly. Use the command line arguments listed below to configure the underlying service.
--listen: Specify the IP address to listen on for connections to Gradio.
--username: Set a username for authentication.
--password: Set a password for authentication.
--server_port: Define the port to run the server listener on.
--inbrowser: Open the Gradio UI in a web browser.
--share: Share the Gradio UI.
--language: Set custom language
On Windows, you can use either the gui.ps1
or gui.bat
script located in the root directory. Choose the script that suits your preference and run it in a terminal, providing the desired command line arguments. Here's an example:
gui.ps1 --listen 127.0.0.1 --server_port 7860 --inbrowser --share
or
gui.bat --listen 127.0.0.1 --server_port 7860 --inbrowser --share
To launch the GUI on Linux or macOS, run the gui.sh
script located in the root directory. Provide the desired command line arguments as follows:
gui.sh --listen 127.0.0.1 --server_port 7860 --inbrowser --share
The repository now provides a default configuration file named config.toml
. This file is a template that you can customize to suit your needs.
To use the default configuration file, follow these steps:
- Copy the
config example.toml
file from the root directory of the repository toconfig.toml
. - Open the
config.toml
file in a text editor. - Modify the paths and settings as per your requirements.
This approach allows you to easily adjust the configuration to suit your specific needs to open the desired default folders for each type of folder/file input supported in the GUI.
You can specify the path to your config.toml (or any other name you like) when running the GUI. For instance: ./gui.bat --config c:\my_config.toml
To train a LoRA, you can currently use the train_network.py
code. You can create a LoRA network by using the all-in-one GUI.
Once you have created the LoRA network, you can generate images using auto1111 by installing this extension.
A prompt file might look like this, for example:
# prompt 1
masterpiece, best quality, (1girl), in white shirts, upper body, looking at viewer, simple background --n low quality, worst quality, bad anatomy, bad composition, poor, low effort --w 768 --h 768 --d 1 --l 7.5 --s 28
# prompt 2
masterpiece, best quality, 1boy, in business suit, standing at street, looking back --n (low quality, worst quality), bad anatomy, bad composition, poor, low effort --w 576 --h 832 --d 2 --l 5.5 --s 40
Lines beginning with #
are comments. You can specify options for the generated image with options like --n
after the prompt. The following options can be used:
--n
: Negative prompt up to the next option.--w
: Specifies the width of the generated image.--h
: Specifies the height of the generated image.--d
: Specifies the seed of the generated image.--l
: Specifies the CFG scale of the generated image.--s
: Specifies the number of steps in the generation.
The prompt weighting such as ( )
and [ ]
is working.
If you encounter any issues, refer to the troubleshooting steps below.
If you encounter an X error related to the page file, you may need to increase the page file size limit in Windows.
If you encounter an error indicating that the module tkinter
is not found, try reinstalling Python 3.10 on your system.
When training LORA on a TESLA V100, users reported low GPU utilization. Additionally, there was difficulty in specifying GPUs other than the default for training.
- GPU Selection: Users can specify GPU IDs in the setup configuration to select the desired GPUs for training.
- Improving GPU Load: Utilizing
adamW8bit
optimizer and increasing the batch size can help achieve 70-80% GPU utilization without exceeding GPU memory limits.
The documentation in this section will be moved to a separate document later.
The masked loss is supported in each training script. To enable the masked loss, specify the --masked_loss
option.
The feature is not fully tested, so there may be bugs. If you find any issues, please open an Issue.
ControlNet dataset is used to specify the mask. The mask images should be the RGB images. The pixel value 255 in R channel is treated as the mask (the loss is calculated only for the pixels with the mask), and 0 is treated as the non-mask. The pixel values 0-255 are converted to 0-1 (i.e., the pixel value 128 is treated as the half weight of the loss). See details for the dataset specification in the LLLite documentation.
- Relocate config accordion to the top of the GUI.
- Fix dataset preparation bug.
- Added config.toml support for wd14_caption.
- Added support for Huber loss under the Parameters / Advanced tab.
-
Update sd-scripts to 0.8.7
-
The default value of
huber_schedule
in Scheduled Huber Loss is changed fromexponential
tosnr
, which is expected to give better results. -
Highlights
- The dependent libraries are updated. Please see Upgrade and update the libraries.
- Especially
imagesize
is newly added, so if you cannot update the libraries immediately, please install withpip install imagesize==1.4.1
separately. bitsandbytes==0.43.0
,prodigyopt==1.0
,lion-pytorch==0.0.6
are included in the requirements.txt.bitsandbytes
no longer requires complex procedures as it now officially supports Windows.
- Also, the PyTorch version is updated to 2.1.2 (PyTorch does not need to be updated immediately). In the upgrade procedure, PyTorch is not updated, so please manually install or update torch, torchvision, xformers if necessary (see Upgrade PyTorch).
- Especially
- When logging to wandb is enabled, the entire command line is exposed. Therefore, it is recommended to write wandb API key and HuggingFace token in the configuration file (
.toml
). Thanks to bghira for raising the issue.- A warning is displayed at the start of training if such information is included in the command line.
- Also, if there is an absolute path, the path may be exposed, so it is recommended to specify a relative path or write it in the configuration file. In such cases, an INFO log is displayed.
- See #1123 and PR #1240 for details.
- Colab seems to stop with log output. Try specifying
--console_log_simple
option in the training script to disable rich logging. - Other improvements include the addition of masked loss, scheduled Huber Loss, DeepSpeed support, dataset settings improvements, and image tagging improvements. See below for details.
- The dependent libraries are updated. Please see Upgrade and update the libraries.
-
Training scripts
train_network.py
andsdxl_train_network.py
are modified to record some dataset settings in the metadata of the trained model (caption_prefix
,caption_suffix
,keep_tokens_separator
,secondary_separator
,enable_wildcard
).- Fixed a bug that U-Net and Text Encoders are included in the state in
train_network.py
andsdxl_train_network.py
. The saving and loading of the state are faster, the file size is smaller, and the memory usage when loading is reduced. - DeepSpeed is supported. PR #1101 and #1139 Thanks to BootsofLagrangian! See PR #1101 for details.
- The masked loss is supported in each training script. PR #1207 See Masked loss for details.
- Scheduled Huber Loss has been introduced to each training scripts. PR #1228 Thanks to kabachuha for the PR and cheald, drhead, and others for the discussion! See the PR and Scheduled Huber Loss for details.
- The options
--noise_offset_random_strength
and--ip_noise_gamma_random_strength
are added to each training script. These options can be used to vary the noise offset and ip noise gamma in the range of 0 to the specified value. PR #1177 Thanks to KohakuBlueleaf! - The options
--save_state_on_train_end
are added to each training script. PR #1168 Thanks to gesen2egee! - The options
--sample_every_n_epochs
and--sample_every_n_steps
in each training script now display a warning and ignore them when a number less than or equal to0
is specified. Thanks to S-Del for raising the issue.
-
Dataset settings
- The English version of the dataset settings documentation is added. PR #1175 Thanks to darkstorm2150!
- The
.toml
file for the dataset config is now read in UTF-8 encoding. PR #1167 Thanks to Horizon1704! - Fixed a bug that the last subset settings are applied to all images when multiple subsets of regularization images are specified in the dataset settings. The settings for each subset are correctly applied to each image. PR #1205 Thanks to feffy380!
- Some features are added to the dataset subset settings.
secondary_separator
is added to specify the tag separator that is not the target of shuffling or dropping.- Specify
secondary_separator=";;;"
. When you specifysecondary_separator
, the part is not shuffled or dropped.
- Specify
enable_wildcard
is added. When set totrue
, the wildcard notation{aaa|bbb|ccc}
can be used. The multi-line caption is also enabled.keep_tokens_separator
is updated to be used twice in the caption. When you specifykeep_tokens_separator="|||"
, the part divided by the second|||
is not shuffled or dropped and remains at the end.- The existing features
caption_prefix
andcaption_suffix
can be used together.caption_prefix
andcaption_suffix
are processed first, and thenenable_wildcard
,keep_tokens_separator
, shuffling and dropping, andsecondary_separator
are processed in order. - See Dataset config for details.
- The dataset with DreamBooth method supports caching image information (size, caption). PR #1178 and #1206 Thanks to KohakuBlueleaf! See DreamBooth method specific options for details.
-
Image tagging (not implemented yet in the GUI)
- The support for v3 repositories is added to
tag_image_by_wd14_tagger.py
(--onnx
option only). PR #1192 Thanks to sdbds!- Onnx may need to be updated. Onnx is not installed by default, so please install or update it with
pip install onnx==1.15.0 onnxruntime-gpu==1.17.1
etc. Please also check the comments inrequirements.txt
.
- Onnx may need to be updated. Onnx is not installed by default, so please install or update it with
- The model is now saved in the subdirectory as
--repo_id
intag_image_by_wd14_tagger.py
. This caches multiple repo_id models. Please delete unnecessary files under--model_dir
. - Some options are added to
tag_image_by_wd14_tagger.py
.- Some are added in PR #1216 Thanks to Disty0!
- Output rating tags
--use_rating_tags
and--use_rating_tags_as_last_tag
- Output character tags first
--character_tags_first
- Expand character tags and series
--character_tag_expand
- Specify tags to output first
--always_first_tags
- Replace tags
--tag_replacement
- See Tagging documentation for details.
- Fixed an error when specifying
--beam_search
and a value of 2 or more for--num_beams
inmake_captions.py
.
- The support for v3 repositories is added to
-
About Masked loss The masked loss is supported in each training script. To enable the masked loss, specify the
--masked_loss
option.The feature is not fully tested, so there may be bugs. If you find any issues, please open an Issue.
ControlNet dataset is used to specify the mask. The mask images should be the RGB images. The pixel value 255 in R channel is treated as the mask (the loss is calculated only for the pixels with the mask), and 0 is treated as the non-mask. The pixel values 0-255 are converted to 0-1 (i.e., the pixel value 128 is treated as the half weight of the loss). See details for the dataset specification in the LLLite documentation.
-
About Scheduled Huber Loss Scheduled Huber Loss has been introduced to each training scripts. This is a method to improve robustness against outliers or anomalies (data corruption) in the training data.
With the traditional MSE (L2) loss function, the impact of outliers could be significant, potentially leading to a degradation in the quality of generated images. On the other hand, while the Huber loss function can suppress the influence of outliers, it tends to compromise the reproduction of fine details in images.
To address this, the proposed method employs a clever application of the Huber loss function. By scheduling the use of Huber loss in the early stages of training (when noise is high) and MSE in the later stages, it strikes a balance between outlier robustness and fine detail reproduction.
Experimental results have confirmed that this method achieves higher accuracy on data containing outliers compared to pure Huber loss or MSE. The increase in computational cost is minimal.
The newly added arguments loss_type, huber_schedule, and huber_c allow for the selection of the loss function type (Huber, smooth L1, MSE), scheduling method (exponential, constant, SNR), and Huber's parameter. This enables optimization based on the characteristics of the dataset.
See PR #1228 for details.
loss_type
: Specify the loss function type. Choosehuber
for Huber loss,smooth_l1
for smooth L1 loss, andl2
for MSE loss. The default isl2
, which is the same as before.huber_schedule
: Specify the scheduling method. Chooseexponential
,constant
, orsnr
. The default issnr
.huber_c
: Specify the Huber's parameter. The default is0.1
.
Please read Releases for recent updates.`
-
-
Added GUI support for the new parameters listed above.
-
Moved accelerate launch parameters to a new
Accelerate launch
accordion above theModel
accordion. -
Added support for
Debiased Estimation loss
to Dreambooth settings. -
Added support for "Dataset Preparation" defaults via the config.toml file.
-
Added a field to allow for the input of extra accelerate launch arguments.
-
Added new caption tool from https://github.com/kainatquaderee
- Add support for toml dataset configuration fole to all trainers
- Add new setup menu option to install Triton 2.1.0 for Windows
- Add support for LyCORIS BOFT and DoRA and QLyCORIS options for LoHA, LoKr and LoCon
- Fix issue with vae path validation
- Other fixes
- Fix blip caption issue
- Fix issue with image samples.
-
Enhanced Logging and Tracking Capabilities
- Added support for configuring advanced logging and tracking:
wandb_run_name
: Set a custom name for your Weights & Biases runs to easily identify and organize your experiments.log_tracker_name
andlog_tracker_config
: Integrate custom logging trackers with your projects. Specify the tracker name and provide its configuration to enable detailed monitoring and logging of your runs.
- Added support for configuring advanced logging and tracking:
-
Custom Path Defaults
- You can now specify custom paths more easily:
- Simply copy the
config example.toml
file located in the root directory of the repository toconfig.toml
. - Edit the
config.toml
file to adjust paths and settings according to your preferences.
- Simply copy the
- You can now specify custom paths more easily:
-
sd-scripts updated to v0.8.5
- Bug Fixes:
- Corrected an issue where the value of timestep embedding was incorrect during SDXL training. This fix ensures accurate training progress and results.
- Addressed a related inference issue with the generation script, improving the reliability of SDXL model outputs.
- Note: The exact impact of this bug is currently unknown, but it's recommended to update to v0.8.5 for anyone engaged in SDXL training to ensure optimal performance and results.
- Bug Fixes:
-
Upgrade of
lycoris_lora
Python Module- Updated the
lycoris_lora
module to version 2.2.0.post3. This update may include bug fixes, performance improvements, and new features.
- Updated the
- To benefit from the latest features and improvements, users are encouraged to update their installations and configurations accordingly.
- Increase icon size.
- More setup fixes.