Recommended Training Parameters:

1 Epochs and Repeats

1.1 Epochs:

Number of dataset image training cycles. We suggest 10 for beginners. The value can be raised if the training seems insufficient due to a small dataset or lowered if the dataset is huge.

1.2 Repeats:

Number of times an image is learned. Higher values lead to better effects and more complex image compositions. Setting it too high may increase the risk of overfitting. Therefore, we suggest using 10 to achieve good training results while minimizing the chance of overfitting. Note: You can increase the epochs and repeats if the training results do not resemble.

2 Learning Rate and Optimizer:

2.1 learning_rate (Overall Learning Rate):

Degree of change in each repeat. Higher values mean faster learning but may cause model crashes or inability to converge. Lower values mean slower learning but may achieve optimal state. This value becomes ineffective after setting separate learning rates for U-Net and Text Encoder.

2.2 unet_lr (U-Net Learning Rate):

U-Net guides noise images generated by random seeds to determine denoising direction, find areas needing change, and provide the required data. Higher values mean faster fitting but risk missing details, while lower values cause underfitting and no resemblance among generated images and materials. The value is set accordingly based on the model type and dataset. We suggest 0.0002 for character training.

2.3 text_encoder_lr (Text Encoder Learning Rate):

It converts tags to embedding form for U-Net to understand. Since the text encoder of SDXL is already well-trained, there is usually no need for further training, and default values are fine unless there are special needs.

2.4 Optimizer:

An algorithm in deep learning that adjusts model parameters to minimize the loss function. During neural network training, the optimizer updates the model's weight based on the gradient information of the loss function so the model can better fit the training data. The default optimizer, AdamW, can be used for SDXL training, and other optimizers, like the easy-to-use Prodigy with adaptive learning rates, can also be chosen based on specific requirements.

2.5 lr_scheduler (Learning Rate Scheduler Settings):

Refers to a strategy or algorithm for dynamically adjusting the learning rate during training. Choosing Constant is sufficient under normal circumstances.

3 Network Settings:

https://media.discordapp.net/attachments/1174196809497849966/1174197353037713429/image.png?ex=65826700&is=656ff200&hm=69a1ae84035fca2f96d5b4777a59bf566622a061af93cf0dad89c26a3766892b&=&format=webp&quality=lossless&width=550&height=260

3.1 network_dim (Network Dimension):