From ef4397963bdfc7882addc12e0a4510868a4b1f33 Mon Sep 17 00:00:00 2001 From: rockerBOO Date: Mon, 8 Sep 2025 14:16:33 -0400 Subject: [PATCH 1/3] Fix validation dataset documentation to not use subsets --- docs/flux_train_network.md | 20 +++++++++++++++++--- 1 file changed, 17 insertions(+), 3 deletions(-) diff --git a/docs/flux_train_network.md b/docs/flux_train_network.md index 23828eb7..f5b67a7e 100644 --- a/docs/flux_train_network.md +++ b/docs/flux_train_network.md @@ -550,18 +550,32 @@ You can calculate validation loss during training using a validation dataset to To set up validation, add a `validation_split` and optionally `validation_seed` to your dataset configuration TOML file. ```toml +validation_seed = 42 # [Optional] Validation seed, otherwise uses training seed for validation split . + [[datasets]] enable_bucket = true resolution = [1024, 1024] -validation_seed = 42 # [Optional] Validation seed, otherwise uses training seed for validation split . [[datasets.subsets]] image_dir = "path/to/image/directory" - validation_split = 0.1 # Split between 0.0 and 1.0 where 1.0 will use the full subset as a validation dataset + +[[datasets]] +enable_bucket = true +resolution = [1024, 1024] +validation_split = 0.1 # Split between 0.0 and 1.0 where 1.0 will use the full subset as a validation dataset [[datasets.subsets]] + # This directory will split 10% to validation and 90% to training + image_dir = "path/to/image/second-directory" + +[[datasets]] +enable_bucket = true +resolution = [1024, 1024] +validation_split = 1.0 # Will use this full subset as a validation subset. + + [[datasets.subsets]] + # This directory will use the 100% to validation and 0% to training image_dir = "path/to/image/full_validation" - validation_split = 1.0 # Will use this full subset as a validation subset. ``` **Notes:** From 78685b9c5f2141c99a1478ff3f4d59c276828dd1 Mon Sep 17 00:00:00 2001 From: rockerBOO Date: Mon, 8 Sep 2025 14:18:50 -0400 Subject: [PATCH 2/3] Move general settings to top to make more clear the validation bits --- docs/flux_train_network.md | 8 ++------ 1 file changed, 2 insertions(+), 6 deletions(-) diff --git a/docs/flux_train_network.md b/docs/flux_train_network.md index f5b67a7e..ccf6dff7 100644 --- a/docs/flux_train_network.md +++ b/docs/flux_train_network.md @@ -551,17 +551,15 @@ To set up validation, add a `validation_split` and optionally `validation_seed` ```toml validation_seed = 42 # [Optional] Validation seed, otherwise uses training seed for validation split . - -[[datasets]] enable_bucket = true resolution = [1024, 1024] +[[datasets]] [[datasets.subsets]] + # This directory will use 100% of the images for training image_dir = "path/to/image/directory" [[datasets]] -enable_bucket = true -resolution = [1024, 1024] validation_split = 0.1 # Split between 0.0 and 1.0 where 1.0 will use the full subset as a validation dataset [[datasets.subsets]] @@ -569,8 +567,6 @@ validation_split = 0.1 # Split between 0.0 and 1.0 where 1.0 will use the full s image_dir = "path/to/image/second-directory" [[datasets]] -enable_bucket = true -resolution = [1024, 1024] validation_split = 1.0 # Will use this full subset as a validation subset. [[datasets.subsets]] From fe4c18934c2d34ff2eb3eb65ea1eaa8ecec207cd Mon Sep 17 00:00:00 2001 From: rockerBOO Date: Mon, 8 Sep 2025 14:28:55 -0400 Subject: [PATCH 3/3] blocks_to_swap is supported for validation loss now --- docs/flux_train_network.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/docs/flux_train_network.md b/docs/flux_train_network.md index ccf6dff7..3d61f830 100644 --- a/docs/flux_train_network.md +++ b/docs/flux_train_network.md @@ -577,7 +577,7 @@ validation_split = 1.0 # Will use this full subset as a validation subset. **Notes:** * Validation loss calculation uses fixed timestep sampling and random seeds to reduce loss variation due to randomness for more stable evaluation. -* Currently, validation loss is not supported when using `--blocks_to_swap` or Schedule-Free optimizers (`AdamWScheduleFree`, `RAdamScheduleFree`, `ProdigyScheduleFree`). +* Currently, validation loss is not supported when using Schedule-Free optimizers (`AdamWScheduleFree`, `RAdamScheduleFree`, `ProdigyScheduleFree`).
日本語