rockerBOO
7b83d50dc0
Merge branch 'sd3' into lumina
2025-02-26 22:13:56 -05:00
sdbds
fc772affbe
1、Implement cfg_trunc calculation directly using timesteps, without intermediate steps.
...
2、Deprecate and remove the guidance_scale parameter because it used in inference not train
3、Add inference command-line arguments --ct for cfg_trunc_ratio and --rc for renorm_cfg to control CFG truncation and renormalization during inference.
2025-02-24 14:10:24 +08:00
rockerBOO
42a801514c
Fix system prompt in datasets
2025-02-23 13:48:37 -05:00
rockerBOO
025cca699b
Fix samples, LoRA training. Add system prompt, use_flash_attn
2025-02-23 01:29:18 -05:00
Kohya S
efb2a128cd
fix wandb val logging
2025-02-21 22:07:35 +09:00
Kohya S
dc7d5fb459
Merge branch 'sd3' into val-loss-improvement
2025-02-18 21:34:30 +09:00
rockerBOO
9436b41061
Fix validation split and add test
2025-02-17 14:28:41 -05:00
rockerBOO
3ed7606f88
Clear sizes for validation reg images to be consistent
2025-02-17 12:07:23 -05:00
rockerBOO
3365cfadd7
Fix sizes for validation split
2025-02-17 12:07:23 -05:00
rockerBOO
f3a010978c
Clear sizes for validation reg images to be consistent
2025-02-16 22:28:34 -05:00
rockerBOO
3c7496ae3f
Fix sizes for validation split
2025-02-16 22:18:14 -05:00
sdbds
d154e76c45
init
2025-02-12 16:30:05 +08:00
Kohya S
a24db1d532
fix: validation timestep generation fails on SD/SDXL training
2025-02-04 22:02:42 +09:00
Kohya S
58b82a576e
Fix to work with validation dataset
2025-01-26 21:21:21 +09:00
rockerBOO
c04e5dfe92
Fix loss recorder on 0. Fix validation for cached runs. Assert on validation dataset
2025-01-23 09:57:24 -05:00
rockerBOO
b489082495
Disable repeats for validation datasets
2025-01-12 16:42:04 -05:00
rockerBOO
2bbb40ce51
Fix regularization images with validation
...
Adding metadata recording for validation arguments
Add comments about the validation split for clarity of intention
2025-01-12 14:29:50 -05:00
rockerBOO
556f3f1696
Fix documentation, remove unused function, fix bucket reso for sd1.5, fix multiple datasets
2025-01-08 13:41:15 -05:00
rockerBOO
1231f5114c
Remove unused train_util code, fix accelerate.log for wandb, add init_trackers library code
2025-01-07 22:31:41 -05:00
rockerBOO
695f38962c
Move get_huber_threshold_if_needed
2025-01-03 15:25:12 -05:00
rockerBOO
0522070d19
Fix training, validation split, revert to using upstream implemenation
2025-01-03 15:20:25 -05:00
rockerBOO
6604b36044
Remove duplicate assignment
2025-01-03 02:04:59 -05:00
rockerBOO
fbfc2753eb
Update text for train/reg with repeats
2025-01-03 01:53:12 -05:00
rockerBOO
c8c3569df2
Cleanup order, types, print to logger
2025-01-03 01:26:45 -05:00
rockerBOO
534059dea5
Typos and lingering is_train
2025-01-03 01:18:15 -05:00
rockerBOO
d23c7322ee
Merge remote-tracking branch 'hina/feature/val-loss' into validation-loss-upstream
...
Modified implementation for process_batch and cleanup validation
recording
2025-01-03 00:48:08 -05:00
rockerBOO
7f6e124c7c
Merge branch 'gesen2egee/val' into validation-loss-upstream
...
Modified various implementations to restore original behavior
2025-01-02 23:04:38 -05:00
rockerBOO
449c1c5c50
Adding modified train_util and config_util
2025-01-02 15:59:20 -05:00
gesen2egee
8743532963
val
2025-01-02 15:57:12 -05:00
Hina Chen
05bb9183fa
Add Validation loss for LoRA training
2024-12-27 16:47:59 +08:00
nhamanasu
8e378cf03d
add RAdamScheduleFree support
2024-12-11 19:43:44 +09:00
Kohya S.
e3fd6c52a0
Merge pull request #1812 from rockerBOO/tests
...
Add pytest testing
2024-12-02 21:38:43 +09:00
Kohya S
1dc873d9b4
update README and clean up code for schedulefree optimizer
2024-12-01 22:00:44 +09:00
Kohya S.
14c9ba925f
Merge pull request #1811 from rockerBOO/schedule-free-prodigy
...
Allow unknown schedule-free optimizers to continue to module loader
2024-12-01 21:51:25 +09:00
Kohya S
1476040787
fix: update help text for huber loss parameters in train_util.py
2024-12-01 21:26:39 +09:00
Kohya S
cc11989755
fix: refactor huber-loss calculation in multiple training scripts
2024-12-01 21:20:28 +09:00
Kohya S
14f642f88b
fix: huber_schedule exponential not working on sd3_train.py
2024-12-01 13:30:35 +09:00
recris
7b61e9eb58
Fix issues found in review (pt 2)
2024-11-30 11:36:40 +00:00
rockerBOO
c7cadbc8c7
Add pytest testing
2024-11-29 15:52:03 -05:00
rockerBOO
928b9393da
Allow unknown schedule-free optimizers to continue to module loader
2024-11-29 14:12:34 -05:00
recris
740ec1d526
Fix issues found in review
2024-11-28 20:38:32 +00:00
recris
420a180d93
Implement pseudo Huber loss for Flux and SD3
2024-11-27 18:37:09 +00:00
Kohya S
2bb0f547d7
update grad hook creation to fix TE lr in sd3 fine tuning
2024-11-14 19:33:12 +09:00
Kohya S
2cb7a6db02
feat: add block swap for FLUX.1/SD3 LoRA training
2024-11-12 21:39:13 +09:00
Kohya S
3fe94b058a
update comment
2024-11-12 08:09:07 +09:00
sdbds
26bd4540a6
init
2024-11-11 09:25:28 +08:00
feffy380
b3248a8eef
fix: sort order when getting image size from cache file
2024-11-07 14:31:05 +01:00
Kohya S
9aa6f52ac3
Fix memory leak in latent caching. bmp failed to cache
2024-11-01 21:43:21 +09:00
Kohya S
1434d8506f
Support SD3.5M multi resolutional training
2024-10-31 19:58:22 +09:00
kohya-ss
d4f7849592
prevent unintended cast for disk cached TE outputs
2024-10-27 19:35:56 +09:00