rockerBOO
|
de830b8941
|
Move progress bar to account for sampling image first
|
2025-01-29 00:02:45 -05:00 |
|
Kohya S
|
0778dd9b1d
|
fix Text Encoder only LoRA training
|
2025-01-27 22:03:42 +09:00 |
|
rockerBOO
|
c04e5dfe92
|
Fix loss recorder on 0. Fix validation for cached runs. Assert on validation dataset
|
2025-01-23 09:57:24 -05:00 |
|
rockerBOO
|
25929dd0d7
|
Remove Validating... print to fix output layout
|
2025-01-12 15:38:57 -05:00 |
|
rockerBOO
|
ee9265cf26
|
Fix validate_every_n_steps for gradient accumulation
|
2025-01-12 14:56:35 -05:00 |
|
rockerBOO
|
0456858992
|
Fix validate_every_n_steps always running first step
|
2025-01-12 14:47:49 -05:00 |
|
rockerBOO
|
2bbb40ce51
|
Fix regularization images with validation
Adding metadata recording for validation arguments
Add comments about the validation split for clarity of intention
|
2025-01-12 14:29:50 -05:00 |
|
rockerBOO
|
4c61adc996
|
Add divergence to logs
Divergence is the difference between training and validation to
allow a clear value to indicate the difference between the two
in the logs.
|
2025-01-12 13:18:26 -05:00 |
|
rockerBOO
|
1e61392cf2
|
Revert bucket_reso_steps to correct 64
|
2025-01-08 18:43:26 -05:00 |
|
rockerBOO
|
556f3f1696
|
Fix documentation, remove unused function, fix bucket reso for sd1.5, fix multiple datasets
|
2025-01-08 13:41:15 -05:00 |
|
rockerBOO
|
1231f5114c
|
Remove unused train_util code, fix accelerate.log for wandb, add init_trackers library code
|
2025-01-07 22:31:41 -05:00 |
|
rockerBOO
|
742bee9738
|
Set validation steps in multiple lines for readability
|
2025-01-06 17:34:23 -05:00 |
|
rockerBOO
|
fcb2ff010c
|
Clean up some validation help documentation
|
2025-01-06 11:39:32 -05:00 |
|
rockerBOO
|
f8850296c8
|
Fix validate epoch, cleanup imports
|
2025-01-06 11:34:10 -05:00 |
|
rockerBOO
|
c64d1a22fc
|
Add validate_every_n_epochs, change name validate_every_n_steps
|
2025-01-06 11:30:21 -05:00 |
|
rockerBOO
|
1c63e7cc49
|
Cleanup unused code and formatting
|
2025-01-06 11:07:47 -05:00 |
|
rockerBOO
|
bbf6bbd5ea
|
Use self.get_noise_pred_and_target and drop fixed timesteps
|
2025-01-06 10:48:38 -05:00 |
|
rockerBOO
|
1c0ae306e5
|
Add missing functions for training batch
|
2025-01-03 15:43:02 -05:00 |
|
rockerBOO
|
1f9ba40b8b
|
Add step break for validation epoch. Remove unused variable
|
2025-01-03 15:32:07 -05:00 |
|
rockerBOO
|
0522070d19
|
Fix training, validation split, revert to using upstream implemenation
|
2025-01-03 15:20:25 -05:00 |
|
rockerBOO
|
58bfa36d02
|
Add seed help clarifying info
|
2025-01-03 02:00:28 -05:00 |
|
rockerBOO
|
534059dea5
|
Typos and lingering is_train
|
2025-01-03 01:18:15 -05:00 |
|
rockerBOO
|
d23c7322ee
|
Merge remote-tracking branch 'hina/feature/val-loss' into validation-loss-upstream
Modified implementation for process_batch and cleanup validation
recording
|
2025-01-03 00:48:08 -05:00 |
|
rockerBOO
|
7f6e124c7c
|
Merge branch 'gesen2egee/val' into validation-loss-upstream
Modified various implementations to restore original behavior
|
2025-01-02 23:04:38 -05:00 |
|
gesen2egee
|
8743532963
|
val
|
2025-01-02 15:57:12 -05:00 |
|
Hina Chen
|
cb89e0284e
|
Change val latent loss compare
|
2024-12-28 11:57:04 +08:00 |
|
Hina Chen
|
64bd5317dc
|
Split val latents/batch and pick up val latents shape size which equal to training batch.
|
2024-12-28 11:42:15 +08:00 |
|
Hina Chen
|
62164e5792
|
Change val loss calculate method
|
2024-12-27 17:28:05 +08:00 |
|
Hina Chen
|
05bb9183fa
|
Add Validation loss for LoRA training
|
2024-12-27 16:47:59 +08:00 |
|
Kohya S.
|
14c9ba925f
|
Merge pull request #1811 from rockerBOO/schedule-free-prodigy
Allow unknown schedule-free optimizers to continue to module loader
|
2024-12-01 21:51:25 +09:00 |
|
Kohya S
|
cc11989755
|
fix: refactor huber-loss calculation in multiple training scripts
|
2024-12-01 21:20:28 +09:00 |
|
rockerBOO
|
6593cfbec1
|
Fix d * lr step log
|
2024-11-29 14:16:24 -05:00 |
|
rockerBOO
|
87f5224e2d
|
Support d*lr for ProdigyPlus optimizer
|
2024-11-29 14:16:00 -05:00 |
|
recris
|
420a180d93
|
Implement pseudo Huber loss for Flux and SD3
|
2024-11-27 18:37:09 +00:00 |
|
Kohya S
|
2cb7a6db02
|
feat: add block swap for FLUX.1/SD3 LoRA training
|
2024-11-12 21:39:13 +09:00 |
|
Kohya S
|
cde90b8903
|
feat: implement block swapping for FLUX.1 LoRA (WIP)
|
2024-11-12 08:49:05 +09:00 |
|
kohya-ss
|
1065dd1b56
|
Fix to work dropout_rate for TEs
|
2024-10-27 19:36:36 +09:00 |
|
kohya-ss
|
a1255d637f
|
Fix SD3 LoRA training to work (WIP)
|
2024-10-27 17:03:36 +09:00 |
|
Kohya S
|
db2b4d41b9
|
Add dropout rate arguments for CLIP-L, CLIP-G, and T5, fix Text Encoders LoRA not trained
|
2024-10-27 16:42:58 +09:00 |
|
kohya-ss
|
d2c549d7b2
|
support SD3 LoRA
|
2024-10-25 21:58:31 +09:00 |
|
Kohya S
|
5fba6f514a
|
Merge branch 'dev' into sd3
|
2024-10-25 19:03:27 +09:00 |
|
catboxanon
|
e1b63c2249
|
Only add warning for deprecated scaling vpred loss function
|
2024-10-21 08:12:53 -04:00 |
|
catboxanon
|
8fc30f8205
|
Fix training for V-pred and ztSNR
1) Updates debiased estimation loss function for V-pred.
2) Prevents now-deprecated scaling of loss if ztSNR is enabled.
|
2024-10-21 07:34:33 -04:00 |
|
Kohya S
|
3cc5b8db99
|
Diff Output Preserv loss for SDXL
|
2024-10-18 20:57:13 +09:00 |
|
kohya-ss
|
c80c304779
|
Refactor caching in train scripts
|
2024-10-12 20:18:41 +09:00 |
|
kohya-ss
|
ff4083b910
|
Merge branch 'sd3' into multi-gpu-caching
|
2024-10-12 16:39:36 +09:00 |
|
Kohya S
|
886f75345c
|
support weighted captions for sdxl LoRA and fine tuning
|
2024-10-10 08:27:15 +09:00 |
|
Kohya S
|
ba08a89894
|
call optimizer eval/train for sample_at_first, also set train after resuming closes #1667
|
2024-10-04 20:35:16 +09:00 |
|
gesen2egee
|
3028027e07
|
Update train_network.py
|
2024-10-04 16:41:41 +08:00 |
|
Kohya S
|
56a63f01ae
|
Merge branch 'sd3' into multi-gpu-caching
|
2024-09-29 10:12:18 +09:00 |
|