DKnight54
9dcdd163fb
Update train_network.py
2025-03-28 19:54:56 +08:00
DKnight54
c6d986e868
Update train_network.py
2025-03-28 19:46:23 +08:00
DKnight54
7c94d386c1
Update train_network.py
2025-03-28 19:45:30 +08:00
DKnight54
4456065d29
Update train_network.py
2025-03-28 19:29:59 +08:00
DKnight54
666468857b
Update train_network.py
2025-03-28 17:48:47 +08:00
DKnight54
8d14893598
Update train_network.py
2025-02-07 01:28:26 +08:00
DKnight54
c1e387f9f9
Update train_network.py
2025-02-07 01:18:43 +08:00
DKnight54
f36b545480
Update train_network.py
2025-02-06 19:14:05 +08:00
DKnight54
e6433ec035
Update train_network.py
2025-02-03 16:37:44 +08:00
DKnight54
4c5ec694f7
Update train_network.py
2025-02-03 05:22:00 +08:00
DKnight54
e250e7aad0
Update train_network.py
2025-02-03 05:09:14 +08:00
DKnight54
639a60de8a
Update train_network.py
2025-02-03 04:05:55 +08:00
DKnight54
b3b856393c
Update train_network.py
2025-02-03 03:38:24 +08:00
DKnight54
83a171d3dc
Update train_network.py
2025-02-02 04:56:22 +08:00
DKnight54
aea7cada42
Update train_network.py
2025-02-02 04:54:36 +08:00
DKnight54
896ade6f92
Update train_network.py
2025-02-02 04:46:53 +08:00
DKnight54
6fc38477c8
Update train_network.py
2025-02-02 04:39:15 +08:00
DKnight54
13f764dbbe
Update train_network.py
2025-02-01 02:06:24 +08:00
DKnight54
2cdaa33147
Update train_network.py
2025-01-31 21:47:15 +08:00
DKnight54
6de0051eb2
Update train_network.py
2025-01-31 15:07:18 +08:00
DKnight54
f73db401fa
Update train_network.py
2025-01-31 03:12:06 +08:00
DKnight54
5559d1b711
Update train_network.py
2025-01-31 01:56:30 +08:00
DKnight54
39a375139a
adding example generation
2025-01-29 18:46:52 +08:00
catboxanon
e1b63c2249
Only add warning for deprecated scaling vpred loss function
2024-10-21 08:12:53 -04:00
catboxanon
8fc30f8205
Fix training for V-pred and ztSNR
...
1) Updates debiased estimation loss function for V-pred.
2) Prevents now-deprecated scaling of loss if ztSNR is enabled.
2024-10-21 07:34:33 -04:00
Kohya S
fe2aa32484
adjust min/max bucket reso divisible by reso steps #1632
2024-09-29 09:49:25 +09:00
Kohya S
4dbcef429b
update for corner cases
2024-06-04 21:26:55 +09:00
Kohaku-Blueleaf
3eb27ced52
Skip the final 1 step
2024-05-31 12:24:15 +08:00
Kohaku-Blueleaf
b2363f1021
Final implementation
2024-05-31 12:20:20 +08:00
Kohya S
da6fea3d97
simplify and update alpha mask to work with various cases
2024-05-19 21:26:18 +09:00
u-haru
db6752901f
画像のアルファチャンネルをlossのマスクとして使用するオプションを追加 ( #1223 )
...
* Add alpha_mask parameter and apply masked loss
* Fix type hint in trim_and_resize_if_required function
* Refactor code to use keyword arguments in train_util.py
* Fix alpha mask flipping logic
* Fix alpha mask initialization
* Fix alpha_mask transformation
* Cache alpha_mask
* Update alpha_masks to be on CPU
* Set flipped_alpha_masks to Null if option disabled
* Check if alpha_mask is None
* Set alpha_mask to None if option disabled
* Add description of alpha_mask option to docs
2024-05-19 19:07:25 +09:00
Kohya S
c68baae480
add --log_config option to enable/disable output training config
2024-05-19 17:21:04 +09:00
Kohya S
47187f7079
Merge pull request #1285 from ccharest93/main
...
Hyperparameter tracking
2024-05-19 16:31:33 +09:00
Kohya S
52e64c69cf
add debug log
2024-05-04 18:43:52 +09:00
Kohya S
58c2d856ae
support block dim/lr for sdxl
2024-05-03 22:18:20 +09:00
Kohya S
969f82ab47
move loraplus args from args to network_args, simplify log lr desc
2024-04-29 20:04:25 +09:00
Kohya S
834445a1d6
Merge pull request #1233 from rockerBOO/lora-plus
...
Add LoRA+ support
2024-04-29 18:05:12 +09:00
Kohya S
0540c33aca
pop weights if available #1247
2024-04-21 17:45:29 +09:00
Kohya S
52652cba1a
disable main process check for deepspeed #1247
2024-04-21 17:41:32 +09:00
Maatra
2c9db5d9f2
passing filtered hyperparameters to accelerate
2024-04-20 14:11:43 +01:00
rockerBOO
75833e84a1
Fix default LR, Add overall LoRA+ ratio, Add log
...
`--loraplus_ratio` added for both TE and UNet
Add log for lora+
2024-04-08 19:23:02 -04:00
Kohya S
d30ebb205c
update readme, add metadata for network module
2024-04-07 14:58:17 +09:00
kabachuha
90b18795fc
Add option to use Scheduled Huber Loss in all training pipelines to improve resilience to data corruption ( #1228 )
...
* add huber loss and huber_c compute to train_util
* add reduction modes
* add huber_c retrieval from timestep getter
* move get timesteps and huber to own function
* add conditional loss to all training scripts
* add cond loss to train network
* add (scheduled) huber_loss to args
* fixup twice timesteps getting
* PHL-schedule should depend on noise scheduler's num timesteps
* *2 multiplier to huber loss cause of 1/2 a^2 conv.
The Taylor expansion of sqrt near zero gives 1/2 a^2, which differs from a^2 of the standard MSE loss. This change scales them better against one another
* add option for smooth l1 (huber / delta)
* unify huber scheduling
* add snr huber scheduler
---------
Co-authored-by: Kohya S <52813779+kohya-ss@users.noreply.github.com >
2024-04-07 13:54:21 +09:00
ykume
cd587ce62c
verify command line args if wandb is enabled
2024-04-05 08:23:03 +09:00
rockerBOO
f99fe281cb
Add LoRA+ support
2024-04-01 15:38:26 -04:00
Kohya S
2258a1b753
add save/load hook to remove U-Net/TEs from state
2024-03-31 15:50:35 +09:00
Kohya S
c86e356013
Merge branch 'dev' into dataset-cache
2024-03-26 19:43:40 +09:00
Kohya S
ab1e389347
Merge branch 'dev' into masked-loss
2024-03-26 19:39:30 +09:00
Kohya S
a2b8531627
make each script consistent, fix to work w/o DeepSpeed
2024-03-25 22:28:46 +09:00
Kohya S
025347214d
refactor metadata caching for DreamBooth dataset
2024-03-24 18:09:32 +09:00