Disty0
a6a2b5a867
Fix IPEX support and add XPU device to device_utils
2024-01-31 17:32:37 +03:00
Aarni Koskela
afc38707d5
Refactor memory cleaning into a single function
2024-01-23 14:28:50 +02:00
Aarni Koskela
6f3f701d3d
Deduplicate ipex initialization code
2024-01-19 18:07:36 +02:00
Kohya S
32b759a328
Add wandb_run_name parameter to init_kwargs #1032
2024-01-14 22:02:03 +09:00
Kohya S
04ef8d395f
speed up nan replace in sdxl training ref #1009
2023-12-21 21:44:03 +09:00
Isotr0py
bb5ae389f7
fix DDP SDXL training
2023-12-12 19:58:44 +08:00
Kohya S
42750f7846
fix error on pool_workaround in sdxl TE training ref #994
2023-12-10 09:18:33 +09:00
Kohya S
912dca8f65
fix duplicated sample gen for every epoch ref #907
2023-12-07 22:13:38 +09:00
Isotr0py
db84530074
Fix gradients synchronization for multi-GPUs training ( #989 )
...
* delete DDP wrapper
* fix train_db vae and train_network
* fix train_db vae and train_network unwrap
* network grad sync
---------
Co-authored-by: Kohya S <52813779+kohya-ss@users.noreply.github.com >
2023-12-07 22:01:42 +09:00
Kohya S
383b4a2c3e
Merge pull request #907 from shirayu/add_option_sample_at_first
...
Add option --sample_at_first
2023-12-03 21:00:32 +09:00
Kohya S
6231aa91e2
common lr logging, set default None to ddp_timeout
2023-11-05 19:09:17 +09:00
Yuta Hayashibe
fea810b437
Added --sample_at_first to generate sample images before training
2023-10-29 21:44:57 +09:00
Kohya S
01d929ee2a
support separate learning rates for TE1/2
2023-10-29 20:38:01 +09:00
Kohya S
9d6a5a0c79
Merge pull request #899 from shirayu/use_moving_average
...
Show moving average loss in the progress bar
2023-10-29 14:37:58 +09:00
Kohaku-Blueleaf
1cefb2a753
Better implementation for te autocast ( #895 )
...
* Better implementation for te
* Fix some misunderstanding
* as same as unet, add explicit convert
* Better cache TE and TE lr
* Fix with list
* Add timeout settings
* Fix arg style
2023-10-28 15:49:59 +09:00
Yuta Hayashibe
63992b81c8
Fix initialize place of loss_recorder
2023-10-27 21:13:29 +09:00
Yuta Hayashibe
0d21925bdf
Use @property
2023-10-27 18:14:27 +09:00
Yuta Hayashibe
efef5c8ead
Show "avr_loss" instead of "loss" because it is moving average
2023-10-27 17:59:58 +09:00
Yuta Hayashibe
3d2bb1a8f1
Add LossRecorder and use moving average in all places
2023-10-27 17:49:49 +09:00
青龍聖者@bdsqlsz
202f2c3292
Debias Estimation loss ( #889 )
...
* update for bnb 0.41.1
* fixed generate_controlnet_subsets_config for training
* Revert "update for bnb 0.41.1"
This reverts commit 70bd3612d8 .
* add debiased_estimation_loss
* add train_network
* Revert "add train_network"
This reverts commit 6539363c5c .
* Update train_network.py
2023-10-23 22:59:14 +09:00
Yuta Hayashibe
27f9b6ffeb
updated typos to v1.16.15 and fix typos
2023-10-01 21:51:24 +09:00
Disty0
b64389c8a9
Intel ARC support with IPEX
2023-09-19 18:05:05 +03:00
Kohya S
e2c2689f5c
support block lr for U-Net
2023-08-12 13:13:59 +09:00
Kohya S
6889ee2b85
add warning for bucket_reso_steps with SDXL
2023-08-11 19:02:36 +09:00
Kohya S
92f41f1051
update sdxl ver in lora metadata from v0-9 to v1-0
2023-08-06 22:06:48 +09:00
Kohya S
0636399c8c
add adding v-pred like loss for noise pred
2023-07-31 08:23:28 +09:00
Kohya S
b1e44e96bc
fix to show batch size for each dataset refs #637
2023-07-23 15:39:56 +09:00
Kohya S
73a08c0be0
Merge pull request #630 from ddPn08/sdxl
...
make tracker init_kwargs configurable
2023-07-20 22:05:55 +09:00
Kohya S
acf16c063a
make to work with PyTorch 1.12
2023-07-20 21:41:16 +09:00
Kohya S
225e871819
enable full bf16 trainint in train_network
2023-07-19 08:41:42 +09:00
Kohya S
6d2d8dfd2f
add zero_terminal_snr option
2023-07-18 23:17:23 +09:00
Kohya S
516f64f4d9
add caching to disk for text encoder outputs
2023-07-16 14:53:47 +09:00
Kohya S
8fa5fb2816
support diffusers format for SDXL
2023-07-12 21:57:14 +09:00
Kohya S
3c67e595b8
fix gradient accumulation doesn't work
2023-07-12 21:35:57 +09:00
ddPn08
b841dd78fe
make tracker init_kwargs configurable
2023-07-11 10:21:45 +09:00
Kohya S
0416f26a76
support multi gpu in caching text encoder outputs
2023-07-09 16:02:56 +09:00
Kohaku-Blueleaf
5f348579d1
Update sdxl_train.py
2023-07-09 12:46:35 +08:00
Kohya S
3d0375daa6
fix to work sdxl state dict without logit_scale
2023-07-05 21:45:30 +09:00
Kohya S
ea182461d3
add min/max_timestep
2023-07-03 20:44:42 +09:00
Kohya S
64cf922841
add feature to sample images during sdxl training
2023-07-02 16:42:19 +09:00
Kohya S
227a62e4c4
fix to work with dreambooth ds without toml
2023-06-30 07:40:22 +09:00
Kohya S
d395bc0647
fix max_token_length not works for sdxl
2023-06-29 13:02:19 +09:00
Kohya S
a751dc25d6
use CLIPTextModelWithProjection
2023-06-27 20:48:06 +09:00
Kohya S
2c461e4ad3
Add no_half_vae for SDXL training, add nan check
2023-06-26 20:38:09 +09:00
Kohya S
747af145ed
add sdxl fine-tuning and LoRA
2023-06-26 08:07:24 +09:00