Kohya S
4a2cef887c
fix lllite training not working ref #913
2023-12-10 09:23:37 +09:00
Kohya S
42750f7846
fix error on pool_workaround in sdxl TE training ref #994
2023-12-10 09:18:33 +09:00
Kohya S
e8c3a02830
Merge branch 'dev' into gradual_latent_hires_fix
2023-12-08 08:23:53 +09:00
Isotr0py
db84530074
Fix gradients synchronization for multi-GPUs training ( #989 )
...
* delete DDP wrapper
* fix train_db vae and train_network
* fix train_db vae and train_network unwrap
* network grad sync
---------
Co-authored-by: Kohya S <52813779+kohya-ss@users.noreply.github.com >
2023-12-07 22:01:42 +09:00
Kohya S
72bbaac96d
Merge pull request #985 from Disty0/dev
...
Update IPEX hijacks
2023-12-07 21:39:24 +09:00
Kohya S
5713d63dc5
add temporary workaround for playground-v2
2023-12-06 23:08:02 +09:00
Disty0
dd7bb33ab6
IPEX fix torch.UntypedStorage.is_cuda
2023-12-05 22:18:47 +03:00
Disty0
a9c6182b3f
Cleanup IPEX libs
2023-12-05 19:52:31 +03:00
Disty0
3d70137d31
Disable IPEX attention if the GPU supports 64 bit
2023-12-05 19:40:16 +03:00
Disty0
bce9a081db
Update IPEX hijacks
2023-12-05 14:17:31 +03:00
Kohya S
46cf41cc93
Merge pull request #961 from rockerBOO/attention-processor
...
Add attention processor
2023-12-03 21:24:12 +09:00
Kohya S
81a440c8e8
Merge pull request #955 from xzuyn/paged_adamw
...
Add PagedAdamW
2023-12-03 21:22:38 +09:00
Kohya S
f24a3b5282
show seed in generating samples
2023-12-03 21:15:30 +09:00
Kohya S
383b4a2c3e
Merge pull request #907 from shirayu/add_option_sample_at_first
...
Add option --sample_at_first
2023-12-03 21:00:32 +09:00
Kohya S
df59822a27
Merge pull request #906 from shirayu/accept_scheduler_designation_in_training
...
Accept sampler designation in sampling of training
2023-12-03 20:46:16 +09:00
Kohya S
7a4e50705c
add target_x flag (not sure this impl is correct)
2023-12-03 17:59:41 +09:00
Kohya S
29b6fa6212
add unsharp mask
2023-11-28 22:33:22 +09:00
Furqanil Taqwa
1bdd83a85f
remove unnecessary debug print
2023-11-28 17:26:27 +07:00
Furqanil Taqwa
1624c239c2
added keep_tokens_separator to dynamically keep token for being shuffled
2023-11-28 17:23:55 +07:00
Furqanil Taqwa
4a913ce61e
initialize keep_tokens_separator to dataset config
2023-11-28 17:22:35 +07:00
Kohya S
764e333fa2
make slicing vae compatible with latest diffusers
2023-11-26 18:12:04 +09:00
Kohya S
c61e3bf4c9
make separate U-Net for inference
2023-11-26 18:11:30 +09:00
Kohya S
fc8649d80f
Merge pull request #934 from feffy380/fix-minsnr-vpred-zsnr
...
Fix min-snr-gamma for v-prediction and ZSNR.
2023-11-25 21:19:39 +09:00
Kohya S
6d6d86260b
add Deep Shrink
2023-11-23 19:40:48 +09:00
rockerBOO
c856ea4249
Add attention processor
2023-11-19 12:11:36 -05:00
Kohya S
f312522cef
Merge pull request #913 from KohakuBlueleaf/custom-seperator
...
Add custom seperator for shuffle caption
2023-11-19 21:32:01 +09:00
xzuyn
da5a144589
Add PagedAdamW
2023-11-18 07:47:27 -05:00
feffy380
6b3148fd3f
Fix min-snr-gamma for v-prediction and ZSNR.
...
This fixes min-snr for vpred+zsnr by dividing directly by SNR+1.
The old implementation did it in two steps: (min-snr/snr) * (snr/(snr+1)), which causes division by zero when combined with --zero_terminal_snr
2023-11-07 23:02:25 +01:00
rockerBOO
9c591bdb12
Remove unnecessary subset line from collate
2023-11-05 16:58:20 -05:00
rockerBOO
3de9e6c443
Add validation split of datasets
2023-11-05 12:37:44 -05:00
rockerBOO
5b19bda85c
Add validation loss
2023-11-05 12:35:46 -05:00
Kohya S
6231aa91e2
common lr logging, set default None to ddp_timeout
2023-11-05 19:09:17 +09:00
Kohaku-Blueleaf
489b728dbc
Fix typo again
2023-10-30 20:19:51 +08:00
Kohaku-Blueleaf
583e2b2d01
Fix typo
2023-10-30 20:02:04 +08:00
Kohaku-Blueleaf
5dc2a0d3fd
Add custom seperator
2023-10-30 19:55:30 +08:00
Yuta Hayashibe
fea810b437
Added --sample_at_first to generate sample images before training
2023-10-29 21:44:57 +09:00
Yuta Hayashibe
40d917b0fe
Removed incorrect comments
2023-10-29 21:02:44 +09:00
Yuta Hayashibe
cf876fcdb4
Accept --ss to set sample_sampler dynamically
2023-10-29 20:15:04 +09:00
Yuta Hayashibe
291c29caaf
Added a function line_to_prompt_dict() and removed duplicated initializations
2023-10-29 19:57:25 +09:00
Yuta Hayashibe
01e00ac1b0
Make a function get_my_scheduler()
2023-10-29 19:46:02 +09:00
Kohya S
a9ed4ed8a8
Merge pull request #900 from xzuyn/paged_adamw_32bit
...
Add PagedAdamW32bit
2023-10-29 15:01:55 +09:00
Kohya S
9d6a5a0c79
Merge pull request #899 from shirayu/use_moving_average
...
Show moving average loss in the progress bar
2023-10-29 14:37:58 +09:00
Kohya S
fb97a7aab1
Merge pull request #898 from shirayu/update_repare_buckets_latents
...
Fix a typo and add assertions in making buckets
2023-10-29 14:29:53 +09:00
Kohaku-Blueleaf
1cefb2a753
Better implementation for te autocast ( #895 )
...
* Better implementation for te
* Fix some misunderstanding
* as same as unet, add explicit convert
* Better cache TE and TE lr
* Fix with list
* Add timeout settings
* Fix arg style
2023-10-28 15:49:59 +09:00
xzuyn
d8f68674fb
Update train_util.py
2023-10-27 07:05:53 -04:00
Yuta Hayashibe
0d21925bdf
Use @property
2023-10-27 18:14:27 +09:00
Yuta Hayashibe
3d2bb1a8f1
Add LossRecorder and use moving average in all places
2023-10-27 17:49:49 +09:00
Yuta Hayashibe
837a4dddb8
Added assertions
2023-10-26 13:34:36 +09:00
青龍聖者@bdsqlsz
202f2c3292
Debias Estimation loss ( #889 )
...
* update for bnb 0.41.1
* fixed generate_controlnet_subsets_config for training
* Revert "update for bnb 0.41.1"
This reverts commit 70bd3612d8 .
* add debiased_estimation_loss
* add train_network
* Revert "add train_network"
This reverts commit 6539363c5c .
* Update train_network.py
2023-10-23 22:59:14 +09:00
Isotr0py
592014923f
Support JPEG-XL on windows
2023-10-04 21:48:25 +08:00