rockerBOO
24ab4c0c4a
Support loading more checkpoint types
2025-03-26 16:35:04 -04:00
rockerBOO
c0f2808763
Support more checkpoint files for flux
2025-03-26 16:34:47 -04:00
Kohya S
6364379f17
Merge branch 'dev' into sd3
2025-03-21 22:07:50 +09:00
Kohya S
5253a38783
Merge branch 'main' into dev
2025-03-21 22:07:03 +09:00
Kohya S
8f4ee8fc34
doc: update README for latest
v0.9.1
2025-03-21 22:05:48 +09:00
Kohya S.
367f348430
Merge pull request #1964 from Nekotekina/main
...
Fix missing text encoder attn modules
2025-03-21 21:59:03 +09:00
Kohya S
d151833526
docs: update README with recent changes and specify version for pytorch-optimizer
2025-03-20 22:05:29 +09:00
Kohya S.
936d333ff4
Merge pull request #1985 from gesen2egee/pytorch-optimizer
...
Support pytorch_optimizer
2025-03-20 22:01:03 +09:00
gesen2egee
5b210ad717
update prodigyopt and prodigy-plus-schedule-free
2025-03-19 10:49:06 +08:00
gesen2egee
e5b5c7e1db
Update requirements.txt
2025-03-15 13:29:32 +08:00
Kohya S.
75933d70a1
Merge pull request #1960 from kohya-ss/sd3_safetensors_merge
...
Sd3 safetensors merge
2025-03-05 23:28:38 +09:00
Kohya S
aa2bde7ece
docs: add utility script for merging SD3 weights into a single .safetensors file
2025-03-05 23:24:52 +09:00
Ivan Chikish
acdca2abb7
Fix [occasionally] missing text encoder attn modules
...
Should fix #1952
I added alternative name for CLIPAttention.
I have no idea why this name changed.
Now it should accept both names.
2025-03-01 20:35:45 +03:00
Kohya S
ba5251168a
fix: save tensors as is dtype, add save_precision option
2025-03-01 10:31:39 +09:00
Kohya S
272f4c3775
Merge branch 'sd3' into sd3_safetensors_merge
2025-02-28 23:52:36 +09:00
Kohya S
734333d0c9
feat: enhance merging logic for safetensors models to handle key prefixes correctly
2025-02-28 23:52:29 +09:00
Kohya S
3d79239be4
docs: update README to include recent improvements in validation loss calculation
2025-02-26 21:21:04 +09:00
Kohya S
ec350c83eb
Merge branch 'dev' into sd3
2025-02-26 21:17:29 +09:00
Kohya S.
49651892ce
Merge pull request #1903 from kohya-ss/val-loss-improvement
...
Val loss improvement
2025-02-26 21:15:14 +09:00
Kohya S
1fcac98280
Merge branch 'sd3' into val-loss-improvement
2025-02-26 21:09:10 +09:00
Kohya S.
b286304e5f
Merge pull request #1953 from Disty0/dev
...
Update IPEX libs
2025-02-26 21:03:09 +09:00
Kohya S
ae409e83c9
fix: FLUX/SD3 network training not working without caching latents closes #1954
2025-02-26 20:56:32 +09:00
Kohya S
5228db1548
feat: add script to merge multiple safetensors files into a single file for SD3
2025-02-26 20:50:58 +09:00
Kohya S
f4a0047865
feat: support metadata loading in MemoryEfficientSafeOpen
2025-02-26 20:50:44 +09:00
Disty0
f68702f71c
Update IPEX libs
2025-02-25 21:27:41 +03:00
Kohya S.
6e90c0f86c
Merge pull request #1909 from rockerBOO/progress_bar
...
Move progress bar to account for sampling image first
2025-02-24 18:57:44 +09:00
Kohya S
67fde015f7
Merge branch 'dev' into sd3
2025-02-24 18:56:15 +09:00
Kohya S.
386b7332c6
Merge pull request #1918 from tsukimiya/fix_vperd_warning
...
Remove v-pred warning.
2025-02-24 18:55:25 +09:00
Kohya S
905f081798
Merge branch 'dev' into sd3
2025-02-24 18:54:28 +09:00
Kohya S.
59ae9ea20c
Merge pull request #1945 from yidiq7/dev
...
Remove position_ids for V2
2025-02-24 18:53:46 +09:00
Kohya S
efb2a128cd
fix wandb val logging
2025-02-21 22:07:35 +09:00
Yidi
13df47516d
Remove position_ids for V2
...
The postions_ids cause errors for the newer version of transformer.
This has already been fixed in convert_ldm_clip_checkpoint_v1() but
not in v2.
The new code applies the same fix to convert_ldm_clip_checkpoint_v2().
2025-02-20 04:49:51 -05:00
Kohya S
4a36996134
modify log step calculation
2025-02-18 22:05:08 +09:00
Kohya S
dc7d5fb459
Merge branch 'sd3' into val-loss-improvement
2025-02-18 21:34:30 +09:00
Kohya S.
894037f2c6
Merge pull request #1943 from rockerBOO/validation-split-test
...
Fix validation split and add test
2025-02-18 21:30:32 +09:00
rockerBOO
9436b41061
Fix validation split and add test
2025-02-17 14:28:41 -05:00
Kohya S.
7c22e12a39
Merge pull request #1938 from rockerBOO/validation-epoch-loss-recorder
...
Fix validation epoch loss to check epoch average
2025-02-17 21:45:04 +09:00
Kohya S.
6051fa8217
Merge pull request #1940 from rockerBOO/split-size-fix
...
Fix sizes for validation split
2025-02-17 21:42:10 +09:00
rockerBOO
f3a010978c
Clear sizes for validation reg images to be consistent
2025-02-16 22:28:34 -05:00
rockerBOO
3c7496ae3f
Fix sizes for validation split
2025-02-16 22:18:14 -05:00
rockerBOO
4671e23778
Fix validation epoch loss to check epoch average
2025-02-16 01:42:44 -05:00
Kohya S
63337d9fe4
Merge branch 'sd3' into val-loss-improvement
2025-02-15 21:41:07 +09:00
Kohya S.
ee295c7d9f
Merge pull request #1935 from rockerBOO/validation-epoch-fix
...
Fix validation epoch divergence
2025-02-15 21:39:28 +09:00
rockerBOO
ab88b431b0
Fix validation epoch divergence
2025-02-14 11:14:38 -05:00
Kohya S
76b761943b
fix: simplify validation step condition in NetworkTrainer
2025-02-11 21:53:57 +09:00
Kohya S
cd80752175
fix: remove unused parameter 'accelerator' from encode_images_to_latents method
2025-02-11 21:42:58 +09:00
Kohya S
177203818a
fix: unpause training progress bar after vaidation
2025-02-11 21:42:46 +09:00
Kohya S
344845b429
fix: validation with block swap
2025-02-09 21:25:40 +09:00
Kohya S
0911683717
set python random state
2025-02-09 20:53:49 +09:00
Kohya S
a24db1d532
fix: validation timestep generation fails on SD/SDXL training
2025-02-04 22:02:42 +09:00