Commit Graph

991 Commits

Author SHA1 Message Date
Kohya S
326c425a5b feat: enhance LoRA weight handling in model loading and add text encoder loading function 2026-02-12 22:24:48 +09:00
Kohya S
cc1f57bc70 fix: review with Copilot 2026-02-12 08:05:41 +09:00
Kohya S.
57aa70ea9e Update library/anima_train_utils.py
Remove unused import.

Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com>
2026-02-12 08:02:05 +09:00
Kohya S.
07df0e54a0 Update library/safetensors_utils.py
Fix duplicated new_key for concat_hook.

Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com>
2026-02-11 22:36:19 +09:00
Kohya S.
82f963650b Update library/strategy_anima.py
Use torch.all instead of all.

Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com>
2026-02-11 22:28:00 +09:00
Kohya S
59267d19f3 feat: optimize RMSNorm forward method and remove unused torch_attention_op 2026-02-11 22:06:47 +09:00
Kohya S
4b2283491e feat: add VAE chunking and caching options to reduce memory usage 2026-02-11 21:32:00 +09:00
Kohya S
a7cd38dcaf feat: split CFG processing in do_sample function to reduce memory usage 2026-02-11 18:00:10 +09:00
Kohya S
9349c91c89 fix: remove unused code 2026-02-11 17:23:58 +09:00
Kohya S
b67cc5a457 feat: update Anima SAI model spec metadata handling 2026-02-11 14:56:23 +09:00
Kohya S
dbb40ae4c0 fix: wait all block trasfer before siwtching offloader mode 2026-02-10 21:46:58 +09:00
Kohya S
8d2d286a13 feat: refactor Anima training script 2026-02-10 21:27:22 +09:00
Kohya S
6d08c93b23 feat: enhance block swap functionality for inference and training in Anima model 2026-02-10 21:26:50 +09:00
Kohya S
02a75944b3 fix: keep latents 4D except DiT call 2026-02-10 21:26:20 +09:00
Kohya S
58db77a488 fix: update to work fp8_scaled option 2026-02-09 23:31:33 +09:00
Kohya S
0f413974b7 feat: support LoRA format without net. prefix 2026-02-09 23:21:04 +09:00
kohya-ss
8f5b298906 feat: simplify Anima training script and update empty caption handling 2026-02-09 12:46:53 +09:00
kohya-ss
06dcb30016 fix: update argument names in anima_train_utils to align with other archtectures 2026-02-09 12:46:04 +09:00
kohya-ss
bf9806648c feat: support 5d tensor in get_noisy_model_input_and_timesteps 2026-02-09 12:43:37 +09:00
kohya-ss
2774e7757b feat: add dtype property and all-zero mask handling in cross-attention in LLMAdapterTransformerBlock 2026-02-09 12:43:11 +09:00
kohya-ss
4f6511bf28 feat: kept caption dropout rate in cache and handle in training script 2026-02-08 15:35:53 +09:00
kohya-ss
d992037984 format: format 2026-02-08 12:22:54 +09:00
kohya-ss
5c4643516c feat: loading with dynamic fp8 optimization and LoRA support 2026-02-08 12:17:02 +09:00
kohya-ss
44b8d79577 feat: use unified attention module, add wrapper for state dict compatibility 2026-02-08 12:16:00 +09:00
kohya-ss
10445ff660 feat: simplify encode_tokens 2026-02-08 12:06:12 +09:00
kohya-ss
7b0ed3269a feat: add Qwen-Image VAE 2026-02-08 11:13:09 +09:00
duongve13112002
e21a7736f8 Support Anima model (#2260)
* Support Anima model

* Update document and fix bug

* Fix latent normlization

* Fix typo

* Fix cache embedding

* fix typo in tests/test_anima_cache.py

* Remove redundant argument apply_t5_attn_mask

* Improving caching with argument caption_dropout_rate

* Fix W&B logging bugs

* Fix discrete_flow_shift default value
2026-02-08 10:18:55 +09:00
Kohya S.
c6bc632ec6 fix: metadata dataset degradation and make it work (#2186)
* fix: support dataset with metadata

* feat: support another tagger model

* fix: improve handling of image size and caption/tag processing in FineTuningDataset

* fix: enhance metadata loading to support JSONL format in FineTuningDataset

* feat: enhance image loading and processing in ImageLoadingPrepDataset with batch support and output options

* fix: improve image path handling and memory management in dataset classes

* Update finetune/tag_images_by_wd14_tagger.py

Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com>

* fix: add return type annotation for process_tag_replacement function and ensure tags are returned

* feat: add artist category threshold for tagging

* doc: add comment for clarification

---------

Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com>
2026-01-18 15:17:07 +09:00
urlesistiana
f7fc7ddda2 fix #2201: lumina 2 timesteps handling 2025-10-13 16:08:28 +08:00
Kohya S
5462a6bb24 Merge branch 'dev' into sd3 2025-09-29 21:02:02 +09:00
Kohya S
63711390a0 Merge branch 'main' into dev 2025-09-29 20:56:07 +09:00
Kohya S
60bfa97b19 fix: disable_mmap_safetensors not defined in SDXL TI training 2025-09-29 20:52:48 +09:00
Kohya S.
e7b89826c5 Update library/custom_offloading_utils.py
Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com>
2025-09-21 13:29:58 +09:00
Kohya S
806d535ef1 fix: block-wise scaling is overwritten by per-tensor scaling 2025-09-21 13:10:41 +09:00
Kohya S
3876343fad fix: remove print statement for guidance rescale in AdaptiveProjectedGuidance 2025-09-21 13:09:38 +09:00
Kohya S
040d976597 feat: add guidance rescale options for Adaptive Projected Guidance in inference 2025-09-21 13:03:14 +09:00
Kohya S
9621d9d637 feat: add Adaptive Projected Guidance parameters and noise rescaling 2025-09-21 12:34:40 +09:00
Kohya S
f41e9e2b58 feat: add vae_chunk_size argument for memory-efficient VAE decoding and processing 2025-09-21 11:09:37 +09:00
Kohya S
b090d15f7d feat: add multi backend attention and related update for HI2.1 models and scripts 2025-09-20 19:45:33 +09:00
Kohya S
f834b2e0d4 fix: --fp8_vl to work 2025-09-18 23:46:18 +09:00
Kohya S
f6b4bdc83f feat: block-wise fp8 quantization 2025-09-18 21:20:54 +09:00
Kohya S
f5b004009e fix: correct tensor indexing in HunyuanVAE2D class for blending and encoding functions 2025-09-17 21:54:25 +09:00
Kohya S
4e2a80a6ca refactor: update imports to use safetensors_utils for memory-efficient operations 2025-09-13 21:07:11 +09:00
Kohya S
d831c88832 fix: sample generation doesn't work with block swap 2025-09-13 21:06:04 +09:00
Kohya S
bae7fa74eb Merge branch 'sd3' into feat-hunyuan-image-2.1-inference 2025-09-13 20:13:58 +09:00
Kohya S.
e1c666e97f Update library/safetensors_utils.py
Co-authored-by: Copilot <175728472+Copilot@users.noreply.github.com>
2025-09-13 20:03:55 +09:00
Kohya S
8783f8aed3 feat: faster safetensors load and split safetensor utils 2025-09-13 19:51:38 +09:00
Kohya S
209c02dbb6 feat: HunyuanImage LoRA training 2025-09-12 21:40:42 +09:00
Kohya S
a0f0afbb46 fix: revert constructor signature update 2025-09-11 22:27:00 +09:00
Kohya S
7f983c558d feat: block swap for inference and initial impl for HunyuanImage LoRA (not working) 2025-09-11 22:15:22 +09:00