Commit Graph

8 Commits

Author SHA1 Message Date
Kohya S
59267d19f3 feat: optimize RMSNorm forward method and remove unused torch_attention_op 2026-02-11 22:06:47 +09:00
Kohya S
9349c91c89 fix: remove unused code 2026-02-11 17:23:58 +09:00
Kohya S
6d08c93b23 feat: enhance block swap functionality for inference and training in Anima model 2026-02-10 21:26:50 +09:00
Kohya S
0f413974b7 feat: support LoRA format without net. prefix 2026-02-09 23:21:04 +09:00
kohya-ss
2774e7757b feat: add dtype property and all-zero mask handling in cross-attention in LLMAdapterTransformerBlock 2026-02-09 12:43:11 +09:00
kohya-ss
d992037984 format: format 2026-02-08 12:22:54 +09:00
kohya-ss
44b8d79577 feat: use unified attention module, add wrapper for state dict compatibility 2026-02-08 12:16:00 +09:00
duongve13112002
e21a7736f8 Support Anima model (#2260)
* Support Anima model

* Update document and fix bug

* Fix latent normlization

* Fix typo

* Fix cache embedding

* fix typo in tests/test_anima_cache.py

* Remove redundant argument apply_t5_attn_mask

* Improving caching with argument caption_dropout_rate

* Fix W&B logging bugs

* Fix discrete_flow_shift default value
2026-02-08 10:18:55 +09:00