Kohya S
|
59267d19f3
|
feat: optimize RMSNorm forward method and remove unused torch_attention_op
|
2026-02-11 22:06:47 +09:00 |
|
Kohya S
|
9349c91c89
|
fix: remove unused code
|
2026-02-11 17:23:58 +09:00 |
|
Kohya S
|
6d08c93b23
|
feat: enhance block swap functionality for inference and training in Anima model
|
2026-02-10 21:26:50 +09:00 |
|
Kohya S
|
0f413974b7
|
feat: support LoRA format without net. prefix
|
2026-02-09 23:21:04 +09:00 |
|
kohya-ss
|
2774e7757b
|
feat: add dtype property and all-zero mask handling in cross-attention in LLMAdapterTransformerBlock
|
2026-02-09 12:43:11 +09:00 |
|
kohya-ss
|
d992037984
|
format: format
|
2026-02-08 12:22:54 +09:00 |
|
kohya-ss
|
44b8d79577
|
feat: use unified attention module, add wrapper for state dict compatibility
|
2026-02-08 12:16:00 +09:00 |
|
duongve13112002
|
e21a7736f8
|
Support Anima model (#2260)
* Support Anima model
* Update document and fix bug
* Fix latent normlization
* Fix typo
* Fix cache embedding
* fix typo in tests/test_anima_cache.py
* Remove redundant argument apply_t5_attn_mask
* Improving caching with argument caption_dropout_rate
* Fix W&B logging bugs
* Fix discrete_flow_shift default value
|
2026-02-08 10:18:55 +09:00 |
|