Kohya S
6731d8a57f
fix: update system prompt handling
2025-06-29 22:21:48 +09:00
Kohya S
935e0037dc
feat: update lumina system prompt handling
2025-06-29 21:33:09 +09:00
rockerBOO
61f7283167
Fix non-cache vae encode
2025-03-21 20:38:43 -04:00
青龍聖者@bdsqlsz
09c4710d1e
Merge pull request #22 from rockerBOO/sage_attn
...
Add Sage Attention for Lumina
2025-03-03 10:26:02 +08:00
青龍聖者@bdsqlsz
b6e4194ea5
Merge pull request #20 from rockerBOO/lumina-system-prompt-special-token
...
Lumina system prompt special token
2025-03-02 18:30:49 +08:00
rockerBOO
a69884a209
Add Sage Attention for Lumina
2025-03-01 20:37:45 -05:00
rockerBOO
d6f7e2e20c
Fix block swap for sample images
2025-02-28 14:08:27 -05:00
rockerBOO
9647f1e324
Fix validation block swap. Add custom offloading tests
2025-02-27 20:36:36 -05:00
rockerBOO
42fe22f5a2
Enable block swap for Lumina
2025-02-27 03:21:24 -05:00
rockerBOO
ce2610d29b
Change system prompt to inject Prompt Start special token
2025-02-27 02:47:04 -05:00
sdbds
ce37c08b9a
clean code and add finetune code
2025-02-26 11:20:03 +08:00
sdbds
fc772affbe
1、Implement cfg_trunc calculation directly using timesteps, without intermediate steps.
...
2、Deprecate and remove the guidance_scale parameter because it used in inference not train
3、Add inference command-line arguments --ct for cfg_trunc_ratio and --rc for renorm_cfg to control CFG truncation and renormalization during inference.
2025-02-24 14:10:24 +08:00
rockerBOO
025cca699b
Fix samples, LoRA training. Add system prompt, use_flash_attn
2025-02-23 01:29:18 -05:00
rockerBOO
98efbc3bb7
Add documentation to model, use SDPA attention, sample images
2025-02-18 00:58:53 -05:00
sdbds
aa36c48685
update for always use gemma2 mask
2025-02-17 19:00:18 +08:00
rockerBOO
60a76ebb72
Add caching gemma2, add gradient checkpointing, refactor lumina model code
2025-02-16 01:06:34 -05:00
rockerBOO
a00b06bc97
Lumina 2 and Gemma 2 model loading
2025-02-15 14:56:11 -05:00
sdbds
c0caf33e3f
update
2025-02-15 16:38:59 +08:00
sdbds
d154e76c45
init
2025-02-12 16:30:05 +08:00