From b35b053b8d7c75c64dec7dd7d94d9a9b8ef27e66 Mon Sep 17 00:00:00 2001 From: michaelgzhang <49577754+mgz-dev@users.noreply.github.com> Date: Sat, 11 Feb 2023 03:14:43 -0600 Subject: [PATCH] clean up print formatting --- networks/resize_lora.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/networks/resize_lora.py b/networks/resize_lora.py index 29f87e6d..e21bdabd 100644 --- a/networks/resize_lora.py +++ b/networks/resize_lora.py @@ -101,7 +101,7 @@ def resize_lora_model(lora_sd, new_rank, save_dtype, device, verbose): s_sum = torch.sum(torch.abs(S)) s_rank = torch.sum(torch.abs(S[:new_rank])) verbose_str+=f"{block_down_name:76} | " - verbose_str+=f"sum(S) retained: {(s_rank)/s_sum:.1%}%, max(S) to max(S_dropped) ratio: {S[0]/S[new_rank]:0.1f}\n" + verbose_str+=f"sum(S) retained: {(s_rank)/s_sum:.1%}, max(S) ratio: {S[0]/S[new_rank]:0.1f}\n" U = U[:, :new_rank] S = S[:new_rank]