diff --git a/library/fp8_optimization_utils.py b/library/fp8_optimization_utils.py index 82ec6bfc..02f99ab6 100644 --- a/library/fp8_optimization_utils.py +++ b/library/fp8_optimization_utils.py @@ -220,10 +220,6 @@ def quantize_weight( tensor_max = torch.max(torch.abs(tensor).view(-1)) scale = tensor_max / max_value - # Calculate scale factor - scale = torch.max(torch.abs(tensor.flatten())) / max_value - # print(f"Optimizing {key} with scale: {scale}") - # numerical safety scale = torch.clamp(scale, min=1e-8) scale = scale.to(torch.float32) # ensure scale is in float32 for division