common lr checking for dadaptation and prodigy

This commit is contained in:
Kohya S
2023-06-15 21:47:37 +09:00
parent e97d67a681
commit 5845de7d7c

View File

@@ -2752,15 +2752,7 @@ def get_optimizer(args, trainable_params):
optimizer_class = torch.optim.SGD optimizer_class = torch.optim.SGD
optimizer = optimizer_class(trainable_params, lr=lr, nesterov=True, **optimizer_kwargs) optimizer = optimizer_class(trainable_params, lr=lr, nesterov=True, **optimizer_kwargs)
elif optimizer_type.startswith("DAdapt".lower()): elif optimizer_type.startswith("DAdapt".lower()) or optimizer_type == "Prodigy".lower():
# DAdaptation family
# check dadaptation is installed
try:
import dadaptation
import dadaptation.experimental as experimental
except ImportError:
raise ImportError("No dadaptation / dadaptation がインストールされていないようです")
# check lr and lr_count, and print warning # check lr and lr_count, and print warning
actual_lr = lr actual_lr = lr
lr_count = 1 lr_count = 1
@@ -2773,72 +2765,60 @@ def get_optimizer(args, trainable_params):
if actual_lr <= 0.1: if actual_lr <= 0.1:
print( print(
f"learning rate is too low. If using dadaptation, set learning rate around 1.0 / 学習率が低すぎるようです。1.0前後の値を指定してください: lr={actual_lr}" f"learning rate is too low. If using D-Adaptation or Prodigy, set learning rate around 1.0 / 学習率が低すぎるようです。D-AdaptationまたはProdigyの使用時は1.0前後の値を指定してください: lr={actual_lr}"
) )
print("recommend option: lr=1.0 / 推奨は1.0です") print("recommend option: lr=1.0 / 推奨は1.0です")
if lr_count > 1: if lr_count > 1:
print( print(
f"when multiple learning rates are specified with dadaptation (e.g. for Text Encoder and U-Net), only the first one will take effect / D-Adaptationで複数の学習率を指定した場合Text EncoderとU-Netなど、最初の学習率のみが有効になります: lr={actual_lr}" f"when multiple learning rates are specified with dadaptation (e.g. for Text Encoder and U-Net), only the first one will take effect / D-AdaptationまたはProdigyで複数の学習率を指定した場合Text EncoderとU-Netなど、最初の学習率のみが有効になります: lr={actual_lr}"
) )
# set optimizer if optimizer_type.startswith("DAdapt".lower()):
if optimizer_type == "DAdaptation".lower() or optimizer_type == "DAdaptAdamPreprint".lower(): # DAdaptation family
optimizer_class = experimental.DAdaptAdamPreprint # check dadaptation is installed
print(f"use D-Adaptation AdamPreprint optimizer | {optimizer_kwargs}") try:
elif optimizer_type == "DAdaptAdaGrad".lower(): import dadaptation
optimizer_class = dadaptation.DAdaptAdaGrad import dadaptation.experimental as experimental
print(f"use D-Adaptation AdaGrad optimizer | {optimizer_kwargs}") except ImportError:
elif optimizer_type == "DAdaptAdam".lower(): raise ImportError("No dadaptation / dadaptation がインストールされていないようです")
optimizer_class = dadaptation.DAdaptAdam
print(f"use D-Adaptation Adam optimizer | {optimizer_kwargs}") # set optimizer
elif optimizer_type == "DAdaptAdan".lower(): if optimizer_type == "DAdaptation".lower() or optimizer_type == "DAdaptAdamPreprint".lower():
optimizer_class = dadaptation.DAdaptAdan optimizer_class = experimental.DAdaptAdamPreprint
print(f"use D-Adaptation Adan optimizer | {optimizer_kwargs}") print(f"use D-Adaptation AdamPreprint optimizer | {optimizer_kwargs}")
elif optimizer_type == "DAdaptAdanIP".lower(): elif optimizer_type == "DAdaptAdaGrad".lower():
optimizer_class = experimental.DAdaptAdanIP optimizer_class = dadaptation.DAdaptAdaGrad
print(f"use D-Adaptation AdanIP optimizer | {optimizer_kwargs}") print(f"use D-Adaptation AdaGrad optimizer | {optimizer_kwargs}")
elif optimizer_type == "DAdaptLion".lower(): elif optimizer_type == "DAdaptAdam".lower():
optimizer_class = dadaptation.DAdaptLion optimizer_class = dadaptation.DAdaptAdam
print(f"use D-Adaptation Lion optimizer | {optimizer_kwargs}") print(f"use D-Adaptation Adam optimizer | {optimizer_kwargs}")
elif optimizer_type == "DAdaptSGD".lower(): elif optimizer_type == "DAdaptAdan".lower():
optimizer_class = dadaptation.DAdaptSGD optimizer_class = dadaptation.DAdaptAdan
print(f"use D-Adaptation SGD optimizer | {optimizer_kwargs}") print(f"use D-Adaptation Adan optimizer | {optimizer_kwargs}")
elif optimizer_type == "DAdaptAdanIP".lower():
optimizer_class = experimental.DAdaptAdanIP
print(f"use D-Adaptation AdanIP optimizer | {optimizer_kwargs}")
elif optimizer_type == "DAdaptLion".lower():
optimizer_class = dadaptation.DAdaptLion
print(f"use D-Adaptation Lion optimizer | {optimizer_kwargs}")
elif optimizer_type == "DAdaptSGD".lower():
optimizer_class = dadaptation.DAdaptSGD
print(f"use D-Adaptation SGD optimizer | {optimizer_kwargs}")
else:
raise ValueError(f"Unknown optimizer type: {optimizer_type}")
optimizer = optimizer_class(trainable_params, lr=lr, **optimizer_kwargs)
else: else:
raise ValueError(f"Unknown optimizer type: {optimizer_type}") # Prodigy
# check Prodigy is installed
try:
import prodigyopt
except ImportError:
raise ImportError("No Prodigy / Prodigy がインストールされていないようです")
optimizer = optimizer_class(trainable_params, lr=lr, **optimizer_kwargs) print(f"use Prodigy optimizer | {optimizer_kwargs}")
optimizer_class = prodigyopt.Prodigy
elif optimizer_type == "Prodigy".lower(): optimizer = optimizer_class(trainable_params, lr=lr, **optimizer_kwargs)
# Prodigy
# check Prodigy is installed
try:
import prodigyopt
except ImportError:
raise ImportError("No Prodigy / Prodigy がインストールされていないようです")
# check lr and lr_count, and print warning
actual_lr = lr
lr_count = 1
if type(trainable_params) == list and type(trainable_params[0]) == dict:
lrs = set()
actual_lr = trainable_params[0].get("lr", actual_lr)
for group in trainable_params:
lrs.add(group.get("lr", actual_lr))
lr_count = len(lrs)
if actual_lr <= 0.1:
print(
f"learning rate is too low. If using Prodigy, set learning rate around 1.0 / 学習率が低すぎるようです。1.0前後の値を指定してください: lr={actual_lr}"
)
print("recommend option: lr=1.0 / 推奨は1.0です")
if lr_count > 1:
print(
f"when multiple learning rates are specified with Prodigy (e.g. for Text Encoder and U-Net), only the first one will take effect / Prodigyで複数の学習率を指定した場合Text EncoderとU-Netなど、最初の学習率のみが有効になります: lr={actual_lr}"
)
print(f"use Prodigy optimizer | {optimizer_kwargs}")
optimizer_class = prodigyopt.Prodigy
optimizer = optimizer_class(trainable_params, lr=lr, **optimizer_kwargs)
elif optimizer_type == "Adafactor".lower(): elif optimizer_type == "Adafactor".lower():
# 引数を確認して適宜補正する # 引数を確認して適宜補正する