Merge pull request #630 from ddPn08/sdxl

make tracker init_kwargs configurable
This commit is contained in:
Kohya S
2023-07-20 22:05:55 +09:00
committed by GitHub
8 changed files with 40 additions and 7 deletions

View File

@@ -5,6 +5,7 @@ import gc
import math
import os
from multiprocessing import Value
import toml
from tqdm import tqdm
import torch
@@ -355,7 +356,10 @@ def train(args):
custom_train_functions.fix_noise_scheduler_betas_for_zero_terminal_snr(noise_scheduler)
if accelerator.is_main_process:
accelerator.init_trackers("finetuning" if args.log_tracker_name is None else args.log_tracker_name)
init_kwargs = {}
if args.log_tracker_config is not None:
init_kwargs = toml.load(args.log_tracker_config)
accelerator.init_trackers("finetuning" if args.log_tracker_name is None else args.log_tracker_name, init_kwargs=init_kwargs)
for epoch in range(num_train_epochs):
accelerator.print(f"\nepoch {epoch+1}/{num_train_epochs}")