From 2cb41dfe19d9e3f01f79c68e3888454d84bb3523 Mon Sep 17 00:00:00 2001 From: NickKolok Date: Tue, 10 Dec 2024 01:15:29 +0300 Subject: [PATCH] [typo] Fix two more similar typos: 'gradient ccumulation' -> 'gradient accumulation' --- train_textual_inversion.py | 2 +- train_textual_inversion_XTI.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/train_textual_inversion.py b/train_textual_inversion.py index 6b6e7f5a..faa2ff61 100644 --- a/train_textual_inversion.py +++ b/train_textual_inversion.py @@ -489,7 +489,7 @@ class TextualInversionTrainer: accelerator.print( f" total train batch size (with parallel & distributed & accumulation) / 総バッチサイズ(並列学習、勾配合計含む): {total_batch_size}" ) - accelerator.print(f" gradient ccumulation steps / 勾配を合計するステップ数 = {args.gradient_accumulation_steps}") + accelerator.print(f" gradient accumulation steps / 勾配を合計するステップ数 = {args.gradient_accumulation_steps}") accelerator.print(f" total optimization steps / 学習ステップ数: {args.max_train_steps}") progress_bar = tqdm(range(args.max_train_steps), smoothing=0, disable=not accelerator.is_local_main_process, desc="steps") diff --git a/train_textual_inversion_XTI.py b/train_textual_inversion_XTI.py index 8dd5c672..66474ce7 100644 --- a/train_textual_inversion_XTI.py +++ b/train_textual_inversion_XTI.py @@ -382,7 +382,7 @@ def train(args): print(f" num epochs / epoch数: {num_train_epochs}") print(f" batch size per device / バッチサイズ: {args.train_batch_size}") print(f" total train batch size (with parallel & distributed & accumulation) / 総バッチサイズ(並列学習、勾配合計含む): {total_batch_size}") - print(f" gradient ccumulation steps / 勾配を合計するステップ数 = {args.gradient_accumulation_steps}") + print(f" gradient accumulation steps / 勾配を合計するステップ数 = {args.gradient_accumulation_steps}") print(f" total optimization steps / 学習ステップ数: {args.max_train_steps}") progress_bar = tqdm(range(args.max_train_steps), smoothing=0, disable=not accelerator.is_local_main_process, desc="steps")