From b32bcb8e58243197fd4dc585456812bdefb7a2e9 Mon Sep 17 00:00:00 2001 From: eliphatfs Date: Fri, 7 Jul 2023 14:29:24 +0800 Subject: [PATCH 1/2] Multiply lr scheduler steps by `num_processes`. --- examples/controlnet/train_controlnet.py | 4 ++-- examples/custom_diffusion/train_custom_diffusion.py | 4 ++-- examples/dreambooth/train_dreambooth.py | 4 ++-- examples/dreambooth/train_dreambooth_lora.py | 4 ++-- examples/instruct_pix2pix/train_instruct_pix2pix.py | 4 ++-- .../dreambooth_inpaint/train_dreambooth_inpaint.py | 4 ++-- .../dreambooth_inpaint/train_dreambooth_inpaint_lora.py | 4 ++-- .../intel_opts/textual_inversion/textual_inversion_bf16.py | 4 ++-- .../intel_opts/textual_inversion_dfq/textual_inversion.py | 4 ++-- examples/research_projects/lora/train_text_to_image_lora.py | 4 ++-- .../mulit_token_textual_inversion/textual_inversion.py | 4 ++-- .../train_multi_subject_dreambooth.py | 4 ++-- .../onnxruntime/text_to_image/train_text_to_image.py | 4 ++-- .../onnxruntime/textual_inversion/textual_inversion.py | 4 ++-- examples/text_to_image/train_text_to_image.py | 4 ++-- examples/text_to_image/train_text_to_image_lora.py | 4 ++-- examples/textual_inversion/textual_inversion.py | 4 ++-- 17 files changed, 34 insertions(+), 34 deletions(-) diff --git a/examples/controlnet/train_controlnet.py b/examples/controlnet/train_controlnet.py index 21ab38a3d7a9..c46fb2a8fa3e 100644 --- a/examples/controlnet/train_controlnet.py +++ b/examples/controlnet/train_controlnet.py @@ -897,8 +897,8 @@ def load_model_hook(models, input_dir): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, + num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, num_cycles=args.lr_num_cycles, power=args.lr_power, ) diff --git a/examples/custom_diffusion/train_custom_diffusion.py b/examples/custom_diffusion/train_custom_diffusion.py index e0ec56eca1f3..b7063beabd60 100644 --- a/examples/custom_diffusion/train_custom_diffusion.py +++ b/examples/custom_diffusion/train_custom_diffusion.py @@ -1007,8 +1007,8 @@ def main(args): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, + num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, ) # Prepare everything with our `accelerator`. diff --git a/examples/dreambooth/train_dreambooth.py b/examples/dreambooth/train_dreambooth.py index 797cfbd0e5d7..2061c2bf7f96 100644 --- a/examples/dreambooth/train_dreambooth.py +++ b/examples/dreambooth/train_dreambooth.py @@ -1075,8 +1075,8 @@ def compute_text_embeddings(prompt): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, + num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, num_cycles=args.lr_num_cycles, power=args.lr_power, ) diff --git a/examples/dreambooth/train_dreambooth_lora.py b/examples/dreambooth/train_dreambooth_lora.py index 72fcfa648b48..35623abf2523 100644 --- a/examples/dreambooth/train_dreambooth_lora.py +++ b/examples/dreambooth/train_dreambooth_lora.py @@ -1038,8 +1038,8 @@ def compute_text_embeddings(prompt): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, + num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, num_cycles=args.lr_num_cycles, power=args.lr_power, ) diff --git a/examples/instruct_pix2pix/train_instruct_pix2pix.py b/examples/instruct_pix2pix/train_instruct_pix2pix.py index e84698a8f215..b14c1ed032cd 100644 --- a/examples/instruct_pix2pix/train_instruct_pix2pix.py +++ b/examples/instruct_pix2pix/train_instruct_pix2pix.py @@ -690,8 +690,8 @@ def collate_fn(examples): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, + num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, ) # Prepare everything with our `accelerator`. diff --git a/examples/research_projects/dreambooth_inpaint/train_dreambooth_inpaint.py b/examples/research_projects/dreambooth_inpaint/train_dreambooth_inpaint.py index a9ff5e90ed51..a07a430bed14 100644 --- a/examples/research_projects/dreambooth_inpaint/train_dreambooth_inpaint.py +++ b/examples/research_projects/dreambooth_inpaint/train_dreambooth_inpaint.py @@ -600,8 +600,8 @@ def collate_fn(examples): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, + num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, ) if args.train_text_encoder: diff --git a/examples/research_projects/dreambooth_inpaint/train_dreambooth_inpaint_lora.py b/examples/research_projects/dreambooth_inpaint/train_dreambooth_inpaint_lora.py index b3f37c53eac8..385db3a9e92d 100644 --- a/examples/research_projects/dreambooth_inpaint/train_dreambooth_inpaint_lora.py +++ b/examples/research_projects/dreambooth_inpaint/train_dreambooth_inpaint_lora.py @@ -644,8 +644,8 @@ def collate_fn(examples): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, + num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, ) # Prepare everything with our `accelerator`. diff --git a/examples/research_projects/intel_opts/textual_inversion/textual_inversion_bf16.py b/examples/research_projects/intel_opts/textual_inversion/textual_inversion_bf16.py index 3c2dd396ca47..5af6b2675799 100644 --- a/examples/research_projects/intel_opts/textual_inversion/textual_inversion_bf16.py +++ b/examples/research_projects/intel_opts/textual_inversion/textual_inversion_bf16.py @@ -481,8 +481,8 @@ def main(): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, + num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, ) text_encoder, optimizer, train_dataloader, lr_scheduler = accelerator.prepare( diff --git a/examples/research_projects/intel_opts/textual_inversion_dfq/textual_inversion.py b/examples/research_projects/intel_opts/textual_inversion_dfq/textual_inversion.py index 2188b17b68d7..6926b8f948ac 100644 --- a/examples/research_projects/intel_opts/textual_inversion_dfq/textual_inversion.py +++ b/examples/research_projects/intel_opts/textual_inversion_dfq/textual_inversion.py @@ -588,8 +588,8 @@ def main(): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, + num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, ) if not train_unet: diff --git a/examples/research_projects/lora/train_text_to_image_lora.py b/examples/research_projects/lora/train_text_to_image_lora.py index e2e0dbaaa4db..bcf3c8d1e51e 100644 --- a/examples/research_projects/lora/train_text_to_image_lora.py +++ b/examples/research_projects/lora/train_text_to_image_lora.py @@ -701,8 +701,8 @@ def collate_fn(examples): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, + num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, ) # Prepare everything with our `accelerator`. diff --git a/examples/research_projects/mulit_token_textual_inversion/textual_inversion.py b/examples/research_projects/mulit_token_textual_inversion/textual_inversion.py index 2aaebd729c80..171133904e56 100644 --- a/examples/research_projects/mulit_token_textual_inversion/textual_inversion.py +++ b/examples/research_projects/mulit_token_textual_inversion/textual_inversion.py @@ -690,8 +690,8 @@ def main(): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, + num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, ) # Prepare everything with our `accelerator`. diff --git a/examples/research_projects/multi_subject_dreambooth/train_multi_subject_dreambooth.py b/examples/research_projects/multi_subject_dreambooth/train_multi_subject_dreambooth.py index c75a0a9acc64..7b7570c6a215 100644 --- a/examples/research_projects/multi_subject_dreambooth/train_multi_subject_dreambooth.py +++ b/examples/research_projects/multi_subject_dreambooth/train_multi_subject_dreambooth.py @@ -970,8 +970,8 @@ def main(args): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, + num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, num_cycles=args.lr_num_cycles, power=args.lr_power, ) diff --git a/examples/research_projects/onnxruntime/text_to_image/train_text_to_image.py b/examples/research_projects/onnxruntime/text_to_image/train_text_to_image.py index f1e28c7e2caa..044d3390de21 100644 --- a/examples/research_projects/onnxruntime/text_to_image/train_text_to_image.py +++ b/examples/research_projects/onnxruntime/text_to_image/train_text_to_image.py @@ -732,8 +732,8 @@ def collate_fn(examples): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, + num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, ) # Prepare everything with our `accelerator`. diff --git a/examples/research_projects/onnxruntime/textual_inversion/textual_inversion.py b/examples/research_projects/onnxruntime/textual_inversion/textual_inversion.py index 1013dfa8aba4..5770260d1b54 100644 --- a/examples/research_projects/onnxruntime/textual_inversion/textual_inversion.py +++ b/examples/research_projects/onnxruntime/textual_inversion/textual_inversion.py @@ -741,8 +741,8 @@ def main(): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, + num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, ) # Prepare everything with our `accelerator`. diff --git a/examples/text_to_image/train_text_to_image.py b/examples/text_to_image/train_text_to_image.py index 8f4cdba2bfb6..ca511294d202 100644 --- a/examples/text_to_image/train_text_to_image.py +++ b/examples/text_to_image/train_text_to_image.py @@ -819,8 +819,8 @@ def collate_fn(examples): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, + num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, ) # Prepare everything with our `accelerator`. diff --git a/examples/text_to_image/train_text_to_image_lora.py b/examples/text_to_image/train_text_to_image_lora.py index 29259e408eff..e233b5ae5bd9 100644 --- a/examples/text_to_image/train_text_to_image_lora.py +++ b/examples/text_to_image/train_text_to_image_lora.py @@ -662,8 +662,8 @@ def collate_fn(examples): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, + num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, ) # Prepare everything with our `accelerator`. diff --git a/examples/textual_inversion/textual_inversion.py b/examples/textual_inversion/textual_inversion.py index 14b0997862d2..1bd90cbdcce2 100644 --- a/examples/textual_inversion/textual_inversion.py +++ b/examples/textual_inversion/textual_inversion.py @@ -737,8 +737,8 @@ def main(): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps, + num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, num_cycles=args.lr_num_cycles * args.gradient_accumulation_steps, ) From 660c55194a44442eb73ad03bc1b8e6d3e6a276e8 Mon Sep 17 00:00:00 2001 From: eliphatfs Date: Thu, 13 Jul 2023 20:07:27 +0800 Subject: [PATCH 2/2] Stop multiplying steps by gradient accumulation. --- examples/controlnet/train_controlnet.py | 4 ++-- examples/custom_diffusion/train_custom_diffusion.py | 4 ++-- examples/dreambooth/train_dreambooth.py | 4 ++-- examples/dreambooth/train_dreambooth_lora.py | 4 ++-- examples/instruct_pix2pix/train_instruct_pix2pix.py | 4 ++-- .../dreambooth_inpaint/train_dreambooth_inpaint.py | 4 ++-- .../dreambooth_inpaint/train_dreambooth_inpaint_lora.py | 4 ++-- .../intel_opts/textual_inversion/textual_inversion_bf16.py | 4 ++-- .../intel_opts/textual_inversion_dfq/textual_inversion.py | 4 ++-- examples/research_projects/lora/train_text_to_image_lora.py | 4 ++-- .../mulit_token_textual_inversion/textual_inversion.py | 4 ++-- .../train_multi_subject_dreambooth.py | 4 ++-- .../onnxruntime/text_to_image/train_text_to_image.py | 4 ++-- .../onnxruntime/textual_inversion/textual_inversion.py | 4 ++-- examples/text_to_image/train_text_to_image.py | 4 ++-- examples/text_to_image/train_text_to_image_lora.py | 4 ++-- examples/textual_inversion/textual_inversion.py | 6 +++--- 17 files changed, 35 insertions(+), 35 deletions(-) diff --git a/examples/controlnet/train_controlnet.py b/examples/controlnet/train_controlnet.py index c46fb2a8fa3e..39f1b0754d86 100644 --- a/examples/controlnet/train_controlnet.py +++ b/examples/controlnet/train_controlnet.py @@ -897,8 +897,8 @@ def load_model_hook(models, input_dir): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_warmup_steps=args.lr_warmup_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * accelerator.num_processes, num_cycles=args.lr_num_cycles, power=args.lr_power, ) diff --git a/examples/custom_diffusion/train_custom_diffusion.py b/examples/custom_diffusion/train_custom_diffusion.py index b7063beabd60..166473a9d74c 100644 --- a/examples/custom_diffusion/train_custom_diffusion.py +++ b/examples/custom_diffusion/train_custom_diffusion.py @@ -1007,8 +1007,8 @@ def main(args): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_warmup_steps=args.lr_warmup_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * accelerator.num_processes, ) # Prepare everything with our `accelerator`. diff --git a/examples/dreambooth/train_dreambooth.py b/examples/dreambooth/train_dreambooth.py index 2061c2bf7f96..423584443746 100644 --- a/examples/dreambooth/train_dreambooth.py +++ b/examples/dreambooth/train_dreambooth.py @@ -1075,8 +1075,8 @@ def compute_text_embeddings(prompt): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_warmup_steps=args.lr_warmup_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * accelerator.num_processes, num_cycles=args.lr_num_cycles, power=args.lr_power, ) diff --git a/examples/dreambooth/train_dreambooth_lora.py b/examples/dreambooth/train_dreambooth_lora.py index 35623abf2523..9c3ef325db3e 100644 --- a/examples/dreambooth/train_dreambooth_lora.py +++ b/examples/dreambooth/train_dreambooth_lora.py @@ -1038,8 +1038,8 @@ def compute_text_embeddings(prompt): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_warmup_steps=args.lr_warmup_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * accelerator.num_processes, num_cycles=args.lr_num_cycles, power=args.lr_power, ) diff --git a/examples/instruct_pix2pix/train_instruct_pix2pix.py b/examples/instruct_pix2pix/train_instruct_pix2pix.py index b14c1ed032cd..0a1aa0f3a77c 100644 --- a/examples/instruct_pix2pix/train_instruct_pix2pix.py +++ b/examples/instruct_pix2pix/train_instruct_pix2pix.py @@ -690,8 +690,8 @@ def collate_fn(examples): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_warmup_steps=args.lr_warmup_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * accelerator.num_processes, ) # Prepare everything with our `accelerator`. diff --git a/examples/research_projects/dreambooth_inpaint/train_dreambooth_inpaint.py b/examples/research_projects/dreambooth_inpaint/train_dreambooth_inpaint.py index a07a430bed14..a3eaba014cf6 100644 --- a/examples/research_projects/dreambooth_inpaint/train_dreambooth_inpaint.py +++ b/examples/research_projects/dreambooth_inpaint/train_dreambooth_inpaint.py @@ -600,8 +600,8 @@ def collate_fn(examples): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_warmup_steps=args.lr_warmup_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * accelerator.num_processes, ) if args.train_text_encoder: diff --git a/examples/research_projects/dreambooth_inpaint/train_dreambooth_inpaint_lora.py b/examples/research_projects/dreambooth_inpaint/train_dreambooth_inpaint_lora.py index 385db3a9e92d..d25c6d22f8e7 100644 --- a/examples/research_projects/dreambooth_inpaint/train_dreambooth_inpaint_lora.py +++ b/examples/research_projects/dreambooth_inpaint/train_dreambooth_inpaint_lora.py @@ -644,8 +644,8 @@ def collate_fn(examples): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_warmup_steps=args.lr_warmup_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * accelerator.num_processes, ) # Prepare everything with our `accelerator`. diff --git a/examples/research_projects/intel_opts/textual_inversion/textual_inversion_bf16.py b/examples/research_projects/intel_opts/textual_inversion/textual_inversion_bf16.py index 5af6b2675799..ff24130c9b61 100644 --- a/examples/research_projects/intel_opts/textual_inversion/textual_inversion_bf16.py +++ b/examples/research_projects/intel_opts/textual_inversion/textual_inversion_bf16.py @@ -481,8 +481,8 @@ def main(): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_warmup_steps=args.lr_warmup_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * accelerator.num_processes, ) text_encoder, optimizer, train_dataloader, lr_scheduler = accelerator.prepare( diff --git a/examples/research_projects/intel_opts/textual_inversion_dfq/textual_inversion.py b/examples/research_projects/intel_opts/textual_inversion_dfq/textual_inversion.py index 6926b8f948ac..675b16f30d78 100644 --- a/examples/research_projects/intel_opts/textual_inversion_dfq/textual_inversion.py +++ b/examples/research_projects/intel_opts/textual_inversion_dfq/textual_inversion.py @@ -588,8 +588,8 @@ def main(): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_warmup_steps=args.lr_warmup_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * accelerator.num_processes, ) if not train_unet: diff --git a/examples/research_projects/lora/train_text_to_image_lora.py b/examples/research_projects/lora/train_text_to_image_lora.py index bcf3c8d1e51e..d69284042af4 100644 --- a/examples/research_projects/lora/train_text_to_image_lora.py +++ b/examples/research_projects/lora/train_text_to_image_lora.py @@ -701,8 +701,8 @@ def collate_fn(examples): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_warmup_steps=args.lr_warmup_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * accelerator.num_processes, ) # Prepare everything with our `accelerator`. diff --git a/examples/research_projects/mulit_token_textual_inversion/textual_inversion.py b/examples/research_projects/mulit_token_textual_inversion/textual_inversion.py index 171133904e56..63b6c3860a29 100644 --- a/examples/research_projects/mulit_token_textual_inversion/textual_inversion.py +++ b/examples/research_projects/mulit_token_textual_inversion/textual_inversion.py @@ -690,8 +690,8 @@ def main(): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_warmup_steps=args.lr_warmup_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * accelerator.num_processes, ) # Prepare everything with our `accelerator`. diff --git a/examples/research_projects/multi_subject_dreambooth/train_multi_subject_dreambooth.py b/examples/research_projects/multi_subject_dreambooth/train_multi_subject_dreambooth.py index 7b7570c6a215..4e03e23fc128 100644 --- a/examples/research_projects/multi_subject_dreambooth/train_multi_subject_dreambooth.py +++ b/examples/research_projects/multi_subject_dreambooth/train_multi_subject_dreambooth.py @@ -970,8 +970,8 @@ def main(args): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_warmup_steps=args.lr_warmup_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * accelerator.num_processes, num_cycles=args.lr_num_cycles, power=args.lr_power, ) diff --git a/examples/research_projects/onnxruntime/text_to_image/train_text_to_image.py b/examples/research_projects/onnxruntime/text_to_image/train_text_to_image.py index 044d3390de21..0cd915b423c7 100644 --- a/examples/research_projects/onnxruntime/text_to_image/train_text_to_image.py +++ b/examples/research_projects/onnxruntime/text_to_image/train_text_to_image.py @@ -732,8 +732,8 @@ def collate_fn(examples): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_warmup_steps=args.lr_warmup_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * accelerator.num_processes, ) # Prepare everything with our `accelerator`. diff --git a/examples/research_projects/onnxruntime/textual_inversion/textual_inversion.py b/examples/research_projects/onnxruntime/textual_inversion/textual_inversion.py index 5770260d1b54..59b5089d07b4 100644 --- a/examples/research_projects/onnxruntime/textual_inversion/textual_inversion.py +++ b/examples/research_projects/onnxruntime/textual_inversion/textual_inversion.py @@ -741,8 +741,8 @@ def main(): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_warmup_steps=args.lr_warmup_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * accelerator.num_processes, ) # Prepare everything with our `accelerator`. diff --git a/examples/text_to_image/train_text_to_image.py b/examples/text_to_image/train_text_to_image.py index ca511294d202..94bbb008ced1 100644 --- a/examples/text_to_image/train_text_to_image.py +++ b/examples/text_to_image/train_text_to_image.py @@ -819,8 +819,8 @@ def collate_fn(examples): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_warmup_steps=args.lr_warmup_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * accelerator.num_processes, ) # Prepare everything with our `accelerator`. diff --git a/examples/text_to_image/train_text_to_image_lora.py b/examples/text_to_image/train_text_to_image_lora.py index e233b5ae5bd9..d8226b650bc1 100644 --- a/examples/text_to_image/train_text_to_image_lora.py +++ b/examples/text_to_image/train_text_to_image_lora.py @@ -662,8 +662,8 @@ def collate_fn(examples): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, + num_warmup_steps=args.lr_warmup_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * accelerator.num_processes, ) # Prepare everything with our `accelerator`. diff --git a/examples/textual_inversion/textual_inversion.py b/examples/textual_inversion/textual_inversion.py index 1bd90cbdcce2..d36b8b325995 100644 --- a/examples/textual_inversion/textual_inversion.py +++ b/examples/textual_inversion/textual_inversion.py @@ -737,9 +737,9 @@ def main(): lr_scheduler = get_scheduler( args.lr_scheduler, optimizer=optimizer, - num_warmup_steps=args.lr_warmup_steps * args.gradient_accumulation_steps * accelerator.num_processes, - num_training_steps=args.max_train_steps * args.gradient_accumulation_steps * accelerator.num_processes, - num_cycles=args.lr_num_cycles * args.gradient_accumulation_steps, + num_warmup_steps=args.lr_warmup_steps * accelerator.num_processes, + num_training_steps=args.max_train_steps * accelerator.num_processes, + num_cycles=args.lr_num_cycles, ) # Prepare everything with our `accelerator`.