Skip to content

Commit b0c6ff5

Browse files
we1559SunMarc
andauthored
fix issue that some example with no trainer use accelerator.end_train… (#37435)
* fix issue that some example with no trainer use accelerator.end_training in a wrong way * reformat code --------- Co-authored-by: Marc Sun <[email protected]>
1 parent 6f5014a commit b0c6ff5

File tree

12 files changed

+36
-36
lines changed

12 files changed

+36
-36
lines changed

examples/pytorch/image-classification/run_image_classification_no_trainer.py

+3-3
Original file line numberDiff line numberDiff line change
@@ -617,9 +617,6 @@ def collate_fn(examples):
617617
output_dir = os.path.join(args.output_dir, output_dir)
618618
accelerator.save_state(output_dir)
619619

620-
if args.with_tracking:
621-
accelerator.end_training()
622-
623620
if args.output_dir is not None:
624621
accelerator.wait_for_everyone()
625622
unwrapped_model = accelerator.unwrap_model(model)
@@ -640,6 +637,9 @@ def collate_fn(examples):
640637
with open(os.path.join(args.output_dir, "all_results.json"), "w") as f:
641638
json.dump(all_results, f)
642639

640+
accelerator.wait_for_everyone()
641+
accelerator.end_training()
642+
643643

644644
if __name__ == "__main__":
645645
main()

examples/pytorch/image-pretraining/run_mim_no_trainer.py

+3-3
Original file line numberDiff line numberDiff line change
@@ -778,9 +778,6 @@ def preprocess_images(examples):
778778
output_dir = os.path.join(args.output_dir, output_dir)
779779
accelerator.save_state(output_dir)
780780

781-
if args.with_tracking:
782-
accelerator.end_training()
783-
784781
if args.output_dir is not None:
785782
accelerator.wait_for_everyone()
786783
unwrapped_model = accelerator.unwrap_model(model)
@@ -798,6 +795,9 @@ def preprocess_images(examples):
798795
token=args.hub_token,
799796
)
800797

798+
accelerator.wait_for_everyone()
799+
accelerator.end_training()
800+
801801

802802
if __name__ == "__main__":
803803
main()

examples/pytorch/instance-segmentation/run_instance_segmentation_no_trainer.py

+3-3
Original file line numberDiff line numberDiff line change
@@ -714,9 +714,6 @@ def main():
714714

715715
logger.info(f"Test metrics: {metrics}")
716716

717-
if args.with_tracking:
718-
accelerator.end_training()
719-
720717
if args.output_dir is not None:
721718
accelerator.wait_for_everyone()
722719
unwrapped_model = accelerator.unwrap_model(model)
@@ -739,6 +736,9 @@ def main():
739736
ignore_patterns=["epoch_*"],
740737
)
741738

739+
accelerator.wait_for_everyone()
740+
accelerator.end_training()
741+
742742

743743
if __name__ == "__main__":
744744
main()

examples/pytorch/language-modeling/run_clm_no_trainer.py

+3-3
Original file line numberDiff line numberDiff line change
@@ -697,9 +697,6 @@ def group_texts(examples):
697697
output_dir = os.path.join(args.output_dir, output_dir)
698698
accelerator.save_state(output_dir)
699699

700-
if args.with_tracking:
701-
accelerator.end_training()
702-
703700
if args.output_dir is not None:
704701
accelerator.wait_for_everyone()
705702
unwrapped_model = accelerator.unwrap_model(model)
@@ -719,6 +716,9 @@ def group_texts(examples):
719716
with open(os.path.join(args.output_dir, "all_results.json"), "w") as f:
720717
json.dump({"perplexity": perplexity}, f)
721718

719+
accelerator.wait_for_everyone()
720+
accelerator.end_training()
721+
722722

723723
if __name__ == "__main__":
724724
main()

examples/pytorch/language-modeling/run_fim_no_trainer.py

+3-3
Original file line numberDiff line numberDiff line change
@@ -891,9 +891,6 @@ def apply_fim(examples):
891891
output_dir = os.path.join(args.output_dir, output_dir)
892892
accelerator.save_state(output_dir)
893893

894-
if args.with_tracking:
895-
accelerator.end_training()
896-
897894
if args.output_dir is not None:
898895
accelerator.wait_for_everyone()
899896
unwrapped_model = accelerator.unwrap_model(model)
@@ -908,6 +905,9 @@ def apply_fim(examples):
908905
with open(os.path.join(args.output_dir, "all_results.json"), "w") as f:
909906
json.dump({"perplexity": perplexity}, f)
910907

908+
accelerator.wait_for_everyone()
909+
accelerator.end_training()
910+
911911

912912
if __name__ == "__main__":
913913
main()

examples/pytorch/language-modeling/run_mlm_no_trainer.py

+3-3
Original file line numberDiff line numberDiff line change
@@ -735,9 +735,6 @@ def group_texts(examples):
735735
output_dir = os.path.join(args.output_dir, output_dir)
736736
accelerator.save_state(output_dir)
737737

738-
if args.with_tracking:
739-
accelerator.end_training()
740-
741738
if args.output_dir is not None:
742739
accelerator.wait_for_everyone()
743740
unwrapped_model = accelerator.unwrap_model(model)
@@ -757,6 +754,9 @@ def group_texts(examples):
757754
with open(os.path.join(args.output_dir, "all_results.json"), "w") as f:
758755
json.dump({"perplexity": perplexity}, f)
759756

757+
accelerator.wait_for_everyone()
758+
accelerator.end_training()
759+
760760

761761
if __name__ == "__main__":
762762
main()

examples/pytorch/multiple-choice/run_swag_no_trainer.py

+3-3
Original file line numberDiff line numberDiff line change
@@ -622,9 +622,6 @@ def preprocess_function(examples):
622622
output_dir = os.path.join(args.output_dir, output_dir)
623623
accelerator.save_state(output_dir)
624624

625-
if args.with_tracking:
626-
accelerator.end_training()
627-
628625
if args.output_dir is not None:
629626
accelerator.wait_for_everyone()
630627
unwrapped_model = accelerator.unwrap_model(model)
@@ -645,6 +642,9 @@ def preprocess_function(examples):
645642
with open(os.path.join(args.output_dir, "all_results.json"), "w") as f:
646643
json.dump(all_results, f)
647644

645+
accelerator.wait_for_everyone()
646+
accelerator.end_training()
647+
648648

649649
if __name__ == "__main__":
650650
main()

examples/pytorch/object-detection/run_object_detection_no_trainer.py

+3-3
Original file line numberDiff line numberDiff line change
@@ -759,9 +759,6 @@ def main():
759759

760760
logger.info(f"Test metrics: {metrics}")
761761

762-
if args.with_tracking:
763-
accelerator.end_training()
764-
765762
if args.output_dir is not None:
766763
accelerator.wait_for_everyone()
767764
unwrapped_model = accelerator.unwrap_model(model)
@@ -784,6 +781,9 @@ def main():
784781
ignore_patterns=["epoch_*"],
785782
)
786783

784+
accelerator.wait_for_everyone()
785+
accelerator.end_training()
786+
787787

788788
if __name__ == "__main__":
789789
main()

examples/pytorch/semantic-segmentation/run_semantic_segmentation_no_trainer.py

+3-3
Original file line numberDiff line numberDiff line change
@@ -602,9 +602,6 @@ def preprocess_batch(example_batch, transforms: A.Compose):
602602
output_dir = os.path.join(args.output_dir, output_dir)
603603
accelerator.save_state(output_dir)
604604

605-
if args.with_tracking:
606-
accelerator.end_training()
607-
608605
if args.output_dir is not None:
609606
accelerator.wait_for_everyone()
610607
unwrapped_model = accelerator.unwrap_model(model)
@@ -628,6 +625,9 @@ def preprocess_batch(example_batch, transforms: A.Compose):
628625
with open(os.path.join(args.output_dir, "all_results.json"), "w") as f:
629626
json.dump(all_results, f, indent=2)
630627

628+
accelerator.wait_for_everyone()
629+
accelerator.end_training()
630+
631631

632632
if __name__ == "__main__":
633633
main()

examples/pytorch/text-classification/run_glue_no_trainer.py

+3-3
Original file line numberDiff line numberDiff line change
@@ -634,9 +634,6 @@ def preprocess_function(examples):
634634
output_dir = os.path.join(args.output_dir, output_dir)
635635
accelerator.save_state(output_dir)
636636

637-
if args.with_tracking:
638-
accelerator.end_training()
639-
640637
if args.output_dir is not None:
641638
accelerator.wait_for_everyone()
642639
unwrapped_model = accelerator.unwrap_model(model)
@@ -679,6 +676,9 @@ def preprocess_function(examples):
679676
with open(os.path.join(args.output_dir, "all_results.json"), "w") as f:
680677
json.dump(all_results, f)
681678

679+
accelerator.wait_for_everyone()
680+
accelerator.end_training()
681+
682682

683683
if __name__ == "__main__":
684684
main()

examples/pytorch/token-classification/run_ner_no_trainer.py

+3-3
Original file line numberDiff line numberDiff line change
@@ -794,9 +794,6 @@ def compute_metrics():
794794
output_dir = os.path.join(args.output_dir, output_dir)
795795
accelerator.save_state(output_dir)
796796

797-
if args.with_tracking:
798-
accelerator.end_training()
799-
800797
if args.output_dir is not None:
801798
accelerator.wait_for_everyone()
802799
unwrapped_model = accelerator.unwrap_model(model)
@@ -826,6 +823,9 @@ def compute_metrics():
826823
all_results[key] = int(value)
827824
json.dump(all_results, f)
828825

826+
accelerator.wait_for_everyone()
827+
accelerator.end_training()
828+
829829

830830
if __name__ == "__main__":
831831
main()

examples/pytorch/translation/run_translation_no_trainer.py

+3-3
Original file line numberDiff line numberDiff line change
@@ -762,9 +762,6 @@ def postprocess_text(preds, labels):
762762
output_dir = os.path.join(args.output_dir, output_dir)
763763
accelerator.save_state(output_dir)
764764

765-
if args.with_tracking:
766-
accelerator.end_training()
767-
768765
if args.output_dir is not None:
769766
accelerator.wait_for_everyone()
770767
unwrapped_model = accelerator.unwrap_model(model)
@@ -784,6 +781,9 @@ def postprocess_text(preds, labels):
784781
with open(os.path.join(args.output_dir, "all_results.json"), "w") as f:
785782
json.dump({"eval_bleu": eval_metric["score"]}, f)
786783

784+
accelerator.wait_for_everyone()
785+
accelerator.end_training()
786+
787787

788788
if __name__ == "__main__":
789789
main()

0 commit comments

Comments
 (0)