Skip to content

Commit f53cf60

Browse files
committed
fix newlines in mock data generation
1 parent 113eca4 commit f53cf60

File tree

1 file changed

+14
-14
lines changed

1 file changed

+14
-14
lines changed

test/builtin_dataset_mocks.py

Lines changed: 14 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -534,7 +534,7 @@ def imagenet(info, root, config):
534534
devkit_root.mkdir()
535535
data_root = devkit_root / "data"
536536
data_root.mkdir()
537-
with open(data_root / "ILSVRC2012_validation_ground_truth.txt", "w") as file:
537+
with open(data_root / "ILSVRC2012_validation_ground_truth.txt", "w", newline="") as file:
538538
for label in torch.randint(0, len(wnids), (num_samples,)).tolist():
539539
file.write(f"{label}\n")
540540
make_tar(root, f"{devkit_root}.tar.gz", devkit_root, compression="gz")
@@ -672,7 +672,7 @@ def _make_split_files(cls, root_map):
672672
}
673673

674674
for split, ids in ids_map.items():
675-
with open(root_map[split] / f"{split}.txt", "w") as fh:
675+
with open(root_map[split] / f"{split}.txt", "w", newline="") as fh:
676676
fh.writelines(f"{id}\n" for id in ids)
677677

678678
return sorted(set(itertools.chain(*ids_map.values()))), {split: len(ids) for split, ids in ids_map.items()}
@@ -738,11 +738,11 @@ def semeion(info, root, config):
738738

739739
images = torch.rand(num_samples, 256)
740740
labels = one_hot(torch.randint(len(info.categories), size=(num_samples,)))
741-
with open(root / "semeion.data", "w") as fh:
741+
with open(root / "semeion.data", "w", newline="") as fh:
742742
for image, one_hot_label in zip(images, labels):
743743
image_columns = " ".join([f"{pixel.item():.4f}" for pixel in image])
744744
labels_columns = " ".join([str(label.item()) for label in one_hot_label])
745-
fh.write(f"{image_columns} {labels_columns}\n")
745+
fh.write(f"{image_columns} {labels_columns}\r\n")
746746

747747
return num_samples
748748

@@ -780,7 +780,7 @@ def _make_split_files(cls, root, *, year, trainval):
780780
task_folder = split_folder / task_sub_folder
781781
task_folder.mkdir(parents=True, exist_ok=True)
782782
for split, ids in ids_map.items():
783-
with open(task_folder / f"{split}.txt", "w") as fh:
783+
with open(task_folder / f"{split}.txt", "w", newline="") as fh:
784784
fh.writelines(f"{id}\n" for id in ids)
785785

786786
return sorted(set(itertools.chain(*ids_map.values()))), {split: len(ids) for split, ids in ids_map.items()}
@@ -857,7 +857,7 @@ def voc(info, root, config):
857857
class CelebAMockData:
858858
@classmethod
859859
def _make_ann_file(cls, root, name, data, *, field_names=None):
860-
with open(root / name, "w") as file:
860+
with open(root / name, "w", newline="") as file:
861861
if field_names:
862862
file.write(f"{len(data)}\r\n")
863863
file.write(" ".join(field_names) + "\r\n")
@@ -971,13 +971,13 @@ def dtd(info, root, _):
971971
meta_folder = data_folder / "labels"
972972
meta_folder.mkdir()
973973

974-
with open(meta_folder / "labels_joint_anno.txt", "w") as file:
974+
with open(meta_folder / "labels_joint_anno.txt", "w", newline="") as file:
975975
for cls, image_ids in image_ids_per_category.items():
976976
for image_id in image_ids:
977977
joint_categories = random.choices(
978978
list(categories - {cls}), k=int(torch.randint(len(categories) - 1, ()))
979979
)
980-
file.write(" ".join([image_id, *sorted([cls, *joint_categories])]) + "\n")
980+
file.write(" ".join([image_id, *sorted([cls, *joint_categories])]) + " \n")
981981

982982
image_ids = list(itertools.chain(*image_ids_per_category.values()))
983983
splits = ("train", "val", "test")
@@ -986,7 +986,7 @@ def dtd(info, root, _):
986986
random.shuffle(image_ids)
987987
for offset, split in enumerate(splits):
988988
image_ids_in_config = image_ids[offset :: len(splits)]
989-
with open(meta_folder / f"{split}{fold}.txt", "w") as file:
989+
with open(meta_folder / f"{split}{fold}.txt", "w", newline="") as file:
990990
file.write("\n".join(image_ids_in_config) + "\n")
991991

992992
num_samples_map[info.make_config(split=split, fold=str(fold))] = len(image_ids_in_config)
@@ -1104,7 +1104,7 @@ def generate(self, root):
11041104
num_samples_map = {}
11051105
for offset, split in enumerate(splits):
11061106
split_and_classification_anns_in_split = split_and_classification_anns[offset :: len(splits)]
1107-
with open(anns_folder / f"{split}.txt", "w") as file:
1107+
with open(anns_folder / f"{split}.txt", "w", newline="") as file:
11081108
writer = csv.writer(file, delimiter=" ")
11091109
for split_and_classification_ann in split_and_classification_anns_in_split:
11101110
writer.writerow(split_and_classification_ann)
@@ -1171,7 +1171,7 @@ def _make_archive(cls, root):
11711171
image_files = cls._make_images(images_folder)
11721172
image_ids = list(range(1, len(image_files) + 1))
11731173

1174-
with open(archive_folder / "images.txt", "w") as file:
1174+
with open(archive_folder / "images.txt", "w", newline="") as file:
11751175
file.write(
11761176
"\n".join(
11771177
f"{id} {path.relative_to(images_folder).as_posix()}" for id, path in zip(image_ids, image_files)
@@ -1181,10 +1181,10 @@ def _make_archive(cls, root):
11811181
split_ids = torch.randint(2, (len(image_ids),)).tolist()
11821182
counts = Counter(split_ids)
11831183
num_samples_map = {"train": counts[1], "test": counts[0]}
1184-
with open(archive_folder / "train_test_split.txt", "w") as file:
1184+
with open(archive_folder / "train_test_split.txt", "w", newline="") as file:
11851185
file.write("\n".join(f"{image_id} {split_id}" for image_id, split_id in zip(image_ids, split_ids)))
11861186

1187-
with open(archive_folder / "bounding_boxes.txt", "w") as file:
1187+
with open(archive_folder / "bounding_boxes.txt", "w", newline="") as file:
11881188
file.write(
11891189
"\n".join(
11901190
" ".join(
@@ -1237,7 +1237,7 @@ def _make_splits(cls, root, image_files):
12371237
image_files_in_split = image_files[offset :: len(splits)]
12381238

12391239
split_file = split_folder / f"{split}.txt"
1240-
with open(split_file, "w") as file:
1240+
with open(split_file, "w", newline="") as file:
12411241
file.write(
12421242
"\n".join(
12431243
sorted(

0 commit comments

Comments
 (0)