From f4bc801b471d21612cd8522b73ab3016cfc6f86c Mon Sep 17 00:00:00 2001 From: Philipp Donn <30521025+phinik@users.noreply.github.com> Date: Tue, 9 Jan 2024 08:35:17 +0100 Subject: [PATCH] remove script --- scripts/old_dataset_script.py | 247 ---------------------------------- 1 file changed, 247 deletions(-) delete mode 100644 scripts/old_dataset_script.py diff --git a/scripts/old_dataset_script.py b/scripts/old_dataset_script.py deleted file mode 100644 index 641ec39..0000000 --- a/scripts/old_dataset_script.py +++ /dev/null @@ -1,247 +0,0 @@ -#!/usr/bin/env python3 - -import os -import yaml -import cv2 -import random -import argparse -import numpy as np - - -# Available classes for YOEO -CLASSES = { - 'bb_classes': ['ball', 'robot'], - 'segmentation_classes': ['field edge', 'lines'], - 'ignored_classes': ['goalpost', 'obstacle', 'L-Intersection', 'X-Intersection', 'T-Intersection'] - } - -#ROBOT_CLASSES = ["robot_red", "robot_blue", "robot_unknown"] -#ROBOT_NUMBER = [None, 1, 2, 3, 4, 5, 6] -ROBOT_COLOR_NUMBERS = [ - "red_None", "red_1", "red_2", "red_3", "red_4", "red_5", "red_6", - "blue_None", "blue_1", "blue_2", "blue_3", "blue_4", "blue_5", "blue_6", - "unknown_None", "unknown_1", "unknown_2", "unknown_3", "unknown_4", "unknown_5", "unknown_6" - ] - -""" -This script reads annotations in the expected yaml format below -to generate the corresponding yolo .txt files and the segmentation masks. - - -Expected YAML format (Example): -=============================== - -Please refer to the TORSO-21 documentation for this: https://github.com/bit-bots/TORSO_21_dataset#structure - - -Expects following file tree (Example): -====================================== - -We expect to be given a subdirectory of the structure documented here: https://github.com/bit-bots/TORSO_21_dataset#structure - - # TORSO-21 -> reality|simulation -> train|test -├── annotations.yaml -├── images -│ ├── image1.jpg -│ ├── image2.png -│ └── ... -└── segmentations - ├── image1.png - ├── image2.png - └── ... - -Produces the following file tree (Example): -=========================================== - - # TORSO-21 -> reality|simulation -> train|test -├── train.txt -├── test.txt -├── yoeo.names -├── yoeo.data -├── images # Images already exist in dataset; symlinks are created in destination-dir case -│ ├── image1.jpg -│ ├── image2.png -│ └── ... -├── labels -│ ├── image1.txt -│ ├── image2.txt -│ └── ... -└── segmentations - ├── image1.png - ├── image2.png - └── ... - -with train.txt and test.txt containing absolute image-paths for training and evaluation respectively -with yoeo.names containing the class names of bounding boxes -with yoeo.data: containing number of bounding box classes as well as absolute path to train.txt, test.txt and yoeo.names -""" - - -def range_limited_float_type_0_to_1(arg): - """Type function for argparse - a float within some predefined bounds - Derived from 'https://stackoverflow.com/questions/55324449/how-to-specify-a-minimum-or-maximum-float-value-with-argparse/55410582#55410582'. - """ - minimum = 0.0 - maximum = 1.0 - try: - f = float(arg) - except ValueError: - raise argparse.ArgumentTypeError("Must be a floating point number") - if f < minimum or f > maximum: - raise argparse.ArgumentTypeError(f"Argument must be between {minimum} and {maximum}") - return f - - -parser = argparse.ArgumentParser(description="Create YOEO labels from yaml files.") -parser.add_argument("dataset_dir", type=str, help="Directory to a dataset. Output will be written here, unless --destination-dir is given.") -parser.add_argument("annotation_file", type=str, help="Full path of annotation file") -parser.add_argument("testsplit", type=range_limited_float_type_0_to_1, help="Amount of test images from total images: train/test split (between 0.0 and 1.0)") -parser.add_argument("-s", "--seed", type=int, default=random.randint(0, (2**64)-1), help="Seed, that controls the train/test split (integer)") -parser.add_argument("--destination-dir", type=str, default="", help="Writes output files to specified directory.") -parser.add_argument("--create-symlinks", action="store_true", help="Create symlinks for image files to destination-dir. Useful, when using read-only datasets. Requires --destination-dir") -parser.add_argument("--ignore-blurred", action="store_true", help="Ignore blurred labels") -parser.add_argument("--ignore-concealed", action="store_true", help="Ignore concealed labels") -parser.add_argument("--ignore-classes", nargs="+", default=[], help="Append class names, to be ignored") -args = parser.parse_args() - -# Remove ignored classes from CLASSES list -for ignore_class in args.ignore_classes: - for category in CLASSES.keys(): - if ignore_class in CLASSES[category]: - CLASSES[category].remove(ignore_class) - print(f"Ignoring class '{ignore_class}'") - -# Defaults -create_symlinks = False -dataset_dir = args.dataset_dir -destination_dir = args.dataset_dir -image_names = [] # Collect image paths for train/test split - -# Overwrite defaults, if destination path is given -if args.destination_dir: - create_symlinks = args.create_symlinks - destination_dir = args.destination_dir - -# Create output directories if needed -images_dir = os.path.join(destination_dir, "images") -if not os.path.exists(images_dir): - os.makedirs(images_dir) - -labels_dir = os.path.join(destination_dir, "labels") -if not os.path.exists(labels_dir): - os.makedirs(labels_dir) - -masks_dir = os.path.join(destination_dir, "segmentations") -if not os.path.exists(masks_dir): - os.makedirs(masks_dir) - -# Load annotation data from yaml file -annotations_file = args.annotation_file #os.path.join(dataset_dir, "annotations.yaml") -with open(annotations_file) as f: - export = yaml.safe_load(f) - -for img_name, frame in export['images'].items(): - image_names.append(img_name) # Collect image names - - # Generate segmentations in correct format - seg_path = os.path.join(dataset_dir, "segmentations", os.path.splitext(img_name)[0] + ".png") - seg_in = cv2.imread(seg_path) - if seg_in is not None: - mask = np.zeros(seg_in.shape[:2], dtype=np.uint8) - mask += ((seg_in == (127, 127, 127)).all(axis=2)).astype(np.uint8) # Lines - mask += (((seg_in == (254, 254, 254)).all(axis=2)).astype(np.uint8) * 2) # Field - seg_out = np.zeros(seg_in.shape, dtype=np.uint8) - seg_out[..., 0] = mask - seg_out[..., 1] = mask - seg_out[..., 2] = mask - cv2.imwrite(os.path.join(masks_dir, os.path.splitext(img_name)[0] + ".png"), seg_out) - else: - print(f"No segmentation found: '{seg_path}'") - continue - - name = os.path.splitext(img_name)[0] # Remove file extension - imgwidth = frame['width'] - imgheight = frame['height'] - annotations = [] - - for annotation in frame['annotations']: - # Ignore if blurred or concealed and should be ignored - if not ((args.ignore_blurred and annotation['blurred']) or - (args.ignore_concealed and annotation['concealed'])): - - if annotation['type'] in CLASSES['bb_classes']: # Handle bounding boxes - if annotation['in_image']: - min_x = min(map(lambda x: x[0], annotation['vector'])) - max_x = max(map(lambda x: x[0], annotation['vector'])) - min_y = min(map(lambda x: x[1], annotation['vector'])) - max_y = max(map(lambda x: x[1], annotation['vector'])) - - annowidth = max_x - min_x - annoheight = max_y - min_y - relannowidth = annowidth / imgwidth - relannoheight = annoheight / imgheight - - center_x = min_x + (annowidth / 2) - center_y = min_y + (annoheight / 2) - relcenter_x = center_x / imgwidth - relcenter_y = center_y / imgheight - - if annotation['type'] != "robot": - classID = CLASSES['bb_classes'].index(annotation['type']) # Derive classID from index in predefined classes - else: - if annotation["number"] is None: - number = "None" - else: - number = str(annotation["number"]) - classID = ROBOT_COLOR_NUMBERS.index(f"{annotation['color']}_{number}") + 1 - annotations.append(f"{classID} {relcenter_x} {relcenter_y} {relannowidth} {relannoheight}") # Append to store it later - else: # Annotation is not in image - continue - elif annotation['type'] in CLASSES['segmentation_classes']: # Handle segmentations - continue - elif annotation['type'] in CLASSES['ignored_classes']: # Ignore this annotation - continue - else: - print(f"The annotation type '{annotation['type']}' is not supported or should be ignored. Image: '{img_name}'") - - # Store BB annotations in .txt file - with open(os.path.join(labels_dir, name + ".txt"), "w") as output: - output.writelines([annotation + "\n" for annotation in annotations]) - -# Create symlinks for images to destination directory -# This is helpful, if dataset directory is read-only -if create_symlinks: - for image_name in image_names: - link_path = os.path.join(images_dir, image_name) - target_path = os.path.join(dataset_dir, "images", image_name) - os.symlink(target_path, link_path) - -# Seed is used for train/test split -random.seed(args.seed) -print(f"Using seed: {args.seed}") - -# Generate train/testsplit of images -random.shuffle(sorted(image_names)) # Sort for consistent order then shuffle with seed -train_images = image_names[0:round(len(image_names) * (1 - args.testsplit))] # Split first range -test_images = image_names[round(len(image_names) * (1 - args.testsplit)) + 1:-1] # Split last range - -# Generate meta files -train_images = set(train_images) # Prevent images from showing up twice -train_path = os.path.join(destination_dir, "train.txt") -with open(train_path, "w") as train_file: - train_file.writelines([str(os.path.join(destination_dir, image_name)) + "\n" for image_name in train_images]) - -test_images = set(test_images) # Prevent images from showing up twice -test_path = os.path.join(destination_dir, "test.txt") -with open(test_path, "w") as test_file: - test_file.writelines([str(os.path.join(destination_dir, image_name)) + "\n" for image_name in test_images]) - -names_path = os.path.join(destination_dir, "yoeo.names") -with open(names_path, "w") as names_file: - names_file.writelines([class_name + "\n" for class_name in CLASSES['bb_classes']]) - -data_path = os.path.join(destination_dir, "yoeo.data") -with open(data_path, "w") as data_file: - data_file.write(f"train={train_path}\n") - data_file.write(f"valid={test_path}\n") - data_file.write(f"names={names_path}\n")