| | import os |
| | from PIL import Image, ImageOps |
| | import math |
| | import tqdm |
| |
|
| | from modules import paths, shared, images, deepbooru |
| | from modules.textual_inversion import autocrop |
| |
|
| |
|
| | def preprocess(id_task, process_src, process_dst, process_width, process_height, preprocess_txt_action, process_keep_original_size, process_flip, process_split, process_caption, process_caption_deepbooru=False, split_threshold=0.5, overlap_ratio=0.2, process_focal_crop=False, process_focal_crop_face_weight=0.9, process_focal_crop_entropy_weight=0.15, process_focal_crop_edges_weight=0.5, process_focal_crop_debug=False, process_multicrop=None, process_multicrop_mindim=None, process_multicrop_maxdim=None, process_multicrop_minarea=None, process_multicrop_maxarea=None, process_multicrop_objective=None, process_multicrop_threshold=None): |
| | try: |
| | if process_caption: |
| | shared.interrogator.load() |
| |
|
| | if process_caption_deepbooru: |
| | deepbooru.model.start() |
| |
|
| | preprocess_work(process_src, process_dst, process_width, process_height, preprocess_txt_action, process_keep_original_size, process_flip, process_split, process_caption, process_caption_deepbooru, split_threshold, overlap_ratio, process_focal_crop, process_focal_crop_face_weight, process_focal_crop_entropy_weight, process_focal_crop_edges_weight, process_focal_crop_debug, process_multicrop, process_multicrop_mindim, process_multicrop_maxdim, process_multicrop_minarea, process_multicrop_maxarea, process_multicrop_objective, process_multicrop_threshold) |
| |
|
| | finally: |
| |
|
| | if process_caption: |
| | shared.interrogator.send_blip_to_ram() |
| |
|
| | if process_caption_deepbooru: |
| | deepbooru.model.stop() |
| |
|
| |
|
| | def listfiles(dirname): |
| | return os.listdir(dirname) |
| |
|
| |
|
| | class PreprocessParams: |
| | src = None |
| | dstdir = None |
| | subindex = 0 |
| | flip = False |
| | process_caption = False |
| | process_caption_deepbooru = False |
| | preprocess_txt_action = None |
| |
|
| |
|
| | def save_pic_with_caption(image, index, params: PreprocessParams, existing_caption=None): |
| | caption = "" |
| |
|
| | if params.process_caption: |
| | caption += shared.interrogator.generate_caption(image) |
| |
|
| | if params.process_caption_deepbooru: |
| | if caption: |
| | caption += ", " |
| | caption += deepbooru.model.tag_multi(image) |
| |
|
| | filename_part = params.src |
| | filename_part = os.path.splitext(filename_part)[0] |
| | filename_part = os.path.basename(filename_part) |
| |
|
| | basename = f"{index:05}-{params.subindex}-{filename_part}" |
| | image.save(os.path.join(params.dstdir, f"{basename}.png")) |
| |
|
| | if params.preprocess_txt_action == 'prepend' and existing_caption: |
| | caption = f"{existing_caption} {caption}" |
| | elif params.preprocess_txt_action == 'append' and existing_caption: |
| | caption = f"{caption} {existing_caption}" |
| | elif params.preprocess_txt_action == 'copy' and existing_caption: |
| | caption = existing_caption |
| |
|
| | caption = caption.strip() |
| |
|
| | if caption: |
| | with open(os.path.join(params.dstdir, f"{basename}.txt"), "w", encoding="utf8") as file: |
| | file.write(caption) |
| |
|
| | params.subindex += 1 |
| |
|
| |
|
| | def save_pic(image, index, params, existing_caption=None): |
| | save_pic_with_caption(image, index, params, existing_caption=existing_caption) |
| |
|
| | if params.flip: |
| | save_pic_with_caption(ImageOps.mirror(image), index, params, existing_caption=existing_caption) |
| |
|
| |
|
| | def split_pic(image, inverse_xy, width, height, overlap_ratio): |
| | if inverse_xy: |
| | from_w, from_h = image.height, image.width |
| | to_w, to_h = height, width |
| | else: |
| | from_w, from_h = image.width, image.height |
| | to_w, to_h = width, height |
| | h = from_h * to_w // from_w |
| | if inverse_xy: |
| | image = image.resize((h, to_w)) |
| | else: |
| | image = image.resize((to_w, h)) |
| |
|
| | split_count = math.ceil((h - to_h * overlap_ratio) / (to_h * (1.0 - overlap_ratio))) |
| | y_step = (h - to_h) / (split_count - 1) |
| | for i in range(split_count): |
| | y = int(y_step * i) |
| | if inverse_xy: |
| | splitted = image.crop((y, 0, y + to_h, to_w)) |
| | else: |
| | splitted = image.crop((0, y, to_w, y + to_h)) |
| | yield splitted |
| |
|
| | |
| | def center_crop(image: Image, w: int, h: int): |
| | iw, ih = image.size |
| | if ih / h < iw / w: |
| | sw = w * ih / h |
| | box = (iw - sw) / 2, 0, iw - (iw - sw) / 2, ih |
| | else: |
| | sh = h * iw / w |
| | box = 0, (ih - sh) / 2, iw, ih - (ih - sh) / 2 |
| | return image.resize((w, h), Image.Resampling.LANCZOS, box) |
| |
|
| |
|
| | def multicrop_pic(image: Image, mindim, maxdim, minarea, maxarea, objective, threshold): |
| | iw, ih = image.size |
| | err = lambda w, h: 1-(lambda x: x if x < 1 else 1/x)(iw/ih/(w/h)) |
| | wh = max(((w, h) for w in range(mindim, maxdim+1, 64) for h in range(mindim, maxdim+1, 64) |
| | if minarea <= w * h <= maxarea and err(w, h) <= threshold), |
| | key= lambda wh: (wh[0]*wh[1], -err(*wh))[::1 if objective=='Maximize area' else -1], |
| | default=None |
| | ) |
| | return wh and center_crop(image, *wh) |
| |
|
| |
|
| | def preprocess_work(process_src, process_dst, process_width, process_height, preprocess_txt_action, process_keep_original_size, process_flip, process_split, process_caption, process_caption_deepbooru=False, split_threshold=0.5, overlap_ratio=0.2, process_focal_crop=False, process_focal_crop_face_weight=0.9, process_focal_crop_entropy_weight=0.3, process_focal_crop_edges_weight=0.5, process_focal_crop_debug=False, process_multicrop=None, process_multicrop_mindim=None, process_multicrop_maxdim=None, process_multicrop_minarea=None, process_multicrop_maxarea=None, process_multicrop_objective=None, process_multicrop_threshold=None): |
| | width = process_width |
| | height = process_height |
| | src = os.path.abspath(process_src) |
| | dst = os.path.abspath(process_dst) |
| | split_threshold = max(0.0, min(1.0, split_threshold)) |
| | overlap_ratio = max(0.0, min(0.9, overlap_ratio)) |
| |
|
| | assert src != dst, 'same directory specified as source and destination' |
| |
|
| | os.makedirs(dst, exist_ok=True) |
| |
|
| | files = listfiles(src) |
| |
|
| | shared.state.job = "preprocess" |
| | shared.state.textinfo = "Preprocessing..." |
| | shared.state.job_count = len(files) |
| |
|
| | params = PreprocessParams() |
| | params.dstdir = dst |
| | params.flip = process_flip |
| | params.process_caption = process_caption |
| | params.process_caption_deepbooru = process_caption_deepbooru |
| | params.preprocess_txt_action = preprocess_txt_action |
| |
|
| | pbar = tqdm.tqdm(files) |
| | for index, imagefile in enumerate(pbar): |
| | params.subindex = 0 |
| | filename = os.path.join(src, imagefile) |
| | try: |
| | img = Image.open(filename) |
| | img = ImageOps.exif_transpose(img) |
| | img = img.convert("RGB") |
| | except Exception: |
| | continue |
| |
|
| | description = f"Preprocessing [Image {index}/{len(files)}]" |
| | pbar.set_description(description) |
| | shared.state.textinfo = description |
| |
|
| | params.src = filename |
| |
|
| | existing_caption = None |
| | existing_caption_filename = f"{os.path.splitext(filename)[0]}.txt" |
| | if os.path.exists(existing_caption_filename): |
| | with open(existing_caption_filename, 'r', encoding="utf8") as file: |
| | existing_caption = file.read() |
| |
|
| | if shared.state.interrupted: |
| | break |
| |
|
| | if img.height > img.width: |
| | ratio = (img.width * height) / (img.height * width) |
| | inverse_xy = False |
| | else: |
| | ratio = (img.height * width) / (img.width * height) |
| | inverse_xy = True |
| |
|
| | process_default_resize = True |
| |
|
| | if process_split and ratio < 1.0 and ratio <= split_threshold: |
| | for splitted in split_pic(img, inverse_xy, width, height, overlap_ratio): |
| | save_pic(splitted, index, params, existing_caption=existing_caption) |
| | process_default_resize = False |
| |
|
| | if process_focal_crop and img.height != img.width: |
| |
|
| | dnn_model_path = None |
| | try: |
| | dnn_model_path = autocrop.download_and_cache_models(os.path.join(paths.models_path, "opencv")) |
| | except Exception as e: |
| | print("Unable to load face detection model for auto crop selection. Falling back to lower quality haar method.", e) |
| |
|
| | autocrop_settings = autocrop.Settings( |
| | crop_width = width, |
| | crop_height = height, |
| | face_points_weight = process_focal_crop_face_weight, |
| | entropy_points_weight = process_focal_crop_entropy_weight, |
| | corner_points_weight = process_focal_crop_edges_weight, |
| | annotate_image = process_focal_crop_debug, |
| | dnn_model_path = dnn_model_path, |
| | ) |
| | for focal in autocrop.crop_image(img, autocrop_settings): |
| | save_pic(focal, index, params, existing_caption=existing_caption) |
| | process_default_resize = False |
| |
|
| | if process_multicrop: |
| | cropped = multicrop_pic(img, process_multicrop_mindim, process_multicrop_maxdim, process_multicrop_minarea, process_multicrop_maxarea, process_multicrop_objective, process_multicrop_threshold) |
| | if cropped is not None: |
| | save_pic(cropped, index, params, existing_caption=existing_caption) |
| | else: |
| | print(f"skipped {img.width}x{img.height} image {filename} (can't find suitable size within error threshold)") |
| | process_default_resize = False |
| |
|
| | if process_keep_original_size: |
| | save_pic(img, index, params, existing_caption=existing_caption) |
| | process_default_resize = False |
| |
|
| | if process_default_resize: |
| | img = images.resize_image(1, img, width, height) |
| | save_pic(img, index, params, existing_caption=existing_caption) |
| |
|
| | shared.state.nextjob() |
| |
|