| | import torch |
| | from modules import devices, shared |
| |
|
| | module_in_gpu = None |
| | cpu = torch.device("cpu") |
| |
|
| |
|
| | def send_everything_to_cpu(): |
| | global module_in_gpu |
| |
|
| | if module_in_gpu is not None: |
| | module_in_gpu.to(cpu) |
| |
|
| | module_in_gpu = None |
| |
|
| |
|
| | def is_needed(sd_model): |
| | return shared.cmd_opts.lowvram or shared.cmd_opts.medvram or shared.cmd_opts.medvram_sdxl and hasattr(sd_model, 'conditioner') |
| |
|
| |
|
| | def apply(sd_model): |
| | enable = is_needed(sd_model) |
| | shared.parallel_processing_allowed = not enable |
| |
|
| | if enable: |
| | setup_for_low_vram(sd_model, not shared.cmd_opts.lowvram) |
| | else: |
| | sd_model.lowvram = False |
| |
|
| |
|
| | def setup_for_low_vram(sd_model, use_medvram): |
| | if getattr(sd_model, 'lowvram', False): |
| | return |
| |
|
| | sd_model.lowvram = True |
| |
|
| | parents = {} |
| |
|
| | def send_me_to_gpu(module, _): |
| | """send this module to GPU; send whatever tracked module was previous in GPU to CPU; |
| | we add this as forward_pre_hook to a lot of modules and this way all but one of them will |
| | be in CPU |
| | """ |
| | global module_in_gpu |
| |
|
| | module = parents.get(module, module) |
| |
|
| | if module_in_gpu == module: |
| | return |
| |
|
| | if module_in_gpu is not None: |
| | module_in_gpu.to(cpu) |
| |
|
| | module.to(devices.device) |
| | module_in_gpu = module |
| |
|
| | |
| | |
| | |
| |
|
| | first_stage_model = sd_model.first_stage_model |
| | first_stage_model_encode = sd_model.first_stage_model.encode |
| | first_stage_model_decode = sd_model.first_stage_model.decode |
| |
|
| | def first_stage_model_encode_wrap(x): |
| | send_me_to_gpu(first_stage_model, None) |
| | return first_stage_model_encode(x) |
| |
|
| | def first_stage_model_decode_wrap(z): |
| | send_me_to_gpu(first_stage_model, None) |
| | return first_stage_model_decode(z) |
| |
|
| | to_remain_in_cpu = [ |
| | (sd_model, 'first_stage_model'), |
| | (sd_model, 'depth_model'), |
| | (sd_model, 'embedder'), |
| | (sd_model, 'model'), |
| | (sd_model, 'embedder'), |
| | ] |
| |
|
| | is_sdxl = hasattr(sd_model, 'conditioner') |
| | is_sd2 = not is_sdxl and hasattr(sd_model.cond_stage_model, 'model') |
| |
|
| | if is_sdxl: |
| | to_remain_in_cpu.append((sd_model, 'conditioner')) |
| | elif is_sd2: |
| | to_remain_in_cpu.append((sd_model.cond_stage_model, 'model')) |
| | else: |
| | to_remain_in_cpu.append((sd_model.cond_stage_model, 'transformer')) |
| |
|
| | |
| | stored = [] |
| | for obj, field in to_remain_in_cpu: |
| | module = getattr(obj, field, None) |
| | stored.append(module) |
| | setattr(obj, field, None) |
| |
|
| | |
| | sd_model.to(devices.device) |
| |
|
| | |
| | for (obj, field), module in zip(to_remain_in_cpu, stored): |
| | setattr(obj, field, module) |
| |
|
| | |
| | if is_sdxl: |
| | sd_model.conditioner.register_forward_pre_hook(send_me_to_gpu) |
| | elif is_sd2: |
| | sd_model.cond_stage_model.model.register_forward_pre_hook(send_me_to_gpu) |
| | sd_model.cond_stage_model.model.token_embedding.register_forward_pre_hook(send_me_to_gpu) |
| | parents[sd_model.cond_stage_model.model] = sd_model.cond_stage_model |
| | parents[sd_model.cond_stage_model.model.token_embedding] = sd_model.cond_stage_model |
| | else: |
| | sd_model.cond_stage_model.transformer.register_forward_pre_hook(send_me_to_gpu) |
| | parents[sd_model.cond_stage_model.transformer] = sd_model.cond_stage_model |
| |
|
| | sd_model.first_stage_model.register_forward_pre_hook(send_me_to_gpu) |
| | sd_model.first_stage_model.encode = first_stage_model_encode_wrap |
| | sd_model.first_stage_model.decode = first_stage_model_decode_wrap |
| | if sd_model.depth_model: |
| | sd_model.depth_model.register_forward_pre_hook(send_me_to_gpu) |
| | if sd_model.embedder: |
| | sd_model.embedder.register_forward_pre_hook(send_me_to_gpu) |
| |
|
| | if use_medvram: |
| | sd_model.model.register_forward_pre_hook(send_me_to_gpu) |
| | else: |
| | diff_model = sd_model.model.diffusion_model |
| |
|
| | |
| | |
| | stored = diff_model.input_blocks, diff_model.middle_block, diff_model.output_blocks, diff_model.time_embed |
| | diff_model.input_blocks, diff_model.middle_block, diff_model.output_blocks, diff_model.time_embed = None, None, None, None |
| | sd_model.model.to(devices.device) |
| | diff_model.input_blocks, diff_model.middle_block, diff_model.output_blocks, diff_model.time_embed = stored |
| |
|
| | |
| | diff_model.time_embed.register_forward_pre_hook(send_me_to_gpu) |
| | for block in diff_model.input_blocks: |
| | block.register_forward_pre_hook(send_me_to_gpu) |
| | diff_model.middle_block.register_forward_pre_hook(send_me_to_gpu) |
| | for block in diff_model.output_blocks: |
| | block.register_forward_pre_hook(send_me_to_gpu) |
| |
|
| |
|
| | def is_enabled(sd_model): |
| | return sd_model.lowvram |
| |
|