mirror of
https://github.com/AUTOMATIC1111/stable-diffusion-webui.git
synced 2024-11-21 03:11:40 +08:00
50 lines
1.9 KiB
Python
50 lines
1.9 KiB
Python
import os
|
|
|
|
import torch
|
|
|
|
from modules import shared
|
|
from modules.shared import cmd_opts
|
|
|
|
|
|
def initialize():
|
|
"""Initializes fields inside the shared module in a controlled manner.
|
|
|
|
Should be called early because some other modules you can import mingt need these fields to be already set.
|
|
"""
|
|
|
|
os.makedirs(cmd_opts.hypernetwork_dir, exist_ok=True)
|
|
|
|
from modules import options, shared_options
|
|
shared.options_templates = shared_options.options_templates
|
|
shared.opts = options.Options(shared_options.options_templates, shared_options.restricted_opts)
|
|
shared.restricted_opts = shared_options.restricted_opts
|
|
if os.path.exists(shared.config_filename):
|
|
shared.opts.load(shared.config_filename)
|
|
|
|
from modules import devices
|
|
devices.device, devices.device_interrogate, devices.device_gfpgan, devices.device_esrgan, devices.device_codeformer = \
|
|
(devices.cpu if any(y in cmd_opts.use_cpu for y in [x, 'all']) else devices.get_optimal_device() for x in ['sd', 'interrogate', 'gfpgan', 'esrgan', 'codeformer'])
|
|
|
|
devices.dtype = torch.float32 if cmd_opts.no_half else torch.float16
|
|
devices.dtype_vae = torch.float32 if cmd_opts.no_half or cmd_opts.no_half_vae else torch.float16
|
|
|
|
shared.device = devices.device
|
|
shared.weight_load_location = None if cmd_opts.lowram else "cpu"
|
|
|
|
from modules import shared_state
|
|
shared.state = shared_state.State()
|
|
|
|
from modules import styles
|
|
shared.prompt_styles = styles.StyleDatabase(shared.styles_filename)
|
|
|
|
from modules import interrogate
|
|
shared.interrogator = interrogate.InterrogateModels("interrogate")
|
|
|
|
from modules import shared_total_tqdm
|
|
shared.total_tqdm = shared_total_tqdm.TotalTQDM()
|
|
|
|
from modules import memmon, devices
|
|
shared.mem_mon = memmon.MemUsageMonitor("MemMon", devices.device, shared.opts)
|
|
shared.mem_mon.start()
|
|
|