import os DEFAULT_MODEL = "lama" AVAILABLE_MODELS = [ "lama", "ldm", "zits", "mat", "fcf", "sd1.5", "cv2", "manga", "sd2", "paint_by_example" ] AVAILABLE_DEVICES = ["cuda", "cpu", "mps"] DEFAULT_DEVICE = 'cuda' NO_HALF_HELP = """ Using full precision model. If your generate result is always black or green, use this argument. (sd/paint_by_exmaple) """ CPU_OFFLOAD_HELP = """ Offloads all models to CPU, significantly reducing vRAM usage. (sd/paint_by_example) """ DISABLE_NSFW_HELP = """ Disable NSFW checker. (sd/paint_by_example) """ SD_CPU_TEXTENCODER_HELP = """ Run Stable Diffusion text encoder model on CPU to save GPU memory. """ LOCAL_FILES_ONLY_HELP = """ Use local files only, not connect to Hugging Face server. (sd/paint_by_example) """ ENABLE_XFORMERS_HELP = """ Enable xFormers optimizations. Requires xformers package has been installed. See: https://github.com/facebookresearch/xformers (sd/paint_by_example) """ DEFAULT_MODEL_DIR = os.getenv( "XDG_CACHE_HOME", os.path.join(os.path.expanduser("~"), ".cache") ) MODEL_DIR_HELP = """ Model download directory (by setting XDG_CACHE_HOME environment variable), by default model downloaded to ~/.cache """ OUTPUT_DIR_HELP = """ Only required when --input is directory. Result images will be saved to output directory automatically. """ INPUT_HELP = """ If input is image, it will be loaded by default. If input is directory, you can browse and select image in file manager. """ GUI_HELP = """ Launch Lama Cleaner as desktop app """ NO_GUI_AUTO_CLOSE_HELP = """ Prevent backend auto close after the GUI window closed. """