diff --git a/extensions-builtin/sdnext-modernui b/extensions-builtin/sdnext-modernui index a9300638a..c1c7dfda4 160000 --- a/extensions-builtin/sdnext-modernui +++ b/extensions-builtin/sdnext-modernui @@ -1 +1 @@ -Subproject commit a9300638a86f77be92d80a7f18dae40ecd5c4a81 +Subproject commit c1c7dfda4a2bd75f8f1924046135b11a73eec459 diff --git a/javascript/ibmplexmono-nerdfont-medium.ttf b/javascript/ibmplexmono-nerdfont-medium.ttf new file mode 100644 index 000000000..39f178db7 Binary files /dev/null and b/javascript/ibmplexmono-nerdfont-medium.ttf differ diff --git a/modules/images.py b/modules/images.py index 4e294d7a0..6b3bba16a 100644 --- a/modules/images.py +++ b/modules/images.py @@ -242,7 +242,7 @@ def parse_comfy_metadata(data: dict): version = dct.get('extra', {}).get('frontendVersion', 'unknown') if version is not None: res = f" | Version: {version} | Nodes: {nodes}" - except: + except Exception: pass return res @@ -257,7 +257,7 @@ def parse_comfy_metadata(data: dict): model = inp.get('model', None) if isinstance(model, str) and len(model) > 0: res += f" | Model: {model} | Class: {val.get('class_type', '')}" - except: + except Exception: pass return res @@ -280,7 +280,7 @@ def parse_invoke_metadata(data: dict): version = dct['app_version'] if isinstance(version, str) and len(version) > 0: res += f" | Version: {version}" - except: + except Exception: pass return res @@ -299,7 +299,7 @@ def parse_novelai_metadata(data: dict): dct = json.loads(data["Comment"]) sampler = sd_samplers.samplers_map.get(dct["sampler"], "Euler a") geninfo = f'{data["Description"]} Negative prompt: {dct["uc"]} Steps: {dct["steps"]}, Sampler: {sampler}, CFG scale: {dct["scale"]}, Seed: {dct["seed"]}, Clip skip: 2, ENSD: 31337' - except Exception as e: + except Exception: pass return geninfo diff --git a/modules/prompt_parser_diffusers.py b/modules/prompt_parser_diffusers.py index e72099b98..5e087dd9a 100644 --- a/modules/prompt_parser_diffusers.py +++ b/modules/prompt_parser_diffusers.py @@ -6,7 +6,7 @@ from collections import OrderedDict import torch from compel.embeddings_provider import BaseTextualInversionManager, EmbeddingsProvider from transformers import PreTrainedTokenizer -from modules import shared, prompt_parser, devices, sd_models, errors +from modules import shared, prompt_parser, devices, sd_models from modules.prompt_parser_xhinker import get_weighted_text_embeddings_sd15, get_weighted_text_embeddings_sdxl_2p, get_weighted_text_embeddings_sd3, get_weighted_text_embeddings_flux1, get_weighted_text_embeddings_chroma debug_enabled = os.environ.get('SD_PROMPT_DEBUG', None) @@ -253,7 +253,6 @@ class PromptEmbedder: return torch.cat(res) except Exception as e: shared.log.error(f"Prompt encode: {e}") - # errors.display(e, 'encode') return None diff --git a/wiki b/wiki index 9b30cf154..8df9dbd32 160000 --- a/wiki +++ b/wiki @@ -1 +1 @@ -Subproject commit 9b30cf154e3f715da69a8317511465fe8b7bf13f +Subproject commit 8df9dbd32e89d51698b73fc00ca6f649be35b39d