mirror of https://github.com/vladmandic/automatic
Merge pull request #4556 from CalamitousFelicitousness/feat/f2preq-ref
feat(flux2-klein): add SDNQ pre-quantized model support and reference imagespull/4566/head
commit
6f5e576ea7
|
|
@ -168,7 +168,7 @@
|
|||
"desc": "FLUX.2-klein-4B is a 4 billion parameter size-distilled version of FLUX.2-dev optimized for consumer GPUs. Achieves sub-second inference with 4 steps. Supports both text-to-image generation and multi-reference image editing. Apache 2.0 licensed.",
|
||||
"skip": true,
|
||||
"tags": "distilled",
|
||||
"extras": "sampler: Default, cfg_scale: 4.0, steps: 4",
|
||||
"extras": "sampler: Default, cfg_scale: 1.0, steps: 4",
|
||||
"size": 8.5,
|
||||
"date": "2025 January"
|
||||
},
|
||||
|
|
@ -178,7 +178,7 @@
|
|||
"desc": "FLUX.2-klein-9B is a 9 billion parameter size-distilled version of FLUX.2-dev. Higher quality than 4B variant with sub-second inference using 4 steps. Supports text-to-image and multi-reference editing. Non-commercial license.",
|
||||
"skip": true,
|
||||
"tags": "distilled",
|
||||
"extras": "sampler: Default, cfg_scale: 4.0, steps: 4",
|
||||
"extras": "sampler: Default, cfg_scale: 1.0, steps: 4",
|
||||
"size": 18.5,
|
||||
"date": "2025 January"
|
||||
}
|
||||
|
|
|
|||
|
|
@ -51,20 +51,20 @@
|
|||
},
|
||||
"Black Forest Labs FLUX.2 Klein 4B sdnq-uint4-dynamic": {
|
||||
"path": "Disty0/FLUX.2-klein-4B-SDNQ-4bit-dynamic",
|
||||
"preview": "black-forest-labs--FLUX.2-klein-base-4B.jpg",
|
||||
"desc": "Dynamic 4 bit quantization of black-forest-labs/FLUX.2-klein-9B using SDNQ.",
|
||||
"preview": "Disty0--FLUX.2-klein-4B-SDNQ-4bit-dynamic.jpg",
|
||||
"desc": "Dynamic 4-bit quantization of black-forest-labs/FLUX.2-klein-4B using SDNQ.",
|
||||
"skip": true,
|
||||
"extras": "sampler: Default, cfg_scale: 4.0, steps: 50",
|
||||
"extras": "sampler: Default, cfg_scale: 1.0, steps: 4",
|
||||
"tags": "quantized",
|
||||
"size": 5.1,
|
||||
"date": "2025 January"
|
||||
},
|
||||
"Black Forest Labs FLUX.2 Klein 9B sdnq-uint4-dynamic-svd": {
|
||||
"path": "Disty0/FLUX.2-klein-9B-SDNQ-4bit-dynamic-svd-r32",
|
||||
"preview": "black-forest-labs--FLUX.2-klein-base-9B.jpg",
|
||||
"desc": "FLUX.2-klein-base-9B is the undistilled 9 billion parameter base model of FLUX.2-klein. Requires 50 inference steps for full quality but offers flexibility for fine-tuning. Supports text-to-image and multi-reference editing. Non-commercial license.",
|
||||
"preview": "Disty0--FLUX.2-klein-9B-SDNQ-4bit-dynamic-svd-r32.jpg",
|
||||
"desc": "Dynamic 4-bit quantization of black-forest-labs/FLUX.2-klein-9B using SDNQ with SVD rank 32.",
|
||||
"skip": true,
|
||||
"extras": "sampler: Default, cfg_scale: 4.0, steps: 50",
|
||||
"extras": "sampler: Default, cfg_scale: 1.0, steps: 4",
|
||||
"tags": "quantized",
|
||||
"size": 11.7,
|
||||
"date": "2025 January"
|
||||
|
|
|
|||
|
|
@ -203,7 +203,8 @@ def load_text_encoder(repo_id, cls_name, load_config=None, subfolder="text_encod
|
|||
# Qwen3ForCausalLM - shared text encoders by hidden_size:
|
||||
# - Z-Image, Klein-4B: Qwen3-4B (hidden_size=2560)
|
||||
# - Klein-9B: Qwen3-8B (hidden_size=4096)
|
||||
elif cls_name == transformers.Qwen3ForCausalLM and allow_shared and shared.opts.te_shared_t5:
|
||||
# SDNQ repos for Klein and Z-Image contain text encoders pre-quantized with different quantization methods, skip shared loading
|
||||
elif cls_name == transformers.Qwen3ForCausalLM and allow_shared and shared.opts.te_shared_t5 and 'sdnq' not in repo_id.lower():
|
||||
if '-9b' in repo_id.lower():
|
||||
shared_repo = 'black-forest-labs/FLUX.2-klein-9B' # 9B variants use Qwen3-8B
|
||||
else:
|
||||
|
|
|
|||
Loading…
Reference in New Issue