Merge pull request #4556 from CalamitousFelicitousness/feat/f2preq-ref

feat(flux2-klein): add SDNQ pre-quantized model support and reference images
pull/4566/head
Vladimir Mandic 2026-01-18 07:14:23 +01:00 committed by GitHub
commit 6f5e576ea7
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
3 changed files with 10 additions and 9 deletions

View File

@ -168,7 +168,7 @@
"desc": "FLUX.2-klein-4B is a 4 billion parameter size-distilled version of FLUX.2-dev optimized for consumer GPUs. Achieves sub-second inference with 4 steps. Supports both text-to-image generation and multi-reference image editing. Apache 2.0 licensed.",
"skip": true,
"tags": "distilled",
"extras": "sampler: Default, cfg_scale: 4.0, steps: 4",
"extras": "sampler: Default, cfg_scale: 1.0, steps: 4",
"size": 8.5,
"date": "2025 January"
},
@ -178,7 +178,7 @@
"desc": "FLUX.2-klein-9B is a 9 billion parameter size-distilled version of FLUX.2-dev. Higher quality than 4B variant with sub-second inference using 4 steps. Supports text-to-image and multi-reference editing. Non-commercial license.",
"skip": true,
"tags": "distilled",
"extras": "sampler: Default, cfg_scale: 4.0, steps: 4",
"extras": "sampler: Default, cfg_scale: 1.0, steps: 4",
"size": 18.5,
"date": "2025 January"
}

View File

@ -51,20 +51,20 @@
},
"Black Forest Labs FLUX.2 Klein 4B sdnq-uint4-dynamic": {
"path": "Disty0/FLUX.2-klein-4B-SDNQ-4bit-dynamic",
"preview": "black-forest-labs--FLUX.2-klein-base-4B.jpg",
"desc": "Dynamic 4 bit quantization of black-forest-labs/FLUX.2-klein-9B using SDNQ.",
"preview": "Disty0--FLUX.2-klein-4B-SDNQ-4bit-dynamic.jpg",
"desc": "Dynamic 4-bit quantization of black-forest-labs/FLUX.2-klein-4B using SDNQ.",
"skip": true,
"extras": "sampler: Default, cfg_scale: 4.0, steps: 50",
"extras": "sampler: Default, cfg_scale: 1.0, steps: 4",
"tags": "quantized",
"size": 5.1,
"date": "2025 January"
},
"Black Forest Labs FLUX.2 Klein 9B sdnq-uint4-dynamic-svd": {
"path": "Disty0/FLUX.2-klein-9B-SDNQ-4bit-dynamic-svd-r32",
"preview": "black-forest-labs--FLUX.2-klein-base-9B.jpg",
"desc": "FLUX.2-klein-base-9B is the undistilled 9 billion parameter base model of FLUX.2-klein. Requires 50 inference steps for full quality but offers flexibility for fine-tuning. Supports text-to-image and multi-reference editing. Non-commercial license.",
"preview": "Disty0--FLUX.2-klein-9B-SDNQ-4bit-dynamic-svd-r32.jpg",
"desc": "Dynamic 4-bit quantization of black-forest-labs/FLUX.2-klein-9B using SDNQ with SVD rank 32.",
"skip": true,
"extras": "sampler: Default, cfg_scale: 4.0, steps: 50",
"extras": "sampler: Default, cfg_scale: 1.0, steps: 4",
"tags": "quantized",
"size": 11.7,
"date": "2025 January"

View File

@ -203,7 +203,8 @@ def load_text_encoder(repo_id, cls_name, load_config=None, subfolder="text_encod
# Qwen3ForCausalLM - shared text encoders by hidden_size:
# - Z-Image, Klein-4B: Qwen3-4B (hidden_size=2560)
# - Klein-9B: Qwen3-8B (hidden_size=4096)
elif cls_name == transformers.Qwen3ForCausalLM and allow_shared and shared.opts.te_shared_t5:
# SDNQ repos for Klein and Z-Image contain text encoders pre-quantized with different quantization methods, skip shared loading
elif cls_name == transformers.Qwen3ForCausalLM and allow_shared and shared.opts.te_shared_t5 and 'sdnq' not in repo_id.lower():
if '-9b' in repo_id.lower():
shared_repo = 'black-forest-labs/FLUX.2-klein-9B' # 9B variants use Qwen3-8B
else: