stable-diffusion-webui-mode.../components/LoRA-v1A-CLIP.txt

216 lines
15 KiB
Plaintext

lora_te_text_model_encoder_layers_0_mlp_fc1.alpha []
lora_te_text_model_encoder_layers_0_mlp_fc1.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_0_mlp_fc1.lora_up.weight [3072,-1]
lora_te_text_model_encoder_layers_0_mlp_fc2.alpha []
lora_te_text_model_encoder_layers_0_mlp_fc2.lora_down.weight [-1,3072]
lora_te_text_model_encoder_layers_0_mlp_fc2.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_0_self_attn_k_proj.alpha []
lora_te_text_model_encoder_layers_0_self_attn_k_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_0_self_attn_k_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_0_self_attn_out_proj.alpha []
lora_te_text_model_encoder_layers_0_self_attn_out_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_0_self_attn_out_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_0_self_attn_q_proj.alpha []
lora_te_text_model_encoder_layers_0_self_attn_q_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_0_self_attn_q_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_0_self_attn_v_proj.alpha []
lora_te_text_model_encoder_layers_0_self_attn_v_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_0_self_attn_v_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_10_mlp_fc1.alpha []
lora_te_text_model_encoder_layers_10_mlp_fc1.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_10_mlp_fc1.lora_up.weight [3072,-1]
lora_te_text_model_encoder_layers_10_mlp_fc2.alpha []
lora_te_text_model_encoder_layers_10_mlp_fc2.lora_down.weight [-1,3072]
lora_te_text_model_encoder_layers_10_mlp_fc2.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_10_self_attn_k_proj.alpha []
lora_te_text_model_encoder_layers_10_self_attn_k_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_10_self_attn_k_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_10_self_attn_out_proj.alpha []
lora_te_text_model_encoder_layers_10_self_attn_out_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_10_self_attn_out_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_10_self_attn_q_proj.alpha []
lora_te_text_model_encoder_layers_10_self_attn_q_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_10_self_attn_q_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_10_self_attn_v_proj.alpha []
lora_te_text_model_encoder_layers_10_self_attn_v_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_10_self_attn_v_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_11_mlp_fc1.alpha []
lora_te_text_model_encoder_layers_11_mlp_fc1.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_11_mlp_fc1.lora_up.weight [3072,-1]
lora_te_text_model_encoder_layers_11_mlp_fc2.alpha []
lora_te_text_model_encoder_layers_11_mlp_fc2.lora_down.weight [-1,3072]
lora_te_text_model_encoder_layers_11_mlp_fc2.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_11_self_attn_k_proj.alpha []
lora_te_text_model_encoder_layers_11_self_attn_k_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_11_self_attn_k_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_11_self_attn_out_proj.alpha []
lora_te_text_model_encoder_layers_11_self_attn_out_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_11_self_attn_out_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_11_self_attn_q_proj.alpha []
lora_te_text_model_encoder_layers_11_self_attn_q_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_11_self_attn_q_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_11_self_attn_v_proj.alpha []
lora_te_text_model_encoder_layers_11_self_attn_v_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_11_self_attn_v_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_1_mlp_fc1.alpha []
lora_te_text_model_encoder_layers_1_mlp_fc1.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_1_mlp_fc1.lora_up.weight [3072,-1]
lora_te_text_model_encoder_layers_1_mlp_fc2.alpha []
lora_te_text_model_encoder_layers_1_mlp_fc2.lora_down.weight [-1,3072]
lora_te_text_model_encoder_layers_1_mlp_fc2.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_1_self_attn_k_proj.alpha []
lora_te_text_model_encoder_layers_1_self_attn_k_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_1_self_attn_k_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_1_self_attn_out_proj.alpha []
lora_te_text_model_encoder_layers_1_self_attn_out_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_1_self_attn_out_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_1_self_attn_q_proj.alpha []
lora_te_text_model_encoder_layers_1_self_attn_q_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_1_self_attn_q_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_1_self_attn_v_proj.alpha []
lora_te_text_model_encoder_layers_1_self_attn_v_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_1_self_attn_v_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_2_mlp_fc1.alpha []
lora_te_text_model_encoder_layers_2_mlp_fc1.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_2_mlp_fc1.lora_up.weight [3072,-1]
lora_te_text_model_encoder_layers_2_mlp_fc2.alpha []
lora_te_text_model_encoder_layers_2_mlp_fc2.lora_down.weight [-1,3072]
lora_te_text_model_encoder_layers_2_mlp_fc2.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_2_self_attn_k_proj.alpha []
lora_te_text_model_encoder_layers_2_self_attn_k_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_2_self_attn_k_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_2_self_attn_out_proj.alpha []
lora_te_text_model_encoder_layers_2_self_attn_out_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_2_self_attn_out_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_2_self_attn_q_proj.alpha []
lora_te_text_model_encoder_layers_2_self_attn_q_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_2_self_attn_q_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_2_self_attn_v_proj.alpha []
lora_te_text_model_encoder_layers_2_self_attn_v_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_2_self_attn_v_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_3_mlp_fc1.alpha []
lora_te_text_model_encoder_layers_3_mlp_fc1.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_3_mlp_fc1.lora_up.weight [3072,-1]
lora_te_text_model_encoder_layers_3_mlp_fc2.alpha []
lora_te_text_model_encoder_layers_3_mlp_fc2.lora_down.weight [-1,3072]
lora_te_text_model_encoder_layers_3_mlp_fc2.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_3_self_attn_k_proj.alpha []
lora_te_text_model_encoder_layers_3_self_attn_k_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_3_self_attn_k_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_3_self_attn_out_proj.alpha []
lora_te_text_model_encoder_layers_3_self_attn_out_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_3_self_attn_out_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_3_self_attn_q_proj.alpha []
lora_te_text_model_encoder_layers_3_self_attn_q_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_3_self_attn_q_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_3_self_attn_v_proj.alpha []
lora_te_text_model_encoder_layers_3_self_attn_v_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_3_self_attn_v_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_4_mlp_fc1.alpha []
lora_te_text_model_encoder_layers_4_mlp_fc1.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_4_mlp_fc1.lora_up.weight [3072,-1]
lora_te_text_model_encoder_layers_4_mlp_fc2.alpha []
lora_te_text_model_encoder_layers_4_mlp_fc2.lora_down.weight [-1,3072]
lora_te_text_model_encoder_layers_4_mlp_fc2.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_4_self_attn_k_proj.alpha []
lora_te_text_model_encoder_layers_4_self_attn_k_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_4_self_attn_k_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_4_self_attn_out_proj.alpha []
lora_te_text_model_encoder_layers_4_self_attn_out_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_4_self_attn_out_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_4_self_attn_q_proj.alpha []
lora_te_text_model_encoder_layers_4_self_attn_q_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_4_self_attn_q_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_4_self_attn_v_proj.alpha []
lora_te_text_model_encoder_layers_4_self_attn_v_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_4_self_attn_v_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_5_mlp_fc1.alpha []
lora_te_text_model_encoder_layers_5_mlp_fc1.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_5_mlp_fc1.lora_up.weight [3072,-1]
lora_te_text_model_encoder_layers_5_mlp_fc2.alpha []
lora_te_text_model_encoder_layers_5_mlp_fc2.lora_down.weight [-1,3072]
lora_te_text_model_encoder_layers_5_mlp_fc2.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_5_self_attn_k_proj.alpha []
lora_te_text_model_encoder_layers_5_self_attn_k_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_5_self_attn_k_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_5_self_attn_out_proj.alpha []
lora_te_text_model_encoder_layers_5_self_attn_out_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_5_self_attn_out_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_5_self_attn_q_proj.alpha []
lora_te_text_model_encoder_layers_5_self_attn_q_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_5_self_attn_q_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_5_self_attn_v_proj.alpha []
lora_te_text_model_encoder_layers_5_self_attn_v_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_5_self_attn_v_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_6_mlp_fc1.alpha []
lora_te_text_model_encoder_layers_6_mlp_fc1.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_6_mlp_fc1.lora_up.weight [3072,-1]
lora_te_text_model_encoder_layers_6_mlp_fc2.alpha []
lora_te_text_model_encoder_layers_6_mlp_fc2.lora_down.weight [-1,3072]
lora_te_text_model_encoder_layers_6_mlp_fc2.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_6_self_attn_k_proj.alpha []
lora_te_text_model_encoder_layers_6_self_attn_k_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_6_self_attn_k_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_6_self_attn_out_proj.alpha []
lora_te_text_model_encoder_layers_6_self_attn_out_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_6_self_attn_out_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_6_self_attn_q_proj.alpha []
lora_te_text_model_encoder_layers_6_self_attn_q_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_6_self_attn_q_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_6_self_attn_v_proj.alpha []
lora_te_text_model_encoder_layers_6_self_attn_v_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_6_self_attn_v_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_7_mlp_fc1.alpha []
lora_te_text_model_encoder_layers_7_mlp_fc1.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_7_mlp_fc1.lora_up.weight [3072,-1]
lora_te_text_model_encoder_layers_7_mlp_fc2.alpha []
lora_te_text_model_encoder_layers_7_mlp_fc2.lora_down.weight [-1,3072]
lora_te_text_model_encoder_layers_7_mlp_fc2.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_7_self_attn_k_proj.alpha []
lora_te_text_model_encoder_layers_7_self_attn_k_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_7_self_attn_k_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_7_self_attn_out_proj.alpha []
lora_te_text_model_encoder_layers_7_self_attn_out_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_7_self_attn_out_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_7_self_attn_q_proj.alpha []
lora_te_text_model_encoder_layers_7_self_attn_q_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_7_self_attn_q_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_7_self_attn_v_proj.alpha []
lora_te_text_model_encoder_layers_7_self_attn_v_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_7_self_attn_v_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_8_mlp_fc1.alpha []
lora_te_text_model_encoder_layers_8_mlp_fc1.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_8_mlp_fc1.lora_up.weight [3072,-1]
lora_te_text_model_encoder_layers_8_mlp_fc2.alpha []
lora_te_text_model_encoder_layers_8_mlp_fc2.lora_down.weight [-1,3072]
lora_te_text_model_encoder_layers_8_mlp_fc2.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_8_self_attn_k_proj.alpha []
lora_te_text_model_encoder_layers_8_self_attn_k_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_8_self_attn_k_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_8_self_attn_out_proj.alpha []
lora_te_text_model_encoder_layers_8_self_attn_out_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_8_self_attn_out_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_8_self_attn_q_proj.alpha []
lora_te_text_model_encoder_layers_8_self_attn_q_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_8_self_attn_q_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_8_self_attn_v_proj.alpha []
lora_te_text_model_encoder_layers_8_self_attn_v_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_8_self_attn_v_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_9_mlp_fc1.alpha []
lora_te_text_model_encoder_layers_9_mlp_fc1.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_9_mlp_fc1.lora_up.weight [3072,-1]
lora_te_text_model_encoder_layers_9_mlp_fc2.alpha []
lora_te_text_model_encoder_layers_9_mlp_fc2.lora_down.weight [-1,3072]
lora_te_text_model_encoder_layers_9_mlp_fc2.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_9_self_attn_k_proj.alpha []
lora_te_text_model_encoder_layers_9_self_attn_k_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_9_self_attn_k_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_9_self_attn_out_proj.alpha []
lora_te_text_model_encoder_layers_9_self_attn_out_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_9_self_attn_out_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_9_self_attn_q_proj.alpha []
lora_te_text_model_encoder_layers_9_self_attn_q_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_9_self_attn_q_proj.lora_up.weight [768,-1]
lora_te_text_model_encoder_layers_9_self_attn_v_proj.alpha []
lora_te_text_model_encoder_layers_9_self_attn_v_proj.lora_down.weight [-1,768]
lora_te_text_model_encoder_layers_9_self_attn_v_proj.lora_up.weight [768,-1]