|
|
@ -223,13 +223,28 @@ def model_lora_keys(model, key_map={}):
|
|
|
|
counter += 1
|
|
|
|
counter += 1
|
|
|
|
counter = 0
|
|
|
|
counter = 0
|
|
|
|
text_model_lora_key = "lora_te_text_model_encoder_layers_{}_{}"
|
|
|
|
text_model_lora_key = "lora_te_text_model_encoder_layers_{}_{}"
|
|
|
|
for b in range(24):
|
|
|
|
clip_l_present = False
|
|
|
|
|
|
|
|
for b in range(32):
|
|
|
|
for c in LORA_CLIP_MAP:
|
|
|
|
for c in LORA_CLIP_MAP:
|
|
|
|
k = "transformer.text_model.encoder.layers.{}.{}.weight".format(b, c)
|
|
|
|
k = "transformer.text_model.encoder.layers.{}.{}.weight".format(b, c)
|
|
|
|
if k in sdk:
|
|
|
|
if k in sdk:
|
|
|
|
lora_key = text_model_lora_key.format(b, LORA_CLIP_MAP[c])
|
|
|
|
lora_key = text_model_lora_key.format(b, LORA_CLIP_MAP[c])
|
|
|
|
key_map[lora_key] = k
|
|
|
|
key_map[lora_key] = k
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
k = "clip_l.transformer.text_model.encoder.layers.{}.{}.weight".format(b, c)
|
|
|
|
|
|
|
|
if k in sdk:
|
|
|
|
|
|
|
|
lora_key = "lora_te1_text_model_encoder_layers_{}_{}".format(b, LORA_CLIP_MAP[c]) #SDXL base
|
|
|
|
|
|
|
|
key_map[lora_key] = k
|
|
|
|
|
|
|
|
clip_l_present = True
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
k = "clip_g.transformer.text_model.encoder.layers.{}.{}.weight".format(b, c)
|
|
|
|
|
|
|
|
if k in sdk:
|
|
|
|
|
|
|
|
if clip_l_present:
|
|
|
|
|
|
|
|
lora_key = "lora_te2_text_model_encoder_layers_{}_{}".format(b, LORA_CLIP_MAP[c]) #SDXL base
|
|
|
|
|
|
|
|
else:
|
|
|
|
|
|
|
|
lora_key = "lora_te_text_model_encoder_layers_{}_{}".format(b, LORA_CLIP_MAP[c]) #TODO: test if this is correct for SDXL-Refiner
|
|
|
|
|
|
|
|
key_map[lora_key] = k
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
#Locon stuff
|
|
|
|
#Locon stuff
|
|
|
|
ds_counter = 0
|
|
|
|
ds_counter = 0
|
|
|
|