fix weights
parent
73de4b8f0f
commit
3a608ead87
|
|
@ -449,7 +449,12 @@ def lora_Linear_forward(self, input):
|
|||
torch.nn.Linear_forward_before_lora = backup_Linear_forward
|
||||
return result
|
||||
return lycoris.lyco_Linear_forward(self, input)
|
||||
clear_cache_lora(self, False)
|
||||
if lora_ext.is_sd_1_5:
|
||||
import networks
|
||||
networks.network_restore_weights_from_backup(self)
|
||||
networks.network_reset_cached_weight(self)
|
||||
else:
|
||||
clear_cache_lora(self, False)
|
||||
if (not self.weight.is_cuda) and input.is_cuda: #if variables not on the same device (between cpu and gpu)
|
||||
self_weight_cuda = self.weight.to(device=devices.device) #pass to GPU
|
||||
to_del = self.weight
|
||||
|
|
@ -488,7 +493,12 @@ def lora_Conv2d_forward(self, input):
|
|||
return result
|
||||
|
||||
return lycoris.lyco_Conv2d_forward(self, input)
|
||||
clear_cache_lora(self, False)
|
||||
if lora_ext.is_sd_1_5:
|
||||
import networks
|
||||
networks.network_restore_weights_from_backup(self)
|
||||
networks.network_reset_cached_weight(self)
|
||||
else:
|
||||
clear_cache_lora(self, False)
|
||||
if (not self.weight.is_cuda) and input.is_cuda:
|
||||
self_weight_cuda = self.weight.to(device=devices.device)
|
||||
to_del = self.weight
|
||||
|
|
@ -528,7 +538,12 @@ def lora_MultiheadAttention_forward(self, input):
|
|||
return result
|
||||
|
||||
return lycoris.lyco_MultiheadAttention_forward(self, input)
|
||||
clear_cache_lora(self, False)
|
||||
if lora_ext.is_sd_1_5:
|
||||
import networks
|
||||
networks.network_restore_weights_from_backup(self)
|
||||
networks.network_reset_cached_weight(self)
|
||||
else:
|
||||
clear_cache_lora(self, False)
|
||||
if (not self.weight.is_cuda) and input.is_cuda:
|
||||
self_weight_cuda = self.weight.to(device=devices.device)
|
||||
to_del = self.weight
|
||||
|
|
|
|||
Loading…
Reference in New Issue