|
|
|
@ -20,8 +20,9 @@ class DisableInitialization:
|
|
|
|
```
|
|
|
|
```
|
|
|
|
"""
|
|
|
|
"""
|
|
|
|
|
|
|
|
|
|
|
|
def __init__(self):
|
|
|
|
def __init__(self, disable_clip=True):
|
|
|
|
self.replaced = []
|
|
|
|
self.replaced = []
|
|
|
|
|
|
|
|
self.disable_clip = disable_clip
|
|
|
|
|
|
|
|
|
|
|
|
def replace(self, obj, field, func):
|
|
|
|
def replace(self, obj, field, func):
|
|
|
|
original = getattr(obj, field, None)
|
|
|
|
original = getattr(obj, field, None)
|
|
|
|
@ -75,6 +76,8 @@ class DisableInitialization:
|
|
|
|
self.replace(torch.nn.init, 'kaiming_uniform_', do_nothing)
|
|
|
|
self.replace(torch.nn.init, 'kaiming_uniform_', do_nothing)
|
|
|
|
self.replace(torch.nn.init, '_no_grad_normal_', do_nothing)
|
|
|
|
self.replace(torch.nn.init, '_no_grad_normal_', do_nothing)
|
|
|
|
self.replace(torch.nn.init, '_no_grad_uniform_', do_nothing)
|
|
|
|
self.replace(torch.nn.init, '_no_grad_uniform_', do_nothing)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
if self.disable_clip:
|
|
|
|
self.create_model_and_transforms = self.replace(open_clip, 'create_model_and_transforms', create_model_and_transforms_without_pretrained)
|
|
|
|
self.create_model_and_transforms = self.replace(open_clip, 'create_model_and_transforms', create_model_and_transforms_without_pretrained)
|
|
|
|
self.CLIPTextModel_from_pretrained = self.replace(ldm.modules.encoders.modules.CLIPTextModel, 'from_pretrained', CLIPTextModel_from_pretrained)
|
|
|
|
self.CLIPTextModel_from_pretrained = self.replace(ldm.modules.encoders.modules.CLIPTextModel, 'from_pretrained', CLIPTextModel_from_pretrained)
|
|
|
|
self.transformers_modeling_utils_load_pretrained_model = self.replace(transformers.modeling_utils.PreTrainedModel, '_load_pretrained_model', transformers_modeling_utils_load_pretrained_model)
|
|
|
|
self.transformers_modeling_utils_load_pretrained_model = self.replace(transformers.modeling_utils.PreTrainedModel, '_load_pretrained_model', transformers_modeling_utils_load_pretrained_model)
|
|
|
|
|