diff --git a/scripts/dataset_tag_editor/interrogators/blip2_captioning.py b/scripts/dataset_tag_editor/interrogators/blip2_captioning.py index b500efe..b309e7d 100644 --- a/scripts/dataset_tag_editor/interrogators/blip2_captioning.py +++ b/scripts/dataset_tag_editor/interrogators/blip2_captioning.py @@ -12,12 +12,8 @@ class BLIP2Captioning: def load(self): if self.model is None or self.processor is None: - self.processor = Blip2Processor.from_pretrained( - self.MODEL_REPO, cache_dir=paths.model_path - ) - self.model = Blip2ForConditionalGeneration.from_pretrained( - self.MODEL_REPO, cache_dir=paths.model_path - ).to(devices.device) + self.processor = Blip2Processor.from_pretrained(self.MODEL_REPO) + self.model = Blip2ForConditionalGeneration.from_pretrained(self.MODEL_REPO).to(devices.device) def unload(self): if not shared.opts.interrogate_keep_models_in_memory: diff --git a/scripts/dataset_tag_editor/interrogators/git_large_captioning.py b/scripts/dataset_tag_editor/interrogators/git_large_captioning.py index bf9f223..ddcc1e1 100644 --- a/scripts/dataset_tag_editor/interrogators/git_large_captioning.py +++ b/scripts/dataset_tag_editor/interrogators/git_large_captioning.py @@ -14,12 +14,8 @@ class GITLargeCaptioning: def load(self): if self.model is None or self.processor is None: - self.processor = AutoProcessor.from_pretrained( - self.MODEL_REPO, cache_dir=paths.model_path - ) - self.model = AutoModelForCausalLM.from_pretrained( - self.MODEL_REPO, cache_dir=paths.model_path - ).to(shared.device) + self.processor = AutoProcessor.from_pretrained(self.MODEL_REPO) + self.model = AutoModelForCausalLM.from_pretrained(self.MODEL_REPO).to(shared.device) lowvram.send_everything_to_cpu() def unload(self): diff --git a/scripts/dataset_tag_editor/interrogators/waifu_diffusion_tagger.py b/scripts/dataset_tag_editor/interrogators/waifu_diffusion_tagger.py index b66d41b..3897af5 100644 --- a/scripts/dataset_tag_editor/interrogators/waifu_diffusion_tagger.py +++ b/scripts/dataset_tag_editor/interrogators/waifu_diffusion_tagger.py @@ -26,7 +26,7 @@ class WaifuDiffusionTagger: if not self.model: path_model = huggingface_hub.hf_hub_download( - self.MODEL_REPO, self.MODEL_FILENAME, cache_dir=paths.model_path + self.MODEL_REPO, self.MODEL_FILENAME ) if ( "all" in shared.cmd_opts.use_cpu