diff --git a/threestudio/models/prompt_processors/deepfloyd_prompt_processor.py b/threestudio/models/prompt_processors/deepfloyd_prompt_processor.py index b4316ee..751b1e3 100644 --- a/threestudio/models/prompt_processors/deepfloyd_prompt_processor.py +++ b/threestudio/models/prompt_processors/deepfloyd_prompt_processor.py @@ -58,7 +58,6 @@ class DeepFloydPromptProcessor(PromptProcessor): tokenizer = T5Tokenizer.from_pretrained( pretrained_model_name_or_path, subfolder="tokenizer", - local_files_only=True ) text_encoder = T5EncoderModel.from_pretrained( pretrained_model_name_or_path, @@ -67,7 +66,6 @@ class DeepFloydPromptProcessor(PromptProcessor): load_in_8bit=True, variant="8bit", device_map="auto", - local_files_only=True ) with torch.no_grad(): text_inputs = tokenizer( diff --git a/threestudio/models/prompt_processors/stable_diffusion_prompt_processor.py b/threestudio/models/prompt_processors/stable_diffusion_prompt_processor.py index 8423b20..e2e334e 100644 --- a/threestudio/models/prompt_processors/stable_diffusion_prompt_processor.py +++ b/threestudio/models/prompt_processors/stable_diffusion_prompt_processor.py @@ -75,13 +75,11 @@ class StableDiffusionPromptProcessor(PromptProcessor): tokenizer = AutoTokenizer.from_pretrained( pretrained_model_name_or_path, subfolder="tokenizer", - local_files_only=True, ) text_encoder = CLIPTextModel.from_pretrained( pretrained_model_name_or_path, subfolder="text_encoder", device_map="auto", - local_files_only=True, ) with torch.no_grad():