From 8d9f2e69d2a19ad31634553cc75d298a456e0a07 Mon Sep 17 00:00:00 2001 From: sayakpaul Date: Fri, 27 Dec 2024 17:43:59 +0530 Subject: [PATCH] fix comment --- tests/lora/test_lora_layers_flux.py | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) diff --git a/tests/lora/test_lora_layers_flux.py b/tests/lora/test_lora_layers_flux.py index b7352a4c2ad8..8d3b3245ed4b 100644 --- a/tests/lora/test_lora_layers_flux.py +++ b/tests/lora/test_lora_layers_flux.py @@ -194,8 +194,7 @@ def test_lora_expansion_works_for_absent_keys(self): pipe.unload_lora_weights() pipe.load_lora_weights(os.path.join(tmpdirname, "pytorch_lora_weights.safetensors")) - # modify the state dict to have alpha values following - # https://huggingface.co/TheLastBen/Jon_Snow_Flux_LoRA/blob/main/jon_snow.safetensors + # Modify the state dict to exclude "x_embedder" related LoRA params. lora_state_dict = safetensors.torch.load_file(os.path.join(tmpdirname, "pytorch_lora_weights.safetensors")) lora_state_dict_without_xembedder = {k: v for k, v in lora_state_dict.items() if "x_embedder" not in k}