linoyts HF Staff commited on
Commit
26bc0e1
·
verified ·
1 Parent(s): f1ee2b7

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +0 -28
app.py CHANGED
@@ -46,34 +46,6 @@ pipe = WanPipeline.from_pretrained(MODEL_ID,
46
  torch_dtype=torch.bfloat16,
47
  ).to('cuda')
48
 
49
- # load, fuse, unload before compilation
50
- # pipe.load_lora_weights(
51
- # "vrgamedevgirl84/Wan14BT2VFusioniX",
52
- # weight_name="FusionX_LoRa/Phantom_Wan_14B_FusionX_LoRA.safetensors",
53
- # adapter_name="phantom"
54
- # )
55
-
56
- # pipe.set_adapters(["phantom"], adapter_weights=[0.95])
57
- # pipe.fuse_lora(adapter_names=["phantom"], lora_scale=1.0)
58
- # pipe.unload_lora_weights()
59
-
60
-
61
- # pipe.load_lora_weights(
62
- # "vrgamedevgirl84/Wan14BT2VFusioniX",
63
- # weight_name="FusionX_LoRa/Phantom_Wan_14B_FusionX_LoRA.safetensors",
64
- # adapter_name="phantom"
65
- # )
66
- # kwargs = {}
67
- # kwargs["load_into_transformer_2"] = True
68
- # pipe.load_lora_weights(
69
- # "vrgamedevgirl84/Wan14BT2VFusioniX",
70
- # weight_name="FusionX_LoRa/Phantom_Wan_14B_FusionX_LoRA.safetensors",
71
- # adapter_name="phantom_2", **kwargs
72
- # )
73
- # pipe.set_adapters(["phantom", "phantom_2"], adapter_weights=[1., 1.])
74
- # pipe.fuse_lora(adapter_names=["phantom"], lora_scale=3., components=["transformer"])
75
- # pipe.fuse_lora(adapter_names=["phantom_2"], lora_scale=1., components=["transformer_2"])
76
- # pipe.unload_lora_weights()
77
 
78
  for i in range(3):
79
  gc.collect()
 
46
  torch_dtype=torch.bfloat16,
47
  ).to('cuda')
48
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
49
 
50
  for i in range(3):
51
  gc.collect()