Pre-Compiled AoTI

#68
by multimodalart HF Staff - opened
Files changed (2) hide show
  1. app.py +9 -6
  2. requirements.txt +2 -1
app.py CHANGED
@@ -6,10 +6,10 @@ import spaces
6
 
7
  from PIL import Image
8
 
9
- from optimization import optimize_pipeline_
10
  from qwenimage.pipeline_qwen_image_edit import QwenImageEditPipeline
11
- from qwenimage.transformer_qwenimage import QwenImageTransformer2DModel
12
- from qwenimage.qwen_fa3_processor import QwenDoubleStreamAttnProcessorFA3
13
 
14
  import os
15
  import base64
@@ -144,11 +144,14 @@ device = "cuda" if torch.cuda.is_available() else "cpu"
144
 
145
  # Load the model pipeline
146
  pipe = QwenImageEditPipeline.from_pretrained("Qwen/Qwen-Image-Edit", torch_dtype=dtype).to(device)
147
- pipe.transformer.__class__ = QwenImageTransformer2DModel
148
- pipe.transformer.set_attn_processor(QwenDoubleStreamAttnProcessorFA3())
 
 
 
149
 
150
  # --- Ahead-of-time compilation ---
151
- optimize_pipeline_(pipe, image=Image.new("RGB", (1024, 1024)), prompt="prompt")
152
 
153
  # --- UI Constants and Helpers ---
154
  MAX_SEED = np.iinfo(np.int32).max
 
6
 
7
  from PIL import Image
8
 
9
+ # from optimization import optimize_pipeline_
10
  from qwenimage.pipeline_qwen_image_edit import QwenImageEditPipeline
11
+ # from qwenimage.transformer_qwenimage import QwenImageTransformer2DModel
12
+ # from qwenimage.qwen_fa3_processor import QwenDoubleStreamAttnProcessorFA3
13
 
14
  import os
15
  import base64
 
144
 
145
  # Load the model pipeline
146
  pipe = QwenImageEditPipeline.from_pretrained("Qwen/Qwen-Image-Edit", torch_dtype=dtype).to(device)
147
+
148
+ spaces.aoti_blocks_load(pipe.transformer, "zerogpu-aoti/Qwen-Image", variant="fa3")
149
+
150
+ # pipe.transformer.__class__ = QwenImageTransformer2DModel
151
+ # pipe.transformer.set_attn_processor(QwenDoubleStreamAttnProcessorFA3())
152
 
153
  # --- Ahead-of-time compilation ---
154
+ # optimize_pipeline_(pipe, image=Image.new("RGB", (1024, 1024)), prompt="prompt")
155
 
156
  # --- UI Constants and Helpers ---
157
  MAX_SEED = np.iinfo(np.int32).max
requirements.txt CHANGED
@@ -7,4 +7,5 @@ safetensors
7
  sentencepiece
8
  dashscope
9
  torchvision
10
- peft
 
 
7
  sentencepiece
8
  dashscope
9
  torchvision
10
+ peft
11
+ torch==2.8