thanks!
@bot /style
Style fixes have been applied. View the workflow run here.
The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.
Thank you!
205 | 223 | image = self.image_processor(images=image, return_tensors="pt").to(self.device) | |
206 | 224 | image_embeds = self.image_encoder(**image, output_hidden_states=True) | |
207 | return image_embeds.hidden_states[-1] | ||
225 | return image_embeds.hidden_states[-2] |
@a-r-r-o-w the change of output is probably from here
138 | 156 | tokenizer: AutoTokenizer, | |
139 | 157 | text_encoder: UMT5EncoderModel, | |
140 | image_encoder: CLIPVisionModelWithProjection, | ||
158 | image_encoder: CLIPVisionModel, |
need to match with here https://huggingface.co/Wan-AI/Wan2.1-I2V-14B-480P-Diffusers/blob/main/model_index.json#L6
Login to write a write a comment.
I fix some bug in wan i2v pipeline and add a new example