Skip to content

Conversation

@ParamThakkar123
Copy link
Contributor

No description provided.

@ParamThakkar123 ParamThakkar123 marked this pull request as draft November 29, 2025 05:34
@ParamThakkar123
Copy link
Contributor Author

I added support for video generation models starting with Wan2.1 models will later extend to more models if this works, currently I added support for WanPipeline in supported model architectures in Image_diffusion plugin but it gives an error where its trying to find /workspace/jobs/None which doesn't exist

@codecov-commenter
Copy link

codecov-commenter commented Nov 29, 2025

Codecov Report

❌ Patch coverage is 2.15054% with 182 lines in your changes missing coverage. Please review.

Files with missing lines Patch % Lines
api/transformerlab/plugins/image_diffusion/main.py 0.00% 153 Missing ⚠️
api/transformerlab/routers/experiment/diffusion.py 12.12% 27 Missing and 2 partials ⚠️

📢 Thoughts on this report? Let us know!

@deep1401
Copy link
Member

deep1401 commented Dec 2, 2025

I added support for video generation models starting with Wan2.1 models will later extend to more models if this works, currently I added support for WanPipeline in supported model architectures in Image_diffusion plugin but it gives an error where its trying to find /workspace/jobs/None which doesn't exist

There was an issue with running diffusion models, could you update and check if that issue is resolved now? The issue was caused by our new system with the workspace under orgs

@deep1401
Copy link
Member

deep1401 commented Dec 3, 2025

I added support for video generation models starting with Wan2.1 models will later extend to more models if this works, currently I added support for WanPipeline in supported model architectures in Image_diffusion plugin but it gives an error where its trying to find /workspace/jobs/None which doesn't exist

There was an issue with running diffusion models, could you update and check if that issue is resolved now? The issue was caused by our new system with the workspace under orgs

I figure this is still a WIP but I would also suggest not adding things to the existing image_diffusion plugin and creating a new video diffusion plugin.

@ParamThakkar123 ParamThakkar123 marked this pull request as ready for review December 5, 2025 18:19
Copy link
Member

@deep1401 deep1401 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Added some comments, we might have to restructure the PR and break it down into 2 and test these

"url": "",
"files": ["main.py", "video_diffusion_worker.py", "setup.sh"],
"model_architectures": [
"DiffusionPipeline",
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

There is no architecture called DiffusionPipeline, this should not be here

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Also our image_diffusion plugin has no architectures in there so lets also not put any in here.

@@ -1,0 +1,3 @@
#!/bin/bash
uv pip install git+https://github.com/huggingface/diffusers transformers accelerate safetensors
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

What version of diffusers is required? The latest release on pypi is 0.35.2 and I assumed Wan2.1 would work with that but maybe you'd know better

"amd"
],
"files": ["main.py", "diffusion_worker.py", "setup.sh"],
"model_architectures": ["DiffusionPipeline", "WanPipeline"],
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I dont think adding architectures here works. Did you try running the model?
You'd need to add architecture in the massive list in the diffusion router

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This wouldnt work because we've not made any frontend changes right? Maybe lets split up PRs and you have a separate WIP for video diffusion. You could implement API and the plugin and add a video for a working model in that PR.

Lets not add anything without testing since diffusion has a lot of things which might break by minor changes here or there

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Also lets make this PR image_diffusion changes only and also add the ZTurbo model additions in here

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I dont think this works since the structure of the wan model is different. Reference: https://huggingface.co/Wan-AI/Wan2.1-T2V-14B/tree/main

This file in image diffusion was added specifically for FLUX as that required sharding. Lets try to test everything we add :)

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Yeah. Currently testing all the things I added. The models are quite huge :)

Copy link
Member

@deep1401 deep1401 Dec 5, 2025

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

okay no worries I converted the PR to draft for now and we can split it up and test things.
Let me know if you need my help testing anything or if you dont get something on diffusion about how it works, I can help you figure out!

I mentioned a text to video reference model below but you can look at WanAI's collections and maybe select smaller models for testing. They usually put the small ones in their diffusers collections

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Sure! Thanks deep. I will try this out and get it done earliest 🫡

@@ -0,0 +1,3 @@
#!/bin/bash
uv pip install git+https://github.com/huggingface/diffusers transformers accelerate safetensors
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Same comment as above as to why we need dev version and for what model

Copy link
Member

@deep1401 deep1401 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Sorry I forgot to add this earlier but lets use this as a reference model for Wan support. This feels small enough to run on our hardware and diffusers required version is 0.33.0.dev so just updating to current diffusers prod version should work
https://huggingface.co/Wan-AI/Wan2.1-T2V-1.3B-Diffusers

@deep1401 deep1401 marked this pull request as draft December 5, 2025 20:17
@ParamThakkar123 ParamThakkar123 changed the title Added Video Generation support Added Video Generation support for generating images Dec 17, 2025
@ParamThakkar123 ParamThakkar123 marked this pull request as ready for review December 17, 2025 10:15
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants