-
Notifications
You must be signed in to change notification settings - Fork 3.7k
[diffusion] pipeline: fix error when enable torch compile #14509
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Summary of ChangesHello @zcnrex, I'm Gemini Code Assist1! I'm currently reviewing this pull request and will post my feedback shortly. In the meantime, here's a summary to help you and other reviewers quickly get up to speed! This pull request addresses a critical bug in the diffusion pipeline that previously prevented the successful utilization of Highlights
Using Gemini Code AssistThe full guide for Gemini Code Assist can be found on our documentation page, here are some quick tips. Invoking Gemini You can request assistance from Gemini at any point by creating a comment using either
Customization To customize Gemini Code Assist for GitHub experience, repository maintainers can create a configuration file and/or provide a custom code review style guide (such as PEP-8 for Python) by creating and adding files to a Limitations & Feedback Gemini Code Assist may make mistakes. Please leave feedback on any instances where its feedback is incorrect or counter productive. You can react with 👍 and 👎 on @gemini-code-assist comments. If you're interested in giving your feedback about your experience with Gemini Code Assist for Github and other Google products, sign up here. You can also get AI-powered code generation, chat, as well as code reviews directly in the IDE at no cost with the Gemini Code Assist IDE Extension. Footnotes
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Code Review
This pull request addresses a crash that occurs when torch.compile is enabled by separating the compiled function from the original model module. While this correctly resolves the AttributeError, the implementation has a couple of significant issues.
First, in the case of lazy model loading within _prepare_denoising_loop, the code still overwrites self.transformer with the compiled function, which reintroduces the original bug under that specific code path. I've added a specific comment with a suggested fix for this.
Second, and more critically, the compiled functions (self.transformer_compiled_func and self.transformer_2_compiled_func) are never actually used for the main model forward pass. The code continues to call the original, un-compiled modules. This means that while the crash is avoided, the performance benefits of torch.compile (as shown in the PR description) are not realized with the current changes. To fix this, the forward method in DenoisingStage needs to be updated to select and call the appropriate compiled function when torch.compile is enabled. Since this is outside the current diff, I couldn't leave a direct comment, but it's a crucial change to make this PR effective.
Motivation
Error when setting
--enable-torch-compile trueglang serve --model-path Wan-AI/Wan2.1-T2V-1.3B-Diffusers --port 3000 --enable-torch-compile trueModifications
Updated the denoising.py now it runs without error.
Accuracy Tests
Checked output video is the same.
Benchmarking and Profiling
Tested on H100
enabled
disabled
Checklist