-
Notifications
You must be signed in to change notification settings - Fork 50
Expand configs with undefined dtypes to allowed WMMA ones for attention #2123
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: develop
Are you sure you want to change the base?
Conversation
|
@umangyadav @dhernandez0 We can move this expanding logic (which i tried directly in Jenkinsfile) for example to handleNewConfigs script (or other script used in CI) and call that function in CI before tuning attn on navi? Then we would reuse defined dtypes from perfRunner in handleNewConfigs, and it won't be hardcoded as in case I did in Jenkinsfile direclty in this PR (did it like that for testing purposes) |
I'm happy with this solution as a temporary fix until we find how to properly fix it. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Pull request overview
This PR modifies the Jenkins pipeline to handle attention configurations for gfx1 chips by expanding entries without explicit dtypes to include all allowed WMMA data types (i8, f16, bf16), while filtering out f32 configurations.
Key Changes:
- Replaces simple
grepfiltering with a Python script that processes attention configuration files - Expands configuration entries lacking explicit
-tdtype flags to multiple entries with each allowed dtype - Preserves comments and empty lines in the configuration file
💡 Add Copilot custom instructions for smarter, more guided reviews. Learn how to get started.
|
i am okay with keeping logic here in jenkinsfile but drawback is that we have to keep same logic in multiple different places now. |
Once I finish resolving it on MITuna we can remove redundant code for this |
Motivation
Resolves known problems we have in MITuna until we fix them on MITuna so that our CI doesn't get blocked for dtypes problems...
Technical Details
Expand configs with undefined dtypes to allowed WMMA ones for attention
Test Plan
Tuning Attention
(Commit: 3d9f673 - with embedded py script right before tuning attn)
✅ https://ml-ci-internal.amd.com/blue/organizations/jenkins/MLIR%2Fmlir/detail/PR-2123/5/pipeline/667 (fails for other reasons)
(Commit: f3322d2 - after moving filtering logic in groovy function)
⌛https://ml-ci-internal.amd.com/job/MLIR/job/mlir/job/PR-2123/
Test Result
Submission Checklist