Skip to content

Conversation

@pytorchbot
Copy link
Collaborator

Currently Efficient attention and Flash attention share the same set of GPU
kernels on ROCM and have common limitations on head sizes.

Fixes #132004

cc @jeffdaily @sunway513 @jithunnair-amd @pruthvistony @ROCmSupport @dllehr-amd @jataylo @hongxiayang

Currently Efficient attention and Flash attention share the same set of GPU
kernels on ROCM and have common limitations on head sizes.

Fixes #132004

Pull Request resolved: #133331
Approved by: https://github.com/malfet, https://github.com/jithunnair-amd

(cherry picked from commit 46ecc67)
@pytorch-bot
Copy link

pytorch-bot bot commented Aug 27, 2024

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/134531

Note: Links to docs will display an error until the docs builds have been completed.

❌ 2 New Failures, 1 Unrelated Failure

As of commit ae4d2fd with merge base b66e3f0 (image):

NEW FAILURES - The following jobs have failed:

FLAKY - The following job failed but was likely due to flakiness present on trunk:

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@pytorch-bot pytorch-bot bot added ciflow/rocm Trigger "default" config CI on ROCm module: rocm AMD GPU support for Pytorch labels Aug 27, 2024
@pruthvistony pruthvistony requested a review from atalman August 27, 2024 00:10
@pruthvistony
Copy link
Collaborator

pruthvistony commented Aug 27, 2024

@atalman , @malfet
Can you please help on this PR cherry-pick.

The code changes are ROCm related and fixes - #132004

@pruthvistony pruthvistony requested a review from malfet August 27, 2024 00:11
@atalman atalman merged commit 6a79d4a into release/2.4 Aug 27, 2024
@atalman atalman deleted the cherry-pick-133331-by-pytorch_bot_bot_ branch August 27, 2024 11:36
xinyazhang added a commit to ROCm/pytorch that referenced this pull request Aug 28, 2024
…#134531)

[ROCm] Prevent accidental enablement of efficient attention. (pytorch#133331)

Currently Efficient attention and Flash attention share the same set of GPU
kernels on ROCM and have common limitations on head sizes.

Fixes pytorch#132004

Pull Request resolved: pytorch#133331
Approved by: https://github.com/malfet, https://github.com/jithunnair-amd

(cherry picked from commit 46ecc67)

Co-authored-by: Xinya Zhang <[email protected]>
pruthvistony pushed a commit to ROCm/pytorch that referenced this pull request Sep 9, 2024
…#134531) (#1565)

[ROCm] Prevent accidental enablement of efficient attention. (pytorch#133331)

Currently Efficient attention and Flash attention share the same set of
GPU kernels on ROCM and have common limitations on head sizes.

Pull Request resolved: pytorch#133331
Approved by: https://github.com/malfet,
https://github.com/jithunnair-amd

(cherry picked from commit 46ecc67)

Fixes pytorch#132004

Co-authored-by: pytorchbot <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ciflow/rocm Trigger "default" config CI on ROCm module: rocm AMD GPU support for Pytorch open source

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants