Skip to content

Conversation

@vivekkhandelwal1
Copy link
Collaborator

No description provided.

Copy link
Member

@pashu123 pashu123 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM! Are there any reasons why we aren't decomposing this torch op into another set of torch ops?

@vivekkhandelwal1
Copy link
Collaborator Author

LGTM! Are there any reasons why we aren't decomposing this torch op into another set of torch ops?

AFAIK, the reason for not doing that is since we want the attention op as a single kernel. Hence, we just lower it to tm_tensor.attention and the rest is taken care of during the codegen.

@vivekkhandelwal1 vivekkhandelwal1 merged commit 25aa0c6 into llvm:main Feb 5, 2025
3 checks passed
@vivekkhandelwal1 vivekkhandelwal1 deleted the aten-sdpa-gqa branch February 5, 2025 05:28
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants