Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merge OpenAI Triton commit 3613bf4 #2574

Merged
merged 9 commits into from
Oct 25, 2024
Merged

Conversation

whitneywhtsang
Copy link
Contributor

@whitneywhtsang whitneywhtsang commented Oct 25, 2024

This PR change the Triton base from 1064b59 to 3613bf4 (Oct 24).
Pass rate: 98.98%

Please do not squash and merge this PR.

makslevental and others added 8 commits October 23, 2024 08:15
Note, there are no uses of `nvgpu::` in this lib. Unblocks building
`*-opt` tools with "custom" LLVM that was built with
`-DLLVM_TARGETS_TO_BUILD="host;AMDGPU"` (i.e., no `NVPTX`).
This PR implements general conversion of MFMA dot operand
to Linear Layout.
Hopper supports vectorized atomics for add, max, and min. This PR adds
support for generating these instructions.

Note: atomic add/min/max also have packed instructions for f16x2 and
bf16x2. Packed instructions were used prior to this PR, but vectorized
instructions weren't. When vectorized instructions are available, this
PR switches to using vectorized instructions (like .v2.f16 instead of
.f16x2, or .v8.f16 instead of .v4.f16x2). When vectorized instructions
aren't available, packed instructions will be used instead.

This PR also adds a check for mask alignment, which wasn't previously
checked.
…4974)

This is a quick follow-up for the recent autotuner/testing changes as in
triton-lang/triton#4496. This PR moves the empty
cache creation into the driver code to make the code more device
independent.
…d (#4980)

The bitwidth is unimplemented in LLVM for pointer types so it throws an exception when evaluating the condition `tensorTy.getElementType().getIntOrFloatBitWidth()`
This commit refactors the AccelerateAMDMatmul patterns
in prep for mxfp support.
@whitneywhtsang whitneywhtsang self-assigned this Oct 25, 2024
@whitneywhtsang whitneywhtsang marked this pull request as ready for review October 25, 2024 11:58
@whitneywhtsang whitneywhtsang changed the title Merge OpenAI Triton commit 3c13f09 Merge OpenAI Triton commit 3613bf4 Oct 25, 2024
@whitneywhtsang whitneywhtsang merged commit be47a27 into main Oct 25, 2024
5 checks passed
@whitneywhtsang whitneywhtsang deleted the whitneywhtsang/merge2 branch October 25, 2024 14:38
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

9 participants