-
Notifications
You must be signed in to change notification settings - Fork 6.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
triton + sageattention error: RuntimeError: PassManager::run failed #6228
Comments
the 1080 is shader model 6.7, which is not supported by triton. it requires 7.5 or higher |
Ok, why then I get this test program (written lower) working fine for me, without any errors? import torch
import triton
import triton.language as tl
@triton.jit
def add_kernel(x_ptr, y_ptr, output_ptr, n_elements, BLOCK_SIZE: tl.constexpr):
pid = tl.program_id(axis=0)
block_start = pid * BLOCK_SIZE
offsets = block_start + tl.arange(0, BLOCK_SIZE)
mask = offsets < n_elements
x = tl.load(x_ptr + offsets, mask=mask)
y = tl.load(y_ptr + offsets, mask=mask)
output = x + y
tl.store(output_ptr + offsets, output, mask=mask)
def add(x: torch.Tensor, y: torch.Tensor):
output = torch.empty_like(x)
assert x.is_cuda and y.is_cuda and output.is_cuda
n_elements = output.numel()
grid = lambda meta: (triton.cdiv(n_elements, meta["BLOCK_SIZE"]),)
add_kernel[grid](x, y, output, n_elements, BLOCK_SIZE=1024)
return output
a = torch.rand(3, device="cuda")
b = a + a
b_compiled = add(a, a)
print(b_compiled - b)
print("If you see tensor([0., 0., 0.], device='cuda:0'), then it works") |
I am trying to tell you that your error is caused by the fact that your GPU is too old. CUDA features are added progressively so it's possible for simple test programs to pass but your actual application to fail. https://github.com/triton-lang/triton?tab=readme-ov-file#compatibility SM 8.0 is actually required by Triton.
the int8-int8 op that is referenced there and that Sage Attention uses isn't implemented by Triton for SM 6.5 in principle pascal can have it just fine. it's just triton doesn't implement it. |
Expected Behavior
Expected image gen to start as intended
Actual Behavior
Image gen crashed on 1st step.
Steps to Reproduce
added --use-sage-attention tag
made a default workflow
Debug Logs
Other
GPU: Nvidia GTX1080
Steps:
The text was updated successfully, but these errors were encountered: