Skip to content

[decoding_attention] Fix broken flash_attention and xformers #302

[decoding_attention] Fix broken flash_attention and xformers

[decoding_attention] Fix broken flash_attention and xformers #302

Triggered via pull request December 9, 2024 13:12
Status Failure
Total duration 4h 0m 35s
Artifacts

pr.yaml

on: pull_request
h100-pytorch-test  /  linux-test-h100
4h 0m
h100-pytorch-test / linux-test-h100
h100-triton-main-test  /  linux-test-h100
4h 0m
h100-triton-main-test / linux-test-h100
Fit to window
Zoom out
Zoom in

Annotations

4 errors
h100-pytorch-test / linux-test-h100
The job running on runner gcp-h100-runner-hl226-runner-qfcgj has exceeded the maximum execution time of 240 minutes.
h100-pytorch-test / linux-test-h100
The operation was canceled.
h100-triton-main-test / linux-test-h100
The job running on runner gcp-h100-runner-hl226-runner-7rwdn has exceeded the maximum execution time of 240 minutes.
h100-triton-main-test / linux-test-h100
The operation was canceled.