Skip to content

Commit

Permalink
Fix CI
Browse files Browse the repository at this point in the history
  • Loading branch information
xuzhao9 committed Dec 10, 2024
1 parent 94e670c commit 6afeccf
Show file tree
Hide file tree
Showing 2 changed files with 17 additions and 4 deletions.
3 changes: 3 additions & 0 deletions test/test_gpu/skip_tests_h100_pytorch.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -37,3 +37,6 @@ ragged_attention:
- hstu_triton_ragged_attention_persistent
# cpu-op for testing
test_op:
# TODO: decoding attention requires updated xformers and flash_attn
# Which will RAM OOM on the CI machine
decoding_attention:
18 changes: 14 additions & 4 deletions test/test_gpu/skip_tests_h100_triton_main.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -16,10 +16,10 @@ fp8_attention:
- colfax_fmha
# fb-only kernels
fp8_fused_quant_gemm_rowwise:
gemm:
# internal only kernels
- hstu_triton_matmul
- colfax_cutlass_matmul
# gemm:
# # internal only kernels
# - hstu_triton_matmul
# - colfax_cutlass_matmul
# jagged tests are slow, so disable them in OSS
jagged_layer_norm:
jagged_mean:
Expand All @@ -35,3 +35,13 @@ ragged_attention:
- hstu_triton_ragged_attention_persistent
# cpu-op for testing
test_op:
# TODO: decoding attention requires updated xformers and flash_attn
# Which will RAM OOM on the CI machine
decoding_attention:
# FIXME: PT2 is broken with Triton-main
launch_latency:
addmm:
gemm:
flash_attention:
gather_gemv:
layer_norm:

0 comments on commit 6afeccf

Please sign in to comment.