Skip to content

Commit 7ab695c

Browse files
committed
fix flex attention
1 parent dfd6faf commit 7ab695c

File tree

2 files changed

+8
-0
lines changed

2 files changed

+8
-0
lines changed

benchmarks/nightly/autogen.yaml

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -155,3 +155,7 @@ vector_exp_bwd:
155155
welford_fwd:
156156
args: --op welford --baseline eager_layer_norm --metrics latency,speedup --only
157157
test_no_welford,triton_welford,eager_layer_norm
158+
bf16_flex_attention_fwd:
159+
args: --op flex_attention --metrics latency,tflops --only compiled
160+
bf16_flex_attention_bwd:
161+
args: --op flex_attention --metrics latency,tflops --only compiled

benchmarks/nightly/manual.yaml

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -22,3 +22,7 @@ extra_args:
2222
# flash_attention triton_tutorial_flash_v2 impl only supports causal in backward
2323
bf16_flash_attention_bwd:
2424
args: --op flash_attention --baseline flash_v3 --metrics tflops,speedup --bwd --only triton_tutorial_flash_v2,flash_v3 --causal
25+
bf16_flex_attention_fwd:
26+
args: --op flex_attention --metrics latency,tflops --only compiled
27+
bf16_flex_attention_bwd:
28+
args: --op flex_attention --metrics latency,tflops --only compiled

0 commit comments

Comments
 (0)