Skip to content

Commit 7c00634

Browse files
committed
Remove layer_norm from TPU benchmark (OOB slice bug, gh#1937)
1 parent d0bf412 commit 7c00634

File tree

2 files changed

+3
-3
lines changed

2 files changed

+3
-3
lines changed

.github/workflows/benchmark_tpu_nightly.yml

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,8 @@ on:
1616
# rms_norm: InductorLoweringError in torch.mean reduction codegen for fori_loop/emit_pipeline
1717
# geglu/swiglu: autotuning takes >15min per kernel (large shape 8x2048x4096), many configs fail to compile
1818
# low_mem_dropout: ~37% element accuracy mismatch on all configs except block_sizes=[128]
19-
default: "exp,add,softmax_two_pass,welford,layer_norm"
19+
# layer_norm: OOB slice when reduction_loops doesn't evenly divide the reduction dim (gh#1937)
20+
default: "exp,add,softmax_two_pass,welford"
2021

2122
permissions:
2223
contents: read
@@ -27,4 +28,4 @@ jobs:
2728
permissions:
2829
contents: read
2930
with:
30-
kernels: ${{ github.event.inputs.kernels || 'exp,add,softmax_two_pass,welford,layer_norm' }}
31+
kernels: ${{ github.event.inputs.kernels || 'exp,add,softmax_two_pass,welford' }}

benchmarks/run_tpu.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -107,7 +107,6 @@ def _softmax_shapes() -> list[tuple[str, tuple[Any, ...]]]:
107107
_softmax_shapes,
108108
),
109109
"welford": ("welford", "welford", None, None),
110-
"layer_norm": ("layer_norm", "layer_norm", None, None),
111110
}
112111

113112

0 commit comments

Comments
 (0)