File tree Expand file tree Collapse file tree 1 file changed +4
-1
lines changed Expand file tree Collapse file tree 1 file changed +4
-1
lines changed Original file line number Diff line number Diff line change 41
41
from transformer_engine_jax import (
42
42
NVTE_Fused_Attn_Backend ,
43
43
get_cudnn_version ,
44
+ get_device_compute_capability ,
44
45
)
45
46
46
47
from distributed_test_base import assert_equal_collectives
@@ -347,7 +348,9 @@ def _check_configs(self):
347
348
pytest .skip (
348
349
"seqlen_q > seqlen_kv is not supported with sliding window attention in cuDNN"
349
350
)
350
-
351
+
352
+ if get_device_compute_capability (0 ) == 100 and self .dropout_prob == 0.1 and self .attn_bias_type is not AttnBiasType .NO_BIAS :
353
+ pytest .skip ("For Blackwell, there is no bprop kernel support for dropout + deterministic (bias) config " )
351
354
# Test the MLA case where head dims for qk differ from head dims for v, only if the tensors
352
355
# are provided in BSHD_BSHD_BSHD or THD_THD_THD formats
353
356
if self .head_dim_qk != self .head_dim_v and not self .qkv_layout .is_separate ():
You can’t perform that action at this time.
0 commit comments