@@ -5745,7 +5745,7 @@ def test_dot_max_num_imprecise_acc(M, N, K, BLOCK_M, BLOCK_N, BLOCK_K, in_type_s
57455745
57465746@pytest .mark .parametrize ("enable_fp_fusion" , [False , True ])
57475747@pytest .mark .parametrize ("default_override" , [False , True ])
5748- def test_enable_fp_fusion (enable_fp_fusion , default_override , device , fresh_knobs_except_libraries ):
5748+ def test_enable_fp_fusion (enable_fp_fusion , default_override , device , fresh_knobs ):
57495749 # Sequential multiply add can be fused by backend
57505750 @triton .jit
57515751 def mul_add (data ):
@@ -5754,7 +5754,7 @@ def mul_add(data):
57545754
57555755 data = torch .randn ((128 , ), device = device , dtype = torch .float32 )
57565756 if default_override :
5757- fresh_knobs_except_libraries .language .default_fp_fusion = enable_fp_fusion
5757+ fresh_knobs .language .default_fp_fusion = enable_fp_fusion
57585758 h = mul_add .warmup (data , grid = (1 , ))
57595759 else :
57605760 h = mul_add .warmup (data , grid = (1 , ), enable_fp_fusion = enable_fp_fusion )
@@ -5772,7 +5772,7 @@ def mul_add(data):
57725772
57735773@pytest .mark .skipif (not is_cuda (), reason = "Requires CUDA" )
57745774@pytest .mark .parametrize ("enable_reflect_ftz" , [False , True ])
5775- def test_enable_reflect_ftz (enable_reflect_ftz , device , fresh_knobs_except_libraries ):
5775+ def test_enable_reflect_ftz (enable_reflect_ftz , device , fresh_knobs ):
57765776
57775777 @triton .jit
57785778 def exp2 (data ):
@@ -5793,7 +5793,7 @@ def exp2(data):
57935793
57945794@pytest .mark .parametrize ("arch" , ["sm70" , "sm80" , "sm90" , "gfx942" , "gfx950" , "gfx1200" ])
57955795@pytest .mark .parametrize ("env_var_override" , [False , True ])
5796- def test_override_arch (arch , env_var_override , device , fresh_knobs_except_libraries ):
5796+ def test_override_arch (arch , env_var_override , device , fresh_knobs ):
57975797 if arch .startswith ("sm" ) and not is_cuda ():
57985798 pytest .skip (f"{ arch } arch only for CUDA" )
57995799 elif arch .startswith ("gfx" ) and not is_hip ():
@@ -5810,7 +5810,7 @@ def simple(data, out):
58105810
58115811 if is_cuda ():
58125812 if env_var_override :
5813- fresh_knobs_except_libraries .runtime .override_arch = str (arch )
5813+ fresh_knobs .runtime .override_arch = str (arch )
58145814 h = simple .warmup (data , out , grid = (1 , ))
58155815 else :
58165816 h = simple .warmup (data , out , arch = arch , grid = (1 , ))
@@ -5820,7 +5820,7 @@ def simple(data, out):
58205820 # For HIP, the generated kernel is a binary containing the final ISA. So we cannot run
58215821 # them like CUDA side if the chip doesn't match. Here we just check generated ISA.
58225822 if env_var_override :
5823- fresh_knobs_except_libraries .runtime .override_arch = str (arch )
5823+ fresh_knobs .runtime .override_arch = str (arch )
58245824 h = simple .warmup (data , out , grid = (1 , ))
58255825 else :
58265826 h = simple .warmup (data , out , arch = arch , grid = (1 , ))
0 commit comments