@@ -226,15 +226,15 @@ class TORCH_API Context {
226226 bool userEnabledMkldnn () const ;
227227 void setUserEnabledMkldnn (bool e);
228228 bool benchmarkCuDNN () const ;
229- void setBenchmarkCuDNN (bool );
229+ void setBenchmarkCuDNN (bool /* b */ );
230230 int benchmarkLimitCuDNN () const ;
231- void setBenchmarkLimitCuDNN (int );
231+ void setBenchmarkLimitCuDNN (int /* b */ );
232232 bool immediateMiopen () const ;
233- void setImmediateMiopen (bool );
233+ void setImmediateMiopen (bool /* b */ );
234234 bool deterministicCuDNN () const ;
235- void setDeterministicCuDNN (bool );
235+ void setDeterministicCuDNN (bool /* b */ );
236236 bool deterministicMkldnn () const ;
237- void setDeterministicMkldnn (bool );
237+ void setDeterministicMkldnn (bool /* b */ );
238238 bool userEnabledNNPACK () const ;
239239 void setUserEnabledNNPACK (bool e);
240240
@@ -252,32 +252,32 @@ class TORCH_API Context {
252252 void setSDPPriorityOrder (const std::vector<int64_t >& order);
253253 std::array<at::SDPBackend, at::num_sdp_backends> sDPPriorityOrder ();
254254
255- void setSDPUseFlash (bool );
255+ void setSDPUseFlash (bool /* e */ );
256256 bool userEnabledFlashSDP () const ;
257257
258- void setSDPUseMemEfficient (bool );
258+ void setSDPUseMemEfficient (bool /* e */ );
259259 bool userEnabledMemEfficientSDP () const ;
260260
261- void setSDPUseMath (bool );
261+ void setSDPUseMath (bool /* e */ );
262262 bool userEnabledMathSDP () const ;
263263
264- void setSDPUseCuDNN (bool );
264+ void setSDPUseCuDNN (bool /* e */ );
265265 bool userEnabledCuDNNSDP () const ;
266266
267- void setAllowFP16BF16ReductionMathSDP (bool );
267+ void setAllowFP16BF16ReductionMathSDP (bool /* e */ );
268268 bool allowFP16BF16ReductionMathSDP () const ;
269269
270- void setSDPUseOverrideable (bool );
270+ void setSDPUseOverrideable (bool /* e */ );
271271 bool userEnabledOverrideableSDP () const ;
272272
273273 at::LinalgBackend linalgPreferredBackend () const ;
274- void setLinalgPreferredBackend (at::LinalgBackend);
274+ void setLinalgPreferredBackend (at::LinalgBackend /* b */ );
275275
276276 at::BlasBackend blasPreferredBackend ();
277- void setBlasPreferredBackend (at::BlasBackend);
277+ void setBlasPreferredBackend (at::BlasBackend /* b */ );
278278
279279 at::ROCmFABackend getROCmFAPreferredBackend ();
280- void setROCmFAPreferredBackend (at::ROCmFABackend);
280+ void setROCmFAPreferredBackend (at::ROCmFABackend /* b */ );
281281
282282 // Note [Enabling Deterministic Operations]
283283 // ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
@@ -310,9 +310,9 @@ class TORCH_API Context {
310310
311311 bool deterministicAlgorithms () const ;
312312 bool deterministicAlgorithmsWarnOnly () const ;
313- void setDeterministicAlgorithms (bool , bool );
313+ void setDeterministicAlgorithms (bool /* b */ , bool /* warn_only */ );
314314 bool deterministicFillUninitializedMemory () const ;
315- void setDeterministicFillUninitializedMemory (bool );
315+ void setDeterministicFillUninitializedMemory (bool /* b */ );
316316
317317 // Note [Writing Nondeterministic Operations]
318318 // ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
@@ -356,11 +356,11 @@ class TORCH_API Context {
356356 Float32Op op,
357357 Float32Precision p);
358358 bool allowTF32CuDNN (std::optional<Float32Op> op = std::nullopt ) const ;
359- void setAllowTF32CuDNN (bool );
359+ void setAllowTF32CuDNN (bool /* b */ );
360360 bool allowTF32OneDNN () const ;
361- void setAllowTF32OneDNN (bool );
361+ void setAllowTF32OneDNN (bool /* b */ );
362362 bool allowTF32CuBLAS () const ;
363- void setAllowTF32CuBLAS (bool );
363+ void setAllowTF32CuBLAS (bool /* b */ );
364364 Float32MatmulPrecision float32MatmulPrecision () const ;
365365 Float32Precision float32Precision (Float32Backend backend, Float32Op op) const ;
366366 CuBLASReductionOption allowFP16ReductionCuBLAS () const ;
@@ -372,7 +372,7 @@ class TORCH_API Context {
372372 bool allow_reduced_precision,
373373 bool allow_splitk = true );
374374 bool allowFP16AccumulationCuBLAS () const ;
375- void setAllowFP16AccumulationCuBLAS (bool );
375+ void setAllowFP16AccumulationCuBLAS (bool /* b */ );
376376
377377 // Matmuls can use a so-called "persistent" kernel which launches one CUDA
378378 // block for each SM on the GPU, and each block then iterates over multiple
@@ -384,7 +384,7 @@ class TORCH_API Context {
384384 // to make matmuls target only a subset of the SMs, so they can fully schedule
385385 // even next to a comms kernel, and only be a few percent slower.
386386 std::optional<int32_t > _SMCarveout_EXPERIMENTAL () const ;
387- void _setSMCarveout_EXPERIMENTAL (std::optional<int32_t >);
387+ void _setSMCarveout_EXPERIMENTAL (std::optional<int32_t > /* c */ );
388388
389389 at::QEngine qEngine () const ;
390390 void setQEngine (at::QEngine e);
@@ -405,7 +405,7 @@ class TORCH_API Context {
405405 void setDefaultMobileCPUAllocator ();
406406 void unsetDefaultMobileCPUAllocator ();
407407 bool allowFP16ReductionCPU () const ;
408- void setAllowFP16ReductionCPU (bool );
408+ void setAllowFP16ReductionCPU (bool /* b */ );
409409
410410 // Preserved for BC
411411 void lazyInitCUDA () {
0 commit comments