diff --git a/deep_gemm/include/deep_gemm/fp8_gemm.cuh b/deep_gemm/include/deep_gemm/fp8_gemm.cuh index 386139c..4029aa8 100644 --- a/deep_gemm/include/deep_gemm/fp8_gemm.cuh +++ b/deep_gemm/include/deep_gemm/fp8_gemm.cuh @@ -271,7 +271,7 @@ fp8_gemm_kernel(float* scales_b, int* grouped_layout, } }; - if (!scheduler.is_valid_m(math_wg_idx * WGMMA::M, m_block_idx)) { + if (!scheduler.is_m_valid(math_wg_idx * WGMMA::M, m_block_idx)) { // Skip useless computation for unaligned Ms launch_k_iterations([&](int k_iter, auto type, auto _) { #pragma unroll diff --git a/deep_gemm/include/deep_gemm/scheduler.cuh b/deep_gemm/include/deep_gemm/scheduler.cuh index 8b6dbf3..97e7ee5 100644 --- a/deep_gemm/include/deep_gemm/scheduler.cuh +++ b/deep_gemm/include/deep_gemm/scheduler.cuh @@ -48,7 +48,7 @@ struct Scheduler { } } - __device__ __forceinline__ bool is_valid_m(const uint32_t m_offset, const uint32_t& m_block_idx) const { + __device__ __forceinline__ bool is_m_valid(const uint32_t m_offset, const uint32_t& m_block_idx) const { if constexpr (kGemmType == GemmType::Normal) { return true; } else if constexpr (kGemmType == GemmType::GroupedContiguous) {