From d352c74880d27c88a9923b697798074210cb8c8f Mon Sep 17 00:00:00 2001 From: Anton Bondarenko Date: Tue, 18 Feb 2025 23:28:30 +0100 Subject: [PATCH 1/2] Add message when skip tests Analyzing skip test w/o a proper report message is hard. Providing more details helps with that. Signed-off-by: Anton Bondarenko --- test/tests/bfloat16_test.cpp | 2 +- test/tests/float16_test.cpp | 2 +- .../matmul_clamp_f16_bf16p_bf16p_test.cpp | 6 ++-- .../matmul_clamp_f32_bf16p_bf16p_test.cpp | 6 ++-- test/tests/matmul_clamp_f32_f32_f32p_test.cpp | 2 +- ...matmul_clamp_f32_qai8dxp_qsi4c32p_test.cpp | 16 +++++----- .../matmul_clamp_f32_qai8dxp_qsi4cxp_test.cpp | 32 +++++++++---------- .../matmul_clamp_f32_qai8dxp_qsi8cxp_test.cpp | 12 +++---- ...atmul_clamp_f32_qsi8d32p_qsi4c32p_test.cpp | 12 +++---- test/tests/matmul_test.cpp | 24 +++++++------- 10 files changed, 57 insertions(+), 57 deletions(-) diff --git a/test/tests/bfloat16_test.cpp b/test/tests/bfloat16_test.cpp index a8d4bb88..03334a2e 100644 --- a/test/tests/bfloat16_test.cpp +++ b/test/tests/bfloat16_test.cpp @@ -14,7 +14,7 @@ namespace kai::test { TEST(BFloat16, SimpleTest) { if (!cpu_has_bf16()) { - GTEST_SKIP(); + GTEST_SKIP() << "No CPU support for BFloat16"; } ASSERT_EQ(static_cast(BFloat16()), 0.0F); diff --git a/test/tests/float16_test.cpp b/test/tests/float16_test.cpp index 66abe88c..6e286971 100644 --- a/test/tests/float16_test.cpp +++ b/test/tests/float16_test.cpp @@ -14,7 +14,7 @@ namespace kai::test { TEST(Float16, SimpleTest) { if (!cpu_has_fp16()) { - GTEST_SKIP(); + GTEST_SKIP() << "No CPU support for FP16"; } ASSERT_EQ(static_cast(Float16()), 0.0F); diff --git a/test/tests/matmul_clamp_f16_bf16p_bf16p_test.cpp b/test/tests/matmul_clamp_f16_bf16p_bf16p_test.cpp index cb73302d..2e3469c5 100644 --- a/test/tests/matmul_clamp_f16_bf16p_bf16p_test.cpp +++ b/test/tests/matmul_clamp_f16_bf16p_bf16p_test.cpp @@ -242,11 +242,11 @@ TEST_P(MatMulTestBf16OutFp16, Output) { const auto& [method, info, portion] = GetParam(); if (method.fn_is_supported && !method.fn_is_supported()) { - GTEST_SKIP(); + GTEST_SKIP() << "CPU features are not supported by current CPU"; } if (!method.has_main_kernel()) { - GTEST_SKIP(); + GTEST_SKIP() << "No main kernel available"; } const auto& data = test_data(); @@ -260,7 +260,7 @@ TEST_P(MatMulTestBf16OutFp16, Output) { const auto rect = portion.compute_portion(info.m, info.n, method.m0, method.n0); if (rect.height() == 0 || rect.width() == 0) { - GTEST_SKIP(); + GTEST_SKIP() << "Empty dimension of matrix(" << rect.width() << "," << rect.height() << ")"; } const size_t lhs_w = info.k; diff --git a/test/tests/matmul_clamp_f32_bf16p_bf16p_test.cpp b/test/tests/matmul_clamp_f32_bf16p_bf16p_test.cpp index f59174a6..9fff8005 100644 --- a/test/tests/matmul_clamp_f32_bf16p_bf16p_test.cpp +++ b/test/tests/matmul_clamp_f32_bf16p_bf16p_test.cpp @@ -497,11 +497,11 @@ TEST_P(MatMulTestBf16, Output) { const auto& [method, info, portion] = GetParam(); if (method.fn_is_supported && !method.fn_is_supported()) { - GTEST_SKIP(); + GTEST_SKIP() << "CPU features are not supported by current CPU"; } if (!method.has_main_kernel()) { - GTEST_SKIP(); + GTEST_SKIP() << "No main kernel available"; } const auto& data = test_data(); @@ -514,7 +514,7 @@ TEST_P(MatMulTestBf16, Output) { const auto rect = portion.compute_portion(info.m, info.n, m_step, n_step); if (rect.height() == 0 || rect.width() == 0) { - GTEST_SKIP(); + GTEST_SKIP() << "Empty dimension of matrix(" << rect.width() << "," << rect.height() << ")"; } const size_t lhs_w = info.k; diff --git a/test/tests/matmul_clamp_f32_f32_f32p_test.cpp b/test/tests/matmul_clamp_f32_f32_f32p_test.cpp index 4ac1f986..94f61255 100644 --- a/test/tests/matmul_clamp_f32_f32_f32p_test.cpp +++ b/test/tests/matmul_clamp_f32_f32_f32p_test.cpp @@ -68,7 +68,7 @@ TEST_P(MatMulTest_f32_f32_f32p, EndToEnd) // NOLINT(google-readability-avoid-un const auto& ukernel_variant = ukernel_variants.at(variant_idx); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP(); + GTEST_SKIP() << "CPU features are not supported by current CPU"; } constexpr uint32_t seed = 0; diff --git a/test/tests/matmul_clamp_f32_qai8dxp_qsi4c32p_test.cpp b/test/tests/matmul_clamp_f32_qai8dxp_qsi4c32p_test.cpp index 06b50043..954496bf 100644 --- a/test/tests/matmul_clamp_f32_qai8dxp_qsi4c32p_test.cpp +++ b/test/tests/matmul_clamp_f32_qai8dxp_qsi4c32p_test.cpp @@ -91,7 +91,7 @@ TEST_P(MatMulTest_f32_qmatmul_clamp_f32_qai8dxp_qsi4c32p, Offset_RHS) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4c32p.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP(); + GTEST_SKIP() << "CPU features are not supported by current CPU"; } const size_t M = matmul_shape.m; @@ -103,7 +103,7 @@ TEST_P(MatMulTest_f32_qmatmul_clamp_f32_qai8dxp_qsi4c32p, Offset_RHS) { const auto rect = portion.compute_portion(M, N, m_step, n_step); if (rect.height() == 0 || rect.width() == 0) { - GTEST_SKIP(); + GTEST_SKIP() << "Empty dimension of matrix(" << rect.width() << "," << rect.height() << ")"; } const auto nr = ukernel_variant.interface.get_nr(); @@ -128,7 +128,7 @@ TEST_P(MatMulTest_f32_qmatmul_clamp_f32_qai8dxp_qsi4c32p, Offset_LHS) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4c32p.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP(); + GTEST_SKIP() << "CPU features are not supported by current CPU"; } const size_t M = matmul_shape.m; @@ -140,7 +140,7 @@ TEST_P(MatMulTest_f32_qmatmul_clamp_f32_qai8dxp_qsi4c32p, Offset_LHS) { const auto rect = portion.compute_portion(M, N, m_step, n_step); if (rect.height() == 0 || rect.width() == 0) { - GTEST_SKIP(); + GTEST_SKIP() << "Empty dimension of matrix(" << rect.width() << "," << rect.height() << ")"; } const auto mr = ukernel_variant.interface.get_mr(); @@ -159,7 +159,7 @@ TEST_P(MatMulTest_f32_qmatmul_clamp_f32_qai8dxp_qsi4c32p, EndToEnd_RHS_nxk) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4c32p.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP(); + GTEST_SKIP() << "CPU features are not supported by current CPU"; } constexpr uint32_t seed = 0; @@ -201,7 +201,7 @@ TEST_P(MatMulTest_f32_qmatmul_clamp_f32_qai8dxp_qsi4c32p, EndToEnd_RHS_nxk) { const auto rect = portion.compute_portion(M, N, m_step, n_step); if (rect.height() == 0 || rect.width() == 0) { - GTEST_SKIP(); + GTEST_SKIP() << "Empty dimension of matrix(" << rect.width() << "," << rect.height() << ")"; } const auto lhs_start_row = rect.start_row(); @@ -293,7 +293,7 @@ TEST_P(MatMulTest_f32_qmatmul_clamp_f32_qai8dxp_qsi4c32p, EndToEnd_RHS_kxn) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4c32p.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP(); + GTEST_SKIP() << "CPU features are not supported by current CPU"; } const uint32_t seed = 0; @@ -347,7 +347,7 @@ TEST_P(MatMulTest_f32_qmatmul_clamp_f32_qai8dxp_qsi4c32p, EndToEnd_RHS_kxn) { const auto rect = portion.compute_portion(M, N, m_step, n_step); if (rect.height() == 0 || rect.width() == 0) { - GTEST_SKIP(); + GTEST_SKIP() << "Empty dimension of matrix(" << rect.width() << "," << rect.height() << ")"; } const auto lhs_start_row = rect.start_row(); diff --git a/test/tests/matmul_clamp_f32_qai8dxp_qsi4cxp_test.cpp b/test/tests/matmul_clamp_f32_qai8dxp_qsi4cxp_test.cpp index 22edc8c1..eaaa6b06 100644 --- a/test/tests/matmul_clamp_f32_qai8dxp_qsi4cxp_test.cpp +++ b/test/tests/matmul_clamp_f32_qai8dxp_qsi4cxp_test.cpp @@ -210,7 +210,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi4cxp, Offset_RHS) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP(); + GTEST_SKIP() << "CPU features are not supported by current CPU"; } const size_t M = matmul_shape.m; @@ -222,7 +222,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi4cxp, Offset_RHS) { const auto rect = portion.compute_portion(M, N, m_step, n_step); if (rect.height() == 0 || rect.width() == 0) { - GTEST_SKIP(); + GTEST_SKIP() << "Empty dimension of matrix(" << rect.width() << "," << rect.height() << ")"; } const auto nr = ukernel_variant.interface.get_nr(); @@ -240,7 +240,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi4cxp, Offset_LHS) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP(); + GTEST_SKIP() << "CPU features are not supported by current CPU"; } const size_t M = matmul_shape.m; @@ -252,7 +252,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi4cxp, Offset_LHS) { const auto rect = portion.compute_portion(M, N, m_step, n_step); if (rect.height() == 0 || rect.width() == 0) { - GTEST_SKIP(); + GTEST_SKIP() << "Empty dimension of matrix(" << rect.width() << "," << rect.height() << ")"; } const auto mr = ukernel_variant.interface.get_mr(); @@ -271,10 +271,10 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi4cxp, EndToEnd_RHS_nxk_qsi4cx) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP(); + GTEST_SKIP() << "CPU features are not supported by current CPU"; } if (ukernel_variant.rhs_pack_type == RhsPackType::KxN) { - GTEST_SKIP(); // Wrong type. This test for NxK + GTEST_SKIP() << "Wrong type. This test for NxK"; } if (!ukernel_variant.signed_integer_support) { GTEST_SKIP() << "Signed integer input unsupported"; @@ -318,7 +318,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi4cxp, EndToEnd_RHS_nxk_qsi4cx) { const auto rect = portion.compute_portion(M, N, m_step, n_step); if (rect.height() == 0 || rect.width() == 0) { - GTEST_SKIP(); + GTEST_SKIP() << "Empty dimension of matrix(" << rect.width() << "," << rect.height() << ")"; } const auto lhs_start_row = rect.start_row(); @@ -398,10 +398,10 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi4cxp, EndToEnd_RHS_nxk_qsu4cx) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP(); + GTEST_SKIP() << "CPU features are not supported by current CPU"; } if (ukernel_variant.rhs_pack_type == RhsPackType::KxN) { - GTEST_SKIP(); // Wrong type. This test for NxK + GTEST_SKIP() << "Wrong type. This test for NxK"; } const uint32_t seed = 0; @@ -442,7 +442,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi4cxp, EndToEnd_RHS_nxk_qsu4cx) { const auto rect = portion.compute_portion(M, N, m_step, n_step); if (rect.height() == 0 || rect.width() == 0) { - GTEST_SKIP(); + GTEST_SKIP() << "Empty dimension of matrix(" << rect.width() << "," << rect.height() << ")"; } const auto lhs_start_row = rect.start_row(); @@ -522,10 +522,10 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi4cxp, EndToEnd_RHS_kxn_qsi4cx) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP(); + GTEST_SKIP() << "CPU features are not supported by current CPU"; } if (ukernel_variant.rhs_pack_type == RhsPackType::NxK) { - GTEST_SKIP(); // Wrong type. This test for KxN + GTEST_SKIP() << "Wrong type. This test for KxN"; } if (!ukernel_variant.signed_integer_support) { GTEST_SKIP() << "Signed integer input unsupported"; @@ -580,7 +580,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi4cxp, EndToEnd_RHS_kxn_qsi4cx) { const auto rect = portion.compute_portion(M, N, m_step, n_step); if (rect.height() == 0 || rect.width() == 0) { - GTEST_SKIP(); + GTEST_SKIP() << "Empty dimension of matrix(" << rect.width() << "," << rect.height() << ")"; } const auto lhs_start_row = rect.start_row(); @@ -653,10 +653,10 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi4cxp, EndToEnd_RHS_kxn_qsu4cx) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP(); + GTEST_SKIP() << "CPU features are not supported by current CPU"; } if (ukernel_variant.rhs_pack_type == RhsPackType::NxK) { - GTEST_SKIP(); // Wrong type. This test for KxN + GTEST_SKIP() << "Wrong type. This test for KxN"; } const uint32_t seed = 0; @@ -709,7 +709,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi4cxp, EndToEnd_RHS_kxn_qsu4cx) { const auto rect = portion.compute_portion(M, N, m_step, n_step); if (rect.height() == 0 || rect.width() == 0) { - GTEST_SKIP(); + GTEST_SKIP() << "Empty dimension of matrix(" << rect.width() << "," << rect.height() << ")"; } const auto lhs_start_row = rect.start_row(); diff --git a/test/tests/matmul_clamp_f32_qai8dxp_qsi8cxp_test.cpp b/test/tests/matmul_clamp_f32_qai8dxp_qsi8cxp_test.cpp index eab1e6a7..f55213de 100644 --- a/test/tests/matmul_clamp_f32_qai8dxp_qsi8cxp_test.cpp +++ b/test/tests/matmul_clamp_f32_qai8dxp_qsi8cxp_test.cpp @@ -53,7 +53,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi8cxp, Offset_RHS) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi8cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP(); + GTEST_SKIP() << "CPU features are not supported by current CPU"; } const size_t K = matmul_shape.k; @@ -77,7 +77,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi8cxp, Offset_LHS) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi8cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP(); + GTEST_SKIP() << "CPU features are not supported by current CPU"; } const size_t K = matmul_shape.k; @@ -98,7 +98,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi8cxp, EndToEnd_RHS_nxk_qsi8cx) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi8cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP(); + GTEST_SKIP() << "CPU features are not supported by current CPU"; } const uint32_t seed = 0; @@ -139,7 +139,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi8cxp, EndToEnd_RHS_nxk_qsi8cx) { const auto rect = portion.compute_portion(M, N, m_step, n_step); if (rect.height() == 0 || rect.width() == 0) { - GTEST_SKIP() << "Skipping empty portion."; + GTEST_SKIP() << "Empty dimension of matrix(" << rect.width() << "," << rect.height() << ")"; } // Runs the LHS packing micro-kernel. @@ -212,7 +212,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi8cxp, EndToEnd_RHS_kxn_qsi8cx) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi8cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP(); + GTEST_SKIP() << "CPU features are not supported by current CPU"; } const uint32_t seed = 0; @@ -264,7 +264,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi8cxp, EndToEnd_RHS_kxn_qsi8cx) { const auto rect = portion.compute_portion(M, N, m_step, n_step); if (rect.height() == 0 || rect.width() == 0) { - GTEST_SKIP() << "Skipping empty portion."; + GTEST_SKIP() << "Empty dimension of matrix(" << rect.width() << "," << rect.height() << ")"; } const auto lhs_start_row = rect.start_row(); diff --git a/test/tests/matmul_clamp_f32_qsi8d32p_qsi4c32p_test.cpp b/test/tests/matmul_clamp_f32_qsi8d32p_qsi4c32p_test.cpp index b91c1fbc..153decb5 100644 --- a/test/tests/matmul_clamp_f32_qsi8d32p_qsi4c32p_test.cpp +++ b/test/tests/matmul_clamp_f32_qsi8d32p_qsi4c32p_test.cpp @@ -102,7 +102,7 @@ TEST_P(MatMulTest_f32_qsi8d32p_qsi4c32p, Offset_RHS) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qsi8d32p_qsi4c32p.at(variant_index); if (ukernel_variant.ukernel.fn_is_supported && !ukernel_variant.ukernel.fn_is_supported()) { - GTEST_SKIP(); + GTEST_SKIP() << "CPU features are not supported by current CPU"; } const size_t bl = 32; @@ -118,7 +118,7 @@ TEST_P(MatMulTest_f32_qsi8d32p_qsi4c32p, Offset_RHS) { const auto rect = portion.compute_portion(M, N, m_step, n_step); if (rect.height() == 0 || rect.width() == 0) { - GTEST_SKIP() << "Test Portion size is 0!"; + GTEST_SKIP() << "Empty dimension of matrix(" << rect.width() << "," << rect.height() << ")"; } const auto rhs_start_row = rect.start_col(); @@ -132,7 +132,7 @@ TEST_P(MatMulTest_f32_qsi8d32p_qsi4c32p, Offset_LHS) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qsi8d32p_qsi4c32p.at(variant_index); if (ukernel_variant.ukernel.fn_is_supported && !ukernel_variant.ukernel.fn_is_supported()) { - GTEST_SKIP(); + GTEST_SKIP() << "CPU features are not supported by current CPU"; } const size_t bl = 32; @@ -149,7 +149,7 @@ TEST_P(MatMulTest_f32_qsi8d32p_qsi4c32p, Offset_LHS) { const auto rect = portion.compute_portion(M, N, m_step, n_step); if (rect.height() == 0 || rect.width() == 0) { - GTEST_SKIP() << "Test Portion size is 0!"; + GTEST_SKIP() << "Empty dimension of matrix(" << rect.width() << "," << rect.height() << ")"; } const auto lhs_start_row = rect.start_row(); @@ -164,7 +164,7 @@ TEST_P(MatMulTest_f32_qsi8d32p_qsi4c32p, EndToEnd) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qsi8d32p_qsi4c32p.at(variant_index); if (ukernel_variant.ukernel.fn_is_supported && !ukernel_variant.ukernel.fn_is_supported()) { - GTEST_SKIP(); + GTEST_SKIP() << "CPU features are not supported by current CPU"; } const std::uint32_t seed = 0; @@ -191,7 +191,7 @@ TEST_P(MatMulTest_f32_qsi8d32p_qsi4c32p, EndToEnd) { const auto rect = portion.compute_portion(M, N, m_step, n_step); if (rect.height() == 0 || rect.width() == 0) { - GTEST_SKIP() << "Test Portion size is 0!"; + GTEST_SKIP() << "Empty dimension of matrix(" << rect.width() << "," << rect.height() << ")"; } // Generates input data. const auto ref_lhs = fill_random(M * K, seed + 0); diff --git a/test/tests/matmul_test.cpp b/test/tests/matmul_test.cpp index 387f9e6e..d9f20665 100644 --- a/test/tests/matmul_test.cpp +++ b/test/tests/matmul_test.cpp @@ -456,11 +456,11 @@ TEST_P(MatMulTest, PackedLhs) { const auto& [method, info, portion] = GetParam(); if (method.fn_is_supported && !method.fn_is_supported()) { - GTEST_SKIP(); + GTEST_SKIP() << "CPU features are not supported by current CPU"; } if (!method.is_pack_lhs_needed()) { - GTEST_SKIP(); + GTEST_SKIP() << "Test not valid w/o LHS pack"; } const auto& data = test_data(); @@ -472,7 +472,7 @@ TEST_P(MatMulTest, PackedLhs) { lhs_w); // LHS packing micro-kernel API doesn't support scheduling over K dimension. if (rect.height() == 0 || rect.width() == 0) { - GTEST_SKIP(); + GTEST_SKIP() << "Empty dimension of matrix(" << rect.width() << "," << rect.height() << ")"; } const auto mr = method.fn_get_mr(); @@ -509,11 +509,11 @@ TEST_P(MatMulTest, PackedRhs) { const auto& [method, info, portion] = GetParam(); if (method.fn_is_supported && !method.fn_is_supported()) { - GTEST_SKIP(); + GTEST_SKIP() << "CPU features are not supported by current CPU"; } if (!method.is_pack_rhs_needed()) { - GTEST_SKIP(); + GTEST_SKIP() << "Test not valid w/o RHS pack"; } const auto& data = test_data(); @@ -526,7 +526,7 @@ TEST_P(MatMulTest, PackedRhs) { const Rect rect = portion.compute_portion(rhs_full_width, rhs_full_height, block_height, block_width); if (rect.height() == 0 || rect.width() == 0) { - GTEST_SKIP(); + GTEST_SKIP() << "Empty dimension of matrix(" << rect.width() << "," << rect.height() << ")"; } const auto rhs_start_row = rect.start_row(); @@ -581,11 +581,11 @@ TEST_P(MatMulTest, PackedTransposedRhs) { const auto& [method, info, portion] = GetParam(); if (method.fn_is_supported && !method.fn_is_supported()) { - GTEST_SKIP(); + GTEST_SKIP() << "CPU features are not supported by current CPU"; } if (!method.is_pack_rhs_nxk_needed()) { - GTEST_SKIP(); + GTEST_SKIP() << "Test not valid w/o pre-processing of transposed RHS matrix"; } const auto& data = test_data(); @@ -598,7 +598,7 @@ TEST_P(MatMulTest, PackedTransposedRhs) { method.packed_rhs_format.scheduler_block_width(info.k)); if (rect.height() == 0 || rect.width() == 0) { - GTEST_SKIP(); + GTEST_SKIP() << "Empty dimension of matrix(" << rect.width() << "," << rect.height() << ")"; } const auto ref_rhs_row_stride = method.rhs_format.default_row_stride(info.k); @@ -643,11 +643,11 @@ TEST_P(MatMulTest, Output) { const auto& [method, info, portion] = GetParam(); if (method.fn_is_supported && !method.fn_is_supported()) { - GTEST_SKIP(); + GTEST_SKIP() << "CPU features are not supported by current CPU"; } if (!method.has_main_kernel()) { - GTEST_SKIP(); + GTEST_SKIP() << "No main kernel available"; } const auto& data = test_data(); @@ -660,7 +660,7 @@ TEST_P(MatMulTest, Output) { const auto rect = portion.compute_portion(info.m, info.n, method.m0, method.n0); if (rect.height() == 0 || rect.width() == 0) { - GTEST_SKIP(); + GTEST_SKIP() << "Empty dimension of matrix(" << rect.width() << "," << rect.height() << ")"; } const auto lhs_w = info.k; -- GitLab From a516c260c072850af24603149e401e67976ff5ab Mon Sep 17 00:00:00 2001 From: Anton Bondarenko Date: Fri, 7 Mar 2025 14:24:07 +0100 Subject: [PATCH 2/2] Fix copyright years Signed-off-by: Anton Bondarenko --- test/tests/bfloat16_test.cpp | 2 +- test/tests/float16_test.cpp | 2 +- test/tests/matmul_clamp_f16_bf16p_bf16p_test.cpp | 2 +- test/tests/matmul_clamp_f32_f32_f32p_test.cpp | 2 +- test/tests/matmul_test.cpp | 2 +- 5 files changed, 5 insertions(+), 5 deletions(-) diff --git a/test/tests/bfloat16_test.cpp b/test/tests/bfloat16_test.cpp index 03334a2e..8a2886c8 100644 --- a/test/tests/bfloat16_test.cpp +++ b/test/tests/bfloat16_test.cpp @@ -1,5 +1,5 @@ // -// SPDX-FileCopyrightText: Copyright 2024 Arm Limited and/or its affiliates +// SPDX-FileCopyrightText: Copyright 2024-2025 Arm Limited and/or its affiliates // // SPDX-License-Identifier: Apache-2.0 // diff --git a/test/tests/float16_test.cpp b/test/tests/float16_test.cpp index 6e286971..ea919fdd 100644 --- a/test/tests/float16_test.cpp +++ b/test/tests/float16_test.cpp @@ -1,5 +1,5 @@ // -// SPDX-FileCopyrightText: Copyright 2024 Arm Limited and/or its affiliates +// SPDX-FileCopyrightText: Copyright 2024-2025 Arm Limited and/or its affiliates // // SPDX-License-Identifier: Apache-2.0 // diff --git a/test/tests/matmul_clamp_f16_bf16p_bf16p_test.cpp b/test/tests/matmul_clamp_f16_bf16p_bf16p_test.cpp index 2e3469c5..038e3615 100644 --- a/test/tests/matmul_clamp_f16_bf16p_bf16p_test.cpp +++ b/test/tests/matmul_clamp_f16_bf16p_bf16p_test.cpp @@ -1,5 +1,5 @@ // -// SPDX-FileCopyrightText: Copyright 2024 Arm Limited and/or its affiliates +// SPDX-FileCopyrightText: Copyright 2024-2025 Arm Limited and/or its affiliates // // SPDX-License-Identifier: Apache-2.0 // diff --git a/test/tests/matmul_clamp_f32_f32_f32p_test.cpp b/test/tests/matmul_clamp_f32_f32_f32p_test.cpp index 94f61255..2f6fd131 100644 --- a/test/tests/matmul_clamp_f32_f32_f32p_test.cpp +++ b/test/tests/matmul_clamp_f32_f32_f32p_test.cpp @@ -1,5 +1,5 @@ // -// SPDX-FileCopyrightText: Copyright 2024 Arm Limited and/or its affiliates +// SPDX-FileCopyrightText: Copyright 2024-2025 Arm Limited and/or its affiliates // // SPDX-License-Identifier: Apache-2.0 // diff --git a/test/tests/matmul_test.cpp b/test/tests/matmul_test.cpp index d9f20665..dd4b0dd8 100644 --- a/test/tests/matmul_test.cpp +++ b/test/tests/matmul_test.cpp @@ -1,5 +1,5 @@ // -// SPDX-FileCopyrightText: Copyright 2024 Arm Limited and/or its affiliates +// SPDX-FileCopyrightText: Copyright 2024-2025 Arm Limited and/or its affiliates // // SPDX-License-Identifier: Apache-2.0 // -- GitLab