From e2140a49197937d14a629bcc53ecd255ad2c30a7 Mon Sep 17 00:00:00 2001 From: Dan Johansson Date: Thu, 3 Jul 2025 13:29:16 +0200 Subject: [PATCH] Reduce test name redundancy Long test names cause the JUnit report files to exceed the size limit. By removing redundancies in the test names, the report size is reduced by 15%. Signed-off-by: Dan Johansson --- test/common/matmul_test_common.cpp | 12 ++++++------ test/tests/imatmul_test.cpp | 8 ++++---- test/tests/matmul_clamp_f16_bf16p_bf16p_test.cpp | 2 +- test/tests/matmul_clamp_f16_qai8dxp_qsi4cxp_test.cpp | 2 +- test/tests/matmul_clamp_f16_qai8dxp_qsi8cxp_test.cpp | 2 +- .../matmul_clamp_f16_qsi8d32p_qai4c32p_test.cpp | 2 +- test/tests/matmul_clamp_f32_bf16p_bf16p_test.cpp | 2 +- test/tests/matmul_clamp_f32_f32_f32p_test.cpp | 2 +- .../tests/matmul_clamp_f32_qai8dxp_qsi4c32p_test.cpp | 8 ++++---- test/tests/matmul_clamp_f32_qai8dxp_qsi4cxp_test.cpp | 12 ++++++------ test/tests/matmul_clamp_f32_qai8dxp_qsi8cxp_test.cpp | 8 ++++---- .../matmul_clamp_f32_qsi8d32p_qai4c32p_test.cpp | 2 +- .../matmul_clamp_f32_qsi8d32p_qsi4c32p_test.cpp | 6 +++--- test/tests/matmul_clamp_qai8_qai8p_qsi8cxp_test.cpp | 9 ++++----- test/tests/matmul_test.cpp | 8 ++++---- 15 files changed, 42 insertions(+), 43 deletions(-) diff --git a/test/common/matmul_test_common.cpp b/test/common/matmul_test_common.cpp index 371f7920..1cb4b5a0 100644 --- a/test/common/matmul_test_common.cpp +++ b/test/common/matmul_test_common.cpp @@ -17,7 +17,7 @@ std::ostream& operator<<(std::ostream& os, const MatMulShape& shape) { void PrintTo(const MatMulTestParams& param, std::ostream* os) { const auto& [method, shape, portion] = param; - *os << "Method_" << method.name << "__"; + *os << method.name << "__"; PrintTo(shape, os); *os << "__"; PrintTo(portion, os); @@ -28,17 +28,17 @@ void PrintTo(const MatMulShape& shape, std::ostream* os) { } void PrintTo(const MatrixPortion& portion, std::ostream* os) { - *os << "PortionStartRow_" << static_cast(portion.start_row() * 1000) // - << "__PortionStartCol_" << static_cast(portion.start_col() * 1000) // - << "__PortionHeight_" << static_cast(portion.height() * 1000) // - << "__PortionWidth_" << static_cast(portion.width() * 1000); + *os << "Portion__R_" << static_cast(portion.start_row() * 1000) // + << "__C_" << static_cast(portion.start_col() * 1000) // + << "__H_" << static_cast(portion.height() * 1000) // + << "__W_" << static_cast(portion.width() * 1000); } std::string test_description( const std::string_view& name, const MatMulShape& shape, const MatrixPortion& portion, bool bias) { std::ostringstream os; - os << "Method_" << name << "__"; + os << name << "__"; PrintTo(shape, &os); os << "__"; PrintTo(portion, &os); diff --git a/test/tests/imatmul_test.cpp b/test/tests/imatmul_test.cpp index 8f4922dd..ac2ceeb4 100644 --- a/test/tests/imatmul_test.cpp +++ b/test/tests/imatmul_test.cpp @@ -162,7 +162,7 @@ const IndirectMatMulArray& get_indirect_matmul_methods() { static IndirectMatMulArray indirect_matmul_methods{}; // F16 IMATMUL //////////////////////////////////////////////////////////// - indirect_matmul_methods[0].name = "indirect_matmul_f16_f16p_f16p_2vlx2vl_sme2_mopa"; + indirect_matmul_methods[0].name = "imatmul_f16_f16p_f16p_2vlx2vl_sme2_mopa"; indirect_matmul_methods[0].is_supported = cpu_has_sme2; indirect_matmul_methods[0].pack_shape.m = 2 * get_sme_vector_length(); indirect_matmul_methods[0].pack_shape.n = 2 * get_sme_vector_length(); @@ -201,7 +201,7 @@ const IndirectMatMulArray& get_indirect_matmul_methods() { indirect_matmul_methods[0].imatmul.imatmul = ukernel_f16.run_imatmul; // F32 IMATMUL //////////////////////////////////////////////////////////// - indirect_matmul_methods[1].name = "indirect_matmul_f32_f32p_f32p_2vlx2vl_sme2_mopa"; + indirect_matmul_methods[1].name = "imatmul_f32_f32p_f32p_2vlx2vl_sme2_mopa"; indirect_matmul_methods[1].is_supported = cpu_has_sme2; indirect_matmul_methods[1].pack_shape.m = 2 * get_sme_vector_length(); indirect_matmul_methods[1].pack_shape.n = 2 * get_sme_vector_length(); @@ -462,7 +462,7 @@ Buffer imatmul( TEST_P(IndirectMatMulTest, Output) { const auto& [method, shape, k_chunk_length, output_portion, clamp_rate] = GetParam(); if (not method.is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const KChunk k_chunk{shape.k, k_chunk_length}; @@ -493,7 +493,7 @@ TEST_P(IndirectMatMulTest, Output) { /// Name generator for test case [[maybe_unused]] static void PrintTo(const IndirectMatMulTestParams& param, std::ostream* os) { const auto& [method, shape, k_chunk_length, portion, clamp_rate] = param; - *os << "Method_" << method.name << "__"; + *os << method.name << "__"; PrintTo(shape, os); *os << "__K_chunk_length_" << k_chunk_length; *os << "__clamp_rate_" << static_cast(clamp_rate * 100) << "__"; diff --git a/test/tests/matmul_clamp_f16_bf16p_bf16p_test.cpp b/test/tests/matmul_clamp_f16_bf16p_bf16p_test.cpp index 66f92189..56a7bcbf 100644 --- a/test/tests/matmul_clamp_f16_bf16p_bf16p_test.cpp +++ b/test/tests/matmul_clamp_f16_bf16p_bf16p_test.cpp @@ -227,7 +227,7 @@ TEST_P(MatMulTestBf16OutFp16, Output) { const auto& [method, info, portion] = GetParam(); if (method.fn_is_supported && !method.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } if (!method.has_main_kernel()) { diff --git a/test/tests/matmul_clamp_f16_qai8dxp_qsi4cxp_test.cpp b/test/tests/matmul_clamp_f16_qai8dxp_qsi4cxp_test.cpp index 5b529876..a60cf6ee 100644 --- a/test/tests/matmul_clamp_f16_qai8dxp_qsi4cxp_test.cpp +++ b/test/tests/matmul_clamp_f16_qai8dxp_qsi4cxp_test.cpp @@ -61,7 +61,7 @@ TEST_P(MatMulTest_f16_qai8dxp_qsi4cxp, EndToEnd) { const auto& ukernel_variant = variants_kai_matmul_clamp_f16_qai8dxp_qsi4cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const std::uint32_t seed = 0; diff --git a/test/tests/matmul_clamp_f16_qai8dxp_qsi8cxp_test.cpp b/test/tests/matmul_clamp_f16_qai8dxp_qsi8cxp_test.cpp index 3a1710f8..e42586d1 100644 --- a/test/tests/matmul_clamp_f16_qai8dxp_qsi8cxp_test.cpp +++ b/test/tests/matmul_clamp_f16_qai8dxp_qsi8cxp_test.cpp @@ -62,7 +62,7 @@ TEST_P(MatMulTest_f16_qai8dxp_qsi8cxp, EndToEnd) { const auto& ukernel_variant = variants_kai_matmul_clamp_f16_qai8dxp_qsi8cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const std::uint32_t seed = 0; diff --git a/test/tests/matmul_clamp_f16_qsi8d32p_qai4c32p_test.cpp b/test/tests/matmul_clamp_f16_qsi8d32p_qai4c32p_test.cpp index 1a1f3acb..f9845f50 100644 --- a/test/tests/matmul_clamp_f16_qsi8d32p_qai4c32p_test.cpp +++ b/test/tests/matmul_clamp_f16_qsi8d32p_qai4c32p_test.cpp @@ -58,7 +58,7 @@ TEST_P(MatMulTest_f16_qsi8d32p_qai4c32p, EndToEnd) { const auto& ukernel_variant = variants_kai_matmul_clamp_f16_qsi8d32p_qai4c32p.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const std::uint32_t seed = 0; diff --git a/test/tests/matmul_clamp_f32_bf16p_bf16p_test.cpp b/test/tests/matmul_clamp_f32_bf16p_bf16p_test.cpp index 76ea30b7..fc4460ba 100644 --- a/test/tests/matmul_clamp_f32_bf16p_bf16p_test.cpp +++ b/test/tests/matmul_clamp_f32_bf16p_bf16p_test.cpp @@ -436,7 +436,7 @@ TEST_P(MatMulTestBf16, Output) { const auto& [method, info, portion] = GetParam(); if (method.fn_is_supported && !method.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } if (!method.has_main_kernel()) { diff --git a/test/tests/matmul_clamp_f32_f32_f32p_test.cpp b/test/tests/matmul_clamp_f32_f32_f32p_test.cpp index 3b37518a..3fdef040 100644 --- a/test/tests/matmul_clamp_f32_f32_f32p_test.cpp +++ b/test/tests/matmul_clamp_f32_f32_f32p_test.cpp @@ -75,7 +75,7 @@ TEST_P(MatMulTest_f32_f32_f32p, EndToEnd) // NOLINT(google-readability-avoid-un const auto& ukernel_variant = ukernel_variants.at(variant_idx); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } constexpr uint32_t seed = 0; diff --git a/test/tests/matmul_clamp_f32_qai8dxp_qsi4c32p_test.cpp b/test/tests/matmul_clamp_f32_qai8dxp_qsi4c32p_test.cpp index 5cd255d5..be8e2b71 100644 --- a/test/tests/matmul_clamp_f32_qai8dxp_qsi4c32p_test.cpp +++ b/test/tests/matmul_clamp_f32_qai8dxp_qsi4c32p_test.cpp @@ -88,7 +88,7 @@ TEST_P(MatMulTest_f32_qmatmul_clamp_f32_qai8dxp_qsi4c32p, Offset_RHS) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4c32p.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const size_t M = matmul_shape.m; @@ -125,7 +125,7 @@ TEST_P(MatMulTest_f32_qmatmul_clamp_f32_qai8dxp_qsi4c32p, Offset_LHS) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4c32p.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const size_t M = matmul_shape.m; @@ -156,7 +156,7 @@ TEST_P(MatMulTest_f32_qmatmul_clamp_f32_qai8dxp_qsi4c32p, EndToEnd_RHS_nxk) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4c32p.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } constexpr uint32_t seed = 0; @@ -348,7 +348,7 @@ TEST_P(MatMulTest_f32_qmatmul_clamp_f32_qai8dxp_qsi4c32p, EndToEnd_RHS_kxn) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4c32p.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const uint32_t seed = 0; diff --git a/test/tests/matmul_clamp_f32_qai8dxp_qsi4cxp_test.cpp b/test/tests/matmul_clamp_f32_qai8dxp_qsi4cxp_test.cpp index 37c37c39..d492943c 100644 --- a/test/tests/matmul_clamp_f32_qai8dxp_qsi4cxp_test.cpp +++ b/test/tests/matmul_clamp_f32_qai8dxp_qsi4cxp_test.cpp @@ -209,7 +209,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi4cxp, Offset_RHS) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const size_t M = matmul_shape.m; @@ -239,7 +239,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi4cxp, Offset_LHS) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const size_t M = matmul_shape.m; @@ -270,7 +270,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi4cxp, EndToEnd_RHS_nxk_qsi4cx) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } if (ukernel_variant.rhs_pack_type == RhsPackType::KxN) { GTEST_SKIP() << "Wrong type. This test for NxK"; @@ -401,7 +401,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi4cxp, EndToEnd_RHS_nxk_qsu4cx) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } if (ukernel_variant.rhs_pack_type == RhsPackType::KxN) { GTEST_SKIP() << "Wrong type. This test for NxK"; @@ -532,7 +532,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi4cxp, EndToEnd_RHS_kxn_qsi4cx) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } if (ukernel_variant.rhs_pack_type == RhsPackType::NxK) { GTEST_SKIP() << "Wrong type. This test for KxN"; @@ -667,7 +667,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi4cxp, EndToEnd_RHS_kxn_qsu4cx) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } if (ukernel_variant.rhs_pack_type == RhsPackType::NxK) { GTEST_SKIP() << "Wrong type. This test for KxN"; diff --git a/test/tests/matmul_clamp_f32_qai8dxp_qsi8cxp_test.cpp b/test/tests/matmul_clamp_f32_qai8dxp_qsi8cxp_test.cpp index d8db748a..315b5b0c 100644 --- a/test/tests/matmul_clamp_f32_qai8dxp_qsi8cxp_test.cpp +++ b/test/tests/matmul_clamp_f32_qai8dxp_qsi8cxp_test.cpp @@ -61,7 +61,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi8cxp, Offset_RHS) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi8cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const size_t K = matmul_shape.k; @@ -85,7 +85,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi8cxp, Offset_LHS) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi8cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const size_t K = matmul_shape.k; @@ -106,7 +106,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi8cxp, EndToEnd_RHS_nxk_qsi8cx) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi8cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const uint32_t seed = 0; @@ -220,7 +220,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi8cxp, EndToEnd_RHS_kxn_qsi8cx) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi8cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const uint32_t seed = 0; diff --git a/test/tests/matmul_clamp_f32_qsi8d32p_qai4c32p_test.cpp b/test/tests/matmul_clamp_f32_qsi8d32p_qai4c32p_test.cpp index fc703432..dfb5b2fb 100644 --- a/test/tests/matmul_clamp_f32_qsi8d32p_qai4c32p_test.cpp +++ b/test/tests/matmul_clamp_f32_qsi8d32p_qai4c32p_test.cpp @@ -129,7 +129,7 @@ TEST_P(MatMulTest_f32_qsi8d32p_qai4c32p, EndToEnd) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qsi8d32p_qai4c32p.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const std::uint32_t seed = 0; diff --git a/test/tests/matmul_clamp_f32_qsi8d32p_qsi4c32p_test.cpp b/test/tests/matmul_clamp_f32_qsi8d32p_qsi4c32p_test.cpp index 24326307..ac69e119 100644 --- a/test/tests/matmul_clamp_f32_qsi8d32p_qsi4c32p_test.cpp +++ b/test/tests/matmul_clamp_f32_qsi8d32p_qsi4c32p_test.cpp @@ -100,7 +100,7 @@ TEST_P(MatMulTest_f32_qsi8d32p_qsi4c32p, Offset_RHS) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qsi8d32p_qsi4c32p.at(variant_index); if (ukernel_variant.ukernel.fn_is_supported && !ukernel_variant.ukernel.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const size_t bl = 32; @@ -130,7 +130,7 @@ TEST_P(MatMulTest_f32_qsi8d32p_qsi4c32p, Offset_LHS) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qsi8d32p_qsi4c32p.at(variant_index); if (ukernel_variant.ukernel.fn_is_supported && !ukernel_variant.ukernel.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const size_t bl = 32; @@ -162,7 +162,7 @@ TEST_P(MatMulTest_f32_qsi8d32p_qsi4c32p, EndToEnd) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qsi8d32p_qsi4c32p.at(variant_index); if (ukernel_variant.ukernel.fn_is_supported && !ukernel_variant.ukernel.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const std::uint32_t seed = 0; diff --git a/test/tests/matmul_clamp_qai8_qai8p_qsi8cxp_test.cpp b/test/tests/matmul_clamp_qai8_qai8p_qsi8cxp_test.cpp index 05044804..b787fc38 100644 --- a/test/tests/matmul_clamp_qai8_qai8p_qsi8cxp_test.cpp +++ b/test/tests/matmul_clamp_qai8_qai8p_qsi8cxp_test.cpp @@ -278,7 +278,7 @@ const std::array& get_indirect_gemm_variants() { static const kai_imatmul_clamp_qai8_qai8p_qsi8cxp_ukernel& ukernel = get_imatmul_clamp_qai8_qai8_qsi8cxp2vlx4sb_1x16vl_sme2_dot_interface(); - variants[0].name = "indirect_matmul_qai8_qai8p_qsi8cxp"; + variants[0].name = "imatmul_qai8_qai8p_qsi8cxp"; variants[0].acc_pack.m = 2 * get_sme_vector_length(); variants[0].acc_pack.n = 2 * get_sme_vector_length(); variants[0].acc_pack.k = sizeof(int32_t) / sizeof(int8_t); @@ -733,8 +733,7 @@ static std::string test_description( const MatrixPortion& portion, size_t k_chunk_len, float clamp_ratio) { std::ostringstream sstream; - sstream << "Method_" << variant.name << "__M_" << shape.m << "__N_" << shape.n << "__k_chunk_count_" << shape.k - << "__"; + sstream << variant.name << "__M_" << shape.m << "__N_" << shape.n << "__k_chunk_count_" << shape.k << "__"; PrintTo(portion, &sstream); sstream << "__k_chunk_len_" << k_chunk_len << "__clamp_ratio_" << static_cast(clamp_ratio * 100); @@ -745,7 +744,7 @@ TEST_P(MatMulQuantizedTest, EndToEnd) { const auto& [variant, shape, output_portion, clamp_ratio] = GetParam(); if (!variant.is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } TestDataId test_data_id{shape, variant.acc_pack, shape.k, false, clamp_ratio}; @@ -879,7 +878,7 @@ TEST_P(IndirectMatMulQuantizedTest, EndToEnd) { MatMulShape shape{shape_k_chunk.m, shape_k_chunk.n, k_chunk.count * k_chunk.length}; if (!variant.is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } // Toggle padding testst when LHS has more than one row diff --git a/test/tests/matmul_test.cpp b/test/tests/matmul_test.cpp index dd893ec3..01404603 100644 --- a/test/tests/matmul_test.cpp +++ b/test/tests/matmul_test.cpp @@ -473,7 +473,7 @@ TEST_P(MatMulTest, PackedLhs) { const auto& [method, info, portion] = GetParam(); if (method.fn_is_supported && !method.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } if (!method.is_pack_lhs_needed()) { @@ -525,7 +525,7 @@ TEST_P(MatMulTest, PackedRhs) { const auto& [method, info, portion] = GetParam(); if (method.fn_is_supported && !method.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } if (!method.is_pack_rhs_needed()) { @@ -597,7 +597,7 @@ TEST_P(MatMulTest, PackedTransposedRhs) { const auto& [method, info, portion] = GetParam(); if (method.fn_is_supported && !method.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } if (!method.is_pack_rhs_nxk_needed()) { @@ -658,7 +658,7 @@ TEST_P(MatMulTest, Output) { const auto& [method, info, portion] = GetParam(); if (method.fn_is_supported && !method.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } if (!method.has_main_kernel()) { -- GitLab