diff --git a/test/common/matmul_test_common.cpp b/test/common/matmul_test_common.cpp index 371f792035f635bb2f22c78db08bd531d9d72d75..1cb4b5a02e7217d91fe65dcd488a738fa3874252 100644 --- a/test/common/matmul_test_common.cpp +++ b/test/common/matmul_test_common.cpp @@ -17,7 +17,7 @@ std::ostream& operator<<(std::ostream& os, const MatMulShape& shape) { void PrintTo(const MatMulTestParams& param, std::ostream* os) { const auto& [method, shape, portion] = param; - *os << "Method_" << method.name << "__"; + *os << method.name << "__"; PrintTo(shape, os); *os << "__"; PrintTo(portion, os); @@ -28,17 +28,17 @@ void PrintTo(const MatMulShape& shape, std::ostream* os) { } void PrintTo(const MatrixPortion& portion, std::ostream* os) { - *os << "PortionStartRow_" << static_cast(portion.start_row() * 1000) // - << "__PortionStartCol_" << static_cast(portion.start_col() * 1000) // - << "__PortionHeight_" << static_cast(portion.height() * 1000) // - << "__PortionWidth_" << static_cast(portion.width() * 1000); + *os << "Portion__R_" << static_cast(portion.start_row() * 1000) // + << "__C_" << static_cast(portion.start_col() * 1000) // + << "__H_" << static_cast(portion.height() * 1000) // + << "__W_" << static_cast(portion.width() * 1000); } std::string test_description( const std::string_view& name, const MatMulShape& shape, const MatrixPortion& portion, bool bias) { std::ostringstream os; - os << "Method_" << name << "__"; + os << name << "__"; PrintTo(shape, &os); os << "__"; PrintTo(portion, &os); diff --git a/test/tests/imatmul_test.cpp b/test/tests/imatmul_test.cpp index 8f4922dd51f8e7575cc36e8239bab123dd2d9d5f..ac2ceeb4c61ca547bc8185b4d4526d20ef77d2e7 100644 --- a/test/tests/imatmul_test.cpp +++ b/test/tests/imatmul_test.cpp @@ -162,7 +162,7 @@ const IndirectMatMulArray& get_indirect_matmul_methods() { static IndirectMatMulArray indirect_matmul_methods{}; // F16 IMATMUL //////////////////////////////////////////////////////////// - indirect_matmul_methods[0].name = "indirect_matmul_f16_f16p_f16p_2vlx2vl_sme2_mopa"; + indirect_matmul_methods[0].name = "imatmul_f16_f16p_f16p_2vlx2vl_sme2_mopa"; indirect_matmul_methods[0].is_supported = cpu_has_sme2; indirect_matmul_methods[0].pack_shape.m = 2 * get_sme_vector_length(); indirect_matmul_methods[0].pack_shape.n = 2 * get_sme_vector_length(); @@ -201,7 +201,7 @@ const IndirectMatMulArray& get_indirect_matmul_methods() { indirect_matmul_methods[0].imatmul.imatmul = ukernel_f16.run_imatmul; // F32 IMATMUL //////////////////////////////////////////////////////////// - indirect_matmul_methods[1].name = "indirect_matmul_f32_f32p_f32p_2vlx2vl_sme2_mopa"; + indirect_matmul_methods[1].name = "imatmul_f32_f32p_f32p_2vlx2vl_sme2_mopa"; indirect_matmul_methods[1].is_supported = cpu_has_sme2; indirect_matmul_methods[1].pack_shape.m = 2 * get_sme_vector_length(); indirect_matmul_methods[1].pack_shape.n = 2 * get_sme_vector_length(); @@ -462,7 +462,7 @@ Buffer imatmul( TEST_P(IndirectMatMulTest, Output) { const auto& [method, shape, k_chunk_length, output_portion, clamp_rate] = GetParam(); if (not method.is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const KChunk k_chunk{shape.k, k_chunk_length}; @@ -493,7 +493,7 @@ TEST_P(IndirectMatMulTest, Output) { /// Name generator for test case [[maybe_unused]] static void PrintTo(const IndirectMatMulTestParams& param, std::ostream* os) { const auto& [method, shape, k_chunk_length, portion, clamp_rate] = param; - *os << "Method_" << method.name << "__"; + *os << method.name << "__"; PrintTo(shape, os); *os << "__K_chunk_length_" << k_chunk_length; *os << "__clamp_rate_" << static_cast(clamp_rate * 100) << "__"; diff --git a/test/tests/matmul_clamp_f16_bf16p_bf16p_test.cpp b/test/tests/matmul_clamp_f16_bf16p_bf16p_test.cpp index 66f92189bb1a3490b3770ef0309d077eba9da17f..56a7bcbf80122070b6968dd63925427222f9277e 100644 --- a/test/tests/matmul_clamp_f16_bf16p_bf16p_test.cpp +++ b/test/tests/matmul_clamp_f16_bf16p_bf16p_test.cpp @@ -227,7 +227,7 @@ TEST_P(MatMulTestBf16OutFp16, Output) { const auto& [method, info, portion] = GetParam(); if (method.fn_is_supported && !method.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } if (!method.has_main_kernel()) { diff --git a/test/tests/matmul_clamp_f16_qai8dxp_qsi4cxp_test.cpp b/test/tests/matmul_clamp_f16_qai8dxp_qsi4cxp_test.cpp index 5b529876b4928604f592b9483035e53617cdd834..a60cf6eee2b625ea406e0ec39708edba5360c2cb 100644 --- a/test/tests/matmul_clamp_f16_qai8dxp_qsi4cxp_test.cpp +++ b/test/tests/matmul_clamp_f16_qai8dxp_qsi4cxp_test.cpp @@ -61,7 +61,7 @@ TEST_P(MatMulTest_f16_qai8dxp_qsi4cxp, EndToEnd) { const auto& ukernel_variant = variants_kai_matmul_clamp_f16_qai8dxp_qsi4cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const std::uint32_t seed = 0; diff --git a/test/tests/matmul_clamp_f16_qai8dxp_qsi8cxp_test.cpp b/test/tests/matmul_clamp_f16_qai8dxp_qsi8cxp_test.cpp index 3a1710f8999e315822f30489590da8ac211cd4b3..e42586d1e053ed5faa156e3512bc94be3aed8409 100644 --- a/test/tests/matmul_clamp_f16_qai8dxp_qsi8cxp_test.cpp +++ b/test/tests/matmul_clamp_f16_qai8dxp_qsi8cxp_test.cpp @@ -62,7 +62,7 @@ TEST_P(MatMulTest_f16_qai8dxp_qsi8cxp, EndToEnd) { const auto& ukernel_variant = variants_kai_matmul_clamp_f16_qai8dxp_qsi8cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const std::uint32_t seed = 0; diff --git a/test/tests/matmul_clamp_f16_qsi8d32p_qai4c32p_test.cpp b/test/tests/matmul_clamp_f16_qsi8d32p_qai4c32p_test.cpp index 1a1f3acb3eb1439f3790213a6e556c773162ad36..f9845f5034136c4daa36aeba67ffb0498053723e 100644 --- a/test/tests/matmul_clamp_f16_qsi8d32p_qai4c32p_test.cpp +++ b/test/tests/matmul_clamp_f16_qsi8d32p_qai4c32p_test.cpp @@ -58,7 +58,7 @@ TEST_P(MatMulTest_f16_qsi8d32p_qai4c32p, EndToEnd) { const auto& ukernel_variant = variants_kai_matmul_clamp_f16_qsi8d32p_qai4c32p.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const std::uint32_t seed = 0; diff --git a/test/tests/matmul_clamp_f32_bf16p_bf16p_test.cpp b/test/tests/matmul_clamp_f32_bf16p_bf16p_test.cpp index 76ea30b7913c71830c5f25c2167aee01d960ea2c..fc4460ba0c76154ae58442f0927e61860fa19cb0 100644 --- a/test/tests/matmul_clamp_f32_bf16p_bf16p_test.cpp +++ b/test/tests/matmul_clamp_f32_bf16p_bf16p_test.cpp @@ -436,7 +436,7 @@ TEST_P(MatMulTestBf16, Output) { const auto& [method, info, portion] = GetParam(); if (method.fn_is_supported && !method.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } if (!method.has_main_kernel()) { diff --git a/test/tests/matmul_clamp_f32_f32_f32p_test.cpp b/test/tests/matmul_clamp_f32_f32_f32p_test.cpp index 3b37518a89b54c5267340ede7964c7a606e6ab9d..3fdef0406a848eb403e6bbecbad45f32fb914187 100644 --- a/test/tests/matmul_clamp_f32_f32_f32p_test.cpp +++ b/test/tests/matmul_clamp_f32_f32_f32p_test.cpp @@ -75,7 +75,7 @@ TEST_P(MatMulTest_f32_f32_f32p, EndToEnd) // NOLINT(google-readability-avoid-un const auto& ukernel_variant = ukernel_variants.at(variant_idx); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } constexpr uint32_t seed = 0; diff --git a/test/tests/matmul_clamp_f32_qai8dxp_qsi4c32p_test.cpp b/test/tests/matmul_clamp_f32_qai8dxp_qsi4c32p_test.cpp index 5cd255d5bdf91da650d834c19f355acca341a939..be8e2b717e057081b7bfcd1d92f0defdb2db66a8 100644 --- a/test/tests/matmul_clamp_f32_qai8dxp_qsi4c32p_test.cpp +++ b/test/tests/matmul_clamp_f32_qai8dxp_qsi4c32p_test.cpp @@ -88,7 +88,7 @@ TEST_P(MatMulTest_f32_qmatmul_clamp_f32_qai8dxp_qsi4c32p, Offset_RHS) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4c32p.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const size_t M = matmul_shape.m; @@ -125,7 +125,7 @@ TEST_P(MatMulTest_f32_qmatmul_clamp_f32_qai8dxp_qsi4c32p, Offset_LHS) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4c32p.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const size_t M = matmul_shape.m; @@ -156,7 +156,7 @@ TEST_P(MatMulTest_f32_qmatmul_clamp_f32_qai8dxp_qsi4c32p, EndToEnd_RHS_nxk) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4c32p.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } constexpr uint32_t seed = 0; @@ -348,7 +348,7 @@ TEST_P(MatMulTest_f32_qmatmul_clamp_f32_qai8dxp_qsi4c32p, EndToEnd_RHS_kxn) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4c32p.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const uint32_t seed = 0; diff --git a/test/tests/matmul_clamp_f32_qai8dxp_qsi4cxp_test.cpp b/test/tests/matmul_clamp_f32_qai8dxp_qsi4cxp_test.cpp index 37c37c39c601c8196fc7c8433b7cb5970dd65bc0..d492943cc9feabc025077679a11dcc66ae8403d6 100644 --- a/test/tests/matmul_clamp_f32_qai8dxp_qsi4cxp_test.cpp +++ b/test/tests/matmul_clamp_f32_qai8dxp_qsi4cxp_test.cpp @@ -209,7 +209,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi4cxp, Offset_RHS) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const size_t M = matmul_shape.m; @@ -239,7 +239,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi4cxp, Offset_LHS) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const size_t M = matmul_shape.m; @@ -270,7 +270,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi4cxp, EndToEnd_RHS_nxk_qsi4cx) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } if (ukernel_variant.rhs_pack_type == RhsPackType::KxN) { GTEST_SKIP() << "Wrong type. This test for NxK"; @@ -401,7 +401,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi4cxp, EndToEnd_RHS_nxk_qsu4cx) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } if (ukernel_variant.rhs_pack_type == RhsPackType::KxN) { GTEST_SKIP() << "Wrong type. This test for NxK"; @@ -532,7 +532,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi4cxp, EndToEnd_RHS_kxn_qsi4cx) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } if (ukernel_variant.rhs_pack_type == RhsPackType::NxK) { GTEST_SKIP() << "Wrong type. This test for KxN"; @@ -667,7 +667,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi4cxp, EndToEnd_RHS_kxn_qsu4cx) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi4cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } if (ukernel_variant.rhs_pack_type == RhsPackType::NxK) { GTEST_SKIP() << "Wrong type. This test for KxN"; diff --git a/test/tests/matmul_clamp_f32_qai8dxp_qsi8cxp_test.cpp b/test/tests/matmul_clamp_f32_qai8dxp_qsi8cxp_test.cpp index d8db748a5b4ab38e4e582dd990950eebd157d2ed..315b5b0c6afda632185ba4a3e4e835b77d162758 100644 --- a/test/tests/matmul_clamp_f32_qai8dxp_qsi8cxp_test.cpp +++ b/test/tests/matmul_clamp_f32_qai8dxp_qsi8cxp_test.cpp @@ -61,7 +61,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi8cxp, Offset_RHS) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi8cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const size_t K = matmul_shape.k; @@ -85,7 +85,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi8cxp, Offset_LHS) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi8cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const size_t K = matmul_shape.k; @@ -106,7 +106,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi8cxp, EndToEnd_RHS_nxk_qsi8cx) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi8cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const uint32_t seed = 0; @@ -220,7 +220,7 @@ TEST_P(MatMulTest_f32_qai8dxp_qsi8cxp, EndToEnd_RHS_kxn_qsi8cx) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qai8dxp_qsi8cxp.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const uint32_t seed = 0; diff --git a/test/tests/matmul_clamp_f32_qsi8d32p_qai4c32p_test.cpp b/test/tests/matmul_clamp_f32_qsi8d32p_qai4c32p_test.cpp index fc703432069f9bf32e73b7c5367a0b4db84cd9a3..dfb5b2fb00688bad6dd9b24ab68a939908e43e71 100644 --- a/test/tests/matmul_clamp_f32_qsi8d32p_qai4c32p_test.cpp +++ b/test/tests/matmul_clamp_f32_qsi8d32p_qai4c32p_test.cpp @@ -129,7 +129,7 @@ TEST_P(MatMulTest_f32_qsi8d32p_qai4c32p, EndToEnd) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qsi8d32p_qai4c32p.at(variant_index); if (ukernel_variant.fn_is_supported && !ukernel_variant.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const std::uint32_t seed = 0; diff --git a/test/tests/matmul_clamp_f32_qsi8d32p_qsi4c32p_test.cpp b/test/tests/matmul_clamp_f32_qsi8d32p_qsi4c32p_test.cpp index 24326307066e32190fad46eeccc6207245228e69..ac69e11926b465eaed659be032d7a9ce2d862b1d 100644 --- a/test/tests/matmul_clamp_f32_qsi8d32p_qsi4c32p_test.cpp +++ b/test/tests/matmul_clamp_f32_qsi8d32p_qsi4c32p_test.cpp @@ -100,7 +100,7 @@ TEST_P(MatMulTest_f32_qsi8d32p_qsi4c32p, Offset_RHS) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qsi8d32p_qsi4c32p.at(variant_index); if (ukernel_variant.ukernel.fn_is_supported && !ukernel_variant.ukernel.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const size_t bl = 32; @@ -130,7 +130,7 @@ TEST_P(MatMulTest_f32_qsi8d32p_qsi4c32p, Offset_LHS) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qsi8d32p_qsi4c32p.at(variant_index); if (ukernel_variant.ukernel.fn_is_supported && !ukernel_variant.ukernel.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const size_t bl = 32; @@ -162,7 +162,7 @@ TEST_P(MatMulTest_f32_qsi8d32p_qsi4c32p, EndToEnd) { const auto& ukernel_variant = variants_kai_matmul_clamp_f32_qsi8d32p_qsi4c32p.at(variant_index); if (ukernel_variant.ukernel.fn_is_supported && !ukernel_variant.ukernel.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } const std::uint32_t seed = 0; diff --git a/test/tests/matmul_clamp_qai8_qai8p_qsi8cxp_test.cpp b/test/tests/matmul_clamp_qai8_qai8p_qsi8cxp_test.cpp index 050448042ca11aea65e91f53810c9ec3326bf569..b787fc386d5dbbdfd082876b6778c40b100fd572 100644 --- a/test/tests/matmul_clamp_qai8_qai8p_qsi8cxp_test.cpp +++ b/test/tests/matmul_clamp_qai8_qai8p_qsi8cxp_test.cpp @@ -278,7 +278,7 @@ const std::array& get_indirect_gemm_variants() { static const kai_imatmul_clamp_qai8_qai8p_qsi8cxp_ukernel& ukernel = get_imatmul_clamp_qai8_qai8_qsi8cxp2vlx4sb_1x16vl_sme2_dot_interface(); - variants[0].name = "indirect_matmul_qai8_qai8p_qsi8cxp"; + variants[0].name = "imatmul_qai8_qai8p_qsi8cxp"; variants[0].acc_pack.m = 2 * get_sme_vector_length(); variants[0].acc_pack.n = 2 * get_sme_vector_length(); variants[0].acc_pack.k = sizeof(int32_t) / sizeof(int8_t); @@ -733,8 +733,7 @@ static std::string test_description( const MatrixPortion& portion, size_t k_chunk_len, float clamp_ratio) { std::ostringstream sstream; - sstream << "Method_" << variant.name << "__M_" << shape.m << "__N_" << shape.n << "__k_chunk_count_" << shape.k - << "__"; + sstream << variant.name << "__M_" << shape.m << "__N_" << shape.n << "__k_chunk_count_" << shape.k << "__"; PrintTo(portion, &sstream); sstream << "__k_chunk_len_" << k_chunk_len << "__clamp_ratio_" << static_cast(clamp_ratio * 100); @@ -745,7 +744,7 @@ TEST_P(MatMulQuantizedTest, EndToEnd) { const auto& [variant, shape, output_portion, clamp_ratio] = GetParam(); if (!variant.is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } TestDataId test_data_id{shape, variant.acc_pack, shape.k, false, clamp_ratio}; @@ -879,7 +878,7 @@ TEST_P(IndirectMatMulQuantizedTest, EndToEnd) { MatMulShape shape{shape_k_chunk.m, shape_k_chunk.n, k_chunk.count * k_chunk.length}; if (!variant.is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } // Toggle padding testst when LHS has more than one row diff --git a/test/tests/matmul_test.cpp b/test/tests/matmul_test.cpp index dd893ec3a67bc359fd4dfc9796105de3d1595800..01404603a9b9b60fb335a056abd3419250e9fe31 100644 --- a/test/tests/matmul_test.cpp +++ b/test/tests/matmul_test.cpp @@ -473,7 +473,7 @@ TEST_P(MatMulTest, PackedLhs) { const auto& [method, info, portion] = GetParam(); if (method.fn_is_supported && !method.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } if (!method.is_pack_lhs_needed()) { @@ -525,7 +525,7 @@ TEST_P(MatMulTest, PackedRhs) { const auto& [method, info, portion] = GetParam(); if (method.fn_is_supported && !method.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } if (!method.is_pack_rhs_needed()) { @@ -597,7 +597,7 @@ TEST_P(MatMulTest, PackedTransposedRhs) { const auto& [method, info, portion] = GetParam(); if (method.fn_is_supported && !method.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } if (!method.is_pack_rhs_nxk_needed()) { @@ -658,7 +658,7 @@ TEST_P(MatMulTest, Output) { const auto& [method, info, portion] = GetParam(); if (method.fn_is_supported && !method.fn_is_supported()) { - GTEST_SKIP() << "CPU features are not supported by current CPU"; + GTEST_SKIP() << "Unsupported CPU feature"; } if (!method.has_main_kernel()) {