Lines Matching refs:unspec
52 /* Return the UNSPEC_CMLA* unspec for rotation amount ROT. */
66 /* Return the UNSPEC_FCMLA* unspec for rotation amount ROT. */
80 /* Return the UNSPEC_COND_FCMLA* unspec for rotation amount ROT. */
114 /* Expand a call to svmla_lane or svmls_lane using floating-point unspec
115 UNSPEC. */
117 expand_mla_mls_lane (function_expander &e, int unspec)
123 insn_code icode = code_for_aarch64_lane (unspec, e.vector_mode (0));
179 CONSTEXPR svac_impl (int unspec) : m_unspec (unspec) {}
189 /* The unspec code for the underlying comparison. */
261 CONSTEXPR svbrk_binary_impl (int unspec) : m_unspec (unspec) {}
269 /* The unspec code associated with the operation. */
277 CONSTEXPR svbrk_unary_impl (int unspec) : m_unspec (unspec) {}
285 /* The unspec code associated with the operation. */
295 /* Convert the rotation amount into a specific unspec. */
311 CONSTEXPR svclast_impl (int unspec) : m_unspec (unspec) {}
327 /* The unspec code associated with the operation. */
337 /* Convert the rotation amount into a specific unspec. */
360 /* Convert the rotation amount into a specific unspec. */
431 /* The unspec code to use for floating-point comparisons. */
466 int unspec = (unsigned_p ? m_unspec_for_uint : m_unspec_for_sint);
467 return e.use_exact_insn (code_for_aarch64_pred_cmp_wide (unspec, mode));
473 /* The unspec codes for signed and unsigned wide comparisons
679 int unspec = (e.type_suffix (1).unsigned_p
684 ? code_for_aarch64_sve_nonextend (unspec, mode1, mode0)
685 : code_for_cond_nonextend (unspec, mode1, mode0));
688 ? code_for_aarch64_sve_extend (unspec, mode1, mode0)
689 : code_for_cond_extend (unspec, mode1, mode0));
693 int unspec = (!e.type_suffix (0).integer_p ? UNSPEC_COND_FCVT
698 ? code_for_aarch64_sve_nontrunc (unspec, mode1, mode0)
699 : code_for_cond_nontrunc (unspec, mode1, mode0));
702 ? code_for_aarch64_sve_trunc (unspec, mode1, mode0)
703 : code_for_cond_trunc (unspec, mode1, mode0));
744 int unspec = unspec_for (e);
746 return e.use_exact_insn (code_for_aarch64_dot_prod_lane (unspec, mode));
1075 CONSTEXPR svlast_impl (int unspec) : m_unspec (unspec) {}
1083 /* The unspec
1367 CONSTEXPR svldxf1_impl (int unspec) : m_unspec (unspec) {}
1386 /* The unspec associated with the load. */
1394 CONSTEXPR svldxf1_extend_impl (type_suffix_index memory_type, int unspec)
1395 : extending_load (memory_type), m_unspec (unspec) {}
1416 /* The unspec associated with the load. */
1632 CONSTEXPR svpfirst_svpnext_impl (int unspec) : m_unspec (unspec) {}
1642 /* The unspec associated with the operation. */
1802 (const:V16BI (unspec:V16BI [(const_int PATTERN)
2346 int unspec = e.type_suffix (0).unsigned_p ? unpacku : unpacks;
2347 icode = code_for_aarch64_sve_unpk (unspec, unspec, mode);