From: Tamar Christina <Tamar.Christina@arm.com>
To: Tamar Christina <Tamar.Christina@arm.com>,
Richard Sandiford <Richard.Sandiford@arm.com>
Cc: "gcc-patches@gcc.gnu.org" <gcc-patches@gcc.gnu.org>,
nd <nd@arm.com>, Richard Earnshaw <Richard.Earnshaw@arm.com>,
Marcus Shawcroft <Marcus.Shawcroft@arm.com>,
Kyrylo Tkachov <Kyrylo.Tkachov@arm.com>
Subject: RE: [PATCH 5/8]AArch64 aarch64: Make existing V2HF be usable.
Date: Tue, 22 Nov 2022 16:01:18 +0000 [thread overview]
Message-ID: <VI1PR08MB5325C0DE24F64251620FFC49FF0D9@VI1PR08MB5325.eurprd08.prod.outlook.com> (raw)
In-Reply-To: <VI1PR08MB5325335D195073D1E5B2AB33FF009@VI1PR08MB5325.eurprd08.prod.outlook.com>
Ping
> -----Original Message-----
> From: Gcc-patches <gcc-patches-
> bounces+tamar.christina=arm.com@gcc.gnu.org> On Behalf Of Tamar
> Christina via Gcc-patches
> Sent: Friday, November 11, 2022 2:40 PM
> To: Richard Sandiford <Richard.Sandiford@arm.com>
> Cc: gcc-patches@gcc.gnu.org; nd <nd@arm.com>; Richard Earnshaw
> <Richard.Earnshaw@arm.com>; Marcus Shawcroft
> <Marcus.Shawcroft@arm.com>; Kyrylo Tkachov <Kyrylo.Tkachov@arm.com>
> Subject: RE: [PATCH 5/8]AArch64 aarch64: Make existing V2HF be usable.
>
> Hi,
>
>
> > This name might cause confusion with the SVE iterators, where FULL
> > means "every bit of the register is used". How about something like
> > VMOVE instead?
> >
> > With this change, I guess VALL_F16 represents "The set of all modes
> > for which the vld1 intrinsics are provided" and VMOVE or whatever is
> > "All Advanced SIMD modes suitable for moving, loading, and storing".
> > That is, VMOVE extends VALL_F16 with modes that are not manifested via
> > intrinsics.
> >
>
> Done.
>
> > Where is the 2h used, and is it valid syntax in that context?
> >
> > Same for later instances of 2h.
>
> They are, but they weren't meant to be in this patch. They belong in a
> separate FP16 series that I won't get to finish for GCC 13 due not being able
> to finish writing all the tests. I have moved them to that patch series though.
>
> While the addp patch series has been killed, this patch is still good standalone
> and improves codegen as shown in the updated testcase.
>
> Bootstrapped Regtested on aarch64-none-linux-gnu and no issues.
>
> Ok for master?
>
> Thanks,
> Tamar
>
> gcc/ChangeLog:
>
> * config/aarch64/aarch64-simd.md (*aarch64_simd_movv2hf): New.
> (mov<mode>, movmisalign<mode>, aarch64_dup_lane<mode>,
> aarch64_store_lane0<mode>, aarch64_simd_vec_set<mode>,
> @aarch64_simd_vec_copy_lane<mode>, vec_set<mode>,
> reduc_<optab>_scal_<mode>, reduc_<fmaxmin>_scal_<mode>,
> aarch64_reduc_<optab>_internal<mode>,
> aarch64_get_lane<mode>,
> vec_init<mode><Vel>, vec_extract<mode><Vel>): Support V2HF.
> (aarch64_simd_dupv2hf): New.
> * config/aarch64/aarch64.cc (aarch64_classify_vector_mode):
> Add E_V2HFmode.
> * config/aarch64/iterators.md (VHSDF_P): New.
> (V2F, VMOVE, nunits, Vtype, Vmtype, Vetype, stype, VEL,
> Vel, q, vp): Add V2HF.
> * config/arm/types.md (neon_fp_reduc_add_h): New.
>
> gcc/testsuite/ChangeLog:
>
> * gcc.target/aarch64/sve/slp_1.c: Update testcase.
>
> --- inline copy of patch ---
>
> diff --git a/gcc/config/aarch64/aarch64-simd.md
> b/gcc/config/aarch64/aarch64-simd.md
> index
> f4152160084d6b6f34bd69f0ba6386c1ab50f77e..487a31010245accec28e779661
> e6c2d578fca4b7 100644
> --- a/gcc/config/aarch64/aarch64-simd.md
> +++ b/gcc/config/aarch64/aarch64-simd.md
> @@ -19,10 +19,10 @@
> ;; <http://www.gnu.org/licenses/>.
>
> (define_expand "mov<mode>"
> - [(set (match_operand:VALL_F16 0 "nonimmediate_operand")
> - (match_operand:VALL_F16 1 "general_operand"))]
> + [(set (match_operand:VMOVE 0 "nonimmediate_operand")
> + (match_operand:VMOVE 1 "general_operand"))]
> "TARGET_SIMD"
> - "
> +{
> /* Force the operand into a register if it is not an
> immediate whose use can be replaced with xzr.
> If the mode is 16 bytes wide, then we will be doing @@ -46,12 +46,11 @@
> (define_expand "mov<mode>"
> aarch64_expand_vector_init (operands[0], operands[1]);
> DONE;
> }
> - "
> -)
> +})
>
> (define_expand "movmisalign<mode>"
> - [(set (match_operand:VALL_F16 0 "nonimmediate_operand")
> - (match_operand:VALL_F16 1 "general_operand"))]
> + [(set (match_operand:VMOVE 0 "nonimmediate_operand")
> + (match_operand:VMOVE 1 "general_operand"))]
> "TARGET_SIMD && !STRICT_ALIGNMENT"
> {
> /* This pattern is not permitted to fail during expansion: if both arguments
> @@ -73,6 +72,16 @@ (define_insn "aarch64_simd_dup<mode>"
> [(set_attr "type" "neon_dup<q>, neon_from_gp<q>")]
> )
>
> +(define_insn "aarch64_simd_dupv2hf"
> + [(set (match_operand:V2HF 0 "register_operand" "=w")
> + (vec_duplicate:V2HF
> + (match_operand:HF 1 "register_operand" "0")))]
> + "TARGET_SIMD"
> + "@
> + sli\\t%d0, %d1, 16"
> + [(set_attr "type" "neon_shift_imm")]
> +)
> +
> (define_insn "aarch64_simd_dup<mode>"
> [(set (match_operand:VDQF_F16 0 "register_operand" "=w,w")
> (vec_duplicate:VDQF_F16
> @@ -85,10 +94,10 @@ (define_insn "aarch64_simd_dup<mode>"
> )
>
> (define_insn "aarch64_dup_lane<mode>"
> - [(set (match_operand:VALL_F16 0 "register_operand" "=w")
> - (vec_duplicate:VALL_F16
> + [(set (match_operand:VMOVE 0 "register_operand" "=w")
> + (vec_duplicate:VMOVE
> (vec_select:<VEL>
> - (match_operand:VALL_F16 1 "register_operand" "w")
> + (match_operand:VMOVE 1 "register_operand" "w")
> (parallel [(match_operand:SI 2 "immediate_operand" "i")])
> )))]
> "TARGET_SIMD"
> @@ -142,6 +151,29 @@ (define_insn
> "*aarch64_simd_mov<VDMOV:mode>"
> mov_reg, neon_move<q>")]
> )
>
> +(define_insn "*aarch64_simd_movv2hf"
> + [(set (match_operand:V2HF 0 "nonimmediate_operand"
> + "=w, m, m, w, ?r, ?w, ?r, w, w")
> + (match_operand:V2HF 1 "general_operand"
> + "m, Dz, w, w, w, r, r, Dz, Dn"))]
> + "TARGET_SIMD_F16INST
> + && (register_operand (operands[0], V2HFmode)
> + || aarch64_simd_reg_or_zero (operands[1], V2HFmode))"
> + "@
> + ldr\\t%s0, %1
> + str\\twzr, %0
> + str\\t%s1, %0
> + mov\\t%0.2s[0], %1.2s[0]
> + umov\\t%w0, %1.s[0]
> + fmov\\t%s0, %1
> + mov\\t%0, %1
> + movi\\t%d0, 0
> + * return aarch64_output_simd_mov_immediate (operands[1], 32);"
> + [(set_attr "type" "neon_load1_1reg, store_8, neon_store1_1reg,\
> + neon_logic, neon_to_gp, f_mcr,\
> + mov_reg, neon_move, neon_move")]
> +)
> +
> (define_insn "*aarch64_simd_mov<VQMOV:mode>"
> [(set (match_operand:VQMOV 0 "nonimmediate_operand"
> "=w, Umn, m, w, ?r, ?w, ?r, w")
> @@ -182,7 +214,7 @@ (define_insn "*aarch64_simd_mov<VQMOV:mode>"
>
> (define_insn "aarch64_store_lane0<mode>"
> [(set (match_operand:<VEL> 0 "memory_operand" "=m")
> - (vec_select:<VEL> (match_operand:VALL_F16 1 "register_operand"
> "w")
> + (vec_select:<VEL> (match_operand:VMOVE 1 "register_operand"
> "w")
> (parallel [(match_operand 2 "const_int_operand"
> "n")])))]
> "TARGET_SIMD
> && ENDIAN_LANE_N (<nunits>, INTVAL (operands[2])) == 0"
> @@ -1035,11 +1067,11 @@ (define_insn "one_cmpl<mode>2"
> )
>
> (define_insn "aarch64_simd_vec_set<mode>"
> - [(set (match_operand:VALL_F16 0 "register_operand" "=w,w,w")
> - (vec_merge:VALL_F16
> - (vec_duplicate:VALL_F16
> + [(set (match_operand:VMOVE 0 "register_operand" "=w,w,w")
> + (vec_merge:VMOVE
> + (vec_duplicate:VMOVE
> (match_operand:<VEL> 1
> "aarch64_simd_nonimmediate_operand" "w,?r,Utv"))
> - (match_operand:VALL_F16 3 "register_operand" "0,0,0")
> + (match_operand:VMOVE 3 "register_operand" "0,0,0")
> (match_operand:SI 2 "immediate_operand" "i,i,i")))]
> "TARGET_SIMD"
> {
> @@ -1061,14 +1093,14 @@ (define_insn "aarch64_simd_vec_set<mode>"
> )
>
> (define_insn "@aarch64_simd_vec_copy_lane<mode>"
> - [(set (match_operand:VALL_F16 0 "register_operand" "=w")
> - (vec_merge:VALL_F16
> - (vec_duplicate:VALL_F16
> + [(set (match_operand:VMOVE 0 "register_operand" "=w")
> + (vec_merge:VMOVE
> + (vec_duplicate:VMOVE
> (vec_select:<VEL>
> - (match_operand:VALL_F16 3 "register_operand" "w")
> + (match_operand:VMOVE 3 "register_operand" "w")
> (parallel
> [(match_operand:SI 4 "immediate_operand" "i")])))
> - (match_operand:VALL_F16 1 "register_operand" "0")
> + (match_operand:VMOVE 1 "register_operand" "0")
> (match_operand:SI 2 "immediate_operand" "i")))]
> "TARGET_SIMD"
> {
> @@ -1376,7 +1408,7 @@ (define_insn "vec_shr_<mode>"
> )
>
> (define_expand "vec_set<mode>"
> - [(match_operand:VALL_F16 0 "register_operand")
> + [(match_operand:VMOVE 0 "register_operand")
> (match_operand:<VEL> 1 "aarch64_simd_nonimmediate_operand")
> (match_operand:SI 2 "immediate_operand")]
> "TARGET_SIMD"
> @@ -3495,7 +3527,7 @@ (define_insn "popcount<mode>2"
> ;; gimple_fold'd to the IFN_REDUC_(MAX|MIN) function. (This is FP
> smax/smin).
> (define_expand "reduc_<optab>_scal_<mode>"
> [(match_operand:<VEL> 0 "register_operand")
> - (unspec:<VEL> [(match_operand:VHSDF 1 "register_operand")]
> + (unspec:<VEL> [(match_operand:VHSDF_P 1 "register_operand")]
> FMAXMINV)]
> "TARGET_SIMD"
> {
> @@ -3510,7 +3542,7 @@ (define_expand "reduc_<optab>_scal_<mode>"
>
> (define_expand "reduc_<fmaxmin>_scal_<mode>"
> [(match_operand:<VEL> 0 "register_operand")
> - (unspec:<VEL> [(match_operand:VHSDF 1 "register_operand")]
> + (unspec:<VEL> [(match_operand:VHSDF_P 1 "register_operand")]
> FMAXMINNMV)]
> "TARGET_SIMD"
> {
> @@ -3554,8 +3586,8 @@ (define_insn
> "aarch64_reduc_<optab>_internalv2si"
> )
>
> (define_insn "aarch64_reduc_<optab>_internal<mode>"
> - [(set (match_operand:VHSDF 0 "register_operand" "=w")
> - (unspec:VHSDF [(match_operand:VHSDF 1 "register_operand" "w")]
> + [(set (match_operand:VHSDF_P 0 "register_operand" "=w")
> + (unspec:VHSDF_P [(match_operand:VHSDF_P 1 "register_operand"
> + "w")]
> FMAXMINV))]
> "TARGET_SIMD"
> "<maxmin_uns_op><vp>\\t%<Vetype>0, %1.<Vtype>"
> @@ -4200,7 +4232,7 @@ (define_insn
> "*aarch64_get_lane_zero_extend<GPI:mode><VDQQH:mode>"
> (define_insn_and_split "aarch64_get_lane<mode>"
> [(set (match_operand:<VEL> 0 "aarch64_simd_nonimmediate_operand"
> "=?r, w, Utv")
> (vec_select:<VEL>
> - (match_operand:VALL_F16 1 "register_operand" "w, w, w")
> + (match_operand:VMOVE 1 "register_operand" "w, w, w")
> (parallel [(match_operand:SI 2 "immediate_operand" "i, i, i")])))]
> "TARGET_SIMD"
> {
> @@ -7981,7 +8013,7 @@ (define_expand "aarch64_st1<VALL_F16:mode>"
> ;; Standard pattern name vec_init<mode><Vel>.
>
> (define_expand "vec_init<mode><Vel>"
> - [(match_operand:VALL_F16 0 "register_operand")
> + [(match_operand:VMOVE 0 "register_operand")
> (match_operand 1 "" "")]
> "TARGET_SIMD"
> {
> @@ -8060,7 +8092,7 @@ (define_insn "aarch64_urecpe<mode>"
>
> (define_expand "vec_extract<mode><Vel>"
> [(match_operand:<VEL> 0 "aarch64_simd_nonimmediate_operand")
> - (match_operand:VALL_F16 1 "register_operand")
> + (match_operand:VMOVE 1 "register_operand")
> (match_operand:SI 2 "immediate_operand")]
> "TARGET_SIMD"
> {
> diff --git a/gcc/config/aarch64/aarch64.cc b/gcc/config/aarch64/aarch64.cc
> index
> 84dbe2f4ea7d03b424602ed98a34e7824217dc91..35671cb86e374f9ded21d0e4
> 944c63bc2cbc0901 100644
> --- a/gcc/config/aarch64/aarch64.cc
> +++ b/gcc/config/aarch64/aarch64.cc
> @@ -3566,6 +3566,7 @@ aarch64_classify_vector_mode (machine_mode
> mode)
> case E_V8BFmode:
> case E_V4SFmode:
> case E_V2DFmode:
> + case E_V2HFmode:
> return TARGET_SIMD ? VEC_ADVSIMD : 0;
>
> default:
> diff --git a/gcc/config/aarch64/iterators.md
> b/gcc/config/aarch64/iterators.md index
> 37d8161a33b1c399d80be82afa67613a087389d4..dfcf86a440e316c2abdbcc6463
> 63d39e458d1a91 100644
> --- a/gcc/config/aarch64/iterators.md
> +++ b/gcc/config/aarch64/iterators.md
> @@ -160,6 +160,10 @@ (define_mode_iterator VDQF [V2SF V4SF V2DF])
> (define_mode_iterator VHSDF [(V4HF "TARGET_SIMD_F16INST")
> (V8HF "TARGET_SIMD_F16INST")
> V2SF V4SF V2DF])
> +;; Advanced SIMD Float modes suitable for pairwise operations.
> +(define_mode_iterator VHSDF_P [(V4HF "TARGET_SIMD_F16INST")
> + (V8HF "TARGET_SIMD_F16INST")
> + V2SF V4SF V2DF (V2HF
> "TARGET_SIMD_F16INST")])
>
> ;; Advanced SIMD Float modes, and DF.
> (define_mode_iterator VDQF_DF [V2SF V4SF V2DF DF]) @@ -188,15 +192,23
> @@ (define_mode_iterator VDQF_COND [V2SF V2SI V4SF V4SI V2DF V2DI])
> (define_mode_iterator VALLF [V2SF V4SF V2DF SF DF])
>
> ;; Advanced SIMD Float modes with 2 elements.
> -(define_mode_iterator V2F [V2SF V2DF])
> +(define_mode_iterator V2F [V2SF V2DF V2HF])
>
> ;; All Advanced SIMD modes on which we support any arithmetic operations.
> (define_mode_iterator VALL [V8QI V16QI V4HI V8HI V2SI V4SI V2DI V2SF
> V4SF V2DF])
>
> -;; All Advanced SIMD modes suitable for moving, loading, and storing.
> +;; All Advanced SIMD modes suitable for moving, loading, and storing ;;
> +except V2HF.
> (define_mode_iterator VALL_F16 [V8QI V16QI V4HI V8HI V2SI V4SI V2DI
> V4HF V8HF V4BF V8BF V2SF V4SF V2DF])
>
> +;; All Advanced SIMD modes suitable for moving, loading, and storing ;;
> +including V2HF (define_mode_iterator VMOVE [V8QI V16QI V4HI V8HI V2SI
> +V4SI V2DI
> + V4HF V8HF V4BF V8BF V2SF V4SF V2DF
> + (V2HF "TARGET_SIMD_F16INST")])
> +
> +
> ;; The VALL_F16 modes except the 128-bit 2-element ones.
> (define_mode_iterator VALL_F16_NO_V2Q [V8QI V16QI V4HI V8HI V2SI
> V4SI
> V4HF V8HF V2SF V4SF])
> @@ -1076,7 +1088,7 @@ (define_mode_attr nunits [(V8QI "8") (V16QI "16")
> (V2SF "2") (V4SF "4")
> (V1DF "1") (V2DF "2")
> (DI "1") (DF "1")
> - (V8DI "8")])
> + (V8DI "8") (V2HF "2")])
>
> ;; Map a mode to the number of bits in it, if the size of the mode ;; is
> constant.
> @@ -1090,6 +1102,7 @@ (define_mode_attr s [(HF "h") (SF "s") (DF "d") (SI
> "s") (DI "d")])
>
> ;; Give the length suffix letter for a sign- or zero-extension.
> (define_mode_attr size [(QI "b") (HI "h") (SI "w")])
> +(define_mode_attr sizel [(QI "b") (HI "h") (SI "")])
>
> ;; Give the number of bits in the mode
> (define_mode_attr sizen [(QI "8") (HI "16") (SI "32") (DI "64")]) @@ -1193,7
> +1206,7 @@ (define_mode_attr Vmntype [(V8HI ".8b") (V4SI ".4h")
> (define_mode_attr Vetype [(V8QI "b") (V16QI "b")
> (V4HI "h") (V8HI "h")
> (V2SI "s") (V4SI "s")
> - (V2DI "d")
> + (V2DI "d") (V2HF "h")
> (V4HF "h") (V8HF "h")
> (V2SF "s") (V4SF "s")
> (V2DF "d")
> @@ -1285,7 +1298,7 @@ (define_mode_attr Vcwtype [(VNx16QI "b")
> (VNx8QI "h") (VNx4QI "w") (VNx2QI "d") ;; more accurately.
> (define_mode_attr stype [(V8QI "b") (V16QI "b") (V4HI "s") (V8HI "s")
> (V2SI "s") (V4SI "s") (V2DI "d") (V4HF "s")
> - (V8HF "s") (V2SF "s") (V4SF "s") (V2DF "d")
> + (V8HF "s") (V2SF "s") (V4SF "s") (V2DF "d") (V2HF
> "s")
> (HF "s") (SF "s") (DF "d") (QI "b") (HI "s")
> (SI "s") (DI "d")])
>
> @@ -1360,8 +1373,8 @@ (define_mode_attr VEL [(V8QI "QI") (V16QI "QI")
> (V4HF "HF") (V8HF "HF")
> (V2SF "SF") (V4SF "SF")
> (DF "DF") (V2DF "DF")
> - (SI "SI") (HI "HI")
> - (QI "QI")
> + (SI "SI") (V2HF "HF")
> + (QI "QI") (HI "HI")
> (V4BF "BF") (V8BF "BF")
> (VNx16QI "QI") (VNx8QI "QI") (VNx4QI "QI") (VNx2QI
> "QI")
> (VNx8HI "HI") (VNx4HI "HI") (VNx2HI "HI") @@ -1381,7
> +1394,7 @@ (define_mode_attr Vel [(V8QI "qi") (V16QI "qi")
> (V2SF "sf") (V4SF "sf")
> (V2DF "df") (DF "df")
> (SI "si") (HI "hi")
> - (QI "qi")
> + (QI "qi") (V2HF "hf")
> (V4BF "bf") (V8BF "bf")
> (VNx16QI "qi") (VNx8QI "qi") (VNx4QI "qi") (VNx2QI "qi")
> (VNx8HI "hi") (VNx4HI "hi") (VNx2HI "hi") @@ -1866,7
> +1879,7 @@ (define_mode_attr q [(V8QI "") (V16QI "_q")
> (V4HF "") (V8HF "_q")
> (V4BF "") (V8BF "_q")
> (V2SF "") (V4SF "_q")
> - (V2DF "_q")
> + (V2HF "") (V2DF "_q")
> (QI "") (HI "") (SI "") (DI "") (HF "") (SF "") (DF "")
> (V2x8QI "") (V2x16QI "_q")
> (V2x4HI "") (V2x8HI "_q")
> @@ -1905,6 +1918,7 @@ (define_mode_attr vp [(V8QI "v") (V16QI "v")
> (V2SI "p") (V4SI "v")
> (V2DI "p") (V2DF "p")
> (V2SF "p") (V4SF "v")
> + (V2HF "p")
> (V4HF "v") (V8HF "v")])
>
> (define_mode_attr vsi2qi [(V2SI "v8qi") (V4SI "v16qi") diff --git
> a/gcc/config/arm/types.md b/gcc/config/arm/types.md index
> 7d0504bdd944e9c0d1b545b0b66a9a1adc808714..3cfbc7a93cca1bea4925853e5
> 1d0a147c5722247 100644
> --- a/gcc/config/arm/types.md
> +++ b/gcc/config/arm/types.md
> @@ -483,6 +483,7 @@ (define_attr "autodetect_type"
> ; neon_fp_minmax_s_q
> ; neon_fp_minmax_d
> ; neon_fp_minmax_d_q
> +; neon_fp_reduc_add_h
> ; neon_fp_reduc_add_s
> ; neon_fp_reduc_add_s_q
> ; neon_fp_reduc_add_d
> @@ -1033,6 +1034,7 @@ (define_attr "type"
> neon_fp_minmax_d,\
> neon_fp_minmax_d_q,\
> \
> + neon_fp_reduc_add_h,\
> neon_fp_reduc_add_s,\
> neon_fp_reduc_add_s_q,\
> neon_fp_reduc_add_d,\
> @@ -1257,8 +1259,8 @@ (define_attr "is_neon_type" "yes,no"
> neon_fp_compare_d, neon_fp_compare_d_q, neon_fp_minmax_s,\
> neon_fp_minmax_s_q, neon_fp_minmax_d, neon_fp_minmax_d_q,\
> neon_fp_neg_s, neon_fp_neg_s_q, neon_fp_neg_d,
> neon_fp_neg_d_q,\
> - neon_fp_reduc_add_s, neon_fp_reduc_add_s_q,
> neon_fp_reduc_add_d,\
> - neon_fp_reduc_add_d_q, neon_fp_reduc_minmax_s,
> + neon_fp_reduc_add_h, neon_fp_reduc_add_s,
> neon_fp_reduc_add_s_q,\
> + neon_fp_reduc_add_d, neon_fp_reduc_add_d_q,
> + neon_fp_reduc_minmax_s,\
> neon_fp_reduc_minmax_s_q, neon_fp_reduc_minmax_d,\
> neon_fp_reduc_minmax_d_q,\
> neon_fp_cvt_narrow_s_q, neon_fp_cvt_narrow_d_q,\ diff --git
> a/gcc/testsuite/gcc.target/aarch64/sve/slp_1.c
> b/gcc/testsuite/gcc.target/aarch64/sve/slp_1.c
> index
> 07d71a63414b1066ea431e287286ad048515711a..e6021c5a42748701e5326a5c3
> 87a39a0bbadc9e5 100644
> --- a/gcc/testsuite/gcc.target/aarch64/sve/slp_1.c
> +++ b/gcc/testsuite/gcc.target/aarch64/sve/slp_1.c
> @@ -30,11 +30,9 @@ vec_slp_##TYPE (TYPE *restrict a, TYPE b, TYPE c, int n)
> \
> TEST_ALL (VEC_PERM)
>
> /* We should use one DUP for each of the 8-, 16- and 32-bit types,
> - although we currently use LD1RW for _Float16. We should use two
> - DUPs for each of the three 64-bit types. */
> + We should use two DUPs for each of the three 64-bit types. */
> /* { dg-final { scan-assembler-times {\tmov\tz[0-9]+\.h, [hw]} 2 } } */
> -/* { dg-final { scan-assembler-times {\tmov\tz[0-9]+\.s, [sw]} 2 } } */
> -/* { dg-final { scan-assembler-times {\tld1rw\tz[0-9]+\.s, } 1 } } */
> +/* { dg-final { scan-assembler-times {\tmov\tz[0-9]+\.s, [sw]} 3 } } */
> /* { dg-final { scan-assembler-times {\tmov\tz[0-9]+\.d, [dx]} 9 } } */
> /* { dg-final { scan-assembler-times {\tzip1\tz[0-9]+\.d, z[0-9]+\.d, z[0-
> 9]+\.d\n} 3 } } */
> /* { dg-final { scan-assembler-not {\tzip2\t} } } */ @@ -53,7 +51,7 @@
> TEST_ALL (VEC_PERM)
> /* { dg-final { scan-assembler-times {\twhilelo\tp[0-7]\.s} 6 } } */
> /* { dg-final { scan-assembler-times {\twhilelo\tp[0-7]\.d} 6 } } */
> /* { dg-final { scan-assembler-not {\tldr} } } */
> -/* { dg-final { scan-assembler-times {\tstr} 2 } } */
> -/* { dg-final { scan-assembler-times {\tstr\th[0-9]+} 2 } } */
> +/* { dg-final { scan-assembler-not {\tstr} } } */
> +/* { dg-final { scan-assembler-not {\tstr\th[0-9]+} } } */
>
> /* { dg-final { scan-assembler-not {\tuqdec} } } */
next prev parent reply other threads:[~2022-11-22 16:01 UTC|newest]
Thread overview: 50+ messages / expand[flat|nested] mbox.gz Atom feed top
2022-10-31 11:56 [PATCH 1/8]middle-end: Recognize scalar reductions from bitfields and array_refs Tamar Christina
2022-10-31 11:57 ` [PATCH 2/8]middle-end: Recognize scalar widening reductions Tamar Christina
2022-10-31 21:42 ` Jeff Law
2022-11-07 13:21 ` Richard Biener
2022-10-31 11:57 ` [PATCH 3/8]middle-end: Support extractions of subvectors from arbitrary element position inside a vector Tamar Christina
2022-10-31 21:44 ` Jeff Law
2022-11-01 14:25 ` Richard Sandiford
2022-11-11 14:33 ` Tamar Christina
2022-11-15 8:35 ` Hongtao Liu
2022-11-15 8:51 ` Tamar Christina
2022-11-15 9:37 ` Hongtao Liu
2022-11-15 10:00 ` Tamar Christina
2022-11-15 17:39 ` Richard Sandiford
2022-11-17 8:04 ` Hongtao Liu
2022-11-17 9:39 ` Richard Sandiford
2022-11-17 10:20 ` Hongtao Liu
2022-11-17 13:59 ` Richard Sandiford
2022-11-18 2:31 ` Hongtao Liu
2022-11-18 9:16 ` Richard Sandiford
2022-10-31 11:58 ` [PATCH 4/8]AArch64 aarch64: Implement widening reduction patterns Tamar Christina
2022-11-01 14:41 ` Richard Sandiford
2022-10-31 11:58 ` [PATCH 5/8]AArch64 aarch64: Make existing V2HF be usable Tamar Christina
2022-11-01 14:58 ` Richard Sandiford
2022-11-01 15:11 ` Tamar Christina
2022-11-11 14:39 ` Tamar Christina
2022-11-22 16:01 ` Tamar Christina [this message]
2022-11-30 4:26 ` Tamar Christina
2022-12-06 10:28 ` Richard Sandiford
2022-12-06 10:58 ` Tamar Christina
2022-12-06 11:05 ` Richard Sandiford
2022-10-31 11:59 ` [PATCH 6/8]AArch64: Add peephole and scheduling logic for pairwise operations that appear late in RTL Tamar Christina
2022-10-31 11:59 ` [PATCH 7/8]AArch64: Consolidate zero and sign extension patterns and add missing ones Tamar Christina
2022-11-30 4:28 ` Tamar Christina
2022-12-06 15:59 ` Richard Sandiford
2022-10-31 12:00 ` [PATCH 8/8]AArch64: Have reload not choose to do add on the scalar side if both values exist on the SIMD side Tamar Christina
2022-11-01 15:04 ` Richard Sandiford
2022-11-01 15:20 ` Tamar Christina
2022-10-31 21:41 ` [PATCH 1/8]middle-end: Recognize scalar reductions from bitfields and array_refs Jeff Law
2022-11-05 11:32 ` Richard Biener
2022-11-07 7:16 ` Tamar Christina
2022-11-07 10:17 ` Richard Biener
2022-11-07 11:00 ` Tamar Christina
2022-11-07 11:22 ` Richard Biener
2022-11-07 11:56 ` Tamar Christina
2022-11-22 10:36 ` Richard Sandiford
2022-11-22 10:58 ` Richard Biener
2022-11-22 11:02 ` Tamar Christina
2022-11-22 11:06 ` Richard Sandiford
2022-11-22 11:08 ` Richard Biener
2022-11-22 14:33 ` Jeff Law
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=VI1PR08MB5325C0DE24F64251620FFC49FF0D9@VI1PR08MB5325.eurprd08.prod.outlook.com \
--to=tamar.christina@arm.com \
--cc=Kyrylo.Tkachov@arm.com \
--cc=Marcus.Shawcroft@arm.com \
--cc=Richard.Earnshaw@arm.com \
--cc=Richard.Sandiford@arm.com \
--cc=gcc-patches@gcc.gnu.org \
--cc=nd@arm.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for read-only IMAP folder(s) and NNTP newsgroup(s).