From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from foss.arm.com (foss.arm.com [217.140.110.172]) by sourceware.org (Postfix) with ESMTP id AD3A43858D1E for ; Fri, 30 Sep 2022 11:51:39 +0000 (GMT) DMARC-Filter: OpenDMARC Filter v1.4.1 sourceware.org AD3A43858D1E Authentication-Results: sourceware.org; dmarc=none (p=none dis=none) header.from=foss.arm.com Authentication-Results: sourceware.org; spf=none smtp.mailfrom=foss.arm.com Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.121.207.14]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id 04E9312FC; Fri, 30 Sep 2022 04:51:46 -0700 (PDT) Received: from [10.2.78.76] (e126323.cambridge.arm.com [10.2.78.76]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id BD3043F73B; Fri, 30 Sep 2022 04:51:38 -0700 (PDT) Message-ID: <2f6776db-d787-cb66-1518-1c70b9d04b8c@foss.arm.com> Date: Fri, 30 Sep 2022 12:51:37 +0100 MIME-Version: 1.0 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:91.0) Gecko/20100101 Thunderbird/91.11.0 Subject: Re: [PATCH v2] testsuite: [arm] Relax expected register names in MVE tests Content-Language: en-GB To: Christophe Lyon , gcc-patches@gcc.gnu.org References: <20220930111938.354664-1-christophe.lyon@arm.com> From: Richard Earnshaw In-Reply-To: <20220930111938.354664-1-christophe.lyon@arm.com> Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 7bit X-Spam-Status: No, score=-3496.4 required=5.0 tests=BAYES_00,GIT_PATCH_0,KAM_DMARC_STATUS,KAM_LAZY_DOMAIN_SECURITY,KAM_SHORT,NICE_REPLY_A,SPF_HELO_NONE,SPF_NONE,TXREP autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on server2.sourceware.org List-Id: On 30/09/2022 12:19, Christophe Lyon via Gcc-patches wrote: > These two tests have hardcoded q0 as destination/source of load/store > instructions, but this register is actually used only under > -mfloat-abi=hard. When using -mfloat-abi=softfp, other registers > (eg. q3) can be used to transfer function arguments from core > registers to MVE registers, making the expected regexp fail. > > This small patch replaces q0 with q[0-7] to accept any 'q' register. > In several places where we had q[0-9]+, replace it with q[0-7] as MVE > only has q0-q7 registers. > > OK for trunk? > > gcc/testsuite/ChangeLog: > > * gcc.target/arm/mve/mve_load_memory_modes.c: Update expected > registers. > * gcc.target/arm/mve/mve_store_memory_modes.c: Likewise. OK. R. > --- > .../arm/mve/mve_load_memory_modes.c | 58 +++++++++---------- > .../arm/mve/mve_store_memory_modes.c | 58 +++++++++---------- > 2 files changed, 58 insertions(+), 58 deletions(-) > > diff --git a/gcc/testsuite/gcc.target/arm/mve/mve_load_memory_modes.c b/gcc/testsuite/gcc.target/arm/mve/mve_load_memory_modes.c > index e35eb1108aa..816980d1203 100644 > --- a/gcc/testsuite/gcc.target/arm/mve/mve_load_memory_modes.c > +++ b/gcc/testsuite/gcc.target/arm/mve/mve_load_memory_modes.c > @@ -7,7 +7,7 @@ > /* > **off_load8_0: > ** ... > -** vldrb.8 q0, \[r0, #16\] > +** vldrb.8 q[0-7], \[r0, #16\] > ** ... > */ > int8x16_t off_load8_0 (int8_t * a) > @@ -18,7 +18,7 @@ int8x16_t off_load8_0 (int8_t * a) > /* > **off_load8_1: > ** ... > -** vldrb.u16 q0, \[r0, #1\] > +** vldrb.u16 q[0-7], \[r0, #1\] > ** ... > */ > uint16x8_t off_load8_1 (uint8_t * a) > @@ -29,7 +29,7 @@ uint16x8_t off_load8_1 (uint8_t * a) > /* > **off_load8_2: > ** ... > -** vldrb.s32 q0, \[r0, #127\] > +** vldrb.s32 q[0-7], \[r0, #127\] > ** ... > */ > int32x4_t off_load8_2 (int8_t * a) > @@ -40,7 +40,7 @@ int32x4_t off_load8_2 (int8_t * a) > /* > **off_load8_3: > ** ... > -** vldrb.8 q0, \[r0, #-127\] > +** vldrb.8 q[0-7], \[r0, #-127\] > ** ... > */ > uint8x16_t off_load8_3 (uint8_t * a) > @@ -51,7 +51,7 @@ uint8x16_t off_load8_3 (uint8_t * a) > /* > **not_off_load8_0: > ** ... > -** vldrb.8 q0, \[r[0-9]+\] > +** vldrb.8 q[0-7], \[r[0-7]+\] > ** ... > */ > int8x16_t not_off_load8_0 (int8_t * a) > @@ -62,7 +62,7 @@ int8x16_t not_off_load8_0 (int8_t * a) > /* > **off_loadfp16_0: > ** ... > -** vldrh.16 q0, \[r0, #-244\] > +** vldrh.16 q[0-7], \[r0, #-244\] > ** ... > */ > float16x8_t off_loadfp16_0 (float16_t *a) > @@ -73,7 +73,7 @@ float16x8_t off_loadfp16_0 (float16_t *a) > /* > **off_load16_0: > ** ... > -** vldrh.16 q0, \[r0, #-2\] > +** vldrh.16 q[0-7], \[r0, #-2\] > ** ... > */ > uint16x8_t off_load16_0 (uint16_t * a) > @@ -84,7 +84,7 @@ uint16x8_t off_load16_0 (uint16_t * a) > /* > **off_load16_1: > ** ... > -** vldrh.u32 q0, \[r0, #254\] > +** vldrh.u32 q[0-7], \[r0, #254\] > ** ... > */ > uint32x4_t off_load16_1 (uint16_t * a) > @@ -95,7 +95,7 @@ uint32x4_t off_load16_1 (uint16_t * a) > /* > **not_off_load16_0: > ** ... > -** vldrh.16 q0, \[r[0-9]+\] > +** vldrh.16 q[0-7], \[r[0-7]+\] > ** ... > */ > int16x8_t not_off_load16_0 (int8_t * a) > @@ -106,7 +106,7 @@ int16x8_t not_off_load16_0 (int8_t * a) > /* > **not_off_load16_1: > ** ... > -** vldrh.u32 q0, \[r[0-9]+\] > +** vldrh.u32 q[0-7], \[r[0-7]+\] > ** ... > */ > uint32x4_t not_off_load16_1 (uint16_t * a) > @@ -117,7 +117,7 @@ uint32x4_t not_off_load16_1 (uint16_t * a) > /* > **off_loadfp32_0: > ** ... > -** vldrw.32 q0, \[r0, #24\] > +** vldrw.32 q[0-7], \[r0, #24\] > ** ... > */ > float32x4_t off_loadfp32_0 (float32_t *a) > @@ -128,7 +128,7 @@ float32x4_t off_loadfp32_0 (float32_t *a) > /* > **off_load32_0: > ** ... > -** vldrw.32 q0, \[r0, #4\] > +** vldrw.32 q[0-7], \[r0, #4\] > ** ... > */ > uint32x4_t off_load32_0 (uint32_t * a) > @@ -139,7 +139,7 @@ uint32x4_t off_load32_0 (uint32_t * a) > /* > **off_load32_1: > ** ... > -** vldrw.32 q0, \[r0, #-508\] > +** vldrw.32 q[0-7], \[r0, #-508\] > ** ... > */ > int32x4_t off_load32_1 (int32_t * a) > @@ -149,7 +149,7 @@ int32x4_t off_load32_1 (int32_t * a) > /* > **pre_load8_0: > ** ... > -** vldrb.8 q[0-9]+, \[r0, #16\]! > +** vldrb.8 q[0-7], \[r0, #16\]! > ** ... > */ > int8_t* pre_load8_0 (int8_t * a, int8x16_t *v) > @@ -162,7 +162,7 @@ int8_t* pre_load8_0 (int8_t * a, int8x16_t *v) > /* > **pre_load8_1: > ** ... > -** vldrb.u16 q[0-9]+, \[r0, #4\]! > +** vldrb.u16 q[0-7], \[r0, #4\]! > ** ... > */ > uint8_t* pre_load8_1 (uint8_t * a, uint16x8_t *v) > @@ -175,7 +175,7 @@ uint8_t* pre_load8_1 (uint8_t * a, uint16x8_t *v) > /* > **pre_loadfp16_0: > ** ... > -** vldrh.16 q[0-9]+, \[r0, #128\]! > +** vldrh.16 q[0-7], \[r0, #128\]! > ** ... > */ > float16_t* pre_loadfp16_0 (float16_t *a, float16x8_t *v) > @@ -188,7 +188,7 @@ float16_t* pre_loadfp16_0 (float16_t *a, float16x8_t *v) > /* > **pre_load16_0: > ** ... > -** vldrh.16 q[0-9]+, \[r0, #-254\]! > +** vldrh.16 q[0-7], \[r0, #-254\]! > ** ... > */ > int16_t* pre_load16_0 (int16_t * a, int16x8_t *v) > @@ -201,7 +201,7 @@ int16_t* pre_load16_0 (int16_t * a, int16x8_t *v) > /* > **pre_load16_1: > ** ... > -** vldrh.s32 q[0-9]+, \[r0, #52\]! > +** vldrh.s32 q[0-7], \[r0, #52\]! > ** ... > */ > int16_t* pre_load16_1 (int16_t * a, int32x4_t *v) > @@ -214,7 +214,7 @@ int16_t* pre_load16_1 (int16_t * a, int32x4_t *v) > /* > **pre_loadfp32_0: > ** ... > -** vldrw.32 q[0-9]+, \[r0, #-72\]! > +** vldrw.32 q[0-7], \[r0, #-72\]! > ** ... > */ > float32_t* pre_loadfp32_0 (float32_t *a, float32x4_t *v) > @@ -228,7 +228,7 @@ float32_t* pre_loadfp32_0 (float32_t *a, float32x4_t *v) > /* > **pre_load32_0: > ** ... > -** vldrw.32 q[0-9]+, \[r0, #-4\]! > +** vldrw.32 q[0-7], \[r0, #-4\]! > ** ... > */ > uint32_t* pre_load32_0 (uint32_t * a, uint32x4_t *v) > @@ -242,7 +242,7 @@ uint32_t* pre_load32_0 (uint32_t * a, uint32x4_t *v) > /* > **post_load8_0: > ** ... > -** vldrb.8 q[0-9]+, \[r0\], #26 > +** vldrb.8 q[0-7], \[r0\], #26 > ** ... > */ > uint8_t* post_load8_0 (uint8_t * a, uint8x16_t *v) > @@ -255,7 +255,7 @@ uint8_t* post_load8_0 (uint8_t * a, uint8x16_t *v) > /* > **post_load8_1: > ** ... > -** vldrb.s16 q[0-9]+, \[r0\], #-1 > +** vldrb.s16 q[0-7], \[r0\], #-1 > ** ... > */ > int8_t* post_load8_1 (int8_t * a, int16x8_t *v) > @@ -268,7 +268,7 @@ int8_t* post_load8_1 (int8_t * a, int16x8_t *v) > /* > **post_load8_2: > ** ... > -** vldrb.8 q[0-9]+, \[r0\], #26 > +** vldrb.8 q[0-7], \[r0\], #26 > ** ... > */ > uint8_t* post_load8_2 (uint8_t * a, uint8x16_t *v) > @@ -281,7 +281,7 @@ uint8_t* post_load8_2 (uint8_t * a, uint8x16_t *v) > /* > **post_load8_3: > ** ... > -** vldrb.s16 q[0-9]+, \[r0\], #-1 > +** vldrb.s16 q[0-7], \[r0\], #-1 > ** ... > */ > int8_t* post_load8_3 (int8_t * a, int16x8_t *v) > @@ -294,7 +294,7 @@ int8_t* post_load8_3 (int8_t * a, int16x8_t *v) > /* > **post_loadfp16_0: > ** ... > -** vldrh.16 q[0-9]+, \[r0\], #-24 > +** vldrh.16 q[0-7], \[r0\], #-24 > ** ... > */ > float16_t* post_loadfp16_0 (float16_t *a, float16x8_t *v) > @@ -307,7 +307,7 @@ float16_t* post_loadfp16_0 (float16_t *a, float16x8_t *v) > /* > **post_load16_0: > ** ... > -** vldrh.16 q[0-9]+, \[r0\], #-126 > +** vldrh.16 q[0-7], \[r0\], #-126 > ** ... > */ > uint16_t* post_load16_0 (uint16_t * a, uint16x8_t *v) > @@ -320,7 +320,7 @@ uint16_t* post_load16_0 (uint16_t * a, uint16x8_t *v) > /* > **post_load16_1: > ** ... > -** vldrh.u32 q[0-9]+, \[r0\], #16 > +** vldrh.u32 q[0-7], \[r0\], #16 > ** ... > */ > uint16_t* post_load16_1 (uint16_t * a, uint32x4_t *v) > @@ -333,7 +333,7 @@ uint16_t* post_load16_1 (uint16_t * a, uint32x4_t *v) > /* > **post_loadfp32_0: > ** ... > -** vldrw.32 q[0-9]+, \[r0\], #4 > +** vldrw.32 q[0-7], \[r0\], #4 > ** ... > */ > float32_t* post_loadfp32_0 (float32_t *a, float32x4_t *v) > @@ -346,7 +346,7 @@ float32_t* post_loadfp32_0 (float32_t *a, float32x4_t *v) > /* > **post_load32_0: > ** ... > -** vldrw.32 q[0-9]+, \[r0\], #-16 > +** vldrw.32 q[0-7], \[r0\], #-16 > ** ... > */ > int32_t* post_load32_0 (int32_t * a, int32x4_t *v) > diff --git a/gcc/testsuite/gcc.target/arm/mve/mve_store_memory_modes.c b/gcc/testsuite/gcc.target/arm/mve/mve_store_memory_modes.c > index 632f5b44f0b..11df1144064 100644 > --- a/gcc/testsuite/gcc.target/arm/mve/mve_store_memory_modes.c > +++ b/gcc/testsuite/gcc.target/arm/mve/mve_store_memory_modes.c > @@ -7,7 +7,7 @@ > /* > **off_store8_0: > ** ... > -** vstrb.8 q0, \[r0, #16\] > +** vstrb.8 q[0-7], \[r0, #16\] > ** ... > */ > uint8_t *off_store8_0 (uint8_t * a, uint8x16_t v) > @@ -19,7 +19,7 @@ uint8_t *off_store8_0 (uint8_t * a, uint8x16_t v) > /* > **off_store8_1: > ** ... > -** vstrb.16 q0, \[r0, #-1\] > +** vstrb.16 q[0-7], \[r0, #-1\] > ** ... > */ > int8_t *off_store8_1 (int8_t * a, int16x8_t v) > @@ -31,7 +31,7 @@ int8_t *off_store8_1 (int8_t * a, int16x8_t v) > /* > **off_store8_2: > ** ... > -** vstrb.32 q0, \[r0, #-127\] > +** vstrb.32 q[0-7], \[r0, #-127\] > ** ... > */ > uint8_t *off_store8_2 (uint8_t * a, uint32x4_t v) > @@ -43,7 +43,7 @@ uint8_t *off_store8_2 (uint8_t * a, uint32x4_t v) > /* > **off_store8_3: > ** ... > -** vstrb.8 q0, \[r0, #127\] > +** vstrb.8 q[0-7], \[r0, #127\] > ** ... > */ > int8_t *off_store8_3 (int8_t * a, int8x16_t v) > @@ -55,7 +55,7 @@ int8_t *off_store8_3 (int8_t * a, int8x16_t v) > /* > **not_off_store8_0: > ** ... > -** vstrb.8 q0, \[r[0-9]+\] > +** vstrb.8 q[0-7], \[r[0-7]+\] > ** ... > */ > uint8_t *not_off_store8_0 (uint8_t * a, uint8x16_t v) > @@ -67,7 +67,7 @@ uint8_t *not_off_store8_0 (uint8_t * a, uint8x16_t v) > /* > **off_storefp16_0: > ** ... > -** vstrh.16 q0, \[r0, #250\] > +** vstrh.16 q[0-7], \[r0, #250\] > ** ... > */ > float16_t *off_storefp16_0 (float16_t *a, float16x8_t v) > @@ -79,7 +79,7 @@ float16_t *off_storefp16_0 (float16_t *a, float16x8_t v) > /* > **off_store16_0: > ** ... > -** vstrh.16 q0, \[r0, #4\] > +** vstrh.16 q[0-7], \[r0, #4\] > ** ... > */ > int16_t *off_store16_0 (int16_t * a, int16x8_t v) > @@ -91,7 +91,7 @@ int16_t *off_store16_0 (int16_t * a, int16x8_t v) > /* > **off_store16_1: > ** ... > -** vstrh.32 q0, \[r0, #-254\] > +** vstrh.32 q[0-7], \[r0, #-254\] > ** ... > */ > int16_t *off_store16_1 (int16_t * a, int32x4_t v) > @@ -103,7 +103,7 @@ int16_t *off_store16_1 (int16_t * a, int32x4_t v) > /* > **not_off_store16_0: > ** ... > -** vstrh.16 q0, \[r[0-9]+\] > +** vstrh.16 q[0-7], \[r[0-7]+\] > ** ... > */ > uint8_t *not_off_store16_0 (uint8_t * a, uint16x8_t v) > @@ -115,7 +115,7 @@ uint8_t *not_off_store16_0 (uint8_t * a, uint16x8_t v) > /* > **not_off_store16_1: > ** ... > -** vstrh.32 q0, \[r[0-9]+\] > +** vstrh.32 q[0-7], \[r[0-7]+\] > ** ... > */ > int16_t *not_off_store16_1 (int16_t * a, int32x4_t v) > @@ -127,7 +127,7 @@ int16_t *not_off_store16_1 (int16_t * a, int32x4_t v) > /* > **off_storefp32_0: > ** ... > -** vstrw.32 q0, \[r0, #-412\] > +** vstrw.32 q[0-7], \[r0, #-412\] > ** ... > */ > float32_t *off_storefp32_0 (float32_t *a, float32x4_t v) > @@ -139,7 +139,7 @@ float32_t *off_storefp32_0 (float32_t *a, float32x4_t v) > /* > **off_store32_0: > ** ... > -** vstrw.32 q0, \[r0, #-4\] > +** vstrw.32 q[0-7], \[r0, #-4\] > ** ... > */ > int32_t *off_store32_0 (int32_t * a, int32x4_t v) > @@ -151,7 +151,7 @@ int32_t *off_store32_0 (int32_t * a, int32x4_t v) > /* > **off_store32_1: > ** ... > -** vstrw.32 q0, \[r0, #508\] > +** vstrw.32 q[0-7], \[r0, #508\] > ** ... > */ > uint32_t *off_store32_1 (uint32_t * a, uint32x4_t v) > @@ -163,7 +163,7 @@ uint32_t *off_store32_1 (uint32_t * a, uint32x4_t v) > /* > **pre_store8_0: > ** ... > -** vstrb.8 q[0-9]+, \[r0, #-16\]! > +** vstrb.8 q[0-7], \[r0, #-16\]! > ** ... > */ > uint8_t* pre_store8_0 (uint8_t * a, uint8x16_t v) > @@ -176,7 +176,7 @@ uint8_t* pre_store8_0 (uint8_t * a, uint8x16_t v) > /* > **pre_store8_1: > ** ... > -** vstrb.16 q[0-9]+, \[r0, #4\]! > +** vstrb.16 q[0-7], \[r0, #4\]! > ** ... > */ > int8_t* pre_store8_1 (int8_t * a, int16x8_t v) > @@ -189,7 +189,7 @@ int8_t* pre_store8_1 (int8_t * a, int16x8_t v) > /* > **pre_storefp16_0: > ** ... > -** vstrh.16 q0, \[r0, #8\]! > +** vstrh.16 q[0-7], \[r0, #8\]! > ** ... > */ > float16_t *pre_storefp16_0 (float16_t *a, float16x8_t v) > @@ -202,7 +202,7 @@ float16_t *pre_storefp16_0 (float16_t *a, float16x8_t v) > /* > **pre_store16_0: > ** ... > -** vstrh.16 q[0-9]+, \[r0, #254\]! > +** vstrh.16 q[0-7], \[r0, #254\]! > ** ... > */ > uint16_t* pre_store16_0 (uint16_t * a, uint16x8_t v) > @@ -215,7 +215,7 @@ uint16_t* pre_store16_0 (uint16_t * a, uint16x8_t v) > /* > **pre_store16_1: > ** ... > -** vstrh.32 q[0-9]+, \[r0, #-52\]! > +** vstrh.32 q[0-7], \[r0, #-52\]! > ** ... > */ > int16_t* pre_store16_1 (int16_t * a, int32x4_t v) > @@ -228,7 +228,7 @@ int16_t* pre_store16_1 (int16_t * a, int32x4_t v) > /* > **pre_storefp32_0: > ** ... > -** vstrw.32 q0, \[r0, #-4\]! > +** vstrw.32 q[0-7], \[r0, #-4\]! > ** ... > */ > float32_t *pre_storefp32_0 (float32_t *a, float32x4_t v) > @@ -241,7 +241,7 @@ float32_t *pre_storefp32_0 (float32_t *a, float32x4_t v) > /* > **pre_store32_0: > ** ... > -** vstrw.32 q[0-9]+, \[r0, #4\]! > +** vstrw.32 q[0-7], \[r0, #4\]! > ** ... > */ > int32_t* pre_store32_0 (int32_t * a, int32x4_t v) > @@ -255,7 +255,7 @@ int32_t* pre_store32_0 (int32_t * a, int32x4_t v) > /* > **post_store8_0: > ** ... > -** vstrb.8 q[0-9]+, \[r0\], #-26 > +** vstrb.8 q[0-7], \[r0\], #-26 > ** ... > */ > int8_t* post_store8_0 (int8_t * a, int8x16_t v) > @@ -268,7 +268,7 @@ int8_t* post_store8_0 (int8_t * a, int8x16_t v) > /* > **post_store8_1: > ** ... > -** vstrb.16 q[0-9]+, \[r0\], #1 > +** vstrb.16 q[0-7], \[r0\], #1 > ** ... > */ > uint8_t* post_store8_1 (uint8_t * a, uint16x8_t v) > @@ -281,7 +281,7 @@ uint8_t* post_store8_1 (uint8_t * a, uint16x8_t v) > /* > **post_store8_2: > ** ... > -** vstrb.8 q[0-9]+, \[r0\], #-26 > +** vstrb.8 q[0-7], \[r0\], #-26 > ** ... > */ > int8_t* post_store8_2 (int8_t * a, int8x16_t v) > @@ -294,7 +294,7 @@ int8_t* post_store8_2 (int8_t * a, int8x16_t v) > /* > **post_store8_3: > ** ... > -** vstrb.16 q[0-9]+, \[r0\], #7 > +** vstrb.16 q[0-7], \[r0\], #7 > ** ... > */ > uint8_t* post_store8_3 (uint8_t * a, uint16x8_t v) > @@ -307,7 +307,7 @@ uint8_t* post_store8_3 (uint8_t * a, uint16x8_t v) > /* > **post_storefp16_0: > ** ... > -** vstrh.16 q[0-9]+, \[r0\], #-16 > +** vstrh.16 q[0-7], \[r0\], #-16 > ** ... > */ > float16_t *post_storefp16_0 (float16_t *a, float16x8_t v) > @@ -320,7 +320,7 @@ float16_t *post_storefp16_0 (float16_t *a, float16x8_t v) > /* > **post_store16_0: > ** ... > -** vstrh.16 q[0-9]+, \[r0\], #126 > +** vstrh.16 q[0-7], \[r0\], #126 > ** ... > */ > int16_t* post_store16_0 (int16_t * a, int16x8_t v) > @@ -333,7 +333,7 @@ int16_t* post_store16_0 (int16_t * a, int16x8_t v) > /* > **post_store16_1: > ** ... > -** vstrh.32 q[0-9]+, \[r0\], #-16 > +** vstrh.32 q[0-7], \[r0\], #-16 > ** ... > */ > uint16_t* post_store16_1 (uint16_t * a, uint32x4_t v) > @@ -346,7 +346,7 @@ uint16_t* post_store16_1 (uint16_t * a, uint32x4_t v) > /* > **post_storefp32_0: > ** ... > -** vstrw.32 q[0-9]+, \[r0\], #-16 > +** vstrw.32 q[0-7], \[r0\], #-16 > ** ... > */ > float32_t* post_storefp32_0 (float32_t * a, float32x4_t v) > @@ -359,7 +359,7 @@ float32_t* post_storefp32_0 (float32_t * a, float32x4_t v) > /* > **post_store32_0: > ** ... > -** vstrw.32 q[0-9]+, \[r0\], #16 > +** vstrw.32 q[0-7], \[r0\], #16 > ** ... > */ > int32_t* post_store32_0 (int32_t * a, int32x4_t v)