public inbox for gcc-patches@gcc.gnu.org
 help / color / mirror / Atom feed
From: Palmer Dabbelt <palmer@dabbelt.com>
To: Vineet Gupta <vineetg@rivosinc.com>, jeffreyalaw@gmail.com
Cc: cmuellner@gcc.gnu.org, Andrea Parri <andrea@rivosinc.com>,
	gcc-patches@gcc.gnu.org, kito.cheng@sifive.com,
	gnu-toolchain@rivosinc.com
Subject: Re: [PATCH v2 00/10] [RISC-V] Atomics improvements [PR100265/PR100266]
Date: Tue, 11 Oct 2022 17:15:43 -0700 (PDT)	[thread overview]
Message-ID: <mhng-d009d127-c4f4-45d6-80da-e085ae3e6a12@palmer-ri-x1c9> (raw)
In-Reply-To: <3862a109-8083-7a36-3d85-8f9e5e10627c@rivosinc.com>

On Tue, 11 Oct 2022 16:31:25 PDT (-0700), Vineet Gupta wrote:
>
>
> On 10/11/22 13:46, Christoph Müllner wrote:
>> On Tue, Oct 11, 2022 at 9:31 PM Palmer Dabbelt <palmer@dabbelt.com> wrote:
>>
>>     On Tue, 11 Oct 2022 12:06:27 PDT (-0700), Vineet Gupta wrote:
>>     > Hi Christoph, Kito,
>>     >
>>     > On 5/5/21 12:36, Christoph Muellner via Gcc-patches wrote:
>>     >> This series provides a cleanup of the current atomics
>>     implementation
>>     >> of RISC-V:
>>     >>
>>     >> * PR100265: Use proper fences for atomic load/store
>>     >> * PR100266: Provide programmatic implementation of CAS
>>     >>
>>     >> As both are very related, I merged the patches into one series.
>>     >>
>>     >> The first patch could be squashed into the following patches,
>>     >> but I found it easier to understand the chances with it in place.
>>     >>
>>     >> The series has been tested as follows:
>>     >> * Building and testing a multilib RV32/64 toolchain
>>     >>    (bootstrapped with riscv-gnu-toolchain repo)
>>     >> * Manual review of generated sequences for GCC's atomic
>>     builtins API
>>     >>
>>     >> The programmatic re-implementation of CAS benefits from a REE
>>     improvement
>>     >> (see PR100264):
>>     >> https://gcc.gnu.org/pipermail/gcc-patches/2021-April/568680.html
>>     >> If this patch is not in place, then an additional extension
>>     instruction
>>     >> is emitted after the SC.W (in case of RV64 and CAS for uint32_t).
>>     >>
>>     >> Further, the new CAS code requires cbranch INSN helpers to be
>>     present:
>>     >> https://gcc.gnu.org/pipermail/gcc-patches/2021-May/569689.html
>>     >
>>     > I was wondering is this patchset is blocked on some technical
>>     grounds.
>>
>>     There's a v3 (though I can't find all of it, so not quite sure what
>>     happened), but IIUC that still has the same fundamental problems that
>>     all these have had: changing over to the new fence model may by an
>>     ABI
>>     break and the split CAS implementation doesn't ensure eventual
>>     success
>>     (see Jim's comments).  Not sure if there's other comments floating
>>     around, though, that's just what I remember.
>>
>>
>> v3 was sent on May 27, 2022, when I rebased this on an internal tree:
>> https://gcc.gnu.org/pipermail/gcc-patches/2022-May/595712.html
>> I dropped the CAS patch in v3 (issue: stack spilling under extreme 
>> register pressure instead of erroring out) as I thought that this was 
>> the blocker for the series.
>> I just learned a few weeks ago, when I asked Palmer at the GNU 
>> Cauldron about this series, that the ABI break is the blocker.
>
> Yeah I was confused about the ABI aspect as I didn't see any mention of 
> that in the public reviews of v1 and v2.

Sorry, I thought we'd talked about it somewhere but it must have just 
been in meetings and such.  Patrick was writing a similar patch set 
around the same time so it probably just got tied up in that, we ended 
up reducing it to just the strong CAS inline stuff because we couldn't 
sort out the correctness of the rest of it.

>> My initial understanding was that fixing something broken cannot be an 
>> ABI break.
>> And that the mismatch of the implementation in 2021 and the 
>> recommended mappings in the ratified specification from 2019 is 
>> something that is broken. I still don't know the background here, but 
>> I guess this assumption is incorrect from a historical point of view.

We agreed that we wouldn't break binaries back when we submitted the 
port.  The ISA has changed many times since then, including adding the 
recommended mappings, but those binaries exist and we can't just 
silently break things for users.

>> However, I'm sure that I am not the only one that assumes the mappings 
>> in the specification to be implemented in compilers and tools. 
>> Therefore I still consider the implementation of the RISC-V atomics in 
>> GCC as broken (at least w.r.t. user expectation from people that lack 
>> the historical background and just read the RISC-V specification).

You can't just read one of those RISC-V PDFs and assume that 
implementations that match those words will function correctly.  Those 
words regularly change in ways where reasonable readers would end up 
with incompatible implementations due to those differences.  That's why 
we're so explicit about versions and such these days, we're just getting 
burned by these old mappings because they're from back when we though 
the RISC-V definition of compatibility was going to match the more 
common one and we didn't build in fallbacks.

>>     +Andrea, in case he has time to look at the memory model / ABI
>>     issues.
>>
>>     We'd still need to sort out the CAS issues, though, and it's not
>>     abundantly clear it's worth the work: we're essentailly
>>     constrained to
>>     just emitting those fixed CAS sequences due to the eventual success
>>     rules, so it's not clear what the benefit of splitting those up is.
>>     With WRS there are some routines we might want to generate code for
>>     (cond_read_acquire() in Linux, for example) but we'd really need
>>     to dig
>>     into those to see if it's even sane/fast.
>>
>>     There's another patch set to fix the lack of inline atomic routines
>>     without breaking stuff, there were some minor comments from Kito and
>>     IIRC I had some test failures that I needed to chase down as well.
>>     That's a much safer fix in the short term, we'll need to deal with
>>     this
>>     eventually but at least we can stop the libatomic issues for the
>>     distro
>>     folks.
>>
>>
>> I expect that the pressure for a proper fix upstream (instead of a 
>> backward compatible compromise) will increase over time (once people 
>> start building big iron based on RISC-V and start hunting performance 
>> bottlenecks in multithreaded workloads to be competitive).
>> What could be done to get some relief is to enable the new atomics ABI 
>> by a command line switch and promote its use. And at one point in the 
>> future (if there are enough fixes to justify a break) the new ABI can 
>> be enabled by default with a new flag to enable the old ABI.
>
> Indeed we are stuck with inefficiencies with status quo. The new abi 
> option sounds like a reasonable plan going fwd.

I don't think we're just stuck with the status quo, we really just need 
to go through the mappings and figure out which can be made both fast 
and ABI-compatible.  Then we can fix those and see where we stand, maybe 
it's good enough or maybe we need to introduce some sort of 
compatibility break to make things faster (and/or compatible with LLVM, 
where I suspect we're broken right now).

If we do need a break then I think it's probably possible to do it in 
phases, where we have a middle-ground compatibility mode that works for 
both the old and new mappings so distros can gradually move over as they 
rebuild packages.

Issues like the libstdc++ shared_ptr/mutex fallback don't map well to 
that, though.  There's also some stuff like the IO fence bits that we 
can probably just add an argument for now, those were likely just a bad 
idea at the time and should be safe to turn off for the vast majority of 
users (though those are more of an API break).

> Also my understand is that while the considerations are ABI centric, the 
> option to faciliate this need not be tied to canonical -mabi=lp32, lp64d 
> etc. It might just be a toggle as -matomic=legacy,2019 etc (this is not 
> suggestive just indicative). Otherwise there's another level of blowup 
> in multilib testing etc.

The psABI doesn't mention memory ordering at all.  IIUC that's a pretty 
standard hole in psABI documents, but it means we're in a grey area 
here.

+Jeff, who was offering to help when the threads got crossed.  I'd 
punted on a lot of this in the hope Andrea could help out, as I'm not 
really a memory model guy and this is pretty far down the rabbit hole.  
Happy to have the help if you're offering, though, as what's there is 
likely a pretty big performance issue for anyone with a reasonable 
memory system.

> -Vineet

  reply	other threads:[~2022-10-12  0:15 UTC|newest]

Thread overview: 31+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2021-05-05 19:36 Christoph Muellner
2021-05-05 19:36 ` [PATCH v2 01/10] RISC-V: Simplify memory model code [PR 100265] Christoph Muellner
2021-05-05 19:36 ` [PATCH v2 02/10] RISC-V: Emit proper memory ordering suffixes for AMOs " Christoph Muellner
2021-05-05 19:36 ` [PATCH v2 03/10] RISC-V: Eliminate %F specifier from riscv_print_operand() " Christoph Muellner
2021-05-05 19:36 ` [PATCH v2 04/10] RISC-V: Use STORE instead of AMOSWAP for atomic stores " Christoph Muellner
2021-05-05 19:36 ` [PATCH v2 05/10] RISC-V: Emit fences according to chosen memory model " Christoph Muellner
2021-05-05 19:36 ` [PATCH v2 06/10] RISC-V: Implement atomic_{load,store} " Christoph Muellner
2021-05-05 19:36 ` [PATCH v2 07/10] RISC-V: Model INSNs for LR and SC [PR 100266] Christoph Muellner
2021-05-05 19:36 ` [PATCH v2 08/10] RISC-V: Add s.ext-consuming " Christoph Muellner
2021-05-05 19:36 ` [PATCH v2 09/10] RISC-V: Provide programmatic implementation of CAS " Christoph Muellner
2021-05-06  0:27   ` Jim Wilson
2021-05-05 19:36 ` [PATCH v2 10/10] RISC-V: Introduce predicate "riscv_sync_memory_operand" " Christoph Muellner
2022-10-11 19:06 ` [PATCH v2 00/10] [RISC-V] Atomics improvements [PR100265/PR100266] Vineet Gupta
2022-10-11 19:31   ` Palmer Dabbelt
2022-10-11 20:46     ` Christoph Müllner
2022-10-11 23:31       ` Vineet Gupta
2022-10-12  0:15         ` Palmer Dabbelt [this message]
2022-10-12  8:03           ` Christoph Müllner
2022-10-13 23:11             ` Jeff Law
2022-10-12 17:16           ` Andrea Parri
2022-10-20 19:01             ` Andrea Parri
2022-10-29  5:02               ` Jeff Law
2022-10-13 23:04           ` Jeff Law
2022-10-13 22:39         ` Jeff Law
2022-10-13 23:14           ` Palmer Dabbelt
2022-10-14 11:03             ` Christoph Müllner
2022-10-14 20:39               ` Jeff Law
2022-10-14 21:57                 ` Palmer Dabbelt
2022-10-15  0:31                   ` Palmer Dabbelt
2022-10-14  0:14           ` Vineet Gupta
2022-10-11 23:14     ` Jeff Law

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=mhng-d009d127-c4f4-45d6-80da-e085ae3e6a12@palmer-ri-x1c9 \
    --to=palmer@dabbelt.com \
    --cc=andrea@rivosinc.com \
    --cc=cmuellner@gcc.gnu.org \
    --cc=gcc-patches@gcc.gnu.org \
    --cc=gnu-toolchain@rivosinc.com \
    --cc=jeffreyalaw@gmail.com \
    --cc=kito.cheng@sifive.com \
    --cc=vineetg@rivosinc.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for read-only IMAP folder(s) and NNTP newsgroup(s).