Read_once(), Write_once(), but Not for Rust

(lwn.net)

59 points | by todsacerdoti 3 hours ago

4 comments

  • bheadmaster 1 hour ago

        An understanding of READ_ONCE() and WRITE_ONCE() is important for kernel developers who will be dealing with any sort of concurrent access to data. So, naturally, they are almost entirely absent from the kernel's documentation.
    
    Made me chuckle.
    • semiquaver 1 hour ago
      More chuckles from the source:

         /*
          * Yes, this permits 64-bit accesses on 32-bit architectures. These will
          * actually be atomic in some cases (namely Armv7 + LPAE), but for others we
          * rely on the access being split into 2x32-bit accesses for a 32-bit quantity
          * (e.g. a virtual address) and a strong prevailing wind.
          */
  • gpderetta 2 hours ago
    Very interesting. AFAIK the kernel explicitly gives consume semantics to read_once (and in fact it is not just a compiler barrier on alpha), so technically lowering it to a relaxed operation is wrong.

    Does rust have or need the equivalent of std::memory_order_consume? Famously this was deemed unimplementable in C++.

    • steveklabnik 2 hours ago
      It wasn’t implemented for the same reason. Rust uses C++20 ordering.
      • gpderetta 1 hour ago
        right, so I would expect that the equivalent of READ_ONCE is converted to an acquire in rust, even if slightly pessimal.

        But the article says that the suggestion is to convert them to relaxed loads. Is the expectation to YOLO it and hope that the compiler doesn't break control and data dependencies?

        • bonzini 1 hour ago
          There is a yolo way that actually works, which would be to change it to a relaxed load followed by an acquire signal fence.
          • loeg 1 hour ago
            Is that any better than just using an acquire load?
            • gpderetta 1 hour ago
              It is cheaper on ARM and POWER. But I'm not sure it is always safe. The standard has very complex rules for consume to make sure that the compiler didn't break the dependencies.

              edit: and those rules where so complex that compilers decided where not implementable or not worth it.

    • loeg 1 hour ago
      Does anything care about Alpha? The platform hasn't been sold in 20 years.
      • jcranmer 1 hour ago
        It's a persistent misunderstanding that release-consume is about Alpha. It's not; in fact, Alpha is one of the few architectures where release-consume doesn't help.

        In a TSO architecture like x86 or SPARC, every "regular" memory load/store is effectively a release/acquire by default. Using release/consume or relaxed provides no extra speedup on these architectures. In weak memory models, you need to add in acquire barriers to get release/acquire architectures. But also, most weak memory models have a basic rule that a data-dependent load has an implicit ordering dependency on the values that computed it (most notably, loading *p has an implicit dependency on p).

        The goal of release/consume is to be able to avoid having an acquire fence if you have only those dependencies--to promote a hardware data dependency semantic rule to a language-level semantic rule. For Alpha's ultra-weak model, you still need the acquire fence in this mode, it doesn't help Alpha one whit. Unfortunately, for various reasons, no one has been able to work out a language-level semantics for consume that compilers are willing to implement (preserving data dependencies through optimizations is a lot more difficult than it appears), so all compilers have remapped consume to acquire, making it useless.

      • gpderetta 1 hour ago
        consume is trivial on alpha, it is the same as acquire (always needs a #LoadLoad). It is also the same as acquire (and relaxed) on x86 and SPARC (a plain load, #LoadLoad is always implied).

        The only place where consume matters is on relaxed but not too relaxed architectures like ARM and POWER, where consume relies on the implicit #LoadLoad of controls and data dependencies.

        • bonzini 1 hour ago
          Also on alpha there's only store-store and full memory barriers. Acquire is very expensive.
  • epolanski 1 hour ago
    What is your take on their names instead of "atomic_read" and "atomic_write"?
    • gpm 1 hour ago
      The problem with atomic_read and atomic_write is that some people will interpret that as "atomic with a sequentially consistent ordering" and some as "atomic with a relaxed ordering" and everything in between. It's a fine name for a function that takes an argument that specifies memory ordering [1]. It's not great for anything else.

      Read_once and Write_once convey that there's more nuance than that, and tries to convey the nuance.

      [1] E.g. in rust anything that takes https://doc.rust-lang.org/std/sync/atomic/enum.Ordering.html

    • kccqzy 30 minutes ago
      I think “atomic” implies something more than just “once” because for atomic we customarily consider the memory order with that memory access, but “once” just implies reading and writing exactly once. Neither are good names because the kernel developers clearly assumed some kind of atomicity with some kind of memory ordering here but just calling it “atomic” doesn’t convey that.
  • chrismsimpson 2 hours ago
    > The truth of the matter, though, is that the Rust community seems to want to take a different approach to concurrent data access.

    Not knowing anything about development of the kernel, does this kind of thing create a two tier Linux development experience?

    • zaphar 2 hours ago
      Not sure if it introduces a tiered experience or not. But reading the article it appears that the Rust devs advocated for an api that is clearer in it's semantics with the tradeoff that now understanding how it interacts with C code requires understanding two APIs. How this shakes out in practice remains to be seen.