[PATCH v2 02/12] mm: introduce execmem_text_alloc() and jit_text_alloc()
Mark Rutland
mark.rutland at arm.com
Mon Jun 26 23:01:55 AEST 2023
On Mon, Jun 19, 2023 at 10:09:02AM -0700, Andy Lutomirski wrote:
> On Sun, Jun 18, 2023, at 1:00 AM, Mike Rapoport wrote:
> > On Sat, Jun 17, 2023 at 01:38:29PM -0700, Andy Lutomirski wrote:
> >> On Fri, Jun 16, 2023, at 1:50 AM, Mike Rapoport wrote:
> >> > From: "Mike Rapoport (IBM)" <rppt at kernel.org>
> >> >
> >> > module_alloc() is used everywhere as a mean to allocate memory for code.
> >> >
> >> > Beside being semantically wrong, this unnecessarily ties all subsystems
> >> > that need to allocate code, such as ftrace, kprobes and BPF to modules
> >> > and puts the burden of code allocation to the modules code.
> >> >
> >> > Several architectures override module_alloc() because of various
> >> > constraints where the executable memory can be located and this causes
> >> > additional obstacles for improvements of code allocation.
> >> >
> >> > Start splitting code allocation from modules by introducing
> >> > execmem_text_alloc(), execmem_free(), jit_text_alloc(), jit_free() APIs.
> >> >
> >> > Initially, execmem_text_alloc() and jit_text_alloc() are wrappers for
> >> > module_alloc() and execmem_free() and jit_free() are replacements of
> >> > module_memfree() to allow updating all call sites to use the new APIs.
> >> >
> >> > The intention semantics for new allocation APIs:
> >> >
> >> > * execmem_text_alloc() should be used to allocate memory that must reside
> >> > close to the kernel image, like loadable kernel modules and generated
> >> > code that is restricted by relative addressing.
> >> >
> >> > * jit_text_alloc() should be used to allocate memory for generated code
> >> > when there are no restrictions for the code placement. For
> >> > architectures that require that any code is within certain distance
> >> > from the kernel image, jit_text_alloc() will be essentially aliased to
> >> > execmem_text_alloc().
> >> >
> >>
> >> Is there anything in this series to help users do the appropriate
> >> synchronization when the actually populate the allocated memory with
> >> code? See here, for example:
> >
> > This series only factors out the executable allocations from modules and
> > puts them in a central place.
> > Anything else would go on top after this lands.
>
> Hmm.
>
> On the one hand, there's nothing wrong with factoring out common code. On the
> other hand, this is probably the right time to at least start thinking about
> synchronization, at least to the extent that it might make us want to change
> this API. (I'm not at all saying that this series should require changes --
> I'm just saying that this is a good time to think about how this should
> work.)
>
> The current APIs, *and* the proposed jit_text_alloc() API, don't actually
> look like the one think in the Linux ecosystem that actually intelligently
> and efficiently maps new text into an address space: mmap().
>
> On x86, you can mmap() an existing file full of executable code PROT_EXEC and
> jump to it with minimal synchronization (just the standard implicit ordering
> in the kernel that populates the pages before setting up the PTEs and
> whatever user synchronization is needed to avoid jumping into the mapping
> before mmap() finishes). It works across CPUs, and the only possible way
> userspace can screw it up (for a read-only mapping of read-only text, anyway)
> is to jump to the mapping too early, in which case userspace gets a page
> fault. Incoherence is impossible, and no one needs to "serialize" (in the
> SDM sense).
>
> I think the same sequence (from userspace's perspective) works on other
> architectures, too, although I think more cache management is needed on the
> kernel's end. As far as I know, no Linux SMP architecture needs an IPI to
> map executable text into usermode, but I could easily be wrong. (IIRC RISC-V
> has very developer-unfriendly icache management, but I don't remember the
> details.)
That's my understanding too, with a couple of details:
1) After the copy we perform and complete all the data + instruction cache
maintenance *before* marking the mapping as executable.
2) Even *after* the mapping is marked executable, a thread could take a
spurious fault on an instruction fetch for the new instructions. One way to
think about this is that the CPU attempted to speculate the instructions
earlier, saw that the mapping was faulting, and placed a "generate a fault
here" operation into its pipeline to generate that later.
The CPU pipeline/OoO-engine/whatever is effectively a transient cache for
operations in-flight which is only ever "invalidated" by a
context-synchronization-event (akin to an x86 serializing effect).
We're only guarnateed to have a new instruction fetch (from the I-cache into
the CPU pipeline) after the next context synchronization event (akin to an x86
serializing effect), and luckily out exception entry/exit is architecturally
guarnateed to provide that (unless we explicitly opt out via a control bit).
I know we're a bit lax with that today: I think we omit the
context-synchronization-event when enabling ftrace callsites, and worse, for
static keys. Those are both on my TODO list of nasty problems that require
careful auditing...
> Of course, using ptrace or any other FOLL_FORCE to modify text on x86 is
> rather fraught, and I bet many things do it wrong when userspace is
> multithreaded. But not in production because it's mostly not used in
> production.)
I suspect uprobes needs a look too...
I'll need to go dig into all that a bit before I have more of an opinion on the
shape of the API.
Thanks,
Mark.
> But jit_text_alloc() can't do this, because the order of operations doesn't
> match. With jit_text_alloc(), the executable mapping shows up before the
> text is populated, so there is no atomic change from not-there to
> populated-and-executable. Which means that there is an opportunity for CPUs,
> speculatively or otherwise, to start filling various caches with intermediate
> states of the text, which means that various architectures (even x86!) may
> need serialization.
>
> For eBPF- and module- like use cases, where JITting/code gen is quite
> coarse-grained, perhaps something vaguely like:
>
> jit_text_alloc() -> returns a handle and an executable virtual address, but does *not* map it there
> jit_text_write() -> write to that handle
> jit_text_map() -> map it and synchronize if needed (no sync needed on x86, I think)
>
> could be more efficient and/or safer.
>
> (Modules could use this too. Getting alternatives right might take some
> fiddling, because off the top of my head, this doesn't match how it works
> now.)
>
> To make alternatives easier, this could work, maybe (haven't fully thought it through):
>
> jit_text_alloc()
> jit_text_map_rw_inplace() -> map at the target address, but RW, !X
>
> write the text and apply alternatives
>
> jit_text_finalize() -> change from RW to RX *and synchronize*
>
> jit_text_finalize() would either need to wait for RCU (possibly extra heavy
> weight RCU to get "serialization") or send an IPI.
>
> This is slower than the alloc, write, map solution, but allows alternatives
> to be applied at the final address.
>
>
> Even fancier variants where the writing is some using something like
> use_temporary_mm() might even make sense.
>
>
> To what extent does performance matter for the various users? module loading
> is slow, and I don't think we care that much. eBPF loaded is not super fast,
> and we care to a limited extent. I *think* the bcachefs use case needs to be
> very fast, but I'm not sure it can be fast and supportable.
>
> Anyway, food for thought.
>
More information about the Linuxppc-dev
mailing list