Follow @Openwall on Twitter for new release announcements and other news
[<prev] [next>] [thread-next>] [day] [month] [year] [list]
Date: Tue, 20 Nov 2018 16:26:19 -0700
From: Tycho Andersen <>
To: Julian Stecklina <>
Cc:, Liran Alon <>,
	Jonathan Adams <>,
	David Woodhouse <>,
	Igor Stoppa <>
Subject: Re: [RFC PATCH 0/6] Process-local memory allocations

On Tue, Nov 20, 2018 at 03:07:59PM +0100, Julian Stecklina wrote:
> In a world with processor information leak vulnerabilities, having a treasure
> trove of information available for leaking in the global kernel address space is
> starting to be a liability. The biggest offender is the linear mapping of all
> physical memory and there are already efforts (XPFO) to start addressing this.
> In this patch series, I'd like to propose breaking up the kernel address space
> further and introduce process-local mappings in the kernel.
> The rationale is that there are allocations in the kernel containing data that
> should only be accessible when the kernel is executing in the context of a
> specific process. A prime example is KVM vCPU state. This patch series
> introduces process-local memory in the kernel address space by claiming a PGD
> entry for this specific purpose. Then it converts KVM on x86 to use these new
> primitives to store GPR and FPU registers of vCPUs. KVM is a good testing
> ground, because it makes sure userspace can only interact with a VM from a
> single process.
> Process-local allocations in the kernel can be part of a robust L1TF mitigation
> strategy that works even with SMT enabled. The specific goal here is to make it
> harder for a random thread using cache load gadget (usually a bounds check of a
> system call argument plus array access suffices) to prefetch interesting data
> into the L1 cache and use L1TF to leak this data.
> The patch set applies to kvm/next [1]. Feedback is very welcome, both about the
> general approach and the actual implementation. As far as testing goes, the KVM
> unit tests seem happy on Intel. AMD is only compile tested at the moment.

This seems similar in spirit to prmem:

Basically, we have some special memory that we want to leave unmapped
(read only) most of the time, but map it (writable) sometimes. I
wonder if we should merge the APIs in to one

spmemalloc(size, flags, PRLOCAL)

type thing? Could we share some infrastructure then? (I also didn't
follow what happened to the patches Nadav was going to send that might
replace prmem somehow.)


Powered by blists - more mailing lists

Confused about mailing lists and their use? Read about mailing lists on Wikipedia and check out these guidelines on proper formatting of your messages.