[Xenial PATCH 2/3] UBUNTU: SAUCE: s390/mm: fix local TLB flushing vs. detach of an mm address space
Colin Ian King
colin.king at canonical.com
Tue Sep 12 11:58:20 UTC 2017
On 12/09/17 10:45, Stefan Bader wrote:
> From: Martin Schwidefsky <schwidefsky at de.ibm.com>
>
> BugLink: http://bugs.launchpad.net/bugs/1708399
>
> The local TLB flushing code keeps an additional mask in the mm.context,
> the cpu_attach_mask. At the time a global flush of an address space is
> done the cpu_attach_mask is copied to the mm_cpumask in order to avoid
> future global flushes in case the mm is used by a single CPU only after
> the flush.
>
> Trouble is that the reset of the mm_cpumask is racy against the detach
> of an mm address space by switch_mm. The current order is first the
> global TLB flush and then the copy of the cpu_attach_mask to the
> mm_cpumask. The order needs to be the other way around.
>
> Cc: <stable at vger.kernel.org>
> Reviewed-by: Heiko Carstens <heiko.carstens at de.ibm.com>
> Signed-off-by: Martin Schwidefsky <schwidefsky at de.ibm.com>
> (backported from b3e5dc45fd1ec2aa1de6b80008f9295eb17e0659 linux-next)
> [merged with "s390/mm,kvm: flush gmap address space with IDTE"]
> Signed-off-by: Stefan Bader <stefan.bader at canonical.com>
> ---
> arch/s390/include/asm/tlbflush.h | 56 ++++++++++++----------------------------
> 1 file changed, 16 insertions(+), 40 deletions(-)
>
> diff --git a/arch/s390/include/asm/tlbflush.h b/arch/s390/include/asm/tlbflush.h
> index 80868c84..d54cc83 100644
> --- a/arch/s390/include/asm/tlbflush.h
> +++ b/arch/s390/include/asm/tlbflush.h
> @@ -47,47 +47,31 @@ static inline void __tlb_flush_global(void)
> }
>
> /*
> - * Flush TLB entries for a specific mm on all CPUs (in case gmap is used
> - * this implicates multiple ASCEs!).
> + * Flush TLB entries for a specific ASCE on all CPUs.
> */
> -static inline void __tlb_flush_full(struct mm_struct *mm)
> +static inline void __tlb_flush_mm(struct mm_struct * mm)
> {
> + /*
> + * If the machine has IDTE we prefer to do a per mm flush
> + * on all cpus instead of doing a local flush if the mm
> + * only ran on the local cpu.
> + */
> preempt_disable();
> atomic_add(0x10000, &mm->context.attach_count);
> - if (cpumask_equal(mm_cpumask(mm), cpumask_of(smp_processor_id()))) {
> - /* Local TLB flush */
> - __tlb_flush_local();
> + /* Reset TLB flush mask */
> + if (MACHINE_HAS_TLB_LC)
> + cpumask_copy(mm_cpumask(mm), &mm->context.cpu_attach_mask);
> + barrier();
> + if (MACHINE_HAS_IDTE && list_empty(&mm->context.gmap_list)) {
> + __tlb_flush_idte(mm->context.asce);
> } else {
> /* Global TLB flush */
> __tlb_flush_global();
> - /* Reset TLB flush mask */
> - if (MACHINE_HAS_TLB_LC)
> - cpumask_copy(mm_cpumask(mm),
> - &mm->context.cpu_attach_mask);
> }
> atomic_sub(0x10000, &mm->context.attach_count);
> preempt_enable();
> }
>
> -/*
> - * Flush TLB entries for a specific ASCE on all CPUs. Should never be used
> - * when more than one asce (e.g. gmap) ran on this mm.
> - */
> -static inline void __tlb_flush_asce(struct mm_struct *mm, unsigned long asce)
> -{
> - preempt_disable();
> - atomic_add(0x10000, &mm->context.attach_count);
> - if (MACHINE_HAS_IDTE)
> - __tlb_flush_idte(asce);
> - else
> - __tlb_flush_global();
> - /* Reset TLB flush mask */
> - if (MACHINE_HAS_TLB_LC)
> - cpumask_copy(mm_cpumask(mm), &mm->context.cpu_attach_mask);
> - atomic_sub(0x10000, &mm->context.attach_count);
> - preempt_enable();
> -}
> -
> static inline void __tlb_flush_kernel(void)
> {
> if (MACHINE_HAS_IDTE)
> @@ -97,7 +81,6 @@ static inline void __tlb_flush_kernel(void)
> }
> #else
> #define __tlb_flush_global() __tlb_flush_local()
> -#define __tlb_flush_full(mm) __tlb_flush_local()
>
> /*
> * Flush TLB entries for a specific ASCE on all CPUs.
> @@ -111,21 +94,14 @@ static inline void __tlb_flush_kernel(void)
> {
> __tlb_flush_local();
> }
> -#endif
>
> static inline void __tlb_flush_mm(struct mm_struct * mm)
> {
> - /*
> - * If the machine has IDTE we prefer to do a per mm flush
> - * on all cpus instead of doing a local flush if the mm
> - * only ran on the local cpu.
> - */
> - if (MACHINE_HAS_IDTE && list_empty(&mm->context.gmap_list))
> - __tlb_flush_asce(mm, mm->context.asce);
> - else
> - __tlb_flush_full(mm);
> + __tlb_flush_local();
> }
>
> +#endif
> +
> static inline void __tlb_flush_mm_lazy(struct mm_struct * mm)
> {
> if (mm->context.flush_mm) {
>
Maybe I'm being a bit slow here, but I'm finding this hard to see how
this is related to linux-next commit
b3e5dc45fd1ec2aa1de6b80008f9295eb17e0659
Do you mind explaining this backport?
Colin
More information about the kernel-team
mailing list