ACK: [SRU][Cosmic][PATCH 1/4] kvm: svm: Ensure an IBPB on all affected CPUs when freeing a vmcb
Tyler Hicks
tyhicks at canonical.com
Mon Mar 25 18:18:22 UTC 2019
On 2019-03-25 14:55:52, Juerg Haefliger wrote:
> From: Jim Mattson <jmattson at google.com>
>
> Previously, we only called indirect_branch_prediction_barrier on the
> logical CPU that freed a vmcb. This function should be called on all
> logical CPUs that last loaded the vmcb in question.
>
> Fixes: 15d45071523d ("KVM/x86: Add IBPB support")
> Reported-by: Neel Natu <neelnatu at google.com>
> Signed-off-by: Jim Mattson <jmattson at google.com>
> Reviewed-by: Konrad Rzeszutek Wilk <konrad.wilk at oracle.com>
> Cc: stable at vger.kernel.org
> Signed-off-by: Paolo Bonzini <pbonzini at redhat.com>
>
> CVE-2017-5715
>
> (cherry picked from commit fd65d3142f734bc4376053c8d75670041903134d)
> Signed-off-by: Juerg Haefliger <juergh at canonical.com>
Acked-by: Tyler Hicks <tyhicks at canonical.com>
Tyler
> ---
> arch/x86/kvm/svm.c | 20 +++++++++++++++-----
> 1 file changed, 15 insertions(+), 5 deletions(-)
>
> diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c
> index 3e59a187fe30..75d5f180ffa5 100644
> --- a/arch/x86/kvm/svm.c
> +++ b/arch/x86/kvm/svm.c
> @@ -2188,21 +2188,31 @@ static struct kvm_vcpu *svm_create_vcpu(struct kvm *kvm, unsigned int id)
> return ERR_PTR(err);
> }
>
> +static void svm_clear_current_vmcb(struct vmcb *vmcb)
> +{
> + int i;
> +
> + for_each_online_cpu(i)
> + cmpxchg(&per_cpu(svm_data, i)->current_vmcb, vmcb, NULL);
> +}
> +
> static void svm_free_vcpu(struct kvm_vcpu *vcpu)
> {
> struct vcpu_svm *svm = to_svm(vcpu);
>
> + /*
> + * The vmcb page can be recycled, causing a false negative in
> + * svm_vcpu_load(). So, ensure that no logical CPU has this
> + * vmcb page recorded as its current vmcb.
> + */
> + svm_clear_current_vmcb(svm->vmcb);
> +
> __free_page(pfn_to_page(__sme_clr(svm->vmcb_pa) >> PAGE_SHIFT));
> __free_pages(virt_to_page(svm->msrpm), MSRPM_ALLOC_ORDER);
> __free_page(virt_to_page(svm->nested.hsave));
> __free_pages(virt_to_page(svm->nested.msrpm), MSRPM_ALLOC_ORDER);
> kvm_vcpu_uninit(vcpu);
> kmem_cache_free(kvm_vcpu_cache, svm);
> - /*
> - * The vmcb page can be recycled, causing a false negative in
> - * svm_vcpu_load(). So do a full IBPB now.
> - */
> - indirect_branch_prediction_barrier();
> }
>
> static void svm_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
> --
> 2.19.1
>
>
> --
> kernel-team mailing list
> kernel-team at lists.ubuntu.com
> https://lists.ubuntu.com/mailman/listinfo/kernel-team
More information about the kernel-team
mailing list