ACK: [SRU][Cosmic][PATCH 1/4] kvm: svm: Ensure an IBPB on all affected CPUs when freeing a vmcb

Tyler Hicks tyhicks at canonical.com
Mon Mar 25 18:18:22 UTC 2019


On 2019-03-25 14:55:52, Juerg Haefliger wrote:
> From: Jim Mattson <jmattson at google.com>
> 
> Previously, we only called indirect_branch_prediction_barrier on the
> logical CPU that freed a vmcb. This function should be called on all
> logical CPUs that last loaded the vmcb in question.
> 
> Fixes: 15d45071523d ("KVM/x86: Add IBPB support")
> Reported-by: Neel Natu <neelnatu at google.com>
> Signed-off-by: Jim Mattson <jmattson at google.com>
> Reviewed-by: Konrad Rzeszutek Wilk <konrad.wilk at oracle.com>
> Cc: stable at vger.kernel.org
> Signed-off-by: Paolo Bonzini <pbonzini at redhat.com>
> 
> CVE-2017-5715
> 
> (cherry picked from commit fd65d3142f734bc4376053c8d75670041903134d)
> Signed-off-by: Juerg Haefliger <juergh at canonical.com>

Acked-by: Tyler Hicks <tyhicks at canonical.com>

Tyler

> ---
>  arch/x86/kvm/svm.c | 20 +++++++++++++++-----
>  1 file changed, 15 insertions(+), 5 deletions(-)
> 
> diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c
> index 3e59a187fe30..75d5f180ffa5 100644
> --- a/arch/x86/kvm/svm.c
> +++ b/arch/x86/kvm/svm.c
> @@ -2188,21 +2188,31 @@ static struct kvm_vcpu *svm_create_vcpu(struct kvm *kvm, unsigned int id)
>  	return ERR_PTR(err);
>  }
>  
> +static void svm_clear_current_vmcb(struct vmcb *vmcb)
> +{
> +	int i;
> +
> +	for_each_online_cpu(i)
> +		cmpxchg(&per_cpu(svm_data, i)->current_vmcb, vmcb, NULL);
> +}
> +
>  static void svm_free_vcpu(struct kvm_vcpu *vcpu)
>  {
>  	struct vcpu_svm *svm = to_svm(vcpu);
>  
> +	/*
> +	 * The vmcb page can be recycled, causing a false negative in
> +	 * svm_vcpu_load(). So, ensure that no logical CPU has this
> +	 * vmcb page recorded as its current vmcb.
> +	 */
> +	svm_clear_current_vmcb(svm->vmcb);
> +
>  	__free_page(pfn_to_page(__sme_clr(svm->vmcb_pa) >> PAGE_SHIFT));
>  	__free_pages(virt_to_page(svm->msrpm), MSRPM_ALLOC_ORDER);
>  	__free_page(virt_to_page(svm->nested.hsave));
>  	__free_pages(virt_to_page(svm->nested.msrpm), MSRPM_ALLOC_ORDER);
>  	kvm_vcpu_uninit(vcpu);
>  	kmem_cache_free(kvm_vcpu_cache, svm);
> -	/*
> -	 * The vmcb page can be recycled, causing a false negative in
> -	 * svm_vcpu_load(). So do a full IBPB now.
> -	 */
> -	indirect_branch_prediction_barrier();
>  }
>  
>  static void svm_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
> -- 
> 2.19.1
> 
> 
> -- 
> kernel-team mailing list
> kernel-team at lists.ubuntu.com
> https://lists.ubuntu.com/mailman/listinfo/kernel-team



More information about the kernel-team mailing list