Re: [PATCH A] KVM: Simplify kvm->tlbs_dirty handling

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Il 18/02/2014 09:22, Takuya Yoshikawa ha scritto:
> When this was introduced, kvm_flush_remote_tlbs() could be called
> without holding mmu_lock.  It is now acknowledged that the function
> must be called before releasing mmu_lock, and all callers have already
> been changed to do so.
> 
> There is no need to use smp_mb() and cmpxchg() any more.
> 
> Signed-off-by: Takuya Yoshikawa <yoshikawa_takuya_b1@xxxxxxxxxxxxx>

I prefer this patch, and in fact we can make it even simpler:

diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h
index ed1cc89..9816b68 100644
--- a/include/linux/kvm_host.h
+++ b/include/linux/kvm_host.h
@@ -401,7 +401,9 @@ struct kvm {
 	unsigned long mmu_notifier_seq;
 	long mmu_notifier_count;
 #endif
+	/* Protected by mmu_lock */
 	bool tlbs_dirty;
+
 	struct list_head devices;
 };
 
diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c
index 51744da..f5668a4 100644
--- a/virt/kvm/kvm_main.c
+++ b/virt/kvm/kvm_main.c
@@ -188,12 +188,6 @@ void kvm_flush_remote_tlbs(struct kvm *kvm)
 {
 	if (make_all_cpus_request(kvm, KVM_REQ_TLB_FLUSH))
 		++kvm->stat.remote_tlb_flush;
-	/*
-	 * tlbs_dirty is used only for optimizing x86's shadow paging code with
-	 * mmu notifiers in mind, see the note on sync_page().  Since it is
-	 * always protected with mmu_lock there, should kvm_flush_remote_tlbs()
-	 * be called before releasing mmu_lock, this is safe.
-	 */
 	kvm->tlbs_dirty = false;
 }
 EXPORT_SYMBOL_GPL(kvm_flush_remote_tlbs);


What do you think?

Paolo
> ---
>  arch/x86/kvm/paging_tmpl.h |    7 ++++---
>  include/linux/kvm_host.h   |    2 +-
>  virt/kvm/kvm_main.c        |   11 +++++++----
>  3 files changed, 12 insertions(+), 8 deletions(-)
> 
> diff --git a/arch/x86/kvm/paging_tmpl.h b/arch/x86/kvm/paging_tmpl.h
> index cba218a..b1e6c1b 100644
> --- a/arch/x86/kvm/paging_tmpl.h
> +++ b/arch/x86/kvm/paging_tmpl.h
> @@ -913,7 +913,8 @@ static gpa_t FNAME(gva_to_gpa_nested)(struct kvm_vcpu *vcpu, gva_t vaddr,
>   *   and kvm_mmu_notifier_invalidate_range_start detect the mapping page isn't
>   *   used by guest then tlbs are not flushed, so guest is allowed to access the
>   *   freed pages.
> - *   And we increase kvm->tlbs_dirty to delay tlbs flush in this case.
> + *   We set tlbs_dirty to let the notifier know this change and delay the flush
> + *   until such a case actually happens.
>   */
>  static int FNAME(sync_page)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp)
>  {
> @@ -942,7 +943,7 @@ static int FNAME(sync_page)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp)
>  			return -EINVAL;
>  
>  		if (FNAME(prefetch_invalid_gpte)(vcpu, sp, &sp->spt[i], gpte)) {
> -			vcpu->kvm->tlbs_dirty++;
> +			vcpu->kvm->tlbs_dirty = true;
>  			continue;
>  		}
>  
> @@ -957,7 +958,7 @@ static int FNAME(sync_page)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp)
>  
>  		if (gfn != sp->gfns[i]) {
>  			drop_spte(vcpu->kvm, &sp->spt[i]);
> -			vcpu->kvm->tlbs_dirty++;
> +			vcpu->kvm->tlbs_dirty = true;
>  			continue;
>  		}
>  
> diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h
> index f5937b8..ed1cc89 100644
> --- a/include/linux/kvm_host.h
> +++ b/include/linux/kvm_host.h
> @@ -401,7 +401,7 @@ struct kvm {
>  	unsigned long mmu_notifier_seq;
>  	long mmu_notifier_count;
>  #endif
> -	long tlbs_dirty;
> +	bool tlbs_dirty;
>  	struct list_head devices;
>  };
>  
> diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c
> index a9e999a..51744da 100644
> --- a/virt/kvm/kvm_main.c
> +++ b/virt/kvm/kvm_main.c
> @@ -186,12 +186,15 @@ static bool make_all_cpus_request(struct kvm *kvm, unsigned int req)
>  
>  void kvm_flush_remote_tlbs(struct kvm *kvm)
>  {
> -	long dirty_count = kvm->tlbs_dirty;
> -
> -	smp_mb();
>  	if (make_all_cpus_request(kvm, KVM_REQ_TLB_FLUSH))
>  		++kvm->stat.remote_tlb_flush;
> -	cmpxchg(&kvm->tlbs_dirty, dirty_count, 0);
> +	/*
> +	 * tlbs_dirty is used only for optimizing x86's shadow paging code with
> +	 * mmu notifiers in mind, see the note on sync_page().  Since it is
> +	 * always protected with mmu_lock there, should kvm_flush_remote_tlbs()
> +	 * be called before releasing mmu_lock, this is safe.
> +	 */
> +	kvm->tlbs_dirty = false;
>  }
>  EXPORT_SYMBOL_GPL(kvm_flush_remote_tlbs);
>  
> 

--
To unsubscribe from this list: send the line "unsubscribe kvm" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [KVM ARM]     [KVM ia64]     [KVM ppc]     [Virtualization Tools]     [Spice Development]     [Libvirt]     [Libvirt Users]     [Linux USB Devel]     [Linux Audio Users]     [Yosemite Questions]     [Linux Kernel]     [Linux SCSI]     [XFree86]
  Powered by Linux