From patchwork Fri Nov 25 02:35:22 2011 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Pingfan Liu X-Patchwork-Id: 127635 Return-Path: X-Original-To: incoming@patchwork.ozlabs.org Delivered-To: patchwork-incoming@bilbo.ozlabs.org Received: from lists.gnu.org (lists.gnu.org [140.186.70.17]) (using TLSv1 with cipher AES256-SHA (256/256 bits)) (Client did not present a certificate) by ozlabs.org (Postfix) with ESMTPS id 9F1BC1007D1 for ; Fri, 25 Nov 2011 13:36:09 +1100 (EST) Received: from localhost ([::1]:50213 helo=lists.gnu.org) by lists.gnu.org with esmtp (Exim 4.71) (envelope-from ) id 1RTle2-000487-Vx for incoming@patchwork.ozlabs.org; Thu, 24 Nov 2011 21:36:06 -0500 Received: from eggs.gnu.org ([140.186.70.92]:38235) by lists.gnu.org with esmtp (Exim 4.71) (envelope-from ) id 1RTldu-000431-Q9 for qemu-devel@nongnu.org; Thu, 24 Nov 2011 21:35:59 -0500 Received: from Debian-exim by eggs.gnu.org with spam-scanned (Exim 4.71) (envelope-from ) id 1RTldt-0007GB-QX for qemu-devel@nongnu.org; Thu, 24 Nov 2011 21:35:58 -0500 Received: from mail-iy0-f173.google.com ([209.85.210.173]:39535) by eggs.gnu.org with esmtp (Exim 4.71) (envelope-from ) id 1RTldt-0007DM-L5 for qemu-devel@nongnu.org; Thu, 24 Nov 2011 21:35:57 -0500 Received: by mail-iy0-f173.google.com with SMTP id k32so4328882iak.4 for ; Thu, 24 Nov 2011 18:35:57 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=from:to:cc:subject:date:message-id:x-mailer:in-reply-to:references; bh=0tEQ8/4OW3bCBJXX4nEe1a8N2+sJjGZOPnP2d1obvJQ=; b=sJpEJ7fu42Bh8vouGUUJ4YQFIem7h1ANqn9J1ypkZGngmxlkULm+9TURjgghqtLYnw bCCxL1z5xHkFhJEinn8PxA0gHi946Wu6oF7mrgqZxtXFIIFkueUpKG4U/fSa0YjUyslL uLasyMKbxvnsHi0vihLlDzNkE2t79hT6SHNS8= Received: by 10.50.158.227 with SMTP id wx3mr35045648igb.52.1322188557394; Thu, 24 Nov 2011 18:35:57 -0800 (PST) Received: from localhost ([202.108.130.138]) by mx.google.com with ESMTPS id ft1sm53243969igc.3.2011.11.24.18.35.55 (version=TLSv1/SSLv3 cipher=OTHER); Thu, 24 Nov 2011 18:35:56 -0800 (PST) From: Liu Ping Fan To: kvm@vger.kernel.org, qemu-devel@nongnu.org Date: Fri, 25 Nov 2011 10:35:22 +0800 Message-Id: <1322188529-11609-2-git-send-email-kernelfans@gmail.com> X-Mailer: git-send-email 1.7.4.4 In-Reply-To: <1322188529-11609-1-git-send-email-kernelfans@gmail.com> References: <1322188529-11609-1-git-send-email-kernelfans@gmail.com> X-detected-operating-system: by eggs.gnu.org: GNU/Linux 2.6 (newer, 2) X-Received-From: 209.85.210.173 Cc: aliguori@us.ibm.com, Liu Ping Fan , linux-kernel@vger.kernel.org, ryanh@us.ibm.com, jan.kiszka@web.de, avi@redhat.com Subject: [Qemu-devel] [PATCH 1/2] kvm: make vcpu life cycle separated from kvm instance X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.14 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Errors-To: qemu-devel-bounces+incoming=patchwork.ozlabs.org@nongnu.org Sender: qemu-devel-bounces+incoming=patchwork.ozlabs.org@nongnu.org From: Liu Ping Fan Currently, vcpu can be destructed only when kvm instance destroyed. Change this to vcpu as a refer to kvm, and then vcpu MUST and CAN be destroyed before kvm's destroy. Qemu will take advantage of this to exit the vcpu thread if the thread is no longer in use by guest. Signed-off-by: Liu Ping Fan --- arch/x86/kvm/x86.c | 28 ++++++++-------------------- include/linux/kvm_host.h | 2 ++ virt/kvm/kvm_main.c | 31 +++++++++++++++++++++++++++++-- 3 files changed, 39 insertions(+), 22 deletions(-) diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c index c38efd7..ea2315a 100644 --- a/arch/x86/kvm/x86.c +++ b/arch/x86/kvm/x86.c @@ -6560,27 +6560,16 @@ static void kvm_unload_vcpu_mmu(struct kvm_vcpu *vcpu) vcpu_put(vcpu); } -static void kvm_free_vcpus(struct kvm *kvm) +void kvm_arch_vcpu_zap(struct kref *ref) { - unsigned int i; - struct kvm_vcpu *vcpu; - - /* - * Unpin any mmu pages first. - */ - kvm_for_each_vcpu(i, vcpu, kvm) { - kvm_clear_async_pf_completion_queue(vcpu); - kvm_unload_vcpu_mmu(vcpu); - } - kvm_for_each_vcpu(i, vcpu, kvm) - kvm_arch_vcpu_free(vcpu); - - mutex_lock(&kvm->lock); - for (i = 0; i < atomic_read(&kvm->online_vcpus); i++) - kvm->vcpus[i] = NULL; + struct kvm_vcpu *vcpu = container_of(ref, struct kvm_vcpu, refcount); + struct kvm *kvm = vcpu->kvm; - atomic_set(&kvm->online_vcpus, 0); - mutex_unlock(&kvm->lock); + printk(KERN_INFO "%s, zap vcpu:0x%x\n", __func__, vcpu->vcpu_id); + kvm_clear_async_pf_completion_queue(vcpu); + kvm_unload_vcpu_mmu(vcpu); + kvm_arch_vcpu_free(vcpu); + kvm_put_kvm(kvm); } void kvm_arch_sync_events(struct kvm *kvm) @@ -6594,7 +6583,6 @@ void kvm_arch_destroy_vm(struct kvm *kvm) kvm_iommu_unmap_guest(kvm); kfree(kvm->arch.vpic); kfree(kvm->arch.vioapic); - kvm_free_vcpus(kvm); if (kvm->arch.apic_access_page) put_page(kvm->arch.apic_access_page); if (kvm->arch.ept_identity_pagetable) diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h index d526231..fe35078 100644 --- a/include/linux/kvm_host.h +++ b/include/linux/kvm_host.h @@ -113,6 +113,7 @@ enum { struct kvm_vcpu { struct kvm *kvm; + struct kref refcount; #ifdef CONFIG_PREEMPT_NOTIFIERS struct preempt_notifier preempt_notifier; #endif @@ -460,6 +461,7 @@ void kvm_arch_exit(void); int kvm_arch_vcpu_init(struct kvm_vcpu *vcpu); void kvm_arch_vcpu_uninit(struct kvm_vcpu *vcpu); +void kvm_arch_vcpu_zap(struct kref *ref); void kvm_arch_vcpu_free(struct kvm_vcpu *vcpu); void kvm_arch_vcpu_load(struct kvm_vcpu *vcpu, int cpu); void kvm_arch_vcpu_put(struct kvm_vcpu *vcpu); diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c index d9cfb78..f166bc8 100644 --- a/virt/kvm/kvm_main.c +++ b/virt/kvm/kvm_main.c @@ -580,6 +580,7 @@ static void kvm_destroy_vm(struct kvm *kvm) kvm_arch_free_vm(kvm); hardware_disable_all(); mmdrop(mm); + printk(KERN_INFO "%s finished\n", __func__); } void kvm_get_kvm(struct kvm *kvm) @@ -1503,6 +1504,16 @@ void mark_page_dirty(struct kvm *kvm, gfn_t gfn) mark_page_dirty_in_slot(kvm, memslot, gfn); } +void kvm_vcpu_get(struct kvm_vcpu *vcpu) +{ + kref_get(&vcpu->refcount); +} + +void kvm_vcpu_put(struct kvm_vcpu *vcpu) +{ + kref_put(&vcpu->refcount, kvm_arch_vcpu_zap); +} + /* * The vCPU has executed a HLT instruction with in-kernel mode enabled. */ @@ -1623,8 +1634,13 @@ static int kvm_vcpu_mmap(struct file *file, struct vm_area_struct *vma) static int kvm_vcpu_release(struct inode *inode, struct file *filp) { struct kvm_vcpu *vcpu = filp->private_data; + struct kvm *kvm = vcpu->kvm; - kvm_put_kvm(vcpu->kvm); + filp->private_data = NULL; + mutex_lock(&kvm->lock); + atomic_sub(1, &kvm->online_vcpus); + mutex_unlock(&kvm->lock); + kvm_vcpu_put(vcpu); return 0; } @@ -1646,6 +1662,17 @@ static int create_vcpu_fd(struct kvm_vcpu *vcpu) return anon_inode_getfd("kvm-vcpu", &kvm_vcpu_fops, vcpu, O_RDWR); } +static struct kvm_vcpu *kvm_vcpu_create(struct kvm *kvm, u32 id) +{ + struct kvm_vcpu *vcpu; + vcpu = kvm_arch_vcpu_create(kvm, id); + if (IS_ERR(vcpu)) + return vcpu; + + kref_init(&vcpu->refcount); + return vcpu; +} + /* * Creates some virtual cpus. Good luck creating more than one. */ @@ -1654,7 +1681,7 @@ static int kvm_vm_ioctl_create_vcpu(struct kvm *kvm, u32 id) int r; struct kvm_vcpu *vcpu, *v; - vcpu = kvm_arch_vcpu_create(kvm, id); + vcpu = kvm_vcpu_create(kvm, id); if (IS_ERR(vcpu)) return PTR_ERR(vcpu);