From patchwork Wed Dec 11 08:47:54 2013 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Pingfan Liu X-Patchwork-Id: 299880 Return-Path: X-Original-To: incoming@patchwork.ozlabs.org Delivered-To: patchwork-incoming@bilbo.ozlabs.org Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by ozlabs.org (Postfix) with ESMTP id 3E4B92C03E4 for ; Wed, 11 Dec 2013 19:47:17 +1100 (EST) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1750873Ab3LKIrQ (ORCPT ); Wed, 11 Dec 2013 03:47:16 -0500 Received: from mail-pd0-f173.google.com ([209.85.192.173]:56117 "EHLO mail-pd0-f173.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1750779Ab3LKIrQ (ORCPT ); Wed, 11 Dec 2013 03:47:16 -0500 Received: by mail-pd0-f173.google.com with SMTP id p10so9065405pdj.18 for ; Wed, 11 Dec 2013 00:47:15 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20120113; h=from:to:cc:subject:date:message-id:in-reply-to:references; bh=CZA5yiuTfU2c/UDZsCsVVY+VH8EZHPELWizHwwpt9cM=; b=i7SDuBva8L+hdKQFegQpnLdZ+8ap5DB83HdZ9bIdPwx7uWxdNslXrm11+yyOWq8lm3 CoyL2GsR6nLxHSupRziOwLJGbdZjHZCVDS+fbfXon+nrN0sCfUiCW62ymZoAa/I6NEO3 ewQNUxLa5bfUezd8+wo+jXHby0voKoXkSqHkkfq9wc/Ce+IJ2jo8XcVrrzlg9SF6E1ZN 9X741JqTWTWzH7xgEmtv732I75Q+MDEIWoLhzDnXwtpx4mhk0I1dUtvnDPTJ9iQAlWC9 cchj8KTmjBJnQ38Mz1JwgW4iwtc7H8iah659k1X7VxhiVTPF8W39hYZ6ZJAbLtVHvPbV PyKw== X-Received: by 10.69.12.36 with SMTP id en4mr375807pbd.54.1386751635808; Wed, 11 Dec 2013 00:47:15 -0800 (PST) Received: from localhost ([202.108.130.138]) by mx.google.com with ESMTPSA id bh6sm42797866pad.20.2013.12.11.00.47.14 for (version=TLSv1.2 cipher=RC4-SHA bits=128/128); Wed, 11 Dec 2013 00:47:15 -0800 (PST) From: Liu Ping Fan To: linuxppc-dev@lists.ozlabs.org, kvm-ppc@vger.kernel.org Cc: "Aneesh Kumar K.V" , Benjamin Herrenschmidt , Paul Mackerras , Alexander Graf Subject: [PATCH 4/4] powernv: kvm: make the handling of _PAGE_NUMA faster for guest Date: Wed, 11 Dec 2013 16:47:54 +0800 Message-Id: <1386751674-14136-5-git-send-email-pingfank@linux.vnet.ibm.com> X-Mailer: git-send-email 1.8.1.4 In-Reply-To: <1386751674-14136-1-git-send-email-pingfank@linux.vnet.ibm.com> References: <1386751674-14136-1-git-send-email-pingfank@linux.vnet.ibm.com> Sender: kvm-ppc-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: kvm-ppc@vger.kernel.org The period check of _PAGE_NUMA can probably happen on the correctly placed page. For this case, when guest try to setup hpte in real mode, we try to resolve the numa fault in real mode, since the switch between guest context and host context costs too much. Signed-off-by: Liu Ping Fan --- arch/powerpc/kvm/book3s_hv_rm_mmu.c | 32 ++++++++++++++++++++++++++++++-- 1 file changed, 30 insertions(+), 2 deletions(-) diff --git a/arch/powerpc/kvm/book3s_hv_rm_mmu.c b/arch/powerpc/kvm/book3s_hv_rm_mmu.c index ae46052..a06b199 100644 --- a/arch/powerpc/kvm/book3s_hv_rm_mmu.c +++ b/arch/powerpc/kvm/book3s_hv_rm_mmu.c @@ -179,6 +179,11 @@ long kvmppc_do_h_enter(struct kvm *kvm, unsigned long flags, unsigned int writing; unsigned long mmu_seq; unsigned long rcbits; + struct mm_struct *mm = kvm->mm; + struct vm_area_struct *vma; + int page_nid, target_nid; + struct page *test_page; + pte_t *ptep; psize = hpte_page_size(pteh, ptel); if (!psize) @@ -234,8 +239,26 @@ long kvmppc_do_h_enter(struct kvm *kvm, unsigned long flags, /* Look up the Linux PTE for the backing page */ pte_size = psize; - pte = lookup_linux_pte(pgdir, hva, writing, &pte_size, NULL); - if (pte_present(pte) && !pte_numa(pte)) { + pte = lookup_linux_pte(pgdir, hva, writing, &pte_size, &ptep); + if (pte_present(pte)) { + if (pte_numa(pte)) { + /* If fail, let gup handle it */ + if (unlikely(!down_read_trylock(&mm->mmap_sem))) + goto pte_check; + + vma = find_vma(mm, hva); + up_read(&mm->mmap_sem); + test_page = pte_page(pte); + page_nid = page_to_nid(test_page); + target_nid = numa_migrate_prep(test_page, vma, + hva, page_nid); + put_page(test_page); + if (unlikely(target_nid != -1)) { + /* If fail, let gup handle it */ + goto pte_check; + } + } + if (writing && !pte_write(pte)) /* make the actual HPTE be read-only */ ptel = hpte_make_readonly(ptel); @@ -244,6 +267,7 @@ long kvmppc_do_h_enter(struct kvm *kvm, unsigned long flags, } } +pte_check: if (pte_size < psize) return H_PARAMETER; if (pa && pte_size > psize) @@ -339,6 +363,10 @@ long kvmppc_do_h_enter(struct kvm *kvm, unsigned long flags, pteh &= ~HPTE_V_VALID; unlock_rmap(rmap); } else { + if (pte_numa(pte) && pa) { + pte = pte_mknonnuma(pte); + *ptep = pte; + } kvmppc_add_revmap_chain(kvm, rev, rmap, pte_index, realmode); /* Only set R/C in real HPTE if already set in *rmap */