diff mbox series

[v3,02/10] mm/ppc: Replace pXd_is_leaf() with pXd_leaf()

Message ID 20240305043750.93762-3-peterx@redhat.com (mailing list archive)
State Handled Elsewhere
Headers show
Series mm/treewide: Replace pXd_large() with pXd_leaf() | expand

Commit Message

Peter Xu March 5, 2024, 4:37 a.m. UTC
From: Peter Xu <peterx@redhat.com>

They're the same macros underneath.  Drop pXd_is_leaf(), instead always use
pXd_leaf().

At the meantime, instead of renames, drop the pXd_is_leaf() fallback
definitions directly in arch/powerpc/include/asm/pgtable.h. because similar
fallback macros for pXd_leaf() are already defined in
include/linux/pgtable.h.

Cc: Michael Ellerman <mpe@ellerman.id.au>
Cc: Nicholas Piggin <npiggin@gmail.com>
Cc: "Aneesh Kumar K.V" <aneesh.kumar@kernel.org>
Cc: "Naveen N. Rao" <naveen.n.rao@linux.ibm.com>
Cc: linuxppc-dev@lists.ozlabs.org
Suggested-by: Christophe Leroy <christophe.leroy@csgroup.eu>
Reviewed-by: Jason Gunthorpe <jgg@nvidia.com>
Signed-off-by: Peter Xu <peterx@redhat.com>
---
 arch/powerpc/include/asm/book3s/64/pgtable.h | 10 ++++----
 arch/powerpc/include/asm/pgtable.h           | 24 --------------------
 arch/powerpc/kvm/book3s_64_mmu_radix.c       | 12 +++++-----
 arch/powerpc/mm/book3s64/radix_pgtable.c     | 14 ++++++------
 arch/powerpc/mm/pgtable.c                    |  6 ++---
 arch/powerpc/mm/pgtable_64.c                 |  6 ++---
 arch/powerpc/xmon/xmon.c                     |  6 ++---
 7 files changed, 26 insertions(+), 52 deletions(-)

Comments

Christophe Leroy March 5, 2024, 5:31 p.m. UTC | #1
Le 05/03/2024 à 05:37, peterx@redhat.com a écrit :
> From: Peter Xu <peterx@redhat.com>
> 
> They're the same macros underneath.  Drop pXd_is_leaf(), instead always use
> pXd_leaf().
> 
> At the meantime, instead of renames, drop the pXd_is_leaf() fallback
> definitions directly in arch/powerpc/include/asm/pgtable.h. because similar
> fallback macros for pXd_leaf() are already defined in
> include/linux/pgtable.h.
> 
> Cc: Michael Ellerman <mpe@ellerman.id.au>
> Cc: Nicholas Piggin <npiggin@gmail.com>
> Cc: "Aneesh Kumar K.V" <aneesh.kumar@kernel.org>
> Cc: "Naveen N. Rao" <naveen.n.rao@linux.ibm.com>
> Cc: linuxppc-dev@lists.ozlabs.org
> Suggested-by: Christophe Leroy <christophe.leroy@csgroup.eu>
> Reviewed-by: Jason Gunthorpe <jgg@nvidia.com>
> Signed-off-by: Peter Xu <peterx@redhat.com>

Reviewed-by: Christophe Leroy <christophe.leroy@csgroup.eu>

In case you post a new version, in the subject, usually with use powerpc 
not ppc

> ---
>   arch/powerpc/include/asm/book3s/64/pgtable.h | 10 ++++----
>   arch/powerpc/include/asm/pgtable.h           | 24 --------------------
>   arch/powerpc/kvm/book3s_64_mmu_radix.c       | 12 +++++-----
>   arch/powerpc/mm/book3s64/radix_pgtable.c     | 14 ++++++------
>   arch/powerpc/mm/pgtable.c                    |  6 ++---
>   arch/powerpc/mm/pgtable_64.c                 |  6 ++---
>   arch/powerpc/xmon/xmon.c                     |  6 ++---
>   7 files changed, 26 insertions(+), 52 deletions(-)
> 
> diff --git a/arch/powerpc/include/asm/book3s/64/pgtable.h b/arch/powerpc/include/asm/book3s/64/pgtable.h
> index d1318e8582ac..3e99e409774a 100644
> --- a/arch/powerpc/include/asm/book3s/64/pgtable.h
> +++ b/arch/powerpc/include/asm/book3s/64/pgtable.h
> @@ -1439,18 +1439,16 @@ static inline bool is_pte_rw_upgrade(unsigned long old_val, unsigned long new_va
>   /*
>    * Like pmd_huge() and pmd_large(), but works regardless of config options
>    */
> -#define pmd_is_leaf pmd_is_leaf
> -#define pmd_leaf pmd_is_leaf
> +#define pmd_leaf pmd_leaf
>   #define pmd_large pmd_leaf
> -static inline bool pmd_is_leaf(pmd_t pmd)
> +static inline bool pmd_leaf(pmd_t pmd)
>   {
>   	return !!(pmd_raw(pmd) & cpu_to_be64(_PAGE_PTE));
>   }
>   
> -#define pud_is_leaf pud_is_leaf
> -#define pud_leaf pud_is_leaf
> +#define pud_leaf pud_leaf
>   #define pud_large pud_leaf
> -static inline bool pud_is_leaf(pud_t pud)
> +static inline bool pud_leaf(pud_t pud)
>   {
>   	return !!(pud_raw(pud) & cpu_to_be64(_PAGE_PTE));
>   }
> diff --git a/arch/powerpc/include/asm/pgtable.h b/arch/powerpc/include/asm/pgtable.h
> index 5928b3c1458d..e6edf1cdbc5b 100644
> --- a/arch/powerpc/include/asm/pgtable.h
> +++ b/arch/powerpc/include/asm/pgtable.h
> @@ -182,30 +182,6 @@ static inline void pte_frag_set(mm_context_t *ctx, void *p)
>   }
>   #endif
>   
> -#ifndef pmd_is_leaf
> -#define pmd_is_leaf pmd_is_leaf
> -static inline bool pmd_is_leaf(pmd_t pmd)
> -{
> -	return false;
> -}
> -#endif
> -
> -#ifndef pud_is_leaf
> -#define pud_is_leaf pud_is_leaf
> -static inline bool pud_is_leaf(pud_t pud)
> -{
> -	return false;
> -}
> -#endif
> -
> -#ifndef p4d_is_leaf
> -#define p4d_is_leaf p4d_is_leaf
> -static inline bool p4d_is_leaf(p4d_t p4d)
> -{
> -	return false;
> -}
> -#endif
> -
>   #define pmd_pgtable pmd_pgtable
>   static inline pgtable_t pmd_pgtable(pmd_t pmd)
>   {
> diff --git a/arch/powerpc/kvm/book3s_64_mmu_radix.c b/arch/powerpc/kvm/book3s_64_mmu_radix.c
> index 4a1abb9f7c05..408d98f8a514 100644
> --- a/arch/powerpc/kvm/book3s_64_mmu_radix.c
> +++ b/arch/powerpc/kvm/book3s_64_mmu_radix.c
> @@ -503,7 +503,7 @@ static void kvmppc_unmap_free_pmd(struct kvm *kvm, pmd_t *pmd, bool full,
>   	for (im = 0; im < PTRS_PER_PMD; ++im, ++p) {
>   		if (!pmd_present(*p))
>   			continue;
> -		if (pmd_is_leaf(*p)) {
> +		if (pmd_leaf(*p)) {
>   			if (full) {
>   				pmd_clear(p);
>   			} else {
> @@ -532,7 +532,7 @@ static void kvmppc_unmap_free_pud(struct kvm *kvm, pud_t *pud,
>   	for (iu = 0; iu < PTRS_PER_PUD; ++iu, ++p) {
>   		if (!pud_present(*p))
>   			continue;
> -		if (pud_is_leaf(*p)) {
> +		if (pud_leaf(*p)) {
>   			pud_clear(p);
>   		} else {
>   			pmd_t *pmd;
> @@ -635,12 +635,12 @@ int kvmppc_create_pte(struct kvm *kvm, pgd_t *pgtable, pte_t pte,
>   		new_pud = pud_alloc_one(kvm->mm, gpa);
>   
>   	pmd = NULL;
> -	if (pud && pud_present(*pud) && !pud_is_leaf(*pud))
> +	if (pud && pud_present(*pud) && !pud_leaf(*pud))
>   		pmd = pmd_offset(pud, gpa);
>   	else if (level <= 1)
>   		new_pmd = kvmppc_pmd_alloc();
>   
> -	if (level == 0 && !(pmd && pmd_present(*pmd) && !pmd_is_leaf(*pmd)))
> +	if (level == 0 && !(pmd && pmd_present(*pmd) && !pmd_leaf(*pmd)))
>   		new_ptep = kvmppc_pte_alloc();
>   
>   	/* Check if we might have been invalidated; let the guest retry if so */
> @@ -658,7 +658,7 @@ int kvmppc_create_pte(struct kvm *kvm, pgd_t *pgtable, pte_t pte,
>   		new_pud = NULL;
>   	}
>   	pud = pud_offset(p4d, gpa);
> -	if (pud_is_leaf(*pud)) {
> +	if (pud_leaf(*pud)) {
>   		unsigned long hgpa = gpa & PUD_MASK;
>   
>   		/* Check if we raced and someone else has set the same thing */
> @@ -709,7 +709,7 @@ int kvmppc_create_pte(struct kvm *kvm, pgd_t *pgtable, pte_t pte,
>   		new_pmd = NULL;
>   	}
>   	pmd = pmd_offset(pud, gpa);
> -	if (pmd_is_leaf(*pmd)) {
> +	if (pmd_leaf(*pmd)) {
>   		unsigned long lgpa = gpa & PMD_MASK;
>   
>   		/* Check if we raced and someone else has set the same thing */
> diff --git a/arch/powerpc/mm/book3s64/radix_pgtable.c b/arch/powerpc/mm/book3s64/radix_pgtable.c
> index c6a4ac766b2b..1f8db10693e3 100644
> --- a/arch/powerpc/mm/book3s64/radix_pgtable.c
> +++ b/arch/powerpc/mm/book3s64/radix_pgtable.c
> @@ -204,14 +204,14 @@ static void radix__change_memory_range(unsigned long start, unsigned long end,
>   		pudp = pud_alloc(&init_mm, p4dp, idx);
>   		if (!pudp)
>   			continue;
> -		if (pud_is_leaf(*pudp)) {
> +		if (pud_leaf(*pudp)) {
>   			ptep = (pte_t *)pudp;
>   			goto update_the_pte;
>   		}
>   		pmdp = pmd_alloc(&init_mm, pudp, idx);
>   		if (!pmdp)
>   			continue;
> -		if (pmd_is_leaf(*pmdp)) {
> +		if (pmd_leaf(*pmdp)) {
>   			ptep = pmdp_ptep(pmdp);
>   			goto update_the_pte;
>   		}
> @@ -767,7 +767,7 @@ static void __meminit remove_pmd_table(pmd_t *pmd_start, unsigned long addr,
>   		if (!pmd_present(*pmd))
>   			continue;
>   
> -		if (pmd_is_leaf(*pmd)) {
> +		if (pmd_leaf(*pmd)) {
>   			if (IS_ALIGNED(addr, PMD_SIZE) &&
>   			    IS_ALIGNED(next, PMD_SIZE)) {
>   				if (!direct)
> @@ -807,7 +807,7 @@ static void __meminit remove_pud_table(pud_t *pud_start, unsigned long addr,
>   		if (!pud_present(*pud))
>   			continue;
>   
> -		if (pud_is_leaf(*pud)) {
> +		if (pud_leaf(*pud)) {
>   			if (!IS_ALIGNED(addr, PUD_SIZE) ||
>   			    !IS_ALIGNED(next, PUD_SIZE)) {
>   				WARN_ONCE(1, "%s: unaligned range\n", __func__);
> @@ -845,7 +845,7 @@ remove_pagetable(unsigned long start, unsigned long end, bool direct,
>   		if (!p4d_present(*p4d))
>   			continue;
>   
> -		if (p4d_is_leaf(*p4d)) {
> +		if (p4d_leaf(*p4d)) {
>   			if (!IS_ALIGNED(addr, P4D_SIZE) ||
>   			    !IS_ALIGNED(next, P4D_SIZE)) {
>   				WARN_ONCE(1, "%s: unaligned range\n", __func__);
> @@ -1554,7 +1554,7 @@ int pud_set_huge(pud_t *pud, phys_addr_t addr, pgprot_t prot)
>   
>   int pud_clear_huge(pud_t *pud)
>   {
> -	if (pud_is_leaf(*pud)) {
> +	if (pud_leaf(*pud)) {
>   		pud_clear(pud);
>   		return 1;
>   	}
> @@ -1601,7 +1601,7 @@ int pmd_set_huge(pmd_t *pmd, phys_addr_t addr, pgprot_t prot)
>   
>   int pmd_clear_huge(pmd_t *pmd)
>   {
> -	if (pmd_is_leaf(*pmd)) {
> +	if (pmd_leaf(*pmd)) {
>   		pmd_clear(pmd);
>   		return 1;
>   	}
> diff --git a/arch/powerpc/mm/pgtable.c b/arch/powerpc/mm/pgtable.c
> index 549a440ed7f6..9e7ba9c3851f 100644
> --- a/arch/powerpc/mm/pgtable.c
> +++ b/arch/powerpc/mm/pgtable.c
> @@ -410,7 +410,7 @@ pte_t *__find_linux_pte(pgd_t *pgdir, unsigned long ea,
>   	if (p4d_none(p4d))
>   		return NULL;
>   
> -	if (p4d_is_leaf(p4d)) {
> +	if (p4d_leaf(p4d)) {
>   		ret_pte = (pte_t *)p4dp;
>   		goto out;
>   	}
> @@ -432,7 +432,7 @@ pte_t *__find_linux_pte(pgd_t *pgdir, unsigned long ea,
>   	if (pud_none(pud))
>   		return NULL;
>   
> -	if (pud_is_leaf(pud)) {
> +	if (pud_leaf(pud)) {
>   		ret_pte = (pte_t *)pudp;
>   		goto out;
>   	}
> @@ -471,7 +471,7 @@ pte_t *__find_linux_pte(pgd_t *pgdir, unsigned long ea,
>   		goto out;
>   	}
>   
> -	if (pmd_is_leaf(pmd)) {
> +	if (pmd_leaf(pmd)) {
>   		ret_pte = (pte_t *)pmdp;
>   		goto out;
>   	}
> diff --git a/arch/powerpc/mm/pgtable_64.c b/arch/powerpc/mm/pgtable_64.c
> index 1b366526f4f2..386c6b06eab7 100644
> --- a/arch/powerpc/mm/pgtable_64.c
> +++ b/arch/powerpc/mm/pgtable_64.c
> @@ -100,7 +100,7 @@ EXPORT_SYMBOL(__pte_frag_size_shift);
>   /* 4 level page table */
>   struct page *p4d_page(p4d_t p4d)
>   {
> -	if (p4d_is_leaf(p4d)) {
> +	if (p4d_leaf(p4d)) {
>   		if (!IS_ENABLED(CONFIG_HAVE_ARCH_HUGE_VMAP))
>   			VM_WARN_ON(!p4d_huge(p4d));
>   		return pte_page(p4d_pte(p4d));
> @@ -111,7 +111,7 @@ struct page *p4d_page(p4d_t p4d)
>   
>   struct page *pud_page(pud_t pud)
>   {
> -	if (pud_is_leaf(pud)) {
> +	if (pud_leaf(pud)) {
>   		if (!IS_ENABLED(CONFIG_HAVE_ARCH_HUGE_VMAP))
>   			VM_WARN_ON(!pud_huge(pud));
>   		return pte_page(pud_pte(pud));
> @@ -125,7 +125,7 @@ struct page *pud_page(pud_t pud)
>    */
>   struct page *pmd_page(pmd_t pmd)
>   {
> -	if (pmd_is_leaf(pmd)) {
> +	if (pmd_leaf(pmd)) {
>   		/*
>   		 * vmalloc_to_page may be called on any vmap address (not only
>   		 * vmalloc), and it uses pmd_page() etc., when huge vmap is
> diff --git a/arch/powerpc/xmon/xmon.c b/arch/powerpc/xmon/xmon.c
> index b3b94cd37713..9669c9925225 100644
> --- a/arch/powerpc/xmon/xmon.c
> +++ b/arch/powerpc/xmon/xmon.c
> @@ -3342,7 +3342,7 @@ static void show_pte(unsigned long addr)
>   		return;
>   	}
>   
> -	if (p4d_is_leaf(*p4dp)) {
> +	if (p4d_leaf(*p4dp)) {
>   		format_pte(p4dp, p4d_val(*p4dp));
>   		return;
>   	}
> @@ -3356,7 +3356,7 @@ static void show_pte(unsigned long addr)
>   		return;
>   	}
>   
> -	if (pud_is_leaf(*pudp)) {
> +	if (pud_leaf(*pudp)) {
>   		format_pte(pudp, pud_val(*pudp));
>   		return;
>   	}
> @@ -3370,7 +3370,7 @@ static void show_pte(unsigned long addr)
>   		return;
>   	}
>   
> -	if (pmd_is_leaf(*pmdp)) {
> +	if (pmd_leaf(*pmdp)) {
>   		format_pte(pmdp, pmd_val(*pmdp));
>   		return;
>   	}
Mike Rapoport March 6, 2024, 6:20 a.m. UTC | #2
On Tue, Mar 05, 2024 at 12:37:42PM +0800, peterx@redhat.com wrote:
> From: Peter Xu <peterx@redhat.com>
> 
> They're the same macros underneath.  Drop pXd_is_leaf(), instead always use
> pXd_leaf().
> 
> At the meantime, instead of renames, drop the pXd_is_leaf() fallback
> definitions directly in arch/powerpc/include/asm/pgtable.h. because similar
> fallback macros for pXd_leaf() are already defined in
> include/linux/pgtable.h.
> 
> Cc: Michael Ellerman <mpe@ellerman.id.au>
> Cc: Nicholas Piggin <npiggin@gmail.com>
> Cc: "Aneesh Kumar K.V" <aneesh.kumar@kernel.org>
> Cc: "Naveen N. Rao" <naveen.n.rao@linux.ibm.com>
> Cc: linuxppc-dev@lists.ozlabs.org
> Suggested-by: Christophe Leroy <christophe.leroy@csgroup.eu>
> Reviewed-by: Jason Gunthorpe <jgg@nvidia.com>
> Signed-off-by: Peter Xu <peterx@redhat.com>

Reviewed-by: Mike Rapoport (IBM) <rppt@kernel.org>

> ---
>  arch/powerpc/include/asm/book3s/64/pgtable.h | 10 ++++----
>  arch/powerpc/include/asm/pgtable.h           | 24 --------------------
>  arch/powerpc/kvm/book3s_64_mmu_radix.c       | 12 +++++-----
>  arch/powerpc/mm/book3s64/radix_pgtable.c     | 14 ++++++------
>  arch/powerpc/mm/pgtable.c                    |  6 ++---
>  arch/powerpc/mm/pgtable_64.c                 |  6 ++---
>  arch/powerpc/xmon/xmon.c                     |  6 ++---
>  7 files changed, 26 insertions(+), 52 deletions(-)
> 
> diff --git a/arch/powerpc/include/asm/book3s/64/pgtable.h b/arch/powerpc/include/asm/book3s/64/pgtable.h
> index d1318e8582ac..3e99e409774a 100644
> --- a/arch/powerpc/include/asm/book3s/64/pgtable.h
> +++ b/arch/powerpc/include/asm/book3s/64/pgtable.h
> @@ -1439,18 +1439,16 @@ static inline bool is_pte_rw_upgrade(unsigned long old_val, unsigned long new_va
>  /*
>   * Like pmd_huge() and pmd_large(), but works regardless of config options
>   */
> -#define pmd_is_leaf pmd_is_leaf
> -#define pmd_leaf pmd_is_leaf
> +#define pmd_leaf pmd_leaf
>  #define pmd_large pmd_leaf
> -static inline bool pmd_is_leaf(pmd_t pmd)
> +static inline bool pmd_leaf(pmd_t pmd)
>  {
>  	return !!(pmd_raw(pmd) & cpu_to_be64(_PAGE_PTE));
>  }
>  
> -#define pud_is_leaf pud_is_leaf
> -#define pud_leaf pud_is_leaf
> +#define pud_leaf pud_leaf
>  #define pud_large pud_leaf
> -static inline bool pud_is_leaf(pud_t pud)
> +static inline bool pud_leaf(pud_t pud)
>  {
>  	return !!(pud_raw(pud) & cpu_to_be64(_PAGE_PTE));
>  }
> diff --git a/arch/powerpc/include/asm/pgtable.h b/arch/powerpc/include/asm/pgtable.h
> index 5928b3c1458d..e6edf1cdbc5b 100644
> --- a/arch/powerpc/include/asm/pgtable.h
> +++ b/arch/powerpc/include/asm/pgtable.h
> @@ -182,30 +182,6 @@ static inline void pte_frag_set(mm_context_t *ctx, void *p)
>  }
>  #endif
>  
> -#ifndef pmd_is_leaf
> -#define pmd_is_leaf pmd_is_leaf
> -static inline bool pmd_is_leaf(pmd_t pmd)
> -{
> -	return false;
> -}
> -#endif
> -
> -#ifndef pud_is_leaf
> -#define pud_is_leaf pud_is_leaf
> -static inline bool pud_is_leaf(pud_t pud)
> -{
> -	return false;
> -}
> -#endif
> -
> -#ifndef p4d_is_leaf
> -#define p4d_is_leaf p4d_is_leaf
> -static inline bool p4d_is_leaf(p4d_t p4d)
> -{
> -	return false;
> -}
> -#endif
> -
>  #define pmd_pgtable pmd_pgtable
>  static inline pgtable_t pmd_pgtable(pmd_t pmd)
>  {
> diff --git a/arch/powerpc/kvm/book3s_64_mmu_radix.c b/arch/powerpc/kvm/book3s_64_mmu_radix.c
> index 4a1abb9f7c05..408d98f8a514 100644
> --- a/arch/powerpc/kvm/book3s_64_mmu_radix.c
> +++ b/arch/powerpc/kvm/book3s_64_mmu_radix.c
> @@ -503,7 +503,7 @@ static void kvmppc_unmap_free_pmd(struct kvm *kvm, pmd_t *pmd, bool full,
>  	for (im = 0; im < PTRS_PER_PMD; ++im, ++p) {
>  		if (!pmd_present(*p))
>  			continue;
> -		if (pmd_is_leaf(*p)) {
> +		if (pmd_leaf(*p)) {
>  			if (full) {
>  				pmd_clear(p);
>  			} else {
> @@ -532,7 +532,7 @@ static void kvmppc_unmap_free_pud(struct kvm *kvm, pud_t *pud,
>  	for (iu = 0; iu < PTRS_PER_PUD; ++iu, ++p) {
>  		if (!pud_present(*p))
>  			continue;
> -		if (pud_is_leaf(*p)) {
> +		if (pud_leaf(*p)) {
>  			pud_clear(p);
>  		} else {
>  			pmd_t *pmd;
> @@ -635,12 +635,12 @@ int kvmppc_create_pte(struct kvm *kvm, pgd_t *pgtable, pte_t pte,
>  		new_pud = pud_alloc_one(kvm->mm, gpa);
>  
>  	pmd = NULL;
> -	if (pud && pud_present(*pud) && !pud_is_leaf(*pud))
> +	if (pud && pud_present(*pud) && !pud_leaf(*pud))
>  		pmd = pmd_offset(pud, gpa);
>  	else if (level <= 1)
>  		new_pmd = kvmppc_pmd_alloc();
>  
> -	if (level == 0 && !(pmd && pmd_present(*pmd) && !pmd_is_leaf(*pmd)))
> +	if (level == 0 && !(pmd && pmd_present(*pmd) && !pmd_leaf(*pmd)))
>  		new_ptep = kvmppc_pte_alloc();
>  
>  	/* Check if we might have been invalidated; let the guest retry if so */
> @@ -658,7 +658,7 @@ int kvmppc_create_pte(struct kvm *kvm, pgd_t *pgtable, pte_t pte,
>  		new_pud = NULL;
>  	}
>  	pud = pud_offset(p4d, gpa);
> -	if (pud_is_leaf(*pud)) {
> +	if (pud_leaf(*pud)) {
>  		unsigned long hgpa = gpa & PUD_MASK;
>  
>  		/* Check if we raced and someone else has set the same thing */
> @@ -709,7 +709,7 @@ int kvmppc_create_pte(struct kvm *kvm, pgd_t *pgtable, pte_t pte,
>  		new_pmd = NULL;
>  	}
>  	pmd = pmd_offset(pud, gpa);
> -	if (pmd_is_leaf(*pmd)) {
> +	if (pmd_leaf(*pmd)) {
>  		unsigned long lgpa = gpa & PMD_MASK;
>  
>  		/* Check if we raced and someone else has set the same thing */
> diff --git a/arch/powerpc/mm/book3s64/radix_pgtable.c b/arch/powerpc/mm/book3s64/radix_pgtable.c
> index c6a4ac766b2b..1f8db10693e3 100644
> --- a/arch/powerpc/mm/book3s64/radix_pgtable.c
> +++ b/arch/powerpc/mm/book3s64/radix_pgtable.c
> @@ -204,14 +204,14 @@ static void radix__change_memory_range(unsigned long start, unsigned long end,
>  		pudp = pud_alloc(&init_mm, p4dp, idx);
>  		if (!pudp)
>  			continue;
> -		if (pud_is_leaf(*pudp)) {
> +		if (pud_leaf(*pudp)) {
>  			ptep = (pte_t *)pudp;
>  			goto update_the_pte;
>  		}
>  		pmdp = pmd_alloc(&init_mm, pudp, idx);
>  		if (!pmdp)
>  			continue;
> -		if (pmd_is_leaf(*pmdp)) {
> +		if (pmd_leaf(*pmdp)) {
>  			ptep = pmdp_ptep(pmdp);
>  			goto update_the_pte;
>  		}
> @@ -767,7 +767,7 @@ static void __meminit remove_pmd_table(pmd_t *pmd_start, unsigned long addr,
>  		if (!pmd_present(*pmd))
>  			continue;
>  
> -		if (pmd_is_leaf(*pmd)) {
> +		if (pmd_leaf(*pmd)) {
>  			if (IS_ALIGNED(addr, PMD_SIZE) &&
>  			    IS_ALIGNED(next, PMD_SIZE)) {
>  				if (!direct)
> @@ -807,7 +807,7 @@ static void __meminit remove_pud_table(pud_t *pud_start, unsigned long addr,
>  		if (!pud_present(*pud))
>  			continue;
>  
> -		if (pud_is_leaf(*pud)) {
> +		if (pud_leaf(*pud)) {
>  			if (!IS_ALIGNED(addr, PUD_SIZE) ||
>  			    !IS_ALIGNED(next, PUD_SIZE)) {
>  				WARN_ONCE(1, "%s: unaligned range\n", __func__);
> @@ -845,7 +845,7 @@ remove_pagetable(unsigned long start, unsigned long end, bool direct,
>  		if (!p4d_present(*p4d))
>  			continue;
>  
> -		if (p4d_is_leaf(*p4d)) {
> +		if (p4d_leaf(*p4d)) {
>  			if (!IS_ALIGNED(addr, P4D_SIZE) ||
>  			    !IS_ALIGNED(next, P4D_SIZE)) {
>  				WARN_ONCE(1, "%s: unaligned range\n", __func__);
> @@ -1554,7 +1554,7 @@ int pud_set_huge(pud_t *pud, phys_addr_t addr, pgprot_t prot)
>  
>  int pud_clear_huge(pud_t *pud)
>  {
> -	if (pud_is_leaf(*pud)) {
> +	if (pud_leaf(*pud)) {
>  		pud_clear(pud);
>  		return 1;
>  	}
> @@ -1601,7 +1601,7 @@ int pmd_set_huge(pmd_t *pmd, phys_addr_t addr, pgprot_t prot)
>  
>  int pmd_clear_huge(pmd_t *pmd)
>  {
> -	if (pmd_is_leaf(*pmd)) {
> +	if (pmd_leaf(*pmd)) {
>  		pmd_clear(pmd);
>  		return 1;
>  	}
> diff --git a/arch/powerpc/mm/pgtable.c b/arch/powerpc/mm/pgtable.c
> index 549a440ed7f6..9e7ba9c3851f 100644
> --- a/arch/powerpc/mm/pgtable.c
> +++ b/arch/powerpc/mm/pgtable.c
> @@ -410,7 +410,7 @@ pte_t *__find_linux_pte(pgd_t *pgdir, unsigned long ea,
>  	if (p4d_none(p4d))
>  		return NULL;
>  
> -	if (p4d_is_leaf(p4d)) {
> +	if (p4d_leaf(p4d)) {
>  		ret_pte = (pte_t *)p4dp;
>  		goto out;
>  	}
> @@ -432,7 +432,7 @@ pte_t *__find_linux_pte(pgd_t *pgdir, unsigned long ea,
>  	if (pud_none(pud))
>  		return NULL;
>  
> -	if (pud_is_leaf(pud)) {
> +	if (pud_leaf(pud)) {
>  		ret_pte = (pte_t *)pudp;
>  		goto out;
>  	}
> @@ -471,7 +471,7 @@ pte_t *__find_linux_pte(pgd_t *pgdir, unsigned long ea,
>  		goto out;
>  	}
>  
> -	if (pmd_is_leaf(pmd)) {
> +	if (pmd_leaf(pmd)) {
>  		ret_pte = (pte_t *)pmdp;
>  		goto out;
>  	}
> diff --git a/arch/powerpc/mm/pgtable_64.c b/arch/powerpc/mm/pgtable_64.c
> index 1b366526f4f2..386c6b06eab7 100644
> --- a/arch/powerpc/mm/pgtable_64.c
> +++ b/arch/powerpc/mm/pgtable_64.c
> @@ -100,7 +100,7 @@ EXPORT_SYMBOL(__pte_frag_size_shift);
>  /* 4 level page table */
>  struct page *p4d_page(p4d_t p4d)
>  {
> -	if (p4d_is_leaf(p4d)) {
> +	if (p4d_leaf(p4d)) {
>  		if (!IS_ENABLED(CONFIG_HAVE_ARCH_HUGE_VMAP))
>  			VM_WARN_ON(!p4d_huge(p4d));
>  		return pte_page(p4d_pte(p4d));
> @@ -111,7 +111,7 @@ struct page *p4d_page(p4d_t p4d)
>  
>  struct page *pud_page(pud_t pud)
>  {
> -	if (pud_is_leaf(pud)) {
> +	if (pud_leaf(pud)) {
>  		if (!IS_ENABLED(CONFIG_HAVE_ARCH_HUGE_VMAP))
>  			VM_WARN_ON(!pud_huge(pud));
>  		return pte_page(pud_pte(pud));
> @@ -125,7 +125,7 @@ struct page *pud_page(pud_t pud)
>   */
>  struct page *pmd_page(pmd_t pmd)
>  {
> -	if (pmd_is_leaf(pmd)) {
> +	if (pmd_leaf(pmd)) {
>  		/*
>  		 * vmalloc_to_page may be called on any vmap address (not only
>  		 * vmalloc), and it uses pmd_page() etc., when huge vmap is
> diff --git a/arch/powerpc/xmon/xmon.c b/arch/powerpc/xmon/xmon.c
> index b3b94cd37713..9669c9925225 100644
> --- a/arch/powerpc/xmon/xmon.c
> +++ b/arch/powerpc/xmon/xmon.c
> @@ -3342,7 +3342,7 @@ static void show_pte(unsigned long addr)
>  		return;
>  	}
>  
> -	if (p4d_is_leaf(*p4dp)) {
> +	if (p4d_leaf(*p4dp)) {
>  		format_pte(p4dp, p4d_val(*p4dp));
>  		return;
>  	}
> @@ -3356,7 +3356,7 @@ static void show_pte(unsigned long addr)
>  		return;
>  	}
>  
> -	if (pud_is_leaf(*pudp)) {
> +	if (pud_leaf(*pudp)) {
>  		format_pte(pudp, pud_val(*pudp));
>  		return;
>  	}
> @@ -3370,7 +3370,7 @@ static void show_pte(unsigned long addr)
>  		return;
>  	}
>  
> -	if (pmd_is_leaf(*pmdp)) {
> +	if (pmd_leaf(*pmdp)) {
>  		format_pte(pmdp, pmd_val(*pmdp));
>  		return;
>  	}
> -- 
> 2.44.0
> 
>
diff mbox series

Patch

diff --git a/arch/powerpc/include/asm/book3s/64/pgtable.h b/arch/powerpc/include/asm/book3s/64/pgtable.h
index d1318e8582ac..3e99e409774a 100644
--- a/arch/powerpc/include/asm/book3s/64/pgtable.h
+++ b/arch/powerpc/include/asm/book3s/64/pgtable.h
@@ -1439,18 +1439,16 @@  static inline bool is_pte_rw_upgrade(unsigned long old_val, unsigned long new_va
 /*
  * Like pmd_huge() and pmd_large(), but works regardless of config options
  */
-#define pmd_is_leaf pmd_is_leaf
-#define pmd_leaf pmd_is_leaf
+#define pmd_leaf pmd_leaf
 #define pmd_large pmd_leaf
-static inline bool pmd_is_leaf(pmd_t pmd)
+static inline bool pmd_leaf(pmd_t pmd)
 {
 	return !!(pmd_raw(pmd) & cpu_to_be64(_PAGE_PTE));
 }
 
-#define pud_is_leaf pud_is_leaf
-#define pud_leaf pud_is_leaf
+#define pud_leaf pud_leaf
 #define pud_large pud_leaf
-static inline bool pud_is_leaf(pud_t pud)
+static inline bool pud_leaf(pud_t pud)
 {
 	return !!(pud_raw(pud) & cpu_to_be64(_PAGE_PTE));
 }
diff --git a/arch/powerpc/include/asm/pgtable.h b/arch/powerpc/include/asm/pgtable.h
index 5928b3c1458d..e6edf1cdbc5b 100644
--- a/arch/powerpc/include/asm/pgtable.h
+++ b/arch/powerpc/include/asm/pgtable.h
@@ -182,30 +182,6 @@  static inline void pte_frag_set(mm_context_t *ctx, void *p)
 }
 #endif
 
-#ifndef pmd_is_leaf
-#define pmd_is_leaf pmd_is_leaf
-static inline bool pmd_is_leaf(pmd_t pmd)
-{
-	return false;
-}
-#endif
-
-#ifndef pud_is_leaf
-#define pud_is_leaf pud_is_leaf
-static inline bool pud_is_leaf(pud_t pud)
-{
-	return false;
-}
-#endif
-
-#ifndef p4d_is_leaf
-#define p4d_is_leaf p4d_is_leaf
-static inline bool p4d_is_leaf(p4d_t p4d)
-{
-	return false;
-}
-#endif
-
 #define pmd_pgtable pmd_pgtable
 static inline pgtable_t pmd_pgtable(pmd_t pmd)
 {
diff --git a/arch/powerpc/kvm/book3s_64_mmu_radix.c b/arch/powerpc/kvm/book3s_64_mmu_radix.c
index 4a1abb9f7c05..408d98f8a514 100644
--- a/arch/powerpc/kvm/book3s_64_mmu_radix.c
+++ b/arch/powerpc/kvm/book3s_64_mmu_radix.c
@@ -503,7 +503,7 @@  static void kvmppc_unmap_free_pmd(struct kvm *kvm, pmd_t *pmd, bool full,
 	for (im = 0; im < PTRS_PER_PMD; ++im, ++p) {
 		if (!pmd_present(*p))
 			continue;
-		if (pmd_is_leaf(*p)) {
+		if (pmd_leaf(*p)) {
 			if (full) {
 				pmd_clear(p);
 			} else {
@@ -532,7 +532,7 @@  static void kvmppc_unmap_free_pud(struct kvm *kvm, pud_t *pud,
 	for (iu = 0; iu < PTRS_PER_PUD; ++iu, ++p) {
 		if (!pud_present(*p))
 			continue;
-		if (pud_is_leaf(*p)) {
+		if (pud_leaf(*p)) {
 			pud_clear(p);
 		} else {
 			pmd_t *pmd;
@@ -635,12 +635,12 @@  int kvmppc_create_pte(struct kvm *kvm, pgd_t *pgtable, pte_t pte,
 		new_pud = pud_alloc_one(kvm->mm, gpa);
 
 	pmd = NULL;
-	if (pud && pud_present(*pud) && !pud_is_leaf(*pud))
+	if (pud && pud_present(*pud) && !pud_leaf(*pud))
 		pmd = pmd_offset(pud, gpa);
 	else if (level <= 1)
 		new_pmd = kvmppc_pmd_alloc();
 
-	if (level == 0 && !(pmd && pmd_present(*pmd) && !pmd_is_leaf(*pmd)))
+	if (level == 0 && !(pmd && pmd_present(*pmd) && !pmd_leaf(*pmd)))
 		new_ptep = kvmppc_pte_alloc();
 
 	/* Check if we might have been invalidated; let the guest retry if so */
@@ -658,7 +658,7 @@  int kvmppc_create_pte(struct kvm *kvm, pgd_t *pgtable, pte_t pte,
 		new_pud = NULL;
 	}
 	pud = pud_offset(p4d, gpa);
-	if (pud_is_leaf(*pud)) {
+	if (pud_leaf(*pud)) {
 		unsigned long hgpa = gpa & PUD_MASK;
 
 		/* Check if we raced and someone else has set the same thing */
@@ -709,7 +709,7 @@  int kvmppc_create_pte(struct kvm *kvm, pgd_t *pgtable, pte_t pte,
 		new_pmd = NULL;
 	}
 	pmd = pmd_offset(pud, gpa);
-	if (pmd_is_leaf(*pmd)) {
+	if (pmd_leaf(*pmd)) {
 		unsigned long lgpa = gpa & PMD_MASK;
 
 		/* Check if we raced and someone else has set the same thing */
diff --git a/arch/powerpc/mm/book3s64/radix_pgtable.c b/arch/powerpc/mm/book3s64/radix_pgtable.c
index c6a4ac766b2b..1f8db10693e3 100644
--- a/arch/powerpc/mm/book3s64/radix_pgtable.c
+++ b/arch/powerpc/mm/book3s64/radix_pgtable.c
@@ -204,14 +204,14 @@  static void radix__change_memory_range(unsigned long start, unsigned long end,
 		pudp = pud_alloc(&init_mm, p4dp, idx);
 		if (!pudp)
 			continue;
-		if (pud_is_leaf(*pudp)) {
+		if (pud_leaf(*pudp)) {
 			ptep = (pte_t *)pudp;
 			goto update_the_pte;
 		}
 		pmdp = pmd_alloc(&init_mm, pudp, idx);
 		if (!pmdp)
 			continue;
-		if (pmd_is_leaf(*pmdp)) {
+		if (pmd_leaf(*pmdp)) {
 			ptep = pmdp_ptep(pmdp);
 			goto update_the_pte;
 		}
@@ -767,7 +767,7 @@  static void __meminit remove_pmd_table(pmd_t *pmd_start, unsigned long addr,
 		if (!pmd_present(*pmd))
 			continue;
 
-		if (pmd_is_leaf(*pmd)) {
+		if (pmd_leaf(*pmd)) {
 			if (IS_ALIGNED(addr, PMD_SIZE) &&
 			    IS_ALIGNED(next, PMD_SIZE)) {
 				if (!direct)
@@ -807,7 +807,7 @@  static void __meminit remove_pud_table(pud_t *pud_start, unsigned long addr,
 		if (!pud_present(*pud))
 			continue;
 
-		if (pud_is_leaf(*pud)) {
+		if (pud_leaf(*pud)) {
 			if (!IS_ALIGNED(addr, PUD_SIZE) ||
 			    !IS_ALIGNED(next, PUD_SIZE)) {
 				WARN_ONCE(1, "%s: unaligned range\n", __func__);
@@ -845,7 +845,7 @@  remove_pagetable(unsigned long start, unsigned long end, bool direct,
 		if (!p4d_present(*p4d))
 			continue;
 
-		if (p4d_is_leaf(*p4d)) {
+		if (p4d_leaf(*p4d)) {
 			if (!IS_ALIGNED(addr, P4D_SIZE) ||
 			    !IS_ALIGNED(next, P4D_SIZE)) {
 				WARN_ONCE(1, "%s: unaligned range\n", __func__);
@@ -1554,7 +1554,7 @@  int pud_set_huge(pud_t *pud, phys_addr_t addr, pgprot_t prot)
 
 int pud_clear_huge(pud_t *pud)
 {
-	if (pud_is_leaf(*pud)) {
+	if (pud_leaf(*pud)) {
 		pud_clear(pud);
 		return 1;
 	}
@@ -1601,7 +1601,7 @@  int pmd_set_huge(pmd_t *pmd, phys_addr_t addr, pgprot_t prot)
 
 int pmd_clear_huge(pmd_t *pmd)
 {
-	if (pmd_is_leaf(*pmd)) {
+	if (pmd_leaf(*pmd)) {
 		pmd_clear(pmd);
 		return 1;
 	}
diff --git a/arch/powerpc/mm/pgtable.c b/arch/powerpc/mm/pgtable.c
index 549a440ed7f6..9e7ba9c3851f 100644
--- a/arch/powerpc/mm/pgtable.c
+++ b/arch/powerpc/mm/pgtable.c
@@ -410,7 +410,7 @@  pte_t *__find_linux_pte(pgd_t *pgdir, unsigned long ea,
 	if (p4d_none(p4d))
 		return NULL;
 
-	if (p4d_is_leaf(p4d)) {
+	if (p4d_leaf(p4d)) {
 		ret_pte = (pte_t *)p4dp;
 		goto out;
 	}
@@ -432,7 +432,7 @@  pte_t *__find_linux_pte(pgd_t *pgdir, unsigned long ea,
 	if (pud_none(pud))
 		return NULL;
 
-	if (pud_is_leaf(pud)) {
+	if (pud_leaf(pud)) {
 		ret_pte = (pte_t *)pudp;
 		goto out;
 	}
@@ -471,7 +471,7 @@  pte_t *__find_linux_pte(pgd_t *pgdir, unsigned long ea,
 		goto out;
 	}
 
-	if (pmd_is_leaf(pmd)) {
+	if (pmd_leaf(pmd)) {
 		ret_pte = (pte_t *)pmdp;
 		goto out;
 	}
diff --git a/arch/powerpc/mm/pgtable_64.c b/arch/powerpc/mm/pgtable_64.c
index 1b366526f4f2..386c6b06eab7 100644
--- a/arch/powerpc/mm/pgtable_64.c
+++ b/arch/powerpc/mm/pgtable_64.c
@@ -100,7 +100,7 @@  EXPORT_SYMBOL(__pte_frag_size_shift);
 /* 4 level page table */
 struct page *p4d_page(p4d_t p4d)
 {
-	if (p4d_is_leaf(p4d)) {
+	if (p4d_leaf(p4d)) {
 		if (!IS_ENABLED(CONFIG_HAVE_ARCH_HUGE_VMAP))
 			VM_WARN_ON(!p4d_huge(p4d));
 		return pte_page(p4d_pte(p4d));
@@ -111,7 +111,7 @@  struct page *p4d_page(p4d_t p4d)
 
 struct page *pud_page(pud_t pud)
 {
-	if (pud_is_leaf(pud)) {
+	if (pud_leaf(pud)) {
 		if (!IS_ENABLED(CONFIG_HAVE_ARCH_HUGE_VMAP))
 			VM_WARN_ON(!pud_huge(pud));
 		return pte_page(pud_pte(pud));
@@ -125,7 +125,7 @@  struct page *pud_page(pud_t pud)
  */
 struct page *pmd_page(pmd_t pmd)
 {
-	if (pmd_is_leaf(pmd)) {
+	if (pmd_leaf(pmd)) {
 		/*
 		 * vmalloc_to_page may be called on any vmap address (not only
 		 * vmalloc), and it uses pmd_page() etc., when huge vmap is
diff --git a/arch/powerpc/xmon/xmon.c b/arch/powerpc/xmon/xmon.c
index b3b94cd37713..9669c9925225 100644
--- a/arch/powerpc/xmon/xmon.c
+++ b/arch/powerpc/xmon/xmon.c
@@ -3342,7 +3342,7 @@  static void show_pte(unsigned long addr)
 		return;
 	}
 
-	if (p4d_is_leaf(*p4dp)) {
+	if (p4d_leaf(*p4dp)) {
 		format_pte(p4dp, p4d_val(*p4dp));
 		return;
 	}
@@ -3356,7 +3356,7 @@  static void show_pte(unsigned long addr)
 		return;
 	}
 
-	if (pud_is_leaf(*pudp)) {
+	if (pud_leaf(*pudp)) {
 		format_pte(pudp, pud_val(*pudp));
 		return;
 	}
@@ -3370,7 +3370,7 @@  static void show_pte(unsigned long addr)
 		return;
 	}
 
-	if (pmd_is_leaf(*pmdp)) {
+	if (pmd_leaf(*pmdp)) {
 		format_pte(pmdp, pmd_val(*pmdp));
 		return;
 	}