Message ID | 90594004804c6a9b690b69bdf0e5c4d6c880c5f4.1567117050.git.msuchanek@suse.de (mailing list archive) |
---|---|
State | Superseded |
Headers | show |
Series | Disable compat cruft on ppc64le v5 | expand |
Context | Check | Description |
---|---|---|
snowpatch_ozlabs/apply_patch | warning | Failed to apply on branch next (6ed37b3ef6aef8647b8dad925affa92c4c8cc8a3) |
snowpatch_ozlabs/apply_patch | fail | Failed to apply to any branch |
On 08/29/2019 10:28 PM, Michal Suchanek wrote: > There are numerous references to 32bit functions in generic and 64bit > code so ifdef them out. > > Signed-off-by: Michal Suchanek <msuchanek@suse.de> > --- > v2: > - fix 32bit ifdef condition in signal.c > - simplify the compat ifdef condition in vdso.c - 64bit is redundant > - simplify the compat ifdef condition in callchain.c - 64bit is redundant > v3: > - use IS_ENABLED and maybe_unused where possible > - do not ifdef declarations > - clean up Makefile > v4: > - further makefile cleanup > - simplify is_32bit_task conditions > - avoid ifdef in condition by using return > v5: > - avoid unreachable code on 32bit > - make is_current_64bit constant on !COMPAT > - add stub perf_callchain_user_32 to avoid some ifdefs > --- > arch/powerpc/include/asm/thread_info.h | 4 ++-- > arch/powerpc/kernel/Makefile | 7 +++---- > arch/powerpc/kernel/entry_64.S | 2 ++ > arch/powerpc/kernel/signal.c | 3 +-- > arch/powerpc/kernel/syscall_64.c | 6 ++---- > arch/powerpc/kernel/vdso.c | 5 ++--- > arch/powerpc/perf/callchain.c | 13 +++++++++++-- > 7 files changed, 23 insertions(+), 17 deletions(-) > [...] > diff --git a/arch/powerpc/perf/callchain.c b/arch/powerpc/perf/callchain.c > index c84bbd4298a0..881be5c4e9bb 100644 > --- a/arch/powerpc/perf/callchain.c > +++ b/arch/powerpc/perf/callchain.c > @@ -15,7 +15,7 @@ > #include <asm/sigcontext.h> > #include <asm/ucontext.h> > #include <asm/vdso.h> > -#ifdef CONFIG_PPC64 > +#ifdef CONFIG_COMPAT > #include "../kernel/ppc32.h" > #endif > #include <asm/pte-walk.h> > @@ -291,7 +291,8 @@ static inline int current_is_64bit(void) > * interrupt stack, and the thread flags don't get copied over > * from the thread_info on the main stack to the interrupt stack. > */ > - return !test_ti_thread_flag(task_thread_info(current), TIF_32BIT); > + return !IS_ENABLED(CONFIG_COMPAT) || > + !test_ti_thread_flag(task_thread_info(current), TIF_32BIT); > } > > #else /* CONFIG_PPC64 */ > @@ -341,6 +342,7 @@ static inline int valid_user_sp(unsigned long sp, int is_64) > > #endif /* CONFIG_PPC64 */ > > +#if defined(CONFIG_PPC32) || defined(CONFIG_COMPAT) > /* > * Layout for non-RT signal frames > */ > @@ -482,6 +484,13 @@ static void perf_callchain_user_32(struct perf_callchain_entry_ctx *entry, > sp = next_sp; > } > } > +#else /* 32bit */ > +static void perf_callchain_user_32(struct perf_callchain_entry_ctx *entry, > + struct pt_regs *regs) > +{ > + (void)&read_user_stack_32; /* unused if !COMPAT */ That looks pretty much like a hack. See possible alternative below. > +} > +#endif /* 32bit */ > > void > perf_callchain_user(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs) > --- arch/powerpc/perf/callchain.c | 62 +++++++++++++++++++------------------------ 1 file changed, 27 insertions(+), 35 deletions(-) diff --git a/arch/powerpc/perf/callchain.c b/arch/powerpc/perf/callchain.c index 881be5c4e9bb..1b169b32776a 100644 --- a/arch/powerpc/perf/callchain.c +++ b/arch/powerpc/perf/callchain.c @@ -165,22 +165,6 @@ static int read_user_stack_64(unsigned long __user *ptr, unsigned long *ret) return read_user_stack_slow(ptr, ret, 8); } -static int read_user_stack_32(unsigned int __user *ptr, unsigned int *ret) -{ - if ((unsigned long)ptr > TASK_SIZE - sizeof(unsigned int) || - ((unsigned long)ptr & 3)) - return -EFAULT; - - pagefault_disable(); - if (!__get_user_inatomic(*ret, ptr)) { - pagefault_enable(); - return 0; - } - pagefault_enable(); - - return read_user_stack_slow(ptr, ret, 4); -} - static inline int valid_user_sp(unsigned long sp, int is_64) { if (!sp || (sp & 7) || sp > (is_64 ? TASK_SIZE : 0x100000000UL) - 32) @@ -296,25 +280,10 @@ static inline int current_is_64bit(void) } #else /* CONFIG_PPC64 */ -/* - * On 32-bit we just access the address and let hash_page create a - * HPTE if necessary, so there is no need to fall back to reading - * the page tables. Since this is called at interrupt level, - * do_page_fault() won't treat a DSI as a page fault. - */ -static int read_user_stack_32(unsigned int __user *ptr, unsigned int *ret) -{ - int rc; - - if ((unsigned long)ptr > TASK_SIZE - sizeof(unsigned int) || - ((unsigned long)ptr & 3)) - return -EFAULT; - pagefault_disable(); - rc = __get_user_inatomic(*ret, ptr); - pagefault_enable(); - - return rc; +static int read_user_stack_slow(void __user *ptr, void *buf, int nb) +{ + return 0; } static inline void perf_callchain_user_64(struct perf_callchain_entry_ctx *entry, @@ -344,6 +313,30 @@ static inline int valid_user_sp(unsigned long sp, int is_64) #if defined(CONFIG_PPC32) || defined(CONFIG_COMPAT) /* + * On 32-bit we just access the address and let hash_page create a + * HPTE if necessary, so there is no need to fall back to reading + * the page tables. Since this is called at interrupt level, + * do_page_fault() won't treat a DSI as a page fault. + */ +static int read_user_stack_32(unsigned int __user *ptr, unsigned int *ret) +{ + int rc; + + if ((unsigned long)ptr > TASK_SIZE - sizeof(unsigned int) || + ((unsigned long)ptr & 3)) + return -EFAULT; + + pagefault_disable(); + rc = __get_user_inatomic(*ret, ptr); + pagefault_enable(); + + if (IS_ENABLED(CONFIG_PPC32) || !rc) + return rc; + + return read_user_stack_slow(ptr, ret, 4); +} + +/* * Layout for non-RT signal frames */ struct signal_frame_32 { @@ -488,7 +481,6 @@ static void perf_callchain_user_32(struct perf_callchain_entry_ctx *entry, static void perf_callchain_user_32(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs) { - (void)&read_user_stack_32; /* unused if !COMPAT */ } #endif /* 32bit */
On Fri, 30 Aug 2019 06:35:13 +0000 (UTC) Christophe Leroy <christophe.leroy@c-s.fr> wrote: > On 08/29/2019 10:28 PM, Michal Suchanek wrote: > > There are numerous references to 32bit functions in generic and 64bit > > code so ifdef them out. > > > > Signed-off-by: Michal Suchanek <msuchanek@suse.de> > > --- > > v2: > > - fix 32bit ifdef condition in signal.c > > - simplify the compat ifdef condition in vdso.c - 64bit is redundant > > - simplify the compat ifdef condition in callchain.c - 64bit is redundant > > v3: > > - use IS_ENABLED and maybe_unused where possible > > - do not ifdef declarations > > - clean up Makefile > > v4: > > - further makefile cleanup > > - simplify is_32bit_task conditions > > - avoid ifdef in condition by using return > > v5: > > - avoid unreachable code on 32bit > > - make is_current_64bit constant on !COMPAT > > - add stub perf_callchain_user_32 to avoid some ifdefs > > --- > > arch/powerpc/include/asm/thread_info.h | 4 ++-- > > arch/powerpc/kernel/Makefile | 7 +++---- > > arch/powerpc/kernel/entry_64.S | 2 ++ > > arch/powerpc/kernel/signal.c | 3 +-- > > arch/powerpc/kernel/syscall_64.c | 6 ++---- > > arch/powerpc/kernel/vdso.c | 5 ++--- > > arch/powerpc/perf/callchain.c | 13 +++++++++++-- > > 7 files changed, 23 insertions(+), 17 deletions(-) > > > [...] > > > diff --git a/arch/powerpc/perf/callchain.c b/arch/powerpc/perf/callchain.c > > index c84bbd4298a0..881be5c4e9bb 100644 > > --- a/arch/powerpc/perf/callchain.c > > +++ b/arch/powerpc/perf/callchain.c > > @@ -15,7 +15,7 @@ > > #include <asm/sigcontext.h> > > #include <asm/ucontext.h> > > #include <asm/vdso.h> > > -#ifdef CONFIG_PPC64 > > +#ifdef CONFIG_COMPAT > > #include "../kernel/ppc32.h" > > #endif > > #include <asm/pte-walk.h> > > @@ -291,7 +291,8 @@ static inline int current_is_64bit(void) > > * interrupt stack, and the thread flags don't get copied over > > * from the thread_info on the main stack to the interrupt stack. > > */ > > - return !test_ti_thread_flag(task_thread_info(current), TIF_32BIT); > > + return !IS_ENABLED(CONFIG_COMPAT) || > > + !test_ti_thread_flag(task_thread_info(current), TIF_32BIT); > > } > > > > #else /* CONFIG_PPC64 */ > > @@ -341,6 +342,7 @@ static inline int valid_user_sp(unsigned long sp, int is_64) > > > > #endif /* CONFIG_PPC64 */ > > > > +#if defined(CONFIG_PPC32) || defined(CONFIG_COMPAT) > > /* > > * Layout for non-RT signal frames > > */ > > @@ -482,6 +484,13 @@ static void perf_callchain_user_32(struct perf_callchain_entry_ctx *entry, > > sp = next_sp; > > } > > } > > +#else /* 32bit */ > > +static void perf_callchain_user_32(struct perf_callchain_entry_ctx *entry, > > + struct pt_regs *regs) > > +{ > > + (void)&read_user_stack_32; /* unused if !COMPAT */ > > That looks pretty much like a hack. > > See possible alternative below. > > > +} > > +#endif /* 32bit */ > > > > void > > perf_callchain_user(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs) > > > > --- > arch/powerpc/perf/callchain.c | 62 +++++++++++++++++++------------------------ > 1 file changed, 27 insertions(+), 35 deletions(-) > > diff --git a/arch/powerpc/perf/callchain.c b/arch/powerpc/perf/callchain.c > index 881be5c4e9bb..1b169b32776a 100644 > --- a/arch/powerpc/perf/callchain.c > +++ b/arch/powerpc/perf/callchain.c > @@ -165,22 +165,6 @@ static int read_user_stack_64(unsigned long __user *ptr, unsigned long *ret) > return read_user_stack_slow(ptr, ret, 8); > } > > -static int read_user_stack_32(unsigned int __user *ptr, unsigned int *ret) > -{ > - if ((unsigned long)ptr > TASK_SIZE - sizeof(unsigned int) || > - ((unsigned long)ptr & 3)) > - return -EFAULT; > - > - pagefault_disable(); > - if (!__get_user_inatomic(*ret, ptr)) { > - pagefault_enable(); > - return 0; > - } > - pagefault_enable(); > - > - return read_user_stack_slow(ptr, ret, 4); > -} > - > static inline int valid_user_sp(unsigned long sp, int is_64) > { > if (!sp || (sp & 7) || sp > (is_64 ? TASK_SIZE : 0x100000000UL) - 32) > @@ -296,25 +280,10 @@ static inline int current_is_64bit(void) > } > > #else /* CONFIG_PPC64 */ > -/* > - * On 32-bit we just access the address and let hash_page create a > - * HPTE if necessary, so there is no need to fall back to reading > - * the page tables. Since this is called at interrupt level, > - * do_page_fault() won't treat a DSI as a page fault. > - */ > -static int read_user_stack_32(unsigned int __user *ptr, unsigned int *ret) > -{ > - int rc; > - > - if ((unsigned long)ptr > TASK_SIZE - sizeof(unsigned int) || > - ((unsigned long)ptr & 3)) > - return -EFAULT; > > - pagefault_disable(); > - rc = __get_user_inatomic(*ret, ptr); > - pagefault_enable(); > - > - return rc; > +static int read_user_stack_slow(void __user *ptr, void *buf, int nb) > +{ > + return 0; > } > > static inline void perf_callchain_user_64(struct perf_callchain_entry_ctx *entry, > @@ -344,6 +313,30 @@ static inline int valid_user_sp(unsigned long sp, int is_64) > > #if defined(CONFIG_PPC32) || defined(CONFIG_COMPAT) > /* > + * On 32-bit we just access the address and let hash_page create a > + * HPTE if necessary, so there is no need to fall back to reading > + * the page tables. Since this is called at interrupt level, > + * do_page_fault() won't treat a DSI as a page fault. > + */ > +static int read_user_stack_32(unsigned int __user *ptr, unsigned int *ret) > +{ > + int rc; > + > + if ((unsigned long)ptr > TASK_SIZE - sizeof(unsigned int) || > + ((unsigned long)ptr & 3)) > + return -EFAULT; > + > + pagefault_disable(); > + rc = __get_user_inatomic(*ret, ptr); > + pagefault_enable(); > + > + if (IS_ENABLED(CONFIG_PPC32) || !rc) > + return rc; > + > + return read_user_stack_slow(ptr, ret, 4); Which is not declared here. This is not intended to be the final state, anyway. Thanks Michal > +} > + > +/* > * Layout for non-RT signal frames > */ > struct signal_frame_32 { > @@ -488,7 +481,6 @@ static void perf_callchain_user_32(struct perf_callchain_entry_ctx *entry, > static void perf_callchain_user_32(struct perf_callchain_entry_ctx *entry, > struct pt_regs *regs) > { > - (void)&read_user_stack_32; /* unused if !COMPAT */ > } > #endif /* 32bit */ >
Le 30/08/2019 à 09:54, Michal Suchánek a écrit : > On Fri, 30 Aug 2019 06:35:13 +0000 (UTC) > Christophe Leroy <christophe.leroy@c-s.fr> wrote: > >> On 08/29/2019 10:28 PM, Michal Suchanek wrote: >>> There are numerous references to 32bit functions in generic and 64bit >>> code so ifdef them out. >>> >>> Signed-off-by: Michal Suchanek <msuchanek@suse.de> >>> --- >>> v2: >>> - fix 32bit ifdef condition in signal.c >>> - simplify the compat ifdef condition in vdso.c - 64bit is redundant >>> - simplify the compat ifdef condition in callchain.c - 64bit is redundant >>> v3: >>> - use IS_ENABLED and maybe_unused where possible >>> - do not ifdef declarations >>> - clean up Makefile >>> v4: >>> - further makefile cleanup >>> - simplify is_32bit_task conditions >>> - avoid ifdef in condition by using return >>> v5: >>> - avoid unreachable code on 32bit >>> - make is_current_64bit constant on !COMPAT >>> - add stub perf_callchain_user_32 to avoid some ifdefs >>> --- >>> arch/powerpc/include/asm/thread_info.h | 4 ++-- >>> arch/powerpc/kernel/Makefile | 7 +++---- >>> arch/powerpc/kernel/entry_64.S | 2 ++ >>> arch/powerpc/kernel/signal.c | 3 +-- >>> arch/powerpc/kernel/syscall_64.c | 6 ++---- >>> arch/powerpc/kernel/vdso.c | 5 ++--- >>> arch/powerpc/perf/callchain.c | 13 +++++++++++-- >>> 7 files changed, 23 insertions(+), 17 deletions(-) >>> >> [...] >> >>> diff --git a/arch/powerpc/perf/callchain.c b/arch/powerpc/perf/callchain.c >>> index c84bbd4298a0..881be5c4e9bb 100644 >>> --- a/arch/powerpc/perf/callchain.c >>> +++ b/arch/powerpc/perf/callchain.c >>> @@ -15,7 +15,7 @@ >>> #include <asm/sigcontext.h> >>> #include <asm/ucontext.h> >>> #include <asm/vdso.h> >>> -#ifdef CONFIG_PPC64 >>> +#ifdef CONFIG_COMPAT >>> #include "../kernel/ppc32.h" >>> #endif >>> #include <asm/pte-walk.h> >>> @@ -291,7 +291,8 @@ static inline int current_is_64bit(void) >>> * interrupt stack, and the thread flags don't get copied over >>> * from the thread_info on the main stack to the interrupt stack. >>> */ >>> - return !test_ti_thread_flag(task_thread_info(current), TIF_32BIT); >>> + return !IS_ENABLED(CONFIG_COMPAT) || >>> + !test_ti_thread_flag(task_thread_info(current), TIF_32BIT); >>> } >>> >>> #else /* CONFIG_PPC64 */ >>> @@ -341,6 +342,7 @@ static inline int valid_user_sp(unsigned long sp, int is_64) >>> >>> #endif /* CONFIG_PPC64 */ >>> >>> +#if defined(CONFIG_PPC32) || defined(CONFIG_COMPAT) >>> /* >>> * Layout for non-RT signal frames >>> */ >>> @@ -482,6 +484,13 @@ static void perf_callchain_user_32(struct perf_callchain_entry_ctx *entry, >>> sp = next_sp; >>> } >>> } >>> +#else /* 32bit */ >>> +static void perf_callchain_user_32(struct perf_callchain_entry_ctx *entry, >>> + struct pt_regs *regs) >>> +{ >>> + (void)&read_user_stack_32; /* unused if !COMPAT */ >> >> That looks pretty much like a hack. >> >> See possible alternative below. >> >>> +} >>> +#endif /* 32bit */ >>> >>> void >>> perf_callchain_user(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs) >>> >> >> --- >> arch/powerpc/perf/callchain.c | 62 +++++++++++++++++++------------------------ >> 1 file changed, 27 insertions(+), 35 deletions(-) >> >> diff --git a/arch/powerpc/perf/callchain.c b/arch/powerpc/perf/callchain.c >> index 881be5c4e9bb..1b169b32776a 100644 >> --- a/arch/powerpc/perf/callchain.c >> +++ b/arch/powerpc/perf/callchain.c >> @@ -165,22 +165,6 @@ static int read_user_stack_64(unsigned long __user *ptr, unsigned long *ret) >> return read_user_stack_slow(ptr, ret, 8); >> } >> >> -static int read_user_stack_32(unsigned int __user *ptr, unsigned int *ret) >> -{ >> - if ((unsigned long)ptr > TASK_SIZE - sizeof(unsigned int) || >> - ((unsigned long)ptr & 3)) >> - return -EFAULT; >> - >> - pagefault_disable(); >> - if (!__get_user_inatomic(*ret, ptr)) { >> - pagefault_enable(); >> - return 0; >> - } >> - pagefault_enable(); >> - >> - return read_user_stack_slow(ptr, ret, 4); >> -} >> - >> static inline int valid_user_sp(unsigned long sp, int is_64) >> { >> if (!sp || (sp & 7) || sp > (is_64 ? TASK_SIZE : 0x100000000UL) - 32) >> @@ -296,25 +280,10 @@ static inline int current_is_64bit(void) >> } >> >> #else /* CONFIG_PPC64 */ >> -/* >> - * On 32-bit we just access the address and let hash_page create a >> - * HPTE if necessary, so there is no need to fall back to reading >> - * the page tables. Since this is called at interrupt level, >> - * do_page_fault() won't treat a DSI as a page fault. >> - */ >> -static int read_user_stack_32(unsigned int __user *ptr, unsigned int *ret) >> -{ >> - int rc; >> - >> - if ((unsigned long)ptr > TASK_SIZE - sizeof(unsigned int) || >> - ((unsigned long)ptr & 3)) >> - return -EFAULT; >> >> - pagefault_disable(); >> - rc = __get_user_inatomic(*ret, ptr); >> - pagefault_enable(); >> - >> - return rc; >> +static int read_user_stack_slow(void __user *ptr, void *buf, int nb) >> +{ >> + return 0; Here it is >> } >> >> static inline void perf_callchain_user_64(struct perf_callchain_entry_ctx *entry, >> @@ -344,6 +313,30 @@ static inline int valid_user_sp(unsigned long sp, int is_64) >> >> #if defined(CONFIG_PPC32) || defined(CONFIG_COMPAT) >> /* >> + * On 32-bit we just access the address and let hash_page create a >> + * HPTE if necessary, so there is no need to fall back to reading >> + * the page tables. Since this is called at interrupt level, >> + * do_page_fault() won't treat a DSI as a page fault. >> + */ >> +static int read_user_stack_32(unsigned int __user *ptr, unsigned int *ret) >> +{ >> + int rc; >> + >> + if ((unsigned long)ptr > TASK_SIZE - sizeof(unsigned int) || >> + ((unsigned long)ptr & 3)) >> + return -EFAULT; >> + >> + pagefault_disable(); >> + rc = __get_user_inatomic(*ret, ptr); >> + pagefault_enable(); >> + >> + if (IS_ENABLED(CONFIG_PPC32) || !rc) >> + return rc; >> + >> + return read_user_stack_slow(ptr, ret, 4); > Which is not declared here. This is not intended to be the final state, > anyway. Yes it is declared here, see above Christophe > > Thanks > > Michal >> +} >> + >> +/* >> * Layout for non-RT signal frames >> */ >> struct signal_frame_32 { >> @@ -488,7 +481,6 @@ static void perf_callchain_user_32(struct perf_callchain_entry_ctx *entry, >> static void perf_callchain_user_32(struct perf_callchain_entry_ctx *entry, >> struct pt_regs *regs) >> { >> - (void)&read_user_stack_32; /* unused if !COMPAT */ >> } >> #endif /* 32bit */ >>
diff --git a/arch/powerpc/include/asm/thread_info.h b/arch/powerpc/include/asm/thread_info.h index 8e1d0195ac36..c128d8a48ea3 100644 --- a/arch/powerpc/include/asm/thread_info.h +++ b/arch/powerpc/include/asm/thread_info.h @@ -144,10 +144,10 @@ static inline bool test_thread_local_flags(unsigned int flags) return (ti->local_flags & flags) != 0; } -#ifdef CONFIG_PPC64 +#ifdef CONFIG_COMPAT #define is_32bit_task() (test_thread_flag(TIF_32BIT)) #else -#define is_32bit_task() (1) +#define is_32bit_task() (IS_ENABLED(CONFIG_PPC32)) #endif #if defined(CONFIG_PPC64) diff --git a/arch/powerpc/kernel/Makefile b/arch/powerpc/kernel/Makefile index 1d646a94d96c..9d8772e863b9 100644 --- a/arch/powerpc/kernel/Makefile +++ b/arch/powerpc/kernel/Makefile @@ -44,16 +44,15 @@ CFLAGS_btext.o += -DDISABLE_BRANCH_PROFILING endif obj-y := cputable.o ptrace.o syscalls.o \ - irq.o align.o signal_32.o pmc.o vdso.o \ + irq.o align.o signal_$(BITS).o pmc.o vdso.o \ process.o systbl.o idle.o \ signal.o sysfs.o cacheinfo.o time.o \ prom.o traps.o setup-common.o \ udbg.o misc.o io.o misc_$(BITS).o \ of_platform.o prom_parse.o -obj-$(CONFIG_PPC64) += setup_64.o sys_ppc32.o \ - signal_64.o ptrace32.o \ - paca.o nvram_64.o firmware.o \ +obj-$(CONFIG_PPC64) += setup_64.o paca.o nvram_64.o firmware.o \ syscall_64.o +obj-$(CONFIG_COMPAT) += sys_ppc32.o ptrace32.o signal_32.o obj-$(CONFIG_VDSO32) += vdso32/ obj-$(CONFIG_PPC_WATCHDOG) += watchdog.o obj-$(CONFIG_HAVE_HW_BREAKPOINT) += hw_breakpoint.o diff --git a/arch/powerpc/kernel/entry_64.S b/arch/powerpc/kernel/entry_64.S index 2ec825a85f5b..a2dbf216f607 100644 --- a/arch/powerpc/kernel/entry_64.S +++ b/arch/powerpc/kernel/entry_64.S @@ -51,8 +51,10 @@ SYS_CALL_TABLE: .tc sys_call_table[TC],sys_call_table +#ifdef CONFIG_COMPAT COMPAT_SYS_CALL_TABLE: .tc compat_sys_call_table[TC],compat_sys_call_table +#endif /* This value is used to mark exception frames on the stack. */ exception_marker: diff --git a/arch/powerpc/kernel/signal.c b/arch/powerpc/kernel/signal.c index 60436432399f..61678cb0e6a1 100644 --- a/arch/powerpc/kernel/signal.c +++ b/arch/powerpc/kernel/signal.c @@ -247,7 +247,6 @@ static void do_signal(struct task_struct *tsk) sigset_t *oldset = sigmask_to_save(); struct ksignal ksig = { .sig = 0 }; int ret; - int is32 = is_32bit_task(); BUG_ON(tsk != current); @@ -277,7 +276,7 @@ static void do_signal(struct task_struct *tsk) rseq_signal_deliver(&ksig, tsk->thread.regs); - if (is32) { + if (is_32bit_task()) { if (ksig.ka.sa.sa_flags & SA_SIGINFO) ret = handle_rt_signal32(&ksig, oldset, tsk); else diff --git a/arch/powerpc/kernel/syscall_64.c b/arch/powerpc/kernel/syscall_64.c index 98ed970796d5..0d5cbbe54cf1 100644 --- a/arch/powerpc/kernel/syscall_64.c +++ b/arch/powerpc/kernel/syscall_64.c @@ -38,7 +38,6 @@ typedef long (*syscall_fn)(long, long, long, long, long, long); long system_call_exception(long r3, long r4, long r5, long r6, long r7, long r8, unsigned long r0, struct pt_regs *regs) { - unsigned long ti_flags; syscall_fn f; BUG_ON(!(regs->msr & MSR_PR)); @@ -83,8 +82,7 @@ long system_call_exception(long r3, long r4, long r5, long r6, long r7, long r8, */ regs->softe = IRQS_ENABLED; - ti_flags = current_thread_info()->flags; - if (unlikely(ti_flags & _TIF_SYSCALL_DOTRACE)) { + if (unlikely(current_thread_info()->flags & _TIF_SYSCALL_DOTRACE)) { /* * We use the return value of do_syscall_trace_enter() as the * syscall number. If the syscall was rejected for any reason @@ -100,7 +98,7 @@ long system_call_exception(long r3, long r4, long r5, long r6, long r7, long r8, /* May be faster to do array_index_nospec? */ barrier_nospec(); - if (unlikely(ti_flags & _TIF_32BIT)) { + if (unlikely(is_32bit_task())) { f = (void *)compat_sys_call_table[r0]; r3 &= 0x00000000ffffffffULL; diff --git a/arch/powerpc/kernel/vdso.c b/arch/powerpc/kernel/vdso.c index d60598113a9f..6d4a077f74d6 100644 --- a/arch/powerpc/kernel/vdso.c +++ b/arch/powerpc/kernel/vdso.c @@ -667,9 +667,7 @@ static void __init vdso_setup_syscall_map(void) { unsigned int i; extern unsigned long *sys_call_table; -#ifdef CONFIG_PPC64 extern unsigned long *compat_sys_call_table; -#endif extern unsigned long sys_ni_syscall; @@ -678,7 +676,8 @@ static void __init vdso_setup_syscall_map(void) if (sys_call_table[i] != sys_ni_syscall) vdso_data->syscall_map_64[i >> 5] |= 0x80000000UL >> (i & 0x1f); - if (compat_sys_call_table[i] != sys_ni_syscall) + if (IS_ENABLED(CONFIG_COMPAT) && + compat_sys_call_table[i] != sys_ni_syscall) vdso_data->syscall_map_32[i >> 5] |= 0x80000000UL >> (i & 0x1f); #else /* CONFIG_PPC64 */ diff --git a/arch/powerpc/perf/callchain.c b/arch/powerpc/perf/callchain.c index c84bbd4298a0..881be5c4e9bb 100644 --- a/arch/powerpc/perf/callchain.c +++ b/arch/powerpc/perf/callchain.c @@ -15,7 +15,7 @@ #include <asm/sigcontext.h> #include <asm/ucontext.h> #include <asm/vdso.h> -#ifdef CONFIG_PPC64 +#ifdef CONFIG_COMPAT #include "../kernel/ppc32.h" #endif #include <asm/pte-walk.h> @@ -291,7 +291,8 @@ static inline int current_is_64bit(void) * interrupt stack, and the thread flags don't get copied over * from the thread_info on the main stack to the interrupt stack. */ - return !test_ti_thread_flag(task_thread_info(current), TIF_32BIT); + return !IS_ENABLED(CONFIG_COMPAT) || + !test_ti_thread_flag(task_thread_info(current), TIF_32BIT); } #else /* CONFIG_PPC64 */ @@ -341,6 +342,7 @@ static inline int valid_user_sp(unsigned long sp, int is_64) #endif /* CONFIG_PPC64 */ +#if defined(CONFIG_PPC32) || defined(CONFIG_COMPAT) /* * Layout for non-RT signal frames */ @@ -482,6 +484,13 @@ static void perf_callchain_user_32(struct perf_callchain_entry_ctx *entry, sp = next_sp; } } +#else /* 32bit */ +static void perf_callchain_user_32(struct perf_callchain_entry_ctx *entry, + struct pt_regs *regs) +{ + (void)&read_user_stack_32; /* unused if !COMPAT */ +} +#endif /* 32bit */ void perf_callchain_user(struct perf_callchain_entry_ctx *entry, struct pt_regs *regs)
There are numerous references to 32bit functions in generic and 64bit code so ifdef them out. Signed-off-by: Michal Suchanek <msuchanek@suse.de> --- v2: - fix 32bit ifdef condition in signal.c - simplify the compat ifdef condition in vdso.c - 64bit is redundant - simplify the compat ifdef condition in callchain.c - 64bit is redundant v3: - use IS_ENABLED and maybe_unused where possible - do not ifdef declarations - clean up Makefile v4: - further makefile cleanup - simplify is_32bit_task conditions - avoid ifdef in condition by using return v5: - avoid unreachable code on 32bit - make is_current_64bit constant on !COMPAT - add stub perf_callchain_user_32 to avoid some ifdefs --- arch/powerpc/include/asm/thread_info.h | 4 ++-- arch/powerpc/kernel/Makefile | 7 +++---- arch/powerpc/kernel/entry_64.S | 2 ++ arch/powerpc/kernel/signal.c | 3 +-- arch/powerpc/kernel/syscall_64.c | 6 ++---- arch/powerpc/kernel/vdso.c | 5 ++--- arch/powerpc/perf/callchain.c | 13 +++++++++++-- 7 files changed, 23 insertions(+), 17 deletions(-)