4ad0a4c234
- Add HOTPLUG_SMT support (/sys/devices/system/cpu/smt) and honour the configured SMT state when hotplugging CPUs into the system. - Combine final TLB flush and lazy TLB mm shootdown IPIs when using the Radix MMU to avoid a broadcast TLBIE flush on exit. - Drop the exclusion between ptrace/perf watchpoints, and drop the now unused associated arch hooks. - Add support for the "nohlt" command line option to disable CPU idle. - Add support for -fpatchable-function-entry for ftrace, with GCC >= 13.1. - Rework memory block size determination, and support 256MB size on systems with GPUs that have hotpluggable memory. - Various other small features and fixes. Thanks to: Andrew Donnellan, Aneesh Kumar K.V, Arnd Bergmann, Athira Rajeev, Benjamin Gray, Christophe Leroy, Frederic Barrat, Gautam Menghani, Geoff Levand, Hari Bathini, Immad Mir, Jialin Zhang, Joel Stanley, Jordan Niethe, Justin Stitt, Kajol Jain, Kees Cook, Krzysztof Kozlowski, Laurent Dufour, Liang He, Linus Walleij, Mahesh Salgaonkar, Masahiro Yamada, Michal Suchanek, Nageswara R Sastry, Nathan Chancellor, Nathan Lynch, Naveen N Rao, Nicholas Piggin, Nick Desaulniers, Omar Sandoval, Randy Dunlap, Reza Arbab, Rob Herring, Russell Currey, Sourabh Jain, Thomas Gleixner, Trevor Woerner, Uwe Kleine-König, Vaibhav Jain, Xiongfeng Wang, Yuan Tan, Zhang Rui, Zheng Zengkai. -----BEGIN PGP SIGNATURE----- iQJHBAABCAAxFiEEJFGtCPCthwEv2Y/bUevqPMjhpYAFAmTwgbwTHG1wZUBlbGxl cm1hbi5pZC5hdQAKCRBR6+o8yOGlgFmpD/432vipeoqvkAYsyK0xi/Y3GcY0wcyd WJApLXXadEbtKQrgXQ6sowWqalg5thYnQCRarg/tXKK/po3KfgwkPjGDpOL+cIdr 12QVN2XJm9VmJ1wYJxzk+yXx4F43AdmMdr94qWAGufbTHezwb4UpzVR1NxtFrOE/ X5TNsC2+2mdZY/ZaNHS5vsTIFv3EhQfqgjZPlIAdLn6CGc8xWT514Q/uHA8+ytM/ HL7Hqs33DoPSvgTa5TT/2E0d0k5nO3P5KObzAjpYlireTPaBi51mpKGewcrtm0o2 v3cBlbfx3C7pe9ZhKBK9BH8cjynfiqsVZ9/lCw/7eBNdm9tHuzG0jeS7Db9tCZXS fM7G2R7SoIusPTqxlBmkU5DpYslwrHiVgCyy3ijxkoA/fakVwh/GgTcMsRt73IY6 n6DsUvWwuYHCIeIiHmHQJqCqCRtV+aMzU3AbbBHOjtdIanhlW16M686dEsgCirh7 akRVRD5VqKaqXs34PpkRL89Xv3wZRjl6XZ3hZFfCjSYXfpXDXhgSToIskpHYhKL8 gpY7WtG9YQP05Xz5HRCx6EluaZVeKe0lZi6fezX7Mi9AygJQO8FfXqP1mHBlEq40 ThWtvL9D89RV6lADqqFN20XepgvKNOyAXcE4szvsnIZYUSPmZQZSPxx+DHtROaLP jX3ifxtxJp92pQ== =5g7K -----END PGP SIGNATURE----- Merge tag 'powerpc-6.6-1' of git://git.kernel.org/pub/scm/linux/kernel/git/powerpc/linux Pull powerpc updates from Michael Ellerman: - Add HOTPLUG_SMT support (/sys/devices/system/cpu/smt) and honour the configured SMT state when hotplugging CPUs into the system - Combine final TLB flush and lazy TLB mm shootdown IPIs when using the Radix MMU to avoid a broadcast TLBIE flush on exit - Drop the exclusion between ptrace/perf watchpoints, and drop the now unused associated arch hooks - Add support for the "nohlt" command line option to disable CPU idle - Add support for -fpatchable-function-entry for ftrace, with GCC >= 13.1 - Rework memory block size determination, and support 256MB size on systems with GPUs that have hotpluggable memory - Various other small features and fixes Thanks to Andrew Donnellan, Aneesh Kumar K.V, Arnd Bergmann, Athira Rajeev, Benjamin Gray, Christophe Leroy, Frederic Barrat, Gautam Menghani, Geoff Levand, Hari Bathini, Immad Mir, Jialin Zhang, Joel Stanley, Jordan Niethe, Justin Stitt, Kajol Jain, Kees Cook, Krzysztof Kozlowski, Laurent Dufour, Liang He, Linus Walleij, Mahesh Salgaonkar, Masahiro Yamada, Michal Suchanek, Nageswara R Sastry, Nathan Chancellor, Nathan Lynch, Naveen N Rao, Nicholas Piggin, Nick Desaulniers, Omar Sandoval, Randy Dunlap, Reza Arbab, Rob Herring, Russell Currey, Sourabh Jain, Thomas Gleixner, Trevor Woerner, Uwe Kleine-König, Vaibhav Jain, Xiongfeng Wang, Yuan Tan, Zhang Rui, and Zheng Zengkai. * tag 'powerpc-6.6-1' of git://git.kernel.org/pub/scm/linux/kernel/git/powerpc/linux: (135 commits) macintosh/ams: linux/platform_device.h is needed powerpc/xmon: Reapply "Relax frame size for clang" powerpc/mm/book3s64: Use 256M as the upper limit with coherent device memory attached powerpc/mm/book3s64: Fix build error with SPARSEMEM disabled powerpc/iommu: Fix notifiers being shared by PCI and VIO buses powerpc/mpc5xxx: Add missing fwnode_handle_put() powerpc/config: Disable SLAB_DEBUG_ON in skiroot powerpc/pseries: Remove unused hcall tracing instruction powerpc/pseries: Fix hcall tracepoints with JUMP_LABEL=n powerpc: dts: add missing space before { powerpc/eeh: Use pci_dev_id() to simplify the code powerpc/64s: Move CPU -mtune options into Kconfig powerpc/powermac: Fix unused function warning powerpc/pseries: Rework lppaca_shared_proc() to avoid DEBUG_PREEMPT powerpc: Don't include lppaca.h in paca.h powerpc/pseries: Move hcall_vphn() prototype into vphn.h powerpc/pseries: Move VPHN constants into vphn.h cxl: Drop unused detach_spa() powerpc: Drop zalloc_maybe_bootmem() powerpc/powernv: Use struct opal_prd_msg in more places ...
454 lines
13 KiB
C
454 lines
13 KiB
C
/* SPDX-License-Identifier: GPL-2.0-or-later */
|
|
#ifndef _ASM_POWERPC_PROCESSOR_H
|
|
#define _ASM_POWERPC_PROCESSOR_H
|
|
|
|
/*
|
|
* Copyright (C) 2001 PPC 64 Team, IBM Corp
|
|
*/
|
|
|
|
#include <vdso/processor.h>
|
|
|
|
#include <asm/reg.h>
|
|
|
|
#ifdef CONFIG_VSX
|
|
#define TS_FPRWIDTH 2
|
|
|
|
#ifdef __BIG_ENDIAN__
|
|
#define TS_FPROFFSET 0
|
|
#define TS_VSRLOWOFFSET 1
|
|
#else
|
|
#define TS_FPROFFSET 1
|
|
#define TS_VSRLOWOFFSET 0
|
|
#endif
|
|
|
|
#else
|
|
#define TS_FPRWIDTH 1
|
|
#define TS_FPROFFSET 0
|
|
#endif
|
|
|
|
#ifdef CONFIG_PPC64
|
|
/* Default SMT priority is set to 3. Use 11- 13bits to save priority. */
|
|
#define PPR_PRIORITY 3
|
|
#ifdef __ASSEMBLY__
|
|
#define DEFAULT_PPR (PPR_PRIORITY << 50)
|
|
#else
|
|
#define DEFAULT_PPR ((u64)PPR_PRIORITY << 50)
|
|
#endif /* __ASSEMBLY__ */
|
|
#endif /* CONFIG_PPC64 */
|
|
|
|
#ifndef __ASSEMBLY__
|
|
#include <linux/types.h>
|
|
#include <linux/thread_info.h>
|
|
#include <asm/ptrace.h>
|
|
#include <asm/hw_breakpoint.h>
|
|
|
|
/* We do _not_ want to define new machine types at all, those must die
|
|
* in favor of using the device-tree
|
|
* -- BenH.
|
|
*/
|
|
|
|
/* PREP sub-platform types. Unused */
|
|
#define _PREP_Motorola 0x01 /* motorola prep */
|
|
#define _PREP_Firm 0x02 /* firmworks prep */
|
|
#define _PREP_IBM 0x00 /* ibm prep */
|
|
#define _PREP_Bull 0x03 /* bull prep */
|
|
|
|
/* CHRP sub-platform types. These are arbitrary */
|
|
#define _CHRP_Motorola 0x04 /* motorola chrp, the cobra */
|
|
#define _CHRP_IBM 0x05 /* IBM chrp, the longtrail and longtrail 2 */
|
|
#define _CHRP_Pegasos 0x06 /* Genesi/bplan's Pegasos and Pegasos2 */
|
|
#define _CHRP_briq 0x07 /* TotalImpact's briQ */
|
|
|
|
#if defined(__KERNEL__) && defined(CONFIG_PPC32)
|
|
|
|
extern int _chrp_type;
|
|
|
|
#endif /* defined(__KERNEL__) && defined(CONFIG_PPC32) */
|
|
|
|
#ifdef __KERNEL__
|
|
|
|
#ifdef CONFIG_PPC64
|
|
#include <asm/task_size_64.h>
|
|
#else
|
|
#include <asm/task_size_32.h>
|
|
#endif
|
|
|
|
struct task_struct;
|
|
void start_thread(struct pt_regs *regs, unsigned long fdptr, unsigned long sp);
|
|
|
|
#define TS_FPR(i) fp_state.fpr[i][TS_FPROFFSET]
|
|
#define TS_CKFPR(i) ckfp_state.fpr[i][TS_FPROFFSET]
|
|
|
|
/* FP and VSX 0-31 register set */
|
|
struct thread_fp_state {
|
|
u64 fpr[32][TS_FPRWIDTH] __attribute__((aligned(16)));
|
|
u64 fpscr; /* Floating point status */
|
|
};
|
|
|
|
/* Complete AltiVec register set including VSCR */
|
|
struct thread_vr_state {
|
|
vector128 vr[32] __attribute__((aligned(16)));
|
|
vector128 vscr __attribute__((aligned(16)));
|
|
};
|
|
|
|
struct debug_reg {
|
|
#ifdef CONFIG_PPC_ADV_DEBUG_REGS
|
|
/*
|
|
* The following help to manage the use of Debug Control Registers
|
|
* om the BookE platforms.
|
|
*/
|
|
uint32_t dbcr0;
|
|
uint32_t dbcr1;
|
|
#ifdef CONFIG_BOOKE
|
|
uint32_t dbcr2;
|
|
#endif
|
|
/*
|
|
* The stored value of the DBSR register will be the value at the
|
|
* last debug interrupt. This register can only be read from the
|
|
* user (will never be written to) and has value while helping to
|
|
* describe the reason for the last debug trap. Torez
|
|
*/
|
|
uint32_t dbsr;
|
|
/*
|
|
* The following will contain addresses used by debug applications
|
|
* to help trace and trap on particular address locations.
|
|
* The bits in the Debug Control Registers above help define which
|
|
* of the following registers will contain valid data and/or addresses.
|
|
*/
|
|
unsigned long iac1;
|
|
unsigned long iac2;
|
|
#if CONFIG_PPC_ADV_DEBUG_IACS > 2
|
|
unsigned long iac3;
|
|
unsigned long iac4;
|
|
#endif
|
|
unsigned long dac1;
|
|
unsigned long dac2;
|
|
#if CONFIG_PPC_ADV_DEBUG_DVCS > 0
|
|
unsigned long dvc1;
|
|
unsigned long dvc2;
|
|
#endif
|
|
#endif
|
|
};
|
|
|
|
struct thread_struct {
|
|
unsigned long ksp; /* Kernel stack pointer */
|
|
|
|
#ifdef CONFIG_PPC64
|
|
unsigned long ksp_vsid;
|
|
#endif
|
|
struct pt_regs *regs; /* Pointer to saved register state */
|
|
#ifdef CONFIG_BOOKE
|
|
/* BookE base exception scratch space; align on cacheline */
|
|
unsigned long normsave[8] ____cacheline_aligned;
|
|
#endif
|
|
#ifdef CONFIG_PPC32
|
|
void *pgdir; /* root of page-table tree */
|
|
#ifdef CONFIG_PPC_RTAS
|
|
unsigned long rtas_sp; /* stack pointer for when in RTAS */
|
|
#endif
|
|
#if defined(CONFIG_PPC_BOOK3S_32) && defined(CONFIG_PPC_KUAP)
|
|
unsigned long kuap; /* opened segments for user access */
|
|
#endif
|
|
unsigned long srr0;
|
|
unsigned long srr1;
|
|
unsigned long dar;
|
|
unsigned long dsisr;
|
|
#ifdef CONFIG_PPC_BOOK3S_32
|
|
unsigned long r0, r3, r4, r5, r6, r8, r9, r11;
|
|
unsigned long lr, ctr;
|
|
unsigned long sr0;
|
|
#endif
|
|
#endif /* CONFIG_PPC32 */
|
|
#if defined(CONFIG_BOOKE_OR_40x) && defined(CONFIG_PPC_KUAP)
|
|
unsigned long pid; /* value written in PID reg. at interrupt exit */
|
|
#endif
|
|
/* Debug Registers */
|
|
struct debug_reg debug;
|
|
#ifdef CONFIG_PPC_FPU_REGS
|
|
struct thread_fp_state fp_state;
|
|
struct thread_fp_state *fp_save_area;
|
|
#endif
|
|
int fpexc_mode; /* floating-point exception mode */
|
|
unsigned int align_ctl; /* alignment handling control */
|
|
#ifdef CONFIG_HAVE_HW_BREAKPOINT
|
|
struct perf_event *ptrace_bps[HBP_NUM_MAX];
|
|
#endif /* CONFIG_HAVE_HW_BREAKPOINT */
|
|
struct arch_hw_breakpoint hw_brk[HBP_NUM_MAX]; /* hardware breakpoint info */
|
|
unsigned long trap_nr; /* last trap # on this thread */
|
|
u8 load_slb; /* Ages out SLB preload cache entries */
|
|
u8 load_fp;
|
|
#ifdef CONFIG_ALTIVEC
|
|
u8 load_vec;
|
|
struct thread_vr_state vr_state;
|
|
struct thread_vr_state *vr_save_area;
|
|
unsigned long vrsave;
|
|
int used_vr; /* set if process has used altivec */
|
|
#endif /* CONFIG_ALTIVEC */
|
|
#ifdef CONFIG_VSX
|
|
/* VSR status */
|
|
int used_vsr; /* set if process has used VSX */
|
|
#endif /* CONFIG_VSX */
|
|
#ifdef CONFIG_SPE
|
|
struct_group(spe,
|
|
unsigned long evr[32]; /* upper 32-bits of SPE regs */
|
|
u64 acc; /* Accumulator */
|
|
);
|
|
unsigned long spefscr; /* SPE & eFP status */
|
|
unsigned long spefscr_last; /* SPEFSCR value on last prctl
|
|
call or trap return */
|
|
int used_spe; /* set if process has used spe */
|
|
#endif /* CONFIG_SPE */
|
|
#ifdef CONFIG_PPC_TRANSACTIONAL_MEM
|
|
u8 load_tm;
|
|
u64 tm_tfhar; /* Transaction fail handler addr */
|
|
u64 tm_texasr; /* Transaction exception & summary */
|
|
u64 tm_tfiar; /* Transaction fail instr address reg */
|
|
struct pt_regs ckpt_regs; /* Checkpointed registers */
|
|
|
|
unsigned long tm_tar;
|
|
unsigned long tm_ppr;
|
|
unsigned long tm_dscr;
|
|
unsigned long tm_amr;
|
|
|
|
/*
|
|
* Checkpointed FP and VSX 0-31 register set.
|
|
*
|
|
* When a transaction is active/signalled/scheduled etc., *regs is the
|
|
* most recent set of/speculated GPRs with ckpt_regs being the older
|
|
* checkpointed regs to which we roll back if transaction aborts.
|
|
*
|
|
* These are analogous to how ckpt_regs and pt_regs work
|
|
*/
|
|
struct thread_fp_state ckfp_state; /* Checkpointed FP state */
|
|
struct thread_vr_state ckvr_state; /* Checkpointed VR state */
|
|
unsigned long ckvrsave; /* Checkpointed VRSAVE */
|
|
#endif /* CONFIG_PPC_TRANSACTIONAL_MEM */
|
|
#ifdef CONFIG_KVM_BOOK3S_32_HANDLER
|
|
void* kvm_shadow_vcpu; /* KVM internal data */
|
|
#endif /* CONFIG_KVM_BOOK3S_32_HANDLER */
|
|
#if defined(CONFIG_KVM) && defined(CONFIG_BOOKE)
|
|
struct kvm_vcpu *kvm_vcpu;
|
|
#endif
|
|
#ifdef CONFIG_PPC64
|
|
unsigned long dscr;
|
|
unsigned long fscr;
|
|
/*
|
|
* This member element dscr_inherit indicates that the process
|
|
* has explicitly attempted and changed the DSCR register value
|
|
* for itself. Hence kernel wont use the default CPU DSCR value
|
|
* contained in the PACA structure anymore during process context
|
|
* switch. Once this variable is set, this behaviour will also be
|
|
* inherited to all the children of this process from that point
|
|
* onwards.
|
|
*/
|
|
int dscr_inherit;
|
|
unsigned long tidr;
|
|
#endif
|
|
#ifdef CONFIG_PPC_BOOK3S_64
|
|
unsigned long tar;
|
|
unsigned long ebbrr;
|
|
unsigned long ebbhr;
|
|
unsigned long bescr;
|
|
unsigned long siar;
|
|
unsigned long sdar;
|
|
unsigned long sier;
|
|
unsigned long mmcr2;
|
|
unsigned mmcr0;
|
|
|
|
unsigned used_ebb;
|
|
unsigned long mmcr3;
|
|
unsigned long sier2;
|
|
unsigned long sier3;
|
|
unsigned long hashkeyr;
|
|
|
|
#endif
|
|
};
|
|
|
|
#define ARCH_MIN_TASKALIGN 16
|
|
|
|
#define INIT_SP (sizeof(init_stack) + (unsigned long) &init_stack)
|
|
#define INIT_SP_LIMIT ((unsigned long)&init_stack)
|
|
|
|
#ifdef CONFIG_SPE
|
|
#define SPEFSCR_INIT \
|
|
.spefscr = SPEFSCR_FINVE | SPEFSCR_FDBZE | SPEFSCR_FUNFE | SPEFSCR_FOVFE, \
|
|
.spefscr_last = SPEFSCR_FINVE | SPEFSCR_FDBZE | SPEFSCR_FUNFE | SPEFSCR_FOVFE,
|
|
#else
|
|
#define SPEFSCR_INIT
|
|
#endif
|
|
|
|
#ifdef CONFIG_PPC_BOOK3S_32
|
|
#define SR0_INIT .sr0 = IS_ENABLED(CONFIG_PPC_KUEP) ? SR_NX : 0,
|
|
#else
|
|
#define SR0_INIT
|
|
#endif
|
|
|
|
#if defined(CONFIG_PPC_BOOK3S_32) && defined(CONFIG_PPC_KUAP)
|
|
#define INIT_THREAD { \
|
|
.ksp = INIT_SP, \
|
|
.pgdir = swapper_pg_dir, \
|
|
.kuap = ~0UL, /* KUAP_NONE */ \
|
|
.fpexc_mode = MSR_FE0 | MSR_FE1, \
|
|
SPEFSCR_INIT \
|
|
SR0_INIT \
|
|
}
|
|
#elif defined(CONFIG_PPC32)
|
|
#define INIT_THREAD { \
|
|
.ksp = INIT_SP, \
|
|
.pgdir = swapper_pg_dir, \
|
|
.fpexc_mode = MSR_FE0 | MSR_FE1, \
|
|
SPEFSCR_INIT \
|
|
SR0_INIT \
|
|
}
|
|
#else
|
|
#define INIT_THREAD { \
|
|
.ksp = INIT_SP, \
|
|
.fpexc_mode = 0, \
|
|
}
|
|
#endif
|
|
|
|
#define task_pt_regs(tsk) ((tsk)->thread.regs)
|
|
|
|
unsigned long __get_wchan(struct task_struct *p);
|
|
|
|
#define KSTK_EIP(tsk) ((tsk)->thread.regs? (tsk)->thread.regs->nip: 0)
|
|
#define KSTK_ESP(tsk) ((tsk)->thread.regs? (tsk)->thread.regs->gpr[1]: 0)
|
|
|
|
/* Get/set floating-point exception mode */
|
|
#define GET_FPEXC_CTL(tsk, adr) get_fpexc_mode((tsk), (adr))
|
|
#define SET_FPEXC_CTL(tsk, val) set_fpexc_mode((tsk), (val))
|
|
|
|
extern int get_fpexc_mode(struct task_struct *tsk, unsigned long adr);
|
|
extern int set_fpexc_mode(struct task_struct *tsk, unsigned int val);
|
|
|
|
#define GET_ENDIAN(tsk, adr) get_endian((tsk), (adr))
|
|
#define SET_ENDIAN(tsk, val) set_endian((tsk), (val))
|
|
|
|
extern int get_endian(struct task_struct *tsk, unsigned long adr);
|
|
extern int set_endian(struct task_struct *tsk, unsigned int val);
|
|
|
|
#define GET_UNALIGN_CTL(tsk, adr) get_unalign_ctl((tsk), (adr))
|
|
#define SET_UNALIGN_CTL(tsk, val) set_unalign_ctl((tsk), (val))
|
|
|
|
extern int get_unalign_ctl(struct task_struct *tsk, unsigned long adr);
|
|
extern int set_unalign_ctl(struct task_struct *tsk, unsigned int val);
|
|
|
|
extern void load_fp_state(struct thread_fp_state *fp);
|
|
extern void store_fp_state(struct thread_fp_state *fp);
|
|
extern void load_vr_state(struct thread_vr_state *vr);
|
|
extern void store_vr_state(struct thread_vr_state *vr);
|
|
|
|
static inline unsigned int __unpack_fe01(unsigned long msr_bits)
|
|
{
|
|
return ((msr_bits & MSR_FE0) >> 10) | ((msr_bits & MSR_FE1) >> 8);
|
|
}
|
|
|
|
static inline unsigned long __pack_fe01(unsigned int fpmode)
|
|
{
|
|
return ((fpmode << 10) & MSR_FE0) | ((fpmode << 8) & MSR_FE1);
|
|
}
|
|
|
|
#ifdef CONFIG_PPC64
|
|
|
|
#define spin_begin() \
|
|
asm volatile(ASM_FTR_IFCLR( \
|
|
"or 1,1,1", /* HMT_LOW */ \
|
|
"nop", /* v3.1 uses pause_short in cpu_relax instead */ \
|
|
%0) :: "i" (CPU_FTR_ARCH_31) : "memory")
|
|
|
|
#define spin_cpu_relax() \
|
|
asm volatile(ASM_FTR_IFCLR( \
|
|
"nop", /* Before v3.1 use priority nops in spin_begin/end */ \
|
|
PPC_WAIT(2, 0), /* aka pause_short */ \
|
|
%0) :: "i" (CPU_FTR_ARCH_31) : "memory")
|
|
|
|
#define spin_end() \
|
|
asm volatile(ASM_FTR_IFCLR( \
|
|
"or 2,2,2", /* HMT_MEDIUM */ \
|
|
"nop", \
|
|
%0) :: "i" (CPU_FTR_ARCH_31) : "memory")
|
|
|
|
#endif
|
|
|
|
/*
|
|
* Check that a certain kernel stack pointer is a valid (minimum sized)
|
|
* stack frame in task_struct p.
|
|
*/
|
|
int validate_sp(unsigned long sp, struct task_struct *p);
|
|
|
|
/*
|
|
* validate the stack frame of a particular minimum size, used for when we are
|
|
* looking at a certain object in the stack beyond the minimum.
|
|
*/
|
|
int validate_sp_size(unsigned long sp, struct task_struct *p,
|
|
unsigned long nbytes);
|
|
|
|
/*
|
|
* Prefetch macros.
|
|
*/
|
|
#define ARCH_HAS_PREFETCH
|
|
#define ARCH_HAS_PREFETCHW
|
|
|
|
static inline void prefetch(const void *x)
|
|
{
|
|
if (unlikely(!x))
|
|
return;
|
|
|
|
__asm__ __volatile__ ("dcbt 0,%0" : : "r" (x));
|
|
}
|
|
|
|
static inline void prefetchw(const void *x)
|
|
{
|
|
if (unlikely(!x))
|
|
return;
|
|
|
|
__asm__ __volatile__ ("dcbtst 0,%0" : : "r" (x));
|
|
}
|
|
|
|
/* asm stubs */
|
|
extern unsigned long isa300_idle_stop_noloss(unsigned long psscr_val);
|
|
extern unsigned long isa300_idle_stop_mayloss(unsigned long psscr_val);
|
|
extern unsigned long isa206_idle_insn_mayloss(unsigned long type);
|
|
#ifdef CONFIG_PPC_970_NAP
|
|
extern void power4_idle_nap(void);
|
|
void power4_idle_nap_return(void);
|
|
#endif
|
|
|
|
extern unsigned long cpuidle_disable;
|
|
enum idle_boot_override {IDLE_NO_OVERRIDE = 0, IDLE_POWERSAVE_OFF};
|
|
|
|
extern int powersave_nap; /* set if nap mode can be used in idle loop */
|
|
|
|
extern void power7_idle_type(unsigned long type);
|
|
extern void arch300_idle_type(unsigned long stop_psscr_val,
|
|
unsigned long stop_psscr_mask);
|
|
void pnv_power9_force_smt4_catch(void);
|
|
void pnv_power9_force_smt4_release(void);
|
|
|
|
extern int fix_alignment(struct pt_regs *);
|
|
|
|
#ifdef CONFIG_PPC64
|
|
/*
|
|
* We handle most unaligned accesses in hardware. On the other hand
|
|
* unaligned DMA can be very expensive on some ppc64 IO chips (it does
|
|
* powers of 2 writes until it reaches sufficient alignment).
|
|
*
|
|
* Based on this we disable the IP header alignment in network drivers.
|
|
*/
|
|
#define NET_IP_ALIGN 0
|
|
#endif
|
|
|
|
int do_mathemu(struct pt_regs *regs);
|
|
int do_spe_mathemu(struct pt_regs *regs);
|
|
int speround_handler(struct pt_regs *regs);
|
|
|
|
/* VMX copying */
|
|
int enter_vmx_usercopy(void);
|
|
int exit_vmx_usercopy(void);
|
|
int enter_vmx_ops(void);
|
|
void *exit_vmx_ops(void *dest);
|
|
|
|
#endif /* __KERNEL__ */
|
|
#endif /* __ASSEMBLY__ */
|
|
#endif /* _ASM_POWERPC_PROCESSOR_H */
|