2080ff9493
Add kernel_vstate to keep track of kernel-mode Vector registers when trap introduced context switch happens. Also, provide riscv_v_flags to let context save/restore routine track context status. Context tracking happens whenever the core starts its in-kernel Vector executions. An active (dirty) kernel task's V contexts will be saved to memory whenever a trap-introduced context switch happens. Or, when a softirq, which happens to nest on top of it, uses Vector. Context retoring happens when the execution transfer back to the original Kernel context where it first enable preempt_v. Also, provide a config CONFIG_RISCV_ISA_V_PREEMPTIVE to give users an option to disable preemptible kernel-mode Vector at build time. Users with constraint memory may want to disable this config as preemptible kernel-mode Vector needs extra space for tracking of per thread's kernel-mode V context. Or, users might as well want to disable it if all kernel-mode Vector code is time sensitive and cannot tolerate context switch overhead. Signed-off-by: Andy Chiu <andy.chiu@sifive.com> Tested-by: Björn Töpel <bjorn@rivosinc.com> Tested-by: Lad Prabhakar <prabhakar.mahadev-lad.rj@bp.renesas.com> Link: https://lore.kernel.org/r/20240115055929.4736-11-andy.chiu@sifive.com Signed-off-by: Palmer Dabbelt <palmer@rivosinc.com>
248 lines
5.8 KiB
C
248 lines
5.8 KiB
C
// SPDX-License-Identifier: GPL-2.0-or-later
|
|
/*
|
|
* Copyright (C) 2012 ARM Ltd.
|
|
* Author: Catalin Marinas <catalin.marinas@arm.com>
|
|
* Copyright (C) 2017 Linaro Ltd. <ard.biesheuvel@linaro.org>
|
|
* Copyright (C) 2021 SiFive
|
|
*/
|
|
#include <linux/compiler.h>
|
|
#include <linux/irqflags.h>
|
|
#include <linux/percpu.h>
|
|
#include <linux/preempt.h>
|
|
#include <linux/types.h>
|
|
|
|
#include <asm/vector.h>
|
|
#include <asm/switch_to.h>
|
|
#include <asm/simd.h>
|
|
#ifdef CONFIG_RISCV_ISA_V_PREEMPTIVE
|
|
#include <asm/asm-prototypes.h>
|
|
#endif
|
|
|
|
static inline void riscv_v_flags_set(u32 flags)
|
|
{
|
|
WRITE_ONCE(current->thread.riscv_v_flags, flags);
|
|
}
|
|
|
|
static inline void riscv_v_start(u32 flags)
|
|
{
|
|
int orig;
|
|
|
|
orig = riscv_v_flags();
|
|
BUG_ON((orig & flags) != 0);
|
|
riscv_v_flags_set(orig | flags);
|
|
barrier();
|
|
}
|
|
|
|
static inline void riscv_v_stop(u32 flags)
|
|
{
|
|
int orig;
|
|
|
|
barrier();
|
|
orig = riscv_v_flags();
|
|
BUG_ON((orig & flags) == 0);
|
|
riscv_v_flags_set(orig & ~flags);
|
|
}
|
|
|
|
/*
|
|
* Claim ownership of the CPU vector context for use by the calling context.
|
|
*
|
|
* The caller may freely manipulate the vector context metadata until
|
|
* put_cpu_vector_context() is called.
|
|
*/
|
|
void get_cpu_vector_context(void)
|
|
{
|
|
/*
|
|
* disable softirqs so it is impossible for softirqs to nest
|
|
* get_cpu_vector_context() when kernel is actively using Vector.
|
|
*/
|
|
if (!IS_ENABLED(CONFIG_PREEMPT_RT))
|
|
local_bh_disable();
|
|
else
|
|
preempt_disable();
|
|
|
|
riscv_v_start(RISCV_KERNEL_MODE_V);
|
|
}
|
|
|
|
/*
|
|
* Release the CPU vector context.
|
|
*
|
|
* Must be called from a context in which get_cpu_vector_context() was
|
|
* previously called, with no call to put_cpu_vector_context() in the
|
|
* meantime.
|
|
*/
|
|
void put_cpu_vector_context(void)
|
|
{
|
|
riscv_v_stop(RISCV_KERNEL_MODE_V);
|
|
|
|
if (!IS_ENABLED(CONFIG_PREEMPT_RT))
|
|
local_bh_enable();
|
|
else
|
|
preempt_enable();
|
|
}
|
|
|
|
#ifdef CONFIG_RISCV_ISA_V_PREEMPTIVE
|
|
static __always_inline u32 *riscv_v_flags_ptr(void)
|
|
{
|
|
return ¤t->thread.riscv_v_flags;
|
|
}
|
|
|
|
static inline void riscv_preempt_v_set_dirty(void)
|
|
{
|
|
*riscv_v_flags_ptr() |= RISCV_PREEMPT_V_DIRTY;
|
|
}
|
|
|
|
static inline void riscv_preempt_v_reset_flags(void)
|
|
{
|
|
*riscv_v_flags_ptr() &= ~(RISCV_PREEMPT_V_DIRTY | RISCV_PREEMPT_V_NEED_RESTORE);
|
|
}
|
|
|
|
static inline void riscv_v_ctx_depth_inc(void)
|
|
{
|
|
*riscv_v_flags_ptr() += RISCV_V_CTX_UNIT_DEPTH;
|
|
}
|
|
|
|
static inline void riscv_v_ctx_depth_dec(void)
|
|
{
|
|
*riscv_v_flags_ptr() -= RISCV_V_CTX_UNIT_DEPTH;
|
|
}
|
|
|
|
static inline u32 riscv_v_ctx_get_depth(void)
|
|
{
|
|
return *riscv_v_flags_ptr() & RISCV_V_CTX_DEPTH_MASK;
|
|
}
|
|
|
|
static int riscv_v_stop_kernel_context(void)
|
|
{
|
|
if (riscv_v_ctx_get_depth() != 0 || !riscv_preempt_v_started(current))
|
|
return 1;
|
|
|
|
riscv_preempt_v_clear_dirty(current);
|
|
riscv_v_stop(RISCV_PREEMPT_V);
|
|
return 0;
|
|
}
|
|
|
|
static int riscv_v_start_kernel_context(bool *is_nested)
|
|
{
|
|
struct __riscv_v_ext_state *kvstate, *uvstate;
|
|
|
|
kvstate = ¤t->thread.kernel_vstate;
|
|
if (!kvstate->datap)
|
|
return -ENOENT;
|
|
|
|
if (riscv_preempt_v_started(current)) {
|
|
WARN_ON(riscv_v_ctx_get_depth() == 0);
|
|
*is_nested = true;
|
|
get_cpu_vector_context();
|
|
if (riscv_preempt_v_dirty(current)) {
|
|
__riscv_v_vstate_save(kvstate, kvstate->datap);
|
|
riscv_preempt_v_clear_dirty(current);
|
|
}
|
|
riscv_preempt_v_set_restore(current);
|
|
return 0;
|
|
}
|
|
|
|
/* Transfer the ownership of V from user to kernel, then save */
|
|
riscv_v_start(RISCV_PREEMPT_V | RISCV_PREEMPT_V_DIRTY);
|
|
if ((task_pt_regs(current)->status & SR_VS) == SR_VS_DIRTY) {
|
|
uvstate = ¤t->thread.vstate;
|
|
__riscv_v_vstate_save(uvstate, uvstate->datap);
|
|
}
|
|
riscv_preempt_v_clear_dirty(current);
|
|
return 0;
|
|
}
|
|
|
|
/* low-level V context handling code, called with irq disabled */
|
|
asmlinkage void riscv_v_context_nesting_start(struct pt_regs *regs)
|
|
{
|
|
int depth;
|
|
|
|
if (!riscv_preempt_v_started(current))
|
|
return;
|
|
|
|
depth = riscv_v_ctx_get_depth();
|
|
if (depth == 0 && (regs->status & SR_VS) == SR_VS_DIRTY)
|
|
riscv_preempt_v_set_dirty();
|
|
|
|
riscv_v_ctx_depth_inc();
|
|
}
|
|
|
|
asmlinkage void riscv_v_context_nesting_end(struct pt_regs *regs)
|
|
{
|
|
struct __riscv_v_ext_state *vstate = ¤t->thread.kernel_vstate;
|
|
u32 depth;
|
|
|
|
WARN_ON(!irqs_disabled());
|
|
|
|
if (!riscv_preempt_v_started(current))
|
|
return;
|
|
|
|
riscv_v_ctx_depth_dec();
|
|
depth = riscv_v_ctx_get_depth();
|
|
if (depth == 0) {
|
|
if (riscv_preempt_v_restore(current)) {
|
|
__riscv_v_vstate_restore(vstate, vstate->datap);
|
|
__riscv_v_vstate_clean(regs);
|
|
riscv_preempt_v_reset_flags();
|
|
}
|
|
}
|
|
}
|
|
#else
|
|
#define riscv_v_start_kernel_context(nested) (-ENOENT)
|
|
#define riscv_v_stop_kernel_context() (-ENOENT)
|
|
#endif /* CONFIG_RISCV_ISA_V_PREEMPTIVE */
|
|
|
|
/*
|
|
* kernel_vector_begin(): obtain the CPU vector registers for use by the calling
|
|
* context
|
|
*
|
|
* Must not be called unless may_use_simd() returns true.
|
|
* Task context in the vector registers is saved back to memory as necessary.
|
|
*
|
|
* A matching call to kernel_vector_end() must be made before returning from the
|
|
* calling context.
|
|
*
|
|
* The caller may freely use the vector registers until kernel_vector_end() is
|
|
* called.
|
|
*/
|
|
void kernel_vector_begin(void)
|
|
{
|
|
bool nested = false;
|
|
|
|
if (WARN_ON(!has_vector()))
|
|
return;
|
|
|
|
BUG_ON(!may_use_simd());
|
|
|
|
if (riscv_v_start_kernel_context(&nested)) {
|
|
get_cpu_vector_context();
|
|
riscv_v_vstate_save(¤t->thread.vstate, task_pt_regs(current));
|
|
}
|
|
|
|
if (!nested)
|
|
riscv_v_vstate_set_restore(current, task_pt_regs(current));
|
|
|
|
riscv_v_enable();
|
|
}
|
|
EXPORT_SYMBOL_GPL(kernel_vector_begin);
|
|
|
|
/*
|
|
* kernel_vector_end(): give the CPU vector registers back to the current task
|
|
*
|
|
* Must be called from a context in which kernel_vector_begin() was previously
|
|
* called, with no call to kernel_vector_end() in the meantime.
|
|
*
|
|
* The caller must not use the vector registers after this function is called,
|
|
* unless kernel_vector_begin() is called again in the meantime.
|
|
*/
|
|
void kernel_vector_end(void)
|
|
{
|
|
if (WARN_ON(!has_vector()))
|
|
return;
|
|
|
|
riscv_v_disable();
|
|
|
|
if (riscv_v_stop_kernel_context())
|
|
put_cpu_vector_context();
|
|
}
|
|
EXPORT_SYMBOL_GPL(kernel_vector_end);
|