7ce555252c
An iommu domain is allocated in venus_firmware_init() and is attached to core->fw.dev in the same function. Use iommu_paging_domain_alloc() to make it explicit. Signed-off-by: Lu Baolu <baolu.lu@linux.intel.com> Reviewed-by: Jason Gunthorpe <jgg@nvidia.com> Link: https://lore.kernel.org/r/20240610085555.88197-10-baolu.lu@linux.intel.com Signed-off-by: Stanimir Varbanov <stanimir.k.varbanov@gmail.com> Signed-off-by: Hans Verkuil <hverkuil-cisco@xs4all.nl>
364 lines
8.2 KiB
C
364 lines
8.2 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* Copyright (C) 2017 Linaro Ltd.
|
|
*/
|
|
|
|
#include <linux/device.h>
|
|
#include <linux/firmware.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/iommu.h>
|
|
#include <linux/io.h>
|
|
#include <linux/of.h>
|
|
#include <linux/of_address.h>
|
|
#include <linux/of_reserved_mem.h>
|
|
#include <linux/platform_device.h>
|
|
#include <linux/of_device.h>
|
|
#include <linux/firmware/qcom/qcom_scm.h>
|
|
#include <linux/sizes.h>
|
|
#include <linux/soc/qcom/mdt_loader.h>
|
|
|
|
#include "core.h"
|
|
#include "firmware.h"
|
|
#include "hfi_venus_io.h"
|
|
|
|
#define VENUS_PAS_ID 9
|
|
#define VENUS_FW_MEM_SIZE (6 * SZ_1M)
|
|
#define VENUS_FW_START_ADDR 0x0
|
|
|
|
static void venus_reset_cpu(struct venus_core *core)
|
|
{
|
|
u32 fw_size = core->fw.mapped_mem_size;
|
|
void __iomem *wrapper_base;
|
|
|
|
if (IS_IRIS2_1(core))
|
|
wrapper_base = core->wrapper_tz_base;
|
|
else
|
|
wrapper_base = core->wrapper_base;
|
|
|
|
writel(0, wrapper_base + WRAPPER_FW_START_ADDR);
|
|
writel(fw_size, wrapper_base + WRAPPER_FW_END_ADDR);
|
|
writel(0, wrapper_base + WRAPPER_CPA_START_ADDR);
|
|
writel(fw_size, wrapper_base + WRAPPER_CPA_END_ADDR);
|
|
writel(fw_size, wrapper_base + WRAPPER_NONPIX_START_ADDR);
|
|
writel(fw_size, wrapper_base + WRAPPER_NONPIX_END_ADDR);
|
|
|
|
if (IS_IRIS2_1(core)) {
|
|
/* Bring XTSS out of reset */
|
|
writel(0, wrapper_base + WRAPPER_TZ_XTSS_SW_RESET);
|
|
} else {
|
|
writel(0x0, wrapper_base + WRAPPER_CPU_CGC_DIS);
|
|
writel(0x0, wrapper_base + WRAPPER_CPU_CLOCK_CONFIG);
|
|
|
|
/* Bring ARM9 out of reset */
|
|
writel(0, wrapper_base + WRAPPER_A9SS_SW_RESET);
|
|
}
|
|
}
|
|
|
|
int venus_set_hw_state(struct venus_core *core, bool resume)
|
|
{
|
|
int ret;
|
|
|
|
if (core->use_tz) {
|
|
ret = qcom_scm_set_remote_state(resume, 0);
|
|
if (resume && ret == -EINVAL)
|
|
ret = 0;
|
|
return ret;
|
|
}
|
|
|
|
if (resume) {
|
|
venus_reset_cpu(core);
|
|
} else {
|
|
if (IS_IRIS2_1(core))
|
|
writel(WRAPPER_XTSS_SW_RESET_BIT,
|
|
core->wrapper_tz_base + WRAPPER_TZ_XTSS_SW_RESET);
|
|
else
|
|
writel(WRAPPER_A9SS_SW_RESET_BIT,
|
|
core->wrapper_base + WRAPPER_A9SS_SW_RESET);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int venus_load_fw(struct venus_core *core, const char *fwname,
|
|
phys_addr_t *mem_phys, size_t *mem_size)
|
|
{
|
|
const struct firmware *mdt;
|
|
struct reserved_mem *rmem;
|
|
struct device_node *node;
|
|
struct device *dev;
|
|
ssize_t fw_size;
|
|
void *mem_va;
|
|
int ret;
|
|
|
|
*mem_phys = 0;
|
|
*mem_size = 0;
|
|
|
|
dev = core->dev;
|
|
node = of_parse_phandle(dev->of_node, "memory-region", 0);
|
|
if (!node) {
|
|
dev_err(dev, "no memory-region specified\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
rmem = of_reserved_mem_lookup(node);
|
|
of_node_put(node);
|
|
if (!rmem) {
|
|
dev_err(dev, "failed to lookup reserved memory-region\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
ret = request_firmware(&mdt, fwname, dev);
|
|
if (ret < 0)
|
|
return ret;
|
|
|
|
fw_size = qcom_mdt_get_size(mdt);
|
|
if (fw_size < 0) {
|
|
ret = fw_size;
|
|
goto err_release_fw;
|
|
}
|
|
|
|
*mem_phys = rmem->base;
|
|
*mem_size = rmem->size;
|
|
|
|
if (*mem_size < fw_size || fw_size > VENUS_FW_MEM_SIZE) {
|
|
ret = -EINVAL;
|
|
goto err_release_fw;
|
|
}
|
|
|
|
mem_va = memremap(*mem_phys, *mem_size, MEMREMAP_WC);
|
|
if (!mem_va) {
|
|
dev_err(dev, "unable to map memory region %pa size %#zx\n", mem_phys, *mem_size);
|
|
ret = -ENOMEM;
|
|
goto err_release_fw;
|
|
}
|
|
|
|
if (core->use_tz)
|
|
ret = qcom_mdt_load(dev, mdt, fwname, VENUS_PAS_ID,
|
|
mem_va, *mem_phys, *mem_size, NULL);
|
|
else
|
|
ret = qcom_mdt_load_no_init(dev, mdt, fwname, VENUS_PAS_ID,
|
|
mem_va, *mem_phys, *mem_size, NULL);
|
|
|
|
memunmap(mem_va);
|
|
err_release_fw:
|
|
release_firmware(mdt);
|
|
return ret;
|
|
}
|
|
|
|
static int venus_boot_no_tz(struct venus_core *core, phys_addr_t mem_phys,
|
|
size_t mem_size)
|
|
{
|
|
struct iommu_domain *iommu;
|
|
struct device *dev;
|
|
int ret;
|
|
|
|
dev = core->fw.dev;
|
|
if (!dev)
|
|
return -EPROBE_DEFER;
|
|
|
|
iommu = core->fw.iommu_domain;
|
|
core->fw.mapped_mem_size = mem_size;
|
|
|
|
ret = iommu_map(iommu, VENUS_FW_START_ADDR, mem_phys, mem_size,
|
|
IOMMU_READ | IOMMU_WRITE | IOMMU_PRIV, GFP_KERNEL);
|
|
if (ret) {
|
|
dev_err(dev, "could not map video firmware region\n");
|
|
return ret;
|
|
}
|
|
|
|
venus_reset_cpu(core);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int venus_shutdown_no_tz(struct venus_core *core)
|
|
{
|
|
const size_t mapped = core->fw.mapped_mem_size;
|
|
struct iommu_domain *iommu;
|
|
size_t unmapped;
|
|
u32 reg;
|
|
struct device *dev = core->fw.dev;
|
|
void __iomem *wrapper_base = core->wrapper_base;
|
|
void __iomem *wrapper_tz_base = core->wrapper_tz_base;
|
|
|
|
if (IS_IRIS2_1(core)) {
|
|
/* Assert the reset to XTSS */
|
|
reg = readl(wrapper_tz_base + WRAPPER_TZ_XTSS_SW_RESET);
|
|
reg |= WRAPPER_XTSS_SW_RESET_BIT;
|
|
writel(reg, wrapper_tz_base + WRAPPER_TZ_XTSS_SW_RESET);
|
|
} else {
|
|
/* Assert the reset to ARM9 */
|
|
reg = readl(wrapper_base + WRAPPER_A9SS_SW_RESET);
|
|
reg |= WRAPPER_A9SS_SW_RESET_BIT;
|
|
writel(reg, wrapper_base + WRAPPER_A9SS_SW_RESET);
|
|
}
|
|
|
|
iommu = core->fw.iommu_domain;
|
|
|
|
if (core->fw.mapped_mem_size && iommu) {
|
|
unmapped = iommu_unmap(iommu, VENUS_FW_START_ADDR, mapped);
|
|
|
|
if (unmapped != mapped)
|
|
dev_err(dev, "failed to unmap firmware\n");
|
|
else
|
|
core->fw.mapped_mem_size = 0;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int venus_boot(struct venus_core *core)
|
|
{
|
|
struct device *dev = core->dev;
|
|
const struct venus_resources *res = core->res;
|
|
const char *fwpath = NULL;
|
|
phys_addr_t mem_phys;
|
|
size_t mem_size;
|
|
int ret;
|
|
|
|
if (!IS_ENABLED(CONFIG_QCOM_MDT_LOADER) ||
|
|
(core->use_tz && !qcom_scm_is_available()))
|
|
return -EPROBE_DEFER;
|
|
|
|
ret = of_property_read_string_index(dev->of_node, "firmware-name", 0,
|
|
&fwpath);
|
|
if (ret)
|
|
fwpath = core->res->fwname;
|
|
|
|
ret = venus_load_fw(core, fwpath, &mem_phys, &mem_size);
|
|
if (ret) {
|
|
dev_err(dev, "fail to load video firmware\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
core->fw.mem_size = mem_size;
|
|
core->fw.mem_phys = mem_phys;
|
|
|
|
if (core->use_tz)
|
|
ret = qcom_scm_pas_auth_and_reset(VENUS_PAS_ID);
|
|
else
|
|
ret = venus_boot_no_tz(core, mem_phys, mem_size);
|
|
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (core->use_tz && res->cp_size) {
|
|
/*
|
|
* Clues for porting using downstream data:
|
|
* cp_start = 0
|
|
* cp_size = venus_ns/virtual-addr-pool[0] - yes, address and not size!
|
|
* This works, as the non-secure context bank is placed
|
|
* contiguously right after the Content Protection region.
|
|
*
|
|
* cp_nonpixel_start = venus_sec_non_pixel/virtual-addr-pool[0]
|
|
* cp_nonpixel_size = venus_sec_non_pixel/virtual-addr-pool[1]
|
|
*/
|
|
ret = qcom_scm_mem_protect_video_var(res->cp_start,
|
|
res->cp_size,
|
|
res->cp_nonpixel_start,
|
|
res->cp_nonpixel_size);
|
|
if (ret) {
|
|
qcom_scm_pas_shutdown(VENUS_PAS_ID);
|
|
dev_err(dev, "set virtual address ranges fail (%d)\n",
|
|
ret);
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int venus_shutdown(struct venus_core *core)
|
|
{
|
|
int ret;
|
|
|
|
if (core->use_tz)
|
|
ret = qcom_scm_pas_shutdown(VENUS_PAS_ID);
|
|
else
|
|
ret = venus_shutdown_no_tz(core);
|
|
|
|
return ret;
|
|
}
|
|
|
|
int venus_firmware_init(struct venus_core *core)
|
|
{
|
|
struct platform_device_info info;
|
|
struct iommu_domain *iommu_dom;
|
|
struct platform_device *pdev;
|
|
struct device_node *np;
|
|
int ret;
|
|
|
|
np = of_get_child_by_name(core->dev->of_node, "video-firmware");
|
|
if (!np) {
|
|
core->use_tz = true;
|
|
return 0;
|
|
}
|
|
|
|
memset(&info, 0, sizeof(info));
|
|
info.fwnode = &np->fwnode;
|
|
info.parent = core->dev;
|
|
info.name = np->name;
|
|
info.dma_mask = DMA_BIT_MASK(32);
|
|
|
|
pdev = platform_device_register_full(&info);
|
|
if (IS_ERR(pdev)) {
|
|
of_node_put(np);
|
|
return PTR_ERR(pdev);
|
|
}
|
|
|
|
pdev->dev.of_node = np;
|
|
|
|
ret = of_dma_configure(&pdev->dev, np, true);
|
|
if (ret) {
|
|
dev_err(core->dev, "dma configure fail\n");
|
|
goto err_unregister;
|
|
}
|
|
|
|
core->fw.dev = &pdev->dev;
|
|
|
|
iommu_dom = iommu_paging_domain_alloc(core->fw.dev);
|
|
if (IS_ERR(iommu_dom)) {
|
|
dev_err(core->fw.dev, "Failed to allocate iommu domain\n");
|
|
ret = PTR_ERR(iommu_dom);
|
|
goto err_unregister;
|
|
}
|
|
|
|
ret = iommu_attach_device(iommu_dom, core->fw.dev);
|
|
if (ret) {
|
|
dev_err(core->fw.dev, "could not attach device\n");
|
|
goto err_iommu_free;
|
|
}
|
|
|
|
core->fw.iommu_domain = iommu_dom;
|
|
|
|
of_node_put(np);
|
|
|
|
return 0;
|
|
|
|
err_iommu_free:
|
|
iommu_domain_free(iommu_dom);
|
|
err_unregister:
|
|
platform_device_unregister(pdev);
|
|
of_node_put(np);
|
|
return ret;
|
|
}
|
|
|
|
void venus_firmware_deinit(struct venus_core *core)
|
|
{
|
|
struct iommu_domain *iommu;
|
|
|
|
if (!core->fw.dev)
|
|
return;
|
|
|
|
iommu = core->fw.iommu_domain;
|
|
|
|
iommu_detach_device(iommu, core->fw.dev);
|
|
|
|
if (core->fw.iommu_domain) {
|
|
iommu_domain_free(iommu);
|
|
core->fw.iommu_domain = NULL;
|
|
}
|
|
|
|
platform_device_unregister(to_platform_device(core->fw.dev));
|
|
}
|