ce44d1ff46
All related domains provide MSI parent functionality, so the fallback code to the original platform MSI implementation is not longer required. Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Signed-off-by: Anna-Maria Behnsen <anna-maria@linutronix.de> Signed-off-by: Shivamurthy Shastri <shivamurthy.shastri@linutronix.de> Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Link: https://lore.kernel.org/r/20240623142235.881677325@linutronix.de
369 lines
10 KiB
C
369 lines
10 KiB
C
/*
|
|
* Copyright (C) 2017 Marvell
|
|
*
|
|
* Hanna Hawa <hannah@marvell.com>
|
|
* Thomas Petazzoni <thomas.petazzoni@free-electrons.com>
|
|
*
|
|
* This file is licensed under the terms of the GNU General Public
|
|
* License version 2. This program is licensed "as is" without any
|
|
* warranty of any kind, whether express or implied.
|
|
*/
|
|
|
|
#include <linux/interrupt.h>
|
|
#include <linux/irq.h>
|
|
#include <linux/irqchip.h>
|
|
#include <linux/irqdomain.h>
|
|
#include <linux/jump_label.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/msi.h>
|
|
#include <linux/of_irq.h>
|
|
#include <linux/of_platform.h>
|
|
#include <linux/platform_device.h>
|
|
|
|
#include "irq-msi-lib.h"
|
|
|
|
#include <dt-bindings/interrupt-controller/mvebu-icu.h>
|
|
|
|
/* ICU registers */
|
|
#define ICU_SETSPI_NSR_AL 0x10
|
|
#define ICU_SETSPI_NSR_AH 0x14
|
|
#define ICU_CLRSPI_NSR_AL 0x18
|
|
#define ICU_CLRSPI_NSR_AH 0x1c
|
|
#define ICU_SET_SEI_AL 0x50
|
|
#define ICU_SET_SEI_AH 0x54
|
|
#define ICU_CLR_SEI_AL 0x58
|
|
#define ICU_CLR_SEI_AH 0x5C
|
|
#define ICU_INT_CFG(x) (0x100 + 4 * (x))
|
|
#define ICU_INT_ENABLE BIT(24)
|
|
#define ICU_IS_EDGE BIT(28)
|
|
#define ICU_GROUP_SHIFT 29
|
|
|
|
/* ICU definitions */
|
|
#define ICU_MAX_IRQS 207
|
|
#define ICU_SATA0_ICU_ID 109
|
|
#define ICU_SATA1_ICU_ID 107
|
|
|
|
struct mvebu_icu_subset_data {
|
|
unsigned int icu_group;
|
|
unsigned int offset_set_ah;
|
|
unsigned int offset_set_al;
|
|
unsigned int offset_clr_ah;
|
|
unsigned int offset_clr_al;
|
|
};
|
|
|
|
struct mvebu_icu {
|
|
void __iomem *base;
|
|
struct device *dev;
|
|
};
|
|
|
|
struct mvebu_icu_msi_data {
|
|
struct mvebu_icu *icu;
|
|
atomic_t initialized;
|
|
const struct mvebu_icu_subset_data *subset_data;
|
|
};
|
|
|
|
static DEFINE_STATIC_KEY_FALSE(legacy_bindings);
|
|
|
|
static int mvebu_icu_translate(struct irq_domain *d, struct irq_fwspec *fwspec,
|
|
unsigned long *hwirq, unsigned int *type)
|
|
{
|
|
unsigned int param_count = static_branch_unlikely(&legacy_bindings) ? 3 : 2;
|
|
struct mvebu_icu_msi_data *msi_data = d->host_data;
|
|
struct mvebu_icu *icu = msi_data->icu;
|
|
|
|
/* Check the count of the parameters in dt */
|
|
if (WARN_ON(fwspec->param_count != param_count)) {
|
|
dev_err(icu->dev, "wrong ICU parameter count %d\n",
|
|
fwspec->param_count);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (static_branch_unlikely(&legacy_bindings)) {
|
|
*hwirq = fwspec->param[1];
|
|
*type = fwspec->param[2] & IRQ_TYPE_SENSE_MASK;
|
|
if (fwspec->param[0] != ICU_GRP_NSR) {
|
|
dev_err(icu->dev, "wrong ICU group type %x\n",
|
|
fwspec->param[0]);
|
|
return -EINVAL;
|
|
}
|
|
} else {
|
|
*hwirq = fwspec->param[0];
|
|
*type = fwspec->param[1] & IRQ_TYPE_SENSE_MASK;
|
|
|
|
/*
|
|
* The ICU receives level interrupts. While the NSR are also
|
|
* level interrupts, SEI are edge interrupts. Force the type
|
|
* here in this case. Please note that this makes the interrupt
|
|
* handling unreliable.
|
|
*/
|
|
if (msi_data->subset_data->icu_group == ICU_GRP_SEI)
|
|
*type = IRQ_TYPE_EDGE_RISING;
|
|
}
|
|
|
|
if (*hwirq >= ICU_MAX_IRQS) {
|
|
dev_err(icu->dev, "invalid interrupt number %ld\n", *hwirq);
|
|
return -EINVAL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void mvebu_icu_init(struct mvebu_icu *icu,
|
|
struct mvebu_icu_msi_data *msi_data,
|
|
struct msi_msg *msg)
|
|
{
|
|
const struct mvebu_icu_subset_data *subset = msi_data->subset_data;
|
|
|
|
if (atomic_cmpxchg(&msi_data->initialized, false, true))
|
|
return;
|
|
|
|
/* Set 'SET' ICU SPI message address in AP */
|
|
writel_relaxed(msg[0].address_hi, icu->base + subset->offset_set_ah);
|
|
writel_relaxed(msg[0].address_lo, icu->base + subset->offset_set_al);
|
|
|
|
if (subset->icu_group != ICU_GRP_NSR)
|
|
return;
|
|
|
|
/* Set 'CLEAR' ICU SPI message address in AP (level-MSI only) */
|
|
writel_relaxed(msg[1].address_hi, icu->base + subset->offset_clr_ah);
|
|
writel_relaxed(msg[1].address_lo, icu->base + subset->offset_clr_al);
|
|
}
|
|
|
|
static int mvebu_icu_msi_init(struct irq_domain *domain, struct msi_domain_info *info,
|
|
unsigned int virq, irq_hw_number_t hwirq, msi_alloc_info_t *arg)
|
|
{
|
|
irq_domain_set_hwirq_and_chip(domain, virq, hwirq, info->chip, info->chip_data);
|
|
return irq_set_irqchip_state(virq, IRQCHIP_STATE_PENDING, false);
|
|
}
|
|
|
|
static void mvebu_icu_set_desc(msi_alloc_info_t *arg, struct msi_desc *desc)
|
|
{
|
|
arg->desc = desc;
|
|
arg->hwirq = (u32)desc->data.icookie.value;
|
|
}
|
|
|
|
static void mvebu_icu_write_msi_msg(struct irq_data *d, struct msi_msg *msg)
|
|
{
|
|
struct mvebu_icu_msi_data *msi_data = d->chip_data;
|
|
unsigned int icu_group = msi_data->subset_data->icu_group;
|
|
struct msi_desc *desc = irq_data_get_msi_desc(d);
|
|
struct mvebu_icu *icu = msi_data->icu;
|
|
unsigned int type;
|
|
u32 icu_int;
|
|
|
|
if (msg->address_lo || msg->address_hi) {
|
|
/* One off initialization per domain */
|
|
mvebu_icu_init(icu, msi_data, msg);
|
|
/* Configure the ICU with irq number & type */
|
|
icu_int = msg->data | ICU_INT_ENABLE;
|
|
type = (unsigned int)(desc->data.icookie.value >> 32);
|
|
if (type & IRQ_TYPE_EDGE_RISING)
|
|
icu_int |= ICU_IS_EDGE;
|
|
icu_int |= icu_group << ICU_GROUP_SHIFT;
|
|
} else {
|
|
/* De-configure the ICU */
|
|
icu_int = 0;
|
|
}
|
|
|
|
writel_relaxed(icu_int, icu->base + ICU_INT_CFG(d->hwirq));
|
|
|
|
/*
|
|
* The SATA unit has 2 ports, and a dedicated ICU entry per
|
|
* port. The ahci sata driver supports only one irq interrupt
|
|
* per SATA unit. To solve this conflict, we configure the 2
|
|
* SATA wired interrupts in the south bridge into 1 GIC
|
|
* interrupt in the north bridge. Even if only a single port
|
|
* is enabled, if sata node is enabled, both interrupts are
|
|
* configured (regardless of which port is actually in use).
|
|
*/
|
|
if (d->hwirq == ICU_SATA0_ICU_ID || d->hwirq == ICU_SATA1_ICU_ID) {
|
|
writel_relaxed(icu_int, icu->base + ICU_INT_CFG(ICU_SATA0_ICU_ID));
|
|
writel_relaxed(icu_int, icu->base + ICU_INT_CFG(ICU_SATA1_ICU_ID));
|
|
}
|
|
}
|
|
|
|
static const struct msi_domain_template mvebu_icu_nsr_msi_template = {
|
|
.chip = {
|
|
.name = "ICU-NSR",
|
|
.irq_mask = irq_chip_mask_parent,
|
|
.irq_unmask = irq_chip_unmask_parent,
|
|
.irq_eoi = irq_chip_eoi_parent,
|
|
.irq_set_type = irq_chip_set_type_parent,
|
|
.irq_write_msi_msg = mvebu_icu_write_msi_msg,
|
|
.flags = IRQCHIP_SUPPORTS_LEVEL_MSI,
|
|
},
|
|
|
|
.ops = {
|
|
.msi_translate = mvebu_icu_translate,
|
|
.msi_init = mvebu_icu_msi_init,
|
|
.set_desc = mvebu_icu_set_desc,
|
|
},
|
|
|
|
.info = {
|
|
.bus_token = DOMAIN_BUS_WIRED_TO_MSI,
|
|
.flags = MSI_FLAG_LEVEL_CAPABLE |
|
|
MSI_FLAG_USE_DEV_FWNODE,
|
|
},
|
|
};
|
|
|
|
static const struct msi_domain_template mvebu_icu_sei_msi_template = {
|
|
.chip = {
|
|
.name = "ICU-SEI",
|
|
.irq_mask = irq_chip_mask_parent,
|
|
.irq_unmask = irq_chip_unmask_parent,
|
|
.irq_ack = irq_chip_ack_parent,
|
|
.irq_set_type = irq_chip_set_type_parent,
|
|
.irq_write_msi_msg = mvebu_icu_write_msi_msg,
|
|
.flags = IRQCHIP_SUPPORTS_LEVEL_MSI,
|
|
},
|
|
|
|
.ops = {
|
|
.msi_translate = mvebu_icu_translate,
|
|
.msi_init = mvebu_icu_msi_init,
|
|
.set_desc = mvebu_icu_set_desc,
|
|
},
|
|
|
|
.info = {
|
|
.bus_token = DOMAIN_BUS_WIRED_TO_MSI,
|
|
.flags = MSI_FLAG_LEVEL_CAPABLE |
|
|
MSI_FLAG_USE_DEV_FWNODE,
|
|
},
|
|
};
|
|
|
|
static const struct mvebu_icu_subset_data mvebu_icu_nsr_subset_data = {
|
|
.icu_group = ICU_GRP_NSR,
|
|
.offset_set_ah = ICU_SETSPI_NSR_AH,
|
|
.offset_set_al = ICU_SETSPI_NSR_AL,
|
|
.offset_clr_ah = ICU_CLRSPI_NSR_AH,
|
|
.offset_clr_al = ICU_CLRSPI_NSR_AL,
|
|
};
|
|
|
|
static const struct mvebu_icu_subset_data mvebu_icu_sei_subset_data = {
|
|
.icu_group = ICU_GRP_SEI,
|
|
.offset_set_ah = ICU_SET_SEI_AH,
|
|
.offset_set_al = ICU_SET_SEI_AL,
|
|
};
|
|
|
|
static const struct of_device_id mvebu_icu_subset_of_match[] = {
|
|
{
|
|
.compatible = "marvell,cp110-icu-nsr",
|
|
.data = &mvebu_icu_nsr_subset_data,
|
|
},
|
|
{
|
|
.compatible = "marvell,cp110-icu-sei",
|
|
.data = &mvebu_icu_sei_subset_data,
|
|
},
|
|
{},
|
|
};
|
|
|
|
static int mvebu_icu_subset_probe(struct platform_device *pdev)
|
|
{
|
|
const struct msi_domain_template *tmpl;
|
|
struct mvebu_icu_msi_data *msi_data;
|
|
struct device *dev = &pdev->dev;
|
|
bool sei;
|
|
|
|
msi_data = devm_kzalloc(dev, sizeof(*msi_data), GFP_KERNEL);
|
|
if (!msi_data)
|
|
return -ENOMEM;
|
|
|
|
if (static_branch_unlikely(&legacy_bindings)) {
|
|
msi_data->icu = dev_get_drvdata(dev);
|
|
msi_data->subset_data = &mvebu_icu_nsr_subset_data;
|
|
} else {
|
|
msi_data->icu = dev_get_drvdata(dev->parent);
|
|
msi_data->subset_data = of_device_get_match_data(dev);
|
|
}
|
|
|
|
dev->msi.domain = of_msi_get_domain(dev, dev->of_node, DOMAIN_BUS_PLATFORM_MSI);
|
|
if (!dev->msi.domain)
|
|
return -EPROBE_DEFER;
|
|
|
|
if (!irq_domain_get_of_node(dev->msi.domain))
|
|
return -ENODEV;
|
|
|
|
sei = msi_data->subset_data->icu_group == ICU_GRP_SEI;
|
|
tmpl = sei ? &mvebu_icu_sei_msi_template : &mvebu_icu_nsr_msi_template;
|
|
|
|
if (!msi_create_device_irq_domain(dev, MSI_DEFAULT_DOMAIN, tmpl,
|
|
ICU_MAX_IRQS, NULL, msi_data)) {
|
|
dev_err(dev, "Failed to create ICU MSI domain\n");
|
|
return -ENOMEM;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static struct platform_driver mvebu_icu_subset_driver = {
|
|
.probe = mvebu_icu_subset_probe,
|
|
.driver = {
|
|
.name = "mvebu-icu-subset",
|
|
.of_match_table = mvebu_icu_subset_of_match,
|
|
},
|
|
};
|
|
builtin_platform_driver(mvebu_icu_subset_driver);
|
|
|
|
static int mvebu_icu_probe(struct platform_device *pdev)
|
|
{
|
|
struct mvebu_icu *icu;
|
|
int i;
|
|
|
|
icu = devm_kzalloc(&pdev->dev, sizeof(struct mvebu_icu),
|
|
GFP_KERNEL);
|
|
if (!icu)
|
|
return -ENOMEM;
|
|
|
|
icu->dev = &pdev->dev;
|
|
|
|
icu->base = devm_platform_ioremap_resource(pdev, 0);
|
|
if (IS_ERR(icu->base))
|
|
return PTR_ERR(icu->base);
|
|
|
|
/*
|
|
* Legacy bindings: ICU is one node with one MSI parent: force manually
|
|
* the probe of the NSR interrupts side.
|
|
* New bindings: ICU node has children, one per interrupt controller
|
|
* having its own MSI parent: call platform_populate().
|
|
* All ICU instances should use the same bindings.
|
|
*/
|
|
if (!of_get_child_count(pdev->dev.of_node))
|
|
static_branch_enable(&legacy_bindings);
|
|
|
|
/*
|
|
* Clean all ICU interrupts of type NSR and SEI, required to
|
|
* avoid unpredictable SPI assignments done by firmware.
|
|
*/
|
|
for (i = 0 ; i < ICU_MAX_IRQS ; i++) {
|
|
u32 icu_int, icu_grp;
|
|
|
|
icu_int = readl_relaxed(icu->base + ICU_INT_CFG(i));
|
|
icu_grp = icu_int >> ICU_GROUP_SHIFT;
|
|
|
|
if (icu_grp == ICU_GRP_NSR ||
|
|
(icu_grp == ICU_GRP_SEI &&
|
|
!static_branch_unlikely(&legacy_bindings)))
|
|
writel_relaxed(0x0, icu->base + ICU_INT_CFG(i));
|
|
}
|
|
|
|
platform_set_drvdata(pdev, icu);
|
|
|
|
if (static_branch_unlikely(&legacy_bindings))
|
|
return mvebu_icu_subset_probe(pdev);
|
|
else
|
|
return devm_of_platform_populate(&pdev->dev);
|
|
}
|
|
|
|
static const struct of_device_id mvebu_icu_of_match[] = {
|
|
{ .compatible = "marvell,cp110-icu", },
|
|
{},
|
|
};
|
|
|
|
static struct platform_driver mvebu_icu_driver = {
|
|
.probe = mvebu_icu_probe,
|
|
.driver = {
|
|
.name = "mvebu-icu",
|
|
.of_match_table = mvebu_icu_of_match,
|
|
},
|
|
};
|
|
builtin_platform_driver(mvebu_icu_driver);
|