mirror of
git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-08-05 08:43:31 +00:00
iommu/qcom: Use accessor functions for iommu private data
Make use of dev_iommu_priv_set/get() functions. Signed-off-by: Joerg Roedel <jroedel@suse.de> Link: https://lore.kernel.org/r/20200326150841.10083-15-joro@8bytes.org
This commit is contained in:
parent
3524b5592c
commit
09b5dfff9a
1 changed files with 36 additions and 25 deletions
|
@ -74,16 +74,19 @@ static struct qcom_iommu_domain *to_qcom_iommu_domain(struct iommu_domain *dom)
|
||||||
|
|
||||||
static const struct iommu_ops qcom_iommu_ops;
|
static const struct iommu_ops qcom_iommu_ops;
|
||||||
|
|
||||||
static struct qcom_iommu_dev * to_iommu(struct iommu_fwspec *fwspec)
|
static struct qcom_iommu_dev * to_iommu(struct device *dev)
|
||||||
{
|
{
|
||||||
|
struct iommu_fwspec *fwspec = dev_iommu_fwspec_get(dev);
|
||||||
|
|
||||||
if (!fwspec || fwspec->ops != &qcom_iommu_ops)
|
if (!fwspec || fwspec->ops != &qcom_iommu_ops)
|
||||||
return NULL;
|
return NULL;
|
||||||
return fwspec->iommu_priv;
|
|
||||||
|
return dev_iommu_priv_get(dev);
|
||||||
}
|
}
|
||||||
|
|
||||||
static struct qcom_iommu_ctx * to_ctx(struct iommu_fwspec *fwspec, unsigned asid)
|
static struct qcom_iommu_ctx * to_ctx(struct device *dev, unsigned asid)
|
||||||
{
|
{
|
||||||
struct qcom_iommu_dev *qcom_iommu = to_iommu(fwspec);
|
struct qcom_iommu_dev *qcom_iommu = to_iommu(dev);
|
||||||
if (!qcom_iommu)
|
if (!qcom_iommu)
|
||||||
return NULL;
|
return NULL;
|
||||||
return qcom_iommu->ctxs[asid - 1];
|
return qcom_iommu->ctxs[asid - 1];
|
||||||
|
@ -115,11 +118,14 @@ iommu_readq(struct qcom_iommu_ctx *ctx, unsigned reg)
|
||||||
|
|
||||||
static void qcom_iommu_tlb_sync(void *cookie)
|
static void qcom_iommu_tlb_sync(void *cookie)
|
||||||
{
|
{
|
||||||
struct iommu_fwspec *fwspec = cookie;
|
struct iommu_fwspec *fwspec;
|
||||||
|
struct device *dev = cookie;
|
||||||
unsigned i;
|
unsigned i;
|
||||||
|
|
||||||
|
fwspec = dev_iommu_fwspec_get(dev);
|
||||||
|
|
||||||
for (i = 0; i < fwspec->num_ids; i++) {
|
for (i = 0; i < fwspec->num_ids; i++) {
|
||||||
struct qcom_iommu_ctx *ctx = to_ctx(fwspec, fwspec->ids[i]);
|
struct qcom_iommu_ctx *ctx = to_ctx(dev, fwspec->ids[i]);
|
||||||
unsigned int val, ret;
|
unsigned int val, ret;
|
||||||
|
|
||||||
iommu_writel(ctx, ARM_SMMU_CB_TLBSYNC, 0);
|
iommu_writel(ctx, ARM_SMMU_CB_TLBSYNC, 0);
|
||||||
|
@ -133,11 +139,14 @@ static void qcom_iommu_tlb_sync(void *cookie)
|
||||||
|
|
||||||
static void qcom_iommu_tlb_inv_context(void *cookie)
|
static void qcom_iommu_tlb_inv_context(void *cookie)
|
||||||
{
|
{
|
||||||
struct iommu_fwspec *fwspec = cookie;
|
struct device *dev = cookie;
|
||||||
|
struct iommu_fwspec *fwspec;
|
||||||
unsigned i;
|
unsigned i;
|
||||||
|
|
||||||
|
fwspec = dev_iommu_fwspec_get(dev);
|
||||||
|
|
||||||
for (i = 0; i < fwspec->num_ids; i++) {
|
for (i = 0; i < fwspec->num_ids; i++) {
|
||||||
struct qcom_iommu_ctx *ctx = to_ctx(fwspec, fwspec->ids[i]);
|
struct qcom_iommu_ctx *ctx = to_ctx(dev, fwspec->ids[i]);
|
||||||
iommu_writel(ctx, ARM_SMMU_CB_S1_TLBIASID, ctx->asid);
|
iommu_writel(ctx, ARM_SMMU_CB_S1_TLBIASID, ctx->asid);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -147,13 +156,16 @@ static void qcom_iommu_tlb_inv_context(void *cookie)
|
||||||
static void qcom_iommu_tlb_inv_range_nosync(unsigned long iova, size_t size,
|
static void qcom_iommu_tlb_inv_range_nosync(unsigned long iova, size_t size,
|
||||||
size_t granule, bool leaf, void *cookie)
|
size_t granule, bool leaf, void *cookie)
|
||||||
{
|
{
|
||||||
struct iommu_fwspec *fwspec = cookie;
|
struct device *dev = cookie;
|
||||||
|
struct iommu_fwspec *fwspec;
|
||||||
unsigned i, reg;
|
unsigned i, reg;
|
||||||
|
|
||||||
reg = leaf ? ARM_SMMU_CB_S1_TLBIVAL : ARM_SMMU_CB_S1_TLBIVA;
|
reg = leaf ? ARM_SMMU_CB_S1_TLBIVAL : ARM_SMMU_CB_S1_TLBIVA;
|
||||||
|
|
||||||
|
fwspec = dev_iommu_fwspec_get(dev);
|
||||||
|
|
||||||
for (i = 0; i < fwspec->num_ids; i++) {
|
for (i = 0; i < fwspec->num_ids; i++) {
|
||||||
struct qcom_iommu_ctx *ctx = to_ctx(fwspec, fwspec->ids[i]);
|
struct qcom_iommu_ctx *ctx = to_ctx(dev, fwspec->ids[i]);
|
||||||
size_t s = size;
|
size_t s = size;
|
||||||
|
|
||||||
iova = (iova >> 12) << 12;
|
iova = (iova >> 12) << 12;
|
||||||
|
@ -222,9 +234,10 @@ static irqreturn_t qcom_iommu_fault(int irq, void *dev)
|
||||||
|
|
||||||
static int qcom_iommu_init_domain(struct iommu_domain *domain,
|
static int qcom_iommu_init_domain(struct iommu_domain *domain,
|
||||||
struct qcom_iommu_dev *qcom_iommu,
|
struct qcom_iommu_dev *qcom_iommu,
|
||||||
struct iommu_fwspec *fwspec)
|
struct device *dev)
|
||||||
{
|
{
|
||||||
struct qcom_iommu_domain *qcom_domain = to_qcom_iommu_domain(domain);
|
struct qcom_iommu_domain *qcom_domain = to_qcom_iommu_domain(domain);
|
||||||
|
struct iommu_fwspec *fwspec = dev_iommu_fwspec_get(dev);
|
||||||
struct io_pgtable_ops *pgtbl_ops;
|
struct io_pgtable_ops *pgtbl_ops;
|
||||||
struct io_pgtable_cfg pgtbl_cfg;
|
struct io_pgtable_cfg pgtbl_cfg;
|
||||||
int i, ret = 0;
|
int i, ret = 0;
|
||||||
|
@ -243,7 +256,7 @@ static int qcom_iommu_init_domain(struct iommu_domain *domain,
|
||||||
};
|
};
|
||||||
|
|
||||||
qcom_domain->iommu = qcom_iommu;
|
qcom_domain->iommu = qcom_iommu;
|
||||||
pgtbl_ops = alloc_io_pgtable_ops(ARM_32_LPAE_S1, &pgtbl_cfg, fwspec);
|
pgtbl_ops = alloc_io_pgtable_ops(ARM_32_LPAE_S1, &pgtbl_cfg, dev);
|
||||||
if (!pgtbl_ops) {
|
if (!pgtbl_ops) {
|
||||||
dev_err(qcom_iommu->dev, "failed to allocate pagetable ops\n");
|
dev_err(qcom_iommu->dev, "failed to allocate pagetable ops\n");
|
||||||
ret = -ENOMEM;
|
ret = -ENOMEM;
|
||||||
|
@ -256,7 +269,7 @@ static int qcom_iommu_init_domain(struct iommu_domain *domain,
|
||||||
domain->geometry.force_aperture = true;
|
domain->geometry.force_aperture = true;
|
||||||
|
|
||||||
for (i = 0; i < fwspec->num_ids; i++) {
|
for (i = 0; i < fwspec->num_ids; i++) {
|
||||||
struct qcom_iommu_ctx *ctx = to_ctx(fwspec, fwspec->ids[i]);
|
struct qcom_iommu_ctx *ctx = to_ctx(dev, fwspec->ids[i]);
|
||||||
|
|
||||||
if (!ctx->secure_init) {
|
if (!ctx->secure_init) {
|
||||||
ret = qcom_scm_restore_sec_cfg(qcom_iommu->sec_id, ctx->asid);
|
ret = qcom_scm_restore_sec_cfg(qcom_iommu->sec_id, ctx->asid);
|
||||||
|
@ -363,8 +376,7 @@ static void qcom_iommu_domain_free(struct iommu_domain *domain)
|
||||||
|
|
||||||
static int qcom_iommu_attach_dev(struct iommu_domain *domain, struct device *dev)
|
static int qcom_iommu_attach_dev(struct iommu_domain *domain, struct device *dev)
|
||||||
{
|
{
|
||||||
struct iommu_fwspec *fwspec = dev_iommu_fwspec_get(dev);
|
struct qcom_iommu_dev *qcom_iommu = to_iommu(dev);
|
||||||
struct qcom_iommu_dev *qcom_iommu = to_iommu(fwspec);
|
|
||||||
struct qcom_iommu_domain *qcom_domain = to_qcom_iommu_domain(domain);
|
struct qcom_iommu_domain *qcom_domain = to_qcom_iommu_domain(domain);
|
||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
|
@ -375,7 +387,7 @@ static int qcom_iommu_attach_dev(struct iommu_domain *domain, struct device *dev
|
||||||
|
|
||||||
/* Ensure that the domain is finalized */
|
/* Ensure that the domain is finalized */
|
||||||
pm_runtime_get_sync(qcom_iommu->dev);
|
pm_runtime_get_sync(qcom_iommu->dev);
|
||||||
ret = qcom_iommu_init_domain(domain, qcom_iommu, fwspec);
|
ret = qcom_iommu_init_domain(domain, qcom_iommu, dev);
|
||||||
pm_runtime_put_sync(qcom_iommu->dev);
|
pm_runtime_put_sync(qcom_iommu->dev);
|
||||||
if (ret < 0)
|
if (ret < 0)
|
||||||
return ret;
|
return ret;
|
||||||
|
@ -397,9 +409,9 @@ static int qcom_iommu_attach_dev(struct iommu_domain *domain, struct device *dev
|
||||||
|
|
||||||
static void qcom_iommu_detach_dev(struct iommu_domain *domain, struct device *dev)
|
static void qcom_iommu_detach_dev(struct iommu_domain *domain, struct device *dev)
|
||||||
{
|
{
|
||||||
struct iommu_fwspec *fwspec = dev_iommu_fwspec_get(dev);
|
|
||||||
struct qcom_iommu_dev *qcom_iommu = to_iommu(fwspec);
|
|
||||||
struct qcom_iommu_domain *qcom_domain = to_qcom_iommu_domain(domain);
|
struct qcom_iommu_domain *qcom_domain = to_qcom_iommu_domain(domain);
|
||||||
|
struct iommu_fwspec *fwspec = dev_iommu_fwspec_get(dev);
|
||||||
|
struct qcom_iommu_dev *qcom_iommu = to_iommu(dev);
|
||||||
unsigned i;
|
unsigned i;
|
||||||
|
|
||||||
if (WARN_ON(!qcom_domain->iommu))
|
if (WARN_ON(!qcom_domain->iommu))
|
||||||
|
@ -407,7 +419,7 @@ static void qcom_iommu_detach_dev(struct iommu_domain *domain, struct device *de
|
||||||
|
|
||||||
pm_runtime_get_sync(qcom_iommu->dev);
|
pm_runtime_get_sync(qcom_iommu->dev);
|
||||||
for (i = 0; i < fwspec->num_ids; i++) {
|
for (i = 0; i < fwspec->num_ids; i++) {
|
||||||
struct qcom_iommu_ctx *ctx = to_ctx(fwspec, fwspec->ids[i]);
|
struct qcom_iommu_ctx *ctx = to_ctx(dev, fwspec->ids[i]);
|
||||||
|
|
||||||
/* Disable the context bank: */
|
/* Disable the context bank: */
|
||||||
iommu_writel(ctx, ARM_SMMU_CB_SCTLR, 0);
|
iommu_writel(ctx, ARM_SMMU_CB_SCTLR, 0);
|
||||||
|
@ -514,7 +526,7 @@ static bool qcom_iommu_capable(enum iommu_cap cap)
|
||||||
|
|
||||||
static int qcom_iommu_add_device(struct device *dev)
|
static int qcom_iommu_add_device(struct device *dev)
|
||||||
{
|
{
|
||||||
struct qcom_iommu_dev *qcom_iommu = to_iommu(dev_iommu_fwspec_get(dev));
|
struct qcom_iommu_dev *qcom_iommu = to_iommu(dev);
|
||||||
struct iommu_group *group;
|
struct iommu_group *group;
|
||||||
struct device_link *link;
|
struct device_link *link;
|
||||||
|
|
||||||
|
@ -545,7 +557,7 @@ static int qcom_iommu_add_device(struct device *dev)
|
||||||
|
|
||||||
static void qcom_iommu_remove_device(struct device *dev)
|
static void qcom_iommu_remove_device(struct device *dev)
|
||||||
{
|
{
|
||||||
struct qcom_iommu_dev *qcom_iommu = to_iommu(dev_iommu_fwspec_get(dev));
|
struct qcom_iommu_dev *qcom_iommu = to_iommu(dev);
|
||||||
|
|
||||||
if (!qcom_iommu)
|
if (!qcom_iommu)
|
||||||
return;
|
return;
|
||||||
|
@ -557,7 +569,6 @@ static void qcom_iommu_remove_device(struct device *dev)
|
||||||
|
|
||||||
static int qcom_iommu_of_xlate(struct device *dev, struct of_phandle_args *args)
|
static int qcom_iommu_of_xlate(struct device *dev, struct of_phandle_args *args)
|
||||||
{
|
{
|
||||||
struct iommu_fwspec *fwspec = dev_iommu_fwspec_get(dev);
|
|
||||||
struct qcom_iommu_dev *qcom_iommu;
|
struct qcom_iommu_dev *qcom_iommu;
|
||||||
struct platform_device *iommu_pdev;
|
struct platform_device *iommu_pdev;
|
||||||
unsigned asid = args->args[0];
|
unsigned asid = args->args[0];
|
||||||
|
@ -583,14 +594,14 @@ static int qcom_iommu_of_xlate(struct device *dev, struct of_phandle_args *args)
|
||||||
WARN_ON(asid > qcom_iommu->num_ctxs))
|
WARN_ON(asid > qcom_iommu->num_ctxs))
|
||||||
return -EINVAL;
|
return -EINVAL;
|
||||||
|
|
||||||
if (!fwspec->iommu_priv) {
|
if (!dev_iommu_priv_get(dev)) {
|
||||||
fwspec->iommu_priv = qcom_iommu;
|
dev_iommu_priv_set(dev, qcom_iommu);
|
||||||
} else {
|
} else {
|
||||||
/* make sure devices iommus dt node isn't referring to
|
/* make sure devices iommus dt node isn't referring to
|
||||||
* multiple different iommu devices. Multiple context
|
* multiple different iommu devices. Multiple context
|
||||||
* banks are ok, but multiple devices are not:
|
* banks are ok, but multiple devices are not:
|
||||||
*/
|
*/
|
||||||
if (WARN_ON(qcom_iommu != fwspec->iommu_priv))
|
if (WARN_ON(qcom_iommu != dev_iommu_priv_get(dev)))
|
||||||
return -EINVAL;
|
return -EINVAL;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Add table
Reference in a new issue