On 21/02/18 22:59, Jordan Crouse wrote:
Allow a SMMU device to opt into allocating a TTBR1 pagetable.

The size of the TTBR1 region will be the same as
the TTBR0 size with the sign extension bit set on the highest
bit in the region unless the upstream size is 49 bits and then
the sign-extension bit will be set on the 49th bit.

Um, isn't the 49th bit still "the highest bit" if the address size is 49 bits? ;)

The map/unmap operations will automatically use the appropriate
pagetable based on the specified iova and the existing mask.

Signed-off-by: Jordan Crouse <jcro...@codeaurora.org>
---
  drivers/iommu/arm-smmu-regs.h  |   2 -
  drivers/iommu/arm-smmu.c       |  22 ++++--
  drivers/iommu/io-pgtable-arm.c | 160 ++++++++++++++++++++++++++++++++++++-----
  drivers/iommu/io-pgtable-arm.h |  20 ++++++
  drivers/iommu/io-pgtable.h     |  16 ++++-
  5 files changed, 192 insertions(+), 28 deletions(-)

diff --git a/drivers/iommu/arm-smmu-regs.h b/drivers/iommu/arm-smmu-regs.h
index a1226e4ab5f8..0ce85d5b22e9 100644
--- a/drivers/iommu/arm-smmu-regs.h
+++ b/drivers/iommu/arm-smmu-regs.h
@@ -193,8 +193,6 @@ enum arm_smmu_s2cr_privcfg {
  #define RESUME_RETRY                  (0 << 0)
  #define RESUME_TERMINATE              (1 << 0)
-#define TTBCR2_SEP_SHIFT 15
-#define TTBCR2_SEP_UPSTREAM            (0x7 << TTBCR2_SEP_SHIFT)
  #define TTBCR2_AS                     (1 << 4)
#define TTBRn_ASID_SHIFT 48
diff --git a/drivers/iommu/arm-smmu.c b/drivers/iommu/arm-smmu.c
index 69e7c60792a8..ebfa59b59622 100644
--- a/drivers/iommu/arm-smmu.c
+++ b/drivers/iommu/arm-smmu.c
@@ -248,6 +248,7 @@ struct arm_smmu_domain {
        enum arm_smmu_domain_stage      stage;
        struct mutex                    init_mutex; /* Protects smmu pointer */
        spinlock_t                      cb_lock; /* Serialises ATS1* ops and 
TLB syncs */
+       u32 attributes;
        struct iommu_domain             domain;
  };
@@ -598,7 +599,6 @@ static void arm_smmu_init_context_bank(struct arm_smmu_domain *smmu_domain,
                } else {
                        cb->tcr[0] = pgtbl_cfg->arm_lpae_s1_cfg.tcr;
                        cb->tcr[1] = pgtbl_cfg->arm_lpae_s1_cfg.tcr >> 32;
-                       cb->tcr[1] |= TTBCR2_SEP_UPSTREAM;
                        if (cfg->fmt == ARM_SMMU_CTX_FMT_AARCH64)
                                cb->tcr[1] |= TTBCR2_AS;
                }
@@ -729,6 +729,9 @@ static int arm_smmu_init_domain_context(struct iommu_domain 
*domain,
        enum io_pgtable_fmt fmt;
        struct arm_smmu_domain *smmu_domain = to_smmu_domain(domain);
        struct arm_smmu_cfg *cfg = &smmu_domain->cfg;
+       unsigned int quirks =
+               smmu_domain->attributes & (1 << DOMAIN_ATTR_ENABLE_TTBR1) ?
+                       IO_PGTABLE_QUIRK_ARM_TTBR1 : 0;
mutex_lock(&smmu_domain->init_mutex);
        if (smmu_domain->smmu)
@@ -852,7 +855,11 @@ static int arm_smmu_init_domain_context(struct 
iommu_domain *domain,
        else
                cfg->asid = cfg->cbndx + smmu->cavium_id_base;
+ if (smmu->features & ARM_SMMU_FEAT_COHERENT_WALK)
+               quirks |= IO_PGTABLE_QUIRK_NO_DMA;
+
        pgtbl_cfg = (struct io_pgtable_cfg) {
+               .quirks         = quirks,
                .pgsize_bitmap  = smmu->pgsize_bitmap,
                .ias            = ias,
                .oas            = oas,
@@ -860,9 +867,6 @@ static int arm_smmu_init_domain_context(struct iommu_domain 
*domain,
                .iommu_dev      = smmu->dev,
        };
- if (smmu->features & ARM_SMMU_FEAT_COHERENT_WALK)
-               pgtbl_cfg.quirks = IO_PGTABLE_QUIRK_NO_DMA;
-
        smmu_domain->smmu = smmu;
        pgtbl_ops = alloc_io_pgtable_ops(fmt, &pgtbl_cfg, smmu_domain);
        if (!pgtbl_ops) {
@@ -1477,6 +1481,10 @@ static int arm_smmu_domain_get_attr(struct iommu_domain 
*domain,
        case DOMAIN_ATTR_NESTING:
                *(int *)data = (smmu_domain->stage == ARM_SMMU_DOMAIN_NESTED);
                return 0;
+       case DOMAIN_ATTR_ENABLE_TTBR1:
+               *((int *)data) = !!(smmu_domain->attributes
+                                       & (1 << DOMAIN_ATTR_ENABLE_TTBR1));
+               return 0;
        default:
                return -ENODEV;
        }
@@ -1505,6 +1513,12 @@ static int arm_smmu_domain_set_attr(struct iommu_domain 
*domain,
                else
                        smmu_domain->stage = ARM_SMMU_DOMAIN_S1;
+ break;
+       case DOMAIN_ATTR_ENABLE_TTBR1:
+               if (*((int *)data))
+                       smmu_domain->attributes |=
+                               1 << DOMAIN_ATTR_ENABLE_TTBR1;
+               ret = 0;
                break;
        default:
                ret = -ENODEV;
diff --git a/drivers/iommu/io-pgtable-arm.c b/drivers/iommu/io-pgtable-arm.c
index fff0b6ba0a69..1bd0045f2cb7 100644
--- a/drivers/iommu/io-pgtable-arm.c
+++ b/drivers/iommu/io-pgtable-arm.c
@@ -152,7 +152,7 @@ struct arm_lpae_io_pgtable {
        unsigned long           pg_shift;
        unsigned long           bits_per_level;
- void *pgd;
+       void                    *pgd[2];

This might be reasonable for short-descriptor, but I really don't like it for LPAE. The two tables are more or less independent in terms of size, granule, etc., so this brings in a lot of artificial coupling.

I think it would be a lot cleaner for io-pgtable to have little or no knowledge of this, and it be down to the caller to allocate two tables and merge the TCRs, then dispatch maps/unmaps to the appropriate table by itself.

  };
typedef u64 arm_lpae_iopte;
@@ -394,20 +394,48 @@ static arm_lpae_iopte arm_lpae_prot_to_pte(struct 
arm_lpae_io_pgtable *data,
        return pte;
  }
+static inline arm_lpae_iopte *
+arm_lpae_get_table(struct arm_lpae_io_pgtable *data, unsigned long iova)
+{
+       struct io_pgtable_cfg *cfg = &data->iop.cfg;
+
+       if (cfg->quirks & IO_PGTABLE_QUIRK_ARM_TTBR1)  {
+               unsigned long mask;
+
+               /*
+                * if ias is 48 it really means that bit 48 is the sign
+                * extension bit, otherwise the sign extension bit is ias - 1
+                * (for example, bit 31 for ias 32)
+                */
+               mask = (cfg->ias == 48) ? (1UL << 48) :
+                       (1UL << (cfg->ias - 1));

This would look less silly if it was done in the SMMU driver where the original UBS information is directly to hand, instead of having to reverse-engineer it from the pagetable config (on every operation, no less).

That said, it's still going to be pretty fragile in general, since on SMMUv1 the UBS is merely guessed at from the IPA size, and either way all it tells you is what the SMMU knows about its own interfaces; it doesn't have a clue how many bits the masters connected to said interface(s) are actually capable of driving.

+
+               if (iova & mask)
+                       return data->pgd[1];
+       }
+
+       return data->pgd[0];
+}
+
  static int arm_lpae_map(struct io_pgtable_ops *ops, unsigned long iova,
                        phys_addr_t paddr, size_t size, int iommu_prot)
  {
        struct arm_lpae_io_pgtable *data = io_pgtable_ops_to_data(ops);
-       arm_lpae_iopte *ptep = data->pgd;
+       arm_lpae_iopte *ptep;
        int ret, lvl = ARM_LPAE_START_LVL(data);
        arm_lpae_iopte prot;
+ ptep = arm_lpae_get_table(data, iova);
+
        /* If no access, then nothing to do */
        if (!(iommu_prot & (IOMMU_READ | IOMMU_WRITE)))
                return 0;
- if (WARN_ON(iova >= (1ULL << data->iop.cfg.ias) ||
-                   paddr >= (1ULL << data->iop.cfg.oas)))
+       if (WARN_ON(paddr >= (1ULL << data->iop.cfg.oas)))
+               return -ERANGE;
+
+       if (WARN_ON(!(data->iop.cfg.quirks & IO_PGTABLE_QUIRK_ARM_TTBR1) &&
+                   iova >= (1ULL << data->iop.cfg.ias)))
                return -ERANGE;
prot = arm_lpae_prot_to_pte(data, iommu_prot);
@@ -456,7 +484,10 @@ static void arm_lpae_free_pgtable(struct io_pgtable *iop)
  {
        struct arm_lpae_io_pgtable *data = io_pgtable_to_data(iop);
- __arm_lpae_free_pgtable(data, ARM_LPAE_START_LVL(data), data->pgd);
+       __arm_lpae_free_pgtable(data, ARM_LPAE_START_LVL(data), data->pgd[0]);
+       if (data->pgd[1])
+               __arm_lpae_free_pgtable(data, ARM_LPAE_START_LVL(data),
+                       data->pgd[1]);
        kfree(data);
  }
@@ -564,10 +595,13 @@ static int arm_lpae_unmap(struct io_pgtable_ops *ops, unsigned long iova,
                          size_t size)
  {
        struct arm_lpae_io_pgtable *data = io_pgtable_ops_to_data(ops);
-       arm_lpae_iopte *ptep = data->pgd;
+       arm_lpae_iopte *ptep;
        int lvl = ARM_LPAE_START_LVL(data);
- if (WARN_ON(iova >= (1ULL << data->iop.cfg.ias)))
+       ptep = arm_lpae_get_table(data, iova);
+
+       if (WARN_ON(!(data->iop.cfg.quirks & IO_PGTABLE_QUIRK_ARM_TTBR1) &&
+                   iova >= (1ULL << data->iop.cfg.ias)))
                return 0;
return __arm_lpae_unmap(data, iova, size, lvl, ptep);
@@ -577,9 +611,11 @@ static phys_addr_t arm_lpae_iova_to_phys(struct 
io_pgtable_ops *ops,
                                         unsigned long iova)
  {
        struct arm_lpae_io_pgtable *data = io_pgtable_ops_to_data(ops);
-       arm_lpae_iopte pte, *ptep = data->pgd;
+       arm_lpae_iopte pte, *ptep;
        int lvl = ARM_LPAE_START_LVL(data);
+ ptep = arm_lpae_get_table(data, iova);
+
        do {
                /* Valid IOPTE pointer? */
                if (!ptep)
@@ -689,13 +725,82 @@ arm_lpae_alloc_pgtable(struct io_pgtable_cfg *cfg)
        return data;
  }
+static u64 arm_64_lpae_setup_ttbr1(struct io_pgtable_cfg *cfg,
+               struct arm_lpae_io_pgtable *data)
+
+{
+       u64 reg;
+
+       /* If TTBR1 is disabled, disable speculative walks through the TTBR1 */
+       if (!(cfg->quirks & IO_PGTABLE_QUIRK_ARM_TTBR1)) {
+               reg = ARM_LPAE_TCR_EPD1;
+               reg |= (ARM_LPAE_TCR_SEP_UPSTREAM << ARM_LPAE_TCR_SEP_SHIFT);
+               return reg;
+       }
+
+       reg = (ARM_LPAE_TCR_SH_IS << ARM_LPAE_TCR_SH1_SHIFT) |
+             (ARM_LPAE_TCR_RGN_WBWA << ARM_LPAE_TCR_IRGN1_SHIFT) |
+             (ARM_LPAE_TCR_RGN_WBWA << ARM_LPAE_TCR_ORGN1_SHIFT);
+
+       switch (1 << data->pg_shift) {
+       case SZ_4K:
+               reg |= ARM_LPAE_TCR_TG1_4K;
+               break;
+       case SZ_16K:
+               reg |= ARM_LPAE_TCR_TG1_16K;
+               break;
+       case SZ_64K:
+               reg |= ARM_LPAE_TCR_TG1_64K;
+               break;
+       }
+
+       /* Set T1SZ */
+       reg |= (64ULL - cfg->ias) << ARM_LPAE_TCR_T1SZ_SHIFT;
+
+       /* Set the SEP bit based on the size */
+       switch (cfg->ias) {
+       case 32:
+               reg |= (ARM_LPAE_TCR_SEP_31 << ARM_LPAE_TCR_SEP_SHIFT);
+               break;
+       case 36:
+               reg |= (ARM_LPAE_TCR_SEP_35 << ARM_LPAE_TCR_SEP_SHIFT);
+               break;
+       case 40:
+               reg |= (ARM_LPAE_TCR_SEP_39 << ARM_LPAE_TCR_SEP_SHIFT);
+               break;
+       case 42:
+               reg |= (ARM_LPAE_TCR_SEP_41 << ARM_LPAE_TCR_SEP_SHIFT);
+               break;
+       case 44:
+               reg |= (ARM_LPAE_TCR_SEP_43 << ARM_LPAE_TCR_SEP_SHIFT);
+               break;
+       case 48:
+               /*
+                * If ias is 48 then that probably means that the UBS on the
+                * device was 0101b (49) which is a special case that assumes
+                * bit 48 is the sign extension bit. In this case we are
+                * expected to use ARM_LPAE_TCR_SEP_UPSTREAM to use bit 48 as
+                * the extension bit. One might be confused because there is
+                * also an option to set the SEP to bit 47 but this is probably
+                * not what the arm-smmu driver intended.
+                */

Again, a clear sign that this probably isn't the most appropriate place to be trying to handle this.

+       default:
+               reg |= (ARM_LPAE_TCR_SEP_UPSTREAM << ARM_LPAE_TCR_SEP_SHIFT);
+               break;
+       }
+
+       return reg;
+}
+
  static struct io_pgtable *
  arm_64_lpae_alloc_pgtable_s1(struct io_pgtable_cfg *cfg, void *cookie)
  {
        u64 reg;
        struct arm_lpae_io_pgtable *data;
- if (cfg->quirks & ~(IO_PGTABLE_QUIRK_ARM_NS | IO_PGTABLE_QUIRK_NO_DMA))
+       if (cfg->quirks & ~(IO_PGTABLE_QUIRK_ARM_NS |
+                       IO_PGTABLE_QUIRK_NO_DMA |
+                       IO_PGTABLE_QUIRK_ARM_TTBR1))
                return NULL;
data = arm_lpae_alloc_pgtable(cfg);
@@ -744,8 +849,9 @@ arm_64_lpae_alloc_pgtable_s1(struct io_pgtable_cfg *cfg, 
void *cookie)
reg |= (64ULL - cfg->ias) << ARM_LPAE_TCR_T0SZ_SHIFT; - /* Disable speculative walks through TTBR1 */
-       reg |= ARM_LPAE_TCR_EPD1;
+       /* Bring in the TTBR1 configuration */
+       reg |= arm_64_lpae_setup_ttbr1(cfg, data);
+
        cfg->arm_lpae_s1_cfg.tcr = reg;
/* MAIRs */
@@ -760,16 +866,32 @@ arm_64_lpae_alloc_pgtable_s1(struct io_pgtable_cfg *cfg, 
void *cookie)
        cfg->arm_lpae_s1_cfg.mair[1] = 0;
/* Looking good; allocate a pgd */
-       data->pgd = __arm_lpae_alloc_pages(data->pgd_size, GFP_KERNEL, cfg);
-       if (!data->pgd)
+       data->pgd[0] = __arm_lpae_alloc_pages(data->pgd_size, GFP_KERNEL, cfg);
+       if (!data->pgd[0])
                goto out_free_data;
+
+       if (cfg->quirks & IO_PGTABLE_QUIRK_ARM_TTBR1) {
+               data->pgd[1] = __arm_lpae_alloc_pages(data->pgd_size,
+                       GFP_KERNEL, cfg);
+               if (!data->pgd[1]) {
+                       __arm_lpae_free_pages(data->pgd[0], data->pgd_size,
+                               cfg);
+                       goto out_free_data;
+               }
+       } else {
+               data->pgd[1] = NULL;
+       }
+
        /* Ensure the empty pgd is visible before any actual TTBR write */
        wmb();
/* TTBRs */
-       cfg->arm_lpae_s1_cfg.ttbr[0] = virt_to_phys(data->pgd);
-       cfg->arm_lpae_s1_cfg.ttbr[1] = 0;
+       cfg->arm_lpae_s1_cfg.ttbr[0] = virt_to_phys(data->pgd[0]);
+
+       if (data->pgd[1])
+               cfg->arm_lpae_s1_cfg.ttbr[1] = virt_to_phys(data->pgd[1]);
+
        return &data->iop;
out_free_data:
@@ -854,15 +976,15 @@ arm_64_lpae_alloc_pgtable_s2(struct io_pgtable_cfg *cfg, 
void *cookie)
        cfg->arm_lpae_s2_cfg.vtcr = reg;
/* Allocate pgd pages */
-       data->pgd = __arm_lpae_alloc_pages(data->pgd_size, GFP_KERNEL, cfg);
-       if (!data->pgd)
+       data->pgd[0] = __arm_lpae_alloc_pages(data->pgd_size, GFP_KERNEL, cfg);
+       if (!data->pgd[0])
                goto out_free_data;
/* Ensure the empty pgd is visible before any actual TTBR write */
        wmb();
/* VTTBR */
-       cfg->arm_lpae_s2_cfg.vttbr = virt_to_phys(data->pgd);
+       cfg->arm_lpae_s2_cfg.vttbr = virt_to_phys(data->pgd[0]);
        return &data->iop;
out_free_data:
@@ -960,7 +1082,7 @@ static void __init arm_lpae_dump_ops(struct io_pgtable_ops 
*ops)
                cfg->pgsize_bitmap, cfg->ias);
        pr_err("data: %d levels, 0x%zx pgd_size, %lu pg_shift, %lu bits_per_level, 
pgd @ %p\n",
                data->levels, data->pgd_size, data->pg_shift,
-               data->bits_per_level, data->pgd);
+               data->bits_per_level, data->pgd[0]);
  }
#define __FAIL(ops, i) ({ \
diff --git a/drivers/iommu/io-pgtable-arm.h b/drivers/iommu/io-pgtable-arm.h
index cb31314971ac..6344b1d359a5 100644
--- a/drivers/iommu/io-pgtable-arm.h
+++ b/drivers/iommu/io-pgtable-arm.h
@@ -25,14 +25,21 @@
  #define ARM_LPAE_TCR_TG0_64K          (1 << 14)
  #define ARM_LPAE_TCR_TG0_16K          (2 << 14)
+#define ARM_LPAE_TCR_TG1_16K (1 << 30)
+#define ARM_LPAE_TCR_TG1_4K             (2 << 30)
+#define ARM_LPAE_TCR_TG1_64K            (3 << 30)
+
  #define ARM_LPAE_TCR_SH0_SHIFT                12
+#define ARM_LPAE_TCR_SH1_SHIFT         28
  #define ARM_LPAE_TCR_SH0_MASK         0x3
  #define ARM_LPAE_TCR_SH_NS            0
  #define ARM_LPAE_TCR_SH_OS            2
  #define ARM_LPAE_TCR_SH_IS            3
#define ARM_LPAE_TCR_ORGN0_SHIFT 10
+#define ARM_LPAE_TCR_ORGN1_SHIFT       26
  #define ARM_LPAE_TCR_IRGN0_SHIFT      8
+#define ARM_LPAE_TCR_IRGN1_SHIFT       24
  #define ARM_LPAE_TCR_RGN_MASK         0x3
  #define ARM_LPAE_TCR_RGN_NC           0
  #define ARM_LPAE_TCR_RGN_WBWA         1
@@ -45,6 +52,9 @@
  #define ARM_LPAE_TCR_T0SZ_SHIFT               0
  #define ARM_LPAE_TCR_SZ_MASK          0x3f
+#define ARM_LPAE_TCR_T1SZ_SHIFT 16
+#define ARM_LPAE_TCR_T1SZ_MASK          0x3f
+
  #define ARM_LPAE_TCR_PS_SHIFT         16
  #define ARM_LPAE_TCR_PS_MASK          0x7
@@ -58,6 +68,16 @@
  #define ARM_LPAE_TCR_PS_44_BIT                0x4ULL
  #define ARM_LPAE_TCR_PS_48_BIT                0x5ULL
+#define ARM_LPAE_TCR_SEP_SHIFT (15 + 32)
+
+#define ARM_LPAE_TCR_SEP_31            0x0ULL
+#define ARM_LPAE_TCR_SEP_35            0x1ULL
+#define ARM_LPAE_TCR_SEP_39            0x2ULL
+#define ARM_LPAE_TCR_SEP_41            0x3ULL
+#define ARM_LPAE_TCR_SEP_43            0x4ULL
+#define ARM_LPAE_TCR_SEP_47            0x5ULL
+#define ARM_LPAE_TCR_SEP_UPSTREAM      0x7ULL
+
  #define ARM_LPAE_MAIR_ATTR_SHIFT(n)   ((n) << 3)
  #define ARM_LPAE_MAIR_ATTR_MASK               0xff
  #define ARM_LPAE_MAIR_ATTR_DEVICE     0x04
diff --git a/drivers/iommu/io-pgtable.h b/drivers/iommu/io-pgtable.h
index cd2e1eafffe6..55f7b60cc44d 100644
--- a/drivers/iommu/io-pgtable.h
+++ b/drivers/iommu/io-pgtable.h
@@ -71,12 +71,18 @@ struct io_pgtable_cfg {
         *      be accessed by a fully cache-coherent IOMMU or CPU (e.g. for a
         *      software-emulated IOMMU), such that pagetable updates need not
         *      be treated as explicit DMA data.
+        *
+        * PGTABLE_QUIRK_ARM_TTBR1: Specifies that TTBR1 has been enabled on
+        *      this domain. Set up the configuration registers and dyanmically
+        *      choose which pagetable (TTBR0 or TTBR1) a mapping should go into
+        *      based on the address.
         */
        #define IO_PGTABLE_QUIRK_ARM_NS         BIT(0)
        #define IO_PGTABLE_QUIRK_NO_PERMS       BIT(1)
        #define IO_PGTABLE_QUIRK_TLBI_ON_MAP    BIT(2)
        #define IO_PGTABLE_QUIRK_ARM_MTK_4GB    BIT(3)
        #define IO_PGTABLE_QUIRK_NO_DMA         BIT(4)
+       #define IO_PGTABLE_QUIRK_ARM_TTBR1      BIT(5)
        unsigned long                   quirks;
        unsigned long                   pgsize_bitmap;
        unsigned int                    ias;
@@ -173,18 +179,22 @@ struct io_pgtable {
static inline void io_pgtable_tlb_flush_all(struct io_pgtable *iop)
  {
-       iop->cfg.tlb->tlb_flush_all(iop->cookie);
+       if (iop->cfg.tlb)
+               iop->cfg.tlb->tlb_flush_all(iop->cookie);

What's going on here?

It's not obvious how this change is relevant to TTBR1 support, and either way I can't see how an io_pgtable with no TLB ops wouldn't just be fundamentally broken.

Robin.

  }
static inline void io_pgtable_tlb_add_flush(struct io_pgtable *iop,
                unsigned long iova, size_t size, size_t granule, bool leaf)
  {
-       iop->cfg.tlb->tlb_add_flush(iova, size, granule, leaf, iop->cookie);
+       if (iop->cfg.tlb)
+               iop->cfg.tlb->tlb_add_flush(iova, size, granule, leaf,
+                       iop->cookie);
  }
static inline void io_pgtable_tlb_sync(struct io_pgtable *iop)
  {
-       iop->cfg.tlb->tlb_sync(iop->cookie);
+       if (iop->cfg.tlb)
+               iop->cfg.tlb->tlb_sync(iop->cookie);
  }
/**

_______________________________________________
iommu mailing list
iommu@lists.linux-foundation.org
https://lists.linuxfoundation.org/mailman/listinfo/iommu

Reply via email to