[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [PATCH v2 7/8] xen/arm: Remove Linux specific code that is not usable in XEN


  • To: Stefano Stabellini <sstabellini@xxxxxxxxxx>
  • From: Rahul Singh <Rahul.Singh@xxxxxxx>
  • Date: Wed, 2 Dec 2020 14:34:05 +0000
  • Accept-language: en-US
  • Arc-authentication-results: i=1; mx.microsoft.com 1; spf=pass smtp.mailfrom=arm.com; dmarc=pass action=none header.from=arm.com; dkim=pass header.d=arm.com; arc=none
  • Arc-message-signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=microsoft.com; s=arcselector9901; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-SenderADCheck; bh=pdfQ7/VJjzhu2i+wNCFu7xNJ6fP1Vv/UOiGCcIXYGEY=; b=ZB/iPj0xIyn/UCzb+a16nRl03wJDJeT2BpCcaVU/O3vs2Rfw/JmD8TUcw3l8ZFgHrU4Ou6uSoxIWy9aM52ECvATfaFqDWi5DN8pr94VdAE9vMWBL80TX/Z58YG/4IdQPI2+lD/fXCofHkdzNKTPZByc9ebKuCXP4+AlsOt5M5yUmTWvzFmQSjPzLAClXdbaRzHP/EZzfZZCcyPKBcBRiX1IR0vaCqYunroTh7mcWBVbddAWAhtp8UMkpcFzcOQApbCyFs52/mi3dFrAwaLN94wJ2dRTt7lvvGYdhjaZQmLlKyJ2fH4rfblfukDmGecBlwUObpm7/5YA2mAyt/JFXIQ==
  • Arc-seal: i=1; a=rsa-sha256; s=arcselector9901; d=microsoft.com; cv=none; b=lXIrfVbCgGxmY4t42EfYdDGTv8qyJs0h84NBvrLFzC4o93h4xQ5L/B4kDUCnhA2dE9hcdhtpSnjzWSJKu3g5eFlgYSRAfTfhHXhGFOatH+prUL0LJc+eyzG5MpPJnsB1gmoGl72D7VE2wWDMa0+58KDRCeLdHpyt+I6qGb43+DP3JoY6yuz+eikXSGhp7VWRBZIV7LqLHB/Jf1HK/QLW5ZXkAwWj8h9dq/9w9bqca+eWJZNhvVmaaUZ978h/6CwEImvrcZ/nyHOhKtPfgND0mwxWwkeZXapD8R2CWgQtfRmi5MRrF38+fLSI++0ANO+8fODuMC+BL8GXDg4HmX4vAw==
  • Authentication-results-original: kernel.org; dkim=none (message not signed) header.d=none;kernel.org; dmarc=none action=none header.from=arm.com;
  • Cc: "xen-devel@xxxxxxxxxxxxxxxxxxxx" <xen-devel@xxxxxxxxxxxxxxxxxxxx>, Bertrand Marquis <Bertrand.Marquis@xxxxxxx>, Julien Grall <julien@xxxxxxx>, Volodymyr Babchuk <Volodymyr_Babchuk@xxxxxxxx>
  • Delivery-date: Wed, 02 Dec 2020 14:34:30 +0000
  • List-id: Xen developer discussion <xen-devel.lists.xenproject.org>
  • Nodisclaimer: true
  • Original-authentication-results: kernel.org; dkim=none (message not signed) header.d=none;kernel.org; dmarc=none action=none header.from=arm.com;
  • Thread-index: AQHWxBX830xFguE4YkK2klATz/ucYKnjEigAgADWCQA=
  • Thread-topic: [PATCH v2 7/8] xen/arm: Remove Linux specific code that is not usable in XEN

Hello Stefano,

> On 2 Dec 2020, at 1:48 am, Stefano Stabellini <sstabellini@xxxxxxxxxx> wrote:
> 
> On Thu, 26 Nov 2020, Rahul Singh wrote:
>> struct io_pgtable_ops, struct io_pgtable_cfg, struct iommu_flush_ops,
>> and struct iommu_ops related code are linux specific.
>> 
>> Remove code related to above struct as code is dead code in XEN.
> 
> There are still instances of struct io_pgtable_cfg after applying this
> patch in the following functions:
> - arm_smmu_domain_finalise_s2
> - arm_smmu_domain_finalise
> 

Ok. I will remove the instances.
> 
> 
>> Signed-off-by: Rahul Singh <rahul.singh@xxxxxxx>
>> ---
>> xen/drivers/passthrough/arm/smmu-v3.c | 457 --------------------------
>> 1 file changed, 457 deletions(-)
>> 
>> diff --git a/xen/drivers/passthrough/arm/smmu-v3.c 
>> b/xen/drivers/passthrough/arm/smmu-v3.c
>> index 40e3890a58..55d1cba194 100644
>> --- a/xen/drivers/passthrough/arm/smmu-v3.c
>> +++ b/xen/drivers/passthrough/arm/smmu-v3.c
>> @@ -599,7 +593,6 @@ struct arm_smmu_domain {
>>      struct arm_smmu_device          *smmu;
>>      struct mutex                    init_mutex; /* Protects smmu pointer */
>> 
>> -    struct io_pgtable_ops           *pgtbl_ops;
>>      bool                            non_strict;
>> 
>>      enum arm_smmu_domain_stage      stage;
>> @@ -1297,74 +1290,6 @@ static void arm_smmu_tlb_inv_context(void *cookie)
>>      arm_smmu_cmdq_issue_sync(smmu);
>> }
>> 
>> -static void arm_smmu_tlb_inv_range_nosync(unsigned long iova, size_t size,
>> -                                      size_t granule, bool leaf, void 
>> *cookie)
>> -{
>> -    struct arm_smmu_domain *smmu_domain = cookie;
>> -    struct arm_smmu_device *smmu = smmu_domain->smmu;
>> -    struct arm_smmu_cmdq_ent cmd = {
>> -            .tlbi = {
>> -                    .leaf   = leaf,
>> -                    .addr   = iova,
>> -            },
>> -    };
>> -
>> -    cmd.opcode      = CMDQ_OP_TLBI_S2_IPA;
>> -    cmd.tlbi.vmid   = smmu_domain->s2_cfg.vmid;
>> -
>> -    do {
>> -            arm_smmu_cmdq_issue_cmd(smmu, &cmd);
>> -            cmd.tlbi.addr += granule;
>> -    } while (size -= granule);
>> -}
>> -
>> -static void arm_smmu_tlb_inv_page_nosync(struct iommu_iotlb_gather *gather,
>> -                                     unsigned long iova, size_t granule,
>> -                                     void *cookie)
>> -{
>> -    arm_smmu_tlb_inv_range_nosync(iova, granule, granule, true, cookie);
>> -}
>> -
>> -static void arm_smmu_tlb_inv_walk(unsigned long iova, size_t size,
>> -                              size_t granule, void *cookie)
>> -{
>> -    struct arm_smmu_domain *smmu_domain = cookie;
>> -    struct arm_smmu_device *smmu = smmu_domain->smmu;
>> -
>> -    arm_smmu_tlb_inv_range_nosync(iova, size, granule, false, cookie);
>> -    arm_smmu_cmdq_issue_sync(smmu);
>> -}
>> -
>> -static void arm_smmu_tlb_inv_leaf(unsigned long iova, size_t size,
>> -                              size_t granule, void *cookie)
>> -{
>> -    struct arm_smmu_domain *smmu_domain = cookie;
>> -    struct arm_smmu_device *smmu = smmu_domain->smmu;
>> -
>> -    arm_smmu_tlb_inv_range_nosync(iova, size, granule, true, cookie);
>> -    arm_smmu_cmdq_issue_sync(smmu);
>> -}
>> -
>> -static const struct iommu_flush_ops arm_smmu_flush_ops = {
>> -    .tlb_flush_all  = arm_smmu_tlb_inv_context,
>> -    .tlb_flush_walk = arm_smmu_tlb_inv_walk,
>> -    .tlb_flush_leaf = arm_smmu_tlb_inv_leaf,
>> -    .tlb_add_page   = arm_smmu_tlb_inv_page_nosync,
>> -};
>> -
>> -/* IOMMU API */
>> -static bool arm_smmu_capable(enum iommu_cap cap)
>> -{
>> -    switch (cap) {
>> -    case IOMMU_CAP_CACHE_COHERENCY:
>> -            return true;
>> -    case IOMMU_CAP_NOEXEC:
>> -            return true;
>> -    default:
>> -            return false;
>> -    }
>> -}
>> -
>> static struct iommu_domain *arm_smmu_domain_alloc(unsigned type)
>> {
>>      struct arm_smmu_domain *smmu_domain;
>> @@ -1421,7 +1346,6 @@ static void arm_smmu_domain_free(struct iommu_domain 
>> *domain)
>>      struct arm_smmu_s2_cfg *cfg = &smmu_domain->s2_cfg;
>> 
>>      iommu_put_dma_cookie(domain);
>> -    free_io_pgtable_ops(smmu_domain->pgtbl_ops);
>> 
>>      if (cfg->vmid)
>>              arm_smmu_bitmap_free(smmu->vmid_map, cfg->vmid);
>> @@ -1429,7 +1353,6 @@ static void arm_smmu_domain_free(struct iommu_domain 
>> *domain)
>>      kfree(smmu_domain);
>> }
>> 
>> -
>> static int arm_smmu_domain_finalise_s2(struct arm_smmu_domain *smmu_domain,
>>                                     struct arm_smmu_master *master,
>>                                     struct io_pgtable_cfg *pgtbl_cfg)
>> @@ -1437,7 +1360,6 @@ static int arm_smmu_domain_finalise_s2(struct 
>> arm_smmu_domain *smmu_domain,
>>      int vmid;
>>      struct arm_smmu_device *smmu = smmu_domain->smmu;
>>      struct arm_smmu_s2_cfg *cfg = &smmu_domain->s2_cfg;
>> -    typeof(&pgtbl_cfg->arm_lpae_s2_cfg.vtcr) vtcr;
>> 
>>      vmid = arm_smmu_bitmap_alloc(smmu->vmid_map, smmu->vmid_bits);
>>      if (vmid < 0)
>> @@ -1461,20 +1383,12 @@ static int arm_smmu_domain_finalise(struct 
>> iommu_domain *domain,
>> {
>>      int ret;
>>      unsigned long ias, oas;
>> -    enum io_pgtable_fmt fmt;
>> -    struct io_pgtable_cfg pgtbl_cfg;
>> -    struct io_pgtable_ops *pgtbl_ops;
>>      int (*finalise_stage_fn)(struct arm_smmu_domain *,
>>                               struct arm_smmu_master *,
>>                               struct io_pgtable_cfg *);
>>      struct arm_smmu_domain *smmu_domain = to_smmu_domain(domain);
>>      struct arm_smmu_device *smmu = smmu_domain->smmu;
>> 
>> -    if (domain->type == IOMMU_DOMAIN_IDENTITY) {
>> -            smmu_domain->stage = ARM_SMMU_DOMAIN_BYPASS;
>> -            return 0;
>> -    }
>> -
>>      /* Restrict the stage to what we can actually support */
>>      smmu_domain->stage = ARM_SMMU_DOMAIN_S2;
>> 
>> @@ -1483,40 +1397,17 @@ static int arm_smmu_domain_finalise(struct 
>> iommu_domain *domain,
>>      case ARM_SMMU_DOMAIN_S2:
>>              ias = smmu->ias;
>>              oas = smmu->oas;
>> -            fmt = ARM_64_LPAE_S2;
>>              finalise_stage_fn = arm_smmu_domain_finalise_s2;
>>              break;
>>      default:
>>              return -EINVAL;
>>      }
>> 
>> -    pgtbl_cfg = (struct io_pgtable_cfg) {
>> -            .pgsize_bitmap  = smmu->pgsize_bitmap,
>> -            .ias            = ias,
>> -            .oas            = oas,
>> -            .coherent_walk  = smmu->features & ARM_SMMU_FEAT_COHERENCY,
>> -            .tlb            = &arm_smmu_flush_ops,
>> -            .iommu_dev      = smmu->dev,
>> -    };
>> -
>> -    if (smmu_domain->non_strict)
>> -            pgtbl_cfg.quirks |= IO_PGTABLE_QUIRK_NON_STRICT;
>> -
>> -    pgtbl_ops = alloc_io_pgtable_ops(fmt, &pgtbl_cfg, smmu_domain);
>> -    if (!pgtbl_ops)
>> -            return -ENOMEM;
>> -
>> -    domain->pgsize_bitmap = pgtbl_cfg.pgsize_bitmap;
>> -    domain->geometry.aperture_end = (1UL << pgtbl_cfg.ias) - 1;
>> -    domain->geometry.force_aperture = true;
>> -
>>      ret = finalise_stage_fn(smmu_domain, master, &pgtbl_cfg);
>>      if (ret < 0) {
>> -            free_io_pgtable_ops(pgtbl_ops);
>>              return ret;
>>      }
>> 
>> -    smmu_domain->pgtbl_ops = pgtbl_ops;
>>      return 0;
>> }
>> 
>> @@ -1626,71 +1517,6 @@ out_unlock:
>>      return ret;
>> }
>> 
>> -static int arm_smmu_map(struct iommu_domain *domain, unsigned long iova,
>> -                    phys_addr_t paddr, size_t size, int prot, gfp_t gfp)
>> -{
>> -    struct io_pgtable_ops *ops = to_smmu_domain(domain)->pgtbl_ops;
>> -
>> -    if (!ops)
>> -            return -ENODEV;
>> -
>> -    return ops->map(ops, iova, paddr, size, prot, gfp);
>> -}
>> -
>> -static size_t arm_smmu_unmap(struct iommu_domain *domain, unsigned long 
>> iova,
>> -                         size_t size, struct iommu_iotlb_gather *gather)
>> -{
>> -    struct arm_smmu_domain *smmu_domain = to_smmu_domain(domain);
>> -    struct io_pgtable_ops *ops = smmu_domain->pgtbl_ops;
>> -
>> -    if (!ops)
>> -            return 0;
>> -
>> -    return ops->unmap(ops, iova, size, gather);
>> -}
>> -
>> -static void arm_smmu_flush_iotlb_all(struct iommu_domain *domain)
>> -{
>> -    struct arm_smmu_domain *smmu_domain = to_smmu_domain(domain);
>> -
>> -    if (smmu_domain->smmu)
>> -            arm_smmu_tlb_inv_context(smmu_domain);
>> -}
>> -
>> -static void arm_smmu_iotlb_sync(struct iommu_domain *domain,
>> -                            struct iommu_iotlb_gather *gather)
>> -{
>> -    struct arm_smmu_device *smmu = to_smmu_domain(domain)->smmu;
>> -
>> -    if (smmu)
>> -            arm_smmu_cmdq_issue_sync(smmu);
>> -}
>> -
>> -static phys_addr_t
>> -arm_smmu_iova_to_phys(struct iommu_domain *domain, dma_addr_t iova)
>> -{
>> -    struct io_pgtable_ops *ops = to_smmu_domain(domain)->pgtbl_ops;
>> -
>> -    if (domain->type == IOMMU_DOMAIN_IDENTITY)
>> -            return iova;
>> -
>> -    if (!ops)
>> -            return 0;
>> -
>> -    return ops->iova_to_phys(ops, iova);
>> -}
>> -
>> -static struct platform_driver arm_smmu_driver;
>> -
>> -static
>> -struct arm_smmu_device *arm_smmu_get_by_fwnode(struct fwnode_handle *fwnode)
>> -{
>> -    struct device *dev = 
>> driver_find_device_by_fwnode(&arm_smmu_driver.driver,
>> -                                                      fwnode);
>> -    put_device(dev);
>> -    return dev ? dev_get_drvdata(dev) : NULL;
>> -}
>> -
>> static bool arm_smmu_sid_in_range(struct arm_smmu_device *smmu, u32 sid)
>> {
>>      unsigned long limit = smmu->strtab_cfg.num_l1_ents;
>> @@ -1701,206 +1527,6 @@ static bool arm_smmu_sid_in_range(struct 
>> arm_smmu_device *smmu, u32 sid)
>>      return sid < limit;
>> }
>> 
>> -static struct iommu_ops arm_smmu_ops;
>> -
>> -static struct iommu_device *arm_smmu_probe_device(struct device *dev)
>> -{
>> -    int i, ret;
>> -    struct arm_smmu_device *smmu;
>> -    struct arm_smmu_master *master;
>> -    struct iommu_fwspec *fwspec = dev_iommu_fwspec_get(dev);
>> -
>> -    if (!fwspec || fwspec->ops != &arm_smmu_ops)
>> -            return ERR_PTR(-ENODEV);
>> -
>> -    if (WARN_ON_ONCE(dev_iommu_priv_get(dev)))
>> -            return ERR_PTR(-EBUSY);
>> -
>> -    smmu = arm_smmu_get_by_fwnode(fwspec->iommu_fwnode);
>> -    if (!smmu)
>> -            return ERR_PTR(-ENODEV);
>> -
>> -    master = kzalloc(sizeof(*master), GFP_KERNEL);
>> -    if (!master)
>> -            return ERR_PTR(-ENOMEM);
>> -
>> -    master->dev = dev;
>> -    master->smmu = smmu;
>> -    master->sids = fwspec->ids;
>> -    master->num_sids = fwspec->num_ids;
>> -    dev_iommu_priv_set(dev, master);
>> -
>> -    /* Check the SIDs are in range of the SMMU and our stream table */
>> -    for (i = 0; i < master->num_sids; i++) {
>> -            u32 sid = master->sids[i];
>> -
>> -            if (!arm_smmu_sid_in_range(smmu, sid)) {
>> -                    ret = -ERANGE;
>> -                    goto err_free_master;
>> -            }
>> -
>> -            /* Ensure l2 strtab is initialised */
>> -            if (smmu->features & ARM_SMMU_FEAT_2_LVL_STRTAB) {
>> -                    ret = arm_smmu_init_l2_strtab(smmu, sid);
>> -                    if (ret)
>> -                            goto err_free_master;
>> -            }
>> -    }
>> -
>> -    return &smmu->iommu;
>> -
>> -err_free_master:
>> -    kfree(master);
>> -    dev_iommu_priv_set(dev, NULL);
>> -    return ERR_PTR(ret);
>> -}
>> -
>> -static void arm_smmu_release_device(struct device *dev)
>> -{
>> -    struct iommu_fwspec *fwspec = dev_iommu_fwspec_get(dev);
>> -    struct arm_smmu_master *master;
>> -
>> -    if (!fwspec || fwspec->ops != &arm_smmu_ops)
>> -            return;
>> -
>> -    master = dev_iommu_priv_get(dev);
>> -    arm_smmu_detach_dev(master);
>> -    kfree(master);
>> -    iommu_fwspec_free(dev);
>> -}
>> -
>> -static struct iommu_group *arm_smmu_device_group(struct device *dev)
>> -{
>> -    struct iommu_group *group;
>> -
>> -    /*
>> -     * We don't support devices sharing stream IDs other than PCI RID
>> -     * aliases, since the necessary ID-to-device lookup becomes rather
>> -     * impractical given a potential sparse 32-bit stream ID space.
>> -     */
>> -    if (dev_is_pci(dev))
>> -            group = pci_device_group(dev);
>> -    else
>> -            group = generic_device_group(dev);
>> -
>> -    return group;
>> -}
>> -
>> -static int arm_smmu_domain_get_attr(struct iommu_domain *domain,
>> -                                enum iommu_attr attr, void *data)
>> -{
>> -    struct arm_smmu_domain *smmu_domain = to_smmu_domain(domain);
>> -
>> -    switch (domain->type) {
>> -    case IOMMU_DOMAIN_UNMANAGED:
>> -            switch (attr) {
>> -            case DOMAIN_ATTR_NESTING:
>> -                    *(int *)data = (smmu_domain->stage == 
>> ARM_SMMU_DOMAIN_NESTED);
>> -                    return 0;
>> -            default:
>> -                    return -ENODEV;
>> -            }
>> -            break;
>> -    case IOMMU_DOMAIN_DMA:
>> -            switch (attr) {
>> -            case DOMAIN_ATTR_DMA_USE_FLUSH_QUEUE:
>> -                    *(int *)data = smmu_domain->non_strict;
>> -                    return 0;
>> -            default:
>> -                    return -ENODEV;
>> -            }
>> -            break;
>> -    default:
>> -            return -EINVAL;
>> -    }
>> -}
>> -
>> -static int arm_smmu_domain_set_attr(struct iommu_domain *domain,
>> -                                enum iommu_attr attr, void *data)
>> -{
>> -    int ret = 0;
>> -    struct arm_smmu_domain *smmu_domain = to_smmu_domain(domain);
>> -
>> -    mutex_lock(&smmu_domain->init_mutex);
>> -
>> -    switch (domain->type) {
>> -    case IOMMU_DOMAIN_UNMANAGED:
>> -            switch (attr) {
>> -            case DOMAIN_ATTR_NESTING:
>> -                    if (smmu_domain->smmu) {
>> -                            ret = -EPERM;
>> -                            goto out_unlock;
>> -                    }
>> -
>> -                    if (*(int *)data)
>> -                            smmu_domain->stage = ARM_SMMU_DOMAIN_NESTED;
>> -                    else
>> -                            smmu_domain->stage = ARM_SMMU_DOMAIN_S1;
>> -                    break;
>> -            default:
>> -                    ret = -ENODEV;
>> -            }
>> -            break;
>> -    case IOMMU_DOMAIN_DMA:
>> -            switch(attr) {
>> -            case DOMAIN_ATTR_DMA_USE_FLUSH_QUEUE:
>> -                    smmu_domain->non_strict = *(int *)data;
>> -                    break;
>> -            default:
>> -                    ret = -ENODEV;
>> -            }
>> -            break;
>> -    default:
>> -            ret = -EINVAL;
>> -    }
>> -
>> -out_unlock:
>> -    mutex_unlock(&smmu_domain->init_mutex);
>> -    return ret;
>> -}
>> -
>> -static int arm_smmu_of_xlate(struct device *dev, struct of_phandle_args 
>> *args)
>> -{
>> -    return iommu_fwspec_add_ids(dev, args->args, 1);
>> -}
>> -
>> -static void arm_smmu_get_resv_regions(struct device *dev,
>> -                                  struct list_head *head)
>> -{
>> -    struct iommu_resv_region *region;
>> -    int prot = IOMMU_WRITE | IOMMU_NOEXEC | IOMMU_MMIO;
>> -
>> -    region = iommu_alloc_resv_region(MSI_IOVA_BASE, MSI_IOVA_LENGTH,
>> -                                     prot, IOMMU_RESV_SW_MSI);
>> -    if (!region)
>> -            return;
>> -
>> -    list_add_tail(&region->list, head);
>> -
>> -    iommu_dma_get_resv_regions(dev, head);
>> -}
> 
> Arguably this could have been removed previously as part of the MSI
> patch, but that's OK either way.

Ack. 
> 
> 
>> -static struct iommu_ops arm_smmu_ops = {
>> -    .capable                = arm_smmu_capable,
>> -    .domain_alloc           = arm_smmu_domain_alloc,
>> -    .domain_free            = arm_smmu_domain_free,
>> -    .attach_dev             = arm_smmu_attach_dev,
>> -    .map                    = arm_smmu_map,
>> -    .unmap                  = arm_smmu_unmap,
>> -    .flush_iotlb_all        = arm_smmu_flush_iotlb_all,
>> -    .iotlb_sync             = arm_smmu_iotlb_sync,
>> -    .iova_to_phys           = arm_smmu_iova_to_phys,
>> -    .probe_device           = arm_smmu_probe_device,
>> -    .release_device         = arm_smmu_release_device,
>> -    .device_group           = arm_smmu_device_group,
>> -    .domain_get_attr        = arm_smmu_domain_get_attr,
>> -    .domain_set_attr        = arm_smmu_domain_set_attr,
>> -    .of_xlate               = arm_smmu_of_xlate,
>> -    .get_resv_regions       = arm_smmu_get_resv_regions,
>> -    .put_resv_regions       = generic_iommu_put_resv_regions,
>> -    .pgsize_bitmap          = -1UL, /* Restricted during device attach */
>> -};
>> -
>> /* Probing and initialisation functions */
>> static int arm_smmu_init_one_queue(struct arm_smmu_device *smmu,
>>                                 struct arm_smmu_queue *q,
>> @@ -2515,21 +2139,10 @@ static int arm_smmu_device_hw_probe(struct 
>> arm_smmu_device *smmu)
>>      default:
>>              dev_info(smmu->dev,
>>                      "unknown output address size. Truncating to 48-bit\n");
>> -            fallthrough;
>>      case IDR5_OAS_48_BIT:
>>              smmu->oas = 48;
>>      }
>> 
>> -    if (arm_smmu_ops.pgsize_bitmap == -1UL)
>> -            arm_smmu_ops.pgsize_bitmap = smmu->pgsize_bitmap;
>> -    else
>> -            arm_smmu_ops.pgsize_bitmap |= smmu->pgsize_bitmap;
>> -
>> -    /* Set the DMA mask for our table walker */
>> -    if (dma_set_mask_and_coherent(smmu->dev, DMA_BIT_MASK(smmu->oas)))
>> -            dev_warn(smmu->dev,
>> -                     "failed to set DMA mask for table walker\n");
>> -
>>      smmu->ias = max(smmu->ias, smmu->oas);
>> 
>>      dev_info(smmu->dev, "ias %lu-bit, oas %lu-bit (features 0x%08x)\n",
>> @@ -2595,9 +2208,6 @@ static int arm_smmu_device_dt_probe(struct 
>> platform_device *pdev,
>> 
>>      parse_driver_options(smmu);
>> 
>> -    if (of_dma_is_coherent(dev->of_node))
>> -            smmu->features |= ARM_SMMU_FEAT_COHERENCY;
>> -
> 
> Why this change? The ARM_SMMU_FEAT_COHERENCY flag is still used in
> arm_smmu_device_hw_probe.

I remove this as this is linux specific.  I will remove ARM_SMMU_FEAT_COHERENCY 
flag  used in arm_smmu_device_hw_probe

Regards,
Rahul





 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.