|
@@ -620,6 +620,59 @@ static int init_unity_mappings_for_device(struct dma_ops_domain *dma_dom,
|
|
|
* called with domain->lock held
|
|
|
*/
|
|
|
|
|
|
+/*
|
|
|
+ * This function is used to add a new aperture range to an existing
|
|
|
+ * aperture in case of dma_ops domain allocation or address allocation
|
|
|
+ * failure.
|
|
|
+ */
|
|
|
+static int alloc_new_range(struct dma_ops_domain *dma_dom,
|
|
|
+ bool populate, gfp_t gfp)
|
|
|
+{
|
|
|
+ int index = dma_dom->aperture_size >> APERTURE_RANGE_SHIFT;
|
|
|
+
|
|
|
+ if (index >= APERTURE_MAX_RANGES)
|
|
|
+ return -ENOMEM;
|
|
|
+
|
|
|
+ dma_dom->aperture[index] = kzalloc(sizeof(struct aperture_range), gfp);
|
|
|
+ if (!dma_dom->aperture[index])
|
|
|
+ return -ENOMEM;
|
|
|
+
|
|
|
+ dma_dom->aperture[index]->bitmap = (void *)get_zeroed_page(gfp);
|
|
|
+ if (!dma_dom->aperture[index]->bitmap)
|
|
|
+ goto out_free;
|
|
|
+
|
|
|
+ dma_dom->aperture[index]->offset = dma_dom->aperture_size;
|
|
|
+
|
|
|
+ if (populate) {
|
|
|
+ unsigned long address = dma_dom->aperture_size;
|
|
|
+ int i, num_ptes = APERTURE_RANGE_PAGES / 512;
|
|
|
+ u64 *pte, *pte_page;
|
|
|
+
|
|
|
+ for (i = 0; i < num_ptes; ++i) {
|
|
|
+ pte = alloc_pte(&dma_dom->domain, address,
|
|
|
+ &pte_page, gfp);
|
|
|
+ if (!pte)
|
|
|
+ goto out_free;
|
|
|
+
|
|
|
+ dma_dom->aperture[index]->pte_pages[i] = pte_page;
|
|
|
+
|
|
|
+ address += APERTURE_RANGE_SIZE / 64;
|
|
|
+ }
|
|
|
+ }
|
|
|
+
|
|
|
+ dma_dom->aperture_size += APERTURE_RANGE_SIZE;
|
|
|
+
|
|
|
+ return 0;
|
|
|
+
|
|
|
+out_free:
|
|
|
+ free_page((unsigned long)dma_dom->aperture[index]->bitmap);
|
|
|
+
|
|
|
+ kfree(dma_dom->aperture[index]);
|
|
|
+ dma_dom->aperture[index] = NULL;
|
|
|
+
|
|
|
+ return -ENOMEM;
|
|
|
+}
|
|
|
+
|
|
|
static unsigned long dma_ops_area_alloc(struct device *dev,
|
|
|
struct dma_ops_domain *dom,
|
|
|
unsigned int pages,
|
|
@@ -832,9 +885,6 @@ static struct dma_ops_domain *dma_ops_domain_alloc(struct amd_iommu *iommu,
|
|
|
unsigned order)
|
|
|
{
|
|
|
struct dma_ops_domain *dma_dom;
|
|
|
- unsigned i, num_pte_pages;
|
|
|
- u64 *l2_pde;
|
|
|
- u64 address;
|
|
|
|
|
|
/*
|
|
|
* Currently the DMA aperture must be between 32 MB and 1GB in size
|
|
@@ -846,11 +896,6 @@ static struct dma_ops_domain *dma_ops_domain_alloc(struct amd_iommu *iommu,
|
|
|
if (!dma_dom)
|
|
|
return NULL;
|
|
|
|
|
|
- dma_dom->aperture[0] = kzalloc(sizeof(struct aperture_range),
|
|
|
- GFP_KERNEL);
|
|
|
- if (!dma_dom->aperture[0])
|
|
|
- goto free_dma_dom;
|
|
|
-
|
|
|
spin_lock_init(&dma_dom->domain.lock);
|
|
|
|
|
|
dma_dom->domain.id = domain_id_alloc();
|
|
@@ -862,10 +907,13 @@ static struct dma_ops_domain *dma_ops_domain_alloc(struct amd_iommu *iommu,
|
|
|
dma_dom->domain.priv = dma_dom;
|
|
|
if (!dma_dom->domain.pt_root)
|
|
|
goto free_dma_dom;
|
|
|
- dma_dom->aperture_size = APERTURE_RANGE_SIZE;
|
|
|
- dma_dom->aperture[0]->bitmap = (void *)get_zeroed_page(GFP_KERNEL);
|
|
|
- if (!dma_dom->aperture[0]->bitmap)
|
|
|
+
|
|
|
+ dma_dom->need_flush = false;
|
|
|
+ dma_dom->target_dev = 0xffff;
|
|
|
+
|
|
|
+ if (alloc_new_range(dma_dom, true, GFP_KERNEL))
|
|
|
goto free_dma_dom;
|
|
|
+
|
|
|
/*
|
|
|
* mark the first page as allocated so we never return 0 as
|
|
|
* a valid dma-address. So we can use 0 as error value
|
|
@@ -873,9 +921,6 @@ static struct dma_ops_domain *dma_ops_domain_alloc(struct amd_iommu *iommu,
|
|
|
dma_dom->aperture[0]->bitmap[0] = 1;
|
|
|
dma_dom->next_address = 0;
|
|
|
|
|
|
- dma_dom->need_flush = false;
|
|
|
- dma_dom->target_dev = 0xffff;
|
|
|
-
|
|
|
/* Intialize the exclusion range if necessary */
|
|
|
if (iommu->exclusion_start &&
|
|
|
iommu->exclusion_start < dma_dom->aperture_size) {
|
|
@@ -886,28 +931,6 @@ static struct dma_ops_domain *dma_ops_domain_alloc(struct amd_iommu *iommu,
|
|
|
dma_ops_reserve_addresses(dma_dom, startpage, pages);
|
|
|
}
|
|
|
|
|
|
- /*
|
|
|
- * At the last step, build the page tables so we don't need to
|
|
|
- * allocate page table pages in the dma_ops mapping/unmapping
|
|
|
- * path for the first 128MB of dma address space.
|
|
|
- */
|
|
|
- num_pte_pages = dma_dom->aperture_size / (PAGE_SIZE * 512);
|
|
|
-
|
|
|
- l2_pde = (u64 *)get_zeroed_page(GFP_KERNEL);
|
|
|
- if (l2_pde == NULL)
|
|
|
- goto free_dma_dom;
|
|
|
-
|
|
|
- dma_dom->domain.pt_root[0] = IOMMU_L2_PDE(virt_to_phys(l2_pde));
|
|
|
-
|
|
|
- for (i = 0; i < num_pte_pages; ++i) {
|
|
|
- u64 **pte_page = &dma_dom->aperture[0]->pte_pages[i];
|
|
|
- *pte_page = (u64 *)get_zeroed_page(GFP_KERNEL);
|
|
|
- if (!*pte_page)
|
|
|
- goto free_dma_dom;
|
|
|
- address = virt_to_phys(*pte_page);
|
|
|
- l2_pde[i] = IOMMU_L1_PDE(address);
|
|
|
- }
|
|
|
-
|
|
|
return dma_dom;
|
|
|
|
|
|
free_dma_dom:
|