|
@@ -271,20 +271,21 @@ static struct iovm_struct *alloc_iovm_area(struct iommu *obj, u32 da,
|
|
size_t bytes, u32 flags)
|
|
size_t bytes, u32 flags)
|
|
{
|
|
{
|
|
struct iovm_struct *new, *tmp;
|
|
struct iovm_struct *new, *tmp;
|
|
- u32 start, prev_end, alignement;
|
|
|
|
|
|
+ u32 start, prev_end, alignment;
|
|
|
|
|
|
if (!obj || !bytes)
|
|
if (!obj || !bytes)
|
|
return ERR_PTR(-EINVAL);
|
|
return ERR_PTR(-EINVAL);
|
|
|
|
|
|
start = da;
|
|
start = da;
|
|
- alignement = PAGE_SIZE;
|
|
|
|
|
|
+ alignment = PAGE_SIZE;
|
|
|
|
|
|
- if (flags & IOVMF_DA_ANON) {
|
|
|
|
- start = obj->da_start;
|
|
|
|
|
|
+ if (~flags & IOVMF_DA_FIXED) {
|
|
|
|
+ /* Don't map address 0 */
|
|
|
|
+ start = obj->da_start ? obj->da_start : alignment;
|
|
|
|
|
|
if (flags & IOVMF_LINEAR)
|
|
if (flags & IOVMF_LINEAR)
|
|
- alignement = iopgsz_max(bytes);
|
|
|
|
- start = roundup(start, alignement);
|
|
|
|
|
|
+ alignment = iopgsz_max(bytes);
|
|
|
|
+ start = roundup(start, alignment);
|
|
} else if (start < obj->da_start || start > obj->da_end ||
|
|
} else if (start < obj->da_start || start > obj->da_end ||
|
|
obj->da_end - start < bytes) {
|
|
obj->da_end - start < bytes) {
|
|
return ERR_PTR(-EINVAL);
|
|
return ERR_PTR(-EINVAL);
|
|
@@ -303,8 +304,8 @@ static struct iovm_struct *alloc_iovm_area(struct iommu *obj, u32 da,
|
|
if (tmp->da_start > start && (tmp->da_start - start) >= bytes)
|
|
if (tmp->da_start > start && (tmp->da_start - start) >= bytes)
|
|
goto found;
|
|
goto found;
|
|
|
|
|
|
- if (tmp->da_end >= start && flags & IOVMF_DA_ANON)
|
|
|
|
- start = roundup(tmp->da_end + 1, alignement);
|
|
|
|
|
|
+ if (tmp->da_end >= start && ~flags & IOVMF_DA_FIXED)
|
|
|
|
+ start = roundup(tmp->da_end + 1, alignment);
|
|
|
|
|
|
prev_end = tmp->da_end;
|
|
prev_end = tmp->da_end;
|
|
}
|
|
}
|
|
@@ -650,7 +651,6 @@ u32 iommu_vmap(struct iommu *obj, u32 da, const struct sg_table *sgt,
|
|
flags &= IOVMF_HW_MASK;
|
|
flags &= IOVMF_HW_MASK;
|
|
flags |= IOVMF_DISCONT;
|
|
flags |= IOVMF_DISCONT;
|
|
flags |= IOVMF_MMIO;
|
|
flags |= IOVMF_MMIO;
|
|
- flags |= (da ? IOVMF_DA_FIXED : IOVMF_DA_ANON);
|
|
|
|
|
|
|
|
da = __iommu_vmap(obj, da, sgt, va, bytes, flags);
|
|
da = __iommu_vmap(obj, da, sgt, va, bytes, flags);
|
|
if (IS_ERR_VALUE(da))
|
|
if (IS_ERR_VALUE(da))
|
|
@@ -690,7 +690,7 @@ EXPORT_SYMBOL_GPL(iommu_vunmap);
|
|
* @flags: iovma and page property
|
|
* @flags: iovma and page property
|
|
*
|
|
*
|
|
* Allocate @bytes linearly and creates 1-n-1 mapping and returns
|
|
* Allocate @bytes linearly and creates 1-n-1 mapping and returns
|
|
- * @da again, which might be adjusted if 'IOVMF_DA_ANON' is set.
|
|
|
|
|
|
+ * @da again, which might be adjusted if 'IOVMF_DA_FIXED' is not set.
|
|
*/
|
|
*/
|
|
u32 iommu_vmalloc(struct iommu *obj, u32 da, size_t bytes, u32 flags)
|
|
u32 iommu_vmalloc(struct iommu *obj, u32 da, size_t bytes, u32 flags)
|
|
{
|
|
{
|
|
@@ -709,7 +709,6 @@ u32 iommu_vmalloc(struct iommu *obj, u32 da, size_t bytes, u32 flags)
|
|
flags &= IOVMF_HW_MASK;
|
|
flags &= IOVMF_HW_MASK;
|
|
flags |= IOVMF_DISCONT;
|
|
flags |= IOVMF_DISCONT;
|
|
flags |= IOVMF_ALLOC;
|
|
flags |= IOVMF_ALLOC;
|
|
- flags |= (da ? IOVMF_DA_FIXED : IOVMF_DA_ANON);
|
|
|
|
|
|
|
|
sgt = sgtable_alloc(bytes, flags, da, 0);
|
|
sgt = sgtable_alloc(bytes, flags, da, 0);
|
|
if (IS_ERR(sgt)) {
|
|
if (IS_ERR(sgt)) {
|
|
@@ -780,7 +779,7 @@ static u32 __iommu_kmap(struct iommu *obj, u32 da, u32 pa, void *va,
|
|
* @flags: iovma and page property
|
|
* @flags: iovma and page property
|
|
*
|
|
*
|
|
* Creates 1-1-1 mapping and returns @da again, which can be
|
|
* Creates 1-1-1 mapping and returns @da again, which can be
|
|
- * adjusted if 'IOVMF_DA_ANON' is set.
|
|
|
|
|
|
+ * adjusted if 'IOVMF_DA_FIXED' is not set.
|
|
*/
|
|
*/
|
|
u32 iommu_kmap(struct iommu *obj, u32 da, u32 pa, size_t bytes,
|
|
u32 iommu_kmap(struct iommu *obj, u32 da, u32 pa, size_t bytes,
|
|
u32 flags)
|
|
u32 flags)
|
|
@@ -799,7 +798,6 @@ u32 iommu_kmap(struct iommu *obj, u32 da, u32 pa, size_t bytes,
|
|
flags &= IOVMF_HW_MASK;
|
|
flags &= IOVMF_HW_MASK;
|
|
flags |= IOVMF_LINEAR;
|
|
flags |= IOVMF_LINEAR;
|
|
flags |= IOVMF_MMIO;
|
|
flags |= IOVMF_MMIO;
|
|
- flags |= (da ? IOVMF_DA_FIXED : IOVMF_DA_ANON);
|
|
|
|
|
|
|
|
da = __iommu_kmap(obj, da, pa, va, bytes, flags);
|
|
da = __iommu_kmap(obj, da, pa, va, bytes, flags);
|
|
if (IS_ERR_VALUE(da))
|
|
if (IS_ERR_VALUE(da))
|
|
@@ -838,7 +836,7 @@ EXPORT_SYMBOL_GPL(iommu_kunmap);
|
|
* @flags: iovma and page property
|
|
* @flags: iovma and page property
|
|
*
|
|
*
|
|
* Allocate @bytes linearly and creates 1-1-1 mapping and returns
|
|
* Allocate @bytes linearly and creates 1-1-1 mapping and returns
|
|
- * @da again, which might be adjusted if 'IOVMF_DA_ANON' is set.
|
|
|
|
|
|
+ * @da again, which might be adjusted if 'IOVMF_DA_FIXED' is not set.
|
|
*/
|
|
*/
|
|
u32 iommu_kmalloc(struct iommu *obj, u32 da, size_t bytes, u32 flags)
|
|
u32 iommu_kmalloc(struct iommu *obj, u32 da, size_t bytes, u32 flags)
|
|
{
|
|
{
|
|
@@ -858,7 +856,6 @@ u32 iommu_kmalloc(struct iommu *obj, u32 da, size_t bytes, u32 flags)
|
|
flags &= IOVMF_HW_MASK;
|
|
flags &= IOVMF_HW_MASK;
|
|
flags |= IOVMF_LINEAR;
|
|
flags |= IOVMF_LINEAR;
|
|
flags |= IOVMF_ALLOC;
|
|
flags |= IOVMF_ALLOC;
|
|
- flags |= (da ? IOVMF_DA_FIXED : IOVMF_DA_ANON);
|
|
|
|
|
|
|
|
da = __iommu_kmap(obj, da, pa, va, bytes, flags);
|
|
da = __iommu_kmap(obj, da, pa, va, bytes, flags);
|
|
if (IS_ERR_VALUE(da))
|
|
if (IS_ERR_VALUE(da))
|