|
@@ -134,14 +134,15 @@ unmap_pages:
|
|
|
|
|
|
static int kvm_iommu_map_memslots(struct kvm *kvm)
|
|
|
{
|
|
|
- int i, idx, r = 0;
|
|
|
+ int idx, r = 0;
|
|
|
struct kvm_memslots *slots;
|
|
|
+ struct kvm_memory_slot *memslot;
|
|
|
|
|
|
idx = srcu_read_lock(&kvm->srcu);
|
|
|
slots = kvm_memslots(kvm);
|
|
|
|
|
|
- for (i = 0; i < slots->nmemslots; i++) {
|
|
|
- r = kvm_iommu_map_pages(kvm, &slots->memslots[i]);
|
|
|
+ kvm_for_each_memslot(memslot, slots) {
|
|
|
+ r = kvm_iommu_map_pages(kvm, memslot);
|
|
|
if (r)
|
|
|
break;
|
|
|
}
|
|
@@ -311,16 +312,16 @@ static void kvm_iommu_put_pages(struct kvm *kvm,
|
|
|
|
|
|
static int kvm_iommu_unmap_memslots(struct kvm *kvm)
|
|
|
{
|
|
|
- int i, idx;
|
|
|
+ int idx;
|
|
|
struct kvm_memslots *slots;
|
|
|
+ struct kvm_memory_slot *memslot;
|
|
|
|
|
|
idx = srcu_read_lock(&kvm->srcu);
|
|
|
slots = kvm_memslots(kvm);
|
|
|
|
|
|
- for (i = 0; i < slots->nmemslots; i++) {
|
|
|
- kvm_iommu_put_pages(kvm, slots->memslots[i].base_gfn,
|
|
|
- slots->memslots[i].npages);
|
|
|
- }
|
|
|
+ kvm_for_each_memslot(memslot, slots)
|
|
|
+ kvm_iommu_put_pages(kvm, memslot->base_gfn, memslot->npages);
|
|
|
+
|
|
|
srcu_read_unlock(&kvm->srcu, idx);
|
|
|
|
|
|
return 0;
|