Updated from Linux LTS 3.10.20 to 3.10.21
This commit is contained in:
@@ -103,6 +103,10 @@ int kvm_iommu_map_pages(struct kvm *kvm, struct kvm_memory_slot *slot)
|
||||
while ((gfn << PAGE_SHIFT) & (page_size - 1))
|
||||
page_size >>= 1;
|
||||
|
||||
/* Make sure hva is aligned to the page size we want to map */
|
||||
while (__gfn_to_hva_memslot(slot, gfn) & (page_size - 1))
|
||||
page_size >>= 1;
|
||||
|
||||
/*
|
||||
* Pin all pages we are about to map in memory. This is
|
||||
* important because we unmap and unpin in 4kb steps later.
|
||||
|
||||
Reference in New Issue
Block a user