The movtivation is we want to unify CPU VA and GPU VA.
Signed-off-by: Lang Yu <[email protected]>
---
drivers/hv/dxgkrnl/dxgvmbus.c | 24 ++++++++++++++----------
1 file changed, 14 insertions(+), 10 deletions(-)
diff --git a/drivers/hv/dxgkrnl/dxgvmbus.c b/drivers/hv/dxgkrnl/dxgvmbus.c
index 9320bede3a0a..a4bca27a7cc8 100644
--- a/drivers/hv/dxgkrnl/dxgvmbus.c
+++ b/drivers/hv/dxgkrnl/dxgvmbus.c
@@ -580,7 +580,7 @@ int dxg_unmap_iospace(void *va, u32 size)
return 0;
}
-static u8 *dxg_map_iospace(u64 iospace_address, u32 size,
+static u8 *dxg_map_iospace(u64 iospace_address, u64 user_va, u32 size,
unsigned long protection, bool cached)
{
struct vm_area_struct *vma;
@@ -594,7 +594,12 @@ static u8 *dxg_map_iospace(u64 iospace_address, u32 size,
return NULL;
}
- va = vm_mmap(NULL, 0, size, protection, MAP_SHARED | MAP_ANONYMOUS, 0);
+ if (user_va)
+ va = vm_mmap(untagged_addr(user_va), 0, size, protection,
+ MAP_SHARED | MAP_ANONYMOUS | MAP_FIXED, 0);
+ else
+ va = vm_mmap(NULL, 0, size, protection,
+ MAP_SHARED | MAP_ANONYMOUS, 0);
if ((long)va <= 0) {
DXG_ERR("vm_mmap failed %lx %d", va, size);
return NULL;
@@ -789,9 +794,8 @@ int dxgvmb_send_open_sync_object_nt(struct dxgprocess
*process,
args->sync_object = result.sync_object;
if (syncobj->monitored_fence) {
- void *va = dxg_map_iospace(result.guest_cpu_physical_address,
- PAGE_SIZE, PROT_READ | PROT_WRITE,
- true);
+ void *va = dxg_map_iospace(result.guest_cpu_physical_address, 0,
+ PAGE_SIZE, PROT_READ | PROT_WRITE,
true);
if (va == NULL) {
ret = -ENOMEM;
goto cleanup;
@@ -1315,8 +1319,8 @@ int dxgvmb_send_create_paging_queue(struct dxgprocess
*process,
args->paging_queue = result.paging_queue;
args->sync_object = result.sync_object;
args->fence_cpu_virtual_address =
- dxg_map_iospace(result.fence_storage_physical_address, PAGE_SIZE,
- PROT_READ | PROT_WRITE, true);
+ dxg_map_iospace(result.fence_storage_physical_address, 0,
+ PAGE_SIZE, PROT_READ | PROT_WRITE, true);
if (args->fence_cpu_virtual_address == NULL) {
ret = -ENOMEM;
goto cleanup;
@@ -2867,7 +2871,7 @@ dxgvmb_send_create_sync_object(struct dxgprocess *process,
}
if (syncobj->monitored_fence) {
- va = dxg_map_iospace(result.fence_storage_address, PAGE_SIZE,
+ va = dxg_map_iospace(result.fence_storage_address, 0, PAGE_SIZE,
PROT_READ | PROT_WRITE, true);
if (va == NULL) {
ret = -ENOMEM;
@@ -3156,7 +3160,7 @@ int dxgvmb_send_lock2(struct dxgprocess *process,
} else {
u64 offset = (u64)result.cpu_visible_buffer_offset;
- args->data = dxg_map_iospace(offset,
+ args->data = dxg_map_iospace(offset, args->data,
alloc->num_pages << PAGE_SHIFT,
PROT_READ | PROT_WRITE, alloc->cached);
if (args->data) {
@@ -3712,7 +3716,7 @@ int dxgvmb_send_create_hwqueue(struct dxgprocess *process,
}
hwqueue->progress_fence_mapped_address =
- dxg_map_iospace((u64)command->hwqueue_progress_fence_cpuva,
+ dxg_map_iospace((u64)command->hwqueue_progress_fence_cpuva, 0,
PAGE_SIZE, PROT_READ | PROT_WRITE, true);
if (hwqueue->progress_fence_mapped_address == NULL) {
ret = -ENOMEM;
--
2.25.1