drm/amdgpu: DMA map/unmap when updating GPU mappings
DMA map kfd_mem_attachments in update_gpuvm_pte. This function is called with the BO and page tables reserved, so we can safely update the DMA mapping. DMA unmap when a BO is unmapped from a GPU and before updating mappings in restore workers. Signed-off-by: Felix Kuehling <Felix.Kuehling@amd.com> Acked-by: Oak Zeng <Oak.Zeng@amd.com> Acked-by: Ramesh Errabolu <Ramesh.Errabolu@amd.com> Signed-off-by: Alex Deucher <alexander.deucher@amd.com>
This commit is contained in:
parent
264fb4d332
commit
b72ed8a2de
@ -966,11 +966,12 @@ static int unreserve_bo_and_vms(struct bo_vm_reservation_context *ctx,
|
|||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int unmap_bo_from_gpuvm(struct amdgpu_device *adev,
|
static void unmap_bo_from_gpuvm(struct kgd_mem *mem,
|
||||||
struct kfd_mem_attachment *entry,
|
struct kfd_mem_attachment *entry,
|
||||||
struct amdgpu_sync *sync)
|
struct amdgpu_sync *sync)
|
||||||
{
|
{
|
||||||
struct amdgpu_bo_va *bo_va = entry->bo_va;
|
struct amdgpu_bo_va *bo_va = entry->bo_va;
|
||||||
|
struct amdgpu_device *adev = entry->adev;
|
||||||
struct amdgpu_vm *vm = bo_va->base.vm;
|
struct amdgpu_vm *vm = bo_va->base.vm;
|
||||||
|
|
||||||
amdgpu_vm_bo_unmap(adev, bo_va, entry->va);
|
amdgpu_vm_bo_unmap(adev, bo_va, entry->va);
|
||||||
@ -979,15 +980,20 @@ static int unmap_bo_from_gpuvm(struct amdgpu_device *adev,
|
|||||||
|
|
||||||
amdgpu_sync_fence(sync, bo_va->last_pt_update);
|
amdgpu_sync_fence(sync, bo_va->last_pt_update);
|
||||||
|
|
||||||
return 0;
|
kfd_mem_dmaunmap_attachment(mem, entry);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int update_gpuvm_pte(struct amdgpu_device *adev,
|
static int update_gpuvm_pte(struct kgd_mem *mem,
|
||||||
struct kfd_mem_attachment *entry,
|
struct kfd_mem_attachment *entry,
|
||||||
struct amdgpu_sync *sync)
|
struct amdgpu_sync *sync)
|
||||||
{
|
{
|
||||||
int ret;
|
|
||||||
struct amdgpu_bo_va *bo_va = entry->bo_va;
|
struct amdgpu_bo_va *bo_va = entry->bo_va;
|
||||||
|
struct amdgpu_device *adev = entry->adev;
|
||||||
|
int ret;
|
||||||
|
|
||||||
|
ret = kfd_mem_dmamap_attachment(mem, entry);
|
||||||
|
if (ret)
|
||||||
|
return ret;
|
||||||
|
|
||||||
/* Update the page tables */
|
/* Update the page tables */
|
||||||
ret = amdgpu_vm_bo_update(adev, bo_va, false);
|
ret = amdgpu_vm_bo_update(adev, bo_va, false);
|
||||||
@ -999,14 +1005,15 @@ static int update_gpuvm_pte(struct amdgpu_device *adev,
|
|||||||
return amdgpu_sync_fence(sync, bo_va->last_pt_update);
|
return amdgpu_sync_fence(sync, bo_va->last_pt_update);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int map_bo_to_gpuvm(struct amdgpu_device *adev,
|
static int map_bo_to_gpuvm(struct kgd_mem *mem,
|
||||||
struct kfd_mem_attachment *entry, struct amdgpu_sync *sync,
|
struct kfd_mem_attachment *entry,
|
||||||
bool no_update_pte)
|
struct amdgpu_sync *sync,
|
||||||
|
bool no_update_pte)
|
||||||
{
|
{
|
||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
/* Set virtual address for the allocation */
|
/* Set virtual address for the allocation */
|
||||||
ret = amdgpu_vm_bo_map(adev, entry->bo_va, entry->va, 0,
|
ret = amdgpu_vm_bo_map(entry->adev, entry->bo_va, entry->va, 0,
|
||||||
amdgpu_bo_size(entry->bo_va->base.bo),
|
amdgpu_bo_size(entry->bo_va->base.bo),
|
||||||
entry->pte_flags);
|
entry->pte_flags);
|
||||||
if (ret) {
|
if (ret) {
|
||||||
@ -1018,7 +1025,7 @@ static int map_bo_to_gpuvm(struct amdgpu_device *adev,
|
|||||||
if (no_update_pte)
|
if (no_update_pte)
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
ret = update_gpuvm_pte(adev, entry, sync);
|
ret = update_gpuvm_pte(mem, entry, sync);
|
||||||
if (ret) {
|
if (ret) {
|
||||||
pr_err("update_gpuvm_pte() failed\n");
|
pr_err("update_gpuvm_pte() failed\n");
|
||||||
goto update_gpuvm_pte_failed;
|
goto update_gpuvm_pte_failed;
|
||||||
@ -1027,7 +1034,7 @@ static int map_bo_to_gpuvm(struct amdgpu_device *adev,
|
|||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
update_gpuvm_pte_failed:
|
update_gpuvm_pte_failed:
|
||||||
unmap_bo_from_gpuvm(adev, entry, sync);
|
unmap_bo_from_gpuvm(mem, entry, sync);
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1601,7 +1608,7 @@ int amdgpu_amdkfd_gpuvm_map_memory_to_gpu(
|
|||||||
pr_debug("\t map VA 0x%llx - 0x%llx in entry %p\n",
|
pr_debug("\t map VA 0x%llx - 0x%llx in entry %p\n",
|
||||||
entry->va, entry->va + bo_size, entry);
|
entry->va, entry->va + bo_size, entry);
|
||||||
|
|
||||||
ret = map_bo_to_gpuvm(adev, entry, ctx.sync,
|
ret = map_bo_to_gpuvm(mem, entry, ctx.sync,
|
||||||
is_invalid_userptr);
|
is_invalid_userptr);
|
||||||
if (ret) {
|
if (ret) {
|
||||||
pr_err("Failed to map bo to gpuvm\n");
|
pr_err("Failed to map bo to gpuvm\n");
|
||||||
@ -1640,7 +1647,6 @@ out:
|
|||||||
int amdgpu_amdkfd_gpuvm_unmap_memory_from_gpu(
|
int amdgpu_amdkfd_gpuvm_unmap_memory_from_gpu(
|
||||||
struct kgd_dev *kgd, struct kgd_mem *mem, void *drm_priv)
|
struct kgd_dev *kgd, struct kgd_mem *mem, void *drm_priv)
|
||||||
{
|
{
|
||||||
struct amdgpu_device *adev = get_amdgpu_device(kgd);
|
|
||||||
struct amdgpu_vm *avm = drm_priv_to_vm(drm_priv);
|
struct amdgpu_vm *avm = drm_priv_to_vm(drm_priv);
|
||||||
struct amdkfd_process_info *process_info = avm->process_info;
|
struct amdkfd_process_info *process_info = avm->process_info;
|
||||||
unsigned long bo_size = mem->bo->tbo.base.size;
|
unsigned long bo_size = mem->bo->tbo.base.size;
|
||||||
@ -1675,13 +1681,8 @@ int amdgpu_amdkfd_gpuvm_unmap_memory_from_gpu(
|
|||||||
pr_debug("\t unmap VA 0x%llx - 0x%llx from entry %p\n",
|
pr_debug("\t unmap VA 0x%llx - 0x%llx from entry %p\n",
|
||||||
entry->va, entry->va + bo_size, entry);
|
entry->va, entry->va + bo_size, entry);
|
||||||
|
|
||||||
ret = unmap_bo_from_gpuvm(adev, entry, ctx.sync);
|
unmap_bo_from_gpuvm(mem, entry, ctx.sync);
|
||||||
if (ret == 0) {
|
entry->is_mapped = false;
|
||||||
entry->is_mapped = false;
|
|
||||||
} else {
|
|
||||||
pr_err("failed to unmap VA 0x%llx\n", mem->va);
|
|
||||||
goto unreserve_out;
|
|
||||||
}
|
|
||||||
|
|
||||||
mem->mapped_to_gpu_memory--;
|
mem->mapped_to_gpu_memory--;
|
||||||
pr_debug("\t DEC mapping count %d\n",
|
pr_debug("\t DEC mapping count %d\n",
|
||||||
@ -2058,9 +2059,8 @@ static int validate_invalid_user_pages(struct amdkfd_process_info *process_info)
|
|||||||
if (!attachment->is_mapped)
|
if (!attachment->is_mapped)
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
ret = update_gpuvm_pte((struct amdgpu_device *)
|
kfd_mem_dmaunmap_attachment(mem, attachment);
|
||||||
attachment->adev,
|
ret = update_gpuvm_pte(mem, attachment, &sync);
|
||||||
attachment, &sync);
|
|
||||||
if (ret) {
|
if (ret) {
|
||||||
pr_err("%s: update PTE failed\n", __func__);
|
pr_err("%s: update PTE failed\n", __func__);
|
||||||
/* make sure this gets validated again */
|
/* make sure this gets validated again */
|
||||||
@ -2262,9 +2262,11 @@ int amdgpu_amdkfd_gpuvm_restore_process_bos(void *info, struct dma_fence **ef)
|
|||||||
goto validate_map_fail;
|
goto validate_map_fail;
|
||||||
}
|
}
|
||||||
list_for_each_entry(attachment, &mem->attachments, list) {
|
list_for_each_entry(attachment, &mem->attachments, list) {
|
||||||
ret = update_gpuvm_pte((struct amdgpu_device *)
|
if (!attachment->is_mapped)
|
||||||
attachment->adev, attachment,
|
continue;
|
||||||
&sync_obj);
|
|
||||||
|
kfd_mem_dmaunmap_attachment(mem, attachment);
|
||||||
|
ret = update_gpuvm_pte(mem, attachment, &sync_obj);
|
||||||
if (ret) {
|
if (ret) {
|
||||||
pr_debug("Memory eviction: update PTE failed. Try again\n");
|
pr_debug("Memory eviction: update PTE failed. Try again\n");
|
||||||
goto validate_map_fail;
|
goto validate_map_fail;
|
||||||
|
Loading…
Reference in New Issue
Block a user