Lines Matching refs:ib

300 				   struct amdgpu_ib *ib,  in amdgpu_vm_update_pages()  argument
309 amdgpu_vm_copy_pte(adev, ib, pe, src, count); in amdgpu_vm_update_pages()
312 amdgpu_vm_write_pte(adev, ib, pe, addr, in amdgpu_vm_update_pages()
316 amdgpu_vm_set_pte_pde(adev, ib, pe, addr, in amdgpu_vm_update_pages()
343 struct amdgpu_ib *ib; in amdgpu_vm_clear_bo() local
359 ib = kzalloc(sizeof(struct amdgpu_ib), GFP_KERNEL); in amdgpu_vm_clear_bo()
360 if (!ib) in amdgpu_vm_clear_bo()
363 r = amdgpu_ib_get(ring, NULL, entries * 2 + 64, ib); in amdgpu_vm_clear_bo()
367 ib->length_dw = 0; in amdgpu_vm_clear_bo()
369 amdgpu_vm_update_pages(adev, ib, addr, 0, entries, 0, 0, 0); in amdgpu_vm_clear_bo()
370 amdgpu_vm_pad_ib(adev, ib); in amdgpu_vm_clear_bo()
371 WARN_ON(ib->length_dw > 64); in amdgpu_vm_clear_bo()
372 r = amdgpu_sched_ib_submit_kernel_helper(adev, ring, ib, 1, in amdgpu_vm_clear_bo()
383 amdgpu_ib_free(adev, ib); in amdgpu_vm_clear_bo()
384 kfree(ib); in amdgpu_vm_clear_bo()
436 struct amdgpu_ib *ib; in amdgpu_vm_update_page_directory() local
451 ib = kzalloc(sizeof(struct amdgpu_ib), GFP_KERNEL); in amdgpu_vm_update_page_directory()
452 if (!ib) in amdgpu_vm_update_page_directory()
455 r = amdgpu_ib_get(ring, NULL, ndw * 4, ib); in amdgpu_vm_update_page_directory()
457 kfree(ib); in amdgpu_vm_update_page_directory()
460 ib->length_dw = 0; in amdgpu_vm_update_page_directory()
480 amdgpu_vm_update_pages(adev, ib, last_pde, in amdgpu_vm_update_page_directory()
494 amdgpu_vm_update_pages(adev, ib, last_pde, last_pt, count, in amdgpu_vm_update_page_directory()
497 if (ib->length_dw != 0) { in amdgpu_vm_update_page_directory()
498 amdgpu_vm_pad_ib(adev, ib); in amdgpu_vm_update_page_directory()
499 amdgpu_sync_resv(adev, &ib->sync, pd->tbo.resv, AMDGPU_FENCE_OWNER_VM); in amdgpu_vm_update_page_directory()
500 WARN_ON(ib->length_dw > ndw); in amdgpu_vm_update_page_directory()
501 r = amdgpu_sched_ib_submit_kernel_helper(adev, ring, ib, 1, in amdgpu_vm_update_page_directory()
514 if (!amdgpu_enable_scheduler || ib->length_dw == 0) { in amdgpu_vm_update_page_directory()
515 amdgpu_ib_free(adev, ib); in amdgpu_vm_update_page_directory()
516 kfree(ib); in amdgpu_vm_update_page_directory()
522 amdgpu_ib_free(adev, ib); in amdgpu_vm_update_page_directory()
523 kfree(ib); in amdgpu_vm_update_page_directory()
541 struct amdgpu_ib *ib, in amdgpu_vm_frag_ptes() argument
579 amdgpu_vm_update_pages(adev, ib, pe_start, addr, count, in amdgpu_vm_frag_ptes()
587 amdgpu_vm_update_pages(adev, ib, pe_start, addr, count, in amdgpu_vm_frag_ptes()
594 amdgpu_vm_update_pages(adev, ib, frag_start, addr, count, in amdgpu_vm_frag_ptes()
602 amdgpu_vm_update_pages(adev, ib, frag_end, addr, count, in amdgpu_vm_frag_ptes()
623 struct amdgpu_ib *ib, in amdgpu_vm_update_ptes() argument
646 amdgpu_sync_resv(adev, &ib->sync, pt->tbo.resv, owner); in amdgpu_vm_update_ptes()
662 amdgpu_vm_frag_ptes(adev, ib, last_pte, in amdgpu_vm_update_ptes()
680 amdgpu_vm_frag_ptes(adev, ib, last_pte, in amdgpu_vm_update_ptes()
712 struct amdgpu_ib *ib; in amdgpu_vm_bo_update_mapping() local
760 ib = kzalloc(sizeof(struct amdgpu_ib), GFP_KERNEL); in amdgpu_vm_bo_update_mapping()
761 if (!ib) in amdgpu_vm_bo_update_mapping()
764 r = amdgpu_ib_get(ring, NULL, ndw * 4, ib); in amdgpu_vm_bo_update_mapping()
766 kfree(ib); in amdgpu_vm_bo_update_mapping()
770 ib->length_dw = 0; in amdgpu_vm_bo_update_mapping()
772 r = amdgpu_vm_update_ptes(adev, vm, ib, mapping->it.start, in amdgpu_vm_bo_update_mapping()
777 amdgpu_ib_free(adev, ib); in amdgpu_vm_bo_update_mapping()
778 kfree(ib); in amdgpu_vm_bo_update_mapping()
782 amdgpu_vm_pad_ib(adev, ib); in amdgpu_vm_bo_update_mapping()
783 WARN_ON(ib->length_dw > ndw); in amdgpu_vm_bo_update_mapping()
784 r = amdgpu_sched_ib_submit_kernel_helper(adev, ring, ib, 1, in amdgpu_vm_bo_update_mapping()
798 amdgpu_ib_free(adev, ib); in amdgpu_vm_bo_update_mapping()
799 kfree(ib); in amdgpu_vm_bo_update_mapping()
804 amdgpu_ib_free(adev, ib); in amdgpu_vm_bo_update_mapping()
805 kfree(ib); in amdgpu_vm_bo_update_mapping()