/openbmc/linux/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_gem.c | 92 if (robj) { in amdgpu_gem_object_free() 93 amdgpu_hmm_unregister(robj); in amdgpu_gem_object_free() 94 amdgpu_bo_unref(&robj); in amdgpu_gem_object_free() 464 struct amdgpu_bo *robj; in amdgpu_mode_dumb_mmap() local 524 struct amdgpu_bo *robj; in amdgpu_gem_wait_idle_ioctl() local 557 struct amdgpu_bo *robj; in amdgpu_gem_metadata_ioctl() local 589 amdgpu_bo_unreserve(robj); in amdgpu_gem_metadata_ioctl() 811 struct amdgpu_bo *robj; in amdgpu_gem_op_ioctl() local 833 amdgpu_bo_unreserve(robj); in amdgpu_gem_op_ioctl() 862 robj->allowed_domains = robj->preferred_domains; in amdgpu_gem_op_ioctl() [all …]
|
H A D | amdgpu_display.c | 1649 amdgpu_display_robj_is_fb(struct amdgpu_device *adev, struct amdgpu_bo *robj) in amdgpu_display_robj_is_fb() argument 1657 if (gem_to_amdgpu_bo(fb_helper->buffer->gem) != robj) in amdgpu_display_robj_is_fb() 1685 struct amdgpu_bo *robj; in amdgpu_display_suspend_helper() local 1700 robj = gem_to_amdgpu_bo(fb->obj[0]); in amdgpu_display_suspend_helper() 1701 if (!amdgpu_display_robj_is_fb(adev, robj)) { in amdgpu_display_suspend_helper() 1702 r = amdgpu_bo_reserve(robj, true); in amdgpu_display_suspend_helper() 1704 amdgpu_bo_unpin(robj); in amdgpu_display_suspend_helper() 1705 amdgpu_bo_unreserve(robj); in amdgpu_display_suspend_helper()
|
/openbmc/linux/drivers/gpu/drm/radeon/ |
H A D | radeon_gem.c | 88 if (robj) { in radeon_gem_object_free() 90 radeon_bo_unref(&robj); in radeon_gem_object_free() 99 struct radeon_bo *robj; in radeon_gem_object_create() local 133 *obj = &robj->tbo.base; in radeon_gem_object_create() 147 struct radeon_bo *robj; in radeon_gem_set_domain() local 472 struct radeon_bo *robj; in radeon_mode_dumb_mmap() local 479 if (radeon_ttm_tt_has_userptr(robj->rdev, robj->tbo.ttm)) { in radeon_mode_dumb_mmap() 501 struct radeon_bo *robj; in radeon_gem_busy_ioctl() local 529 struct radeon_bo *robj; in radeon_gem_wait_idle_ioctl() local 562 struct radeon_bo *robj; in radeon_gem_set_tiling_ioctl() local [all …]
|
H A D | radeon_gart.c | 132 if (rdev->gart.robj == NULL) { in radeon_gart_table_vram_alloc() 160 r = radeon_bo_pin(rdev->gart.robj, in radeon_gart_table_vram_pin() 163 radeon_bo_unreserve(rdev->gart.robj); in radeon_gart_table_vram_pin() 168 radeon_bo_unpin(rdev->gart.robj); in radeon_gart_table_vram_pin() 169 radeon_bo_unreserve(rdev->gart.robj); in radeon_gart_table_vram_pin() 199 if (!rdev->gart.robj) in radeon_gart_table_vram_unpin() 204 radeon_bo_kunmap(rdev->gart.robj); in radeon_gart_table_vram_unpin() 205 radeon_bo_unpin(rdev->gart.robj); in radeon_gart_table_vram_unpin() 206 radeon_bo_unreserve(rdev->gart.robj); in radeon_gart_table_vram_unpin() 222 if (!rdev->gart.robj) in radeon_gart_table_vram_free() [all …]
|
H A D | r100_track.h | 16 struct radeon_bo *robj; member 23 struct radeon_bo *robj; member 28 struct radeon_bo *robj; member 39 struct radeon_bo *robj; member
|
H A D | evergreen_cs.c | 1221 track->db_z_read_bo = reloc->robj; in evergreen_cs_handle_reg() 1233 track->db_z_write_bo = reloc->robj; in evergreen_cs_handle_reg() 1245 track->db_s_read_bo = reloc->robj; in evergreen_cs_handle_reg() 1257 track->db_s_write_bo = reloc->robj; in evergreen_cs_handle_reg() 1281 track->vgt_strmout_bo[tmp] = reloc->robj; in evergreen_cs_handle_reg() 1505 track->cb_color_fmask_bo[tmp] = reloc->robj; in evergreen_cs_handle_reg() 1563 track->cb_color_bo[tmp] = reloc->robj; in evergreen_cs_handle_reg() 1579 track->cb_color_bo[tmp] = reloc->robj; in evergreen_cs_handle_reg() 1591 track->htile_bo = reloc->robj; in evergreen_cs_handle_reg() 2376 texture = reloc->robj; in evergreen_packet3_check() [all …]
|
H A D | radeon_cursor.c | 287 struct radeon_bo *robj; in radeon_crtc_cursor_set2() local 309 robj = gem_to_radeon_bo(obj); in radeon_crtc_cursor_set2() 310 ret = radeon_bo_reserve(robj, false); in radeon_crtc_cursor_set2() 316 ret = radeon_bo_pin_restricted(robj, RADEON_GEM_DOMAIN_VRAM, in radeon_crtc_cursor_set2() 319 radeon_bo_unreserve(robj); in radeon_crtc_cursor_set2() 351 struct radeon_bo *robj = gem_to_radeon_bo(radeon_crtc->cursor_bo); in radeon_crtc_cursor_set2() local 352 ret = radeon_bo_reserve(robj, false); in radeon_crtc_cursor_set2() 354 radeon_bo_unpin(robj); in radeon_crtc_cursor_set2() 355 radeon_bo_unreserve(robj); in radeon_crtc_cursor_set2()
|
H A D | r600_cs.c | 1083 track->vgt_strmout_bo[tmp] = reloc->robj; in r600_cs_check_reg() 1210 track->cb_color_frag_bo[tmp] = reloc->robj; in r600_cs_check_reg() 1241 track->cb_color_tile_bo[tmp] = reloc->robj; in r600_cs_check_reg() 1281 track->cb_color_bo[tmp] = reloc->robj; in r600_cs_check_reg() 1294 track->db_bo = reloc->robj; in r600_cs_check_reg() 1307 track->htile_bo = reloc->robj; in r600_cs_check_reg() 1810 tmp + size, radeon_bo_size(reloc->robj)); in r600_packet3_check() 1840 tmp + size, radeon_bo_size(reloc->robj)); in r600_packet3_check() 1972 texture = reloc->robj; in r600_packet3_check() 1980 mipmap = reloc->robj; in r600_packet3_check() [all …]
|
H A D | r200.c | 188 track->zb.robj = reloc->robj; in r200_packet0_check() 201 track->cb[0].robj = reloc->robj; in r200_packet0_check() 231 track->textures[i].robj = reloc->robj; in r200_packet0_check() 275 track->textures[i].cube_info[face - 1].robj = reloc->robj; in r200_packet0_check()
|
H A D | r100.c | 1342 track->arrays[i + 0].robj = reloc->robj; in r100_packet3_load_vbpntr() 1352 track->arrays[i + 1].robj = reloc->robj; in r100_packet3_load_vbpntr() 1366 track->arrays[i + 0].robj = reloc->robj; in r100_packet3_load_vbpntr() 1603 track->zb.robj = reloc->robj; in r100_packet0_check() 1616 track->cb[0].robj = reloc->robj; in r100_packet0_check() 1643 track->textures[i].robj = reloc->robj; in r100_packet0_check() 1661 track->textures[0].cube_info[i].robj = reloc->robj; in r100_packet0_check() 1679 track->textures[1].cube_info[i].robj = reloc->robj; in r100_packet0_check() 1697 track->textures[2].cube_info[i].robj = reloc->robj; in r100_packet0_check() 1961 track->arrays[0].robj = reloc->robj; in r100_packet3_check() [all …]
|
H A D | radeon_cs.c | 116 p->relocs[i].robj = gem_to_radeon_bo(gobj); in radeon_cs_parser_relocs() 162 if (radeon_ttm_tt_has_userptr(p->rdev, p->relocs[i].robj->tbo.ttm)) { in radeon_cs_parser_relocs() 176 if (p->relocs[i].robj->prime_shared_count) { in radeon_cs_parser_relocs() 185 p->relocs[i].tv.bo = &p->relocs[i].robj->tbo; in radeon_cs_parser_relocs() 259 resv = reloc->robj->tbo.base.resv; in radeon_cs_sync_rings() 404 if (la->robj->tbo.base.size > lb->robj->tbo.base.size) in cmp_size_smaller_first() 406 if (la->robj->tbo.base.size < lb->robj->tbo.base.size) in cmp_size_smaller_first() 447 struct radeon_bo *bo = parser->relocs[i].robj; in radeon_cs_parser_fini() 530 bo = p->relocs[i].robj; in radeon_bo_vm_update_pte()
|
H A D | radeon_device.c | 1579 struct radeon_bo *robj; in radeon_suspend_kms() local 1583 r = radeon_bo_reserve(robj, false); in radeon_suspend_kms() 1585 radeon_bo_unpin(robj); in radeon_suspend_kms() 1586 radeon_bo_unreserve(robj); in radeon_suspend_kms() 1593 robj = gem_to_radeon_bo(fb->obj[0]); in radeon_suspend_kms() 1596 r = radeon_bo_reserve(robj, false); in radeon_suspend_kms() 1598 radeon_bo_unpin(robj); in radeon_suspend_kms() 1599 radeon_bo_unreserve(robj); in radeon_suspend_kms() 1706 r = radeon_bo_reserve(robj, false); in radeon_resume_kms() 1709 r = radeon_bo_pin_restricted(robj, in radeon_resume_kms() [all …]
|
H A D | r300.c | 133 if (rdev->gart.robj) { in rv370_pcie_gart_init() 156 if (rdev->gart.robj == NULL) { in rv370_pcie_gart_enable() 670 track->cb[i].robj = reloc->robj; in r300_packet0_check() 683 track->zb.robj = reloc->robj; in r300_packet0_check() 728 track->textures[i].robj = reloc->robj; in r300_packet0_check() 1127 track->aa.robj = reloc->robj; in r300_packet0_check() 1197 r = r100_cs_track_check_pkt3_indx_buffer(p, pkt, reloc->robj); in r300_packet3_check()
|
H A D | radeon_fbdev.c | 401 bool radeon_fbdev_robj_is_fb(struct radeon_device *rdev, struct radeon_bo *robj) in radeon_fbdev_robj_is_fb() argument 412 if (gobj != &robj->tbo.base) in radeon_fbdev_robj_is_fb()
|
H A D | radeon_object.c | 488 struct radeon_bo *bo = lobj->robj; in radeon_bo_list_validate() 535 lobj->gpu_offset = radeon_bo_gpu_offset(lobj->robj); in radeon_bo_list_validate() 536 lobj->tiling_flags = lobj->robj->tiling_flags; in radeon_bo_list_validate()
|
H A D | r600.c | 1116 if (rdev->gart.robj) { in r600_pcie_gart_init() 1133 if (rdev->gart.robj == NULL) { in r600_pcie_gart_enable() 1511 if (rdev->vram_scratch.robj == NULL) { in r600_vram_scratch_init() 1523 r = radeon_bo_pin(rdev->vram_scratch.robj, in r600_vram_scratch_init() 1529 r = radeon_bo_kmap(rdev->vram_scratch.robj, in r600_vram_scratch_init() 1532 radeon_bo_unpin(rdev->vram_scratch.robj); in r600_vram_scratch_init() 1533 radeon_bo_unreserve(rdev->vram_scratch.robj); in r600_vram_scratch_init() 1542 if (rdev->vram_scratch.robj == NULL) { in r600_vram_scratch_fini() 1547 radeon_bo_kunmap(rdev->vram_scratch.robj); in r600_vram_scratch_fini() 1548 radeon_bo_unpin(rdev->vram_scratch.robj); in r600_vram_scratch_fini() [all …]
|
H A D | radeon_vm.c | 142 list[0].robj = vm->page_directory; in radeon_vm_get_bos() 154 list[idx].robj = vm->page_tables[i].bo; in radeon_vm_get_bos() 157 list[idx].tv.bo = &list[idx].robj->tbo; in radeon_vm_get_bos()
|
H A D | radeon_mode.h | 945 bool radeon_fbdev_robj_is_fb(struct radeon_device *rdev, struct radeon_bo *robj); 951 static inline bool radeon_fbdev_robj_is_fb(struct radeon_device *rdev, struct radeon_bo *robj) in radeon_fbdev_robj_is_fb() argument
|
H A D | radeon_uvd.c | 583 end = start + radeon_bo_size(reloc->robj); in radeon_uvd_cs_reloc() 627 r = radeon_uvd_cs_msg(p, reloc->robj, offset, buf_sizes); in radeon_uvd_cs_reloc()
|
H A D | rs600.c | 550 if (rdev->gart.robj) { in rs600_gart_init() 568 if (rdev->gart.robj == NULL) { in rs600_gart_enable()
|
/openbmc/linux/drivers/gpu/drm/nouveau/ |
H A D | nouveau_prime.c | 46 struct dma_resv *robj = attach->dmabuf->resv; in nouveau_gem_prime_import_sg_table() local 51 dma_resv_lock(robj, NULL); in nouveau_gem_prime_import_sg_table() 74 sg, robj); in nouveau_gem_prime_import_sg_table() 83 dma_resv_unlock(robj); in nouveau_gem_prime_import_sg_table()
|
H A D | nouveau_bo.h | 79 struct sg_table *sg, struct dma_resv *robj); 82 struct dma_resv *robj,
|
H A D | nouveau_bo.c | 336 struct sg_table *sg, struct dma_resv *robj) in nouveau_bo_init() argument 343 .resv = robj, in nouveau_bo_init() 351 sg, robj, nouveau_bo_del_ttm); in nouveau_bo_init() 357 if (!robj) in nouveau_bo_init() 366 struct sg_table *sg, struct dma_resv *robj, in nouveau_bo_new() argument 386 ret = nouveau_bo_init(nvbo, size, align, domain, sg, robj); in nouveau_bo_new()
|
/openbmc/linux/drivers/gpu/drm/etnaviv/ |
H A D | etnaviv_gem.c | 437 struct dma_resv *robj = obj->resv; in etnaviv_gem_describe() local 446 r = dma_resv_lock(robj, NULL); in etnaviv_gem_describe() 450 dma_resv_describe(robj, m); in etnaviv_gem_describe() 451 dma_resv_unlock(robj); in etnaviv_gem_describe()
|
H A D | etnaviv_gem_submit.c | 180 struct dma_resv *robj = bo->obj->base.resv; in submit_fence_sync() local 182 ret = dma_resv_reserve_fences(robj, 1); in submit_fence_sync()
|