/openbmc/linux/drivers/gpu/drm/nouveau/ |
H A D | nouveau_bo.c | 141 struct nouveau_bo *nvbo = nouveau_bo(bo); in nouveau_bo_del_ttm() local 143 WARN_ON(nvbo->bo.pin_count > 0); in nouveau_bo_del_ttm() 145 nv10_bo_put_tile_region(dev, nvbo->tile, NULL); in nouveau_bo_del_ttm() 156 kfree(nvbo); in nouveau_bo_del_ttm() 168 nouveau_bo_fixup_align(struct nouveau_bo *nvbo, int *align, u64 *size) in nouveau_bo_fixup_align() argument 170 struct nouveau_drm *drm = nouveau_bdev(nvbo->bo.bdev); in nouveau_bo_fixup_align() 174 if (nvbo->mode) { in nouveau_bo_fixup_align() 177 *size = roundup_64(*size, 64 * nvbo->mode); in nouveau_bo_fixup_align() 181 *size = roundup_64(*size, 64 * nvbo->mode); in nouveau_bo_fixup_align() 185 *size = roundup_64(*size, 64 * nvbo->mode); in nouveau_bo_fixup_align() [all …]
|
H A D | nouveau_gem.c | 79 struct nouveau_bo *nvbo = nouveau_gem_object(gem); in nouveau_gem_object_del() local 80 struct nouveau_drm *drm = nouveau_bdev(nvbo->bo.bdev); in nouveau_gem_object_del() 91 drm_prime_gem_destroy(gem, nvbo->bo.sg); in nouveau_gem_object_del() 93 ttm_bo_put(&nvbo->bo); in nouveau_gem_object_del() 103 struct nouveau_bo *nvbo = nouveau_gem_object(gem); in nouveau_gem_object_open() local 104 struct nouveau_drm *drm = nouveau_bdev(nvbo->bo.bdev); in nouveau_gem_object_open() 114 if (nvbo->no_share && uvmm && &uvmm->resv != nvbo->bo.base.resv) in nouveau_gem_object_open() 117 ret = ttm_bo_reserve(&nvbo->bo, false, false, NULL); in nouveau_gem_object_open() 129 ret = nouveau_vma_new(nvbo, vmm, &vma); in nouveau_gem_object_open() 135 ttm_bo_unreserve(&nvbo->bo); in nouveau_gem_object_open() [all …]
|
H A D | nouveau_prime.c | 33 struct nouveau_bo *nvbo = nouveau_gem_object(obj); in nouveau_gem_prime_get_sg_table() local 35 return drm_prime_pages_to_sg(obj->dev, nvbo->bo.ttm->pages, in nouveau_gem_prime_get_sg_table() 36 nvbo->bo.ttm->num_pages); in nouveau_gem_prime_get_sg_table() 45 struct nouveau_bo *nvbo; in nouveau_gem_prime_import_sg_table() local 52 nvbo = nouveau_bo_alloc(&drm->client, &size, &align, in nouveau_gem_prime_import_sg_table() 54 if (IS_ERR(nvbo)) { in nouveau_gem_prime_import_sg_table() 55 obj = ERR_CAST(nvbo); in nouveau_gem_prime_import_sg_table() 59 nvbo->valid_domains = NOUVEAU_GEM_DOMAIN_GART; in nouveau_gem_prime_import_sg_table() 61 nvbo->bo.base.funcs = &nouveau_gem_object_funcs; in nouveau_gem_prime_import_sg_table() 65 ret = drm_gem_object_init(dev, &nvbo->bo.base, size); in nouveau_gem_prime_import_sg_table() [all …]
|
H A D | nouveau_vmm.c | 49 nouveau_vma_find(struct nouveau_bo *nvbo, struct nouveau_vmm *vmm) in nouveau_vma_find() argument 53 list_for_each_entry(vma, &nvbo->vma_list, head) { in nouveau_vma_find() 77 nouveau_vma_new(struct nouveau_bo *nvbo, struct nouveau_vmm *vmm, in nouveau_vma_new() argument 80 struct nouveau_mem *mem = nouveau_mem(nvbo->bo.resource); in nouveau_vma_new() 85 if ((vma = *pvma = nouveau_vma_find(nvbo, vmm))) { in nouveau_vma_new() 97 list_add_tail(&vma->head, &nvbo->vma_list); in nouveau_vma_new() 99 if (nvbo->bo.resource->mem_type != TTM_PL_SYSTEM && in nouveau_vma_new() 100 mem->mem.page == nvbo->page) { in nouveau_vma_new()
|
H A D | nouveau_ttm.c | 69 struct nouveau_bo *nvbo = nouveau_bo(bo); in nouveau_vram_manager_new() local 76 ret = nouveau_mem_new(&drm->master, nvbo->kind, nvbo->comp, res); in nouveau_vram_manager_new() 82 ret = nouveau_mem_vram(*res, nvbo->contig, nvbo->page); in nouveau_vram_manager_new() 104 struct nouveau_bo *nvbo = nouveau_bo(bo); in nouveau_gart_manager_new() local 108 ret = nouveau_mem_new(&drm->master, nvbo->kind, nvbo->comp, res); in nouveau_gart_manager_new() 130 struct nouveau_bo *nvbo = nouveau_bo(bo); in nv04_gart_manager_new() local 135 ret = nouveau_mem_new(&drm->master, nvbo->kind, nvbo->comp, res); in nv04_gart_manager_new()
|
H A D | nouveau_display.c | 172 const struct nouveau_bo *nvbo = nouveau_gem_object(fb->obj[0]); in nouveau_framebuffer_get_layout() local 174 *tile_mode = nvbo->mode; in nouveau_framebuffer_get_layout() 175 *kind = nvbo->kind; in nouveau_framebuffer_get_layout() 255 nouveau_check_bl_size(struct nouveau_drm *drm, struct nouveau_bo *nvbo, in nouveau_check_bl_size() argument 285 nvbo->bo.base.size); in nouveau_check_bl_size() 287 if (bl_size + offset > nvbo->bo.base.size) in nouveau_check_bl_size() 300 struct nouveau_bo *nvbo = nouveau_gem_object(gem); in nouveau_framebuffer_new() local 333 tile_mode = nvbo->mode; in nouveau_framebuffer_new() 334 kind = nvbo->kind; in nouveau_framebuffer_new() 345 ret = nouveau_check_bl_size(drm, nvbo, in nouveau_framebuffer_new() [all …]
|
H A D | nouveau_bo.h | 96 void nouveau_bo_sync_for_device(struct nouveau_bo *nvbo); 97 void nouveau_bo_sync_for_cpu(struct nouveau_bo *nvbo); 103 nvbo_kmap_obj_iovirtual(struct nouveau_bo *nvbo) in nvbo_kmap_obj_iovirtual() argument 107 &nvbo->kmap, &is_iomem); in nvbo_kmap_obj_iovirtual()
|
H A D | nouveau_sgdma.c | 71 struct nouveau_bo *nvbo = nouveau_bo(bo); in nouveau_sgdma_create_ttm() local 75 if (nvbo->force_coherent) in nouveau_sgdma_create_ttm()
|
H A D | nouveau_uvmm.c | 773 struct nouveau_bo *nvbo = nouveau_gem_object(uvma->va.gem.obj); in op_map() local 775 nouveau_uvma_map(uvma, nouveau_mem(nvbo->bo.resource)); in op_map() 1323 struct nouveau_bo *nvbo; in nouveau_uvmm_bind_job_submit() local 1334 nvbo = nouveau_gem_object(obj); in nouveau_uvmm_bind_job_submit() 1335 nouveau_bo_placement_set(nvbo, nvbo->valid_domains, 0); in nouveau_uvmm_bind_job_submit() 1336 ret = nouveau_bo_validate(nvbo, true, false); in nouveau_uvmm_bind_job_submit() 1778 nouveau_uvmm_bo_map_all(struct nouveau_bo *nvbo, struct nouveau_mem *mem) in nouveau_uvmm_bo_map_all() argument 1780 struct drm_gem_object *obj = &nvbo->bo.base; in nouveau_uvmm_bo_map_all() 1794 nouveau_uvmm_bo_unmap_all(struct nouveau_bo *nvbo) in nouveau_uvmm_bo_unmap_all() argument 1796 struct drm_gem_object *obj = &nvbo->bo.base; in nouveau_uvmm_bo_unmap_all()
|
H A D | nouveau_crtc.h | 57 struct nouveau_bo *nvbo; member
|
H A D | nouveau_uvmm.h | 90 void nouveau_uvmm_bo_unmap_all(struct nouveau_bo *nvbo);
|
H A D | nouveau_exec.c | 123 struct nouveau_bo *nvbo = nouveau_gem_object(obj); in nouveau_exec_job_submit() local 125 ret = nouveau_bo_validate(nvbo, true, false); in nouveau_exec_job_submit()
|
H A D | nouveau_fence.c | 357 nouveau_fence_sync(struct nouveau_bo *nvbo, struct nouveau_channel *chan, in nouveau_fence_sync() argument 361 struct dma_resv *resv = nvbo->bo.base.resv; in nouveau_fence_sync()
|
/openbmc/linux/drivers/gpu/drm/nouveau/dispnv04/ |
H A D | disp.c | 80 struct nouveau_bo *nvbo; in nv04_display_fini() local 84 nvbo = nouveau_gem_object(fb->obj[0]); in nv04_display_fini() 85 nouveau_bo_unpin(nvbo); in nv04_display_fini() 90 if (nv_crtc->cursor.nvbo) { in nv04_display_fini() 92 nouveau_bo_unmap(nv_crtc->cursor.nvbo); in nv04_display_fini() 93 nouveau_bo_unpin(nv_crtc->cursor.nvbo); in nv04_display_fini() 132 struct nouveau_bo *nvbo; in nv04_display_init() local 136 nvbo = nouveau_gem_object(fb->obj[0]); in nv04_display_init() 137 ret = nouveau_bo_pin(nvbo, NOUVEAU_GEM_DOMAIN_VRAM, true); in nv04_display_init() 144 if (!nv_crtc->cursor.nvbo) in nv04_display_init() [all …]
|
H A D | overlay.c | 126 struct nouveau_bo *nvbo; in nv10_update_plane() local 144 nvbo = nouveau_gem_object(fb->obj[0]); in nv10_update_plane() 145 ret = nouveau_bo_pin(nvbo, NOUVEAU_GEM_DOMAIN_VRAM, false); in nv10_update_plane() 149 nv_plane->cur = nvbo; in nv10_update_plane() 155 nvif_wr32(dev, NV_PVIDEO_OFFSET_BUFF(flip), nvbo->offset); in nv10_update_plane() 177 nvbo->offset + fb->offsets[1]); in nv10_update_plane() 375 struct nouveau_bo *nvbo; in nv04_update_plane() local 390 nvbo = nouveau_gem_object(fb->obj[0]); in nv04_update_plane() 391 ret = nouveau_bo_pin(nvbo, NOUVEAU_GEM_DOMAIN_VRAM, false); in nv04_update_plane() 395 nv_plane->cur = nvbo; in nv04_update_plane() [all …]
|
H A D | crtc.c | 614 struct nouveau_bo *nvbo = nouveau_gem_object(fb->obj[0]); in nv_crtc_swap_fbs() local 618 ret = nouveau_bo_pin(nvbo, NOUVEAU_GEM_DOMAIN_VRAM, false); in nv_crtc_swap_fbs() 622 nouveau_bo_ref(nvbo, &disp->image[nv_crtc->index]); in nv_crtc_swap_fbs() 761 nouveau_bo_unmap(nv_crtc->cursor.nvbo); in nv_crtc_destroy() 762 nouveau_bo_unpin(nv_crtc->cursor.nvbo); in nv_crtc_destroy() 763 nouveau_bo_ref(NULL, &nv_crtc->cursor.nvbo); in nv_crtc_destroy() 833 struct nouveau_bo *nvbo; in nv04_crtc_do_mode_set_base() local 854 nvbo = nouveau_gem_object(drm_fb->obj[0]); in nv04_crtc_do_mode_set_base() 855 nv_crtc->fb.offset = nvbo->offset; in nv04_crtc_do_mode_set_base() 1010 nv11_cursor_upload(dev, cursor, nv_crtc->cursor.nvbo); in nv04_crtc_cursor_set() [all …]
|
/openbmc/linux/drivers/gpu/drm/nouveau/dispnv50/ |
H A D | wndw.c | 521 struct nouveau_bo *nvbo; in nv50_wndw_cleanup_fb() local 527 nvbo = nouveau_gem_object(old_state->fb->obj[0]); in nv50_wndw_cleanup_fb() 528 nouveau_bo_unpin(nvbo); in nv50_wndw_cleanup_fb() 538 struct nouveau_bo *nvbo; in nv50_wndw_prepare_fb() local 547 nvbo = nouveau_gem_object(fb->obj[0]); in nv50_wndw_prepare_fb() 548 ret = nouveau_bo_pin(nvbo, NOUVEAU_GEM_DOMAIN_VRAM, true); in nv50_wndw_prepare_fb() 555 nouveau_bo_unpin(nvbo); in nv50_wndw_prepare_fb() 567 asyw->image.offset[0] = nvbo->offset; in nv50_wndw_prepare_fb()
|