Lines Matching full:v0

49 		struct nvif_vmm_pfnclr_v0 v0;  in nvkm_uvmm_mthd_pfnclr()  member
55 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) { in nvkm_uvmm_mthd_pfnclr()
56 addr = args->v0.addr; in nvkm_uvmm_mthd_pfnclr()
57 size = args->v0.size; in nvkm_uvmm_mthd_pfnclr()
77 struct nvif_vmm_pfnmap_v0 v0; in nvkm_uvmm_mthd_pfnmap() member
84 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, true))) { in nvkm_uvmm_mthd_pfnmap()
85 page = args->v0.page; in nvkm_uvmm_mthd_pfnmap()
86 addr = args->v0.addr; in nvkm_uvmm_mthd_pfnmap()
87 size = args->v0.size; in nvkm_uvmm_mthd_pfnmap()
88 phys = args->v0.phys; in nvkm_uvmm_mthd_pfnmap()
89 if (argc != (size >> page) * sizeof(args->v0.phys[0])) in nvkm_uvmm_mthd_pfnmap()
110 struct nvif_vmm_unmap_v0 v0; in nvkm_uvmm_mthd_unmap() member
117 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) { in nvkm_uvmm_mthd_unmap()
118 addr = args->v0.addr; in nvkm_uvmm_mthd_unmap()
155 struct nvif_vmm_map_v0 v0; in nvkm_uvmm_mthd_map() member
163 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, true))) { in nvkm_uvmm_mthd_map()
164 addr = args->v0.addr; in nvkm_uvmm_mthd_map()
165 size = args->v0.size; in nvkm_uvmm_mthd_map()
166 handle = args->v0.memory; in nvkm_uvmm_mthd_map()
167 offset = args->v0.offset; in nvkm_uvmm_mthd_map()
235 struct nvif_vmm_put_v0 v0; in nvkm_uvmm_mthd_put() member
242 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) { in nvkm_uvmm_mthd_put()
243 addr = args->v0.addr; in nvkm_uvmm_mthd_put()
248 vma = nvkm_vmm_node_search(vmm, args->v0.addr); in nvkm_uvmm_mthd_put()
271 struct nvif_vmm_get_v0 v0; in nvkm_uvmm_mthd_get() member
280 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) { in nvkm_uvmm_mthd_get()
281 getref = args->v0.type == NVIF_VMM_GET_V0_PTES; in nvkm_uvmm_mthd_get()
282 mapref = args->v0.type == NVIF_VMM_GET_V0_ADDR; in nvkm_uvmm_mthd_get()
283 sparse = args->v0.sparse; in nvkm_uvmm_mthd_get()
284 page = args->v0.page; in nvkm_uvmm_mthd_get()
285 align = args->v0.align; in nvkm_uvmm_mthd_get()
286 size = args->v0.size; in nvkm_uvmm_mthd_get()
297 args->v0.addr = vma->addr; in nvkm_uvmm_mthd_get()
305 struct nvif_vmm_page_v0 v0; in nvkm_uvmm_mthd_page() member
314 if (!(nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) { in nvkm_uvmm_mthd_page()
315 if ((index = args->v0.index) >= nr) in nvkm_uvmm_mthd_page()
318 args->v0.shift = page[index].shift; in nvkm_uvmm_mthd_page()
319 args->v0.sparse = !!(type & NVKM_VMM_PAGE_SPARSE); in nvkm_uvmm_mthd_page()
320 args->v0.vram = !!(type & NVKM_VMM_PAGE_VRAM); in nvkm_uvmm_mthd_page()
321 args->v0.host = !!(type & NVKM_VMM_PAGE_HOST); in nvkm_uvmm_mthd_page()
322 args->v0.comp = !!(type & NVKM_VMM_PAGE_COMP); in nvkm_uvmm_mthd_page()
465 struct nvif_vmm_raw_v0 v0; in nvkm_uvmm_mthd_raw() member
472 if ((ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, true))) in nvkm_uvmm_mthd_raw()
475 switch (args->v0.op) { in nvkm_uvmm_mthd_raw()
477 return nvkm_uvmm_mthd_raw_get(uvmm, &args->v0); in nvkm_uvmm_mthd_raw()
479 return nvkm_uvmm_mthd_raw_put(uvmm, &args->v0); in nvkm_uvmm_mthd_raw()
481 return nvkm_uvmm_mthd_raw_map(uvmm, &args->v0); in nvkm_uvmm_mthd_raw()
483 return nvkm_uvmm_mthd_raw_unmap(uvmm, &args->v0); in nvkm_uvmm_mthd_raw()
485 return nvkm_uvmm_mthd_raw_sparse(uvmm, &args->v0); in nvkm_uvmm_mthd_raw()
538 struct nvif_vmm_v0 v0; in nvkm_uvmm_new() member
546 if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, more))) { in nvkm_uvmm_new()
547 managed = args->v0.type == NVIF_VMM_V0_TYPE_MANAGED; in nvkm_uvmm_new()
548 raw = args->v0.type == NVIF_VMM_V0_TYPE_RAW; in nvkm_uvmm_new()
549 addr = args->v0.addr; in nvkm_uvmm_new()
550 size = args->v0.size; in nvkm_uvmm_new()
576 args->v0.page_nr = 0; in nvkm_uvmm_new()
578 args->v0.page_nr++; in nvkm_uvmm_new()
579 args->v0.addr = uvmm->vmm->start; in nvkm_uvmm_new()
580 args->v0.size = uvmm->vmm->limit; in nvkm_uvmm_new()