1 /* 2 * Copyright 2012 Red Hat Inc. 3 * 4 * Permission is hereby granted, free of charge, to any person obtaining a 5 * copy of this software and associated documentation files (the "Software"), 6 * to deal in the Software without restriction, including without limitation 7 * the rights to use, copy, modify, merge, publish, distribute, sublicense, 8 * and/or sell copies of the Software, and to permit persons to whom the 9 * Software is furnished to do so, subject to the following conditions: 10 * 11 * The above copyright notice and this permission notice shall be included in 12 * all copies or substantial portions of the Software. 13 * 14 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 15 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 16 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 17 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR 18 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, 19 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR 20 * OTHER DEALINGS IN THE SOFTWARE. 21 * 22 */ 23 24 #include <nvif/client.h> 25 #include <nvif/driver.h> 26 #include <nvif/ioctl.h> 27 #include <nvif/class.h> 28 29 #include "nouveau_drm.h" 30 #include "nouveau_dma.h" 31 #include "nouveau_gem.h" 32 #include "nouveau_chan.h" 33 #include "nouveau_abi16.h" 34 35 struct nouveau_abi16 * 36 nouveau_abi16_get(struct drm_file *file_priv, struct drm_device *dev) 37 { 38 struct nouveau_cli *cli = nouveau_cli(file_priv); 39 mutex_lock(&cli->mutex); 40 if (!cli->abi16) { 41 struct nouveau_abi16 *abi16; 42 cli->abi16 = abi16 = kzalloc(sizeof(*abi16), GFP_KERNEL); 43 if (cli->abi16) { 44 struct nv_device_v0 args = { 45 .device = ~0ULL, 46 }; 47 48 INIT_LIST_HEAD(&abi16->channels); 49 50 /* allocate device object targeting client's default 51 * device (ie. the one that belongs to the fd it 52 * opened) 53 */ 54 if (nvif_device_init(&cli->base.base, NULL, 55 NOUVEAU_ABI16_DEVICE, NV_DEVICE, 56 &args, sizeof(args), 57 &abi16->device) == 0) 58 return cli->abi16; 59 60 kfree(cli->abi16); 61 cli->abi16 = NULL; 62 } 63 64 mutex_unlock(&cli->mutex); 65 } 66 return cli->abi16; 67 } 68 69 int 70 nouveau_abi16_put(struct nouveau_abi16 *abi16, int ret) 71 { 72 struct nouveau_cli *cli = (void *)nvif_client(&abi16->device.base); 73 mutex_unlock(&cli->mutex); 74 return ret; 75 } 76 77 u16 78 nouveau_abi16_swclass(struct nouveau_drm *drm) 79 { 80 switch (drm->device.info.family) { 81 case NV_DEVICE_INFO_V0_TNT: 82 return 0x006e; 83 case NV_DEVICE_INFO_V0_CELSIUS: 84 case NV_DEVICE_INFO_V0_KELVIN: 85 case NV_DEVICE_INFO_V0_RANKINE: 86 case NV_DEVICE_INFO_V0_CURIE: 87 return 0x016e; 88 case NV_DEVICE_INFO_V0_TESLA: 89 return 0x506e; 90 case NV_DEVICE_INFO_V0_FERMI: 91 case NV_DEVICE_INFO_V0_KEPLER: 92 case NV_DEVICE_INFO_V0_MAXWELL: 93 return 0x906e; 94 } 95 96 return 0x0000; 97 } 98 99 static void 100 nouveau_abi16_ntfy_fini(struct nouveau_abi16_chan *chan, 101 struct nouveau_abi16_ntfy *ntfy) 102 { 103 nvkm_mm_free(&chan->heap, &ntfy->node); 104 list_del(&ntfy->head); 105 kfree(ntfy); 106 } 107 108 static void 109 nouveau_abi16_chan_fini(struct nouveau_abi16 *abi16, 110 struct nouveau_abi16_chan *chan) 111 { 112 struct nouveau_abi16_ntfy *ntfy, *temp; 113 114 /* wait for all activity to stop before releasing notify object, which 115 * may be still in use */ 116 if (chan->chan && chan->ntfy) 117 nouveau_channel_idle(chan->chan); 118 119 /* cleanup notifier state */ 120 list_for_each_entry_safe(ntfy, temp, &chan->notifiers, head) { 121 nouveau_abi16_ntfy_fini(chan, ntfy); 122 } 123 124 if (chan->ntfy) { 125 nouveau_bo_vma_del(chan->ntfy, &chan->ntfy_vma); 126 nouveau_bo_unpin(chan->ntfy); 127 drm_gem_object_unreference_unlocked(&chan->ntfy->gem); 128 } 129 130 if (chan->heap.block_size) 131 nvkm_mm_fini(&chan->heap); 132 133 /* destroy channel object, all children will be killed too */ 134 if (chan->chan) { 135 abi16->handles &= ~(1ULL << (chan->chan->object->handle & 0xffff)); 136 nouveau_channel_del(&chan->chan); 137 } 138 139 list_del(&chan->head); 140 kfree(chan); 141 } 142 143 void 144 nouveau_abi16_fini(struct nouveau_abi16 *abi16) 145 { 146 struct nouveau_cli *cli = (void *)nvif_client(&abi16->device.base); 147 struct nouveau_abi16_chan *chan, *temp; 148 149 /* cleanup channels */ 150 list_for_each_entry_safe(chan, temp, &abi16->channels, head) { 151 nouveau_abi16_chan_fini(abi16, chan); 152 } 153 154 /* destroy the device object */ 155 nvif_device_fini(&abi16->device); 156 157 kfree(cli->abi16); 158 cli->abi16 = NULL; 159 } 160 161 int 162 nouveau_abi16_ioctl_getparam(ABI16_IOCTL_ARGS) 163 { 164 struct nouveau_cli *cli = nouveau_cli(file_priv); 165 struct nouveau_drm *drm = nouveau_drm(dev); 166 struct nvif_device *device = &drm->device; 167 struct nvkm_timer *ptimer = nvxx_timer(device); 168 struct nvkm_gr *gr = nvxx_gr(device); 169 struct drm_nouveau_getparam *getparam = data; 170 171 switch (getparam->param) { 172 case NOUVEAU_GETPARAM_CHIPSET_ID: 173 getparam->value = device->info.chipset; 174 break; 175 case NOUVEAU_GETPARAM_PCI_VENDOR: 176 if (nv_device_is_pci(nvxx_device(device))) 177 getparam->value = dev->pdev->vendor; 178 else 179 getparam->value = 0; 180 break; 181 case NOUVEAU_GETPARAM_PCI_DEVICE: 182 if (nv_device_is_pci(nvxx_device(device))) 183 getparam->value = dev->pdev->device; 184 else 185 getparam->value = 0; 186 break; 187 case NOUVEAU_GETPARAM_BUS_TYPE: 188 if (!nv_device_is_pci(nvxx_device(device))) 189 getparam->value = 3; 190 else 191 if (drm_pci_device_is_agp(dev)) 192 getparam->value = 0; 193 else 194 if (!pci_is_pcie(dev->pdev)) 195 getparam->value = 1; 196 else 197 getparam->value = 2; 198 break; 199 case NOUVEAU_GETPARAM_FB_SIZE: 200 getparam->value = drm->gem.vram_available; 201 break; 202 case NOUVEAU_GETPARAM_AGP_SIZE: 203 getparam->value = drm->gem.gart_available; 204 break; 205 case NOUVEAU_GETPARAM_VM_VRAM_BASE: 206 getparam->value = 0; /* deprecated */ 207 break; 208 case NOUVEAU_GETPARAM_PTIMER_TIME: 209 getparam->value = ptimer->read(ptimer); 210 break; 211 case NOUVEAU_GETPARAM_HAS_BO_USAGE: 212 getparam->value = 1; 213 break; 214 case NOUVEAU_GETPARAM_HAS_PAGEFLIP: 215 getparam->value = 1; 216 break; 217 case NOUVEAU_GETPARAM_GRAPH_UNITS: 218 getparam->value = gr->units ? gr->units(gr) : 0; 219 break; 220 default: 221 NV_PRINTK(debug, cli, "unknown parameter %lld\n", getparam->param); 222 return -EINVAL; 223 } 224 225 return 0; 226 } 227 228 int 229 nouveau_abi16_ioctl_setparam(ABI16_IOCTL_ARGS) 230 { 231 return -EINVAL; 232 } 233 234 int 235 nouveau_abi16_ioctl_channel_alloc(ABI16_IOCTL_ARGS) 236 { 237 struct drm_nouveau_channel_alloc *init = data; 238 struct nouveau_cli *cli = nouveau_cli(file_priv); 239 struct nouveau_drm *drm = nouveau_drm(dev); 240 struct nouveau_abi16 *abi16 = nouveau_abi16_get(file_priv, dev); 241 struct nouveau_abi16_chan *chan; 242 struct nvif_device *device; 243 int ret; 244 245 if (unlikely(!abi16)) 246 return -ENOMEM; 247 248 if (!drm->channel) 249 return nouveau_abi16_put(abi16, -ENODEV); 250 251 device = &abi16->device; 252 253 /* hack to allow channel engine type specification on kepler */ 254 if (device->info.family >= NV_DEVICE_INFO_V0_KEPLER) { 255 if (init->fb_ctxdma_handle != ~0) 256 init->fb_ctxdma_handle = KEPLER_CHANNEL_GPFIFO_A_V0_ENGINE_GR; 257 else 258 init->fb_ctxdma_handle = init->tt_ctxdma_handle; 259 260 /* allow flips to be executed if this is a graphics channel */ 261 init->tt_ctxdma_handle = 0; 262 if (init->fb_ctxdma_handle == KEPLER_CHANNEL_GPFIFO_A_V0_ENGINE_GR) 263 init->tt_ctxdma_handle = 1; 264 } 265 266 if (init->fb_ctxdma_handle == ~0 || init->tt_ctxdma_handle == ~0) 267 return nouveau_abi16_put(abi16, -EINVAL); 268 269 /* allocate "abi16 channel" data and make up a handle for it */ 270 init->channel = __ffs64(~abi16->handles); 271 if (~abi16->handles == 0) 272 return nouveau_abi16_put(abi16, -ENOSPC); 273 274 chan = kzalloc(sizeof(*chan), GFP_KERNEL); 275 if (!chan) 276 return nouveau_abi16_put(abi16, -ENOMEM); 277 278 INIT_LIST_HEAD(&chan->notifiers); 279 list_add(&chan->head, &abi16->channels); 280 abi16->handles |= (1ULL << init->channel); 281 282 /* create channel object and initialise dma and fence management */ 283 ret = nouveau_channel_new(drm, device, 284 NOUVEAU_ABI16_CHAN(init->channel), 285 init->fb_ctxdma_handle, 286 init->tt_ctxdma_handle, &chan->chan); 287 if (ret) 288 goto done; 289 290 if (device->info.family >= NV_DEVICE_INFO_V0_TESLA) 291 init->pushbuf_domains = NOUVEAU_GEM_DOMAIN_VRAM | 292 NOUVEAU_GEM_DOMAIN_GART; 293 else 294 if (chan->chan->push.buffer->bo.mem.mem_type == TTM_PL_VRAM) 295 init->pushbuf_domains = NOUVEAU_GEM_DOMAIN_VRAM; 296 else 297 init->pushbuf_domains = NOUVEAU_GEM_DOMAIN_GART; 298 299 if (device->info.family < NV_DEVICE_INFO_V0_CELSIUS) { 300 init->subchan[0].handle = 0x00000000; 301 init->subchan[0].grclass = 0x0000; 302 init->subchan[1].handle = chan->chan->nvsw.handle; 303 init->subchan[1].grclass = 0x506e; 304 init->nr_subchan = 2; 305 } 306 307 /* Named memory object area */ 308 ret = nouveau_gem_new(dev, PAGE_SIZE, 0, NOUVEAU_GEM_DOMAIN_GART, 309 0, 0, &chan->ntfy); 310 if (ret == 0) 311 ret = nouveau_bo_pin(chan->ntfy, TTM_PL_FLAG_TT, false); 312 if (ret) 313 goto done; 314 315 if (device->info.family >= NV_DEVICE_INFO_V0_TESLA) { 316 ret = nouveau_bo_vma_add(chan->ntfy, cli->vm, 317 &chan->ntfy_vma); 318 if (ret) 319 goto done; 320 } 321 322 ret = drm_gem_handle_create(file_priv, &chan->ntfy->gem, 323 &init->notifier_handle); 324 if (ret) 325 goto done; 326 327 ret = nvkm_mm_init(&chan->heap, 0, PAGE_SIZE, 1); 328 done: 329 if (ret) 330 nouveau_abi16_chan_fini(abi16, chan); 331 return nouveau_abi16_put(abi16, ret); 332 } 333 334 static struct nouveau_abi16_chan * 335 nouveau_abi16_chan(struct nouveau_abi16 *abi16, int channel) 336 { 337 struct nouveau_abi16_chan *chan; 338 339 list_for_each_entry(chan, &abi16->channels, head) { 340 if (chan->chan->object->handle == NOUVEAU_ABI16_CHAN(channel)) 341 return chan; 342 } 343 344 return NULL; 345 } 346 347 int 348 nouveau_abi16_ioctl_channel_free(ABI16_IOCTL_ARGS) 349 { 350 struct drm_nouveau_channel_free *req = data; 351 struct nouveau_abi16 *abi16 = nouveau_abi16_get(file_priv, dev); 352 struct nouveau_abi16_chan *chan; 353 354 if (unlikely(!abi16)) 355 return -ENOMEM; 356 357 chan = nouveau_abi16_chan(abi16, req->channel); 358 if (!chan) 359 return nouveau_abi16_put(abi16, -ENOENT); 360 nouveau_abi16_chan_fini(abi16, chan); 361 return nouveau_abi16_put(abi16, 0); 362 } 363 364 int 365 nouveau_abi16_ioctl_grobj_alloc(ABI16_IOCTL_ARGS) 366 { 367 struct drm_nouveau_grobj_alloc *init = data; 368 struct { 369 struct nvif_ioctl_v0 ioctl; 370 struct nvif_ioctl_new_v0 new; 371 } args = { 372 .ioctl.owner = NVIF_IOCTL_V0_OWNER_ANY, 373 .ioctl.type = NVIF_IOCTL_V0_NEW, 374 .ioctl.path_nr = 3, 375 .ioctl.path[2] = NOUVEAU_ABI16_CLIENT, 376 .ioctl.path[1] = NOUVEAU_ABI16_DEVICE, 377 .ioctl.path[0] = NOUVEAU_ABI16_CHAN(init->channel), 378 .new.route = NVDRM_OBJECT_ABI16, 379 .new.handle = init->handle, 380 .new.oclass = init->class, 381 }; 382 struct nouveau_abi16 *abi16 = nouveau_abi16_get(file_priv, dev); 383 struct nouveau_drm *drm = nouveau_drm(dev); 384 struct nvif_client *client; 385 int ret; 386 387 if (unlikely(!abi16)) 388 return -ENOMEM; 389 390 if (init->handle == ~0) 391 return nouveau_abi16_put(abi16, -EINVAL); 392 client = nvif_client(nvif_object(&abi16->device)); 393 394 /* compatibility with userspace that assumes 506e for all chipsets */ 395 if (init->class == 0x506e) { 396 init->class = nouveau_abi16_swclass(drm); 397 if (init->class == 0x906e) 398 return nouveau_abi16_put(abi16, 0); 399 } 400 401 ret = nvif_client_ioctl(client, &args, sizeof(args)); 402 return nouveau_abi16_put(abi16, ret); 403 } 404 405 int 406 nouveau_abi16_ioctl_notifierobj_alloc(ABI16_IOCTL_ARGS) 407 { 408 struct drm_nouveau_notifierobj_alloc *info = data; 409 struct { 410 struct nvif_ioctl_v0 ioctl; 411 struct nvif_ioctl_new_v0 new; 412 struct nv_dma_v0 ctxdma; 413 } args = { 414 .ioctl.owner = NVIF_IOCTL_V0_OWNER_ANY, 415 .ioctl.type = NVIF_IOCTL_V0_NEW, 416 .ioctl.path_nr = 3, 417 .ioctl.path[2] = NOUVEAU_ABI16_CLIENT, 418 .ioctl.path[1] = NOUVEAU_ABI16_DEVICE, 419 .ioctl.path[0] = NOUVEAU_ABI16_CHAN(info->channel), 420 .new.route = NVDRM_OBJECT_ABI16, 421 .new.handle = info->handle, 422 .new.oclass = NV_DMA_IN_MEMORY, 423 }; 424 struct nouveau_drm *drm = nouveau_drm(dev); 425 struct nouveau_abi16 *abi16 = nouveau_abi16_get(file_priv, dev); 426 struct nouveau_abi16_chan *chan; 427 struct nouveau_abi16_ntfy *ntfy; 428 struct nvif_device *device = &abi16->device; 429 struct nvif_client *client; 430 int ret; 431 432 if (unlikely(!abi16)) 433 return -ENOMEM; 434 435 /* completely unnecessary for these chipsets... */ 436 if (unlikely(device->info.family >= NV_DEVICE_INFO_V0_FERMI)) 437 return nouveau_abi16_put(abi16, -EINVAL); 438 client = nvif_client(nvif_object(&abi16->device)); 439 440 chan = nouveau_abi16_chan(abi16, info->channel); 441 if (!chan) 442 return nouveau_abi16_put(abi16, -ENOENT); 443 444 ntfy = kzalloc(sizeof(*ntfy), GFP_KERNEL); 445 if (!ntfy) 446 return nouveau_abi16_put(abi16, -ENOMEM); 447 448 list_add(&ntfy->head, &chan->notifiers); 449 ntfy->handle = info->handle; 450 451 ret = nvkm_mm_head(&chan->heap, 0, 1, info->size, info->size, 1, 452 &ntfy->node); 453 if (ret) 454 goto done; 455 456 args.ctxdma.start = ntfy->node->offset; 457 args.ctxdma.limit = ntfy->node->offset + ntfy->node->length - 1; 458 if (device->info.family >= NV_DEVICE_INFO_V0_TESLA) { 459 args.ctxdma.target = NV_DMA_V0_TARGET_VM; 460 args.ctxdma.access = NV_DMA_V0_ACCESS_VM; 461 args.ctxdma.start += chan->ntfy_vma.offset; 462 args.ctxdma.limit += chan->ntfy_vma.offset; 463 } else 464 if (drm->agp.stat == ENABLED) { 465 args.ctxdma.target = NV_DMA_V0_TARGET_AGP; 466 args.ctxdma.access = NV_DMA_V0_ACCESS_RDWR; 467 args.ctxdma.start += drm->agp.base + chan->ntfy->bo.offset; 468 args.ctxdma.limit += drm->agp.base + chan->ntfy->bo.offset; 469 client->super = true; 470 } else { 471 args.ctxdma.target = NV_DMA_V0_TARGET_VM; 472 args.ctxdma.access = NV_DMA_V0_ACCESS_RDWR; 473 args.ctxdma.start += chan->ntfy->bo.offset; 474 args.ctxdma.limit += chan->ntfy->bo.offset; 475 } 476 477 ret = nvif_client_ioctl(client, &args, sizeof(args)); 478 client->super = false; 479 if (ret) 480 goto done; 481 482 info->offset = ntfy->node->offset; 483 484 done: 485 if (ret) 486 nouveau_abi16_ntfy_fini(chan, ntfy); 487 return nouveau_abi16_put(abi16, ret); 488 } 489 490 int 491 nouveau_abi16_ioctl_gpuobj_free(ABI16_IOCTL_ARGS) 492 { 493 struct drm_nouveau_gpuobj_free *fini = data; 494 struct { 495 struct nvif_ioctl_v0 ioctl; 496 struct nvif_ioctl_del del; 497 } args = { 498 .ioctl.owner = NVDRM_OBJECT_ABI16, 499 .ioctl.type = NVIF_IOCTL_V0_DEL, 500 .ioctl.path_nr = 4, 501 .ioctl.path[3] = NOUVEAU_ABI16_CLIENT, 502 .ioctl.path[2] = NOUVEAU_ABI16_DEVICE, 503 .ioctl.path[1] = NOUVEAU_ABI16_CHAN(fini->channel), 504 .ioctl.path[0] = fini->handle, 505 }; 506 struct nouveau_abi16 *abi16 = nouveau_abi16_get(file_priv, dev); 507 struct nouveau_abi16_chan *chan; 508 struct nouveau_abi16_ntfy *ntfy; 509 struct nvif_client *client; 510 int ret; 511 512 if (unlikely(!abi16)) 513 return -ENOMEM; 514 515 chan = nouveau_abi16_chan(abi16, fini->channel); 516 if (!chan) 517 return nouveau_abi16_put(abi16, -ENOENT); 518 client = nvif_client(nvif_object(&abi16->device)); 519 520 /* synchronize with the user channel and destroy the gpu object */ 521 nouveau_channel_idle(chan->chan); 522 523 ret = nvif_client_ioctl(client, &args, sizeof(args)); 524 if (ret) 525 return nouveau_abi16_put(abi16, ret); 526 527 /* cleanup extra state if this object was a notifier */ 528 list_for_each_entry(ntfy, &chan->notifiers, head) { 529 if (ntfy->handle == fini->handle) { 530 nvkm_mm_free(&chan->heap, &ntfy->node); 531 list_del(&ntfy->head); 532 break; 533 } 534 } 535 536 return nouveau_abi16_put(abi16, 0); 537 } 538