1*5320918bSDave Airlie /* 2*5320918bSDave Airlie * Copyright (C) 2012 Red Hat 3*5320918bSDave Airlie * 4*5320918bSDave Airlie * based in parts on udlfb.c: 5*5320918bSDave Airlie * Copyright (C) 2009 Roberto De Ioris <roberto@unbit.it> 6*5320918bSDave Airlie * Copyright (C) 2009 Jaya Kumar <jayakumar.lkml@gmail.com> 7*5320918bSDave Airlie * Copyright (C) 2009 Bernie Thompson <bernie@plugable.com> 8*5320918bSDave Airlie * 9*5320918bSDave Airlie * This file is subject to the terms and conditions of the GNU General Public 10*5320918bSDave Airlie * License v2. See the file COPYING in the main directory of this archive for 11*5320918bSDave Airlie * more details. 12*5320918bSDave Airlie */ 13*5320918bSDave Airlie #include "drmP.h" 14*5320918bSDave Airlie #include "udl_drv.h" 15*5320918bSDave Airlie 16*5320918bSDave Airlie /* -BULK_SIZE as per usb-skeleton. Can we get full page and avoid overhead? */ 17*5320918bSDave Airlie #define BULK_SIZE 512 18*5320918bSDave Airlie 19*5320918bSDave Airlie #define MAX_TRANSFER (PAGE_SIZE*16 - BULK_SIZE) 20*5320918bSDave Airlie #define WRITES_IN_FLIGHT (4) 21*5320918bSDave Airlie #define MAX_VENDOR_DESCRIPTOR_SIZE 256 22*5320918bSDave Airlie 23*5320918bSDave Airlie #define GET_URB_TIMEOUT HZ 24*5320918bSDave Airlie #define FREE_URB_TIMEOUT (HZ*2) 25*5320918bSDave Airlie 26*5320918bSDave Airlie static int udl_parse_vendor_descriptor(struct drm_device *dev, 27*5320918bSDave Airlie struct usb_device *usbdev) 28*5320918bSDave Airlie { 29*5320918bSDave Airlie struct udl_device *udl = dev->dev_private; 30*5320918bSDave Airlie char *desc; 31*5320918bSDave Airlie char *buf; 32*5320918bSDave Airlie char *desc_end; 33*5320918bSDave Airlie 34*5320918bSDave Airlie u8 total_len = 0; 35*5320918bSDave Airlie 36*5320918bSDave Airlie buf = kzalloc(MAX_VENDOR_DESCRIPTOR_SIZE, GFP_KERNEL); 37*5320918bSDave Airlie if (!buf) 38*5320918bSDave Airlie return false; 39*5320918bSDave Airlie desc = buf; 40*5320918bSDave Airlie 41*5320918bSDave Airlie total_len = usb_get_descriptor(usbdev, 0x5f, /* vendor specific */ 42*5320918bSDave Airlie 0, desc, MAX_VENDOR_DESCRIPTOR_SIZE); 43*5320918bSDave Airlie if (total_len > 5) { 44*5320918bSDave Airlie DRM_INFO("vendor descriptor length:%x data:%02x %02x %02x %02x" \ 45*5320918bSDave Airlie "%02x %02x %02x %02x %02x %02x %02x\n", 46*5320918bSDave Airlie total_len, desc[0], 47*5320918bSDave Airlie desc[1], desc[2], desc[3], desc[4], desc[5], desc[6], 48*5320918bSDave Airlie desc[7], desc[8], desc[9], desc[10]); 49*5320918bSDave Airlie 50*5320918bSDave Airlie if ((desc[0] != total_len) || /* descriptor length */ 51*5320918bSDave Airlie (desc[1] != 0x5f) || /* vendor descriptor type */ 52*5320918bSDave Airlie (desc[2] != 0x01) || /* version (2 bytes) */ 53*5320918bSDave Airlie (desc[3] != 0x00) || 54*5320918bSDave Airlie (desc[4] != total_len - 2)) /* length after type */ 55*5320918bSDave Airlie goto unrecognized; 56*5320918bSDave Airlie 57*5320918bSDave Airlie desc_end = desc + total_len; 58*5320918bSDave Airlie desc += 5; /* the fixed header we've already parsed */ 59*5320918bSDave Airlie 60*5320918bSDave Airlie while (desc < desc_end) { 61*5320918bSDave Airlie u8 length; 62*5320918bSDave Airlie u16 key; 63*5320918bSDave Airlie 64*5320918bSDave Airlie key = *((u16 *) desc); 65*5320918bSDave Airlie desc += sizeof(u16); 66*5320918bSDave Airlie length = *desc; 67*5320918bSDave Airlie desc++; 68*5320918bSDave Airlie 69*5320918bSDave Airlie switch (key) { 70*5320918bSDave Airlie case 0x0200: { /* max_area */ 71*5320918bSDave Airlie u32 max_area; 72*5320918bSDave Airlie max_area = le32_to_cpu(*((u32 *)desc)); 73*5320918bSDave Airlie DRM_DEBUG("DL chip limited to %d pixel modes\n", 74*5320918bSDave Airlie max_area); 75*5320918bSDave Airlie udl->sku_pixel_limit = max_area; 76*5320918bSDave Airlie break; 77*5320918bSDave Airlie } 78*5320918bSDave Airlie default: 79*5320918bSDave Airlie break; 80*5320918bSDave Airlie } 81*5320918bSDave Airlie desc += length; 82*5320918bSDave Airlie } 83*5320918bSDave Airlie } 84*5320918bSDave Airlie 85*5320918bSDave Airlie goto success; 86*5320918bSDave Airlie 87*5320918bSDave Airlie unrecognized: 88*5320918bSDave Airlie /* allow udlfb to load for now even if firmware unrecognized */ 89*5320918bSDave Airlie DRM_ERROR("Unrecognized vendor firmware descriptor\n"); 90*5320918bSDave Airlie 91*5320918bSDave Airlie success: 92*5320918bSDave Airlie kfree(buf); 93*5320918bSDave Airlie return true; 94*5320918bSDave Airlie } 95*5320918bSDave Airlie 96*5320918bSDave Airlie static void udl_release_urb_work(struct work_struct *work) 97*5320918bSDave Airlie { 98*5320918bSDave Airlie struct urb_node *unode = container_of(work, struct urb_node, 99*5320918bSDave Airlie release_urb_work.work); 100*5320918bSDave Airlie 101*5320918bSDave Airlie up(&unode->dev->urbs.limit_sem); 102*5320918bSDave Airlie } 103*5320918bSDave Airlie 104*5320918bSDave Airlie void udl_urb_completion(struct urb *urb) 105*5320918bSDave Airlie { 106*5320918bSDave Airlie struct urb_node *unode = urb->context; 107*5320918bSDave Airlie struct udl_device *udl = unode->dev; 108*5320918bSDave Airlie unsigned long flags; 109*5320918bSDave Airlie 110*5320918bSDave Airlie /* sync/async unlink faults aren't errors */ 111*5320918bSDave Airlie if (urb->status) { 112*5320918bSDave Airlie if (!(urb->status == -ENOENT || 113*5320918bSDave Airlie urb->status == -ECONNRESET || 114*5320918bSDave Airlie urb->status == -ESHUTDOWN)) { 115*5320918bSDave Airlie DRM_ERROR("%s - nonzero write bulk status received: %d\n", 116*5320918bSDave Airlie __func__, urb->status); 117*5320918bSDave Airlie atomic_set(&udl->lost_pixels, 1); 118*5320918bSDave Airlie } 119*5320918bSDave Airlie } 120*5320918bSDave Airlie 121*5320918bSDave Airlie urb->transfer_buffer_length = udl->urbs.size; /* reset to actual */ 122*5320918bSDave Airlie 123*5320918bSDave Airlie spin_lock_irqsave(&udl->urbs.lock, flags); 124*5320918bSDave Airlie list_add_tail(&unode->entry, &udl->urbs.list); 125*5320918bSDave Airlie udl->urbs.available++; 126*5320918bSDave Airlie spin_unlock_irqrestore(&udl->urbs.lock, flags); 127*5320918bSDave Airlie 128*5320918bSDave Airlie #if 0 129*5320918bSDave Airlie /* 130*5320918bSDave Airlie * When using fb_defio, we deadlock if up() is called 131*5320918bSDave Airlie * while another is waiting. So queue to another process. 132*5320918bSDave Airlie */ 133*5320918bSDave Airlie if (fb_defio) 134*5320918bSDave Airlie schedule_delayed_work(&unode->release_urb_work, 0); 135*5320918bSDave Airlie else 136*5320918bSDave Airlie #endif 137*5320918bSDave Airlie up(&udl->urbs.limit_sem); 138*5320918bSDave Airlie } 139*5320918bSDave Airlie 140*5320918bSDave Airlie static void udl_free_urb_list(struct drm_device *dev) 141*5320918bSDave Airlie { 142*5320918bSDave Airlie struct udl_device *udl = dev->dev_private; 143*5320918bSDave Airlie int count = udl->urbs.count; 144*5320918bSDave Airlie struct list_head *node; 145*5320918bSDave Airlie struct urb_node *unode; 146*5320918bSDave Airlie struct urb *urb; 147*5320918bSDave Airlie int ret; 148*5320918bSDave Airlie unsigned long flags; 149*5320918bSDave Airlie 150*5320918bSDave Airlie DRM_DEBUG("Waiting for completes and freeing all render urbs\n"); 151*5320918bSDave Airlie 152*5320918bSDave Airlie /* keep waiting and freeing, until we've got 'em all */ 153*5320918bSDave Airlie while (count--) { 154*5320918bSDave Airlie 155*5320918bSDave Airlie /* Getting interrupted means a leak, but ok at shutdown*/ 156*5320918bSDave Airlie ret = down_interruptible(&udl->urbs.limit_sem); 157*5320918bSDave Airlie if (ret) 158*5320918bSDave Airlie break; 159*5320918bSDave Airlie 160*5320918bSDave Airlie spin_lock_irqsave(&udl->urbs.lock, flags); 161*5320918bSDave Airlie 162*5320918bSDave Airlie node = udl->urbs.list.next; /* have reserved one with sem */ 163*5320918bSDave Airlie list_del_init(node); 164*5320918bSDave Airlie 165*5320918bSDave Airlie spin_unlock_irqrestore(&udl->urbs.lock, flags); 166*5320918bSDave Airlie 167*5320918bSDave Airlie unode = list_entry(node, struct urb_node, entry); 168*5320918bSDave Airlie urb = unode->urb; 169*5320918bSDave Airlie 170*5320918bSDave Airlie /* Free each separately allocated piece */ 171*5320918bSDave Airlie usb_free_coherent(urb->dev, udl->urbs.size, 172*5320918bSDave Airlie urb->transfer_buffer, urb->transfer_dma); 173*5320918bSDave Airlie usb_free_urb(urb); 174*5320918bSDave Airlie kfree(node); 175*5320918bSDave Airlie } 176*5320918bSDave Airlie udl->urbs.count = 0; 177*5320918bSDave Airlie } 178*5320918bSDave Airlie 179*5320918bSDave Airlie static int udl_alloc_urb_list(struct drm_device *dev, int count, size_t size) 180*5320918bSDave Airlie { 181*5320918bSDave Airlie struct udl_device *udl = dev->dev_private; 182*5320918bSDave Airlie int i = 0; 183*5320918bSDave Airlie struct urb *urb; 184*5320918bSDave Airlie struct urb_node *unode; 185*5320918bSDave Airlie char *buf; 186*5320918bSDave Airlie 187*5320918bSDave Airlie spin_lock_init(&udl->urbs.lock); 188*5320918bSDave Airlie 189*5320918bSDave Airlie udl->urbs.size = size; 190*5320918bSDave Airlie INIT_LIST_HEAD(&udl->urbs.list); 191*5320918bSDave Airlie 192*5320918bSDave Airlie while (i < count) { 193*5320918bSDave Airlie unode = kzalloc(sizeof(struct urb_node), GFP_KERNEL); 194*5320918bSDave Airlie if (!unode) 195*5320918bSDave Airlie break; 196*5320918bSDave Airlie unode->dev = udl; 197*5320918bSDave Airlie 198*5320918bSDave Airlie INIT_DELAYED_WORK(&unode->release_urb_work, 199*5320918bSDave Airlie udl_release_urb_work); 200*5320918bSDave Airlie 201*5320918bSDave Airlie urb = usb_alloc_urb(0, GFP_KERNEL); 202*5320918bSDave Airlie if (!urb) { 203*5320918bSDave Airlie kfree(unode); 204*5320918bSDave Airlie break; 205*5320918bSDave Airlie } 206*5320918bSDave Airlie unode->urb = urb; 207*5320918bSDave Airlie 208*5320918bSDave Airlie buf = usb_alloc_coherent(udl->ddev->usbdev, MAX_TRANSFER, GFP_KERNEL, 209*5320918bSDave Airlie &urb->transfer_dma); 210*5320918bSDave Airlie if (!buf) { 211*5320918bSDave Airlie kfree(unode); 212*5320918bSDave Airlie usb_free_urb(urb); 213*5320918bSDave Airlie break; 214*5320918bSDave Airlie } 215*5320918bSDave Airlie 216*5320918bSDave Airlie /* urb->transfer_buffer_length set to actual before submit */ 217*5320918bSDave Airlie usb_fill_bulk_urb(urb, udl->ddev->usbdev, usb_sndbulkpipe(udl->ddev->usbdev, 1), 218*5320918bSDave Airlie buf, size, udl_urb_completion, unode); 219*5320918bSDave Airlie urb->transfer_flags |= URB_NO_TRANSFER_DMA_MAP; 220*5320918bSDave Airlie 221*5320918bSDave Airlie list_add_tail(&unode->entry, &udl->urbs.list); 222*5320918bSDave Airlie 223*5320918bSDave Airlie i++; 224*5320918bSDave Airlie } 225*5320918bSDave Airlie 226*5320918bSDave Airlie sema_init(&udl->urbs.limit_sem, i); 227*5320918bSDave Airlie udl->urbs.count = i; 228*5320918bSDave Airlie udl->urbs.available = i; 229*5320918bSDave Airlie 230*5320918bSDave Airlie DRM_DEBUG("allocated %d %d byte urbs\n", i, (int) size); 231*5320918bSDave Airlie 232*5320918bSDave Airlie return i; 233*5320918bSDave Airlie } 234*5320918bSDave Airlie 235*5320918bSDave Airlie struct urb *udl_get_urb(struct drm_device *dev) 236*5320918bSDave Airlie { 237*5320918bSDave Airlie struct udl_device *udl = dev->dev_private; 238*5320918bSDave Airlie int ret = 0; 239*5320918bSDave Airlie struct list_head *entry; 240*5320918bSDave Airlie struct urb_node *unode; 241*5320918bSDave Airlie struct urb *urb = NULL; 242*5320918bSDave Airlie unsigned long flags; 243*5320918bSDave Airlie 244*5320918bSDave Airlie /* Wait for an in-flight buffer to complete and get re-queued */ 245*5320918bSDave Airlie ret = down_timeout(&udl->urbs.limit_sem, GET_URB_TIMEOUT); 246*5320918bSDave Airlie if (ret) { 247*5320918bSDave Airlie atomic_set(&udl->lost_pixels, 1); 248*5320918bSDave Airlie DRM_INFO("wait for urb interrupted: %x available: %d\n", 249*5320918bSDave Airlie ret, udl->urbs.available); 250*5320918bSDave Airlie goto error; 251*5320918bSDave Airlie } 252*5320918bSDave Airlie 253*5320918bSDave Airlie spin_lock_irqsave(&udl->urbs.lock, flags); 254*5320918bSDave Airlie 255*5320918bSDave Airlie BUG_ON(list_empty(&udl->urbs.list)); /* reserved one with limit_sem */ 256*5320918bSDave Airlie entry = udl->urbs.list.next; 257*5320918bSDave Airlie list_del_init(entry); 258*5320918bSDave Airlie udl->urbs.available--; 259*5320918bSDave Airlie 260*5320918bSDave Airlie spin_unlock_irqrestore(&udl->urbs.lock, flags); 261*5320918bSDave Airlie 262*5320918bSDave Airlie unode = list_entry(entry, struct urb_node, entry); 263*5320918bSDave Airlie urb = unode->urb; 264*5320918bSDave Airlie 265*5320918bSDave Airlie error: 266*5320918bSDave Airlie return urb; 267*5320918bSDave Airlie } 268*5320918bSDave Airlie 269*5320918bSDave Airlie int udl_submit_urb(struct drm_device *dev, struct urb *urb, size_t len) 270*5320918bSDave Airlie { 271*5320918bSDave Airlie struct udl_device *udl = dev->dev_private; 272*5320918bSDave Airlie int ret; 273*5320918bSDave Airlie 274*5320918bSDave Airlie BUG_ON(len > udl->urbs.size); 275*5320918bSDave Airlie 276*5320918bSDave Airlie urb->transfer_buffer_length = len; /* set to actual payload len */ 277*5320918bSDave Airlie ret = usb_submit_urb(urb, GFP_ATOMIC); 278*5320918bSDave Airlie if (ret) { 279*5320918bSDave Airlie udl_urb_completion(urb); /* because no one else will */ 280*5320918bSDave Airlie atomic_set(&udl->lost_pixels, 1); 281*5320918bSDave Airlie DRM_ERROR("usb_submit_urb error %x\n", ret); 282*5320918bSDave Airlie } 283*5320918bSDave Airlie return ret; 284*5320918bSDave Airlie } 285*5320918bSDave Airlie 286*5320918bSDave Airlie int udl_driver_load(struct drm_device *dev, unsigned long flags) 287*5320918bSDave Airlie { 288*5320918bSDave Airlie struct udl_device *udl; 289*5320918bSDave Airlie int ret; 290*5320918bSDave Airlie 291*5320918bSDave Airlie DRM_DEBUG("\n"); 292*5320918bSDave Airlie udl = kzalloc(sizeof(struct udl_device), GFP_KERNEL); 293*5320918bSDave Airlie if (!udl) 294*5320918bSDave Airlie return -ENOMEM; 295*5320918bSDave Airlie 296*5320918bSDave Airlie udl->ddev = dev; 297*5320918bSDave Airlie dev->dev_private = udl; 298*5320918bSDave Airlie 299*5320918bSDave Airlie if (!udl_parse_vendor_descriptor(dev, dev->usbdev)) { 300*5320918bSDave Airlie DRM_ERROR("firmware not recognized. Assume incompatible device\n"); 301*5320918bSDave Airlie goto err; 302*5320918bSDave Airlie } 303*5320918bSDave Airlie 304*5320918bSDave Airlie if (!udl_alloc_urb_list(dev, WRITES_IN_FLIGHT, MAX_TRANSFER)) { 305*5320918bSDave Airlie ret = -ENOMEM; 306*5320918bSDave Airlie DRM_ERROR("udl_alloc_urb_list failed\n"); 307*5320918bSDave Airlie goto err; 308*5320918bSDave Airlie } 309*5320918bSDave Airlie 310*5320918bSDave Airlie DRM_DEBUG("\n"); 311*5320918bSDave Airlie ret = udl_modeset_init(dev); 312*5320918bSDave Airlie 313*5320918bSDave Airlie ret = udl_fbdev_init(dev); 314*5320918bSDave Airlie return 0; 315*5320918bSDave Airlie err: 316*5320918bSDave Airlie kfree(udl); 317*5320918bSDave Airlie DRM_ERROR("%d\n", ret); 318*5320918bSDave Airlie return ret; 319*5320918bSDave Airlie } 320*5320918bSDave Airlie 321*5320918bSDave Airlie int udl_drop_usb(struct drm_device *dev) 322*5320918bSDave Airlie { 323*5320918bSDave Airlie udl_free_urb_list(dev); 324*5320918bSDave Airlie return 0; 325*5320918bSDave Airlie } 326*5320918bSDave Airlie 327*5320918bSDave Airlie int udl_driver_unload(struct drm_device *dev) 328*5320918bSDave Airlie { 329*5320918bSDave Airlie struct udl_device *udl = dev->dev_private; 330*5320918bSDave Airlie 331*5320918bSDave Airlie if (udl->urbs.count) 332*5320918bSDave Airlie udl_free_urb_list(dev); 333*5320918bSDave Airlie 334*5320918bSDave Airlie udl_fbdev_cleanup(dev); 335*5320918bSDave Airlie udl_modeset_cleanup(dev); 336*5320918bSDave Airlie kfree(udl); 337*5320918bSDave Airlie return 0; 338*5320918bSDave Airlie } 339