1 /* 2 * f_mass_storage.c -- Mass Storage USB Composite Function 3 * 4 * Copyright (C) 2003-2008 Alan Stern 5 * Copyright (C) 2009 Samsung Electronics 6 * Author: Michal Nazarewicz <mina86@mina86.com> 7 * All rights reserved. 8 * 9 * Redistribution and use in source and binary forms, with or without 10 * modification, are permitted provided that the following conditions 11 * are met: 12 * 1. Redistributions of source code must retain the above copyright 13 * notice, this list of conditions, and the following disclaimer, 14 * without modification. 15 * 2. Redistributions in binary form must reproduce the above copyright 16 * notice, this list of conditions and the following disclaimer in the 17 * documentation and/or other materials provided with the distribution. 18 * 3. The names of the above-listed copyright holders may not be used 19 * to endorse or promote products derived from this software without 20 * specific prior written permission. 21 * 22 * ALTERNATIVELY, this software may be distributed under the terms of the 23 * GNU General Public License ("GPL") as published by the Free Software 24 * Foundation, either version 2 of that License or (at your option) any 25 * later version. 26 * 27 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS 28 * IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, 29 * THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR 30 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR 31 * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, 32 * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, 33 * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR 34 * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF 35 * LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING 36 * NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS 37 * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 38 */ 39 40 /* 41 * The Mass Storage Function acts as a USB Mass Storage device, 42 * appearing to the host as a disk drive or as a CD-ROM drive. In 43 * addition to providing an example of a genuinely useful composite 44 * function for a USB device, it also illustrates a technique of 45 * double-buffering for increased throughput. 46 * 47 * For more information about MSF and in particular its module 48 * parameters and sysfs interface read the 49 * <Documentation/usb/mass-storage.txt> file. 50 */ 51 52 /* 53 * MSF is configured by specifying a fsg_config structure. It has the 54 * following fields: 55 * 56 * nluns Number of LUNs function have (anywhere from 1 57 * to FSG_MAX_LUNS). 58 * luns An array of LUN configuration values. This 59 * should be filled for each LUN that 60 * function will include (ie. for "nluns" 61 * LUNs). Each element of the array has 62 * the following fields: 63 * ->filename The path to the backing file for the LUN. 64 * Required if LUN is not marked as 65 * removable. 66 * ->ro Flag specifying access to the LUN shall be 67 * read-only. This is implied if CD-ROM 68 * emulation is enabled as well as when 69 * it was impossible to open "filename" 70 * in R/W mode. 71 * ->removable Flag specifying that LUN shall be indicated as 72 * being removable. 73 * ->cdrom Flag specifying that LUN shall be reported as 74 * being a CD-ROM. 75 * ->nofua Flag specifying that FUA flag in SCSI WRITE(10,12) 76 * commands for this LUN shall be ignored. 77 * 78 * vendor_name 79 * product_name 80 * release Information used as a reply to INQUIRY 81 * request. To use default set to NULL, 82 * NULL, 0xffff respectively. The first 83 * field should be 8 and the second 16 84 * characters or less. 85 * 86 * can_stall Set to permit function to halt bulk endpoints. 87 * Disabled on some USB devices known not 88 * to work correctly. You should set it 89 * to true. 90 * 91 * If "removable" is not set for a LUN then a backing file must be 92 * specified. If it is set, then NULL filename means the LUN's medium 93 * is not loaded (an empty string as "filename" in the fsg_config 94 * structure causes error). The CD-ROM emulation includes a single 95 * data track and no audio tracks; hence there need be only one 96 * backing file per LUN. 97 * 98 * This function is heavily based on "File-backed Storage Gadget" by 99 * Alan Stern which in turn is heavily based on "Gadget Zero" by David 100 * Brownell. The driver's SCSI command interface was based on the 101 * "Information technology - Small Computer System Interface - 2" 102 * document from X3T9.2 Project 375D, Revision 10L, 7-SEP-93, 103 * available at <http://www.t10.org/ftp/t10/drafts/s2/s2-r10l.pdf>. 104 * The single exception is opcode 0x23 (READ FORMAT CAPACITIES), which 105 * was based on the "Universal Serial Bus Mass Storage Class UFI 106 * Command Specification" document, Revision 1.0, December 14, 1998, 107 * available at 108 * <http://www.usb.org/developers/devclass_docs/usbmass-ufi10.pdf>. 109 */ 110 111 /* 112 * Driver Design 113 * 114 * The MSF is fairly straightforward. There is a main kernel 115 * thread that handles most of the work. Interrupt routines field 116 * callbacks from the controller driver: bulk- and interrupt-request 117 * completion notifications, endpoint-0 events, and disconnect events. 118 * Completion events are passed to the main thread by wakeup calls. Many 119 * ep0 requests are handled at interrupt time, but SetInterface, 120 * SetConfiguration, and device reset requests are forwarded to the 121 * thread in the form of "exceptions" using SIGUSR1 signals (since they 122 * should interrupt any ongoing file I/O operations). 123 * 124 * The thread's main routine implements the standard command/data/status 125 * parts of a SCSI interaction. It and its subroutines are full of tests 126 * for pending signals/exceptions -- all this polling is necessary since 127 * the kernel has no setjmp/longjmp equivalents. (Maybe this is an 128 * indication that the driver really wants to be running in userspace.) 129 * An important point is that so long as the thread is alive it keeps an 130 * open reference to the backing file. This will prevent unmounting 131 * the backing file's underlying filesystem and could cause problems 132 * during system shutdown, for example. To prevent such problems, the 133 * thread catches INT, TERM, and KILL signals and converts them into 134 * an EXIT exception. 135 * 136 * In normal operation the main thread is started during the gadget's 137 * fsg_bind() callback and stopped during fsg_unbind(). But it can 138 * also exit when it receives a signal, and there's no point leaving 139 * the gadget running when the thread is dead. As of this moment, MSF 140 * provides no way to deregister the gadget when thread dies -- maybe 141 * a callback functions is needed. 142 * 143 * To provide maximum throughput, the driver uses a circular pipeline of 144 * buffer heads (struct fsg_buffhd). In principle the pipeline can be 145 * arbitrarily long; in practice the benefits don't justify having more 146 * than 2 stages (i.e., double buffering). But it helps to think of the 147 * pipeline as being a long one. Each buffer head contains a bulk-in and 148 * a bulk-out request pointer (since the buffer can be used for both 149 * output and input -- directions always are given from the host's 150 * point of view) as well as a pointer to the buffer and various state 151 * variables. 152 * 153 * Use of the pipeline follows a simple protocol. There is a variable 154 * (fsg->next_buffhd_to_fill) that points to the next buffer head to use. 155 * At any time that buffer head may still be in use from an earlier 156 * request, so each buffer head has a state variable indicating whether 157 * it is EMPTY, FULL, or BUSY. Typical use involves waiting for the 158 * buffer head to be EMPTY, filling the buffer either by file I/O or by 159 * USB I/O (during which the buffer head is BUSY), and marking the buffer 160 * head FULL when the I/O is complete. Then the buffer will be emptied 161 * (again possibly by USB I/O, during which it is marked BUSY) and 162 * finally marked EMPTY again (possibly by a completion routine). 163 * 164 * A module parameter tells the driver to avoid stalling the bulk 165 * endpoints wherever the transport specification allows. This is 166 * necessary for some UDCs like the SuperH, which cannot reliably clear a 167 * halt on a bulk endpoint. However, under certain circumstances the 168 * Bulk-only specification requires a stall. In such cases the driver 169 * will halt the endpoint and set a flag indicating that it should clear 170 * the halt in software during the next device reset. Hopefully this 171 * will permit everything to work correctly. Furthermore, although the 172 * specification allows the bulk-out endpoint to halt when the host sends 173 * too much data, implementing this would cause an unavoidable race. 174 * The driver will always use the "no-stall" approach for OUT transfers. 175 * 176 * One subtle point concerns sending status-stage responses for ep0 177 * requests. Some of these requests, such as device reset, can involve 178 * interrupting an ongoing file I/O operation, which might take an 179 * arbitrarily long time. During that delay the host might give up on 180 * the original ep0 request and issue a new one. When that happens the 181 * driver should not notify the host about completion of the original 182 * request, as the host will no longer be waiting for it. So the driver 183 * assigns to each ep0 request a unique tag, and it keeps track of the 184 * tag value of the request associated with a long-running exception 185 * (device-reset, interface-change, or configuration-change). When the 186 * exception handler is finished, the status-stage response is submitted 187 * only if the current ep0 request tag is equal to the exception request 188 * tag. Thus only the most recently received ep0 request will get a 189 * status-stage response. 190 * 191 * Warning: This driver source file is too long. It ought to be split up 192 * into a header file plus about 3 separate .c files, to handle the details 193 * of the Gadget, USB Mass Storage, and SCSI protocols. 194 */ 195 196 197 /* #define VERBOSE_DEBUG */ 198 /* #define DUMP_MSGS */ 199 200 #include <linux/blkdev.h> 201 #include <linux/completion.h> 202 #include <linux/dcache.h> 203 #include <linux/delay.h> 204 #include <linux/device.h> 205 #include <linux/fcntl.h> 206 #include <linux/file.h> 207 #include <linux/fs.h> 208 #include <linux/kref.h> 209 #include <linux/kthread.h> 210 #include <linux/sched/signal.h> 211 #include <linux/limits.h> 212 #include <linux/rwsem.h> 213 #include <linux/slab.h> 214 #include <linux/spinlock.h> 215 #include <linux/string.h> 216 #include <linux/freezer.h> 217 #include <linux/module.h> 218 #include <linux/uaccess.h> 219 220 #include <linux/usb/ch9.h> 221 #include <linux/usb/gadget.h> 222 #include <linux/usb/composite.h> 223 224 #include "configfs.h" 225 226 227 /*------------------------------------------------------------------------*/ 228 229 #define FSG_DRIVER_DESC "Mass Storage Function" 230 #define FSG_DRIVER_VERSION "2009/09/11" 231 232 static const char fsg_string_interface[] = "Mass Storage"; 233 234 #include "storage_common.h" 235 #include "f_mass_storage.h" 236 237 /* Static strings, in UTF-8 (for simplicity we use only ASCII characters) */ 238 static struct usb_string fsg_strings[] = { 239 {FSG_STRING_INTERFACE, fsg_string_interface}, 240 {} 241 }; 242 243 static struct usb_gadget_strings fsg_stringtab = { 244 .language = 0x0409, /* en-us */ 245 .strings = fsg_strings, 246 }; 247 248 static struct usb_gadget_strings *fsg_strings_array[] = { 249 &fsg_stringtab, 250 NULL, 251 }; 252 253 /*-------------------------------------------------------------------------*/ 254 255 struct fsg_dev; 256 struct fsg_common; 257 258 /* Data shared by all the FSG instances. */ 259 struct fsg_common { 260 struct usb_gadget *gadget; 261 struct usb_composite_dev *cdev; 262 struct fsg_dev *fsg, *new_fsg; 263 wait_queue_head_t fsg_wait; 264 265 /* filesem protects: backing files in use */ 266 struct rw_semaphore filesem; 267 268 /* lock protects: state, all the req_busy's */ 269 spinlock_t lock; 270 271 struct usb_ep *ep0; /* Copy of gadget->ep0 */ 272 struct usb_request *ep0req; /* Copy of cdev->req */ 273 unsigned int ep0_req_tag; 274 275 struct fsg_buffhd *next_buffhd_to_fill; 276 struct fsg_buffhd *next_buffhd_to_drain; 277 struct fsg_buffhd *buffhds; 278 unsigned int fsg_num_buffers; 279 280 int cmnd_size; 281 u8 cmnd[MAX_COMMAND_SIZE]; 282 283 unsigned int lun; 284 struct fsg_lun *luns[FSG_MAX_LUNS]; 285 struct fsg_lun *curlun; 286 287 unsigned int bulk_out_maxpacket; 288 enum fsg_state state; /* For exception handling */ 289 unsigned int exception_req_tag; 290 291 enum data_direction data_dir; 292 u32 data_size; 293 u32 data_size_from_cmnd; 294 u32 tag; 295 u32 residue; 296 u32 usb_amount_left; 297 298 unsigned int can_stall:1; 299 unsigned int free_storage_on_release:1; 300 unsigned int phase_error:1; 301 unsigned int short_packet_received:1; 302 unsigned int bad_lun_okay:1; 303 unsigned int running:1; 304 unsigned int sysfs:1; 305 306 int thread_wakeup_needed; 307 struct completion thread_notifier; 308 struct task_struct *thread_task; 309 310 /* Callback functions. */ 311 const struct fsg_operations *ops; 312 /* Gadget's private data. */ 313 void *private_data; 314 315 char inquiry_string[INQUIRY_STRING_LEN]; 316 317 struct kref ref; 318 }; 319 320 struct fsg_dev { 321 struct usb_function function; 322 struct usb_gadget *gadget; /* Copy of cdev->gadget */ 323 struct fsg_common *common; 324 325 u16 interface_number; 326 327 unsigned int bulk_in_enabled:1; 328 unsigned int bulk_out_enabled:1; 329 330 unsigned long atomic_bitflags; 331 #define IGNORE_BULK_OUT 0 332 333 struct usb_ep *bulk_in; 334 struct usb_ep *bulk_out; 335 }; 336 337 static inline int __fsg_is_set(struct fsg_common *common, 338 const char *func, unsigned line) 339 { 340 if (common->fsg) 341 return 1; 342 ERROR(common, "common->fsg is NULL in %s at %u\n", func, line); 343 WARN_ON(1); 344 return 0; 345 } 346 347 #define fsg_is_set(common) likely(__fsg_is_set(common, __func__, __LINE__)) 348 349 static inline struct fsg_dev *fsg_from_func(struct usb_function *f) 350 { 351 return container_of(f, struct fsg_dev, function); 352 } 353 354 typedef void (*fsg_routine_t)(struct fsg_dev *); 355 356 static int exception_in_progress(struct fsg_common *common) 357 { 358 return common->state > FSG_STATE_IDLE; 359 } 360 361 /* Make bulk-out requests be divisible by the maxpacket size */ 362 static void set_bulk_out_req_length(struct fsg_common *common, 363 struct fsg_buffhd *bh, unsigned int length) 364 { 365 unsigned int rem; 366 367 bh->bulk_out_intended_length = length; 368 rem = length % common->bulk_out_maxpacket; 369 if (rem > 0) 370 length += common->bulk_out_maxpacket - rem; 371 bh->outreq->length = length; 372 } 373 374 375 /*-------------------------------------------------------------------------*/ 376 377 static int fsg_set_halt(struct fsg_dev *fsg, struct usb_ep *ep) 378 { 379 const char *name; 380 381 if (ep == fsg->bulk_in) 382 name = "bulk-in"; 383 else if (ep == fsg->bulk_out) 384 name = "bulk-out"; 385 else 386 name = ep->name; 387 DBG(fsg, "%s set halt\n", name); 388 return usb_ep_set_halt(ep); 389 } 390 391 392 /*-------------------------------------------------------------------------*/ 393 394 /* These routines may be called in process context or in_irq */ 395 396 /* Caller must hold fsg->lock */ 397 static void wakeup_thread(struct fsg_common *common) 398 { 399 /* 400 * Ensure the reading of thread_wakeup_needed 401 * and the writing of bh->state are completed 402 */ 403 smp_mb(); 404 /* Tell the main thread that something has happened */ 405 common->thread_wakeup_needed = 1; 406 if (common->thread_task) 407 wake_up_process(common->thread_task); 408 } 409 410 static void raise_exception(struct fsg_common *common, enum fsg_state new_state) 411 { 412 unsigned long flags; 413 414 /* 415 * Do nothing if a higher-priority exception is already in progress. 416 * If a lower-or-equal priority exception is in progress, preempt it 417 * and notify the main thread by sending it a signal. 418 */ 419 spin_lock_irqsave(&common->lock, flags); 420 if (common->state <= new_state) { 421 common->exception_req_tag = common->ep0_req_tag; 422 common->state = new_state; 423 if (common->thread_task) 424 send_sig_info(SIGUSR1, SEND_SIG_FORCED, 425 common->thread_task); 426 } 427 spin_unlock_irqrestore(&common->lock, flags); 428 } 429 430 431 /*-------------------------------------------------------------------------*/ 432 433 static int ep0_queue(struct fsg_common *common) 434 { 435 int rc; 436 437 rc = usb_ep_queue(common->ep0, common->ep0req, GFP_ATOMIC); 438 common->ep0->driver_data = common; 439 if (rc != 0 && rc != -ESHUTDOWN) { 440 /* We can't do much more than wait for a reset */ 441 WARNING(common, "error in submission: %s --> %d\n", 442 common->ep0->name, rc); 443 } 444 return rc; 445 } 446 447 448 /*-------------------------------------------------------------------------*/ 449 450 /* Completion handlers. These always run in_irq. */ 451 452 static void bulk_in_complete(struct usb_ep *ep, struct usb_request *req) 453 { 454 struct fsg_common *common = ep->driver_data; 455 struct fsg_buffhd *bh = req->context; 456 457 if (req->status || req->actual != req->length) 458 DBG(common, "%s --> %d, %u/%u\n", __func__, 459 req->status, req->actual, req->length); 460 if (req->status == -ECONNRESET) /* Request was cancelled */ 461 usb_ep_fifo_flush(ep); 462 463 /* Hold the lock while we update the request and buffer states */ 464 smp_wmb(); 465 spin_lock(&common->lock); 466 bh->inreq_busy = 0; 467 bh->state = BUF_STATE_EMPTY; 468 wakeup_thread(common); 469 spin_unlock(&common->lock); 470 } 471 472 static void bulk_out_complete(struct usb_ep *ep, struct usb_request *req) 473 { 474 struct fsg_common *common = ep->driver_data; 475 struct fsg_buffhd *bh = req->context; 476 477 dump_msg(common, "bulk-out", req->buf, req->actual); 478 if (req->status || req->actual != bh->bulk_out_intended_length) 479 DBG(common, "%s --> %d, %u/%u\n", __func__, 480 req->status, req->actual, bh->bulk_out_intended_length); 481 if (req->status == -ECONNRESET) /* Request was cancelled */ 482 usb_ep_fifo_flush(ep); 483 484 /* Hold the lock while we update the request and buffer states */ 485 smp_wmb(); 486 spin_lock(&common->lock); 487 bh->outreq_busy = 0; 488 bh->state = BUF_STATE_FULL; 489 wakeup_thread(common); 490 spin_unlock(&common->lock); 491 } 492 493 static int _fsg_common_get_max_lun(struct fsg_common *common) 494 { 495 int i = ARRAY_SIZE(common->luns) - 1; 496 497 while (i >= 0 && !common->luns[i]) 498 --i; 499 500 return i; 501 } 502 503 static int fsg_setup(struct usb_function *f, 504 const struct usb_ctrlrequest *ctrl) 505 { 506 struct fsg_dev *fsg = fsg_from_func(f); 507 struct usb_request *req = fsg->common->ep0req; 508 u16 w_index = le16_to_cpu(ctrl->wIndex); 509 u16 w_value = le16_to_cpu(ctrl->wValue); 510 u16 w_length = le16_to_cpu(ctrl->wLength); 511 512 if (!fsg_is_set(fsg->common)) 513 return -EOPNOTSUPP; 514 515 ++fsg->common->ep0_req_tag; /* Record arrival of a new request */ 516 req->context = NULL; 517 req->length = 0; 518 dump_msg(fsg, "ep0-setup", (u8 *) ctrl, sizeof(*ctrl)); 519 520 switch (ctrl->bRequest) { 521 522 case US_BULK_RESET_REQUEST: 523 if (ctrl->bRequestType != 524 (USB_DIR_OUT | USB_TYPE_CLASS | USB_RECIP_INTERFACE)) 525 break; 526 if (w_index != fsg->interface_number || w_value != 0 || 527 w_length != 0) 528 return -EDOM; 529 530 /* 531 * Raise an exception to stop the current operation 532 * and reinitialize our state. 533 */ 534 DBG(fsg, "bulk reset request\n"); 535 raise_exception(fsg->common, FSG_STATE_RESET); 536 return USB_GADGET_DELAYED_STATUS; 537 538 case US_BULK_GET_MAX_LUN: 539 if (ctrl->bRequestType != 540 (USB_DIR_IN | USB_TYPE_CLASS | USB_RECIP_INTERFACE)) 541 break; 542 if (w_index != fsg->interface_number || w_value != 0 || 543 w_length != 1) 544 return -EDOM; 545 VDBG(fsg, "get max LUN\n"); 546 *(u8 *)req->buf = _fsg_common_get_max_lun(fsg->common); 547 548 /* Respond with data/status */ 549 req->length = min((u16)1, w_length); 550 return ep0_queue(fsg->common); 551 } 552 553 VDBG(fsg, 554 "unknown class-specific control req %02x.%02x v%04x i%04x l%u\n", 555 ctrl->bRequestType, ctrl->bRequest, 556 le16_to_cpu(ctrl->wValue), w_index, w_length); 557 return -EOPNOTSUPP; 558 } 559 560 561 /*-------------------------------------------------------------------------*/ 562 563 /* All the following routines run in process context */ 564 565 /* Use this for bulk or interrupt transfers, not ep0 */ 566 static void start_transfer(struct fsg_dev *fsg, struct usb_ep *ep, 567 struct usb_request *req, int *pbusy, 568 enum fsg_buffer_state *state) 569 { 570 int rc; 571 572 if (ep == fsg->bulk_in) 573 dump_msg(fsg, "bulk-in", req->buf, req->length); 574 575 spin_lock_irq(&fsg->common->lock); 576 *pbusy = 1; 577 *state = BUF_STATE_BUSY; 578 spin_unlock_irq(&fsg->common->lock); 579 580 rc = usb_ep_queue(ep, req, GFP_KERNEL); 581 if (rc == 0) 582 return; /* All good, we're done */ 583 584 *pbusy = 0; 585 *state = BUF_STATE_EMPTY; 586 587 /* We can't do much more than wait for a reset */ 588 589 /* 590 * Note: currently the net2280 driver fails zero-length 591 * submissions if DMA is enabled. 592 */ 593 if (rc != -ESHUTDOWN && !(rc == -EOPNOTSUPP && req->length == 0)) 594 WARNING(fsg, "error in submission: %s --> %d\n", ep->name, rc); 595 } 596 597 static bool start_in_transfer(struct fsg_common *common, struct fsg_buffhd *bh) 598 { 599 if (!fsg_is_set(common)) 600 return false; 601 start_transfer(common->fsg, common->fsg->bulk_in, 602 bh->inreq, &bh->inreq_busy, &bh->state); 603 return true; 604 } 605 606 static bool start_out_transfer(struct fsg_common *common, struct fsg_buffhd *bh) 607 { 608 if (!fsg_is_set(common)) 609 return false; 610 start_transfer(common->fsg, common->fsg->bulk_out, 611 bh->outreq, &bh->outreq_busy, &bh->state); 612 return true; 613 } 614 615 static int sleep_thread(struct fsg_common *common, bool can_freeze) 616 { 617 int rc = 0; 618 619 /* Wait until a signal arrives or we are woken up */ 620 for (;;) { 621 if (can_freeze) 622 try_to_freeze(); 623 set_current_state(TASK_INTERRUPTIBLE); 624 if (signal_pending(current)) { 625 rc = -EINTR; 626 break; 627 } 628 if (common->thread_wakeup_needed) 629 break; 630 schedule(); 631 } 632 __set_current_state(TASK_RUNNING); 633 common->thread_wakeup_needed = 0; 634 635 /* 636 * Ensure the writing of thread_wakeup_needed 637 * and the reading of bh->state are completed 638 */ 639 smp_mb(); 640 return rc; 641 } 642 643 644 /*-------------------------------------------------------------------------*/ 645 646 static int do_read(struct fsg_common *common) 647 { 648 struct fsg_lun *curlun = common->curlun; 649 u32 lba; 650 struct fsg_buffhd *bh; 651 int rc; 652 u32 amount_left; 653 loff_t file_offset, file_offset_tmp; 654 unsigned int amount; 655 ssize_t nread; 656 657 /* 658 * Get the starting Logical Block Address and check that it's 659 * not too big. 660 */ 661 if (common->cmnd[0] == READ_6) 662 lba = get_unaligned_be24(&common->cmnd[1]); 663 else { 664 lba = get_unaligned_be32(&common->cmnd[2]); 665 666 /* 667 * We allow DPO (Disable Page Out = don't save data in the 668 * cache) and FUA (Force Unit Access = don't read from the 669 * cache), but we don't implement them. 670 */ 671 if ((common->cmnd[1] & ~0x18) != 0) { 672 curlun->sense_data = SS_INVALID_FIELD_IN_CDB; 673 return -EINVAL; 674 } 675 } 676 if (lba >= curlun->num_sectors) { 677 curlun->sense_data = SS_LOGICAL_BLOCK_ADDRESS_OUT_OF_RANGE; 678 return -EINVAL; 679 } 680 file_offset = ((loff_t) lba) << curlun->blkbits; 681 682 /* Carry out the file reads */ 683 amount_left = common->data_size_from_cmnd; 684 if (unlikely(amount_left == 0)) 685 return -EIO; /* No default reply */ 686 687 for (;;) { 688 /* 689 * Figure out how much we need to read: 690 * Try to read the remaining amount. 691 * But don't read more than the buffer size. 692 * And don't try to read past the end of the file. 693 */ 694 amount = min(amount_left, FSG_BUFLEN); 695 amount = min((loff_t)amount, 696 curlun->file_length - file_offset); 697 698 /* Wait for the next buffer to become available */ 699 bh = common->next_buffhd_to_fill; 700 while (bh->state != BUF_STATE_EMPTY) { 701 rc = sleep_thread(common, false); 702 if (rc) 703 return rc; 704 } 705 706 /* 707 * If we were asked to read past the end of file, 708 * end with an empty buffer. 709 */ 710 if (amount == 0) { 711 curlun->sense_data = 712 SS_LOGICAL_BLOCK_ADDRESS_OUT_OF_RANGE; 713 curlun->sense_data_info = 714 file_offset >> curlun->blkbits; 715 curlun->info_valid = 1; 716 bh->inreq->length = 0; 717 bh->state = BUF_STATE_FULL; 718 break; 719 } 720 721 /* Perform the read */ 722 file_offset_tmp = file_offset; 723 nread = vfs_read(curlun->filp, 724 (char __user *)bh->buf, 725 amount, &file_offset_tmp); 726 VLDBG(curlun, "file read %u @ %llu -> %d\n", amount, 727 (unsigned long long)file_offset, (int)nread); 728 if (signal_pending(current)) 729 return -EINTR; 730 731 if (nread < 0) { 732 LDBG(curlun, "error in file read: %d\n", (int)nread); 733 nread = 0; 734 } else if (nread < amount) { 735 LDBG(curlun, "partial file read: %d/%u\n", 736 (int)nread, amount); 737 nread = round_down(nread, curlun->blksize); 738 } 739 file_offset += nread; 740 amount_left -= nread; 741 common->residue -= nread; 742 743 /* 744 * Except at the end of the transfer, nread will be 745 * equal to the buffer size, which is divisible by the 746 * bulk-in maxpacket size. 747 */ 748 bh->inreq->length = nread; 749 bh->state = BUF_STATE_FULL; 750 751 /* If an error occurred, report it and its position */ 752 if (nread < amount) { 753 curlun->sense_data = SS_UNRECOVERED_READ_ERROR; 754 curlun->sense_data_info = 755 file_offset >> curlun->blkbits; 756 curlun->info_valid = 1; 757 break; 758 } 759 760 if (amount_left == 0) 761 break; /* No more left to read */ 762 763 /* Send this buffer and go read some more */ 764 bh->inreq->zero = 0; 765 if (!start_in_transfer(common, bh)) 766 /* Don't know what to do if common->fsg is NULL */ 767 return -EIO; 768 common->next_buffhd_to_fill = bh->next; 769 } 770 771 return -EIO; /* No default reply */ 772 } 773 774 775 /*-------------------------------------------------------------------------*/ 776 777 static int do_write(struct fsg_common *common) 778 { 779 struct fsg_lun *curlun = common->curlun; 780 u32 lba; 781 struct fsg_buffhd *bh; 782 int get_some_more; 783 u32 amount_left_to_req, amount_left_to_write; 784 loff_t usb_offset, file_offset, file_offset_tmp; 785 unsigned int amount; 786 ssize_t nwritten; 787 int rc; 788 789 if (curlun->ro) { 790 curlun->sense_data = SS_WRITE_PROTECTED; 791 return -EINVAL; 792 } 793 spin_lock(&curlun->filp->f_lock); 794 curlun->filp->f_flags &= ~O_SYNC; /* Default is not to wait */ 795 spin_unlock(&curlun->filp->f_lock); 796 797 /* 798 * Get the starting Logical Block Address and check that it's 799 * not too big 800 */ 801 if (common->cmnd[0] == WRITE_6) 802 lba = get_unaligned_be24(&common->cmnd[1]); 803 else { 804 lba = get_unaligned_be32(&common->cmnd[2]); 805 806 /* 807 * We allow DPO (Disable Page Out = don't save data in the 808 * cache) and FUA (Force Unit Access = write directly to the 809 * medium). We don't implement DPO; we implement FUA by 810 * performing synchronous output. 811 */ 812 if (common->cmnd[1] & ~0x18) { 813 curlun->sense_data = SS_INVALID_FIELD_IN_CDB; 814 return -EINVAL; 815 } 816 if (!curlun->nofua && (common->cmnd[1] & 0x08)) { /* FUA */ 817 spin_lock(&curlun->filp->f_lock); 818 curlun->filp->f_flags |= O_SYNC; 819 spin_unlock(&curlun->filp->f_lock); 820 } 821 } 822 if (lba >= curlun->num_sectors) { 823 curlun->sense_data = SS_LOGICAL_BLOCK_ADDRESS_OUT_OF_RANGE; 824 return -EINVAL; 825 } 826 827 /* Carry out the file writes */ 828 get_some_more = 1; 829 file_offset = usb_offset = ((loff_t) lba) << curlun->blkbits; 830 amount_left_to_req = common->data_size_from_cmnd; 831 amount_left_to_write = common->data_size_from_cmnd; 832 833 while (amount_left_to_write > 0) { 834 835 /* Queue a request for more data from the host */ 836 bh = common->next_buffhd_to_fill; 837 if (bh->state == BUF_STATE_EMPTY && get_some_more) { 838 839 /* 840 * Figure out how much we want to get: 841 * Try to get the remaining amount, 842 * but not more than the buffer size. 843 */ 844 amount = min(amount_left_to_req, FSG_BUFLEN); 845 846 /* Beyond the end of the backing file? */ 847 if (usb_offset >= curlun->file_length) { 848 get_some_more = 0; 849 curlun->sense_data = 850 SS_LOGICAL_BLOCK_ADDRESS_OUT_OF_RANGE; 851 curlun->sense_data_info = 852 usb_offset >> curlun->blkbits; 853 curlun->info_valid = 1; 854 continue; 855 } 856 857 /* Get the next buffer */ 858 usb_offset += amount; 859 common->usb_amount_left -= amount; 860 amount_left_to_req -= amount; 861 if (amount_left_to_req == 0) 862 get_some_more = 0; 863 864 /* 865 * Except at the end of the transfer, amount will be 866 * equal to the buffer size, which is divisible by 867 * the bulk-out maxpacket size. 868 */ 869 set_bulk_out_req_length(common, bh, amount); 870 if (!start_out_transfer(common, bh)) 871 /* Dunno what to do if common->fsg is NULL */ 872 return -EIO; 873 common->next_buffhd_to_fill = bh->next; 874 continue; 875 } 876 877 /* Write the received data to the backing file */ 878 bh = common->next_buffhd_to_drain; 879 if (bh->state == BUF_STATE_EMPTY && !get_some_more) 880 break; /* We stopped early */ 881 if (bh->state == BUF_STATE_FULL) { 882 smp_rmb(); 883 common->next_buffhd_to_drain = bh->next; 884 bh->state = BUF_STATE_EMPTY; 885 886 /* Did something go wrong with the transfer? */ 887 if (bh->outreq->status != 0) { 888 curlun->sense_data = SS_COMMUNICATION_FAILURE; 889 curlun->sense_data_info = 890 file_offset >> curlun->blkbits; 891 curlun->info_valid = 1; 892 break; 893 } 894 895 amount = bh->outreq->actual; 896 if (curlun->file_length - file_offset < amount) { 897 LERROR(curlun, 898 "write %u @ %llu beyond end %llu\n", 899 amount, (unsigned long long)file_offset, 900 (unsigned long long)curlun->file_length); 901 amount = curlun->file_length - file_offset; 902 } 903 904 /* Don't accept excess data. The spec doesn't say 905 * what to do in this case. We'll ignore the error. 906 */ 907 amount = min(amount, bh->bulk_out_intended_length); 908 909 /* Don't write a partial block */ 910 amount = round_down(amount, curlun->blksize); 911 if (amount == 0) 912 goto empty_write; 913 914 /* Perform the write */ 915 file_offset_tmp = file_offset; 916 nwritten = vfs_write(curlun->filp, 917 (char __user *)bh->buf, 918 amount, &file_offset_tmp); 919 VLDBG(curlun, "file write %u @ %llu -> %d\n", amount, 920 (unsigned long long)file_offset, (int)nwritten); 921 if (signal_pending(current)) 922 return -EINTR; /* Interrupted! */ 923 924 if (nwritten < 0) { 925 LDBG(curlun, "error in file write: %d\n", 926 (int)nwritten); 927 nwritten = 0; 928 } else if (nwritten < amount) { 929 LDBG(curlun, "partial file write: %d/%u\n", 930 (int)nwritten, amount); 931 nwritten = round_down(nwritten, curlun->blksize); 932 } 933 file_offset += nwritten; 934 amount_left_to_write -= nwritten; 935 common->residue -= nwritten; 936 937 /* If an error occurred, report it and its position */ 938 if (nwritten < amount) { 939 curlun->sense_data = SS_WRITE_ERROR; 940 curlun->sense_data_info = 941 file_offset >> curlun->blkbits; 942 curlun->info_valid = 1; 943 break; 944 } 945 946 empty_write: 947 /* Did the host decide to stop early? */ 948 if (bh->outreq->actual < bh->bulk_out_intended_length) { 949 common->short_packet_received = 1; 950 break; 951 } 952 continue; 953 } 954 955 /* Wait for something to happen */ 956 rc = sleep_thread(common, false); 957 if (rc) 958 return rc; 959 } 960 961 return -EIO; /* No default reply */ 962 } 963 964 965 /*-------------------------------------------------------------------------*/ 966 967 static int do_synchronize_cache(struct fsg_common *common) 968 { 969 struct fsg_lun *curlun = common->curlun; 970 int rc; 971 972 /* We ignore the requested LBA and write out all file's 973 * dirty data buffers. */ 974 rc = fsg_lun_fsync_sub(curlun); 975 if (rc) 976 curlun->sense_data = SS_WRITE_ERROR; 977 return 0; 978 } 979 980 981 /*-------------------------------------------------------------------------*/ 982 983 static void invalidate_sub(struct fsg_lun *curlun) 984 { 985 struct file *filp = curlun->filp; 986 struct inode *inode = file_inode(filp); 987 unsigned long rc; 988 989 rc = invalidate_mapping_pages(inode->i_mapping, 0, -1); 990 VLDBG(curlun, "invalidate_mapping_pages -> %ld\n", rc); 991 } 992 993 static int do_verify(struct fsg_common *common) 994 { 995 struct fsg_lun *curlun = common->curlun; 996 u32 lba; 997 u32 verification_length; 998 struct fsg_buffhd *bh = common->next_buffhd_to_fill; 999 loff_t file_offset, file_offset_tmp; 1000 u32 amount_left; 1001 unsigned int amount; 1002 ssize_t nread; 1003 1004 /* 1005 * Get the starting Logical Block Address and check that it's 1006 * not too big. 1007 */ 1008 lba = get_unaligned_be32(&common->cmnd[2]); 1009 if (lba >= curlun->num_sectors) { 1010 curlun->sense_data = SS_LOGICAL_BLOCK_ADDRESS_OUT_OF_RANGE; 1011 return -EINVAL; 1012 } 1013 1014 /* 1015 * We allow DPO (Disable Page Out = don't save data in the 1016 * cache) but we don't implement it. 1017 */ 1018 if (common->cmnd[1] & ~0x10) { 1019 curlun->sense_data = SS_INVALID_FIELD_IN_CDB; 1020 return -EINVAL; 1021 } 1022 1023 verification_length = get_unaligned_be16(&common->cmnd[7]); 1024 if (unlikely(verification_length == 0)) 1025 return -EIO; /* No default reply */ 1026 1027 /* Prepare to carry out the file verify */ 1028 amount_left = verification_length << curlun->blkbits; 1029 file_offset = ((loff_t) lba) << curlun->blkbits; 1030 1031 /* Write out all the dirty buffers before invalidating them */ 1032 fsg_lun_fsync_sub(curlun); 1033 if (signal_pending(current)) 1034 return -EINTR; 1035 1036 invalidate_sub(curlun); 1037 if (signal_pending(current)) 1038 return -EINTR; 1039 1040 /* Just try to read the requested blocks */ 1041 while (amount_left > 0) { 1042 /* 1043 * Figure out how much we need to read: 1044 * Try to read the remaining amount, but not more than 1045 * the buffer size. 1046 * And don't try to read past the end of the file. 1047 */ 1048 amount = min(amount_left, FSG_BUFLEN); 1049 amount = min((loff_t)amount, 1050 curlun->file_length - file_offset); 1051 if (amount == 0) { 1052 curlun->sense_data = 1053 SS_LOGICAL_BLOCK_ADDRESS_OUT_OF_RANGE; 1054 curlun->sense_data_info = 1055 file_offset >> curlun->blkbits; 1056 curlun->info_valid = 1; 1057 break; 1058 } 1059 1060 /* Perform the read */ 1061 file_offset_tmp = file_offset; 1062 nread = vfs_read(curlun->filp, 1063 (char __user *) bh->buf, 1064 amount, &file_offset_tmp); 1065 VLDBG(curlun, "file read %u @ %llu -> %d\n", amount, 1066 (unsigned long long) file_offset, 1067 (int) nread); 1068 if (signal_pending(current)) 1069 return -EINTR; 1070 1071 if (nread < 0) { 1072 LDBG(curlun, "error in file verify: %d\n", (int)nread); 1073 nread = 0; 1074 } else if (nread < amount) { 1075 LDBG(curlun, "partial file verify: %d/%u\n", 1076 (int)nread, amount); 1077 nread = round_down(nread, curlun->blksize); 1078 } 1079 if (nread == 0) { 1080 curlun->sense_data = SS_UNRECOVERED_READ_ERROR; 1081 curlun->sense_data_info = 1082 file_offset >> curlun->blkbits; 1083 curlun->info_valid = 1; 1084 break; 1085 } 1086 file_offset += nread; 1087 amount_left -= nread; 1088 } 1089 return 0; 1090 } 1091 1092 1093 /*-------------------------------------------------------------------------*/ 1094 1095 static int do_inquiry(struct fsg_common *common, struct fsg_buffhd *bh) 1096 { 1097 struct fsg_lun *curlun = common->curlun; 1098 u8 *buf = (u8 *) bh->buf; 1099 1100 if (!curlun) { /* Unsupported LUNs are okay */ 1101 common->bad_lun_okay = 1; 1102 memset(buf, 0, 36); 1103 buf[0] = TYPE_NO_LUN; /* Unsupported, no device-type */ 1104 buf[4] = 31; /* Additional length */ 1105 return 36; 1106 } 1107 1108 buf[0] = curlun->cdrom ? TYPE_ROM : TYPE_DISK; 1109 buf[1] = curlun->removable ? 0x80 : 0; 1110 buf[2] = 2; /* ANSI SCSI level 2 */ 1111 buf[3] = 2; /* SCSI-2 INQUIRY data format */ 1112 buf[4] = 31; /* Additional length */ 1113 buf[5] = 0; /* No special options */ 1114 buf[6] = 0; 1115 buf[7] = 0; 1116 if (curlun->inquiry_string[0]) 1117 memcpy(buf + 8, curlun->inquiry_string, 1118 sizeof(curlun->inquiry_string)); 1119 else 1120 memcpy(buf + 8, common->inquiry_string, 1121 sizeof(common->inquiry_string)); 1122 return 36; 1123 } 1124 1125 static int do_request_sense(struct fsg_common *common, struct fsg_buffhd *bh) 1126 { 1127 struct fsg_lun *curlun = common->curlun; 1128 u8 *buf = (u8 *) bh->buf; 1129 u32 sd, sdinfo; 1130 int valid; 1131 1132 /* 1133 * From the SCSI-2 spec., section 7.9 (Unit attention condition): 1134 * 1135 * If a REQUEST SENSE command is received from an initiator 1136 * with a pending unit attention condition (before the target 1137 * generates the contingent allegiance condition), then the 1138 * target shall either: 1139 * a) report any pending sense data and preserve the unit 1140 * attention condition on the logical unit, or, 1141 * b) report the unit attention condition, may discard any 1142 * pending sense data, and clear the unit attention 1143 * condition on the logical unit for that initiator. 1144 * 1145 * FSG normally uses option a); enable this code to use option b). 1146 */ 1147 #if 0 1148 if (curlun && curlun->unit_attention_data != SS_NO_SENSE) { 1149 curlun->sense_data = curlun->unit_attention_data; 1150 curlun->unit_attention_data = SS_NO_SENSE; 1151 } 1152 #endif 1153 1154 if (!curlun) { /* Unsupported LUNs are okay */ 1155 common->bad_lun_okay = 1; 1156 sd = SS_LOGICAL_UNIT_NOT_SUPPORTED; 1157 sdinfo = 0; 1158 valid = 0; 1159 } else { 1160 sd = curlun->sense_data; 1161 sdinfo = curlun->sense_data_info; 1162 valid = curlun->info_valid << 7; 1163 curlun->sense_data = SS_NO_SENSE; 1164 curlun->sense_data_info = 0; 1165 curlun->info_valid = 0; 1166 } 1167 1168 memset(buf, 0, 18); 1169 buf[0] = valid | 0x70; /* Valid, current error */ 1170 buf[2] = SK(sd); 1171 put_unaligned_be32(sdinfo, &buf[3]); /* Sense information */ 1172 buf[7] = 18 - 8; /* Additional sense length */ 1173 buf[12] = ASC(sd); 1174 buf[13] = ASCQ(sd); 1175 return 18; 1176 } 1177 1178 static int do_read_capacity(struct fsg_common *common, struct fsg_buffhd *bh) 1179 { 1180 struct fsg_lun *curlun = common->curlun; 1181 u32 lba = get_unaligned_be32(&common->cmnd[2]); 1182 int pmi = common->cmnd[8]; 1183 u8 *buf = (u8 *)bh->buf; 1184 1185 /* Check the PMI and LBA fields */ 1186 if (pmi > 1 || (pmi == 0 && lba != 0)) { 1187 curlun->sense_data = SS_INVALID_FIELD_IN_CDB; 1188 return -EINVAL; 1189 } 1190 1191 put_unaligned_be32(curlun->num_sectors - 1, &buf[0]); 1192 /* Max logical block */ 1193 put_unaligned_be32(curlun->blksize, &buf[4]);/* Block length */ 1194 return 8; 1195 } 1196 1197 static int do_read_header(struct fsg_common *common, struct fsg_buffhd *bh) 1198 { 1199 struct fsg_lun *curlun = common->curlun; 1200 int msf = common->cmnd[1] & 0x02; 1201 u32 lba = get_unaligned_be32(&common->cmnd[2]); 1202 u8 *buf = (u8 *)bh->buf; 1203 1204 if (common->cmnd[1] & ~0x02) { /* Mask away MSF */ 1205 curlun->sense_data = SS_INVALID_FIELD_IN_CDB; 1206 return -EINVAL; 1207 } 1208 if (lba >= curlun->num_sectors) { 1209 curlun->sense_data = SS_LOGICAL_BLOCK_ADDRESS_OUT_OF_RANGE; 1210 return -EINVAL; 1211 } 1212 1213 memset(buf, 0, 8); 1214 buf[0] = 0x01; /* 2048 bytes of user data, rest is EC */ 1215 store_cdrom_address(&buf[4], msf, lba); 1216 return 8; 1217 } 1218 1219 static int do_read_toc(struct fsg_common *common, struct fsg_buffhd *bh) 1220 { 1221 struct fsg_lun *curlun = common->curlun; 1222 int msf = common->cmnd[1] & 0x02; 1223 int start_track = common->cmnd[6]; 1224 u8 *buf = (u8 *)bh->buf; 1225 1226 if ((common->cmnd[1] & ~0x02) != 0 || /* Mask away MSF */ 1227 start_track > 1) { 1228 curlun->sense_data = SS_INVALID_FIELD_IN_CDB; 1229 return -EINVAL; 1230 } 1231 1232 memset(buf, 0, 20); 1233 buf[1] = (20-2); /* TOC data length */ 1234 buf[2] = 1; /* First track number */ 1235 buf[3] = 1; /* Last track number */ 1236 buf[5] = 0x16; /* Data track, copying allowed */ 1237 buf[6] = 0x01; /* Only track is number 1 */ 1238 store_cdrom_address(&buf[8], msf, 0); 1239 1240 buf[13] = 0x16; /* Lead-out track is data */ 1241 buf[14] = 0xAA; /* Lead-out track number */ 1242 store_cdrom_address(&buf[16], msf, curlun->num_sectors); 1243 return 20; 1244 } 1245 1246 static int do_mode_sense(struct fsg_common *common, struct fsg_buffhd *bh) 1247 { 1248 struct fsg_lun *curlun = common->curlun; 1249 int mscmnd = common->cmnd[0]; 1250 u8 *buf = (u8 *) bh->buf; 1251 u8 *buf0 = buf; 1252 int pc, page_code; 1253 int changeable_values, all_pages; 1254 int valid_page = 0; 1255 int len, limit; 1256 1257 if ((common->cmnd[1] & ~0x08) != 0) { /* Mask away DBD */ 1258 curlun->sense_data = SS_INVALID_FIELD_IN_CDB; 1259 return -EINVAL; 1260 } 1261 pc = common->cmnd[2] >> 6; 1262 page_code = common->cmnd[2] & 0x3f; 1263 if (pc == 3) { 1264 curlun->sense_data = SS_SAVING_PARAMETERS_NOT_SUPPORTED; 1265 return -EINVAL; 1266 } 1267 changeable_values = (pc == 1); 1268 all_pages = (page_code == 0x3f); 1269 1270 /* 1271 * Write the mode parameter header. Fixed values are: default 1272 * medium type, no cache control (DPOFUA), and no block descriptors. 1273 * The only variable value is the WriteProtect bit. We will fill in 1274 * the mode data length later. 1275 */ 1276 memset(buf, 0, 8); 1277 if (mscmnd == MODE_SENSE) { 1278 buf[2] = (curlun->ro ? 0x80 : 0x00); /* WP, DPOFUA */ 1279 buf += 4; 1280 limit = 255; 1281 } else { /* MODE_SENSE_10 */ 1282 buf[3] = (curlun->ro ? 0x80 : 0x00); /* WP, DPOFUA */ 1283 buf += 8; 1284 limit = 65535; /* Should really be FSG_BUFLEN */ 1285 } 1286 1287 /* No block descriptors */ 1288 1289 /* 1290 * The mode pages, in numerical order. The only page we support 1291 * is the Caching page. 1292 */ 1293 if (page_code == 0x08 || all_pages) { 1294 valid_page = 1; 1295 buf[0] = 0x08; /* Page code */ 1296 buf[1] = 10; /* Page length */ 1297 memset(buf+2, 0, 10); /* None of the fields are changeable */ 1298 1299 if (!changeable_values) { 1300 buf[2] = 0x04; /* Write cache enable, */ 1301 /* Read cache not disabled */ 1302 /* No cache retention priorities */ 1303 put_unaligned_be16(0xffff, &buf[4]); 1304 /* Don't disable prefetch */ 1305 /* Minimum prefetch = 0 */ 1306 put_unaligned_be16(0xffff, &buf[8]); 1307 /* Maximum prefetch */ 1308 put_unaligned_be16(0xffff, &buf[10]); 1309 /* Maximum prefetch ceiling */ 1310 } 1311 buf += 12; 1312 } 1313 1314 /* 1315 * Check that a valid page was requested and the mode data length 1316 * isn't too long. 1317 */ 1318 len = buf - buf0; 1319 if (!valid_page || len > limit) { 1320 curlun->sense_data = SS_INVALID_FIELD_IN_CDB; 1321 return -EINVAL; 1322 } 1323 1324 /* Store the mode data length */ 1325 if (mscmnd == MODE_SENSE) 1326 buf0[0] = len - 1; 1327 else 1328 put_unaligned_be16(len - 2, buf0); 1329 return len; 1330 } 1331 1332 static int do_start_stop(struct fsg_common *common) 1333 { 1334 struct fsg_lun *curlun = common->curlun; 1335 int loej, start; 1336 1337 if (!curlun) { 1338 return -EINVAL; 1339 } else if (!curlun->removable) { 1340 curlun->sense_data = SS_INVALID_COMMAND; 1341 return -EINVAL; 1342 } else if ((common->cmnd[1] & ~0x01) != 0 || /* Mask away Immed */ 1343 (common->cmnd[4] & ~0x03) != 0) { /* Mask LoEj, Start */ 1344 curlun->sense_data = SS_INVALID_FIELD_IN_CDB; 1345 return -EINVAL; 1346 } 1347 1348 loej = common->cmnd[4] & 0x02; 1349 start = common->cmnd[4] & 0x01; 1350 1351 /* 1352 * Our emulation doesn't support mounting; the medium is 1353 * available for use as soon as it is loaded. 1354 */ 1355 if (start) { 1356 if (!fsg_lun_is_open(curlun)) { 1357 curlun->sense_data = SS_MEDIUM_NOT_PRESENT; 1358 return -EINVAL; 1359 } 1360 return 0; 1361 } 1362 1363 /* Are we allowed to unload the media? */ 1364 if (curlun->prevent_medium_removal) { 1365 LDBG(curlun, "unload attempt prevented\n"); 1366 curlun->sense_data = SS_MEDIUM_REMOVAL_PREVENTED; 1367 return -EINVAL; 1368 } 1369 1370 if (!loej) 1371 return 0; 1372 1373 up_read(&common->filesem); 1374 down_write(&common->filesem); 1375 fsg_lun_close(curlun); 1376 up_write(&common->filesem); 1377 down_read(&common->filesem); 1378 1379 return 0; 1380 } 1381 1382 static int do_prevent_allow(struct fsg_common *common) 1383 { 1384 struct fsg_lun *curlun = common->curlun; 1385 int prevent; 1386 1387 if (!common->curlun) { 1388 return -EINVAL; 1389 } else if (!common->curlun->removable) { 1390 common->curlun->sense_data = SS_INVALID_COMMAND; 1391 return -EINVAL; 1392 } 1393 1394 prevent = common->cmnd[4] & 0x01; 1395 if ((common->cmnd[4] & ~0x01) != 0) { /* Mask away Prevent */ 1396 curlun->sense_data = SS_INVALID_FIELD_IN_CDB; 1397 return -EINVAL; 1398 } 1399 1400 if (curlun->prevent_medium_removal && !prevent) 1401 fsg_lun_fsync_sub(curlun); 1402 curlun->prevent_medium_removal = prevent; 1403 return 0; 1404 } 1405 1406 static int do_read_format_capacities(struct fsg_common *common, 1407 struct fsg_buffhd *bh) 1408 { 1409 struct fsg_lun *curlun = common->curlun; 1410 u8 *buf = (u8 *) bh->buf; 1411 1412 buf[0] = buf[1] = buf[2] = 0; 1413 buf[3] = 8; /* Only the Current/Maximum Capacity Descriptor */ 1414 buf += 4; 1415 1416 put_unaligned_be32(curlun->num_sectors, &buf[0]); 1417 /* Number of blocks */ 1418 put_unaligned_be32(curlun->blksize, &buf[4]);/* Block length */ 1419 buf[4] = 0x02; /* Current capacity */ 1420 return 12; 1421 } 1422 1423 static int do_mode_select(struct fsg_common *common, struct fsg_buffhd *bh) 1424 { 1425 struct fsg_lun *curlun = common->curlun; 1426 1427 /* We don't support MODE SELECT */ 1428 if (curlun) 1429 curlun->sense_data = SS_INVALID_COMMAND; 1430 return -EINVAL; 1431 } 1432 1433 1434 /*-------------------------------------------------------------------------*/ 1435 1436 static int halt_bulk_in_endpoint(struct fsg_dev *fsg) 1437 { 1438 int rc; 1439 1440 rc = fsg_set_halt(fsg, fsg->bulk_in); 1441 if (rc == -EAGAIN) 1442 VDBG(fsg, "delayed bulk-in endpoint halt\n"); 1443 while (rc != 0) { 1444 if (rc != -EAGAIN) { 1445 WARNING(fsg, "usb_ep_set_halt -> %d\n", rc); 1446 rc = 0; 1447 break; 1448 } 1449 1450 /* Wait for a short time and then try again */ 1451 if (msleep_interruptible(100) != 0) 1452 return -EINTR; 1453 rc = usb_ep_set_halt(fsg->bulk_in); 1454 } 1455 return rc; 1456 } 1457 1458 static int wedge_bulk_in_endpoint(struct fsg_dev *fsg) 1459 { 1460 int rc; 1461 1462 DBG(fsg, "bulk-in set wedge\n"); 1463 rc = usb_ep_set_wedge(fsg->bulk_in); 1464 if (rc == -EAGAIN) 1465 VDBG(fsg, "delayed bulk-in endpoint wedge\n"); 1466 while (rc != 0) { 1467 if (rc != -EAGAIN) { 1468 WARNING(fsg, "usb_ep_set_wedge -> %d\n", rc); 1469 rc = 0; 1470 break; 1471 } 1472 1473 /* Wait for a short time and then try again */ 1474 if (msleep_interruptible(100) != 0) 1475 return -EINTR; 1476 rc = usb_ep_set_wedge(fsg->bulk_in); 1477 } 1478 return rc; 1479 } 1480 1481 static int throw_away_data(struct fsg_common *common) 1482 { 1483 struct fsg_buffhd *bh; 1484 u32 amount; 1485 int rc; 1486 1487 for (bh = common->next_buffhd_to_drain; 1488 bh->state != BUF_STATE_EMPTY || common->usb_amount_left > 0; 1489 bh = common->next_buffhd_to_drain) { 1490 1491 /* Throw away the data in a filled buffer */ 1492 if (bh->state == BUF_STATE_FULL) { 1493 smp_rmb(); 1494 bh->state = BUF_STATE_EMPTY; 1495 common->next_buffhd_to_drain = bh->next; 1496 1497 /* A short packet or an error ends everything */ 1498 if (bh->outreq->actual < bh->bulk_out_intended_length || 1499 bh->outreq->status != 0) { 1500 raise_exception(common, 1501 FSG_STATE_ABORT_BULK_OUT); 1502 return -EINTR; 1503 } 1504 continue; 1505 } 1506 1507 /* Try to submit another request if we need one */ 1508 bh = common->next_buffhd_to_fill; 1509 if (bh->state == BUF_STATE_EMPTY 1510 && common->usb_amount_left > 0) { 1511 amount = min(common->usb_amount_left, FSG_BUFLEN); 1512 1513 /* 1514 * Except at the end of the transfer, amount will be 1515 * equal to the buffer size, which is divisible by 1516 * the bulk-out maxpacket size. 1517 */ 1518 set_bulk_out_req_length(common, bh, amount); 1519 if (!start_out_transfer(common, bh)) 1520 /* Dunno what to do if common->fsg is NULL */ 1521 return -EIO; 1522 common->next_buffhd_to_fill = bh->next; 1523 common->usb_amount_left -= amount; 1524 continue; 1525 } 1526 1527 /* Otherwise wait for something to happen */ 1528 rc = sleep_thread(common, true); 1529 if (rc) 1530 return rc; 1531 } 1532 return 0; 1533 } 1534 1535 static int finish_reply(struct fsg_common *common) 1536 { 1537 struct fsg_buffhd *bh = common->next_buffhd_to_fill; 1538 int rc = 0; 1539 1540 switch (common->data_dir) { 1541 case DATA_DIR_NONE: 1542 break; /* Nothing to send */ 1543 1544 /* 1545 * If we don't know whether the host wants to read or write, 1546 * this must be CB or CBI with an unknown command. We mustn't 1547 * try to send or receive any data. So stall both bulk pipes 1548 * if we can and wait for a reset. 1549 */ 1550 case DATA_DIR_UNKNOWN: 1551 if (!common->can_stall) { 1552 /* Nothing */ 1553 } else if (fsg_is_set(common)) { 1554 fsg_set_halt(common->fsg, common->fsg->bulk_out); 1555 rc = halt_bulk_in_endpoint(common->fsg); 1556 } else { 1557 /* Don't know what to do if common->fsg is NULL */ 1558 rc = -EIO; 1559 } 1560 break; 1561 1562 /* All but the last buffer of data must have already been sent */ 1563 case DATA_DIR_TO_HOST: 1564 if (common->data_size == 0) { 1565 /* Nothing to send */ 1566 1567 /* Don't know what to do if common->fsg is NULL */ 1568 } else if (!fsg_is_set(common)) { 1569 rc = -EIO; 1570 1571 /* If there's no residue, simply send the last buffer */ 1572 } else if (common->residue == 0) { 1573 bh->inreq->zero = 0; 1574 if (!start_in_transfer(common, bh)) 1575 return -EIO; 1576 common->next_buffhd_to_fill = bh->next; 1577 1578 /* 1579 * For Bulk-only, mark the end of the data with a short 1580 * packet. If we are allowed to stall, halt the bulk-in 1581 * endpoint. (Note: This violates the Bulk-Only Transport 1582 * specification, which requires us to pad the data if we 1583 * don't halt the endpoint. Presumably nobody will mind.) 1584 */ 1585 } else { 1586 bh->inreq->zero = 1; 1587 if (!start_in_transfer(common, bh)) 1588 rc = -EIO; 1589 common->next_buffhd_to_fill = bh->next; 1590 if (common->can_stall) 1591 rc = halt_bulk_in_endpoint(common->fsg); 1592 } 1593 break; 1594 1595 /* 1596 * We have processed all we want from the data the host has sent. 1597 * There may still be outstanding bulk-out requests. 1598 */ 1599 case DATA_DIR_FROM_HOST: 1600 if (common->residue == 0) { 1601 /* Nothing to receive */ 1602 1603 /* Did the host stop sending unexpectedly early? */ 1604 } else if (common->short_packet_received) { 1605 raise_exception(common, FSG_STATE_ABORT_BULK_OUT); 1606 rc = -EINTR; 1607 1608 /* 1609 * We haven't processed all the incoming data. Even though 1610 * we may be allowed to stall, doing so would cause a race. 1611 * The controller may already have ACK'ed all the remaining 1612 * bulk-out packets, in which case the host wouldn't see a 1613 * STALL. Not realizing the endpoint was halted, it wouldn't 1614 * clear the halt -- leading to problems later on. 1615 */ 1616 #if 0 1617 } else if (common->can_stall) { 1618 if (fsg_is_set(common)) 1619 fsg_set_halt(common->fsg, 1620 common->fsg->bulk_out); 1621 raise_exception(common, FSG_STATE_ABORT_BULK_OUT); 1622 rc = -EINTR; 1623 #endif 1624 1625 /* 1626 * We can't stall. Read in the excess data and throw it 1627 * all away. 1628 */ 1629 } else { 1630 rc = throw_away_data(common); 1631 } 1632 break; 1633 } 1634 return rc; 1635 } 1636 1637 static int send_status(struct fsg_common *common) 1638 { 1639 struct fsg_lun *curlun = common->curlun; 1640 struct fsg_buffhd *bh; 1641 struct bulk_cs_wrap *csw; 1642 int rc; 1643 u8 status = US_BULK_STAT_OK; 1644 u32 sd, sdinfo = 0; 1645 1646 /* Wait for the next buffer to become available */ 1647 bh = common->next_buffhd_to_fill; 1648 while (bh->state != BUF_STATE_EMPTY) { 1649 rc = sleep_thread(common, true); 1650 if (rc) 1651 return rc; 1652 } 1653 1654 if (curlun) { 1655 sd = curlun->sense_data; 1656 sdinfo = curlun->sense_data_info; 1657 } else if (common->bad_lun_okay) 1658 sd = SS_NO_SENSE; 1659 else 1660 sd = SS_LOGICAL_UNIT_NOT_SUPPORTED; 1661 1662 if (common->phase_error) { 1663 DBG(common, "sending phase-error status\n"); 1664 status = US_BULK_STAT_PHASE; 1665 sd = SS_INVALID_COMMAND; 1666 } else if (sd != SS_NO_SENSE) { 1667 DBG(common, "sending command-failure status\n"); 1668 status = US_BULK_STAT_FAIL; 1669 VDBG(common, " sense data: SK x%02x, ASC x%02x, ASCQ x%02x;" 1670 " info x%x\n", 1671 SK(sd), ASC(sd), ASCQ(sd), sdinfo); 1672 } 1673 1674 /* Store and send the Bulk-only CSW */ 1675 csw = (void *)bh->buf; 1676 1677 csw->Signature = cpu_to_le32(US_BULK_CS_SIGN); 1678 csw->Tag = common->tag; 1679 csw->Residue = cpu_to_le32(common->residue); 1680 csw->Status = status; 1681 1682 bh->inreq->length = US_BULK_CS_WRAP_LEN; 1683 bh->inreq->zero = 0; 1684 if (!start_in_transfer(common, bh)) 1685 /* Don't know what to do if common->fsg is NULL */ 1686 return -EIO; 1687 1688 common->next_buffhd_to_fill = bh->next; 1689 return 0; 1690 } 1691 1692 1693 /*-------------------------------------------------------------------------*/ 1694 1695 /* 1696 * Check whether the command is properly formed and whether its data size 1697 * and direction agree with the values we already have. 1698 */ 1699 static int check_command(struct fsg_common *common, int cmnd_size, 1700 enum data_direction data_dir, unsigned int mask, 1701 int needs_medium, const char *name) 1702 { 1703 int i; 1704 unsigned int lun = common->cmnd[1] >> 5; 1705 static const char dirletter[4] = {'u', 'o', 'i', 'n'}; 1706 char hdlen[20]; 1707 struct fsg_lun *curlun; 1708 1709 hdlen[0] = 0; 1710 if (common->data_dir != DATA_DIR_UNKNOWN) 1711 sprintf(hdlen, ", H%c=%u", dirletter[(int) common->data_dir], 1712 common->data_size); 1713 VDBG(common, "SCSI command: %s; Dc=%d, D%c=%u; Hc=%d%s\n", 1714 name, cmnd_size, dirletter[(int) data_dir], 1715 common->data_size_from_cmnd, common->cmnd_size, hdlen); 1716 1717 /* 1718 * We can't reply at all until we know the correct data direction 1719 * and size. 1720 */ 1721 if (common->data_size_from_cmnd == 0) 1722 data_dir = DATA_DIR_NONE; 1723 if (common->data_size < common->data_size_from_cmnd) { 1724 /* 1725 * Host data size < Device data size is a phase error. 1726 * Carry out the command, but only transfer as much as 1727 * we are allowed. 1728 */ 1729 common->data_size_from_cmnd = common->data_size; 1730 common->phase_error = 1; 1731 } 1732 common->residue = common->data_size; 1733 common->usb_amount_left = common->data_size; 1734 1735 /* Conflicting data directions is a phase error */ 1736 if (common->data_dir != data_dir && common->data_size_from_cmnd > 0) { 1737 common->phase_error = 1; 1738 return -EINVAL; 1739 } 1740 1741 /* Verify the length of the command itself */ 1742 if (cmnd_size != common->cmnd_size) { 1743 1744 /* 1745 * Special case workaround: There are plenty of buggy SCSI 1746 * implementations. Many have issues with cbw->Length 1747 * field passing a wrong command size. For those cases we 1748 * always try to work around the problem by using the length 1749 * sent by the host side provided it is at least as large 1750 * as the correct command length. 1751 * Examples of such cases would be MS-Windows, which issues 1752 * REQUEST SENSE with cbw->Length == 12 where it should 1753 * be 6, and xbox360 issuing INQUIRY, TEST UNIT READY and 1754 * REQUEST SENSE with cbw->Length == 10 where it should 1755 * be 6 as well. 1756 */ 1757 if (cmnd_size <= common->cmnd_size) { 1758 DBG(common, "%s is buggy! Expected length %d " 1759 "but we got %d\n", name, 1760 cmnd_size, common->cmnd_size); 1761 cmnd_size = common->cmnd_size; 1762 } else { 1763 common->phase_error = 1; 1764 return -EINVAL; 1765 } 1766 } 1767 1768 /* Check that the LUN values are consistent */ 1769 if (common->lun != lun) 1770 DBG(common, "using LUN %u from CBW, not LUN %u from CDB\n", 1771 common->lun, lun); 1772 1773 /* Check the LUN */ 1774 curlun = common->curlun; 1775 if (curlun) { 1776 if (common->cmnd[0] != REQUEST_SENSE) { 1777 curlun->sense_data = SS_NO_SENSE; 1778 curlun->sense_data_info = 0; 1779 curlun->info_valid = 0; 1780 } 1781 } else { 1782 common->bad_lun_okay = 0; 1783 1784 /* 1785 * INQUIRY and REQUEST SENSE commands are explicitly allowed 1786 * to use unsupported LUNs; all others may not. 1787 */ 1788 if (common->cmnd[0] != INQUIRY && 1789 common->cmnd[0] != REQUEST_SENSE) { 1790 DBG(common, "unsupported LUN %u\n", common->lun); 1791 return -EINVAL; 1792 } 1793 } 1794 1795 /* 1796 * If a unit attention condition exists, only INQUIRY and 1797 * REQUEST SENSE commands are allowed; anything else must fail. 1798 */ 1799 if (curlun && curlun->unit_attention_data != SS_NO_SENSE && 1800 common->cmnd[0] != INQUIRY && 1801 common->cmnd[0] != REQUEST_SENSE) { 1802 curlun->sense_data = curlun->unit_attention_data; 1803 curlun->unit_attention_data = SS_NO_SENSE; 1804 return -EINVAL; 1805 } 1806 1807 /* Check that only command bytes listed in the mask are non-zero */ 1808 common->cmnd[1] &= 0x1f; /* Mask away the LUN */ 1809 for (i = 1; i < cmnd_size; ++i) { 1810 if (common->cmnd[i] && !(mask & (1 << i))) { 1811 if (curlun) 1812 curlun->sense_data = SS_INVALID_FIELD_IN_CDB; 1813 return -EINVAL; 1814 } 1815 } 1816 1817 /* If the medium isn't mounted and the command needs to access 1818 * it, return an error. */ 1819 if (curlun && !fsg_lun_is_open(curlun) && needs_medium) { 1820 curlun->sense_data = SS_MEDIUM_NOT_PRESENT; 1821 return -EINVAL; 1822 } 1823 1824 return 0; 1825 } 1826 1827 /* wrapper of check_command for data size in blocks handling */ 1828 static int check_command_size_in_blocks(struct fsg_common *common, 1829 int cmnd_size, enum data_direction data_dir, 1830 unsigned int mask, int needs_medium, const char *name) 1831 { 1832 if (common->curlun) 1833 common->data_size_from_cmnd <<= common->curlun->blkbits; 1834 return check_command(common, cmnd_size, data_dir, 1835 mask, needs_medium, name); 1836 } 1837 1838 static int do_scsi_command(struct fsg_common *common) 1839 { 1840 struct fsg_buffhd *bh; 1841 int rc; 1842 int reply = -EINVAL; 1843 int i; 1844 static char unknown[16]; 1845 1846 dump_cdb(common); 1847 1848 /* Wait for the next buffer to become available for data or status */ 1849 bh = common->next_buffhd_to_fill; 1850 common->next_buffhd_to_drain = bh; 1851 while (bh->state != BUF_STATE_EMPTY) { 1852 rc = sleep_thread(common, true); 1853 if (rc) 1854 return rc; 1855 } 1856 common->phase_error = 0; 1857 common->short_packet_received = 0; 1858 1859 down_read(&common->filesem); /* We're using the backing file */ 1860 switch (common->cmnd[0]) { 1861 1862 case INQUIRY: 1863 common->data_size_from_cmnd = common->cmnd[4]; 1864 reply = check_command(common, 6, DATA_DIR_TO_HOST, 1865 (1<<4), 0, 1866 "INQUIRY"); 1867 if (reply == 0) 1868 reply = do_inquiry(common, bh); 1869 break; 1870 1871 case MODE_SELECT: 1872 common->data_size_from_cmnd = common->cmnd[4]; 1873 reply = check_command(common, 6, DATA_DIR_FROM_HOST, 1874 (1<<1) | (1<<4), 0, 1875 "MODE SELECT(6)"); 1876 if (reply == 0) 1877 reply = do_mode_select(common, bh); 1878 break; 1879 1880 case MODE_SELECT_10: 1881 common->data_size_from_cmnd = 1882 get_unaligned_be16(&common->cmnd[7]); 1883 reply = check_command(common, 10, DATA_DIR_FROM_HOST, 1884 (1<<1) | (3<<7), 0, 1885 "MODE SELECT(10)"); 1886 if (reply == 0) 1887 reply = do_mode_select(common, bh); 1888 break; 1889 1890 case MODE_SENSE: 1891 common->data_size_from_cmnd = common->cmnd[4]; 1892 reply = check_command(common, 6, DATA_DIR_TO_HOST, 1893 (1<<1) | (1<<2) | (1<<4), 0, 1894 "MODE SENSE(6)"); 1895 if (reply == 0) 1896 reply = do_mode_sense(common, bh); 1897 break; 1898 1899 case MODE_SENSE_10: 1900 common->data_size_from_cmnd = 1901 get_unaligned_be16(&common->cmnd[7]); 1902 reply = check_command(common, 10, DATA_DIR_TO_HOST, 1903 (1<<1) | (1<<2) | (3<<7), 0, 1904 "MODE SENSE(10)"); 1905 if (reply == 0) 1906 reply = do_mode_sense(common, bh); 1907 break; 1908 1909 case ALLOW_MEDIUM_REMOVAL: 1910 common->data_size_from_cmnd = 0; 1911 reply = check_command(common, 6, DATA_DIR_NONE, 1912 (1<<4), 0, 1913 "PREVENT-ALLOW MEDIUM REMOVAL"); 1914 if (reply == 0) 1915 reply = do_prevent_allow(common); 1916 break; 1917 1918 case READ_6: 1919 i = common->cmnd[4]; 1920 common->data_size_from_cmnd = (i == 0) ? 256 : i; 1921 reply = check_command_size_in_blocks(common, 6, 1922 DATA_DIR_TO_HOST, 1923 (7<<1) | (1<<4), 1, 1924 "READ(6)"); 1925 if (reply == 0) 1926 reply = do_read(common); 1927 break; 1928 1929 case READ_10: 1930 common->data_size_from_cmnd = 1931 get_unaligned_be16(&common->cmnd[7]); 1932 reply = check_command_size_in_blocks(common, 10, 1933 DATA_DIR_TO_HOST, 1934 (1<<1) | (0xf<<2) | (3<<7), 1, 1935 "READ(10)"); 1936 if (reply == 0) 1937 reply = do_read(common); 1938 break; 1939 1940 case READ_12: 1941 common->data_size_from_cmnd = 1942 get_unaligned_be32(&common->cmnd[6]); 1943 reply = check_command_size_in_blocks(common, 12, 1944 DATA_DIR_TO_HOST, 1945 (1<<1) | (0xf<<2) | (0xf<<6), 1, 1946 "READ(12)"); 1947 if (reply == 0) 1948 reply = do_read(common); 1949 break; 1950 1951 case READ_CAPACITY: 1952 common->data_size_from_cmnd = 8; 1953 reply = check_command(common, 10, DATA_DIR_TO_HOST, 1954 (0xf<<2) | (1<<8), 1, 1955 "READ CAPACITY"); 1956 if (reply == 0) 1957 reply = do_read_capacity(common, bh); 1958 break; 1959 1960 case READ_HEADER: 1961 if (!common->curlun || !common->curlun->cdrom) 1962 goto unknown_cmnd; 1963 common->data_size_from_cmnd = 1964 get_unaligned_be16(&common->cmnd[7]); 1965 reply = check_command(common, 10, DATA_DIR_TO_HOST, 1966 (3<<7) | (0x1f<<1), 1, 1967 "READ HEADER"); 1968 if (reply == 0) 1969 reply = do_read_header(common, bh); 1970 break; 1971 1972 case READ_TOC: 1973 if (!common->curlun || !common->curlun->cdrom) 1974 goto unknown_cmnd; 1975 common->data_size_from_cmnd = 1976 get_unaligned_be16(&common->cmnd[7]); 1977 reply = check_command(common, 10, DATA_DIR_TO_HOST, 1978 (7<<6) | (1<<1), 1, 1979 "READ TOC"); 1980 if (reply == 0) 1981 reply = do_read_toc(common, bh); 1982 break; 1983 1984 case READ_FORMAT_CAPACITIES: 1985 common->data_size_from_cmnd = 1986 get_unaligned_be16(&common->cmnd[7]); 1987 reply = check_command(common, 10, DATA_DIR_TO_HOST, 1988 (3<<7), 1, 1989 "READ FORMAT CAPACITIES"); 1990 if (reply == 0) 1991 reply = do_read_format_capacities(common, bh); 1992 break; 1993 1994 case REQUEST_SENSE: 1995 common->data_size_from_cmnd = common->cmnd[4]; 1996 reply = check_command(common, 6, DATA_DIR_TO_HOST, 1997 (1<<4), 0, 1998 "REQUEST SENSE"); 1999 if (reply == 0) 2000 reply = do_request_sense(common, bh); 2001 break; 2002 2003 case START_STOP: 2004 common->data_size_from_cmnd = 0; 2005 reply = check_command(common, 6, DATA_DIR_NONE, 2006 (1<<1) | (1<<4), 0, 2007 "START-STOP UNIT"); 2008 if (reply == 0) 2009 reply = do_start_stop(common); 2010 break; 2011 2012 case SYNCHRONIZE_CACHE: 2013 common->data_size_from_cmnd = 0; 2014 reply = check_command(common, 10, DATA_DIR_NONE, 2015 (0xf<<2) | (3<<7), 1, 2016 "SYNCHRONIZE CACHE"); 2017 if (reply == 0) 2018 reply = do_synchronize_cache(common); 2019 break; 2020 2021 case TEST_UNIT_READY: 2022 common->data_size_from_cmnd = 0; 2023 reply = check_command(common, 6, DATA_DIR_NONE, 2024 0, 1, 2025 "TEST UNIT READY"); 2026 break; 2027 2028 /* 2029 * Although optional, this command is used by MS-Windows. We 2030 * support a minimal version: BytChk must be 0. 2031 */ 2032 case VERIFY: 2033 common->data_size_from_cmnd = 0; 2034 reply = check_command(common, 10, DATA_DIR_NONE, 2035 (1<<1) | (0xf<<2) | (3<<7), 1, 2036 "VERIFY"); 2037 if (reply == 0) 2038 reply = do_verify(common); 2039 break; 2040 2041 case WRITE_6: 2042 i = common->cmnd[4]; 2043 common->data_size_from_cmnd = (i == 0) ? 256 : i; 2044 reply = check_command_size_in_blocks(common, 6, 2045 DATA_DIR_FROM_HOST, 2046 (7<<1) | (1<<4), 1, 2047 "WRITE(6)"); 2048 if (reply == 0) 2049 reply = do_write(common); 2050 break; 2051 2052 case WRITE_10: 2053 common->data_size_from_cmnd = 2054 get_unaligned_be16(&common->cmnd[7]); 2055 reply = check_command_size_in_blocks(common, 10, 2056 DATA_DIR_FROM_HOST, 2057 (1<<1) | (0xf<<2) | (3<<7), 1, 2058 "WRITE(10)"); 2059 if (reply == 0) 2060 reply = do_write(common); 2061 break; 2062 2063 case WRITE_12: 2064 common->data_size_from_cmnd = 2065 get_unaligned_be32(&common->cmnd[6]); 2066 reply = check_command_size_in_blocks(common, 12, 2067 DATA_DIR_FROM_HOST, 2068 (1<<1) | (0xf<<2) | (0xf<<6), 1, 2069 "WRITE(12)"); 2070 if (reply == 0) 2071 reply = do_write(common); 2072 break; 2073 2074 /* 2075 * Some mandatory commands that we recognize but don't implement. 2076 * They don't mean much in this setting. It's left as an exercise 2077 * for anyone interested to implement RESERVE and RELEASE in terms 2078 * of Posix locks. 2079 */ 2080 case FORMAT_UNIT: 2081 case RELEASE: 2082 case RESERVE: 2083 case SEND_DIAGNOSTIC: 2084 /* Fall through */ 2085 2086 default: 2087 unknown_cmnd: 2088 common->data_size_from_cmnd = 0; 2089 sprintf(unknown, "Unknown x%02x", common->cmnd[0]); 2090 reply = check_command(common, common->cmnd_size, 2091 DATA_DIR_UNKNOWN, ~0, 0, unknown); 2092 if (reply == 0) { 2093 common->curlun->sense_data = SS_INVALID_COMMAND; 2094 reply = -EINVAL; 2095 } 2096 break; 2097 } 2098 up_read(&common->filesem); 2099 2100 if (reply == -EINTR || signal_pending(current)) 2101 return -EINTR; 2102 2103 /* Set up the single reply buffer for finish_reply() */ 2104 if (reply == -EINVAL) 2105 reply = 0; /* Error reply length */ 2106 if (reply >= 0 && common->data_dir == DATA_DIR_TO_HOST) { 2107 reply = min((u32)reply, common->data_size_from_cmnd); 2108 bh->inreq->length = reply; 2109 bh->state = BUF_STATE_FULL; 2110 common->residue -= reply; 2111 } /* Otherwise it's already set */ 2112 2113 return 0; 2114 } 2115 2116 2117 /*-------------------------------------------------------------------------*/ 2118 2119 static int received_cbw(struct fsg_dev *fsg, struct fsg_buffhd *bh) 2120 { 2121 struct usb_request *req = bh->outreq; 2122 struct bulk_cb_wrap *cbw = req->buf; 2123 struct fsg_common *common = fsg->common; 2124 2125 /* Was this a real packet? Should it be ignored? */ 2126 if (req->status || test_bit(IGNORE_BULK_OUT, &fsg->atomic_bitflags)) 2127 return -EINVAL; 2128 2129 /* Is the CBW valid? */ 2130 if (req->actual != US_BULK_CB_WRAP_LEN || 2131 cbw->Signature != cpu_to_le32( 2132 US_BULK_CB_SIGN)) { 2133 DBG(fsg, "invalid CBW: len %u sig 0x%x\n", 2134 req->actual, 2135 le32_to_cpu(cbw->Signature)); 2136 2137 /* 2138 * The Bulk-only spec says we MUST stall the IN endpoint 2139 * (6.6.1), so it's unavoidable. It also says we must 2140 * retain this state until the next reset, but there's 2141 * no way to tell the controller driver it should ignore 2142 * Clear-Feature(HALT) requests. 2143 * 2144 * We aren't required to halt the OUT endpoint; instead 2145 * we can simply accept and discard any data received 2146 * until the next reset. 2147 */ 2148 wedge_bulk_in_endpoint(fsg); 2149 set_bit(IGNORE_BULK_OUT, &fsg->atomic_bitflags); 2150 return -EINVAL; 2151 } 2152 2153 /* Is the CBW meaningful? */ 2154 if (cbw->Lun >= ARRAY_SIZE(common->luns) || 2155 cbw->Flags & ~US_BULK_FLAG_IN || cbw->Length <= 0 || 2156 cbw->Length > MAX_COMMAND_SIZE) { 2157 DBG(fsg, "non-meaningful CBW: lun = %u, flags = 0x%x, " 2158 "cmdlen %u\n", 2159 cbw->Lun, cbw->Flags, cbw->Length); 2160 2161 /* 2162 * We can do anything we want here, so let's stall the 2163 * bulk pipes if we are allowed to. 2164 */ 2165 if (common->can_stall) { 2166 fsg_set_halt(fsg, fsg->bulk_out); 2167 halt_bulk_in_endpoint(fsg); 2168 } 2169 return -EINVAL; 2170 } 2171 2172 /* Save the command for later */ 2173 common->cmnd_size = cbw->Length; 2174 memcpy(common->cmnd, cbw->CDB, common->cmnd_size); 2175 if (cbw->Flags & US_BULK_FLAG_IN) 2176 common->data_dir = DATA_DIR_TO_HOST; 2177 else 2178 common->data_dir = DATA_DIR_FROM_HOST; 2179 common->data_size = le32_to_cpu(cbw->DataTransferLength); 2180 if (common->data_size == 0) 2181 common->data_dir = DATA_DIR_NONE; 2182 common->lun = cbw->Lun; 2183 if (common->lun < ARRAY_SIZE(common->luns)) 2184 common->curlun = common->luns[common->lun]; 2185 else 2186 common->curlun = NULL; 2187 common->tag = cbw->Tag; 2188 return 0; 2189 } 2190 2191 static int get_next_command(struct fsg_common *common) 2192 { 2193 struct fsg_buffhd *bh; 2194 int rc = 0; 2195 2196 /* Wait for the next buffer to become available */ 2197 bh = common->next_buffhd_to_fill; 2198 while (bh->state != BUF_STATE_EMPTY) { 2199 rc = sleep_thread(common, true); 2200 if (rc) 2201 return rc; 2202 } 2203 2204 /* Queue a request to read a Bulk-only CBW */ 2205 set_bulk_out_req_length(common, bh, US_BULK_CB_WRAP_LEN); 2206 if (!start_out_transfer(common, bh)) 2207 /* Don't know what to do if common->fsg is NULL */ 2208 return -EIO; 2209 2210 /* 2211 * We will drain the buffer in software, which means we 2212 * can reuse it for the next filling. No need to advance 2213 * next_buffhd_to_fill. 2214 */ 2215 2216 /* Wait for the CBW to arrive */ 2217 while (bh->state != BUF_STATE_FULL) { 2218 rc = sleep_thread(common, true); 2219 if (rc) 2220 return rc; 2221 } 2222 smp_rmb(); 2223 rc = fsg_is_set(common) ? received_cbw(common->fsg, bh) : -EIO; 2224 bh->state = BUF_STATE_EMPTY; 2225 2226 return rc; 2227 } 2228 2229 2230 /*-------------------------------------------------------------------------*/ 2231 2232 static int alloc_request(struct fsg_common *common, struct usb_ep *ep, 2233 struct usb_request **preq) 2234 { 2235 *preq = usb_ep_alloc_request(ep, GFP_ATOMIC); 2236 if (*preq) 2237 return 0; 2238 ERROR(common, "can't allocate request for %s\n", ep->name); 2239 return -ENOMEM; 2240 } 2241 2242 /* Reset interface setting and re-init endpoint state (toggle etc). */ 2243 static int do_set_interface(struct fsg_common *common, struct fsg_dev *new_fsg) 2244 { 2245 struct fsg_dev *fsg; 2246 int i, rc = 0; 2247 2248 if (common->running) 2249 DBG(common, "reset interface\n"); 2250 2251 reset: 2252 /* Deallocate the requests */ 2253 if (common->fsg) { 2254 fsg = common->fsg; 2255 2256 for (i = 0; i < common->fsg_num_buffers; ++i) { 2257 struct fsg_buffhd *bh = &common->buffhds[i]; 2258 2259 if (bh->inreq) { 2260 usb_ep_free_request(fsg->bulk_in, bh->inreq); 2261 bh->inreq = NULL; 2262 } 2263 if (bh->outreq) { 2264 usb_ep_free_request(fsg->bulk_out, bh->outreq); 2265 bh->outreq = NULL; 2266 } 2267 } 2268 2269 /* Disable the endpoints */ 2270 if (fsg->bulk_in_enabled) { 2271 usb_ep_disable(fsg->bulk_in); 2272 fsg->bulk_in_enabled = 0; 2273 } 2274 if (fsg->bulk_out_enabled) { 2275 usb_ep_disable(fsg->bulk_out); 2276 fsg->bulk_out_enabled = 0; 2277 } 2278 2279 common->fsg = NULL; 2280 wake_up(&common->fsg_wait); 2281 } 2282 2283 common->running = 0; 2284 if (!new_fsg || rc) 2285 return rc; 2286 2287 common->fsg = new_fsg; 2288 fsg = common->fsg; 2289 2290 /* Enable the endpoints */ 2291 rc = config_ep_by_speed(common->gadget, &(fsg->function), fsg->bulk_in); 2292 if (rc) 2293 goto reset; 2294 rc = usb_ep_enable(fsg->bulk_in); 2295 if (rc) 2296 goto reset; 2297 fsg->bulk_in->driver_data = common; 2298 fsg->bulk_in_enabled = 1; 2299 2300 rc = config_ep_by_speed(common->gadget, &(fsg->function), 2301 fsg->bulk_out); 2302 if (rc) 2303 goto reset; 2304 rc = usb_ep_enable(fsg->bulk_out); 2305 if (rc) 2306 goto reset; 2307 fsg->bulk_out->driver_data = common; 2308 fsg->bulk_out_enabled = 1; 2309 common->bulk_out_maxpacket = usb_endpoint_maxp(fsg->bulk_out->desc); 2310 clear_bit(IGNORE_BULK_OUT, &fsg->atomic_bitflags); 2311 2312 /* Allocate the requests */ 2313 for (i = 0; i < common->fsg_num_buffers; ++i) { 2314 struct fsg_buffhd *bh = &common->buffhds[i]; 2315 2316 rc = alloc_request(common, fsg->bulk_in, &bh->inreq); 2317 if (rc) 2318 goto reset; 2319 rc = alloc_request(common, fsg->bulk_out, &bh->outreq); 2320 if (rc) 2321 goto reset; 2322 bh->inreq->buf = bh->outreq->buf = bh->buf; 2323 bh->inreq->context = bh->outreq->context = bh; 2324 bh->inreq->complete = bulk_in_complete; 2325 bh->outreq->complete = bulk_out_complete; 2326 } 2327 2328 common->running = 1; 2329 for (i = 0; i < ARRAY_SIZE(common->luns); ++i) 2330 if (common->luns[i]) 2331 common->luns[i]->unit_attention_data = 2332 SS_RESET_OCCURRED; 2333 return rc; 2334 } 2335 2336 2337 /****************************** ALT CONFIGS ******************************/ 2338 2339 static int fsg_set_alt(struct usb_function *f, unsigned intf, unsigned alt) 2340 { 2341 struct fsg_dev *fsg = fsg_from_func(f); 2342 fsg->common->new_fsg = fsg; 2343 raise_exception(fsg->common, FSG_STATE_CONFIG_CHANGE); 2344 return USB_GADGET_DELAYED_STATUS; 2345 } 2346 2347 static void fsg_disable(struct usb_function *f) 2348 { 2349 struct fsg_dev *fsg = fsg_from_func(f); 2350 fsg->common->new_fsg = NULL; 2351 raise_exception(fsg->common, FSG_STATE_CONFIG_CHANGE); 2352 } 2353 2354 2355 /*-------------------------------------------------------------------------*/ 2356 2357 static void handle_exception(struct fsg_common *common) 2358 { 2359 int i; 2360 struct fsg_buffhd *bh; 2361 enum fsg_state old_state; 2362 struct fsg_lun *curlun; 2363 unsigned int exception_req_tag; 2364 2365 /* 2366 * Clear the existing signals. Anything but SIGUSR1 is converted 2367 * into a high-priority EXIT exception. 2368 */ 2369 for (;;) { 2370 int sig = kernel_dequeue_signal(NULL); 2371 if (!sig) 2372 break; 2373 if (sig != SIGUSR1) { 2374 if (common->state < FSG_STATE_EXIT) 2375 DBG(common, "Main thread exiting on signal\n"); 2376 raise_exception(common, FSG_STATE_EXIT); 2377 } 2378 } 2379 2380 /* Cancel all the pending transfers */ 2381 if (likely(common->fsg)) { 2382 for (i = 0; i < common->fsg_num_buffers; ++i) { 2383 bh = &common->buffhds[i]; 2384 if (bh->inreq_busy) 2385 usb_ep_dequeue(common->fsg->bulk_in, bh->inreq); 2386 if (bh->outreq_busy) 2387 usb_ep_dequeue(common->fsg->bulk_out, 2388 bh->outreq); 2389 } 2390 2391 /* Wait until everything is idle */ 2392 for (;;) { 2393 int num_active = 0; 2394 for (i = 0; i < common->fsg_num_buffers; ++i) { 2395 bh = &common->buffhds[i]; 2396 num_active += bh->inreq_busy + bh->outreq_busy; 2397 } 2398 if (num_active == 0) 2399 break; 2400 if (sleep_thread(common, true)) 2401 return; 2402 } 2403 2404 /* Clear out the controller's fifos */ 2405 if (common->fsg->bulk_in_enabled) 2406 usb_ep_fifo_flush(common->fsg->bulk_in); 2407 if (common->fsg->bulk_out_enabled) 2408 usb_ep_fifo_flush(common->fsg->bulk_out); 2409 } 2410 2411 /* 2412 * Reset the I/O buffer states and pointers, the SCSI 2413 * state, and the exception. Then invoke the handler. 2414 */ 2415 spin_lock_irq(&common->lock); 2416 2417 for (i = 0; i < common->fsg_num_buffers; ++i) { 2418 bh = &common->buffhds[i]; 2419 bh->state = BUF_STATE_EMPTY; 2420 } 2421 common->next_buffhd_to_fill = &common->buffhds[0]; 2422 common->next_buffhd_to_drain = &common->buffhds[0]; 2423 exception_req_tag = common->exception_req_tag; 2424 old_state = common->state; 2425 2426 if (old_state == FSG_STATE_ABORT_BULK_OUT) 2427 common->state = FSG_STATE_STATUS_PHASE; 2428 else { 2429 for (i = 0; i < ARRAY_SIZE(common->luns); ++i) { 2430 curlun = common->luns[i]; 2431 if (!curlun) 2432 continue; 2433 curlun->prevent_medium_removal = 0; 2434 curlun->sense_data = SS_NO_SENSE; 2435 curlun->unit_attention_data = SS_NO_SENSE; 2436 curlun->sense_data_info = 0; 2437 curlun->info_valid = 0; 2438 } 2439 common->state = FSG_STATE_IDLE; 2440 } 2441 spin_unlock_irq(&common->lock); 2442 2443 /* Carry out any extra actions required for the exception */ 2444 switch (old_state) { 2445 case FSG_STATE_ABORT_BULK_OUT: 2446 send_status(common); 2447 spin_lock_irq(&common->lock); 2448 if (common->state == FSG_STATE_STATUS_PHASE) 2449 common->state = FSG_STATE_IDLE; 2450 spin_unlock_irq(&common->lock); 2451 break; 2452 2453 case FSG_STATE_RESET: 2454 /* 2455 * In case we were forced against our will to halt a 2456 * bulk endpoint, clear the halt now. (The SuperH UDC 2457 * requires this.) 2458 */ 2459 if (!fsg_is_set(common)) 2460 break; 2461 if (test_and_clear_bit(IGNORE_BULK_OUT, 2462 &common->fsg->atomic_bitflags)) 2463 usb_ep_clear_halt(common->fsg->bulk_in); 2464 2465 if (common->ep0_req_tag == exception_req_tag) 2466 ep0_queue(common); /* Complete the status stage */ 2467 2468 /* 2469 * Technically this should go here, but it would only be 2470 * a waste of time. Ditto for the INTERFACE_CHANGE and 2471 * CONFIG_CHANGE cases. 2472 */ 2473 /* for (i = 0; i < common->ARRAY_SIZE(common->luns); ++i) */ 2474 /* if (common->luns[i]) */ 2475 /* common->luns[i]->unit_attention_data = */ 2476 /* SS_RESET_OCCURRED; */ 2477 break; 2478 2479 case FSG_STATE_CONFIG_CHANGE: 2480 do_set_interface(common, common->new_fsg); 2481 if (common->new_fsg) 2482 usb_composite_setup_continue(common->cdev); 2483 break; 2484 2485 case FSG_STATE_EXIT: 2486 case FSG_STATE_TERMINATED: 2487 do_set_interface(common, NULL); /* Free resources */ 2488 spin_lock_irq(&common->lock); 2489 common->state = FSG_STATE_TERMINATED; /* Stop the thread */ 2490 spin_unlock_irq(&common->lock); 2491 break; 2492 2493 case FSG_STATE_INTERFACE_CHANGE: 2494 case FSG_STATE_DISCONNECT: 2495 case FSG_STATE_COMMAND_PHASE: 2496 case FSG_STATE_DATA_PHASE: 2497 case FSG_STATE_STATUS_PHASE: 2498 case FSG_STATE_IDLE: 2499 break; 2500 } 2501 } 2502 2503 2504 /*-------------------------------------------------------------------------*/ 2505 2506 static int fsg_main_thread(void *common_) 2507 { 2508 struct fsg_common *common = common_; 2509 2510 /* 2511 * Allow the thread to be killed by a signal, but set the signal mask 2512 * to block everything but INT, TERM, KILL, and USR1. 2513 */ 2514 allow_signal(SIGINT); 2515 allow_signal(SIGTERM); 2516 allow_signal(SIGKILL); 2517 allow_signal(SIGUSR1); 2518 2519 /* Allow the thread to be frozen */ 2520 set_freezable(); 2521 2522 /* 2523 * Arrange for userspace references to be interpreted as kernel 2524 * pointers. That way we can pass a kernel pointer to a routine 2525 * that expects a __user pointer and it will work okay. 2526 */ 2527 set_fs(get_ds()); 2528 2529 /* The main loop */ 2530 while (common->state != FSG_STATE_TERMINATED) { 2531 if (exception_in_progress(common) || signal_pending(current)) { 2532 handle_exception(common); 2533 continue; 2534 } 2535 2536 if (!common->running) { 2537 sleep_thread(common, true); 2538 continue; 2539 } 2540 2541 if (get_next_command(common)) 2542 continue; 2543 2544 spin_lock_irq(&common->lock); 2545 if (!exception_in_progress(common)) 2546 common->state = FSG_STATE_DATA_PHASE; 2547 spin_unlock_irq(&common->lock); 2548 2549 if (do_scsi_command(common) || finish_reply(common)) 2550 continue; 2551 2552 spin_lock_irq(&common->lock); 2553 if (!exception_in_progress(common)) 2554 common->state = FSG_STATE_STATUS_PHASE; 2555 spin_unlock_irq(&common->lock); 2556 2557 if (send_status(common)) 2558 continue; 2559 2560 spin_lock_irq(&common->lock); 2561 if (!exception_in_progress(common)) 2562 common->state = FSG_STATE_IDLE; 2563 spin_unlock_irq(&common->lock); 2564 } 2565 2566 spin_lock_irq(&common->lock); 2567 common->thread_task = NULL; 2568 spin_unlock_irq(&common->lock); 2569 2570 if (!common->ops || !common->ops->thread_exits 2571 || common->ops->thread_exits(common) < 0) { 2572 int i; 2573 2574 down_write(&common->filesem); 2575 for (i = 0; i < ARRAY_SIZE(common->luns); --i) { 2576 struct fsg_lun *curlun = common->luns[i]; 2577 if (!curlun || !fsg_lun_is_open(curlun)) 2578 continue; 2579 2580 fsg_lun_close(curlun); 2581 curlun->unit_attention_data = SS_MEDIUM_NOT_PRESENT; 2582 } 2583 up_write(&common->filesem); 2584 } 2585 2586 /* Let fsg_unbind() know the thread has exited */ 2587 complete_and_exit(&common->thread_notifier, 0); 2588 } 2589 2590 2591 /*************************** DEVICE ATTRIBUTES ***************************/ 2592 2593 static ssize_t ro_show(struct device *dev, struct device_attribute *attr, char *buf) 2594 { 2595 struct fsg_lun *curlun = fsg_lun_from_dev(dev); 2596 2597 return fsg_show_ro(curlun, buf); 2598 } 2599 2600 static ssize_t nofua_show(struct device *dev, struct device_attribute *attr, 2601 char *buf) 2602 { 2603 struct fsg_lun *curlun = fsg_lun_from_dev(dev); 2604 2605 return fsg_show_nofua(curlun, buf); 2606 } 2607 2608 static ssize_t file_show(struct device *dev, struct device_attribute *attr, 2609 char *buf) 2610 { 2611 struct fsg_lun *curlun = fsg_lun_from_dev(dev); 2612 struct rw_semaphore *filesem = dev_get_drvdata(dev); 2613 2614 return fsg_show_file(curlun, filesem, buf); 2615 } 2616 2617 static ssize_t ro_store(struct device *dev, struct device_attribute *attr, 2618 const char *buf, size_t count) 2619 { 2620 struct fsg_lun *curlun = fsg_lun_from_dev(dev); 2621 struct rw_semaphore *filesem = dev_get_drvdata(dev); 2622 2623 return fsg_store_ro(curlun, filesem, buf, count); 2624 } 2625 2626 static ssize_t nofua_store(struct device *dev, struct device_attribute *attr, 2627 const char *buf, size_t count) 2628 { 2629 struct fsg_lun *curlun = fsg_lun_from_dev(dev); 2630 2631 return fsg_store_nofua(curlun, buf, count); 2632 } 2633 2634 static ssize_t file_store(struct device *dev, struct device_attribute *attr, 2635 const char *buf, size_t count) 2636 { 2637 struct fsg_lun *curlun = fsg_lun_from_dev(dev); 2638 struct rw_semaphore *filesem = dev_get_drvdata(dev); 2639 2640 return fsg_store_file(curlun, filesem, buf, count); 2641 } 2642 2643 static DEVICE_ATTR_RW(nofua); 2644 /* mode wil be set in fsg_lun_attr_is_visible() */ 2645 static DEVICE_ATTR(ro, 0, ro_show, ro_store); 2646 static DEVICE_ATTR(file, 0, file_show, file_store); 2647 2648 /****************************** FSG COMMON ******************************/ 2649 2650 static void fsg_common_release(struct kref *ref); 2651 2652 static void fsg_lun_release(struct device *dev) 2653 { 2654 /* Nothing needs to be done */ 2655 } 2656 2657 void fsg_common_get(struct fsg_common *common) 2658 { 2659 kref_get(&common->ref); 2660 } 2661 EXPORT_SYMBOL_GPL(fsg_common_get); 2662 2663 void fsg_common_put(struct fsg_common *common) 2664 { 2665 kref_put(&common->ref, fsg_common_release); 2666 } 2667 EXPORT_SYMBOL_GPL(fsg_common_put); 2668 2669 static struct fsg_common *fsg_common_setup(struct fsg_common *common) 2670 { 2671 if (!common) { 2672 common = kzalloc(sizeof(*common), GFP_KERNEL); 2673 if (!common) 2674 return ERR_PTR(-ENOMEM); 2675 common->free_storage_on_release = 1; 2676 } else { 2677 common->free_storage_on_release = 0; 2678 } 2679 init_rwsem(&common->filesem); 2680 spin_lock_init(&common->lock); 2681 kref_init(&common->ref); 2682 init_completion(&common->thread_notifier); 2683 init_waitqueue_head(&common->fsg_wait); 2684 common->state = FSG_STATE_TERMINATED; 2685 memset(common->luns, 0, sizeof(common->luns)); 2686 2687 return common; 2688 } 2689 2690 void fsg_common_set_sysfs(struct fsg_common *common, bool sysfs) 2691 { 2692 common->sysfs = sysfs; 2693 } 2694 EXPORT_SYMBOL_GPL(fsg_common_set_sysfs); 2695 2696 static void _fsg_common_free_buffers(struct fsg_buffhd *buffhds, unsigned n) 2697 { 2698 if (buffhds) { 2699 struct fsg_buffhd *bh = buffhds; 2700 while (n--) { 2701 kfree(bh->buf); 2702 ++bh; 2703 } 2704 kfree(buffhds); 2705 } 2706 } 2707 2708 int fsg_common_set_num_buffers(struct fsg_common *common, unsigned int n) 2709 { 2710 struct fsg_buffhd *bh, *buffhds; 2711 int i; 2712 2713 buffhds = kcalloc(n, sizeof(*buffhds), GFP_KERNEL); 2714 if (!buffhds) 2715 return -ENOMEM; 2716 2717 /* Data buffers cyclic list */ 2718 bh = buffhds; 2719 i = n; 2720 goto buffhds_first_it; 2721 do { 2722 bh->next = bh + 1; 2723 ++bh; 2724 buffhds_first_it: 2725 bh->buf = kmalloc(FSG_BUFLEN, GFP_KERNEL); 2726 if (unlikely(!bh->buf)) 2727 goto error_release; 2728 } while (--i); 2729 bh->next = buffhds; 2730 2731 _fsg_common_free_buffers(common->buffhds, common->fsg_num_buffers); 2732 common->fsg_num_buffers = n; 2733 common->buffhds = buffhds; 2734 2735 return 0; 2736 2737 error_release: 2738 /* 2739 * "buf"s pointed to by heads after n - i are NULL 2740 * so releasing them won't hurt 2741 */ 2742 _fsg_common_free_buffers(buffhds, n); 2743 2744 return -ENOMEM; 2745 } 2746 EXPORT_SYMBOL_GPL(fsg_common_set_num_buffers); 2747 2748 void fsg_common_remove_lun(struct fsg_lun *lun) 2749 { 2750 if (device_is_registered(&lun->dev)) 2751 device_unregister(&lun->dev); 2752 fsg_lun_close(lun); 2753 kfree(lun); 2754 } 2755 EXPORT_SYMBOL_GPL(fsg_common_remove_lun); 2756 2757 static void _fsg_common_remove_luns(struct fsg_common *common, int n) 2758 { 2759 int i; 2760 2761 for (i = 0; i < n; ++i) 2762 if (common->luns[i]) { 2763 fsg_common_remove_lun(common->luns[i]); 2764 common->luns[i] = NULL; 2765 } 2766 } 2767 2768 void fsg_common_remove_luns(struct fsg_common *common) 2769 { 2770 _fsg_common_remove_luns(common, ARRAY_SIZE(common->luns)); 2771 } 2772 EXPORT_SYMBOL_GPL(fsg_common_remove_luns); 2773 2774 void fsg_common_set_ops(struct fsg_common *common, 2775 const struct fsg_operations *ops) 2776 { 2777 common->ops = ops; 2778 } 2779 EXPORT_SYMBOL_GPL(fsg_common_set_ops); 2780 2781 void fsg_common_free_buffers(struct fsg_common *common) 2782 { 2783 _fsg_common_free_buffers(common->buffhds, common->fsg_num_buffers); 2784 common->buffhds = NULL; 2785 } 2786 EXPORT_SYMBOL_GPL(fsg_common_free_buffers); 2787 2788 int fsg_common_set_cdev(struct fsg_common *common, 2789 struct usb_composite_dev *cdev, bool can_stall) 2790 { 2791 struct usb_string *us; 2792 2793 common->gadget = cdev->gadget; 2794 common->ep0 = cdev->gadget->ep0; 2795 common->ep0req = cdev->req; 2796 common->cdev = cdev; 2797 2798 us = usb_gstrings_attach(cdev, fsg_strings_array, 2799 ARRAY_SIZE(fsg_strings)); 2800 if (IS_ERR(us)) 2801 return PTR_ERR(us); 2802 2803 fsg_intf_desc.iInterface = us[FSG_STRING_INTERFACE].id; 2804 2805 /* 2806 * Some peripheral controllers are known not to be able to 2807 * halt bulk endpoints correctly. If one of them is present, 2808 * disable stalls. 2809 */ 2810 common->can_stall = can_stall && 2811 gadget_is_stall_supported(common->gadget); 2812 2813 return 0; 2814 } 2815 EXPORT_SYMBOL_GPL(fsg_common_set_cdev); 2816 2817 static struct attribute *fsg_lun_dev_attrs[] = { 2818 &dev_attr_ro.attr, 2819 &dev_attr_file.attr, 2820 &dev_attr_nofua.attr, 2821 NULL 2822 }; 2823 2824 static umode_t fsg_lun_dev_is_visible(struct kobject *kobj, 2825 struct attribute *attr, int idx) 2826 { 2827 struct device *dev = kobj_to_dev(kobj); 2828 struct fsg_lun *lun = fsg_lun_from_dev(dev); 2829 2830 if (attr == &dev_attr_ro.attr) 2831 return lun->cdrom ? S_IRUGO : (S_IWUSR | S_IRUGO); 2832 if (attr == &dev_attr_file.attr) 2833 return lun->removable ? (S_IWUSR | S_IRUGO) : S_IRUGO; 2834 return attr->mode; 2835 } 2836 2837 static const struct attribute_group fsg_lun_dev_group = { 2838 .attrs = fsg_lun_dev_attrs, 2839 .is_visible = fsg_lun_dev_is_visible, 2840 }; 2841 2842 static const struct attribute_group *fsg_lun_dev_groups[] = { 2843 &fsg_lun_dev_group, 2844 NULL 2845 }; 2846 2847 int fsg_common_create_lun(struct fsg_common *common, struct fsg_lun_config *cfg, 2848 unsigned int id, const char *name, 2849 const char **name_pfx) 2850 { 2851 struct fsg_lun *lun; 2852 char *pathbuf, *p; 2853 int rc = -ENOMEM; 2854 2855 if (id >= ARRAY_SIZE(common->luns)) 2856 return -ENODEV; 2857 2858 if (common->luns[id]) 2859 return -EBUSY; 2860 2861 if (!cfg->filename && !cfg->removable) { 2862 pr_err("no file given for LUN%d\n", id); 2863 return -EINVAL; 2864 } 2865 2866 lun = kzalloc(sizeof(*lun), GFP_KERNEL); 2867 if (!lun) 2868 return -ENOMEM; 2869 2870 lun->name_pfx = name_pfx; 2871 2872 lun->cdrom = !!cfg->cdrom; 2873 lun->ro = cfg->cdrom || cfg->ro; 2874 lun->initially_ro = lun->ro; 2875 lun->removable = !!cfg->removable; 2876 2877 if (!common->sysfs) { 2878 /* we DON'T own the name!*/ 2879 lun->name = name; 2880 } else { 2881 lun->dev.release = fsg_lun_release; 2882 lun->dev.parent = &common->gadget->dev; 2883 lun->dev.groups = fsg_lun_dev_groups; 2884 dev_set_drvdata(&lun->dev, &common->filesem); 2885 dev_set_name(&lun->dev, "%s", name); 2886 lun->name = dev_name(&lun->dev); 2887 2888 rc = device_register(&lun->dev); 2889 if (rc) { 2890 pr_info("failed to register LUN%d: %d\n", id, rc); 2891 put_device(&lun->dev); 2892 goto error_sysfs; 2893 } 2894 } 2895 2896 common->luns[id] = lun; 2897 2898 if (cfg->filename) { 2899 rc = fsg_lun_open(lun, cfg->filename); 2900 if (rc) 2901 goto error_lun; 2902 } 2903 2904 pathbuf = kmalloc(PATH_MAX, GFP_KERNEL); 2905 p = "(no medium)"; 2906 if (fsg_lun_is_open(lun)) { 2907 p = "(error)"; 2908 if (pathbuf) { 2909 p = file_path(lun->filp, pathbuf, PATH_MAX); 2910 if (IS_ERR(p)) 2911 p = "(error)"; 2912 } 2913 } 2914 pr_info("LUN: %s%s%sfile: %s\n", 2915 lun->removable ? "removable " : "", 2916 lun->ro ? "read only " : "", 2917 lun->cdrom ? "CD-ROM " : "", 2918 p); 2919 kfree(pathbuf); 2920 2921 return 0; 2922 2923 error_lun: 2924 if (device_is_registered(&lun->dev)) 2925 device_unregister(&lun->dev); 2926 fsg_lun_close(lun); 2927 common->luns[id] = NULL; 2928 error_sysfs: 2929 kfree(lun); 2930 return rc; 2931 } 2932 EXPORT_SYMBOL_GPL(fsg_common_create_lun); 2933 2934 int fsg_common_create_luns(struct fsg_common *common, struct fsg_config *cfg) 2935 { 2936 char buf[8]; /* enough for 100000000 different numbers, decimal */ 2937 int i, rc; 2938 2939 fsg_common_remove_luns(common); 2940 2941 for (i = 0; i < cfg->nluns; ++i) { 2942 snprintf(buf, sizeof(buf), "lun%d", i); 2943 rc = fsg_common_create_lun(common, &cfg->luns[i], i, buf, NULL); 2944 if (rc) 2945 goto fail; 2946 } 2947 2948 pr_info("Number of LUNs=%d\n", cfg->nluns); 2949 2950 return 0; 2951 2952 fail: 2953 _fsg_common_remove_luns(common, i); 2954 return rc; 2955 } 2956 EXPORT_SYMBOL_GPL(fsg_common_create_luns); 2957 2958 void fsg_common_set_inquiry_string(struct fsg_common *common, const char *vn, 2959 const char *pn) 2960 { 2961 int i; 2962 2963 /* Prepare inquiryString */ 2964 i = get_default_bcdDevice(); 2965 snprintf(common->inquiry_string, sizeof(common->inquiry_string), 2966 "%-8s%-16s%04x", vn ?: "Linux", 2967 /* Assume product name dependent on the first LUN */ 2968 pn ?: ((*common->luns)->cdrom 2969 ? "File-CD Gadget" 2970 : "File-Stor Gadget"), 2971 i); 2972 } 2973 EXPORT_SYMBOL_GPL(fsg_common_set_inquiry_string); 2974 2975 static void fsg_common_release(struct kref *ref) 2976 { 2977 struct fsg_common *common = container_of(ref, struct fsg_common, ref); 2978 int i; 2979 2980 /* If the thread isn't already dead, tell it to exit now */ 2981 if (common->state != FSG_STATE_TERMINATED) { 2982 raise_exception(common, FSG_STATE_EXIT); 2983 wait_for_completion(&common->thread_notifier); 2984 common->thread_task = NULL; 2985 } 2986 2987 for (i = 0; i < ARRAY_SIZE(common->luns); ++i) { 2988 struct fsg_lun *lun = common->luns[i]; 2989 if (!lun) 2990 continue; 2991 fsg_lun_close(lun); 2992 if (device_is_registered(&lun->dev)) 2993 device_unregister(&lun->dev); 2994 kfree(lun); 2995 } 2996 2997 _fsg_common_free_buffers(common->buffhds, common->fsg_num_buffers); 2998 if (common->free_storage_on_release) 2999 kfree(common); 3000 } 3001 3002 3003 /*-------------------------------------------------------------------------*/ 3004 3005 static int fsg_bind(struct usb_configuration *c, struct usb_function *f) 3006 { 3007 struct fsg_dev *fsg = fsg_from_func(f); 3008 struct fsg_common *common = fsg->common; 3009 struct usb_gadget *gadget = c->cdev->gadget; 3010 int i; 3011 struct usb_ep *ep; 3012 unsigned max_burst; 3013 int ret; 3014 struct fsg_opts *opts; 3015 3016 /* Don't allow to bind if we don't have at least one LUN */ 3017 ret = _fsg_common_get_max_lun(common); 3018 if (ret < 0) { 3019 pr_err("There should be at least one LUN.\n"); 3020 return -EINVAL; 3021 } 3022 3023 opts = fsg_opts_from_func_inst(f->fi); 3024 if (!opts->no_configfs) { 3025 ret = fsg_common_set_cdev(fsg->common, c->cdev, 3026 fsg->common->can_stall); 3027 if (ret) 3028 return ret; 3029 fsg_common_set_inquiry_string(fsg->common, NULL, NULL); 3030 } 3031 3032 if (!common->thread_task) { 3033 common->state = FSG_STATE_IDLE; 3034 common->thread_task = 3035 kthread_create(fsg_main_thread, common, "file-storage"); 3036 if (IS_ERR(common->thread_task)) { 3037 int ret = PTR_ERR(common->thread_task); 3038 common->thread_task = NULL; 3039 common->state = FSG_STATE_TERMINATED; 3040 return ret; 3041 } 3042 DBG(common, "I/O thread pid: %d\n", 3043 task_pid_nr(common->thread_task)); 3044 wake_up_process(common->thread_task); 3045 } 3046 3047 fsg->gadget = gadget; 3048 3049 /* New interface */ 3050 i = usb_interface_id(c, f); 3051 if (i < 0) 3052 goto fail; 3053 fsg_intf_desc.bInterfaceNumber = i; 3054 fsg->interface_number = i; 3055 3056 /* Find all the endpoints we will use */ 3057 ep = usb_ep_autoconfig(gadget, &fsg_fs_bulk_in_desc); 3058 if (!ep) 3059 goto autoconf_fail; 3060 fsg->bulk_in = ep; 3061 3062 ep = usb_ep_autoconfig(gadget, &fsg_fs_bulk_out_desc); 3063 if (!ep) 3064 goto autoconf_fail; 3065 fsg->bulk_out = ep; 3066 3067 /* Assume endpoint addresses are the same for both speeds */ 3068 fsg_hs_bulk_in_desc.bEndpointAddress = 3069 fsg_fs_bulk_in_desc.bEndpointAddress; 3070 fsg_hs_bulk_out_desc.bEndpointAddress = 3071 fsg_fs_bulk_out_desc.bEndpointAddress; 3072 3073 /* Calculate bMaxBurst, we know packet size is 1024 */ 3074 max_burst = min_t(unsigned, FSG_BUFLEN / 1024, 15); 3075 3076 fsg_ss_bulk_in_desc.bEndpointAddress = 3077 fsg_fs_bulk_in_desc.bEndpointAddress; 3078 fsg_ss_bulk_in_comp_desc.bMaxBurst = max_burst; 3079 3080 fsg_ss_bulk_out_desc.bEndpointAddress = 3081 fsg_fs_bulk_out_desc.bEndpointAddress; 3082 fsg_ss_bulk_out_comp_desc.bMaxBurst = max_burst; 3083 3084 ret = usb_assign_descriptors(f, fsg_fs_function, fsg_hs_function, 3085 fsg_ss_function, fsg_ss_function); 3086 if (ret) 3087 goto autoconf_fail; 3088 3089 return 0; 3090 3091 autoconf_fail: 3092 ERROR(fsg, "unable to autoconfigure all endpoints\n"); 3093 i = -ENOTSUPP; 3094 fail: 3095 /* terminate the thread */ 3096 if (fsg->common->state != FSG_STATE_TERMINATED) { 3097 raise_exception(fsg->common, FSG_STATE_EXIT); 3098 wait_for_completion(&fsg->common->thread_notifier); 3099 } 3100 return i; 3101 } 3102 3103 /****************************** ALLOCATE FUNCTION *************************/ 3104 3105 static void fsg_unbind(struct usb_configuration *c, struct usb_function *f) 3106 { 3107 struct fsg_dev *fsg = fsg_from_func(f); 3108 struct fsg_common *common = fsg->common; 3109 3110 DBG(fsg, "unbind\n"); 3111 if (fsg->common->fsg == fsg) { 3112 fsg->common->new_fsg = NULL; 3113 raise_exception(fsg->common, FSG_STATE_CONFIG_CHANGE); 3114 /* FIXME: make interruptible or killable somehow? */ 3115 wait_event(common->fsg_wait, common->fsg != fsg); 3116 } 3117 3118 usb_free_all_descriptors(&fsg->function); 3119 } 3120 3121 static inline struct fsg_lun_opts *to_fsg_lun_opts(struct config_item *item) 3122 { 3123 return container_of(to_config_group(item), struct fsg_lun_opts, group); 3124 } 3125 3126 static inline struct fsg_opts *to_fsg_opts(struct config_item *item) 3127 { 3128 return container_of(to_config_group(item), struct fsg_opts, 3129 func_inst.group); 3130 } 3131 3132 static void fsg_lun_attr_release(struct config_item *item) 3133 { 3134 struct fsg_lun_opts *lun_opts; 3135 3136 lun_opts = to_fsg_lun_opts(item); 3137 kfree(lun_opts); 3138 } 3139 3140 static struct configfs_item_operations fsg_lun_item_ops = { 3141 .release = fsg_lun_attr_release, 3142 }; 3143 3144 static ssize_t fsg_lun_opts_file_show(struct config_item *item, char *page) 3145 { 3146 struct fsg_lun_opts *opts = to_fsg_lun_opts(item); 3147 struct fsg_opts *fsg_opts = to_fsg_opts(opts->group.cg_item.ci_parent); 3148 3149 return fsg_show_file(opts->lun, &fsg_opts->common->filesem, page); 3150 } 3151 3152 static ssize_t fsg_lun_opts_file_store(struct config_item *item, 3153 const char *page, size_t len) 3154 { 3155 struct fsg_lun_opts *opts = to_fsg_lun_opts(item); 3156 struct fsg_opts *fsg_opts = to_fsg_opts(opts->group.cg_item.ci_parent); 3157 3158 return fsg_store_file(opts->lun, &fsg_opts->common->filesem, page, len); 3159 } 3160 3161 CONFIGFS_ATTR(fsg_lun_opts_, file); 3162 3163 static ssize_t fsg_lun_opts_ro_show(struct config_item *item, char *page) 3164 { 3165 return fsg_show_ro(to_fsg_lun_opts(item)->lun, page); 3166 } 3167 3168 static ssize_t fsg_lun_opts_ro_store(struct config_item *item, 3169 const char *page, size_t len) 3170 { 3171 struct fsg_lun_opts *opts = to_fsg_lun_opts(item); 3172 struct fsg_opts *fsg_opts = to_fsg_opts(opts->group.cg_item.ci_parent); 3173 3174 return fsg_store_ro(opts->lun, &fsg_opts->common->filesem, page, len); 3175 } 3176 3177 CONFIGFS_ATTR(fsg_lun_opts_, ro); 3178 3179 static ssize_t fsg_lun_opts_removable_show(struct config_item *item, 3180 char *page) 3181 { 3182 return fsg_show_removable(to_fsg_lun_opts(item)->lun, page); 3183 } 3184 3185 static ssize_t fsg_lun_opts_removable_store(struct config_item *item, 3186 const char *page, size_t len) 3187 { 3188 return fsg_store_removable(to_fsg_lun_opts(item)->lun, page, len); 3189 } 3190 3191 CONFIGFS_ATTR(fsg_lun_opts_, removable); 3192 3193 static ssize_t fsg_lun_opts_cdrom_show(struct config_item *item, char *page) 3194 { 3195 return fsg_show_cdrom(to_fsg_lun_opts(item)->lun, page); 3196 } 3197 3198 static ssize_t fsg_lun_opts_cdrom_store(struct config_item *item, 3199 const char *page, size_t len) 3200 { 3201 struct fsg_lun_opts *opts = to_fsg_lun_opts(item); 3202 struct fsg_opts *fsg_opts = to_fsg_opts(opts->group.cg_item.ci_parent); 3203 3204 return fsg_store_cdrom(opts->lun, &fsg_opts->common->filesem, page, 3205 len); 3206 } 3207 3208 CONFIGFS_ATTR(fsg_lun_opts_, cdrom); 3209 3210 static ssize_t fsg_lun_opts_nofua_show(struct config_item *item, char *page) 3211 { 3212 return fsg_show_nofua(to_fsg_lun_opts(item)->lun, page); 3213 } 3214 3215 static ssize_t fsg_lun_opts_nofua_store(struct config_item *item, 3216 const char *page, size_t len) 3217 { 3218 return fsg_store_nofua(to_fsg_lun_opts(item)->lun, page, len); 3219 } 3220 3221 CONFIGFS_ATTR(fsg_lun_opts_, nofua); 3222 3223 static ssize_t fsg_lun_opts_inquiry_string_show(struct config_item *item, 3224 char *page) 3225 { 3226 return fsg_show_inquiry_string(to_fsg_lun_opts(item)->lun, page); 3227 } 3228 3229 static ssize_t fsg_lun_opts_inquiry_string_store(struct config_item *item, 3230 const char *page, size_t len) 3231 { 3232 return fsg_store_inquiry_string(to_fsg_lun_opts(item)->lun, page, len); 3233 } 3234 3235 CONFIGFS_ATTR(fsg_lun_opts_, inquiry_string); 3236 3237 static struct configfs_attribute *fsg_lun_attrs[] = { 3238 &fsg_lun_opts_attr_file, 3239 &fsg_lun_opts_attr_ro, 3240 &fsg_lun_opts_attr_removable, 3241 &fsg_lun_opts_attr_cdrom, 3242 &fsg_lun_opts_attr_nofua, 3243 &fsg_lun_opts_attr_inquiry_string, 3244 NULL, 3245 }; 3246 3247 static struct config_item_type fsg_lun_type = { 3248 .ct_item_ops = &fsg_lun_item_ops, 3249 .ct_attrs = fsg_lun_attrs, 3250 .ct_owner = THIS_MODULE, 3251 }; 3252 3253 static struct config_group *fsg_lun_make(struct config_group *group, 3254 const char *name) 3255 { 3256 struct fsg_lun_opts *opts; 3257 struct fsg_opts *fsg_opts; 3258 struct fsg_lun_config config; 3259 char *num_str; 3260 u8 num; 3261 int ret; 3262 3263 num_str = strchr(name, '.'); 3264 if (!num_str) { 3265 pr_err("Unable to locate . in LUN.NUMBER\n"); 3266 return ERR_PTR(-EINVAL); 3267 } 3268 num_str++; 3269 3270 ret = kstrtou8(num_str, 0, &num); 3271 if (ret) 3272 return ERR_PTR(ret); 3273 3274 fsg_opts = to_fsg_opts(&group->cg_item); 3275 if (num >= FSG_MAX_LUNS) 3276 return ERR_PTR(-ERANGE); 3277 3278 mutex_lock(&fsg_opts->lock); 3279 if (fsg_opts->refcnt || fsg_opts->common->luns[num]) { 3280 ret = -EBUSY; 3281 goto out; 3282 } 3283 3284 opts = kzalloc(sizeof(*opts), GFP_KERNEL); 3285 if (!opts) { 3286 ret = -ENOMEM; 3287 goto out; 3288 } 3289 3290 memset(&config, 0, sizeof(config)); 3291 config.removable = true; 3292 3293 ret = fsg_common_create_lun(fsg_opts->common, &config, num, name, 3294 (const char **)&group->cg_item.ci_name); 3295 if (ret) { 3296 kfree(opts); 3297 goto out; 3298 } 3299 opts->lun = fsg_opts->common->luns[num]; 3300 opts->lun_id = num; 3301 mutex_unlock(&fsg_opts->lock); 3302 3303 config_group_init_type_name(&opts->group, name, &fsg_lun_type); 3304 3305 return &opts->group; 3306 out: 3307 mutex_unlock(&fsg_opts->lock); 3308 return ERR_PTR(ret); 3309 } 3310 3311 static void fsg_lun_drop(struct config_group *group, struct config_item *item) 3312 { 3313 struct fsg_lun_opts *lun_opts; 3314 struct fsg_opts *fsg_opts; 3315 3316 lun_opts = to_fsg_lun_opts(item); 3317 fsg_opts = to_fsg_opts(&group->cg_item); 3318 3319 mutex_lock(&fsg_opts->lock); 3320 if (fsg_opts->refcnt) { 3321 struct config_item *gadget; 3322 3323 gadget = group->cg_item.ci_parent->ci_parent; 3324 unregister_gadget_item(gadget); 3325 } 3326 3327 fsg_common_remove_lun(lun_opts->lun); 3328 fsg_opts->common->luns[lun_opts->lun_id] = NULL; 3329 lun_opts->lun_id = 0; 3330 mutex_unlock(&fsg_opts->lock); 3331 3332 config_item_put(item); 3333 } 3334 3335 static void fsg_attr_release(struct config_item *item) 3336 { 3337 struct fsg_opts *opts = to_fsg_opts(item); 3338 3339 usb_put_function_instance(&opts->func_inst); 3340 } 3341 3342 static struct configfs_item_operations fsg_item_ops = { 3343 .release = fsg_attr_release, 3344 }; 3345 3346 static ssize_t fsg_opts_stall_show(struct config_item *item, char *page) 3347 { 3348 struct fsg_opts *opts = to_fsg_opts(item); 3349 int result; 3350 3351 mutex_lock(&opts->lock); 3352 result = sprintf(page, "%d", opts->common->can_stall); 3353 mutex_unlock(&opts->lock); 3354 3355 return result; 3356 } 3357 3358 static ssize_t fsg_opts_stall_store(struct config_item *item, const char *page, 3359 size_t len) 3360 { 3361 struct fsg_opts *opts = to_fsg_opts(item); 3362 int ret; 3363 bool stall; 3364 3365 mutex_lock(&opts->lock); 3366 3367 if (opts->refcnt) { 3368 mutex_unlock(&opts->lock); 3369 return -EBUSY; 3370 } 3371 3372 ret = strtobool(page, &stall); 3373 if (!ret) { 3374 opts->common->can_stall = stall; 3375 ret = len; 3376 } 3377 3378 mutex_unlock(&opts->lock); 3379 3380 return ret; 3381 } 3382 3383 CONFIGFS_ATTR(fsg_opts_, stall); 3384 3385 #ifdef CONFIG_USB_GADGET_DEBUG_FILES 3386 static ssize_t fsg_opts_num_buffers_show(struct config_item *item, char *page) 3387 { 3388 struct fsg_opts *opts = to_fsg_opts(item); 3389 int result; 3390 3391 mutex_lock(&opts->lock); 3392 result = sprintf(page, "%d", opts->common->fsg_num_buffers); 3393 mutex_unlock(&opts->lock); 3394 3395 return result; 3396 } 3397 3398 static ssize_t fsg_opts_num_buffers_store(struct config_item *item, 3399 const char *page, size_t len) 3400 { 3401 struct fsg_opts *opts = to_fsg_opts(item); 3402 int ret; 3403 u8 num; 3404 3405 mutex_lock(&opts->lock); 3406 if (opts->refcnt) { 3407 ret = -EBUSY; 3408 goto end; 3409 } 3410 ret = kstrtou8(page, 0, &num); 3411 if (ret) 3412 goto end; 3413 3414 fsg_common_set_num_buffers(opts->common, num); 3415 ret = len; 3416 3417 end: 3418 mutex_unlock(&opts->lock); 3419 return ret; 3420 } 3421 3422 CONFIGFS_ATTR(fsg_opts_, num_buffers); 3423 #endif 3424 3425 static struct configfs_attribute *fsg_attrs[] = { 3426 &fsg_opts_attr_stall, 3427 #ifdef CONFIG_USB_GADGET_DEBUG_FILES 3428 &fsg_opts_attr_num_buffers, 3429 #endif 3430 NULL, 3431 }; 3432 3433 static struct configfs_group_operations fsg_group_ops = { 3434 .make_group = fsg_lun_make, 3435 .drop_item = fsg_lun_drop, 3436 }; 3437 3438 static struct config_item_type fsg_func_type = { 3439 .ct_item_ops = &fsg_item_ops, 3440 .ct_group_ops = &fsg_group_ops, 3441 .ct_attrs = fsg_attrs, 3442 .ct_owner = THIS_MODULE, 3443 }; 3444 3445 static void fsg_free_inst(struct usb_function_instance *fi) 3446 { 3447 struct fsg_opts *opts; 3448 3449 opts = fsg_opts_from_func_inst(fi); 3450 fsg_common_put(opts->common); 3451 kfree(opts); 3452 } 3453 3454 static struct usb_function_instance *fsg_alloc_inst(void) 3455 { 3456 struct fsg_opts *opts; 3457 struct fsg_lun_config config; 3458 int rc; 3459 3460 opts = kzalloc(sizeof(*opts), GFP_KERNEL); 3461 if (!opts) 3462 return ERR_PTR(-ENOMEM); 3463 mutex_init(&opts->lock); 3464 opts->func_inst.free_func_inst = fsg_free_inst; 3465 opts->common = fsg_common_setup(opts->common); 3466 if (IS_ERR(opts->common)) { 3467 rc = PTR_ERR(opts->common); 3468 goto release_opts; 3469 } 3470 3471 rc = fsg_common_set_num_buffers(opts->common, 3472 CONFIG_USB_GADGET_STORAGE_NUM_BUFFERS); 3473 if (rc) 3474 goto release_opts; 3475 3476 pr_info(FSG_DRIVER_DESC ", version: " FSG_DRIVER_VERSION "\n"); 3477 3478 memset(&config, 0, sizeof(config)); 3479 config.removable = true; 3480 rc = fsg_common_create_lun(opts->common, &config, 0, "lun.0", 3481 (const char **)&opts->func_inst.group.cg_item.ci_name); 3482 if (rc) 3483 goto release_buffers; 3484 3485 opts->lun0.lun = opts->common->luns[0]; 3486 opts->lun0.lun_id = 0; 3487 3488 config_group_init_type_name(&opts->func_inst.group, "", &fsg_func_type); 3489 3490 config_group_init_type_name(&opts->lun0.group, "lun.0", &fsg_lun_type); 3491 configfs_add_default_group(&opts->lun0.group, &opts->func_inst.group); 3492 3493 return &opts->func_inst; 3494 3495 release_buffers: 3496 fsg_common_free_buffers(opts->common); 3497 release_opts: 3498 kfree(opts); 3499 return ERR_PTR(rc); 3500 } 3501 3502 static void fsg_free(struct usb_function *f) 3503 { 3504 struct fsg_dev *fsg; 3505 struct fsg_opts *opts; 3506 3507 fsg = container_of(f, struct fsg_dev, function); 3508 opts = container_of(f->fi, struct fsg_opts, func_inst); 3509 3510 mutex_lock(&opts->lock); 3511 opts->refcnt--; 3512 mutex_unlock(&opts->lock); 3513 3514 kfree(fsg); 3515 } 3516 3517 static struct usb_function *fsg_alloc(struct usb_function_instance *fi) 3518 { 3519 struct fsg_opts *opts = fsg_opts_from_func_inst(fi); 3520 struct fsg_common *common = opts->common; 3521 struct fsg_dev *fsg; 3522 3523 fsg = kzalloc(sizeof(*fsg), GFP_KERNEL); 3524 if (unlikely(!fsg)) 3525 return ERR_PTR(-ENOMEM); 3526 3527 mutex_lock(&opts->lock); 3528 opts->refcnt++; 3529 mutex_unlock(&opts->lock); 3530 3531 fsg->function.name = FSG_DRIVER_DESC; 3532 fsg->function.bind = fsg_bind; 3533 fsg->function.unbind = fsg_unbind; 3534 fsg->function.setup = fsg_setup; 3535 fsg->function.set_alt = fsg_set_alt; 3536 fsg->function.disable = fsg_disable; 3537 fsg->function.free_func = fsg_free; 3538 3539 fsg->common = common; 3540 3541 return &fsg->function; 3542 } 3543 3544 DECLARE_USB_FUNCTION_INIT(mass_storage, fsg_alloc_inst, fsg_alloc); 3545 MODULE_LICENSE("GPL"); 3546 MODULE_AUTHOR("Michal Nazarewicz"); 3547 3548 /************************* Module parameters *************************/ 3549 3550 3551 void fsg_config_from_params(struct fsg_config *cfg, 3552 const struct fsg_module_parameters *params, 3553 unsigned int fsg_num_buffers) 3554 { 3555 struct fsg_lun_config *lun; 3556 unsigned i; 3557 3558 /* Configure LUNs */ 3559 cfg->nluns = 3560 min(params->luns ?: (params->file_count ?: 1u), 3561 (unsigned)FSG_MAX_LUNS); 3562 for (i = 0, lun = cfg->luns; i < cfg->nluns; ++i, ++lun) { 3563 lun->ro = !!params->ro[i]; 3564 lun->cdrom = !!params->cdrom[i]; 3565 lun->removable = !!params->removable[i]; 3566 lun->filename = 3567 params->file_count > i && params->file[i][0] 3568 ? params->file[i] 3569 : NULL; 3570 } 3571 3572 /* Let MSF use defaults */ 3573 cfg->vendor_name = NULL; 3574 cfg->product_name = NULL; 3575 3576 cfg->ops = NULL; 3577 cfg->private_data = NULL; 3578 3579 /* Finalise */ 3580 cfg->can_stall = params->stall; 3581 cfg->fsg_num_buffers = fsg_num_buffers; 3582 } 3583 EXPORT_SYMBOL_GPL(fsg_config_from_params); 3584