1 /* 2 * iSCSI over TCP/IP Data-Path lib 3 * 4 * Copyright (C) 2004 Dmitry Yusupov 5 * Copyright (C) 2004 Alex Aizman 6 * Copyright (C) 2005 - 2006 Mike Christie 7 * Copyright (C) 2006 Red Hat, Inc. All rights reserved. 8 * maintained by open-iscsi@googlegroups.com 9 * 10 * This program is free software; you can redistribute it and/or modify 11 * it under the terms of the GNU General Public License as published 12 * by the Free Software Foundation; either version 2 of the License, or 13 * (at your option) any later version. 14 * 15 * This program is distributed in the hope that it will be useful, but 16 * WITHOUT ANY WARRANTY; without even the implied warranty of 17 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 18 * General Public License for more details. 19 * 20 * See the file COPYING included with this distribution for more details. 21 * 22 * Credits: 23 * Christoph Hellwig 24 * FUJITA Tomonori 25 * Arne Redlich 26 * Zhenyu Wang 27 */ 28 29 #include <linux/types.h> 30 #include <linux/list.h> 31 #include <linux/inet.h> 32 #include <linux/slab.h> 33 #include <linux/file.h> 34 #include <linux/blkdev.h> 35 #include <linux/crypto.h> 36 #include <linux/delay.h> 37 #include <linux/kfifo.h> 38 #include <linux/scatterlist.h> 39 #include <net/tcp.h> 40 #include <scsi/scsi_cmnd.h> 41 #include <scsi/scsi_device.h> 42 #include <scsi/scsi_host.h> 43 #include <scsi/scsi.h> 44 #include <scsi/scsi_transport_iscsi.h> 45 46 #include "iscsi_tcp.h" 47 48 MODULE_AUTHOR("Mike Christie <michaelc@cs.wisc.edu>, " 49 "Dmitry Yusupov <dmitry_yus@yahoo.com>, " 50 "Alex Aizman <itn780@yahoo.com>"); 51 MODULE_DESCRIPTION("iSCSI/TCP data-path"); 52 MODULE_LICENSE("GPL"); 53 54 static int iscsi_dbg_libtcp; 55 module_param_named(debug_libiscsi_tcp, iscsi_dbg_libtcp, int, 56 S_IRUGO | S_IWUSR); 57 MODULE_PARM_DESC(debug_libiscsi_tcp, "Turn on debugging for libiscsi_tcp " 58 "module. Set to 1 to turn on, and zero to turn off. Default " 59 "is off."); 60 61 #define ISCSI_DBG_TCP(_conn, dbg_fmt, arg...) \ 62 do { \ 63 if (iscsi_dbg_libtcp) \ 64 iscsi_conn_printk(KERN_INFO, _conn, \ 65 "%s " dbg_fmt, \ 66 __func__, ##arg); \ 67 } while (0); 68 69 static int iscsi_tcp_hdr_recv_done(struct iscsi_tcp_conn *tcp_conn, 70 struct iscsi_segment *segment); 71 72 /* 73 * Scatterlist handling: inside the iscsi_segment, we 74 * remember an index into the scatterlist, and set data/size 75 * to the current scatterlist entry. For highmem pages, we 76 * kmap as needed. 77 * 78 * Note that the page is unmapped when we return from 79 * TCP's data_ready handler, so we may end up mapping and 80 * unmapping the same page repeatedly. The whole reason 81 * for this is that we shouldn't keep the page mapped 82 * outside the softirq. 83 */ 84 85 /** 86 * iscsi_tcp_segment_init_sg - init indicated scatterlist entry 87 * @segment: the buffer object 88 * @sg: scatterlist 89 * @offset: byte offset into that sg entry 90 * 91 * This function sets up the segment so that subsequent 92 * data is copied to the indicated sg entry, at the given 93 * offset. 94 */ 95 static inline void 96 iscsi_tcp_segment_init_sg(struct iscsi_segment *segment, 97 struct scatterlist *sg, unsigned int offset) 98 { 99 segment->sg = sg; 100 segment->sg_offset = offset; 101 segment->size = min(sg->length - offset, 102 segment->total_size - segment->total_copied); 103 segment->data = NULL; 104 } 105 106 /** 107 * iscsi_tcp_segment_map - map the current S/G page 108 * @segment: iscsi_segment 109 * @recv: 1 if called from recv path 110 * 111 * We only need to possibly kmap data if scatter lists are being used, 112 * because the iscsi passthrough and internal IO paths will never use high 113 * mem pages. 114 */ 115 static void iscsi_tcp_segment_map(struct iscsi_segment *segment, int recv) 116 { 117 struct scatterlist *sg; 118 119 if (segment->data != NULL || !segment->sg) 120 return; 121 122 sg = segment->sg; 123 BUG_ON(segment->sg_mapped); 124 BUG_ON(sg->length == 0); 125 126 /* 127 * If the page count is greater than one it is ok to send 128 * to the network layer's zero copy send path. If not we 129 * have to go the slow sendmsg path. We always map for the 130 * recv path. 131 */ 132 if (page_count(sg_page(sg)) >= 1 && !recv) 133 return; 134 135 if (recv) { 136 segment->atomic_mapped = true; 137 segment->sg_mapped = kmap_atomic(sg_page(sg), KM_SOFTIRQ0); 138 } else { 139 segment->atomic_mapped = false; 140 /* the xmit path can sleep with the page mapped so use kmap */ 141 segment->sg_mapped = kmap(sg_page(sg)); 142 } 143 144 segment->data = segment->sg_mapped + sg->offset + segment->sg_offset; 145 } 146 147 void iscsi_tcp_segment_unmap(struct iscsi_segment *segment) 148 { 149 if (segment->sg_mapped) { 150 if (segment->atomic_mapped) 151 kunmap_atomic(segment->sg_mapped, KM_SOFTIRQ0); 152 else 153 kunmap(sg_page(segment->sg)); 154 segment->sg_mapped = NULL; 155 segment->data = NULL; 156 } 157 } 158 EXPORT_SYMBOL_GPL(iscsi_tcp_segment_unmap); 159 160 /* 161 * Splice the digest buffer into the buffer 162 */ 163 static inline void 164 iscsi_tcp_segment_splice_digest(struct iscsi_segment *segment, void *digest) 165 { 166 segment->data = digest; 167 segment->digest_len = ISCSI_DIGEST_SIZE; 168 segment->total_size += ISCSI_DIGEST_SIZE; 169 segment->size = ISCSI_DIGEST_SIZE; 170 segment->copied = 0; 171 segment->sg = NULL; 172 segment->hash = NULL; 173 } 174 175 /** 176 * iscsi_tcp_segment_done - check whether the segment is complete 177 * @tcp_conn: iscsi tcp connection 178 * @segment: iscsi segment to check 179 * @recv: set to one of this is called from the recv path 180 * @copied: number of bytes copied 181 * 182 * Check if we're done receiving this segment. If the receive 183 * buffer is full but we expect more data, move on to the 184 * next entry in the scatterlist. 185 * 186 * If the amount of data we received isn't a multiple of 4, 187 * we will transparently receive the pad bytes, too. 188 * 189 * This function must be re-entrant. 190 */ 191 int iscsi_tcp_segment_done(struct iscsi_tcp_conn *tcp_conn, 192 struct iscsi_segment *segment, int recv, 193 unsigned copied) 194 { 195 struct scatterlist sg; 196 unsigned int pad; 197 198 ISCSI_DBG_TCP(tcp_conn->iscsi_conn, "copied %u %u size %u %s\n", 199 segment->copied, copied, segment->size, 200 recv ? "recv" : "xmit"); 201 if (segment->hash && copied) { 202 /* 203 * If a segment is kmapd we must unmap it before sending 204 * to the crypto layer since that will try to kmap it again. 205 */ 206 iscsi_tcp_segment_unmap(segment); 207 208 if (!segment->data) { 209 sg_init_table(&sg, 1); 210 sg_set_page(&sg, sg_page(segment->sg), copied, 211 segment->copied + segment->sg_offset + 212 segment->sg->offset); 213 } else 214 sg_init_one(&sg, segment->data + segment->copied, 215 copied); 216 crypto_hash_update(segment->hash, &sg, copied); 217 } 218 219 segment->copied += copied; 220 if (segment->copied < segment->size) { 221 iscsi_tcp_segment_map(segment, recv); 222 return 0; 223 } 224 225 segment->total_copied += segment->copied; 226 segment->copied = 0; 227 segment->size = 0; 228 229 /* Unmap the current scatterlist page, if there is one. */ 230 iscsi_tcp_segment_unmap(segment); 231 232 /* Do we have more scatterlist entries? */ 233 ISCSI_DBG_TCP(tcp_conn->iscsi_conn, "total copied %u total size %u\n", 234 segment->total_copied, segment->total_size); 235 if (segment->total_copied < segment->total_size) { 236 /* Proceed to the next entry in the scatterlist. */ 237 iscsi_tcp_segment_init_sg(segment, sg_next(segment->sg), 238 0); 239 iscsi_tcp_segment_map(segment, recv); 240 BUG_ON(segment->size == 0); 241 return 0; 242 } 243 244 /* Do we need to handle padding? */ 245 if (!(tcp_conn->iscsi_conn->session->tt->caps & CAP_PADDING_OFFLOAD)) { 246 pad = iscsi_padding(segment->total_copied); 247 if (pad != 0) { 248 ISCSI_DBG_TCP(tcp_conn->iscsi_conn, 249 "consume %d pad bytes\n", pad); 250 segment->total_size += pad; 251 segment->size = pad; 252 segment->data = segment->padbuf; 253 return 0; 254 } 255 } 256 257 /* 258 * Set us up for transferring the data digest. hdr digest 259 * is completely handled in hdr done function. 260 */ 261 if (segment->hash) { 262 crypto_hash_final(segment->hash, segment->digest); 263 iscsi_tcp_segment_splice_digest(segment, 264 recv ? segment->recv_digest : segment->digest); 265 return 0; 266 } 267 268 return 1; 269 } 270 EXPORT_SYMBOL_GPL(iscsi_tcp_segment_done); 271 272 /** 273 * iscsi_tcp_segment_recv - copy data to segment 274 * @tcp_conn: the iSCSI TCP connection 275 * @segment: the buffer to copy to 276 * @ptr: data pointer 277 * @len: amount of data available 278 * 279 * This function copies up to @len bytes to the 280 * given buffer, and returns the number of bytes 281 * consumed, which can actually be less than @len. 282 * 283 * If hash digest is enabled, the function will update the 284 * hash while copying. 285 * Combining these two operations doesn't buy us a lot (yet), 286 * but in the future we could implement combined copy+crc, 287 * just way we do for network layer checksums. 288 */ 289 static int 290 iscsi_tcp_segment_recv(struct iscsi_tcp_conn *tcp_conn, 291 struct iscsi_segment *segment, const void *ptr, 292 unsigned int len) 293 { 294 unsigned int copy = 0, copied = 0; 295 296 while (!iscsi_tcp_segment_done(tcp_conn, segment, 1, copy)) { 297 if (copied == len) { 298 ISCSI_DBG_TCP(tcp_conn->iscsi_conn, 299 "copied %d bytes\n", len); 300 break; 301 } 302 303 copy = min(len - copied, segment->size - segment->copied); 304 ISCSI_DBG_TCP(tcp_conn->iscsi_conn, "copying %d\n", copy); 305 memcpy(segment->data + segment->copied, ptr + copied, copy); 306 copied += copy; 307 } 308 return copied; 309 } 310 311 inline void 312 iscsi_tcp_dgst_header(struct hash_desc *hash, const void *hdr, size_t hdrlen, 313 unsigned char digest[ISCSI_DIGEST_SIZE]) 314 { 315 struct scatterlist sg; 316 317 sg_init_one(&sg, hdr, hdrlen); 318 crypto_hash_digest(hash, &sg, hdrlen, digest); 319 } 320 EXPORT_SYMBOL_GPL(iscsi_tcp_dgst_header); 321 322 static inline int 323 iscsi_tcp_dgst_verify(struct iscsi_tcp_conn *tcp_conn, 324 struct iscsi_segment *segment) 325 { 326 if (!segment->digest_len) 327 return 1; 328 329 if (memcmp(segment->recv_digest, segment->digest, 330 segment->digest_len)) { 331 ISCSI_DBG_TCP(tcp_conn->iscsi_conn, "digest mismatch\n"); 332 return 0; 333 } 334 335 return 1; 336 } 337 338 /* 339 * Helper function to set up segment buffer 340 */ 341 static inline void 342 __iscsi_segment_init(struct iscsi_segment *segment, size_t size, 343 iscsi_segment_done_fn_t *done, struct hash_desc *hash) 344 { 345 memset(segment, 0, sizeof(*segment)); 346 segment->total_size = size; 347 segment->done = done; 348 349 if (hash) { 350 segment->hash = hash; 351 crypto_hash_init(hash); 352 } 353 } 354 355 inline void 356 iscsi_segment_init_linear(struct iscsi_segment *segment, void *data, 357 size_t size, iscsi_segment_done_fn_t *done, 358 struct hash_desc *hash) 359 { 360 __iscsi_segment_init(segment, size, done, hash); 361 segment->data = data; 362 segment->size = size; 363 } 364 EXPORT_SYMBOL_GPL(iscsi_segment_init_linear); 365 366 inline int 367 iscsi_segment_seek_sg(struct iscsi_segment *segment, 368 struct scatterlist *sg_list, unsigned int sg_count, 369 unsigned int offset, size_t size, 370 iscsi_segment_done_fn_t *done, struct hash_desc *hash) 371 { 372 struct scatterlist *sg; 373 unsigned int i; 374 375 __iscsi_segment_init(segment, size, done, hash); 376 for_each_sg(sg_list, sg, sg_count, i) { 377 if (offset < sg->length) { 378 iscsi_tcp_segment_init_sg(segment, sg, offset); 379 return 0; 380 } 381 offset -= sg->length; 382 } 383 384 return ISCSI_ERR_DATA_OFFSET; 385 } 386 EXPORT_SYMBOL_GPL(iscsi_segment_seek_sg); 387 388 /** 389 * iscsi_tcp_hdr_recv_prep - prep segment for hdr reception 390 * @tcp_conn: iscsi connection to prep for 391 * 392 * This function always passes NULL for the hash argument, because when this 393 * function is called we do not yet know the final size of the header and want 394 * to delay the digest processing until we know that. 395 */ 396 void iscsi_tcp_hdr_recv_prep(struct iscsi_tcp_conn *tcp_conn) 397 { 398 ISCSI_DBG_TCP(tcp_conn->iscsi_conn, 399 "(%s)\n", tcp_conn->iscsi_conn->hdrdgst_en ? 400 "digest enabled" : "digest disabled"); 401 iscsi_segment_init_linear(&tcp_conn->in.segment, 402 tcp_conn->in.hdr_buf, sizeof(struct iscsi_hdr), 403 iscsi_tcp_hdr_recv_done, NULL); 404 } 405 EXPORT_SYMBOL_GPL(iscsi_tcp_hdr_recv_prep); 406 407 /* 408 * Handle incoming reply to any other type of command 409 */ 410 static int 411 iscsi_tcp_data_recv_done(struct iscsi_tcp_conn *tcp_conn, 412 struct iscsi_segment *segment) 413 { 414 struct iscsi_conn *conn = tcp_conn->iscsi_conn; 415 int rc = 0; 416 417 if (!iscsi_tcp_dgst_verify(tcp_conn, segment)) 418 return ISCSI_ERR_DATA_DGST; 419 420 rc = iscsi_complete_pdu(conn, tcp_conn->in.hdr, 421 conn->data, tcp_conn->in.datalen); 422 if (rc) 423 return rc; 424 425 iscsi_tcp_hdr_recv_prep(tcp_conn); 426 return 0; 427 } 428 429 static void 430 iscsi_tcp_data_recv_prep(struct iscsi_tcp_conn *tcp_conn) 431 { 432 struct iscsi_conn *conn = tcp_conn->iscsi_conn; 433 struct hash_desc *rx_hash = NULL; 434 435 if (conn->datadgst_en && 436 !(conn->session->tt->caps & CAP_DIGEST_OFFLOAD)) 437 rx_hash = tcp_conn->rx_hash; 438 439 iscsi_segment_init_linear(&tcp_conn->in.segment, 440 conn->data, tcp_conn->in.datalen, 441 iscsi_tcp_data_recv_done, rx_hash); 442 } 443 444 /** 445 * iscsi_tcp_cleanup_task - free tcp_task resources 446 * @task: iscsi task 447 * 448 * must be called with session lock 449 */ 450 void iscsi_tcp_cleanup_task(struct iscsi_task *task) 451 { 452 struct iscsi_tcp_task *tcp_task = task->dd_data; 453 struct iscsi_r2t_info *r2t; 454 455 /* nothing to do for mgmt */ 456 if (!task->sc) 457 return; 458 459 /* flush task's r2t queues */ 460 while (kfifo_out(&tcp_task->r2tqueue, (void*)&r2t, sizeof(void*))) { 461 kfifo_in(&tcp_task->r2tpool.queue, (void*)&r2t, 462 sizeof(void*)); 463 ISCSI_DBG_TCP(task->conn, "pending r2t dropped\n"); 464 } 465 466 r2t = tcp_task->r2t; 467 if (r2t != NULL) { 468 kfifo_in(&tcp_task->r2tpool.queue, (void*)&r2t, 469 sizeof(void*)); 470 tcp_task->r2t = NULL; 471 } 472 } 473 EXPORT_SYMBOL_GPL(iscsi_tcp_cleanup_task); 474 475 /** 476 * iscsi_tcp_data_in - SCSI Data-In Response processing 477 * @conn: iscsi connection 478 * @task: scsi command task 479 */ 480 static int iscsi_tcp_data_in(struct iscsi_conn *conn, struct iscsi_task *task) 481 { 482 struct iscsi_tcp_conn *tcp_conn = conn->dd_data; 483 struct iscsi_tcp_task *tcp_task = task->dd_data; 484 struct iscsi_data_rsp *rhdr = (struct iscsi_data_rsp *)tcp_conn->in.hdr; 485 int datasn = be32_to_cpu(rhdr->datasn); 486 unsigned total_in_length = scsi_in(task->sc)->length; 487 488 /* 489 * lib iscsi will update this in the completion handling if there 490 * is status. 491 */ 492 if (!(rhdr->flags & ISCSI_FLAG_DATA_STATUS)) 493 iscsi_update_cmdsn(conn->session, (struct iscsi_nopin*)rhdr); 494 495 if (tcp_conn->in.datalen == 0) 496 return 0; 497 498 if (tcp_task->exp_datasn != datasn) { 499 ISCSI_DBG_TCP(conn, "task->exp_datasn(%d) != rhdr->datasn(%d)" 500 "\n", tcp_task->exp_datasn, datasn); 501 return ISCSI_ERR_DATASN; 502 } 503 504 tcp_task->exp_datasn++; 505 506 tcp_task->data_offset = be32_to_cpu(rhdr->offset); 507 if (tcp_task->data_offset + tcp_conn->in.datalen > total_in_length) { 508 ISCSI_DBG_TCP(conn, "data_offset(%d) + data_len(%d) > " 509 "total_length_in(%d)\n", tcp_task->data_offset, 510 tcp_conn->in.datalen, total_in_length); 511 return ISCSI_ERR_DATA_OFFSET; 512 } 513 514 conn->datain_pdus_cnt++; 515 return 0; 516 } 517 518 /** 519 * iscsi_tcp_r2t_rsp - iSCSI R2T Response processing 520 * @conn: iscsi connection 521 * @task: scsi command task 522 */ 523 static int iscsi_tcp_r2t_rsp(struct iscsi_conn *conn, struct iscsi_task *task) 524 { 525 struct iscsi_session *session = conn->session; 526 struct iscsi_tcp_task *tcp_task = task->dd_data; 527 struct iscsi_tcp_conn *tcp_conn = conn->dd_data; 528 struct iscsi_r2t_rsp *rhdr = (struct iscsi_r2t_rsp *)tcp_conn->in.hdr; 529 struct iscsi_r2t_info *r2t; 530 int r2tsn = be32_to_cpu(rhdr->r2tsn); 531 int rc; 532 533 if (tcp_conn->in.datalen) { 534 iscsi_conn_printk(KERN_ERR, conn, 535 "invalid R2t with datalen %d\n", 536 tcp_conn->in.datalen); 537 return ISCSI_ERR_DATALEN; 538 } 539 540 if (tcp_task->exp_datasn != r2tsn){ 541 ISCSI_DBG_TCP(conn, "task->exp_datasn(%d) != rhdr->r2tsn(%d)\n", 542 tcp_task->exp_datasn, r2tsn); 543 return ISCSI_ERR_R2TSN; 544 } 545 546 /* fill-in new R2T associated with the task */ 547 iscsi_update_cmdsn(session, (struct iscsi_nopin*)rhdr); 548 549 if (!task->sc || session->state != ISCSI_STATE_LOGGED_IN) { 550 iscsi_conn_printk(KERN_INFO, conn, 551 "dropping R2T itt %d in recovery.\n", 552 task->itt); 553 return 0; 554 } 555 556 rc = kfifo_out(&tcp_task->r2tpool.queue, (void*)&r2t, sizeof(void*)); 557 if (!rc) { 558 iscsi_conn_printk(KERN_ERR, conn, "Could not allocate R2T. " 559 "Target has sent more R2Ts than it " 560 "negotiated for or driver has has leaked.\n"); 561 return ISCSI_ERR_PROTO; 562 } 563 564 r2t->exp_statsn = rhdr->statsn; 565 r2t->data_length = be32_to_cpu(rhdr->data_length); 566 if (r2t->data_length == 0) { 567 iscsi_conn_printk(KERN_ERR, conn, 568 "invalid R2T with zero data len\n"); 569 kfifo_in(&tcp_task->r2tpool.queue, (void*)&r2t, 570 sizeof(void*)); 571 return ISCSI_ERR_DATALEN; 572 } 573 574 if (r2t->data_length > session->max_burst) 575 ISCSI_DBG_TCP(conn, "invalid R2T with data len %u and max " 576 "burst %u. Attempting to execute request.\n", 577 r2t->data_length, session->max_burst); 578 579 r2t->data_offset = be32_to_cpu(rhdr->data_offset); 580 if (r2t->data_offset + r2t->data_length > scsi_out(task->sc)->length) { 581 iscsi_conn_printk(KERN_ERR, conn, 582 "invalid R2T with data len %u at offset %u " 583 "and total length %d\n", r2t->data_length, 584 r2t->data_offset, scsi_out(task->sc)->length); 585 kfifo_in(&tcp_task->r2tpool.queue, (void*)&r2t, 586 sizeof(void*)); 587 return ISCSI_ERR_DATALEN; 588 } 589 590 r2t->ttt = rhdr->ttt; /* no flip */ 591 r2t->datasn = 0; 592 r2t->sent = 0; 593 594 tcp_task->exp_datasn = r2tsn + 1; 595 kfifo_in(&tcp_task->r2tqueue, (void*)&r2t, sizeof(void*)); 596 conn->r2t_pdus_cnt++; 597 598 iscsi_requeue_task(task); 599 return 0; 600 } 601 602 /* 603 * Handle incoming reply to DataIn command 604 */ 605 static int 606 iscsi_tcp_process_data_in(struct iscsi_tcp_conn *tcp_conn, 607 struct iscsi_segment *segment) 608 { 609 struct iscsi_conn *conn = tcp_conn->iscsi_conn; 610 struct iscsi_hdr *hdr = tcp_conn->in.hdr; 611 int rc; 612 613 if (!iscsi_tcp_dgst_verify(tcp_conn, segment)) 614 return ISCSI_ERR_DATA_DGST; 615 616 /* check for non-exceptional status */ 617 if (hdr->flags & ISCSI_FLAG_DATA_STATUS) { 618 rc = iscsi_complete_pdu(conn, tcp_conn->in.hdr, NULL, 0); 619 if (rc) 620 return rc; 621 } 622 623 iscsi_tcp_hdr_recv_prep(tcp_conn); 624 return 0; 625 } 626 627 /** 628 * iscsi_tcp_hdr_dissect - process PDU header 629 * @conn: iSCSI connection 630 * @hdr: PDU header 631 * 632 * This function analyzes the header of the PDU received, 633 * and performs several sanity checks. If the PDU is accompanied 634 * by data, the receive buffer is set up to copy the incoming data 635 * to the correct location. 636 */ 637 static int 638 iscsi_tcp_hdr_dissect(struct iscsi_conn *conn, struct iscsi_hdr *hdr) 639 { 640 int rc = 0, opcode, ahslen; 641 struct iscsi_tcp_conn *tcp_conn = conn->dd_data; 642 struct iscsi_task *task; 643 644 /* verify PDU length */ 645 tcp_conn->in.datalen = ntoh24(hdr->dlength); 646 if (tcp_conn->in.datalen > conn->max_recv_dlength) { 647 iscsi_conn_printk(KERN_ERR, conn, 648 "iscsi_tcp: datalen %d > %d\n", 649 tcp_conn->in.datalen, conn->max_recv_dlength); 650 return ISCSI_ERR_DATALEN; 651 } 652 653 /* Additional header segments. So far, we don't 654 * process additional headers. 655 */ 656 ahslen = hdr->hlength << 2; 657 658 opcode = hdr->opcode & ISCSI_OPCODE_MASK; 659 /* verify itt (itt encoding: age+cid+itt) */ 660 rc = iscsi_verify_itt(conn, hdr->itt); 661 if (rc) 662 return rc; 663 664 ISCSI_DBG_TCP(conn, "opcode 0x%x ahslen %d datalen %d\n", 665 opcode, ahslen, tcp_conn->in.datalen); 666 667 switch(opcode) { 668 case ISCSI_OP_SCSI_DATA_IN: 669 spin_lock(&conn->session->lock); 670 task = iscsi_itt_to_ctask(conn, hdr->itt); 671 if (!task) 672 rc = ISCSI_ERR_BAD_ITT; 673 else 674 rc = iscsi_tcp_data_in(conn, task); 675 if (rc) { 676 spin_unlock(&conn->session->lock); 677 break; 678 } 679 680 if (tcp_conn->in.datalen) { 681 struct iscsi_tcp_task *tcp_task = task->dd_data; 682 struct hash_desc *rx_hash = NULL; 683 struct scsi_data_buffer *sdb = scsi_in(task->sc); 684 685 /* 686 * Setup copy of Data-In into the Scsi_Cmnd 687 * Scatterlist case: 688 * We set up the iscsi_segment to point to the next 689 * scatterlist entry to copy to. As we go along, 690 * we move on to the next scatterlist entry and 691 * update the digest per-entry. 692 */ 693 if (conn->datadgst_en && 694 !(conn->session->tt->caps & CAP_DIGEST_OFFLOAD)) 695 rx_hash = tcp_conn->rx_hash; 696 697 ISCSI_DBG_TCP(conn, "iscsi_tcp_begin_data_in( " 698 "offset=%d, datalen=%d)\n", 699 tcp_task->data_offset, 700 tcp_conn->in.datalen); 701 task->last_xfer = jiffies; 702 rc = iscsi_segment_seek_sg(&tcp_conn->in.segment, 703 sdb->table.sgl, 704 sdb->table.nents, 705 tcp_task->data_offset, 706 tcp_conn->in.datalen, 707 iscsi_tcp_process_data_in, 708 rx_hash); 709 spin_unlock(&conn->session->lock); 710 return rc; 711 } 712 rc = __iscsi_complete_pdu(conn, hdr, NULL, 0); 713 spin_unlock(&conn->session->lock); 714 break; 715 case ISCSI_OP_SCSI_CMD_RSP: 716 if (tcp_conn->in.datalen) { 717 iscsi_tcp_data_recv_prep(tcp_conn); 718 return 0; 719 } 720 rc = iscsi_complete_pdu(conn, hdr, NULL, 0); 721 break; 722 case ISCSI_OP_R2T: 723 spin_lock(&conn->session->lock); 724 task = iscsi_itt_to_ctask(conn, hdr->itt); 725 if (!task) 726 rc = ISCSI_ERR_BAD_ITT; 727 else if (ahslen) 728 rc = ISCSI_ERR_AHSLEN; 729 else if (task->sc->sc_data_direction == DMA_TO_DEVICE) { 730 task->last_xfer = jiffies; 731 rc = iscsi_tcp_r2t_rsp(conn, task); 732 } else 733 rc = ISCSI_ERR_PROTO; 734 spin_unlock(&conn->session->lock); 735 break; 736 case ISCSI_OP_LOGIN_RSP: 737 case ISCSI_OP_TEXT_RSP: 738 case ISCSI_OP_REJECT: 739 case ISCSI_OP_ASYNC_EVENT: 740 /* 741 * It is possible that we could get a PDU with a buffer larger 742 * than 8K, but there are no targets that currently do this. 743 * For now we fail until we find a vendor that needs it 744 */ 745 if (ISCSI_DEF_MAX_RECV_SEG_LEN < tcp_conn->in.datalen) { 746 iscsi_conn_printk(KERN_ERR, conn, 747 "iscsi_tcp: received buffer of " 748 "len %u but conn buffer is only %u " 749 "(opcode %0x)\n", 750 tcp_conn->in.datalen, 751 ISCSI_DEF_MAX_RECV_SEG_LEN, opcode); 752 rc = ISCSI_ERR_PROTO; 753 break; 754 } 755 756 /* If there's data coming in with the response, 757 * receive it to the connection's buffer. 758 */ 759 if (tcp_conn->in.datalen) { 760 iscsi_tcp_data_recv_prep(tcp_conn); 761 return 0; 762 } 763 /* fall through */ 764 case ISCSI_OP_LOGOUT_RSP: 765 case ISCSI_OP_NOOP_IN: 766 case ISCSI_OP_SCSI_TMFUNC_RSP: 767 rc = iscsi_complete_pdu(conn, hdr, NULL, 0); 768 break; 769 default: 770 rc = ISCSI_ERR_BAD_OPCODE; 771 break; 772 } 773 774 if (rc == 0) { 775 /* Anything that comes with data should have 776 * been handled above. */ 777 if (tcp_conn->in.datalen) 778 return ISCSI_ERR_PROTO; 779 iscsi_tcp_hdr_recv_prep(tcp_conn); 780 } 781 782 return rc; 783 } 784 785 /** 786 * iscsi_tcp_hdr_recv_done - process PDU header 787 * 788 * This is the callback invoked when the PDU header has 789 * been received. If the header is followed by additional 790 * header segments, we go back for more data. 791 */ 792 static int 793 iscsi_tcp_hdr_recv_done(struct iscsi_tcp_conn *tcp_conn, 794 struct iscsi_segment *segment) 795 { 796 struct iscsi_conn *conn = tcp_conn->iscsi_conn; 797 struct iscsi_hdr *hdr; 798 799 /* Check if there are additional header segments 800 * *prior* to computing the digest, because we 801 * may need to go back to the caller for more. 802 */ 803 hdr = (struct iscsi_hdr *) tcp_conn->in.hdr_buf; 804 if (segment->copied == sizeof(struct iscsi_hdr) && hdr->hlength) { 805 /* Bump the header length - the caller will 806 * just loop around and get the AHS for us, and 807 * call again. */ 808 unsigned int ahslen = hdr->hlength << 2; 809 810 /* Make sure we don't overflow */ 811 if (sizeof(*hdr) + ahslen > sizeof(tcp_conn->in.hdr_buf)) 812 return ISCSI_ERR_AHSLEN; 813 814 segment->total_size += ahslen; 815 segment->size += ahslen; 816 return 0; 817 } 818 819 /* We're done processing the header. See if we're doing 820 * header digests; if so, set up the recv_digest buffer 821 * and go back for more. */ 822 if (conn->hdrdgst_en && 823 !(conn->session->tt->caps & CAP_DIGEST_OFFLOAD)) { 824 if (segment->digest_len == 0) { 825 /* 826 * Even if we offload the digest processing we 827 * splice it in so we can increment the skb/segment 828 * counters in preparation for the data segment. 829 */ 830 iscsi_tcp_segment_splice_digest(segment, 831 segment->recv_digest); 832 return 0; 833 } 834 835 iscsi_tcp_dgst_header(tcp_conn->rx_hash, hdr, 836 segment->total_copied - ISCSI_DIGEST_SIZE, 837 segment->digest); 838 839 if (!iscsi_tcp_dgst_verify(tcp_conn, segment)) 840 return ISCSI_ERR_HDR_DGST; 841 } 842 843 tcp_conn->in.hdr = hdr; 844 return iscsi_tcp_hdr_dissect(conn, hdr); 845 } 846 847 /** 848 * iscsi_tcp_recv_segment_is_hdr - tests if we are reading in a header 849 * @tcp_conn: iscsi tcp conn 850 * 851 * returns non zero if we are currently processing or setup to process 852 * a header. 853 */ 854 inline int iscsi_tcp_recv_segment_is_hdr(struct iscsi_tcp_conn *tcp_conn) 855 { 856 return tcp_conn->in.segment.done == iscsi_tcp_hdr_recv_done; 857 } 858 EXPORT_SYMBOL_GPL(iscsi_tcp_recv_segment_is_hdr); 859 860 /** 861 * iscsi_tcp_recv_skb - Process skb 862 * @conn: iscsi connection 863 * @skb: network buffer with header and/or data segment 864 * @offset: offset in skb 865 * @offload: bool indicating if transfer was offloaded 866 * 867 * Will return status of transfer in status. And will return 868 * number of bytes copied. 869 */ 870 int iscsi_tcp_recv_skb(struct iscsi_conn *conn, struct sk_buff *skb, 871 unsigned int offset, bool offloaded, int *status) 872 { 873 struct iscsi_tcp_conn *tcp_conn = conn->dd_data; 874 struct iscsi_segment *segment = &tcp_conn->in.segment; 875 struct skb_seq_state seq; 876 unsigned int consumed = 0; 877 int rc = 0; 878 879 ISCSI_DBG_TCP(conn, "in %d bytes\n", skb->len - offset); 880 /* 881 * Update for each skb instead of pdu, because over slow networks a 882 * data_in's data could take a while to read in. We also want to 883 * account for r2ts. 884 */ 885 conn->last_recv = jiffies; 886 887 if (unlikely(conn->suspend_rx)) { 888 ISCSI_DBG_TCP(conn, "Rx suspended!\n"); 889 *status = ISCSI_TCP_SUSPENDED; 890 return 0; 891 } 892 893 if (offloaded) { 894 segment->total_copied = segment->total_size; 895 goto segment_done; 896 } 897 898 skb_prepare_seq_read(skb, offset, skb->len, &seq); 899 while (1) { 900 unsigned int avail; 901 const u8 *ptr; 902 903 avail = skb_seq_read(consumed, &ptr, &seq); 904 if (avail == 0) { 905 ISCSI_DBG_TCP(conn, "no more data avail. Consumed %d\n", 906 consumed); 907 *status = ISCSI_TCP_SKB_DONE; 908 skb_abort_seq_read(&seq); 909 goto skb_done; 910 } 911 BUG_ON(segment->copied >= segment->size); 912 913 ISCSI_DBG_TCP(conn, "skb %p ptr=%p avail=%u\n", skb, ptr, 914 avail); 915 rc = iscsi_tcp_segment_recv(tcp_conn, segment, ptr, avail); 916 BUG_ON(rc == 0); 917 consumed += rc; 918 919 if (segment->total_copied >= segment->total_size) { 920 skb_abort_seq_read(&seq); 921 goto segment_done; 922 } 923 } 924 925 segment_done: 926 *status = ISCSI_TCP_SEGMENT_DONE; 927 ISCSI_DBG_TCP(conn, "segment done\n"); 928 rc = segment->done(tcp_conn, segment); 929 if (rc != 0) { 930 *status = ISCSI_TCP_CONN_ERR; 931 ISCSI_DBG_TCP(conn, "Error receiving PDU, errno=%d\n", rc); 932 iscsi_conn_failure(conn, rc); 933 return 0; 934 } 935 /* The done() functions sets up the next segment. */ 936 937 skb_done: 938 conn->rxdata_octets += consumed; 939 return consumed; 940 } 941 EXPORT_SYMBOL_GPL(iscsi_tcp_recv_skb); 942 943 /** 944 * iscsi_tcp_task_init - Initialize iSCSI SCSI_READ or SCSI_WRITE commands 945 * @conn: iscsi connection 946 * @task: scsi command task 947 * @sc: scsi command 948 */ 949 int iscsi_tcp_task_init(struct iscsi_task *task) 950 { 951 struct iscsi_tcp_task *tcp_task = task->dd_data; 952 struct iscsi_conn *conn = task->conn; 953 struct scsi_cmnd *sc = task->sc; 954 int err; 955 956 if (!sc) { 957 /* 958 * mgmt tasks do not have a scatterlist since they come 959 * in from the iscsi interface. 960 */ 961 ISCSI_DBG_TCP(conn, "mtask deq [itt 0x%x]\n", task->itt); 962 963 return conn->session->tt->init_pdu(task, 0, task->data_count); 964 } 965 966 BUG_ON(kfifo_len(&tcp_task->r2tqueue)); 967 tcp_task->exp_datasn = 0; 968 969 /* Prepare PDU, optionally w/ immediate data */ 970 ISCSI_DBG_TCP(conn, "task deq [itt 0x%x imm %d unsol %d]\n", 971 task->itt, task->imm_count, task->unsol_r2t.data_length); 972 973 err = conn->session->tt->init_pdu(task, 0, task->imm_count); 974 if (err) 975 return err; 976 task->imm_count = 0; 977 return 0; 978 } 979 EXPORT_SYMBOL_GPL(iscsi_tcp_task_init); 980 981 static struct iscsi_r2t_info *iscsi_tcp_get_curr_r2t(struct iscsi_task *task) 982 { 983 struct iscsi_session *session = task->conn->session; 984 struct iscsi_tcp_task *tcp_task = task->dd_data; 985 struct iscsi_r2t_info *r2t = NULL; 986 987 if (iscsi_task_has_unsol_data(task)) 988 r2t = &task->unsol_r2t; 989 else { 990 spin_lock_bh(&session->lock); 991 if (tcp_task->r2t) { 992 r2t = tcp_task->r2t; 993 /* Continue with this R2T? */ 994 if (r2t->data_length <= r2t->sent) { 995 ISCSI_DBG_TCP(task->conn, 996 " done with r2t %p\n", r2t); 997 kfifo_in(&tcp_task->r2tpool.queue, 998 (void *)&tcp_task->r2t, 999 sizeof(void *)); 1000 tcp_task->r2t = r2t = NULL; 1001 } 1002 } 1003 1004 if (r2t == NULL) { 1005 if (kfifo_out(&tcp_task->r2tqueue, 1006 (void *)&tcp_task->r2t, sizeof(void *)) != 1007 sizeof(void *)) 1008 r2t = NULL; 1009 else 1010 r2t = tcp_task->r2t; 1011 } 1012 spin_unlock_bh(&session->lock); 1013 } 1014 1015 return r2t; 1016 } 1017 1018 /** 1019 * iscsi_tcp_task_xmit - xmit normal PDU task 1020 * @task: iscsi command task 1021 * 1022 * We're expected to return 0 when everything was transmitted successfully, 1023 * -EAGAIN if there's still data in the queue, or != 0 for any other kind 1024 * of error. 1025 */ 1026 int iscsi_tcp_task_xmit(struct iscsi_task *task) 1027 { 1028 struct iscsi_conn *conn = task->conn; 1029 struct iscsi_session *session = conn->session; 1030 struct iscsi_r2t_info *r2t; 1031 int rc = 0; 1032 1033 flush: 1034 /* Flush any pending data first. */ 1035 rc = session->tt->xmit_pdu(task); 1036 if (rc < 0) 1037 return rc; 1038 1039 /* mgmt command */ 1040 if (!task->sc) { 1041 if (task->hdr->itt == RESERVED_ITT) 1042 iscsi_put_task(task); 1043 return 0; 1044 } 1045 1046 /* Are we done already? */ 1047 if (task->sc->sc_data_direction != DMA_TO_DEVICE) 1048 return 0; 1049 1050 r2t = iscsi_tcp_get_curr_r2t(task); 1051 if (r2t == NULL) { 1052 /* Waiting for more R2Ts to arrive. */ 1053 ISCSI_DBG_TCP(conn, "no R2Ts yet\n"); 1054 return 0; 1055 } 1056 1057 rc = conn->session->tt->alloc_pdu(task, ISCSI_OP_SCSI_DATA_OUT); 1058 if (rc) 1059 return rc; 1060 iscsi_prep_data_out_pdu(task, r2t, (struct iscsi_data *) task->hdr); 1061 1062 ISCSI_DBG_TCP(conn, "sol dout %p [dsn %d itt 0x%x doff %d dlen %d]\n", 1063 r2t, r2t->datasn - 1, task->hdr->itt, 1064 r2t->data_offset + r2t->sent, r2t->data_count); 1065 1066 rc = conn->session->tt->init_pdu(task, r2t->data_offset + r2t->sent, 1067 r2t->data_count); 1068 if (rc) { 1069 iscsi_conn_failure(conn, ISCSI_ERR_XMIT_FAILED); 1070 return rc; 1071 } 1072 1073 r2t->sent += r2t->data_count; 1074 goto flush; 1075 } 1076 EXPORT_SYMBOL_GPL(iscsi_tcp_task_xmit); 1077 1078 struct iscsi_cls_conn * 1079 iscsi_tcp_conn_setup(struct iscsi_cls_session *cls_session, int dd_data_size, 1080 uint32_t conn_idx) 1081 1082 { 1083 struct iscsi_conn *conn; 1084 struct iscsi_cls_conn *cls_conn; 1085 struct iscsi_tcp_conn *tcp_conn; 1086 1087 cls_conn = iscsi_conn_setup(cls_session, 1088 sizeof(*tcp_conn) + dd_data_size, conn_idx); 1089 if (!cls_conn) 1090 return NULL; 1091 conn = cls_conn->dd_data; 1092 /* 1093 * due to strange issues with iser these are not set 1094 * in iscsi_conn_setup 1095 */ 1096 conn->max_recv_dlength = ISCSI_DEF_MAX_RECV_SEG_LEN; 1097 1098 tcp_conn = conn->dd_data; 1099 tcp_conn->iscsi_conn = conn; 1100 tcp_conn->dd_data = conn->dd_data + sizeof(*tcp_conn); 1101 return cls_conn; 1102 } 1103 EXPORT_SYMBOL_GPL(iscsi_tcp_conn_setup); 1104 1105 void iscsi_tcp_conn_teardown(struct iscsi_cls_conn *cls_conn) 1106 { 1107 iscsi_conn_teardown(cls_conn); 1108 } 1109 EXPORT_SYMBOL_GPL(iscsi_tcp_conn_teardown); 1110 1111 int iscsi_tcp_r2tpool_alloc(struct iscsi_session *session) 1112 { 1113 int i; 1114 int cmd_i; 1115 1116 /* 1117 * initialize per-task: R2T pool and xmit queue 1118 */ 1119 for (cmd_i = 0; cmd_i < session->cmds_max; cmd_i++) { 1120 struct iscsi_task *task = session->cmds[cmd_i]; 1121 struct iscsi_tcp_task *tcp_task = task->dd_data; 1122 1123 /* 1124 * pre-allocated x2 as much r2ts to handle race when 1125 * target acks DataOut faster than we data_xmit() queues 1126 * could replenish r2tqueue. 1127 */ 1128 1129 /* R2T pool */ 1130 if (iscsi_pool_init(&tcp_task->r2tpool, 1131 session->max_r2t * 2, NULL, 1132 sizeof(struct iscsi_r2t_info))) { 1133 goto r2t_alloc_fail; 1134 } 1135 1136 /* R2T xmit queue */ 1137 if (kfifo_alloc(&tcp_task->r2tqueue, 1138 session->max_r2t * 4 * sizeof(void*), GFP_KERNEL)) { 1139 iscsi_pool_free(&tcp_task->r2tpool); 1140 goto r2t_alloc_fail; 1141 } 1142 } 1143 1144 return 0; 1145 1146 r2t_alloc_fail: 1147 for (i = 0; i < cmd_i; i++) { 1148 struct iscsi_task *task = session->cmds[i]; 1149 struct iscsi_tcp_task *tcp_task = task->dd_data; 1150 1151 kfifo_free(&tcp_task->r2tqueue); 1152 iscsi_pool_free(&tcp_task->r2tpool); 1153 } 1154 return -ENOMEM; 1155 } 1156 EXPORT_SYMBOL_GPL(iscsi_tcp_r2tpool_alloc); 1157 1158 void iscsi_tcp_r2tpool_free(struct iscsi_session *session) 1159 { 1160 int i; 1161 1162 for (i = 0; i < session->cmds_max; i++) { 1163 struct iscsi_task *task = session->cmds[i]; 1164 struct iscsi_tcp_task *tcp_task = task->dd_data; 1165 1166 kfifo_free(&tcp_task->r2tqueue); 1167 iscsi_pool_free(&tcp_task->r2tpool); 1168 } 1169 } 1170 EXPORT_SYMBOL_GPL(iscsi_tcp_r2tpool_free); 1171 1172 void iscsi_tcp_conn_get_stats(struct iscsi_cls_conn *cls_conn, 1173 struct iscsi_stats *stats) 1174 { 1175 struct iscsi_conn *conn = cls_conn->dd_data; 1176 1177 stats->txdata_octets = conn->txdata_octets; 1178 stats->rxdata_octets = conn->rxdata_octets; 1179 stats->scsicmd_pdus = conn->scsicmd_pdus_cnt; 1180 stats->dataout_pdus = conn->dataout_pdus_cnt; 1181 stats->scsirsp_pdus = conn->scsirsp_pdus_cnt; 1182 stats->datain_pdus = conn->datain_pdus_cnt; 1183 stats->r2t_pdus = conn->r2t_pdus_cnt; 1184 stats->tmfcmd_pdus = conn->tmfcmd_pdus_cnt; 1185 stats->tmfrsp_pdus = conn->tmfrsp_pdus_cnt; 1186 } 1187 EXPORT_SYMBOL_GPL(iscsi_tcp_conn_get_stats); 1188