1 /* 2 * bnx2i_iscsi.c: QLogic NetXtreme II iSCSI driver. 3 * 4 * Copyright (c) 2006 - 2013 Broadcom Corporation 5 * Copyright (c) 2007, 2008 Red Hat, Inc. All rights reserved. 6 * Copyright (c) 2007, 2008 Mike Christie 7 * Copyright (c) 2014, QLogic Corporation 8 * 9 * This program is free software; you can redistribute it and/or modify 10 * it under the terms of the GNU General Public License as published by 11 * the Free Software Foundation. 12 * 13 * Written by: Anil Veerabhadrappa (anilgv@broadcom.com) 14 * Previously Maintained by: Eddie Wai (eddie.wai@broadcom.com) 15 * Maintained by: QLogic-Storage-Upstream@qlogic.com 16 */ 17 18 #include <linux/slab.h> 19 #include <scsi/scsi_tcq.h> 20 #include <scsi/libiscsi.h> 21 #include "bnx2i.h" 22 23 struct scsi_transport_template *bnx2i_scsi_xport_template; 24 struct iscsi_transport bnx2i_iscsi_transport; 25 static struct scsi_host_template bnx2i_host_template; 26 27 /* 28 * Global endpoint resource info 29 */ 30 static DEFINE_SPINLOCK(bnx2i_resc_lock); /* protects global resources */ 31 32 DECLARE_PER_CPU(struct bnx2i_percpu_s, bnx2i_percpu); 33 34 static int bnx2i_adapter_ready(struct bnx2i_hba *hba) 35 { 36 int retval = 0; 37 38 if (!hba || !test_bit(ADAPTER_STATE_UP, &hba->adapter_state) || 39 test_bit(ADAPTER_STATE_GOING_DOWN, &hba->adapter_state) || 40 test_bit(ADAPTER_STATE_LINK_DOWN, &hba->adapter_state)) 41 retval = -EPERM; 42 return retval; 43 } 44 45 /** 46 * bnx2i_get_write_cmd_bd_idx - identifies various BD bookmarks 47 * @cmd: iscsi cmd struct pointer 48 * @buf_off: absolute buffer offset 49 * @start_bd_off: u32 pointer to return the offset within the BD 50 * indicated by 'start_bd_idx' on which 'buf_off' falls 51 * @start_bd_idx: index of the BD on which 'buf_off' falls 52 * 53 * identifies & marks various bd info for scsi command's imm data, 54 * unsolicited data and the first solicited data seq. 55 */ 56 static void bnx2i_get_write_cmd_bd_idx(struct bnx2i_cmd *cmd, u32 buf_off, 57 u32 *start_bd_off, u32 *start_bd_idx) 58 { 59 struct iscsi_bd *bd_tbl = cmd->io_tbl.bd_tbl; 60 u32 cur_offset = 0; 61 u32 cur_bd_idx = 0; 62 63 if (buf_off) { 64 while (buf_off >= (cur_offset + bd_tbl->buffer_length)) { 65 cur_offset += bd_tbl->buffer_length; 66 cur_bd_idx++; 67 bd_tbl++; 68 } 69 } 70 71 *start_bd_off = buf_off - cur_offset; 72 *start_bd_idx = cur_bd_idx; 73 } 74 75 /** 76 * bnx2i_setup_write_cmd_bd_info - sets up BD various information 77 * @task: transport layer's cmd struct pointer 78 * 79 * identifies & marks various bd info for scsi command's immediate data, 80 * unsolicited data and first solicited data seq which includes BD start 81 * index & BD buf off. his function takes into account iscsi parameter such 82 * as immediate data and unsolicited data is support on this connection. 83 */ 84 static void bnx2i_setup_write_cmd_bd_info(struct iscsi_task *task) 85 { 86 struct bnx2i_cmd *cmd = task->dd_data; 87 u32 start_bd_offset; 88 u32 start_bd_idx; 89 u32 buffer_offset = 0; 90 u32 cmd_len = cmd->req.total_data_transfer_length; 91 92 /* if ImmediateData is turned off & IntialR2T is turned on, 93 * there will be no immediate or unsolicited data, just return. 94 */ 95 if (!iscsi_task_has_unsol_data(task) && !task->imm_count) 96 return; 97 98 /* Immediate data */ 99 buffer_offset += task->imm_count; 100 if (task->imm_count == cmd_len) 101 return; 102 103 if (iscsi_task_has_unsol_data(task)) { 104 bnx2i_get_write_cmd_bd_idx(cmd, buffer_offset, 105 &start_bd_offset, &start_bd_idx); 106 cmd->req.ud_buffer_offset = start_bd_offset; 107 cmd->req.ud_start_bd_index = start_bd_idx; 108 buffer_offset += task->unsol_r2t.data_length; 109 } 110 111 if (buffer_offset != cmd_len) { 112 bnx2i_get_write_cmd_bd_idx(cmd, buffer_offset, 113 &start_bd_offset, &start_bd_idx); 114 if ((start_bd_offset > task->conn->session->first_burst) || 115 (start_bd_idx > scsi_sg_count(cmd->scsi_cmd))) { 116 int i = 0; 117 118 iscsi_conn_printk(KERN_ALERT, task->conn, 119 "bnx2i- error, buf offset 0x%x " 120 "bd_valid %d use_sg %d\n", 121 buffer_offset, cmd->io_tbl.bd_valid, 122 scsi_sg_count(cmd->scsi_cmd)); 123 for (i = 0; i < cmd->io_tbl.bd_valid; i++) 124 iscsi_conn_printk(KERN_ALERT, task->conn, 125 "bnx2i err, bd[%d]: len %x\n", 126 i, cmd->io_tbl.bd_tbl[i].\ 127 buffer_length); 128 } 129 cmd->req.sd_buffer_offset = start_bd_offset; 130 cmd->req.sd_start_bd_index = start_bd_idx; 131 } 132 } 133 134 135 136 /** 137 * bnx2i_map_scsi_sg - maps IO buffer and prepares the BD table 138 * @hba: adapter instance 139 * @cmd: iscsi cmd struct pointer 140 * 141 * map SG list 142 */ 143 static int bnx2i_map_scsi_sg(struct bnx2i_hba *hba, struct bnx2i_cmd *cmd) 144 { 145 struct scsi_cmnd *sc = cmd->scsi_cmd; 146 struct iscsi_bd *bd = cmd->io_tbl.bd_tbl; 147 struct scatterlist *sg; 148 int byte_count = 0; 149 int bd_count = 0; 150 int sg_count; 151 int sg_len; 152 u64 addr; 153 int i; 154 155 BUG_ON(scsi_sg_count(sc) > ISCSI_MAX_BDS_PER_CMD); 156 157 sg_count = scsi_dma_map(sc); 158 159 scsi_for_each_sg(sc, sg, sg_count, i) { 160 sg_len = sg_dma_len(sg); 161 addr = (u64) sg_dma_address(sg); 162 bd[bd_count].buffer_addr_lo = addr & 0xffffffff; 163 bd[bd_count].buffer_addr_hi = addr >> 32; 164 bd[bd_count].buffer_length = sg_len; 165 bd[bd_count].flags = 0; 166 if (bd_count == 0) 167 bd[bd_count].flags = ISCSI_BD_FIRST_IN_BD_CHAIN; 168 169 byte_count += sg_len; 170 bd_count++; 171 } 172 173 if (bd_count) 174 bd[bd_count - 1].flags |= ISCSI_BD_LAST_IN_BD_CHAIN; 175 176 BUG_ON(byte_count != scsi_bufflen(sc)); 177 return bd_count; 178 } 179 180 /** 181 * bnx2i_iscsi_map_sg_list - maps SG list 182 * @cmd: iscsi cmd struct pointer 183 * 184 * creates BD list table for the command 185 */ 186 static void bnx2i_iscsi_map_sg_list(struct bnx2i_cmd *cmd) 187 { 188 int bd_count; 189 190 bd_count = bnx2i_map_scsi_sg(cmd->conn->hba, cmd); 191 if (!bd_count) { 192 struct iscsi_bd *bd = cmd->io_tbl.bd_tbl; 193 194 bd[0].buffer_addr_lo = bd[0].buffer_addr_hi = 0; 195 bd[0].buffer_length = bd[0].flags = 0; 196 } 197 cmd->io_tbl.bd_valid = bd_count; 198 } 199 200 201 /** 202 * bnx2i_iscsi_unmap_sg_list - unmaps SG list 203 * @cmd: iscsi cmd struct pointer 204 * 205 * unmap IO buffers and invalidate the BD table 206 */ 207 void bnx2i_iscsi_unmap_sg_list(struct bnx2i_cmd *cmd) 208 { 209 struct scsi_cmnd *sc = cmd->scsi_cmd; 210 211 if (cmd->io_tbl.bd_valid && sc) { 212 scsi_dma_unmap(sc); 213 cmd->io_tbl.bd_valid = 0; 214 } 215 } 216 217 static void bnx2i_setup_cmd_wqe_template(struct bnx2i_cmd *cmd) 218 { 219 memset(&cmd->req, 0x00, sizeof(cmd->req)); 220 cmd->req.op_code = 0xFF; 221 cmd->req.bd_list_addr_lo = (u32) cmd->io_tbl.bd_tbl_dma; 222 cmd->req.bd_list_addr_hi = 223 (u32) ((u64) cmd->io_tbl.bd_tbl_dma >> 32); 224 225 } 226 227 228 /** 229 * bnx2i_bind_conn_to_iscsi_cid - bind conn structure to 'iscsi_cid' 230 * @hba: pointer to adapter instance 231 * @bnx2i_conn: pointer to iscsi connection 232 * @iscsi_cid: iscsi context ID, range 0 - (MAX_CONN - 1) 233 * 234 * update iscsi cid table entry with connection pointer. This enables 235 * driver to quickly get hold of connection structure pointer in 236 * completion/interrupt thread using iscsi context ID 237 */ 238 static int bnx2i_bind_conn_to_iscsi_cid(struct bnx2i_hba *hba, 239 struct bnx2i_conn *bnx2i_conn, 240 u32 iscsi_cid) 241 { 242 if (hba && hba->cid_que.conn_cid_tbl[iscsi_cid]) { 243 iscsi_conn_printk(KERN_ALERT, bnx2i_conn->cls_conn->dd_data, 244 "conn bind - entry #%d not free\n", iscsi_cid); 245 return -EBUSY; 246 } 247 248 hba->cid_que.conn_cid_tbl[iscsi_cid] = bnx2i_conn; 249 return 0; 250 } 251 252 253 /** 254 * bnx2i_get_conn_from_id - maps an iscsi cid to corresponding conn ptr 255 * @hba: pointer to adapter instance 256 * @iscsi_cid: iscsi context ID, range 0 - (MAX_CONN - 1) 257 */ 258 struct bnx2i_conn *bnx2i_get_conn_from_id(struct bnx2i_hba *hba, 259 u16 iscsi_cid) 260 { 261 if (!hba->cid_que.conn_cid_tbl) { 262 printk(KERN_ERR "bnx2i: ERROR - missing conn<->cid table\n"); 263 return NULL; 264 265 } else if (iscsi_cid >= hba->max_active_conns) { 266 printk(KERN_ERR "bnx2i: wrong cid #%d\n", iscsi_cid); 267 return NULL; 268 } 269 return hba->cid_que.conn_cid_tbl[iscsi_cid]; 270 } 271 272 273 /** 274 * bnx2i_alloc_iscsi_cid - allocates a iscsi_cid from free pool 275 * @hba: pointer to adapter instance 276 */ 277 static u32 bnx2i_alloc_iscsi_cid(struct bnx2i_hba *hba) 278 { 279 int idx; 280 281 if (!hba->cid_que.cid_free_cnt) 282 return -1; 283 284 idx = hba->cid_que.cid_q_cons_idx; 285 hba->cid_que.cid_q_cons_idx++; 286 if (hba->cid_que.cid_q_cons_idx == hba->cid_que.cid_q_max_idx) 287 hba->cid_que.cid_q_cons_idx = 0; 288 289 hba->cid_que.cid_free_cnt--; 290 return hba->cid_que.cid_que[idx]; 291 } 292 293 294 /** 295 * bnx2i_free_iscsi_cid - returns tcp port to free list 296 * @hba: pointer to adapter instance 297 * @iscsi_cid: iscsi context ID to free 298 */ 299 static void bnx2i_free_iscsi_cid(struct bnx2i_hba *hba, u16 iscsi_cid) 300 { 301 int idx; 302 303 if (iscsi_cid == (u16) -1) 304 return; 305 306 hba->cid_que.cid_free_cnt++; 307 308 idx = hba->cid_que.cid_q_prod_idx; 309 hba->cid_que.cid_que[idx] = iscsi_cid; 310 hba->cid_que.conn_cid_tbl[iscsi_cid] = NULL; 311 hba->cid_que.cid_q_prod_idx++; 312 if (hba->cid_que.cid_q_prod_idx == hba->cid_que.cid_q_max_idx) 313 hba->cid_que.cid_q_prod_idx = 0; 314 } 315 316 317 /** 318 * bnx2i_setup_free_cid_que - sets up free iscsi cid queue 319 * @hba: pointer to adapter instance 320 * 321 * allocates memory for iscsi cid queue & 'cid - conn ptr' mapping table, 322 * and initialize table attributes 323 */ 324 static int bnx2i_setup_free_cid_que(struct bnx2i_hba *hba) 325 { 326 int mem_size; 327 int i; 328 329 mem_size = hba->max_active_conns * sizeof(u32); 330 mem_size = (mem_size + (PAGE_SIZE - 1)) & PAGE_MASK; 331 332 hba->cid_que.cid_que_base = kmalloc(mem_size, GFP_KERNEL); 333 if (!hba->cid_que.cid_que_base) 334 return -ENOMEM; 335 336 mem_size = hba->max_active_conns * sizeof(struct bnx2i_conn *); 337 mem_size = (mem_size + (PAGE_SIZE - 1)) & PAGE_MASK; 338 hba->cid_que.conn_cid_tbl = kmalloc(mem_size, GFP_KERNEL); 339 if (!hba->cid_que.conn_cid_tbl) { 340 kfree(hba->cid_que.cid_que_base); 341 hba->cid_que.cid_que_base = NULL; 342 return -ENOMEM; 343 } 344 345 hba->cid_que.cid_que = (u32 *)hba->cid_que.cid_que_base; 346 hba->cid_que.cid_q_prod_idx = 0; 347 hba->cid_que.cid_q_cons_idx = 0; 348 hba->cid_que.cid_q_max_idx = hba->max_active_conns; 349 hba->cid_que.cid_free_cnt = hba->max_active_conns; 350 351 for (i = 0; i < hba->max_active_conns; i++) { 352 hba->cid_que.cid_que[i] = i; 353 hba->cid_que.conn_cid_tbl[i] = NULL; 354 } 355 return 0; 356 } 357 358 359 /** 360 * bnx2i_release_free_cid_que - releases 'iscsi_cid' queue resources 361 * @hba: pointer to adapter instance 362 */ 363 static void bnx2i_release_free_cid_que(struct bnx2i_hba *hba) 364 { 365 kfree(hba->cid_que.cid_que_base); 366 hba->cid_que.cid_que_base = NULL; 367 368 kfree(hba->cid_que.conn_cid_tbl); 369 hba->cid_que.conn_cid_tbl = NULL; 370 } 371 372 373 /** 374 * bnx2i_alloc_ep - allocates ep structure from global pool 375 * @hba: pointer to adapter instance 376 * 377 * routine allocates a free endpoint structure from global pool and 378 * a tcp port to be used for this connection. Global resource lock, 379 * 'bnx2i_resc_lock' is held while accessing shared global data structures 380 */ 381 static struct iscsi_endpoint *bnx2i_alloc_ep(struct bnx2i_hba *hba) 382 { 383 struct iscsi_endpoint *ep; 384 struct bnx2i_endpoint *bnx2i_ep; 385 u32 ec_div; 386 387 ep = iscsi_create_endpoint(sizeof(*bnx2i_ep)); 388 if (!ep) { 389 printk(KERN_ERR "bnx2i: Could not allocate ep\n"); 390 return NULL; 391 } 392 393 bnx2i_ep = ep->dd_data; 394 bnx2i_ep->cls_ep = ep; 395 INIT_LIST_HEAD(&bnx2i_ep->link); 396 bnx2i_ep->state = EP_STATE_IDLE; 397 bnx2i_ep->ep_iscsi_cid = (u16) -1; 398 bnx2i_ep->hba = hba; 399 bnx2i_ep->hba_age = hba->age; 400 401 ec_div = event_coal_div; 402 while (ec_div >>= 1) 403 bnx2i_ep->ec_shift += 1; 404 405 hba->ofld_conns_active++; 406 init_waitqueue_head(&bnx2i_ep->ofld_wait); 407 return ep; 408 } 409 410 411 /** 412 * bnx2i_free_ep - free endpoint 413 * @ep: pointer to iscsi endpoint structure 414 */ 415 static void bnx2i_free_ep(struct iscsi_endpoint *ep) 416 { 417 struct bnx2i_endpoint *bnx2i_ep = ep->dd_data; 418 unsigned long flags; 419 420 spin_lock_irqsave(&bnx2i_resc_lock, flags); 421 bnx2i_ep->state = EP_STATE_IDLE; 422 bnx2i_ep->hba->ofld_conns_active--; 423 424 if (bnx2i_ep->ep_iscsi_cid != (u16) -1) 425 bnx2i_free_iscsi_cid(bnx2i_ep->hba, bnx2i_ep->ep_iscsi_cid); 426 427 if (bnx2i_ep->conn) { 428 bnx2i_ep->conn->ep = NULL; 429 bnx2i_ep->conn = NULL; 430 } 431 432 bnx2i_ep->hba = NULL; 433 spin_unlock_irqrestore(&bnx2i_resc_lock, flags); 434 iscsi_destroy_endpoint(ep); 435 } 436 437 438 /** 439 * bnx2i_alloc_bdt - allocates buffer descriptor (BD) table for the command 440 * @hba: adapter instance pointer 441 * @session: iscsi session pointer 442 * @cmd: iscsi command structure 443 */ 444 static int bnx2i_alloc_bdt(struct bnx2i_hba *hba, struct iscsi_session *session, 445 struct bnx2i_cmd *cmd) 446 { 447 struct io_bdt *io = &cmd->io_tbl; 448 struct iscsi_bd *bd; 449 450 io->bd_tbl = dma_alloc_coherent(&hba->pcidev->dev, 451 ISCSI_MAX_BDS_PER_CMD * sizeof(*bd), 452 &io->bd_tbl_dma, GFP_KERNEL); 453 if (!io->bd_tbl) { 454 iscsi_session_printk(KERN_ERR, session, "Could not " 455 "allocate bdt.\n"); 456 return -ENOMEM; 457 } 458 io->bd_valid = 0; 459 return 0; 460 } 461 462 /** 463 * bnx2i_destroy_cmd_pool - destroys iscsi command pool and release BD table 464 * @hba: adapter instance pointer 465 * @session: iscsi session pointer 466 */ 467 static void bnx2i_destroy_cmd_pool(struct bnx2i_hba *hba, 468 struct iscsi_session *session) 469 { 470 int i; 471 472 for (i = 0; i < session->cmds_max; i++) { 473 struct iscsi_task *task = session->cmds[i]; 474 struct bnx2i_cmd *cmd = task->dd_data; 475 476 if (cmd->io_tbl.bd_tbl) 477 dma_free_coherent(&hba->pcidev->dev, 478 ISCSI_MAX_BDS_PER_CMD * 479 sizeof(struct iscsi_bd), 480 cmd->io_tbl.bd_tbl, 481 cmd->io_tbl.bd_tbl_dma); 482 } 483 484 } 485 486 487 /** 488 * bnx2i_setup_cmd_pool - sets up iscsi command pool for the session 489 * @hba: adapter instance pointer 490 * @session: iscsi session pointer 491 */ 492 static int bnx2i_setup_cmd_pool(struct bnx2i_hba *hba, 493 struct iscsi_session *session) 494 { 495 int i; 496 497 for (i = 0; i < session->cmds_max; i++) { 498 struct iscsi_task *task = session->cmds[i]; 499 struct bnx2i_cmd *cmd = task->dd_data; 500 501 task->hdr = &cmd->hdr; 502 task->hdr_max = sizeof(struct iscsi_hdr); 503 504 if (bnx2i_alloc_bdt(hba, session, cmd)) 505 goto free_bdts; 506 } 507 508 return 0; 509 510 free_bdts: 511 bnx2i_destroy_cmd_pool(hba, session); 512 return -ENOMEM; 513 } 514 515 516 /** 517 * bnx2i_setup_mp_bdt - allocate BD table resources 518 * @hba: pointer to adapter structure 519 * 520 * Allocate memory for dummy buffer and associated BD 521 * table to be used by middle path (MP) requests 522 */ 523 static int bnx2i_setup_mp_bdt(struct bnx2i_hba *hba) 524 { 525 int rc = 0; 526 struct iscsi_bd *mp_bdt; 527 u64 addr; 528 529 hba->mp_bd_tbl = dma_alloc_coherent(&hba->pcidev->dev, CNIC_PAGE_SIZE, 530 &hba->mp_bd_dma, GFP_KERNEL); 531 if (!hba->mp_bd_tbl) { 532 printk(KERN_ERR "unable to allocate Middle Path BDT\n"); 533 rc = -1; 534 goto out; 535 } 536 537 hba->dummy_buffer = dma_alloc_coherent(&hba->pcidev->dev, 538 CNIC_PAGE_SIZE, 539 &hba->dummy_buf_dma, GFP_KERNEL); 540 if (!hba->dummy_buffer) { 541 printk(KERN_ERR "unable to alloc Middle Path Dummy Buffer\n"); 542 dma_free_coherent(&hba->pcidev->dev, CNIC_PAGE_SIZE, 543 hba->mp_bd_tbl, hba->mp_bd_dma); 544 hba->mp_bd_tbl = NULL; 545 rc = -1; 546 goto out; 547 } 548 549 mp_bdt = (struct iscsi_bd *) hba->mp_bd_tbl; 550 addr = (unsigned long) hba->dummy_buf_dma; 551 mp_bdt->buffer_addr_lo = addr & 0xffffffff; 552 mp_bdt->buffer_addr_hi = addr >> 32; 553 mp_bdt->buffer_length = CNIC_PAGE_SIZE; 554 mp_bdt->flags = ISCSI_BD_LAST_IN_BD_CHAIN | 555 ISCSI_BD_FIRST_IN_BD_CHAIN; 556 out: 557 return rc; 558 } 559 560 561 /** 562 * bnx2i_free_mp_bdt - releases ITT back to free pool 563 * @hba: pointer to adapter instance 564 * 565 * free MP dummy buffer and associated BD table 566 */ 567 static void bnx2i_free_mp_bdt(struct bnx2i_hba *hba) 568 { 569 if (hba->mp_bd_tbl) { 570 dma_free_coherent(&hba->pcidev->dev, CNIC_PAGE_SIZE, 571 hba->mp_bd_tbl, hba->mp_bd_dma); 572 hba->mp_bd_tbl = NULL; 573 } 574 if (hba->dummy_buffer) { 575 dma_free_coherent(&hba->pcidev->dev, CNIC_PAGE_SIZE, 576 hba->dummy_buffer, hba->dummy_buf_dma); 577 hba->dummy_buffer = NULL; 578 } 579 return; 580 } 581 582 /** 583 * bnx2i_drop_session - notifies iscsid of connection error. 584 * @cls_session: iscsi cls session pointer 585 * 586 * This notifies iscsid that there is a error, so it can initiate 587 * recovery. 588 * 589 * This relies on caller using the iscsi class iterator so the object 590 * is refcounted and does not disapper from under us. 591 */ 592 void bnx2i_drop_session(struct iscsi_cls_session *cls_session) 593 { 594 iscsi_session_failure(cls_session->dd_data, ISCSI_ERR_CONN_FAILED); 595 } 596 597 /** 598 * bnx2i_ep_destroy_list_add - add an entry to EP destroy list 599 * @hba: pointer to adapter instance 600 * @ep: pointer to endpoint (transport identifier) structure 601 * 602 * EP destroy queue manager 603 */ 604 static int bnx2i_ep_destroy_list_add(struct bnx2i_hba *hba, 605 struct bnx2i_endpoint *ep) 606 { 607 write_lock_bh(&hba->ep_rdwr_lock); 608 list_add_tail(&ep->link, &hba->ep_destroy_list); 609 write_unlock_bh(&hba->ep_rdwr_lock); 610 return 0; 611 } 612 613 /** 614 * bnx2i_ep_destroy_list_del - add an entry to EP destroy list 615 * 616 * @hba: pointer to adapter instance 617 * @ep: pointer to endpoint (transport identifier) structure 618 * 619 * EP destroy queue manager 620 */ 621 static int bnx2i_ep_destroy_list_del(struct bnx2i_hba *hba, 622 struct bnx2i_endpoint *ep) 623 { 624 write_lock_bh(&hba->ep_rdwr_lock); 625 list_del_init(&ep->link); 626 write_unlock_bh(&hba->ep_rdwr_lock); 627 628 return 0; 629 } 630 631 /** 632 * bnx2i_ep_ofld_list_add - add an entry to ep offload pending list 633 * @hba: pointer to adapter instance 634 * @ep: pointer to endpoint (transport identifier) structure 635 * 636 * pending conn offload completion queue manager 637 */ 638 static int bnx2i_ep_ofld_list_add(struct bnx2i_hba *hba, 639 struct bnx2i_endpoint *ep) 640 { 641 write_lock_bh(&hba->ep_rdwr_lock); 642 list_add_tail(&ep->link, &hba->ep_ofld_list); 643 write_unlock_bh(&hba->ep_rdwr_lock); 644 return 0; 645 } 646 647 /** 648 * bnx2i_ep_ofld_list_del - add an entry to ep offload pending list 649 * @hba: pointer to adapter instance 650 * @ep: pointer to endpoint (transport identifier) structure 651 * 652 * pending conn offload completion queue manager 653 */ 654 static int bnx2i_ep_ofld_list_del(struct bnx2i_hba *hba, 655 struct bnx2i_endpoint *ep) 656 { 657 write_lock_bh(&hba->ep_rdwr_lock); 658 list_del_init(&ep->link); 659 write_unlock_bh(&hba->ep_rdwr_lock); 660 return 0; 661 } 662 663 664 /** 665 * bnx2i_find_ep_in_ofld_list - find iscsi_cid in pending list of endpoints 666 * 667 * @hba: pointer to adapter instance 668 * @iscsi_cid: iscsi context ID to find 669 * 670 */ 671 struct bnx2i_endpoint * 672 bnx2i_find_ep_in_ofld_list(struct bnx2i_hba *hba, u32 iscsi_cid) 673 { 674 struct list_head *list; 675 struct list_head *tmp; 676 struct bnx2i_endpoint *ep = NULL; 677 678 read_lock_bh(&hba->ep_rdwr_lock); 679 list_for_each_safe(list, tmp, &hba->ep_ofld_list) { 680 ep = (struct bnx2i_endpoint *)list; 681 682 if (ep->ep_iscsi_cid == iscsi_cid) 683 break; 684 ep = NULL; 685 } 686 read_unlock_bh(&hba->ep_rdwr_lock); 687 688 if (!ep) 689 printk(KERN_ERR "l5 cid %d not found\n", iscsi_cid); 690 return ep; 691 } 692 693 /** 694 * bnx2i_find_ep_in_destroy_list - find iscsi_cid in destroy list 695 * @hba: pointer to adapter instance 696 * @iscsi_cid: iscsi context ID to find 697 * 698 */ 699 struct bnx2i_endpoint * 700 bnx2i_find_ep_in_destroy_list(struct bnx2i_hba *hba, u32 iscsi_cid) 701 { 702 struct list_head *list; 703 struct list_head *tmp; 704 struct bnx2i_endpoint *ep = NULL; 705 706 read_lock_bh(&hba->ep_rdwr_lock); 707 list_for_each_safe(list, tmp, &hba->ep_destroy_list) { 708 ep = (struct bnx2i_endpoint *)list; 709 710 if (ep->ep_iscsi_cid == iscsi_cid) 711 break; 712 ep = NULL; 713 } 714 read_unlock_bh(&hba->ep_rdwr_lock); 715 716 if (!ep) 717 printk(KERN_ERR "l5 cid %d not found\n", iscsi_cid); 718 719 return ep; 720 } 721 722 /** 723 * bnx2i_ep_active_list_add - add an entry to ep active list 724 * @hba: pointer to adapter instance 725 * @ep: pointer to endpoint (transport identifier) structure 726 * 727 * current active conn queue manager 728 */ 729 static void bnx2i_ep_active_list_add(struct bnx2i_hba *hba, 730 struct bnx2i_endpoint *ep) 731 { 732 write_lock_bh(&hba->ep_rdwr_lock); 733 list_add_tail(&ep->link, &hba->ep_active_list); 734 write_unlock_bh(&hba->ep_rdwr_lock); 735 } 736 737 738 /** 739 * bnx2i_ep_active_list_del - deletes an entry to ep active list 740 * @hba: pointer to adapter instance 741 * @ep: pointer to endpoint (transport identifier) structure 742 * 743 * current active conn queue manager 744 */ 745 static void bnx2i_ep_active_list_del(struct bnx2i_hba *hba, 746 struct bnx2i_endpoint *ep) 747 { 748 write_lock_bh(&hba->ep_rdwr_lock); 749 list_del_init(&ep->link); 750 write_unlock_bh(&hba->ep_rdwr_lock); 751 } 752 753 754 /** 755 * bnx2i_setup_host_queue_size - assigns shost->can_queue param 756 * @hba: pointer to adapter instance 757 * @shost: scsi host pointer 758 * 759 * Initializes 'can_queue' parameter based on how many outstanding commands 760 * the device can handle. Each device 5708/5709/57710 has different 761 * capabilities 762 */ 763 static void bnx2i_setup_host_queue_size(struct bnx2i_hba *hba, 764 struct Scsi_Host *shost) 765 { 766 if (test_bit(BNX2I_NX2_DEV_5708, &hba->cnic_dev_type)) 767 shost->can_queue = ISCSI_MAX_CMDS_PER_HBA_5708; 768 else if (test_bit(BNX2I_NX2_DEV_5709, &hba->cnic_dev_type)) 769 shost->can_queue = ISCSI_MAX_CMDS_PER_HBA_5709; 770 else if (test_bit(BNX2I_NX2_DEV_57710, &hba->cnic_dev_type)) 771 shost->can_queue = ISCSI_MAX_CMDS_PER_HBA_57710; 772 else 773 shost->can_queue = ISCSI_MAX_CMDS_PER_HBA_5708; 774 } 775 776 777 /** 778 * bnx2i_alloc_hba - allocate and init adapter instance 779 * @cnic: cnic device pointer 780 * 781 * allocate & initialize adapter structure and call other 782 * support routines to do per adapter initialization 783 */ 784 struct bnx2i_hba *bnx2i_alloc_hba(struct cnic_dev *cnic) 785 { 786 struct Scsi_Host *shost; 787 struct bnx2i_hba *hba; 788 789 shost = iscsi_host_alloc(&bnx2i_host_template, sizeof(*hba), 0); 790 if (!shost) 791 return NULL; 792 shost->dma_boundary = cnic->pcidev->dma_mask; 793 shost->transportt = bnx2i_scsi_xport_template; 794 shost->max_id = ISCSI_MAX_CONNS_PER_HBA - 1; 795 shost->max_channel = 0; 796 shost->max_lun = 512; 797 shost->max_cmd_len = 16; 798 799 hba = iscsi_host_priv(shost); 800 hba->shost = shost; 801 hba->netdev = cnic->netdev; 802 /* Get PCI related information and update hba struct members */ 803 hba->pcidev = cnic->pcidev; 804 pci_dev_get(hba->pcidev); 805 hba->pci_did = hba->pcidev->device; 806 hba->pci_vid = hba->pcidev->vendor; 807 hba->pci_sdid = hba->pcidev->subsystem_device; 808 hba->pci_svid = hba->pcidev->subsystem_vendor; 809 hba->pci_func = PCI_FUNC(hba->pcidev->devfn); 810 hba->pci_devno = PCI_SLOT(hba->pcidev->devfn); 811 812 bnx2i_identify_device(hba, cnic); 813 bnx2i_setup_host_queue_size(hba, shost); 814 815 hba->reg_base = pci_resource_start(hba->pcidev, 0); 816 if (test_bit(BNX2I_NX2_DEV_5709, &hba->cnic_dev_type)) { 817 hba->regview = pci_iomap(hba->pcidev, 0, BNX2_MQ_CONFIG2); 818 if (!hba->regview) 819 goto ioreg_map_err; 820 } else if (test_bit(BNX2I_NX2_DEV_57710, &hba->cnic_dev_type)) { 821 hba->regview = pci_iomap(hba->pcidev, 0, 4096); 822 if (!hba->regview) 823 goto ioreg_map_err; 824 } 825 826 if (bnx2i_setup_mp_bdt(hba)) 827 goto mp_bdt_mem_err; 828 829 INIT_LIST_HEAD(&hba->ep_ofld_list); 830 INIT_LIST_HEAD(&hba->ep_active_list); 831 INIT_LIST_HEAD(&hba->ep_destroy_list); 832 rwlock_init(&hba->ep_rdwr_lock); 833 834 hba->mtu_supported = BNX2I_MAX_MTU_SUPPORTED; 835 836 /* different values for 5708/5709/57710 */ 837 hba->max_active_conns = ISCSI_MAX_CONNS_PER_HBA; 838 839 if (bnx2i_setup_free_cid_que(hba)) 840 goto cid_que_err; 841 842 /* SQ/RQ/CQ size can be changed via sysfx interface */ 843 if (test_bit(BNX2I_NX2_DEV_57710, &hba->cnic_dev_type)) { 844 if (sq_size && sq_size <= BNX2I_5770X_SQ_WQES_MAX) 845 hba->max_sqes = sq_size; 846 else 847 hba->max_sqes = BNX2I_5770X_SQ_WQES_DEFAULT; 848 } else { /* 5706/5708/5709 */ 849 if (sq_size && sq_size <= BNX2I_570X_SQ_WQES_MAX) 850 hba->max_sqes = sq_size; 851 else 852 hba->max_sqes = BNX2I_570X_SQ_WQES_DEFAULT; 853 } 854 855 hba->max_rqes = rq_size; 856 hba->max_cqes = hba->max_sqes + rq_size; 857 if (test_bit(BNX2I_NX2_DEV_57710, &hba->cnic_dev_type)) { 858 if (hba->max_cqes > BNX2I_5770X_CQ_WQES_MAX) 859 hba->max_cqes = BNX2I_5770X_CQ_WQES_MAX; 860 } else if (hba->max_cqes > BNX2I_570X_CQ_WQES_MAX) 861 hba->max_cqes = BNX2I_570X_CQ_WQES_MAX; 862 863 hba->num_ccell = hba->max_sqes / 2; 864 865 spin_lock_init(&hba->lock); 866 mutex_init(&hba->net_dev_lock); 867 init_waitqueue_head(&hba->eh_wait); 868 if (test_bit(BNX2I_NX2_DEV_57710, &hba->cnic_dev_type)) { 869 hba->hba_shutdown_tmo = 30 * HZ; 870 hba->conn_teardown_tmo = 20 * HZ; 871 hba->conn_ctx_destroy_tmo = 6 * HZ; 872 } else { /* 5706/5708/5709 */ 873 hba->hba_shutdown_tmo = 20 * HZ; 874 hba->conn_teardown_tmo = 10 * HZ; 875 hba->conn_ctx_destroy_tmo = 2 * HZ; 876 } 877 878 #ifdef CONFIG_32BIT 879 spin_lock_init(&hba->stat_lock); 880 #endif 881 memset(&hba->stats, 0, sizeof(struct iscsi_stats_info)); 882 883 if (iscsi_host_add(shost, &hba->pcidev->dev)) 884 goto free_dump_mem; 885 return hba; 886 887 free_dump_mem: 888 bnx2i_release_free_cid_que(hba); 889 cid_que_err: 890 bnx2i_free_mp_bdt(hba); 891 mp_bdt_mem_err: 892 if (hba->regview) { 893 pci_iounmap(hba->pcidev, hba->regview); 894 hba->regview = NULL; 895 } 896 ioreg_map_err: 897 pci_dev_put(hba->pcidev); 898 scsi_host_put(shost); 899 return NULL; 900 } 901 902 /** 903 * bnx2i_free_hba- releases hba structure and resources held by the adapter 904 * @hba: pointer to adapter instance 905 * 906 * free adapter structure and call various cleanup routines. 907 */ 908 void bnx2i_free_hba(struct bnx2i_hba *hba) 909 { 910 struct Scsi_Host *shost = hba->shost; 911 912 iscsi_host_remove(shost); 913 INIT_LIST_HEAD(&hba->ep_ofld_list); 914 INIT_LIST_HEAD(&hba->ep_active_list); 915 INIT_LIST_HEAD(&hba->ep_destroy_list); 916 917 if (hba->regview) { 918 pci_iounmap(hba->pcidev, hba->regview); 919 hba->regview = NULL; 920 } 921 pci_dev_put(hba->pcidev); 922 bnx2i_free_mp_bdt(hba); 923 bnx2i_release_free_cid_que(hba); 924 iscsi_host_free(shost); 925 } 926 927 /** 928 * bnx2i_conn_free_login_resources - free DMA resources used for login process 929 * @hba: pointer to adapter instance 930 * @bnx2i_conn: iscsi connection pointer 931 * 932 * Login related resources, mostly BDT & payload DMA memory is freed 933 */ 934 static void bnx2i_conn_free_login_resources(struct bnx2i_hba *hba, 935 struct bnx2i_conn *bnx2i_conn) 936 { 937 if (bnx2i_conn->gen_pdu.resp_bd_tbl) { 938 dma_free_coherent(&hba->pcidev->dev, CNIC_PAGE_SIZE, 939 bnx2i_conn->gen_pdu.resp_bd_tbl, 940 bnx2i_conn->gen_pdu.resp_bd_dma); 941 bnx2i_conn->gen_pdu.resp_bd_tbl = NULL; 942 } 943 944 if (bnx2i_conn->gen_pdu.req_bd_tbl) { 945 dma_free_coherent(&hba->pcidev->dev, CNIC_PAGE_SIZE, 946 bnx2i_conn->gen_pdu.req_bd_tbl, 947 bnx2i_conn->gen_pdu.req_bd_dma); 948 bnx2i_conn->gen_pdu.req_bd_tbl = NULL; 949 } 950 951 if (bnx2i_conn->gen_pdu.resp_buf) { 952 dma_free_coherent(&hba->pcidev->dev, 953 ISCSI_DEF_MAX_RECV_SEG_LEN, 954 bnx2i_conn->gen_pdu.resp_buf, 955 bnx2i_conn->gen_pdu.resp_dma_addr); 956 bnx2i_conn->gen_pdu.resp_buf = NULL; 957 } 958 959 if (bnx2i_conn->gen_pdu.req_buf) { 960 dma_free_coherent(&hba->pcidev->dev, 961 ISCSI_DEF_MAX_RECV_SEG_LEN, 962 bnx2i_conn->gen_pdu.req_buf, 963 bnx2i_conn->gen_pdu.req_dma_addr); 964 bnx2i_conn->gen_pdu.req_buf = NULL; 965 } 966 } 967 968 /** 969 * bnx2i_conn_alloc_login_resources - alloc DMA resources for login/nop. 970 * @hba: pointer to adapter instance 971 * @bnx2i_conn: iscsi connection pointer 972 * 973 * Mgmt task DNA resources are allocated in this routine. 974 */ 975 static int bnx2i_conn_alloc_login_resources(struct bnx2i_hba *hba, 976 struct bnx2i_conn *bnx2i_conn) 977 { 978 /* Allocate memory for login request/response buffers */ 979 bnx2i_conn->gen_pdu.req_buf = 980 dma_alloc_coherent(&hba->pcidev->dev, 981 ISCSI_DEF_MAX_RECV_SEG_LEN, 982 &bnx2i_conn->gen_pdu.req_dma_addr, 983 GFP_KERNEL); 984 if (bnx2i_conn->gen_pdu.req_buf == NULL) 985 goto login_req_buf_failure; 986 987 bnx2i_conn->gen_pdu.req_buf_size = 0; 988 bnx2i_conn->gen_pdu.req_wr_ptr = bnx2i_conn->gen_pdu.req_buf; 989 990 bnx2i_conn->gen_pdu.resp_buf = 991 dma_alloc_coherent(&hba->pcidev->dev, 992 ISCSI_DEF_MAX_RECV_SEG_LEN, 993 &bnx2i_conn->gen_pdu.resp_dma_addr, 994 GFP_KERNEL); 995 if (bnx2i_conn->gen_pdu.resp_buf == NULL) 996 goto login_resp_buf_failure; 997 998 bnx2i_conn->gen_pdu.resp_buf_size = ISCSI_DEF_MAX_RECV_SEG_LEN; 999 bnx2i_conn->gen_pdu.resp_wr_ptr = bnx2i_conn->gen_pdu.resp_buf; 1000 1001 bnx2i_conn->gen_pdu.req_bd_tbl = 1002 dma_alloc_coherent(&hba->pcidev->dev, CNIC_PAGE_SIZE, 1003 &bnx2i_conn->gen_pdu.req_bd_dma, GFP_KERNEL); 1004 if (bnx2i_conn->gen_pdu.req_bd_tbl == NULL) 1005 goto login_req_bd_tbl_failure; 1006 1007 bnx2i_conn->gen_pdu.resp_bd_tbl = 1008 dma_alloc_coherent(&hba->pcidev->dev, CNIC_PAGE_SIZE, 1009 &bnx2i_conn->gen_pdu.resp_bd_dma, 1010 GFP_KERNEL); 1011 if (bnx2i_conn->gen_pdu.resp_bd_tbl == NULL) 1012 goto login_resp_bd_tbl_failure; 1013 1014 return 0; 1015 1016 login_resp_bd_tbl_failure: 1017 dma_free_coherent(&hba->pcidev->dev, CNIC_PAGE_SIZE, 1018 bnx2i_conn->gen_pdu.req_bd_tbl, 1019 bnx2i_conn->gen_pdu.req_bd_dma); 1020 bnx2i_conn->gen_pdu.req_bd_tbl = NULL; 1021 1022 login_req_bd_tbl_failure: 1023 dma_free_coherent(&hba->pcidev->dev, ISCSI_DEF_MAX_RECV_SEG_LEN, 1024 bnx2i_conn->gen_pdu.resp_buf, 1025 bnx2i_conn->gen_pdu.resp_dma_addr); 1026 bnx2i_conn->gen_pdu.resp_buf = NULL; 1027 login_resp_buf_failure: 1028 dma_free_coherent(&hba->pcidev->dev, ISCSI_DEF_MAX_RECV_SEG_LEN, 1029 bnx2i_conn->gen_pdu.req_buf, 1030 bnx2i_conn->gen_pdu.req_dma_addr); 1031 bnx2i_conn->gen_pdu.req_buf = NULL; 1032 login_req_buf_failure: 1033 iscsi_conn_printk(KERN_ERR, bnx2i_conn->cls_conn->dd_data, 1034 "login resource alloc failed!!\n"); 1035 return -ENOMEM; 1036 1037 } 1038 1039 1040 /** 1041 * bnx2i_iscsi_prep_generic_pdu_bd - prepares BD table. 1042 * @bnx2i_conn: iscsi connection pointer 1043 * 1044 * Allocates buffers and BD tables before shipping requests to cnic 1045 * for PDUs prepared by 'iscsid' daemon 1046 */ 1047 static void bnx2i_iscsi_prep_generic_pdu_bd(struct bnx2i_conn *bnx2i_conn) 1048 { 1049 struct iscsi_bd *bd_tbl; 1050 1051 bd_tbl = (struct iscsi_bd *) bnx2i_conn->gen_pdu.req_bd_tbl; 1052 1053 bd_tbl->buffer_addr_hi = 1054 (u32) ((u64) bnx2i_conn->gen_pdu.req_dma_addr >> 32); 1055 bd_tbl->buffer_addr_lo = (u32) bnx2i_conn->gen_pdu.req_dma_addr; 1056 bd_tbl->buffer_length = bnx2i_conn->gen_pdu.req_wr_ptr - 1057 bnx2i_conn->gen_pdu.req_buf; 1058 bd_tbl->reserved0 = 0; 1059 bd_tbl->flags = ISCSI_BD_LAST_IN_BD_CHAIN | 1060 ISCSI_BD_FIRST_IN_BD_CHAIN; 1061 1062 bd_tbl = (struct iscsi_bd *) bnx2i_conn->gen_pdu.resp_bd_tbl; 1063 bd_tbl->buffer_addr_hi = (u64) bnx2i_conn->gen_pdu.resp_dma_addr >> 32; 1064 bd_tbl->buffer_addr_lo = (u32) bnx2i_conn->gen_pdu.resp_dma_addr; 1065 bd_tbl->buffer_length = ISCSI_DEF_MAX_RECV_SEG_LEN; 1066 bd_tbl->reserved0 = 0; 1067 bd_tbl->flags = ISCSI_BD_LAST_IN_BD_CHAIN | 1068 ISCSI_BD_FIRST_IN_BD_CHAIN; 1069 } 1070 1071 1072 /** 1073 * bnx2i_iscsi_send_generic_request - called to send mgmt tasks. 1074 * @task: transport layer task pointer 1075 * 1076 * called to transmit PDUs prepared by the 'iscsid' daemon. iSCSI login, 1077 * Nop-out and Logout requests flow through this path. 1078 */ 1079 static int bnx2i_iscsi_send_generic_request(struct iscsi_task *task) 1080 { 1081 struct bnx2i_cmd *cmd = task->dd_data; 1082 struct bnx2i_conn *bnx2i_conn = cmd->conn; 1083 int rc = 0; 1084 char *buf; 1085 int data_len; 1086 1087 bnx2i_iscsi_prep_generic_pdu_bd(bnx2i_conn); 1088 switch (task->hdr->opcode & ISCSI_OPCODE_MASK) { 1089 case ISCSI_OP_LOGIN: 1090 bnx2i_send_iscsi_login(bnx2i_conn, task); 1091 break; 1092 case ISCSI_OP_NOOP_OUT: 1093 data_len = bnx2i_conn->gen_pdu.req_buf_size; 1094 buf = bnx2i_conn->gen_pdu.req_buf; 1095 if (data_len) 1096 rc = bnx2i_send_iscsi_nopout(bnx2i_conn, task, 1097 buf, data_len, 1); 1098 else 1099 rc = bnx2i_send_iscsi_nopout(bnx2i_conn, task, 1100 NULL, 0, 1); 1101 break; 1102 case ISCSI_OP_LOGOUT: 1103 rc = bnx2i_send_iscsi_logout(bnx2i_conn, task); 1104 break; 1105 case ISCSI_OP_SCSI_TMFUNC: 1106 rc = bnx2i_send_iscsi_tmf(bnx2i_conn, task); 1107 break; 1108 case ISCSI_OP_TEXT: 1109 rc = bnx2i_send_iscsi_text(bnx2i_conn, task); 1110 break; 1111 default: 1112 iscsi_conn_printk(KERN_ALERT, bnx2i_conn->cls_conn->dd_data, 1113 "send_gen: unsupported op 0x%x\n", 1114 task->hdr->opcode); 1115 } 1116 return rc; 1117 } 1118 1119 1120 /********************************************************************** 1121 * SCSI-ML Interface 1122 **********************************************************************/ 1123 1124 /** 1125 * bnx2i_cpy_scsi_cdb - copies LUN & CDB fields in required format to sq wqe 1126 * @sc: SCSI-ML command pointer 1127 * @cmd: iscsi cmd pointer 1128 */ 1129 static void bnx2i_cpy_scsi_cdb(struct scsi_cmnd *sc, struct bnx2i_cmd *cmd) 1130 { 1131 u32 dword; 1132 int lpcnt; 1133 u8 *srcp; 1134 u32 *dstp; 1135 u32 scsi_lun[2]; 1136 1137 int_to_scsilun(sc->device->lun, (struct scsi_lun *) scsi_lun); 1138 cmd->req.lun[0] = be32_to_cpu(scsi_lun[0]); 1139 cmd->req.lun[1] = be32_to_cpu(scsi_lun[1]); 1140 1141 lpcnt = cmd->scsi_cmd->cmd_len / sizeof(dword); 1142 srcp = (u8 *) sc->cmnd; 1143 dstp = (u32 *) cmd->req.cdb; 1144 while (lpcnt--) { 1145 memcpy(&dword, (const void *) srcp, 4); 1146 *dstp = cpu_to_be32(dword); 1147 srcp += 4; 1148 dstp++; 1149 } 1150 if (sc->cmd_len & 0x3) { 1151 dword = (u32) srcp[0] | ((u32) srcp[1] << 8); 1152 *dstp = cpu_to_be32(dword); 1153 } 1154 } 1155 1156 static void bnx2i_cleanup_task(struct iscsi_task *task) 1157 { 1158 struct iscsi_conn *conn = task->conn; 1159 struct bnx2i_conn *bnx2i_conn = conn->dd_data; 1160 struct bnx2i_hba *hba = bnx2i_conn->hba; 1161 1162 /* 1163 * mgmt task or cmd was never sent to us to transmit. 1164 */ 1165 if (!task->sc || task->state == ISCSI_TASK_PENDING) 1166 return; 1167 /* 1168 * need to clean-up task context to claim dma buffers 1169 */ 1170 if (task->state == ISCSI_TASK_ABRT_TMF) { 1171 bnx2i_send_cmd_cleanup_req(hba, task->dd_data); 1172 1173 spin_unlock_bh(&conn->session->back_lock); 1174 wait_for_completion_timeout(&bnx2i_conn->cmd_cleanup_cmpl, 1175 msecs_to_jiffies(ISCSI_CMD_CLEANUP_TIMEOUT)); 1176 spin_lock_bh(&conn->session->back_lock); 1177 } 1178 bnx2i_iscsi_unmap_sg_list(task->dd_data); 1179 } 1180 1181 /** 1182 * bnx2i_mtask_xmit - transmit mtask to chip for further processing 1183 * @conn: transport layer conn structure pointer 1184 * @task: transport layer command structure pointer 1185 */ 1186 static int 1187 bnx2i_mtask_xmit(struct iscsi_conn *conn, struct iscsi_task *task) 1188 { 1189 struct bnx2i_conn *bnx2i_conn = conn->dd_data; 1190 struct bnx2i_hba *hba = bnx2i_conn->hba; 1191 struct bnx2i_cmd *cmd = task->dd_data; 1192 1193 memset(bnx2i_conn->gen_pdu.req_buf, 0, ISCSI_DEF_MAX_RECV_SEG_LEN); 1194 1195 bnx2i_setup_cmd_wqe_template(cmd); 1196 bnx2i_conn->gen_pdu.req_buf_size = task->data_count; 1197 1198 /* Tx PDU/data length count */ 1199 ADD_STATS_64(hba, tx_pdus, 1); 1200 ADD_STATS_64(hba, tx_bytes, task->data_count); 1201 1202 if (task->data_count) { 1203 memcpy(bnx2i_conn->gen_pdu.req_buf, task->data, 1204 task->data_count); 1205 bnx2i_conn->gen_pdu.req_wr_ptr = 1206 bnx2i_conn->gen_pdu.req_buf + task->data_count; 1207 } 1208 cmd->conn = conn->dd_data; 1209 cmd->scsi_cmd = NULL; 1210 return bnx2i_iscsi_send_generic_request(task); 1211 } 1212 1213 /** 1214 * bnx2i_task_xmit - transmit iscsi command to chip for further processing 1215 * @task: transport layer command structure pointer 1216 * 1217 * maps SG buffers and send request to chip/firmware in the form of SQ WQE 1218 */ 1219 static int bnx2i_task_xmit(struct iscsi_task *task) 1220 { 1221 struct iscsi_conn *conn = task->conn; 1222 struct iscsi_session *session = conn->session; 1223 struct Scsi_Host *shost = iscsi_session_to_shost(session->cls_session); 1224 struct bnx2i_hba *hba = iscsi_host_priv(shost); 1225 struct bnx2i_conn *bnx2i_conn = conn->dd_data; 1226 struct scsi_cmnd *sc = task->sc; 1227 struct bnx2i_cmd *cmd = task->dd_data; 1228 struct iscsi_scsi_req *hdr = (struct iscsi_scsi_req *)task->hdr; 1229 1230 if (atomic_read(&bnx2i_conn->ep->num_active_cmds) + 1 > 1231 hba->max_sqes) 1232 return -ENOMEM; 1233 1234 /* 1235 * If there is no scsi_cmnd this must be a mgmt task 1236 */ 1237 if (!sc) 1238 return bnx2i_mtask_xmit(conn, task); 1239 1240 bnx2i_setup_cmd_wqe_template(cmd); 1241 cmd->req.op_code = ISCSI_OP_SCSI_CMD; 1242 cmd->conn = bnx2i_conn; 1243 cmd->scsi_cmd = sc; 1244 cmd->req.total_data_transfer_length = scsi_bufflen(sc); 1245 cmd->req.cmd_sn = be32_to_cpu(hdr->cmdsn); 1246 1247 bnx2i_iscsi_map_sg_list(cmd); 1248 bnx2i_cpy_scsi_cdb(sc, cmd); 1249 1250 cmd->req.op_attr = ISCSI_ATTR_SIMPLE; 1251 if (sc->sc_data_direction == DMA_TO_DEVICE) { 1252 cmd->req.op_attr |= ISCSI_CMD_REQUEST_WRITE; 1253 cmd->req.itt = task->itt | 1254 (ISCSI_TASK_TYPE_WRITE << ISCSI_CMD_REQUEST_TYPE_SHIFT); 1255 bnx2i_setup_write_cmd_bd_info(task); 1256 } else { 1257 if (scsi_bufflen(sc)) 1258 cmd->req.op_attr |= ISCSI_CMD_REQUEST_READ; 1259 cmd->req.itt = task->itt | 1260 (ISCSI_TASK_TYPE_READ << ISCSI_CMD_REQUEST_TYPE_SHIFT); 1261 } 1262 1263 cmd->req.num_bds = cmd->io_tbl.bd_valid; 1264 if (!cmd->io_tbl.bd_valid) { 1265 cmd->req.bd_list_addr_lo = (u32) hba->mp_bd_dma; 1266 cmd->req.bd_list_addr_hi = (u32) ((u64) hba->mp_bd_dma >> 32); 1267 cmd->req.num_bds = 1; 1268 } 1269 1270 bnx2i_send_iscsi_scsicmd(bnx2i_conn, cmd); 1271 return 0; 1272 } 1273 1274 /** 1275 * bnx2i_session_create - create a new iscsi session 1276 * @ep: pointer to iscsi endpoint 1277 * @cmds_max: user specified maximum commands 1278 * @qdepth: scsi queue depth to support 1279 * @initial_cmdsn: initial iscsi CMDSN to be used for this session 1280 * 1281 * Creates a new iSCSI session instance on given device. 1282 */ 1283 static struct iscsi_cls_session * 1284 bnx2i_session_create(struct iscsi_endpoint *ep, 1285 uint16_t cmds_max, uint16_t qdepth, 1286 uint32_t initial_cmdsn) 1287 { 1288 struct Scsi_Host *shost; 1289 struct iscsi_cls_session *cls_session; 1290 struct bnx2i_hba *hba; 1291 struct bnx2i_endpoint *bnx2i_ep; 1292 1293 if (!ep) { 1294 printk(KERN_ERR "bnx2i: missing ep.\n"); 1295 return NULL; 1296 } 1297 1298 bnx2i_ep = ep->dd_data; 1299 shost = bnx2i_ep->hba->shost; 1300 hba = iscsi_host_priv(shost); 1301 if (bnx2i_adapter_ready(hba)) 1302 return NULL; 1303 1304 /* 1305 * user can override hw limit as long as it is within 1306 * the min/max. 1307 */ 1308 if (cmds_max > hba->max_sqes) 1309 cmds_max = hba->max_sqes; 1310 else if (cmds_max < BNX2I_SQ_WQES_MIN) 1311 cmds_max = BNX2I_SQ_WQES_MIN; 1312 1313 cls_session = iscsi_session_setup(&bnx2i_iscsi_transport, shost, 1314 cmds_max, 0, sizeof(struct bnx2i_cmd), 1315 initial_cmdsn, ISCSI_MAX_TARGET); 1316 if (!cls_session) 1317 return NULL; 1318 1319 if (bnx2i_setup_cmd_pool(hba, cls_session->dd_data)) 1320 goto session_teardown; 1321 return cls_session; 1322 1323 session_teardown: 1324 iscsi_session_teardown(cls_session); 1325 return NULL; 1326 } 1327 1328 1329 /** 1330 * bnx2i_session_destroy - destroys iscsi session 1331 * @cls_session: pointer to iscsi cls session 1332 * 1333 * Destroys previously created iSCSI session instance and releases 1334 * all resources held by it 1335 */ 1336 static void bnx2i_session_destroy(struct iscsi_cls_session *cls_session) 1337 { 1338 struct iscsi_session *session = cls_session->dd_data; 1339 struct Scsi_Host *shost = iscsi_session_to_shost(cls_session); 1340 struct bnx2i_hba *hba = iscsi_host_priv(shost); 1341 1342 bnx2i_destroy_cmd_pool(hba, session); 1343 iscsi_session_teardown(cls_session); 1344 } 1345 1346 1347 /** 1348 * bnx2i_conn_create - create iscsi connection instance 1349 * @cls_session: pointer to iscsi cls session 1350 * @cid: iscsi cid as per rfc (not NX2's CID terminology) 1351 * 1352 * Creates a new iSCSI connection instance for a given session 1353 */ 1354 static struct iscsi_cls_conn * 1355 bnx2i_conn_create(struct iscsi_cls_session *cls_session, uint32_t cid) 1356 { 1357 struct Scsi_Host *shost = iscsi_session_to_shost(cls_session); 1358 struct bnx2i_hba *hba = iscsi_host_priv(shost); 1359 struct bnx2i_conn *bnx2i_conn; 1360 struct iscsi_cls_conn *cls_conn; 1361 struct iscsi_conn *conn; 1362 1363 cls_conn = iscsi_conn_setup(cls_session, sizeof(*bnx2i_conn), 1364 cid); 1365 if (!cls_conn) 1366 return NULL; 1367 conn = cls_conn->dd_data; 1368 1369 bnx2i_conn = conn->dd_data; 1370 bnx2i_conn->cls_conn = cls_conn; 1371 bnx2i_conn->hba = hba; 1372 1373 atomic_set(&bnx2i_conn->work_cnt, 0); 1374 1375 /* 'ep' ptr will be assigned in bind() call */ 1376 bnx2i_conn->ep = NULL; 1377 init_completion(&bnx2i_conn->cmd_cleanup_cmpl); 1378 1379 if (bnx2i_conn_alloc_login_resources(hba, bnx2i_conn)) { 1380 iscsi_conn_printk(KERN_ALERT, conn, 1381 "conn_new: login resc alloc failed!!\n"); 1382 goto free_conn; 1383 } 1384 1385 return cls_conn; 1386 1387 free_conn: 1388 iscsi_conn_teardown(cls_conn); 1389 return NULL; 1390 } 1391 1392 /** 1393 * bnx2i_conn_bind - binds iscsi sess, conn and ep objects together 1394 * @cls_session: pointer to iscsi cls session 1395 * @cls_conn: pointer to iscsi cls conn 1396 * @transport_fd: 64-bit EP handle 1397 * @is_leading: leading connection on this session? 1398 * 1399 * Binds together iSCSI session instance, iSCSI connection instance 1400 * and the TCP connection. This routine returns error code if 1401 * TCP connection does not belong on the device iSCSI sess/conn 1402 * is bound 1403 */ 1404 static int bnx2i_conn_bind(struct iscsi_cls_session *cls_session, 1405 struct iscsi_cls_conn *cls_conn, 1406 uint64_t transport_fd, int is_leading) 1407 { 1408 struct iscsi_conn *conn = cls_conn->dd_data; 1409 struct bnx2i_conn *bnx2i_conn = conn->dd_data; 1410 struct Scsi_Host *shost = iscsi_session_to_shost(cls_session); 1411 struct bnx2i_hba *hba = iscsi_host_priv(shost); 1412 struct bnx2i_endpoint *bnx2i_ep; 1413 struct iscsi_endpoint *ep; 1414 int ret_code; 1415 1416 ep = iscsi_lookup_endpoint(transport_fd); 1417 if (!ep) 1418 return -EINVAL; 1419 /* 1420 * Forcefully terminate all in progress connection recovery at the 1421 * earliest, either in bind(), send_pdu(LOGIN), or conn_start() 1422 */ 1423 if (bnx2i_adapter_ready(hba)) { 1424 ret_code = -EIO; 1425 goto put_ep; 1426 } 1427 1428 bnx2i_ep = ep->dd_data; 1429 if ((bnx2i_ep->state == EP_STATE_TCP_FIN_RCVD) || 1430 (bnx2i_ep->state == EP_STATE_TCP_RST_RCVD)) { 1431 /* Peer disconnect via' FIN or RST */ 1432 ret_code = -EINVAL; 1433 goto put_ep; 1434 } 1435 1436 if (iscsi_conn_bind(cls_session, cls_conn, is_leading)) { 1437 ret_code = -EINVAL; 1438 goto put_ep; 1439 } 1440 1441 if (bnx2i_ep->hba != hba) { 1442 /* Error - TCP connection does not belong to this device 1443 */ 1444 iscsi_conn_printk(KERN_ALERT, cls_conn->dd_data, 1445 "conn bind, ep=0x%p (%s) does not", 1446 bnx2i_ep, bnx2i_ep->hba->netdev->name); 1447 iscsi_conn_printk(KERN_ALERT, cls_conn->dd_data, 1448 "belong to hba (%s)\n", 1449 hba->netdev->name); 1450 ret_code = -EEXIST; 1451 goto put_ep; 1452 } 1453 bnx2i_ep->conn = bnx2i_conn; 1454 bnx2i_conn->ep = bnx2i_ep; 1455 bnx2i_conn->iscsi_conn_cid = bnx2i_ep->ep_iscsi_cid; 1456 bnx2i_conn->fw_cid = bnx2i_ep->ep_cid; 1457 1458 ret_code = bnx2i_bind_conn_to_iscsi_cid(hba, bnx2i_conn, 1459 bnx2i_ep->ep_iscsi_cid); 1460 1461 /* 5706/5708/5709 FW takes RQ as full when initiated, but for 57710 1462 * driver needs to explicitly replenish RQ index during setup. 1463 */ 1464 if (test_bit(BNX2I_NX2_DEV_57710, &bnx2i_ep->hba->cnic_dev_type)) 1465 bnx2i_put_rq_buf(bnx2i_conn, 0); 1466 1467 bnx2i_arm_cq_event_coalescing(bnx2i_conn->ep, CNIC_ARM_CQE); 1468 put_ep: 1469 iscsi_put_endpoint(ep); 1470 return ret_code; 1471 } 1472 1473 1474 /** 1475 * bnx2i_conn_destroy - destroy iscsi connection instance & release resources 1476 * @cls_conn: pointer to iscsi cls conn 1477 * 1478 * Destroy an iSCSI connection instance and release memory resources held by 1479 * this connection 1480 */ 1481 static void bnx2i_conn_destroy(struct iscsi_cls_conn *cls_conn) 1482 { 1483 struct iscsi_conn *conn = cls_conn->dd_data; 1484 struct bnx2i_conn *bnx2i_conn = conn->dd_data; 1485 struct Scsi_Host *shost; 1486 struct bnx2i_hba *hba; 1487 struct bnx2i_work *work, *tmp; 1488 unsigned cpu = 0; 1489 struct bnx2i_percpu_s *p; 1490 1491 shost = iscsi_session_to_shost(iscsi_conn_to_session(cls_conn)); 1492 hba = iscsi_host_priv(shost); 1493 1494 bnx2i_conn_free_login_resources(hba, bnx2i_conn); 1495 1496 if (atomic_read(&bnx2i_conn->work_cnt)) { 1497 for_each_online_cpu(cpu) { 1498 p = &per_cpu(bnx2i_percpu, cpu); 1499 spin_lock_bh(&p->p_work_lock); 1500 list_for_each_entry_safe(work, tmp, 1501 &p->work_list, list) { 1502 if (work->session == conn->session && 1503 work->bnx2i_conn == bnx2i_conn) { 1504 list_del_init(&work->list); 1505 kfree(work); 1506 if (!atomic_dec_and_test( 1507 &bnx2i_conn->work_cnt)) 1508 break; 1509 } 1510 } 1511 spin_unlock_bh(&p->p_work_lock); 1512 } 1513 } 1514 1515 iscsi_conn_teardown(cls_conn); 1516 } 1517 1518 1519 /** 1520 * bnx2i_ep_get_param - return iscsi ep parameter to caller 1521 * @ep: pointer to iscsi endpoint 1522 * @param: parameter type identifier 1523 * @buf: buffer pointer 1524 * 1525 * returns iSCSI ep parameters 1526 */ 1527 static int bnx2i_ep_get_param(struct iscsi_endpoint *ep, 1528 enum iscsi_param param, char *buf) 1529 { 1530 struct bnx2i_endpoint *bnx2i_ep = ep->dd_data; 1531 struct bnx2i_hba *hba = bnx2i_ep->hba; 1532 int len = -ENOTCONN; 1533 1534 if (!hba) 1535 return -ENOTCONN; 1536 1537 switch (param) { 1538 case ISCSI_PARAM_CONN_PORT: 1539 mutex_lock(&hba->net_dev_lock); 1540 if (bnx2i_ep->cm_sk) 1541 len = sprintf(buf, "%hu\n", bnx2i_ep->cm_sk->dst_port); 1542 mutex_unlock(&hba->net_dev_lock); 1543 break; 1544 case ISCSI_PARAM_CONN_ADDRESS: 1545 mutex_lock(&hba->net_dev_lock); 1546 if (bnx2i_ep->cm_sk) 1547 len = sprintf(buf, "%pI4\n", &bnx2i_ep->cm_sk->dst_ip); 1548 mutex_unlock(&hba->net_dev_lock); 1549 break; 1550 default: 1551 return -ENOSYS; 1552 } 1553 1554 return len; 1555 } 1556 1557 /** 1558 * bnx2i_host_get_param - returns host (adapter) related parameters 1559 * @shost: scsi host pointer 1560 * @param: parameter type identifier 1561 * @buf: buffer pointer 1562 */ 1563 static int bnx2i_host_get_param(struct Scsi_Host *shost, 1564 enum iscsi_host_param param, char *buf) 1565 { 1566 struct bnx2i_hba *hba = iscsi_host_priv(shost); 1567 int len = 0; 1568 1569 switch (param) { 1570 case ISCSI_HOST_PARAM_HWADDRESS: 1571 len = sysfs_format_mac(buf, hba->cnic->mac_addr, 6); 1572 break; 1573 case ISCSI_HOST_PARAM_NETDEV_NAME: 1574 len = sprintf(buf, "%s\n", hba->netdev->name); 1575 break; 1576 case ISCSI_HOST_PARAM_IPADDRESS: { 1577 struct list_head *active_list = &hba->ep_active_list; 1578 1579 read_lock_bh(&hba->ep_rdwr_lock); 1580 if (!list_empty(&hba->ep_active_list)) { 1581 struct bnx2i_endpoint *bnx2i_ep; 1582 struct cnic_sock *csk; 1583 1584 bnx2i_ep = list_first_entry(active_list, 1585 struct bnx2i_endpoint, 1586 link); 1587 csk = bnx2i_ep->cm_sk; 1588 if (test_bit(SK_F_IPV6, &csk->flags)) 1589 len = sprintf(buf, "%pI6\n", csk->src_ip); 1590 else 1591 len = sprintf(buf, "%pI4\n", csk->src_ip); 1592 } 1593 read_unlock_bh(&hba->ep_rdwr_lock); 1594 break; 1595 } 1596 default: 1597 return iscsi_host_get_param(shost, param, buf); 1598 } 1599 return len; 1600 } 1601 1602 /** 1603 * bnx2i_conn_start - completes iscsi connection migration to FFP 1604 * @cls_conn: pointer to iscsi cls conn 1605 * 1606 * last call in FFP migration to handover iscsi conn to the driver 1607 */ 1608 static int bnx2i_conn_start(struct iscsi_cls_conn *cls_conn) 1609 { 1610 struct iscsi_conn *conn = cls_conn->dd_data; 1611 struct bnx2i_conn *bnx2i_conn = conn->dd_data; 1612 1613 bnx2i_conn->ep->state = EP_STATE_ULP_UPDATE_START; 1614 bnx2i_update_iscsi_conn(conn); 1615 1616 /* 1617 * this should normally not sleep for a long time so it should 1618 * not disrupt the caller. 1619 */ 1620 timer_setup(&bnx2i_conn->ep->ofld_timer, bnx2i_ep_ofld_timer, 0); 1621 bnx2i_conn->ep->ofld_timer.expires = 1 * HZ + jiffies; 1622 add_timer(&bnx2i_conn->ep->ofld_timer); 1623 /* update iSCSI context for this conn, wait for CNIC to complete */ 1624 wait_event_interruptible(bnx2i_conn->ep->ofld_wait, 1625 bnx2i_conn->ep->state != EP_STATE_ULP_UPDATE_START); 1626 1627 if (signal_pending(current)) 1628 flush_signals(current); 1629 del_timer_sync(&bnx2i_conn->ep->ofld_timer); 1630 1631 iscsi_conn_start(cls_conn); 1632 return 0; 1633 } 1634 1635 1636 /** 1637 * bnx2i_conn_get_stats - returns iSCSI stats 1638 * @cls_conn: pointer to iscsi cls conn 1639 * @stats: pointer to iscsi statistic struct 1640 */ 1641 static void bnx2i_conn_get_stats(struct iscsi_cls_conn *cls_conn, 1642 struct iscsi_stats *stats) 1643 { 1644 struct iscsi_conn *conn = cls_conn->dd_data; 1645 1646 stats->txdata_octets = conn->txdata_octets; 1647 stats->rxdata_octets = conn->rxdata_octets; 1648 stats->scsicmd_pdus = conn->scsicmd_pdus_cnt; 1649 stats->dataout_pdus = conn->dataout_pdus_cnt; 1650 stats->scsirsp_pdus = conn->scsirsp_pdus_cnt; 1651 stats->datain_pdus = conn->datain_pdus_cnt; 1652 stats->r2t_pdus = conn->r2t_pdus_cnt; 1653 stats->tmfcmd_pdus = conn->tmfcmd_pdus_cnt; 1654 stats->tmfrsp_pdus = conn->tmfrsp_pdus_cnt; 1655 stats->digest_err = 0; 1656 stats->timeout_err = 0; 1657 strcpy(stats->custom[0].desc, "eh_abort_cnt"); 1658 stats->custom[0].value = conn->eh_abort_cnt; 1659 stats->custom_length = 1; 1660 } 1661 1662 1663 /** 1664 * bnx2i_check_route - checks if target IP route belongs to one of NX2 devices 1665 * @dst_addr: target IP address 1666 * 1667 * check if route resolves to BNX2 device 1668 */ 1669 static struct bnx2i_hba *bnx2i_check_route(struct sockaddr *dst_addr) 1670 { 1671 struct sockaddr_in *desti = (struct sockaddr_in *) dst_addr; 1672 struct bnx2i_hba *hba; 1673 struct cnic_dev *cnic = NULL; 1674 1675 hba = get_adapter_list_head(); 1676 if (hba && hba->cnic) 1677 cnic = hba->cnic->cm_select_dev(desti, CNIC_ULP_ISCSI); 1678 if (!cnic) { 1679 printk(KERN_ALERT "bnx2i: no route," 1680 "can't connect using cnic\n"); 1681 goto no_nx2_route; 1682 } 1683 hba = bnx2i_find_hba_for_cnic(cnic); 1684 if (!hba) 1685 goto no_nx2_route; 1686 1687 if (bnx2i_adapter_ready(hba)) { 1688 printk(KERN_ALERT "bnx2i: check route, hba not found\n"); 1689 goto no_nx2_route; 1690 } 1691 if (hba->netdev->mtu > hba->mtu_supported) { 1692 printk(KERN_ALERT "bnx2i: %s network i/f mtu is set to %d\n", 1693 hba->netdev->name, hba->netdev->mtu); 1694 printk(KERN_ALERT "bnx2i: iSCSI HBA can support mtu of %d\n", 1695 hba->mtu_supported); 1696 goto no_nx2_route; 1697 } 1698 return hba; 1699 no_nx2_route: 1700 return NULL; 1701 } 1702 1703 1704 /** 1705 * bnx2i_tear_down_conn - tear down iscsi/tcp connection and free resources 1706 * @hba: pointer to adapter instance 1707 * @ep: endpoint (transport identifier) structure 1708 * 1709 * destroys cm_sock structure and on chip iscsi context 1710 */ 1711 static int bnx2i_tear_down_conn(struct bnx2i_hba *hba, 1712 struct bnx2i_endpoint *ep) 1713 { 1714 if (test_bit(BNX2I_CNIC_REGISTERED, &hba->reg_with_cnic) && ep->cm_sk) 1715 hba->cnic->cm_destroy(ep->cm_sk); 1716 1717 if (test_bit(BNX2I_NX2_DEV_57710, &hba->cnic_dev_type) && 1718 ep->state == EP_STATE_DISCONN_TIMEDOUT) { 1719 if (ep->conn && ep->conn->cls_conn && 1720 ep->conn->cls_conn->dd_data) { 1721 struct iscsi_conn *conn = ep->conn->cls_conn->dd_data; 1722 1723 /* Must suspend all rx queue activity for this ep */ 1724 set_bit(ISCSI_CONN_FLAG_SUSPEND_RX, &conn->flags); 1725 } 1726 /* CONN_DISCONNECT timeout may or may not be an issue depending 1727 * on what transcribed in TCP layer, different targets behave 1728 * differently 1729 */ 1730 printk(KERN_ALERT "bnx2i (%s): - WARN - CONN_DISCON timed out, " 1731 "please submit GRC Dump, NW/PCIe trace, " 1732 "driver msgs to developers for analysis\n", 1733 hba->netdev->name); 1734 } 1735 1736 ep->state = EP_STATE_CLEANUP_START; 1737 timer_setup(&ep->ofld_timer, bnx2i_ep_ofld_timer, 0); 1738 ep->ofld_timer.expires = hba->conn_ctx_destroy_tmo + jiffies; 1739 add_timer(&ep->ofld_timer); 1740 1741 bnx2i_ep_destroy_list_add(hba, ep); 1742 1743 /* destroy iSCSI context, wait for it to complete */ 1744 if (bnx2i_send_conn_destroy(hba, ep)) 1745 ep->state = EP_STATE_CLEANUP_CMPL; 1746 1747 wait_event_interruptible(ep->ofld_wait, 1748 (ep->state != EP_STATE_CLEANUP_START)); 1749 1750 if (signal_pending(current)) 1751 flush_signals(current); 1752 del_timer_sync(&ep->ofld_timer); 1753 1754 bnx2i_ep_destroy_list_del(hba, ep); 1755 1756 if (ep->state != EP_STATE_CLEANUP_CMPL) 1757 /* should never happen */ 1758 printk(KERN_ALERT "bnx2i - conn destroy failed\n"); 1759 1760 return 0; 1761 } 1762 1763 1764 /** 1765 * bnx2i_ep_connect - establish TCP connection to target portal 1766 * @shost: scsi host 1767 * @dst_addr: target IP address 1768 * @non_blocking: blocking or non-blocking call 1769 * 1770 * this routine initiates the TCP/IP connection by invoking Option-2 i/f 1771 * with l5_core and the CNIC. This is a multi-step process of resolving 1772 * route to target, create a iscsi connection context, handshaking with 1773 * CNIC module to create/initialize the socket struct and finally 1774 * sending down option-2 request to complete TCP 3-way handshake 1775 */ 1776 static struct iscsi_endpoint *bnx2i_ep_connect(struct Scsi_Host *shost, 1777 struct sockaddr *dst_addr, 1778 int non_blocking) 1779 { 1780 u32 iscsi_cid = BNX2I_CID_RESERVED; 1781 struct sockaddr_in *desti = (struct sockaddr_in *) dst_addr; 1782 struct sockaddr_in6 *desti6; 1783 struct bnx2i_endpoint *bnx2i_ep; 1784 struct bnx2i_hba *hba; 1785 struct cnic_dev *cnic; 1786 struct cnic_sockaddr saddr; 1787 struct iscsi_endpoint *ep; 1788 int rc = 0; 1789 1790 if (shost) { 1791 /* driver is given scsi host to work with */ 1792 hba = iscsi_host_priv(shost); 1793 } else 1794 /* 1795 * check if the given destination can be reached through 1796 * a iscsi capable NetXtreme2 device 1797 */ 1798 hba = bnx2i_check_route(dst_addr); 1799 1800 if (!hba) { 1801 rc = -EINVAL; 1802 goto nohba; 1803 } 1804 mutex_lock(&hba->net_dev_lock); 1805 1806 if (bnx2i_adapter_ready(hba) || !hba->cid_que.cid_free_cnt) { 1807 rc = -EPERM; 1808 goto check_busy; 1809 } 1810 cnic = hba->cnic; 1811 ep = bnx2i_alloc_ep(hba); 1812 if (!ep) { 1813 rc = -ENOMEM; 1814 goto check_busy; 1815 } 1816 bnx2i_ep = ep->dd_data; 1817 1818 atomic_set(&bnx2i_ep->num_active_cmds, 0); 1819 iscsi_cid = bnx2i_alloc_iscsi_cid(hba); 1820 if (iscsi_cid == -1) { 1821 printk(KERN_ALERT "bnx2i (%s): alloc_ep - unable to allocate " 1822 "iscsi cid\n", hba->netdev->name); 1823 rc = -ENOMEM; 1824 bnx2i_free_ep(ep); 1825 goto check_busy; 1826 } 1827 bnx2i_ep->hba_age = hba->age; 1828 1829 rc = bnx2i_alloc_qp_resc(hba, bnx2i_ep); 1830 if (rc != 0) { 1831 printk(KERN_ALERT "bnx2i (%s): ep_conn - alloc QP resc error" 1832 "\n", hba->netdev->name); 1833 rc = -ENOMEM; 1834 goto qp_resc_err; 1835 } 1836 1837 bnx2i_ep->ep_iscsi_cid = (u16)iscsi_cid; 1838 bnx2i_ep->state = EP_STATE_OFLD_START; 1839 bnx2i_ep_ofld_list_add(hba, bnx2i_ep); 1840 1841 timer_setup(&bnx2i_ep->ofld_timer, bnx2i_ep_ofld_timer, 0); 1842 bnx2i_ep->ofld_timer.expires = 2 * HZ + jiffies; 1843 add_timer(&bnx2i_ep->ofld_timer); 1844 1845 if (bnx2i_send_conn_ofld_req(hba, bnx2i_ep)) { 1846 if (bnx2i_ep->state == EP_STATE_OFLD_FAILED_CID_BUSY) { 1847 printk(KERN_ALERT "bnx2i (%s): iscsi cid %d is busy\n", 1848 hba->netdev->name, bnx2i_ep->ep_iscsi_cid); 1849 rc = -EBUSY; 1850 } else 1851 rc = -ENOSPC; 1852 printk(KERN_ALERT "bnx2i (%s): unable to send conn offld kwqe" 1853 "\n", hba->netdev->name); 1854 bnx2i_ep_ofld_list_del(hba, bnx2i_ep); 1855 goto conn_failed; 1856 } 1857 1858 /* Wait for CNIC hardware to setup conn context and return 'cid' */ 1859 wait_event_interruptible(bnx2i_ep->ofld_wait, 1860 bnx2i_ep->state != EP_STATE_OFLD_START); 1861 1862 if (signal_pending(current)) 1863 flush_signals(current); 1864 del_timer_sync(&bnx2i_ep->ofld_timer); 1865 1866 bnx2i_ep_ofld_list_del(hba, bnx2i_ep); 1867 1868 if (bnx2i_ep->state != EP_STATE_OFLD_COMPL) { 1869 if (bnx2i_ep->state == EP_STATE_OFLD_FAILED_CID_BUSY) { 1870 printk(KERN_ALERT "bnx2i (%s): iscsi cid %d is busy\n", 1871 hba->netdev->name, bnx2i_ep->ep_iscsi_cid); 1872 rc = -EBUSY; 1873 } else 1874 rc = -ENOSPC; 1875 goto conn_failed; 1876 } 1877 1878 rc = cnic->cm_create(cnic, CNIC_ULP_ISCSI, bnx2i_ep->ep_cid, 1879 iscsi_cid, &bnx2i_ep->cm_sk, bnx2i_ep); 1880 if (rc) { 1881 rc = -EINVAL; 1882 /* Need to terminate and cleanup the connection */ 1883 goto release_ep; 1884 } 1885 1886 bnx2i_ep->cm_sk->rcv_buf = 256 * 1024; 1887 bnx2i_ep->cm_sk->snd_buf = 256 * 1024; 1888 clear_bit(SK_TCP_TIMESTAMP, &bnx2i_ep->cm_sk->tcp_flags); 1889 1890 memset(&saddr, 0, sizeof(saddr)); 1891 if (dst_addr->sa_family == AF_INET) { 1892 desti = (struct sockaddr_in *) dst_addr; 1893 saddr.remote.v4 = *desti; 1894 saddr.local.v4.sin_family = desti->sin_family; 1895 } else if (dst_addr->sa_family == AF_INET6) { 1896 desti6 = (struct sockaddr_in6 *) dst_addr; 1897 saddr.remote.v6 = *desti6; 1898 saddr.local.v6.sin6_family = desti6->sin6_family; 1899 } 1900 1901 bnx2i_ep->timestamp = jiffies; 1902 bnx2i_ep->state = EP_STATE_CONNECT_START; 1903 if (!test_bit(BNX2I_CNIC_REGISTERED, &hba->reg_with_cnic)) { 1904 rc = -EINVAL; 1905 goto conn_failed; 1906 } else 1907 rc = cnic->cm_connect(bnx2i_ep->cm_sk, &saddr); 1908 if (rc) 1909 goto release_ep; 1910 1911 bnx2i_ep_active_list_add(hba, bnx2i_ep); 1912 1913 rc = bnx2i_map_ep_dbell_regs(bnx2i_ep); 1914 if (rc) 1915 goto del_active_ep; 1916 1917 mutex_unlock(&hba->net_dev_lock); 1918 return ep; 1919 1920 del_active_ep: 1921 bnx2i_ep_active_list_del(hba, bnx2i_ep); 1922 release_ep: 1923 if (bnx2i_tear_down_conn(hba, bnx2i_ep)) { 1924 mutex_unlock(&hba->net_dev_lock); 1925 return ERR_PTR(rc); 1926 } 1927 conn_failed: 1928 bnx2i_free_qp_resc(hba, bnx2i_ep); 1929 qp_resc_err: 1930 bnx2i_free_ep(ep); 1931 check_busy: 1932 mutex_unlock(&hba->net_dev_lock); 1933 nohba: 1934 return ERR_PTR(rc); 1935 } 1936 1937 1938 /** 1939 * bnx2i_ep_poll - polls for TCP connection establishement 1940 * @ep: TCP connection (endpoint) handle 1941 * @timeout_ms: timeout value in milli secs 1942 * 1943 * polls for TCP connect request to complete 1944 */ 1945 static int bnx2i_ep_poll(struct iscsi_endpoint *ep, int timeout_ms) 1946 { 1947 struct bnx2i_endpoint *bnx2i_ep; 1948 int rc = 0; 1949 1950 bnx2i_ep = ep->dd_data; 1951 if ((bnx2i_ep->state == EP_STATE_IDLE) || 1952 (bnx2i_ep->state == EP_STATE_CONNECT_FAILED) || 1953 (bnx2i_ep->state == EP_STATE_OFLD_FAILED)) 1954 return -1; 1955 if (bnx2i_ep->state == EP_STATE_CONNECT_COMPL) 1956 return 1; 1957 1958 rc = wait_event_interruptible_timeout(bnx2i_ep->ofld_wait, 1959 ((bnx2i_ep->state == 1960 EP_STATE_OFLD_FAILED) || 1961 (bnx2i_ep->state == 1962 EP_STATE_CONNECT_FAILED) || 1963 (bnx2i_ep->state == 1964 EP_STATE_CONNECT_COMPL)), 1965 msecs_to_jiffies(timeout_ms)); 1966 if (bnx2i_ep->state == EP_STATE_OFLD_FAILED) 1967 rc = -1; 1968 1969 if (rc > 0) 1970 return 1; 1971 else if (!rc) 1972 return 0; /* timeout */ 1973 else 1974 return rc; 1975 } 1976 1977 1978 /** 1979 * bnx2i_ep_tcp_conn_active - check EP state transition 1980 * @bnx2i_ep: endpoint pointer 1981 * 1982 * check if underlying TCP connection is active 1983 */ 1984 static int bnx2i_ep_tcp_conn_active(struct bnx2i_endpoint *bnx2i_ep) 1985 { 1986 int ret; 1987 int cnic_dev_10g = 0; 1988 1989 if (test_bit(BNX2I_NX2_DEV_57710, &bnx2i_ep->hba->cnic_dev_type)) 1990 cnic_dev_10g = 1; 1991 1992 switch (bnx2i_ep->state) { 1993 case EP_STATE_CLEANUP_FAILED: 1994 case EP_STATE_OFLD_FAILED: 1995 case EP_STATE_DISCONN_TIMEDOUT: 1996 ret = 0; 1997 break; 1998 case EP_STATE_CONNECT_START: 1999 case EP_STATE_CONNECT_FAILED: 2000 case EP_STATE_CONNECT_COMPL: 2001 case EP_STATE_ULP_UPDATE_START: 2002 case EP_STATE_ULP_UPDATE_COMPL: 2003 case EP_STATE_TCP_FIN_RCVD: 2004 case EP_STATE_LOGOUT_SENT: 2005 case EP_STATE_LOGOUT_RESP_RCVD: 2006 case EP_STATE_ULP_UPDATE_FAILED: 2007 ret = 1; 2008 break; 2009 case EP_STATE_TCP_RST_RCVD: 2010 if (cnic_dev_10g) 2011 ret = 0; 2012 else 2013 ret = 1; 2014 break; 2015 default: 2016 ret = 0; 2017 } 2018 2019 return ret; 2020 } 2021 2022 2023 /** 2024 * bnx2i_hw_ep_disconnect - executes TCP connection teardown process in the hw 2025 * @bnx2i_ep: TCP connection (bnx2i endpoint) handle 2026 * 2027 * executes TCP connection teardown process 2028 */ 2029 int bnx2i_hw_ep_disconnect(struct bnx2i_endpoint *bnx2i_ep) 2030 { 2031 struct bnx2i_hba *hba = bnx2i_ep->hba; 2032 struct cnic_dev *cnic; 2033 struct iscsi_session *session = NULL; 2034 struct iscsi_conn *conn = NULL; 2035 int ret = 0; 2036 int close = 0; 2037 int close_ret = 0; 2038 2039 if (!hba) 2040 return 0; 2041 2042 cnic = hba->cnic; 2043 if (!cnic) 2044 return 0; 2045 2046 if (bnx2i_ep->state == EP_STATE_IDLE || 2047 bnx2i_ep->state == EP_STATE_DISCONN_TIMEDOUT) 2048 return 0; 2049 2050 if (!bnx2i_ep_tcp_conn_active(bnx2i_ep)) 2051 goto destroy_conn; 2052 2053 if (bnx2i_ep->conn) { 2054 conn = bnx2i_ep->conn->cls_conn->dd_data; 2055 session = conn->session; 2056 } 2057 2058 timer_setup(&bnx2i_ep->ofld_timer, bnx2i_ep_ofld_timer, 0); 2059 bnx2i_ep->ofld_timer.expires = hba->conn_teardown_tmo + jiffies; 2060 add_timer(&bnx2i_ep->ofld_timer); 2061 2062 if (!test_bit(BNX2I_CNIC_REGISTERED, &hba->reg_with_cnic)) 2063 goto out; 2064 2065 if (session) { 2066 spin_lock_bh(&session->frwd_lock); 2067 if (bnx2i_ep->state != EP_STATE_TCP_FIN_RCVD) { 2068 if (session->state == ISCSI_STATE_LOGGING_OUT) { 2069 if (bnx2i_ep->state == EP_STATE_LOGOUT_SENT) { 2070 /* Logout sent, but no resp */ 2071 printk(KERN_ALERT "bnx2i (%s): WARNING" 2072 " logout response was not " 2073 "received!\n", 2074 bnx2i_ep->hba->netdev->name); 2075 } else if (bnx2i_ep->state == 2076 EP_STATE_LOGOUT_RESP_RCVD) 2077 close = 1; 2078 } 2079 } else 2080 close = 1; 2081 2082 spin_unlock_bh(&session->frwd_lock); 2083 } 2084 2085 bnx2i_ep->state = EP_STATE_DISCONN_START; 2086 2087 if (close) 2088 close_ret = cnic->cm_close(bnx2i_ep->cm_sk); 2089 else 2090 close_ret = cnic->cm_abort(bnx2i_ep->cm_sk); 2091 2092 if (close_ret) 2093 printk(KERN_ALERT "bnx2i (%s): close/abort(%d) returned %d\n", 2094 bnx2i_ep->hba->netdev->name, close, close_ret); 2095 else 2096 /* wait for option-2 conn teardown */ 2097 wait_event_interruptible(bnx2i_ep->ofld_wait, 2098 ((bnx2i_ep->state != EP_STATE_DISCONN_START) 2099 && (bnx2i_ep->state != EP_STATE_TCP_FIN_RCVD))); 2100 2101 if (signal_pending(current)) 2102 flush_signals(current); 2103 del_timer_sync(&bnx2i_ep->ofld_timer); 2104 2105 destroy_conn: 2106 bnx2i_ep_active_list_del(hba, bnx2i_ep); 2107 if (bnx2i_tear_down_conn(hba, bnx2i_ep)) 2108 return -EINVAL; 2109 out: 2110 bnx2i_ep->state = EP_STATE_IDLE; 2111 return ret; 2112 } 2113 2114 2115 /** 2116 * bnx2i_ep_disconnect - executes TCP connection teardown process 2117 * @ep: TCP connection (iscsi endpoint) handle 2118 * 2119 * executes TCP connection teardown process 2120 */ 2121 static void bnx2i_ep_disconnect(struct iscsi_endpoint *ep) 2122 { 2123 struct bnx2i_endpoint *bnx2i_ep; 2124 struct bnx2i_conn *bnx2i_conn = NULL; 2125 struct bnx2i_hba *hba; 2126 2127 bnx2i_ep = ep->dd_data; 2128 2129 /* driver should not attempt connection cleanup until TCP_CONNECT 2130 * completes either successfully or fails. Timeout is 9-secs, so 2131 * wait for it to complete 2132 */ 2133 while ((bnx2i_ep->state == EP_STATE_CONNECT_START) && 2134 !time_after(jiffies, bnx2i_ep->timestamp + (12 * HZ))) 2135 msleep(250); 2136 2137 if (bnx2i_ep->conn) 2138 bnx2i_conn = bnx2i_ep->conn; 2139 hba = bnx2i_ep->hba; 2140 2141 mutex_lock(&hba->net_dev_lock); 2142 2143 if (bnx2i_ep->state == EP_STATE_DISCONN_TIMEDOUT) 2144 goto out; 2145 2146 if (bnx2i_ep->state == EP_STATE_IDLE) 2147 goto free_resc; 2148 2149 if (!test_bit(ADAPTER_STATE_UP, &hba->adapter_state) || 2150 (bnx2i_ep->hba_age != hba->age)) { 2151 bnx2i_ep_active_list_del(hba, bnx2i_ep); 2152 goto free_resc; 2153 } 2154 2155 /* Do all chip cleanup here */ 2156 if (bnx2i_hw_ep_disconnect(bnx2i_ep)) { 2157 mutex_unlock(&hba->net_dev_lock); 2158 return; 2159 } 2160 free_resc: 2161 bnx2i_free_qp_resc(hba, bnx2i_ep); 2162 2163 if (bnx2i_conn) 2164 bnx2i_conn->ep = NULL; 2165 2166 bnx2i_free_ep(ep); 2167 out: 2168 mutex_unlock(&hba->net_dev_lock); 2169 2170 wake_up_interruptible(&hba->eh_wait); 2171 } 2172 2173 2174 /** 2175 * bnx2i_nl_set_path - ISCSI_UEVENT_PATH_UPDATE user message handler 2176 * @shost: scsi host pointer 2177 * @params: pointer to buffer containing iscsi path message 2178 */ 2179 static int bnx2i_nl_set_path(struct Scsi_Host *shost, struct iscsi_path *params) 2180 { 2181 struct bnx2i_hba *hba = iscsi_host_priv(shost); 2182 char *buf = (char *) params; 2183 u16 len = sizeof(*params); 2184 2185 /* handled by cnic driver */ 2186 hba->cnic->iscsi_nl_msg_recv(hba->cnic, ISCSI_UEVENT_PATH_UPDATE, buf, 2187 len); 2188 2189 return 0; 2190 } 2191 2192 static umode_t bnx2i_attr_is_visible(int param_type, int param) 2193 { 2194 switch (param_type) { 2195 case ISCSI_HOST_PARAM: 2196 switch (param) { 2197 case ISCSI_HOST_PARAM_NETDEV_NAME: 2198 case ISCSI_HOST_PARAM_HWADDRESS: 2199 case ISCSI_HOST_PARAM_IPADDRESS: 2200 return S_IRUGO; 2201 default: 2202 return 0; 2203 } 2204 case ISCSI_PARAM: 2205 switch (param) { 2206 case ISCSI_PARAM_MAX_RECV_DLENGTH: 2207 case ISCSI_PARAM_MAX_XMIT_DLENGTH: 2208 case ISCSI_PARAM_HDRDGST_EN: 2209 case ISCSI_PARAM_DATADGST_EN: 2210 case ISCSI_PARAM_CONN_ADDRESS: 2211 case ISCSI_PARAM_CONN_PORT: 2212 case ISCSI_PARAM_EXP_STATSN: 2213 case ISCSI_PARAM_PERSISTENT_ADDRESS: 2214 case ISCSI_PARAM_PERSISTENT_PORT: 2215 case ISCSI_PARAM_PING_TMO: 2216 case ISCSI_PARAM_RECV_TMO: 2217 case ISCSI_PARAM_INITIAL_R2T_EN: 2218 case ISCSI_PARAM_MAX_R2T: 2219 case ISCSI_PARAM_IMM_DATA_EN: 2220 case ISCSI_PARAM_FIRST_BURST: 2221 case ISCSI_PARAM_MAX_BURST: 2222 case ISCSI_PARAM_PDU_INORDER_EN: 2223 case ISCSI_PARAM_DATASEQ_INORDER_EN: 2224 case ISCSI_PARAM_ERL: 2225 case ISCSI_PARAM_TARGET_NAME: 2226 case ISCSI_PARAM_TPGT: 2227 case ISCSI_PARAM_USERNAME: 2228 case ISCSI_PARAM_PASSWORD: 2229 case ISCSI_PARAM_USERNAME_IN: 2230 case ISCSI_PARAM_PASSWORD_IN: 2231 case ISCSI_PARAM_FAST_ABORT: 2232 case ISCSI_PARAM_ABORT_TMO: 2233 case ISCSI_PARAM_LU_RESET_TMO: 2234 case ISCSI_PARAM_TGT_RESET_TMO: 2235 case ISCSI_PARAM_IFACE_NAME: 2236 case ISCSI_PARAM_INITIATOR_NAME: 2237 case ISCSI_PARAM_BOOT_ROOT: 2238 case ISCSI_PARAM_BOOT_NIC: 2239 case ISCSI_PARAM_BOOT_TARGET: 2240 return S_IRUGO; 2241 default: 2242 return 0; 2243 } 2244 } 2245 2246 return 0; 2247 } 2248 2249 /* 2250 * 'Scsi_Host_Template' structure and 'iscsi_tranport' structure template 2251 * used while registering with the scsi host and iSCSI transport module. 2252 */ 2253 static struct scsi_host_template bnx2i_host_template = { 2254 .module = THIS_MODULE, 2255 .name = "QLogic Offload iSCSI Initiator", 2256 .proc_name = "bnx2i", 2257 .queuecommand = iscsi_queuecommand, 2258 .eh_timed_out = iscsi_eh_cmd_timed_out, 2259 .eh_abort_handler = iscsi_eh_abort, 2260 .eh_device_reset_handler = iscsi_eh_device_reset, 2261 .eh_target_reset_handler = iscsi_eh_recover_target, 2262 .change_queue_depth = scsi_change_queue_depth, 2263 .target_alloc = iscsi_target_alloc, 2264 .can_queue = 2048, 2265 .max_sectors = 127, 2266 .cmd_per_lun = 128, 2267 .this_id = -1, 2268 .sg_tablesize = ISCSI_MAX_BDS_PER_CMD, 2269 .shost_groups = bnx2i_dev_groups, 2270 .track_queue_depth = 1, 2271 .cmd_size = sizeof(struct iscsi_cmd), 2272 }; 2273 2274 struct iscsi_transport bnx2i_iscsi_transport = { 2275 .owner = THIS_MODULE, 2276 .name = "bnx2i", 2277 .caps = CAP_RECOVERY_L0 | CAP_HDRDGST | 2278 CAP_MULTI_R2T | CAP_DATADGST | 2279 CAP_DATA_PATH_OFFLOAD | 2280 CAP_TEXT_NEGO, 2281 .create_session = bnx2i_session_create, 2282 .destroy_session = bnx2i_session_destroy, 2283 .create_conn = bnx2i_conn_create, 2284 .bind_conn = bnx2i_conn_bind, 2285 .unbind_conn = iscsi_conn_unbind, 2286 .destroy_conn = bnx2i_conn_destroy, 2287 .attr_is_visible = bnx2i_attr_is_visible, 2288 .set_param = iscsi_set_param, 2289 .get_conn_param = iscsi_conn_get_param, 2290 .get_session_param = iscsi_session_get_param, 2291 .get_host_param = bnx2i_host_get_param, 2292 .start_conn = bnx2i_conn_start, 2293 .stop_conn = iscsi_conn_stop, 2294 .send_pdu = iscsi_conn_send_pdu, 2295 .xmit_task = bnx2i_task_xmit, 2296 .get_stats = bnx2i_conn_get_stats, 2297 /* TCP connect - disconnect - option-2 interface calls */ 2298 .get_ep_param = bnx2i_ep_get_param, 2299 .ep_connect = bnx2i_ep_connect, 2300 .ep_poll = bnx2i_ep_poll, 2301 .ep_disconnect = bnx2i_ep_disconnect, 2302 .set_path = bnx2i_nl_set_path, 2303 /* Error recovery timeout call */ 2304 .session_recovery_timedout = iscsi_session_recovery_timedout, 2305 .cleanup_task = bnx2i_cleanup_task, 2306 }; 2307