0001
0002
0003
0004
0005
0006
0007
0008
0009
0010
0011
0012
0013
0014
0015
0016
0017
0018 #include <linux/slab.h>
0019 #include <scsi/scsi_tcq.h>
0020 #include <scsi/libiscsi.h>
0021 #include "bnx2i.h"
0022
0023 struct scsi_transport_template *bnx2i_scsi_xport_template;
0024 struct iscsi_transport bnx2i_iscsi_transport;
0025 static struct scsi_host_template bnx2i_host_template;
0026
0027
0028
0029
0030 static DEFINE_SPINLOCK(bnx2i_resc_lock);
0031
0032 DECLARE_PER_CPU(struct bnx2i_percpu_s, bnx2i_percpu);
0033
0034 static int bnx2i_adapter_ready(struct bnx2i_hba *hba)
0035 {
0036 int retval = 0;
0037
0038 if (!hba || !test_bit(ADAPTER_STATE_UP, &hba->adapter_state) ||
0039 test_bit(ADAPTER_STATE_GOING_DOWN, &hba->adapter_state) ||
0040 test_bit(ADAPTER_STATE_LINK_DOWN, &hba->adapter_state))
0041 retval = -EPERM;
0042 return retval;
0043 }
0044
0045
0046
0047
0048
0049
0050
0051
0052
0053
0054
0055
0056 static void bnx2i_get_write_cmd_bd_idx(struct bnx2i_cmd *cmd, u32 buf_off,
0057 u32 *start_bd_off, u32 *start_bd_idx)
0058 {
0059 struct iscsi_bd *bd_tbl = cmd->io_tbl.bd_tbl;
0060 u32 cur_offset = 0;
0061 u32 cur_bd_idx = 0;
0062
0063 if (buf_off) {
0064 while (buf_off >= (cur_offset + bd_tbl->buffer_length)) {
0065 cur_offset += bd_tbl->buffer_length;
0066 cur_bd_idx++;
0067 bd_tbl++;
0068 }
0069 }
0070
0071 *start_bd_off = buf_off - cur_offset;
0072 *start_bd_idx = cur_bd_idx;
0073 }
0074
0075
0076
0077
0078
0079
0080
0081
0082
0083
0084 static void bnx2i_setup_write_cmd_bd_info(struct iscsi_task *task)
0085 {
0086 struct bnx2i_cmd *cmd = task->dd_data;
0087 u32 start_bd_offset;
0088 u32 start_bd_idx;
0089 u32 buffer_offset = 0;
0090 u32 cmd_len = cmd->req.total_data_transfer_length;
0091
0092
0093
0094
0095 if (!iscsi_task_has_unsol_data(task) && !task->imm_count)
0096 return;
0097
0098
0099 buffer_offset += task->imm_count;
0100 if (task->imm_count == cmd_len)
0101 return;
0102
0103 if (iscsi_task_has_unsol_data(task)) {
0104 bnx2i_get_write_cmd_bd_idx(cmd, buffer_offset,
0105 &start_bd_offset, &start_bd_idx);
0106 cmd->req.ud_buffer_offset = start_bd_offset;
0107 cmd->req.ud_start_bd_index = start_bd_idx;
0108 buffer_offset += task->unsol_r2t.data_length;
0109 }
0110
0111 if (buffer_offset != cmd_len) {
0112 bnx2i_get_write_cmd_bd_idx(cmd, buffer_offset,
0113 &start_bd_offset, &start_bd_idx);
0114 if ((start_bd_offset > task->conn->session->first_burst) ||
0115 (start_bd_idx > scsi_sg_count(cmd->scsi_cmd))) {
0116 int i = 0;
0117
0118 iscsi_conn_printk(KERN_ALERT, task->conn,
0119 "bnx2i- error, buf offset 0x%x "
0120 "bd_valid %d use_sg %d\n",
0121 buffer_offset, cmd->io_tbl.bd_valid,
0122 scsi_sg_count(cmd->scsi_cmd));
0123 for (i = 0; i < cmd->io_tbl.bd_valid; i++)
0124 iscsi_conn_printk(KERN_ALERT, task->conn,
0125 "bnx2i err, bd[%d]: len %x\n",
0126 i, cmd->io_tbl.bd_tbl[i].\
0127 buffer_length);
0128 }
0129 cmd->req.sd_buffer_offset = start_bd_offset;
0130 cmd->req.sd_start_bd_index = start_bd_idx;
0131 }
0132 }
0133
0134
0135
0136
0137
0138
0139
0140
0141
0142
0143 static int bnx2i_map_scsi_sg(struct bnx2i_hba *hba, struct bnx2i_cmd *cmd)
0144 {
0145 struct scsi_cmnd *sc = cmd->scsi_cmd;
0146 struct iscsi_bd *bd = cmd->io_tbl.bd_tbl;
0147 struct scatterlist *sg;
0148 int byte_count = 0;
0149 int bd_count = 0;
0150 int sg_count;
0151 int sg_len;
0152 u64 addr;
0153 int i;
0154
0155 BUG_ON(scsi_sg_count(sc) > ISCSI_MAX_BDS_PER_CMD);
0156
0157 sg_count = scsi_dma_map(sc);
0158
0159 scsi_for_each_sg(sc, sg, sg_count, i) {
0160 sg_len = sg_dma_len(sg);
0161 addr = (u64) sg_dma_address(sg);
0162 bd[bd_count].buffer_addr_lo = addr & 0xffffffff;
0163 bd[bd_count].buffer_addr_hi = addr >> 32;
0164 bd[bd_count].buffer_length = sg_len;
0165 bd[bd_count].flags = 0;
0166 if (bd_count == 0)
0167 bd[bd_count].flags = ISCSI_BD_FIRST_IN_BD_CHAIN;
0168
0169 byte_count += sg_len;
0170 bd_count++;
0171 }
0172
0173 if (bd_count)
0174 bd[bd_count - 1].flags |= ISCSI_BD_LAST_IN_BD_CHAIN;
0175
0176 BUG_ON(byte_count != scsi_bufflen(sc));
0177 return bd_count;
0178 }
0179
0180
0181
0182
0183
0184
0185
0186 static void bnx2i_iscsi_map_sg_list(struct bnx2i_cmd *cmd)
0187 {
0188 int bd_count;
0189
0190 bd_count = bnx2i_map_scsi_sg(cmd->conn->hba, cmd);
0191 if (!bd_count) {
0192 struct iscsi_bd *bd = cmd->io_tbl.bd_tbl;
0193
0194 bd[0].buffer_addr_lo = bd[0].buffer_addr_hi = 0;
0195 bd[0].buffer_length = bd[0].flags = 0;
0196 }
0197 cmd->io_tbl.bd_valid = bd_count;
0198 }
0199
0200
0201
0202
0203
0204
0205
0206
0207 void bnx2i_iscsi_unmap_sg_list(struct bnx2i_cmd *cmd)
0208 {
0209 struct scsi_cmnd *sc = cmd->scsi_cmd;
0210
0211 if (cmd->io_tbl.bd_valid && sc) {
0212 scsi_dma_unmap(sc);
0213 cmd->io_tbl.bd_valid = 0;
0214 }
0215 }
0216
0217 static void bnx2i_setup_cmd_wqe_template(struct bnx2i_cmd *cmd)
0218 {
0219 memset(&cmd->req, 0x00, sizeof(cmd->req));
0220 cmd->req.op_code = 0xFF;
0221 cmd->req.bd_list_addr_lo = (u32) cmd->io_tbl.bd_tbl_dma;
0222 cmd->req.bd_list_addr_hi =
0223 (u32) ((u64) cmd->io_tbl.bd_tbl_dma >> 32);
0224
0225 }
0226
0227
0228
0229
0230
0231
0232
0233
0234
0235
0236
0237
0238 static int bnx2i_bind_conn_to_iscsi_cid(struct bnx2i_hba *hba,
0239 struct bnx2i_conn *bnx2i_conn,
0240 u32 iscsi_cid)
0241 {
0242 if (hba && hba->cid_que.conn_cid_tbl[iscsi_cid]) {
0243 iscsi_conn_printk(KERN_ALERT, bnx2i_conn->cls_conn->dd_data,
0244 "conn bind - entry #%d not free\n", iscsi_cid);
0245 return -EBUSY;
0246 }
0247
0248 hba->cid_que.conn_cid_tbl[iscsi_cid] = bnx2i_conn;
0249 return 0;
0250 }
0251
0252
0253
0254
0255
0256
0257
0258 struct bnx2i_conn *bnx2i_get_conn_from_id(struct bnx2i_hba *hba,
0259 u16 iscsi_cid)
0260 {
0261 if (!hba->cid_que.conn_cid_tbl) {
0262 printk(KERN_ERR "bnx2i: ERROR - missing conn<->cid table\n");
0263 return NULL;
0264
0265 } else if (iscsi_cid >= hba->max_active_conns) {
0266 printk(KERN_ERR "bnx2i: wrong cid #%d\n", iscsi_cid);
0267 return NULL;
0268 }
0269 return hba->cid_que.conn_cid_tbl[iscsi_cid];
0270 }
0271
0272
0273
0274
0275
0276
0277 static u32 bnx2i_alloc_iscsi_cid(struct bnx2i_hba *hba)
0278 {
0279 int idx;
0280
0281 if (!hba->cid_que.cid_free_cnt)
0282 return -1;
0283
0284 idx = hba->cid_que.cid_q_cons_idx;
0285 hba->cid_que.cid_q_cons_idx++;
0286 if (hba->cid_que.cid_q_cons_idx == hba->cid_que.cid_q_max_idx)
0287 hba->cid_que.cid_q_cons_idx = 0;
0288
0289 hba->cid_que.cid_free_cnt--;
0290 return hba->cid_que.cid_que[idx];
0291 }
0292
0293
0294
0295
0296
0297
0298
0299 static void bnx2i_free_iscsi_cid(struct bnx2i_hba *hba, u16 iscsi_cid)
0300 {
0301 int idx;
0302
0303 if (iscsi_cid == (u16) -1)
0304 return;
0305
0306 hba->cid_que.cid_free_cnt++;
0307
0308 idx = hba->cid_que.cid_q_prod_idx;
0309 hba->cid_que.cid_que[idx] = iscsi_cid;
0310 hba->cid_que.conn_cid_tbl[iscsi_cid] = NULL;
0311 hba->cid_que.cid_q_prod_idx++;
0312 if (hba->cid_que.cid_q_prod_idx == hba->cid_que.cid_q_max_idx)
0313 hba->cid_que.cid_q_prod_idx = 0;
0314 }
0315
0316
0317
0318
0319
0320
0321
0322
0323
0324 static int bnx2i_setup_free_cid_que(struct bnx2i_hba *hba)
0325 {
0326 int mem_size;
0327 int i;
0328
0329 mem_size = hba->max_active_conns * sizeof(u32);
0330 mem_size = (mem_size + (PAGE_SIZE - 1)) & PAGE_MASK;
0331
0332 hba->cid_que.cid_que_base = kmalloc(mem_size, GFP_KERNEL);
0333 if (!hba->cid_que.cid_que_base)
0334 return -ENOMEM;
0335
0336 mem_size = hba->max_active_conns * sizeof(struct bnx2i_conn *);
0337 mem_size = (mem_size + (PAGE_SIZE - 1)) & PAGE_MASK;
0338 hba->cid_que.conn_cid_tbl = kmalloc(mem_size, GFP_KERNEL);
0339 if (!hba->cid_que.conn_cid_tbl) {
0340 kfree(hba->cid_que.cid_que_base);
0341 hba->cid_que.cid_que_base = NULL;
0342 return -ENOMEM;
0343 }
0344
0345 hba->cid_que.cid_que = (u32 *)hba->cid_que.cid_que_base;
0346 hba->cid_que.cid_q_prod_idx = 0;
0347 hba->cid_que.cid_q_cons_idx = 0;
0348 hba->cid_que.cid_q_max_idx = hba->max_active_conns;
0349 hba->cid_que.cid_free_cnt = hba->max_active_conns;
0350
0351 for (i = 0; i < hba->max_active_conns; i++) {
0352 hba->cid_que.cid_que[i] = i;
0353 hba->cid_que.conn_cid_tbl[i] = NULL;
0354 }
0355 return 0;
0356 }
0357
0358
0359
0360
0361
0362
0363 static void bnx2i_release_free_cid_que(struct bnx2i_hba *hba)
0364 {
0365 kfree(hba->cid_que.cid_que_base);
0366 hba->cid_que.cid_que_base = NULL;
0367
0368 kfree(hba->cid_que.conn_cid_tbl);
0369 hba->cid_que.conn_cid_tbl = NULL;
0370 }
0371
0372
0373
0374
0375
0376
0377
0378
0379
0380
0381 static struct iscsi_endpoint *bnx2i_alloc_ep(struct bnx2i_hba *hba)
0382 {
0383 struct iscsi_endpoint *ep;
0384 struct bnx2i_endpoint *bnx2i_ep;
0385 u32 ec_div;
0386
0387 ep = iscsi_create_endpoint(sizeof(*bnx2i_ep));
0388 if (!ep) {
0389 printk(KERN_ERR "bnx2i: Could not allocate ep\n");
0390 return NULL;
0391 }
0392
0393 bnx2i_ep = ep->dd_data;
0394 bnx2i_ep->cls_ep = ep;
0395 INIT_LIST_HEAD(&bnx2i_ep->link);
0396 bnx2i_ep->state = EP_STATE_IDLE;
0397 bnx2i_ep->ep_iscsi_cid = (u16) -1;
0398 bnx2i_ep->hba = hba;
0399 bnx2i_ep->hba_age = hba->age;
0400
0401 ec_div = event_coal_div;
0402 while (ec_div >>= 1)
0403 bnx2i_ep->ec_shift += 1;
0404
0405 hba->ofld_conns_active++;
0406 init_waitqueue_head(&bnx2i_ep->ofld_wait);
0407 return ep;
0408 }
0409
0410
0411
0412
0413
0414
0415 static void bnx2i_free_ep(struct iscsi_endpoint *ep)
0416 {
0417 struct bnx2i_endpoint *bnx2i_ep = ep->dd_data;
0418 unsigned long flags;
0419
0420 spin_lock_irqsave(&bnx2i_resc_lock, flags);
0421 bnx2i_ep->state = EP_STATE_IDLE;
0422 bnx2i_ep->hba->ofld_conns_active--;
0423
0424 if (bnx2i_ep->ep_iscsi_cid != (u16) -1)
0425 bnx2i_free_iscsi_cid(bnx2i_ep->hba, bnx2i_ep->ep_iscsi_cid);
0426
0427 if (bnx2i_ep->conn) {
0428 bnx2i_ep->conn->ep = NULL;
0429 bnx2i_ep->conn = NULL;
0430 }
0431
0432 bnx2i_ep->hba = NULL;
0433 spin_unlock_irqrestore(&bnx2i_resc_lock, flags);
0434 iscsi_destroy_endpoint(ep);
0435 }
0436
0437
0438
0439
0440
0441
0442
0443
0444 static int bnx2i_alloc_bdt(struct bnx2i_hba *hba, struct iscsi_session *session,
0445 struct bnx2i_cmd *cmd)
0446 {
0447 struct io_bdt *io = &cmd->io_tbl;
0448 struct iscsi_bd *bd;
0449
0450 io->bd_tbl = dma_alloc_coherent(&hba->pcidev->dev,
0451 ISCSI_MAX_BDS_PER_CMD * sizeof(*bd),
0452 &io->bd_tbl_dma, GFP_KERNEL);
0453 if (!io->bd_tbl) {
0454 iscsi_session_printk(KERN_ERR, session, "Could not "
0455 "allocate bdt.\n");
0456 return -ENOMEM;
0457 }
0458 io->bd_valid = 0;
0459 return 0;
0460 }
0461
0462
0463
0464
0465
0466
0467 static void bnx2i_destroy_cmd_pool(struct bnx2i_hba *hba,
0468 struct iscsi_session *session)
0469 {
0470 int i;
0471
0472 for (i = 0; i < session->cmds_max; i++) {
0473 struct iscsi_task *task = session->cmds[i];
0474 struct bnx2i_cmd *cmd = task->dd_data;
0475
0476 if (cmd->io_tbl.bd_tbl)
0477 dma_free_coherent(&hba->pcidev->dev,
0478 ISCSI_MAX_BDS_PER_CMD *
0479 sizeof(struct iscsi_bd),
0480 cmd->io_tbl.bd_tbl,
0481 cmd->io_tbl.bd_tbl_dma);
0482 }
0483
0484 }
0485
0486
0487
0488
0489
0490
0491
0492 static int bnx2i_setup_cmd_pool(struct bnx2i_hba *hba,
0493 struct iscsi_session *session)
0494 {
0495 int i;
0496
0497 for (i = 0; i < session->cmds_max; i++) {
0498 struct iscsi_task *task = session->cmds[i];
0499 struct bnx2i_cmd *cmd = task->dd_data;
0500
0501 task->hdr = &cmd->hdr;
0502 task->hdr_max = sizeof(struct iscsi_hdr);
0503
0504 if (bnx2i_alloc_bdt(hba, session, cmd))
0505 goto free_bdts;
0506 }
0507
0508 return 0;
0509
0510 free_bdts:
0511 bnx2i_destroy_cmd_pool(hba, session);
0512 return -ENOMEM;
0513 }
0514
0515
0516
0517
0518
0519
0520
0521
0522
0523 static int bnx2i_setup_mp_bdt(struct bnx2i_hba *hba)
0524 {
0525 int rc = 0;
0526 struct iscsi_bd *mp_bdt;
0527 u64 addr;
0528
0529 hba->mp_bd_tbl = dma_alloc_coherent(&hba->pcidev->dev, CNIC_PAGE_SIZE,
0530 &hba->mp_bd_dma, GFP_KERNEL);
0531 if (!hba->mp_bd_tbl) {
0532 printk(KERN_ERR "unable to allocate Middle Path BDT\n");
0533 rc = -1;
0534 goto out;
0535 }
0536
0537 hba->dummy_buffer = dma_alloc_coherent(&hba->pcidev->dev,
0538 CNIC_PAGE_SIZE,
0539 &hba->dummy_buf_dma, GFP_KERNEL);
0540 if (!hba->dummy_buffer) {
0541 printk(KERN_ERR "unable to alloc Middle Path Dummy Buffer\n");
0542 dma_free_coherent(&hba->pcidev->dev, CNIC_PAGE_SIZE,
0543 hba->mp_bd_tbl, hba->mp_bd_dma);
0544 hba->mp_bd_tbl = NULL;
0545 rc = -1;
0546 goto out;
0547 }
0548
0549 mp_bdt = (struct iscsi_bd *) hba->mp_bd_tbl;
0550 addr = (unsigned long) hba->dummy_buf_dma;
0551 mp_bdt->buffer_addr_lo = addr & 0xffffffff;
0552 mp_bdt->buffer_addr_hi = addr >> 32;
0553 mp_bdt->buffer_length = CNIC_PAGE_SIZE;
0554 mp_bdt->flags = ISCSI_BD_LAST_IN_BD_CHAIN |
0555 ISCSI_BD_FIRST_IN_BD_CHAIN;
0556 out:
0557 return rc;
0558 }
0559
0560
0561
0562
0563
0564
0565
0566
0567 static void bnx2i_free_mp_bdt(struct bnx2i_hba *hba)
0568 {
0569 if (hba->mp_bd_tbl) {
0570 dma_free_coherent(&hba->pcidev->dev, CNIC_PAGE_SIZE,
0571 hba->mp_bd_tbl, hba->mp_bd_dma);
0572 hba->mp_bd_tbl = NULL;
0573 }
0574 if (hba->dummy_buffer) {
0575 dma_free_coherent(&hba->pcidev->dev, CNIC_PAGE_SIZE,
0576 hba->dummy_buffer, hba->dummy_buf_dma);
0577 hba->dummy_buffer = NULL;
0578 }
0579 return;
0580 }
0581
0582
0583
0584
0585
0586
0587
0588
0589
0590
0591
0592 void bnx2i_drop_session(struct iscsi_cls_session *cls_session)
0593 {
0594 iscsi_session_failure(cls_session->dd_data, ISCSI_ERR_CONN_FAILED);
0595 }
0596
0597
0598
0599
0600
0601
0602
0603
0604 static int bnx2i_ep_destroy_list_add(struct bnx2i_hba *hba,
0605 struct bnx2i_endpoint *ep)
0606 {
0607 write_lock_bh(&hba->ep_rdwr_lock);
0608 list_add_tail(&ep->link, &hba->ep_destroy_list);
0609 write_unlock_bh(&hba->ep_rdwr_lock);
0610 return 0;
0611 }
0612
0613
0614
0615
0616
0617
0618
0619
0620
0621 static int bnx2i_ep_destroy_list_del(struct bnx2i_hba *hba,
0622 struct bnx2i_endpoint *ep)
0623 {
0624 write_lock_bh(&hba->ep_rdwr_lock);
0625 list_del_init(&ep->link);
0626 write_unlock_bh(&hba->ep_rdwr_lock);
0627
0628 return 0;
0629 }
0630
0631
0632
0633
0634
0635
0636
0637
0638 static int bnx2i_ep_ofld_list_add(struct bnx2i_hba *hba,
0639 struct bnx2i_endpoint *ep)
0640 {
0641 write_lock_bh(&hba->ep_rdwr_lock);
0642 list_add_tail(&ep->link, &hba->ep_ofld_list);
0643 write_unlock_bh(&hba->ep_rdwr_lock);
0644 return 0;
0645 }
0646
0647
0648
0649
0650
0651
0652
0653
0654 static int bnx2i_ep_ofld_list_del(struct bnx2i_hba *hba,
0655 struct bnx2i_endpoint *ep)
0656 {
0657 write_lock_bh(&hba->ep_rdwr_lock);
0658 list_del_init(&ep->link);
0659 write_unlock_bh(&hba->ep_rdwr_lock);
0660 return 0;
0661 }
0662
0663
0664
0665
0666
0667
0668
0669
0670
0671 struct bnx2i_endpoint *
0672 bnx2i_find_ep_in_ofld_list(struct bnx2i_hba *hba, u32 iscsi_cid)
0673 {
0674 struct list_head *list;
0675 struct list_head *tmp;
0676 struct bnx2i_endpoint *ep = NULL;
0677
0678 read_lock_bh(&hba->ep_rdwr_lock);
0679 list_for_each_safe(list, tmp, &hba->ep_ofld_list) {
0680 ep = (struct bnx2i_endpoint *)list;
0681
0682 if (ep->ep_iscsi_cid == iscsi_cid)
0683 break;
0684 ep = NULL;
0685 }
0686 read_unlock_bh(&hba->ep_rdwr_lock);
0687
0688 if (!ep)
0689 printk(KERN_ERR "l5 cid %d not found\n", iscsi_cid);
0690 return ep;
0691 }
0692
0693
0694
0695
0696
0697
0698
0699 struct bnx2i_endpoint *
0700 bnx2i_find_ep_in_destroy_list(struct bnx2i_hba *hba, u32 iscsi_cid)
0701 {
0702 struct list_head *list;
0703 struct list_head *tmp;
0704 struct bnx2i_endpoint *ep = NULL;
0705
0706 read_lock_bh(&hba->ep_rdwr_lock);
0707 list_for_each_safe(list, tmp, &hba->ep_destroy_list) {
0708 ep = (struct bnx2i_endpoint *)list;
0709
0710 if (ep->ep_iscsi_cid == iscsi_cid)
0711 break;
0712 ep = NULL;
0713 }
0714 read_unlock_bh(&hba->ep_rdwr_lock);
0715
0716 if (!ep)
0717 printk(KERN_ERR "l5 cid %d not found\n", iscsi_cid);
0718
0719 return ep;
0720 }
0721
0722
0723
0724
0725
0726
0727
0728
0729 static void bnx2i_ep_active_list_add(struct bnx2i_hba *hba,
0730 struct bnx2i_endpoint *ep)
0731 {
0732 write_lock_bh(&hba->ep_rdwr_lock);
0733 list_add_tail(&ep->link, &hba->ep_active_list);
0734 write_unlock_bh(&hba->ep_rdwr_lock);
0735 }
0736
0737
0738
0739
0740
0741
0742
0743
0744
0745 static void bnx2i_ep_active_list_del(struct bnx2i_hba *hba,
0746 struct bnx2i_endpoint *ep)
0747 {
0748 write_lock_bh(&hba->ep_rdwr_lock);
0749 list_del_init(&ep->link);
0750 write_unlock_bh(&hba->ep_rdwr_lock);
0751 }
0752
0753
0754
0755
0756
0757
0758
0759
0760
0761
0762
0763 static void bnx2i_setup_host_queue_size(struct bnx2i_hba *hba,
0764 struct Scsi_Host *shost)
0765 {
0766 if (test_bit(BNX2I_NX2_DEV_5708, &hba->cnic_dev_type))
0767 shost->can_queue = ISCSI_MAX_CMDS_PER_HBA_5708;
0768 else if (test_bit(BNX2I_NX2_DEV_5709, &hba->cnic_dev_type))
0769 shost->can_queue = ISCSI_MAX_CMDS_PER_HBA_5709;
0770 else if (test_bit(BNX2I_NX2_DEV_57710, &hba->cnic_dev_type))
0771 shost->can_queue = ISCSI_MAX_CMDS_PER_HBA_57710;
0772 else
0773 shost->can_queue = ISCSI_MAX_CMDS_PER_HBA_5708;
0774 }
0775
0776
0777
0778
0779
0780
0781
0782
0783
0784 struct bnx2i_hba *bnx2i_alloc_hba(struct cnic_dev *cnic)
0785 {
0786 struct Scsi_Host *shost;
0787 struct bnx2i_hba *hba;
0788
0789 shost = iscsi_host_alloc(&bnx2i_host_template, sizeof(*hba), 0);
0790 if (!shost)
0791 return NULL;
0792 shost->dma_boundary = cnic->pcidev->dma_mask;
0793 shost->transportt = bnx2i_scsi_xport_template;
0794 shost->max_id = ISCSI_MAX_CONNS_PER_HBA - 1;
0795 shost->max_channel = 0;
0796 shost->max_lun = 512;
0797 shost->max_cmd_len = 16;
0798
0799 hba = iscsi_host_priv(shost);
0800 hba->shost = shost;
0801 hba->netdev = cnic->netdev;
0802
0803 hba->pcidev = cnic->pcidev;
0804 pci_dev_get(hba->pcidev);
0805 hba->pci_did = hba->pcidev->device;
0806 hba->pci_vid = hba->pcidev->vendor;
0807 hba->pci_sdid = hba->pcidev->subsystem_device;
0808 hba->pci_svid = hba->pcidev->subsystem_vendor;
0809 hba->pci_func = PCI_FUNC(hba->pcidev->devfn);
0810 hba->pci_devno = PCI_SLOT(hba->pcidev->devfn);
0811
0812 bnx2i_identify_device(hba, cnic);
0813 bnx2i_setup_host_queue_size(hba, shost);
0814
0815 hba->reg_base = pci_resource_start(hba->pcidev, 0);
0816 if (test_bit(BNX2I_NX2_DEV_5709, &hba->cnic_dev_type)) {
0817 hba->regview = pci_iomap(hba->pcidev, 0, BNX2_MQ_CONFIG2);
0818 if (!hba->regview)
0819 goto ioreg_map_err;
0820 } else if (test_bit(BNX2I_NX2_DEV_57710, &hba->cnic_dev_type)) {
0821 hba->regview = pci_iomap(hba->pcidev, 0, 4096);
0822 if (!hba->regview)
0823 goto ioreg_map_err;
0824 }
0825
0826 if (bnx2i_setup_mp_bdt(hba))
0827 goto mp_bdt_mem_err;
0828
0829 INIT_LIST_HEAD(&hba->ep_ofld_list);
0830 INIT_LIST_HEAD(&hba->ep_active_list);
0831 INIT_LIST_HEAD(&hba->ep_destroy_list);
0832 rwlock_init(&hba->ep_rdwr_lock);
0833
0834 hba->mtu_supported = BNX2I_MAX_MTU_SUPPORTED;
0835
0836
0837 hba->max_active_conns = ISCSI_MAX_CONNS_PER_HBA;
0838
0839 if (bnx2i_setup_free_cid_que(hba))
0840 goto cid_que_err;
0841
0842
0843 if (test_bit(BNX2I_NX2_DEV_57710, &hba->cnic_dev_type)) {
0844 if (sq_size && sq_size <= BNX2I_5770X_SQ_WQES_MAX)
0845 hba->max_sqes = sq_size;
0846 else
0847 hba->max_sqes = BNX2I_5770X_SQ_WQES_DEFAULT;
0848 } else {
0849 if (sq_size && sq_size <= BNX2I_570X_SQ_WQES_MAX)
0850 hba->max_sqes = sq_size;
0851 else
0852 hba->max_sqes = BNX2I_570X_SQ_WQES_DEFAULT;
0853 }
0854
0855 hba->max_rqes = rq_size;
0856 hba->max_cqes = hba->max_sqes + rq_size;
0857 if (test_bit(BNX2I_NX2_DEV_57710, &hba->cnic_dev_type)) {
0858 if (hba->max_cqes > BNX2I_5770X_CQ_WQES_MAX)
0859 hba->max_cqes = BNX2I_5770X_CQ_WQES_MAX;
0860 } else if (hba->max_cqes > BNX2I_570X_CQ_WQES_MAX)
0861 hba->max_cqes = BNX2I_570X_CQ_WQES_MAX;
0862
0863 hba->num_ccell = hba->max_sqes / 2;
0864
0865 spin_lock_init(&hba->lock);
0866 mutex_init(&hba->net_dev_lock);
0867 init_waitqueue_head(&hba->eh_wait);
0868 if (test_bit(BNX2I_NX2_DEV_57710, &hba->cnic_dev_type)) {
0869 hba->hba_shutdown_tmo = 30 * HZ;
0870 hba->conn_teardown_tmo = 20 * HZ;
0871 hba->conn_ctx_destroy_tmo = 6 * HZ;
0872 } else {
0873 hba->hba_shutdown_tmo = 20 * HZ;
0874 hba->conn_teardown_tmo = 10 * HZ;
0875 hba->conn_ctx_destroy_tmo = 2 * HZ;
0876 }
0877
0878 #ifdef CONFIG_32BIT
0879 spin_lock_init(&hba->stat_lock);
0880 #endif
0881 memset(&hba->stats, 0, sizeof(struct iscsi_stats_info));
0882
0883 if (iscsi_host_add(shost, &hba->pcidev->dev))
0884 goto free_dump_mem;
0885 return hba;
0886
0887 free_dump_mem:
0888 bnx2i_release_free_cid_que(hba);
0889 cid_que_err:
0890 bnx2i_free_mp_bdt(hba);
0891 mp_bdt_mem_err:
0892 if (hba->regview) {
0893 pci_iounmap(hba->pcidev, hba->regview);
0894 hba->regview = NULL;
0895 }
0896 ioreg_map_err:
0897 pci_dev_put(hba->pcidev);
0898 scsi_host_put(shost);
0899 return NULL;
0900 }
0901
0902
0903
0904
0905
0906
0907
0908 void bnx2i_free_hba(struct bnx2i_hba *hba)
0909 {
0910 struct Scsi_Host *shost = hba->shost;
0911
0912 iscsi_host_remove(shost, false);
0913 INIT_LIST_HEAD(&hba->ep_ofld_list);
0914 INIT_LIST_HEAD(&hba->ep_active_list);
0915 INIT_LIST_HEAD(&hba->ep_destroy_list);
0916
0917 if (hba->regview) {
0918 pci_iounmap(hba->pcidev, hba->regview);
0919 hba->regview = NULL;
0920 }
0921 pci_dev_put(hba->pcidev);
0922 bnx2i_free_mp_bdt(hba);
0923 bnx2i_release_free_cid_que(hba);
0924 iscsi_host_free(shost);
0925 }
0926
0927
0928
0929
0930
0931
0932
0933
0934 static void bnx2i_conn_free_login_resources(struct bnx2i_hba *hba,
0935 struct bnx2i_conn *bnx2i_conn)
0936 {
0937 if (bnx2i_conn->gen_pdu.resp_bd_tbl) {
0938 dma_free_coherent(&hba->pcidev->dev, CNIC_PAGE_SIZE,
0939 bnx2i_conn->gen_pdu.resp_bd_tbl,
0940 bnx2i_conn->gen_pdu.resp_bd_dma);
0941 bnx2i_conn->gen_pdu.resp_bd_tbl = NULL;
0942 }
0943
0944 if (bnx2i_conn->gen_pdu.req_bd_tbl) {
0945 dma_free_coherent(&hba->pcidev->dev, CNIC_PAGE_SIZE,
0946 bnx2i_conn->gen_pdu.req_bd_tbl,
0947 bnx2i_conn->gen_pdu.req_bd_dma);
0948 bnx2i_conn->gen_pdu.req_bd_tbl = NULL;
0949 }
0950
0951 if (bnx2i_conn->gen_pdu.resp_buf) {
0952 dma_free_coherent(&hba->pcidev->dev,
0953 ISCSI_DEF_MAX_RECV_SEG_LEN,
0954 bnx2i_conn->gen_pdu.resp_buf,
0955 bnx2i_conn->gen_pdu.resp_dma_addr);
0956 bnx2i_conn->gen_pdu.resp_buf = NULL;
0957 }
0958
0959 if (bnx2i_conn->gen_pdu.req_buf) {
0960 dma_free_coherent(&hba->pcidev->dev,
0961 ISCSI_DEF_MAX_RECV_SEG_LEN,
0962 bnx2i_conn->gen_pdu.req_buf,
0963 bnx2i_conn->gen_pdu.req_dma_addr);
0964 bnx2i_conn->gen_pdu.req_buf = NULL;
0965 }
0966 }
0967
0968
0969
0970
0971
0972
0973
0974
0975 static int bnx2i_conn_alloc_login_resources(struct bnx2i_hba *hba,
0976 struct bnx2i_conn *bnx2i_conn)
0977 {
0978
0979 bnx2i_conn->gen_pdu.req_buf =
0980 dma_alloc_coherent(&hba->pcidev->dev,
0981 ISCSI_DEF_MAX_RECV_SEG_LEN,
0982 &bnx2i_conn->gen_pdu.req_dma_addr,
0983 GFP_KERNEL);
0984 if (bnx2i_conn->gen_pdu.req_buf == NULL)
0985 goto login_req_buf_failure;
0986
0987 bnx2i_conn->gen_pdu.req_buf_size = 0;
0988 bnx2i_conn->gen_pdu.req_wr_ptr = bnx2i_conn->gen_pdu.req_buf;
0989
0990 bnx2i_conn->gen_pdu.resp_buf =
0991 dma_alloc_coherent(&hba->pcidev->dev,
0992 ISCSI_DEF_MAX_RECV_SEG_LEN,
0993 &bnx2i_conn->gen_pdu.resp_dma_addr,
0994 GFP_KERNEL);
0995 if (bnx2i_conn->gen_pdu.resp_buf == NULL)
0996 goto login_resp_buf_failure;
0997
0998 bnx2i_conn->gen_pdu.resp_buf_size = ISCSI_DEF_MAX_RECV_SEG_LEN;
0999 bnx2i_conn->gen_pdu.resp_wr_ptr = bnx2i_conn->gen_pdu.resp_buf;
1000
1001 bnx2i_conn->gen_pdu.req_bd_tbl =
1002 dma_alloc_coherent(&hba->pcidev->dev, CNIC_PAGE_SIZE,
1003 &bnx2i_conn->gen_pdu.req_bd_dma, GFP_KERNEL);
1004 if (bnx2i_conn->gen_pdu.req_bd_tbl == NULL)
1005 goto login_req_bd_tbl_failure;
1006
1007 bnx2i_conn->gen_pdu.resp_bd_tbl =
1008 dma_alloc_coherent(&hba->pcidev->dev, CNIC_PAGE_SIZE,
1009 &bnx2i_conn->gen_pdu.resp_bd_dma,
1010 GFP_KERNEL);
1011 if (bnx2i_conn->gen_pdu.resp_bd_tbl == NULL)
1012 goto login_resp_bd_tbl_failure;
1013
1014 return 0;
1015
1016 login_resp_bd_tbl_failure:
1017 dma_free_coherent(&hba->pcidev->dev, CNIC_PAGE_SIZE,
1018 bnx2i_conn->gen_pdu.req_bd_tbl,
1019 bnx2i_conn->gen_pdu.req_bd_dma);
1020 bnx2i_conn->gen_pdu.req_bd_tbl = NULL;
1021
1022 login_req_bd_tbl_failure:
1023 dma_free_coherent(&hba->pcidev->dev, ISCSI_DEF_MAX_RECV_SEG_LEN,
1024 bnx2i_conn->gen_pdu.resp_buf,
1025 bnx2i_conn->gen_pdu.resp_dma_addr);
1026 bnx2i_conn->gen_pdu.resp_buf = NULL;
1027 login_resp_buf_failure:
1028 dma_free_coherent(&hba->pcidev->dev, ISCSI_DEF_MAX_RECV_SEG_LEN,
1029 bnx2i_conn->gen_pdu.req_buf,
1030 bnx2i_conn->gen_pdu.req_dma_addr);
1031 bnx2i_conn->gen_pdu.req_buf = NULL;
1032 login_req_buf_failure:
1033 iscsi_conn_printk(KERN_ERR, bnx2i_conn->cls_conn->dd_data,
1034 "login resource alloc failed!!\n");
1035 return -ENOMEM;
1036
1037 }
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047 static void bnx2i_iscsi_prep_generic_pdu_bd(struct bnx2i_conn *bnx2i_conn)
1048 {
1049 struct iscsi_bd *bd_tbl;
1050
1051 bd_tbl = (struct iscsi_bd *) bnx2i_conn->gen_pdu.req_bd_tbl;
1052
1053 bd_tbl->buffer_addr_hi =
1054 (u32) ((u64) bnx2i_conn->gen_pdu.req_dma_addr >> 32);
1055 bd_tbl->buffer_addr_lo = (u32) bnx2i_conn->gen_pdu.req_dma_addr;
1056 bd_tbl->buffer_length = bnx2i_conn->gen_pdu.req_wr_ptr -
1057 bnx2i_conn->gen_pdu.req_buf;
1058 bd_tbl->reserved0 = 0;
1059 bd_tbl->flags = ISCSI_BD_LAST_IN_BD_CHAIN |
1060 ISCSI_BD_FIRST_IN_BD_CHAIN;
1061
1062 bd_tbl = (struct iscsi_bd *) bnx2i_conn->gen_pdu.resp_bd_tbl;
1063 bd_tbl->buffer_addr_hi = (u64) bnx2i_conn->gen_pdu.resp_dma_addr >> 32;
1064 bd_tbl->buffer_addr_lo = (u32) bnx2i_conn->gen_pdu.resp_dma_addr;
1065 bd_tbl->buffer_length = ISCSI_DEF_MAX_RECV_SEG_LEN;
1066 bd_tbl->reserved0 = 0;
1067 bd_tbl->flags = ISCSI_BD_LAST_IN_BD_CHAIN |
1068 ISCSI_BD_FIRST_IN_BD_CHAIN;
1069 }
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079 static int bnx2i_iscsi_send_generic_request(struct iscsi_task *task)
1080 {
1081 struct bnx2i_cmd *cmd = task->dd_data;
1082 struct bnx2i_conn *bnx2i_conn = cmd->conn;
1083 int rc = 0;
1084 char *buf;
1085 int data_len;
1086
1087 bnx2i_iscsi_prep_generic_pdu_bd(bnx2i_conn);
1088 switch (task->hdr->opcode & ISCSI_OPCODE_MASK) {
1089 case ISCSI_OP_LOGIN:
1090 bnx2i_send_iscsi_login(bnx2i_conn, task);
1091 break;
1092 case ISCSI_OP_NOOP_OUT:
1093 data_len = bnx2i_conn->gen_pdu.req_buf_size;
1094 buf = bnx2i_conn->gen_pdu.req_buf;
1095 if (data_len)
1096 rc = bnx2i_send_iscsi_nopout(bnx2i_conn, task,
1097 buf, data_len, 1);
1098 else
1099 rc = bnx2i_send_iscsi_nopout(bnx2i_conn, task,
1100 NULL, 0, 1);
1101 break;
1102 case ISCSI_OP_LOGOUT:
1103 rc = bnx2i_send_iscsi_logout(bnx2i_conn, task);
1104 break;
1105 case ISCSI_OP_SCSI_TMFUNC:
1106 rc = bnx2i_send_iscsi_tmf(bnx2i_conn, task);
1107 break;
1108 case ISCSI_OP_TEXT:
1109 rc = bnx2i_send_iscsi_text(bnx2i_conn, task);
1110 break;
1111 default:
1112 iscsi_conn_printk(KERN_ALERT, bnx2i_conn->cls_conn->dd_data,
1113 "send_gen: unsupported op 0x%x\n",
1114 task->hdr->opcode);
1115 }
1116 return rc;
1117 }
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129 static void bnx2i_cpy_scsi_cdb(struct scsi_cmnd *sc, struct bnx2i_cmd *cmd)
1130 {
1131 u32 dword;
1132 int lpcnt;
1133 u8 *srcp;
1134 u32 *dstp;
1135 u32 scsi_lun[2];
1136
1137 int_to_scsilun(sc->device->lun, (struct scsi_lun *) scsi_lun);
1138 cmd->req.lun[0] = be32_to_cpu(scsi_lun[0]);
1139 cmd->req.lun[1] = be32_to_cpu(scsi_lun[1]);
1140
1141 lpcnt = cmd->scsi_cmd->cmd_len / sizeof(dword);
1142 srcp = (u8 *) sc->cmnd;
1143 dstp = (u32 *) cmd->req.cdb;
1144 while (lpcnt--) {
1145 memcpy(&dword, (const void *) srcp, 4);
1146 *dstp = cpu_to_be32(dword);
1147 srcp += 4;
1148 dstp++;
1149 }
1150 if (sc->cmd_len & 0x3) {
1151 dword = (u32) srcp[0] | ((u32) srcp[1] << 8);
1152 *dstp = cpu_to_be32(dword);
1153 }
1154 }
1155
1156 static void bnx2i_cleanup_task(struct iscsi_task *task)
1157 {
1158 struct iscsi_conn *conn = task->conn;
1159 struct bnx2i_conn *bnx2i_conn = conn->dd_data;
1160 struct bnx2i_hba *hba = bnx2i_conn->hba;
1161
1162
1163
1164
1165 if (!task->sc || task->state == ISCSI_TASK_PENDING)
1166 return;
1167
1168
1169
1170 if (task->state == ISCSI_TASK_ABRT_TMF) {
1171 bnx2i_send_cmd_cleanup_req(hba, task->dd_data);
1172
1173 spin_unlock_bh(&conn->session->back_lock);
1174 wait_for_completion_timeout(&bnx2i_conn->cmd_cleanup_cmpl,
1175 msecs_to_jiffies(ISCSI_CMD_CLEANUP_TIMEOUT));
1176 spin_lock_bh(&conn->session->back_lock);
1177 }
1178 bnx2i_iscsi_unmap_sg_list(task->dd_data);
1179 }
1180
1181
1182
1183
1184
1185
1186 static int
1187 bnx2i_mtask_xmit(struct iscsi_conn *conn, struct iscsi_task *task)
1188 {
1189 struct bnx2i_conn *bnx2i_conn = conn->dd_data;
1190 struct bnx2i_hba *hba = bnx2i_conn->hba;
1191 struct bnx2i_cmd *cmd = task->dd_data;
1192
1193 memset(bnx2i_conn->gen_pdu.req_buf, 0, ISCSI_DEF_MAX_RECV_SEG_LEN);
1194
1195 bnx2i_setup_cmd_wqe_template(cmd);
1196 bnx2i_conn->gen_pdu.req_buf_size = task->data_count;
1197
1198
1199 ADD_STATS_64(hba, tx_pdus, 1);
1200 ADD_STATS_64(hba, tx_bytes, task->data_count);
1201
1202 if (task->data_count) {
1203 memcpy(bnx2i_conn->gen_pdu.req_buf, task->data,
1204 task->data_count);
1205 bnx2i_conn->gen_pdu.req_wr_ptr =
1206 bnx2i_conn->gen_pdu.req_buf + task->data_count;
1207 }
1208 cmd->conn = conn->dd_data;
1209 cmd->scsi_cmd = NULL;
1210 return bnx2i_iscsi_send_generic_request(task);
1211 }
1212
1213
1214
1215
1216
1217
1218
1219 static int bnx2i_task_xmit(struct iscsi_task *task)
1220 {
1221 struct iscsi_conn *conn = task->conn;
1222 struct iscsi_session *session = conn->session;
1223 struct Scsi_Host *shost = iscsi_session_to_shost(session->cls_session);
1224 struct bnx2i_hba *hba = iscsi_host_priv(shost);
1225 struct bnx2i_conn *bnx2i_conn = conn->dd_data;
1226 struct scsi_cmnd *sc = task->sc;
1227 struct bnx2i_cmd *cmd = task->dd_data;
1228 struct iscsi_scsi_req *hdr = (struct iscsi_scsi_req *)task->hdr;
1229
1230 if (atomic_read(&bnx2i_conn->ep->num_active_cmds) + 1 >
1231 hba->max_sqes)
1232 return -ENOMEM;
1233
1234
1235
1236
1237 if (!sc)
1238 return bnx2i_mtask_xmit(conn, task);
1239
1240 bnx2i_setup_cmd_wqe_template(cmd);
1241 cmd->req.op_code = ISCSI_OP_SCSI_CMD;
1242 cmd->conn = bnx2i_conn;
1243 cmd->scsi_cmd = sc;
1244 cmd->req.total_data_transfer_length = scsi_bufflen(sc);
1245 cmd->req.cmd_sn = be32_to_cpu(hdr->cmdsn);
1246
1247 bnx2i_iscsi_map_sg_list(cmd);
1248 bnx2i_cpy_scsi_cdb(sc, cmd);
1249
1250 cmd->req.op_attr = ISCSI_ATTR_SIMPLE;
1251 if (sc->sc_data_direction == DMA_TO_DEVICE) {
1252 cmd->req.op_attr |= ISCSI_CMD_REQUEST_WRITE;
1253 cmd->req.itt = task->itt |
1254 (ISCSI_TASK_TYPE_WRITE << ISCSI_CMD_REQUEST_TYPE_SHIFT);
1255 bnx2i_setup_write_cmd_bd_info(task);
1256 } else {
1257 if (scsi_bufflen(sc))
1258 cmd->req.op_attr |= ISCSI_CMD_REQUEST_READ;
1259 cmd->req.itt = task->itt |
1260 (ISCSI_TASK_TYPE_READ << ISCSI_CMD_REQUEST_TYPE_SHIFT);
1261 }
1262
1263 cmd->req.num_bds = cmd->io_tbl.bd_valid;
1264 if (!cmd->io_tbl.bd_valid) {
1265 cmd->req.bd_list_addr_lo = (u32) hba->mp_bd_dma;
1266 cmd->req.bd_list_addr_hi = (u32) ((u64) hba->mp_bd_dma >> 32);
1267 cmd->req.num_bds = 1;
1268 }
1269
1270 bnx2i_send_iscsi_scsicmd(bnx2i_conn, cmd);
1271 return 0;
1272 }
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283 static struct iscsi_cls_session *
1284 bnx2i_session_create(struct iscsi_endpoint *ep,
1285 uint16_t cmds_max, uint16_t qdepth,
1286 uint32_t initial_cmdsn)
1287 {
1288 struct Scsi_Host *shost;
1289 struct iscsi_cls_session *cls_session;
1290 struct bnx2i_hba *hba;
1291 struct bnx2i_endpoint *bnx2i_ep;
1292
1293 if (!ep) {
1294 printk(KERN_ERR "bnx2i: missing ep.\n");
1295 return NULL;
1296 }
1297
1298 bnx2i_ep = ep->dd_data;
1299 shost = bnx2i_ep->hba->shost;
1300 hba = iscsi_host_priv(shost);
1301 if (bnx2i_adapter_ready(hba))
1302 return NULL;
1303
1304
1305
1306
1307
1308 if (cmds_max > hba->max_sqes)
1309 cmds_max = hba->max_sqes;
1310 else if (cmds_max < BNX2I_SQ_WQES_MIN)
1311 cmds_max = BNX2I_SQ_WQES_MIN;
1312
1313 cls_session = iscsi_session_setup(&bnx2i_iscsi_transport, shost,
1314 cmds_max, 0, sizeof(struct bnx2i_cmd),
1315 initial_cmdsn, ISCSI_MAX_TARGET);
1316 if (!cls_session)
1317 return NULL;
1318
1319 if (bnx2i_setup_cmd_pool(hba, cls_session->dd_data))
1320 goto session_teardown;
1321 return cls_session;
1322
1323 session_teardown:
1324 iscsi_session_teardown(cls_session);
1325 return NULL;
1326 }
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336 static void bnx2i_session_destroy(struct iscsi_cls_session *cls_session)
1337 {
1338 struct iscsi_session *session = cls_session->dd_data;
1339 struct Scsi_Host *shost = iscsi_session_to_shost(cls_session);
1340 struct bnx2i_hba *hba = iscsi_host_priv(shost);
1341
1342 bnx2i_destroy_cmd_pool(hba, session);
1343 iscsi_session_teardown(cls_session);
1344 }
1345
1346
1347
1348
1349
1350
1351
1352
1353
1354 static struct iscsi_cls_conn *
1355 bnx2i_conn_create(struct iscsi_cls_session *cls_session, uint32_t cid)
1356 {
1357 struct Scsi_Host *shost = iscsi_session_to_shost(cls_session);
1358 struct bnx2i_hba *hba = iscsi_host_priv(shost);
1359 struct bnx2i_conn *bnx2i_conn;
1360 struct iscsi_cls_conn *cls_conn;
1361 struct iscsi_conn *conn;
1362
1363 cls_conn = iscsi_conn_setup(cls_session, sizeof(*bnx2i_conn),
1364 cid);
1365 if (!cls_conn)
1366 return NULL;
1367 conn = cls_conn->dd_data;
1368
1369 bnx2i_conn = conn->dd_data;
1370 bnx2i_conn->cls_conn = cls_conn;
1371 bnx2i_conn->hba = hba;
1372
1373 atomic_set(&bnx2i_conn->work_cnt, 0);
1374
1375
1376 bnx2i_conn->ep = NULL;
1377 init_completion(&bnx2i_conn->cmd_cleanup_cmpl);
1378
1379 if (bnx2i_conn_alloc_login_resources(hba, bnx2i_conn)) {
1380 iscsi_conn_printk(KERN_ALERT, conn,
1381 "conn_new: login resc alloc failed!!\n");
1382 goto free_conn;
1383 }
1384
1385 return cls_conn;
1386
1387 free_conn:
1388 iscsi_conn_teardown(cls_conn);
1389 return NULL;
1390 }
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404 static int bnx2i_conn_bind(struct iscsi_cls_session *cls_session,
1405 struct iscsi_cls_conn *cls_conn,
1406 uint64_t transport_fd, int is_leading)
1407 {
1408 struct iscsi_conn *conn = cls_conn->dd_data;
1409 struct bnx2i_conn *bnx2i_conn = conn->dd_data;
1410 struct Scsi_Host *shost = iscsi_session_to_shost(cls_session);
1411 struct bnx2i_hba *hba = iscsi_host_priv(shost);
1412 struct bnx2i_endpoint *bnx2i_ep;
1413 struct iscsi_endpoint *ep;
1414 int ret_code;
1415
1416 ep = iscsi_lookup_endpoint(transport_fd);
1417 if (!ep)
1418 return -EINVAL;
1419
1420
1421
1422
1423 if (bnx2i_adapter_ready(hba)) {
1424 ret_code = -EIO;
1425 goto put_ep;
1426 }
1427
1428 bnx2i_ep = ep->dd_data;
1429 if ((bnx2i_ep->state == EP_STATE_TCP_FIN_RCVD) ||
1430 (bnx2i_ep->state == EP_STATE_TCP_RST_RCVD)) {
1431
1432 ret_code = -EINVAL;
1433 goto put_ep;
1434 }
1435
1436 if (iscsi_conn_bind(cls_session, cls_conn, is_leading)) {
1437 ret_code = -EINVAL;
1438 goto put_ep;
1439 }
1440
1441 if (bnx2i_ep->hba != hba) {
1442
1443
1444 iscsi_conn_printk(KERN_ALERT, cls_conn->dd_data,
1445 "conn bind, ep=0x%p (%s) does not",
1446 bnx2i_ep, bnx2i_ep->hba->netdev->name);
1447 iscsi_conn_printk(KERN_ALERT, cls_conn->dd_data,
1448 "belong to hba (%s)\n",
1449 hba->netdev->name);
1450 ret_code = -EEXIST;
1451 goto put_ep;
1452 }
1453 bnx2i_ep->conn = bnx2i_conn;
1454 bnx2i_conn->ep = bnx2i_ep;
1455 bnx2i_conn->iscsi_conn_cid = bnx2i_ep->ep_iscsi_cid;
1456 bnx2i_conn->fw_cid = bnx2i_ep->ep_cid;
1457
1458 ret_code = bnx2i_bind_conn_to_iscsi_cid(hba, bnx2i_conn,
1459 bnx2i_ep->ep_iscsi_cid);
1460
1461
1462
1463
1464 if (test_bit(BNX2I_NX2_DEV_57710, &bnx2i_ep->hba->cnic_dev_type))
1465 bnx2i_put_rq_buf(bnx2i_conn, 0);
1466
1467 bnx2i_arm_cq_event_coalescing(bnx2i_conn->ep, CNIC_ARM_CQE);
1468 put_ep:
1469 iscsi_put_endpoint(ep);
1470 return ret_code;
1471 }
1472
1473
1474
1475
1476
1477
1478
1479
1480
1481 static void bnx2i_conn_destroy(struct iscsi_cls_conn *cls_conn)
1482 {
1483 struct iscsi_conn *conn = cls_conn->dd_data;
1484 struct bnx2i_conn *bnx2i_conn = conn->dd_data;
1485 struct Scsi_Host *shost;
1486 struct bnx2i_hba *hba;
1487 struct bnx2i_work *work, *tmp;
1488 unsigned cpu = 0;
1489 struct bnx2i_percpu_s *p;
1490
1491 shost = iscsi_session_to_shost(iscsi_conn_to_session(cls_conn));
1492 hba = iscsi_host_priv(shost);
1493
1494 bnx2i_conn_free_login_resources(hba, bnx2i_conn);
1495
1496 if (atomic_read(&bnx2i_conn->work_cnt)) {
1497 for_each_online_cpu(cpu) {
1498 p = &per_cpu(bnx2i_percpu, cpu);
1499 spin_lock_bh(&p->p_work_lock);
1500 list_for_each_entry_safe(work, tmp,
1501 &p->work_list, list) {
1502 if (work->session == conn->session &&
1503 work->bnx2i_conn == bnx2i_conn) {
1504 list_del_init(&work->list);
1505 kfree(work);
1506 if (!atomic_dec_and_test(
1507 &bnx2i_conn->work_cnt))
1508 break;
1509 }
1510 }
1511 spin_unlock_bh(&p->p_work_lock);
1512 }
1513 }
1514
1515 iscsi_conn_teardown(cls_conn);
1516 }
1517
1518
1519
1520
1521
1522
1523
1524
1525
1526
1527 static int bnx2i_ep_get_param(struct iscsi_endpoint *ep,
1528 enum iscsi_param param, char *buf)
1529 {
1530 struct bnx2i_endpoint *bnx2i_ep = ep->dd_data;
1531 struct bnx2i_hba *hba = bnx2i_ep->hba;
1532 int len = -ENOTCONN;
1533
1534 if (!hba)
1535 return -ENOTCONN;
1536
1537 switch (param) {
1538 case ISCSI_PARAM_CONN_PORT:
1539 mutex_lock(&hba->net_dev_lock);
1540 if (bnx2i_ep->cm_sk)
1541 len = sprintf(buf, "%hu\n", bnx2i_ep->cm_sk->dst_port);
1542 mutex_unlock(&hba->net_dev_lock);
1543 break;
1544 case ISCSI_PARAM_CONN_ADDRESS:
1545 mutex_lock(&hba->net_dev_lock);
1546 if (bnx2i_ep->cm_sk)
1547 len = sprintf(buf, "%pI4\n", &bnx2i_ep->cm_sk->dst_ip);
1548 mutex_unlock(&hba->net_dev_lock);
1549 break;
1550 default:
1551 return -ENOSYS;
1552 }
1553
1554 return len;
1555 }
1556
1557
1558
1559
1560
1561
1562
1563 static int bnx2i_host_get_param(struct Scsi_Host *shost,
1564 enum iscsi_host_param param, char *buf)
1565 {
1566 struct bnx2i_hba *hba = iscsi_host_priv(shost);
1567 int len = 0;
1568
1569 switch (param) {
1570 case ISCSI_HOST_PARAM_HWADDRESS:
1571 len = sysfs_format_mac(buf, hba->cnic->mac_addr, 6);
1572 break;
1573 case ISCSI_HOST_PARAM_NETDEV_NAME:
1574 len = sprintf(buf, "%s\n", hba->netdev->name);
1575 break;
1576 case ISCSI_HOST_PARAM_IPADDRESS: {
1577 struct list_head *active_list = &hba->ep_active_list;
1578
1579 read_lock_bh(&hba->ep_rdwr_lock);
1580 if (!list_empty(&hba->ep_active_list)) {
1581 struct bnx2i_endpoint *bnx2i_ep;
1582 struct cnic_sock *csk;
1583
1584 bnx2i_ep = list_first_entry(active_list,
1585 struct bnx2i_endpoint,
1586 link);
1587 csk = bnx2i_ep->cm_sk;
1588 if (test_bit(SK_F_IPV6, &csk->flags))
1589 len = sprintf(buf, "%pI6\n", csk->src_ip);
1590 else
1591 len = sprintf(buf, "%pI4\n", csk->src_ip);
1592 }
1593 read_unlock_bh(&hba->ep_rdwr_lock);
1594 break;
1595 }
1596 default:
1597 return iscsi_host_get_param(shost, param, buf);
1598 }
1599 return len;
1600 }
1601
1602
1603
1604
1605
1606
1607
1608 static int bnx2i_conn_start(struct iscsi_cls_conn *cls_conn)
1609 {
1610 struct iscsi_conn *conn = cls_conn->dd_data;
1611 struct bnx2i_conn *bnx2i_conn = conn->dd_data;
1612
1613 bnx2i_conn->ep->state = EP_STATE_ULP_UPDATE_START;
1614 bnx2i_update_iscsi_conn(conn);
1615
1616
1617
1618
1619
1620 timer_setup(&bnx2i_conn->ep->ofld_timer, bnx2i_ep_ofld_timer, 0);
1621 bnx2i_conn->ep->ofld_timer.expires = 1 * HZ + jiffies;
1622 add_timer(&bnx2i_conn->ep->ofld_timer);
1623
1624 wait_event_interruptible(bnx2i_conn->ep->ofld_wait,
1625 bnx2i_conn->ep->state != EP_STATE_ULP_UPDATE_START);
1626
1627 if (signal_pending(current))
1628 flush_signals(current);
1629 del_timer_sync(&bnx2i_conn->ep->ofld_timer);
1630
1631 iscsi_conn_start(cls_conn);
1632 return 0;
1633 }
1634
1635
1636
1637
1638
1639
1640
1641 static void bnx2i_conn_get_stats(struct iscsi_cls_conn *cls_conn,
1642 struct iscsi_stats *stats)
1643 {
1644 struct iscsi_conn *conn = cls_conn->dd_data;
1645
1646 stats->txdata_octets = conn->txdata_octets;
1647 stats->rxdata_octets = conn->rxdata_octets;
1648 stats->scsicmd_pdus = conn->scsicmd_pdus_cnt;
1649 stats->dataout_pdus = conn->dataout_pdus_cnt;
1650 stats->scsirsp_pdus = conn->scsirsp_pdus_cnt;
1651 stats->datain_pdus = conn->datain_pdus_cnt;
1652 stats->r2t_pdus = conn->r2t_pdus_cnt;
1653 stats->tmfcmd_pdus = conn->tmfcmd_pdus_cnt;
1654 stats->tmfrsp_pdus = conn->tmfrsp_pdus_cnt;
1655 stats->digest_err = 0;
1656 stats->timeout_err = 0;
1657 strcpy(stats->custom[0].desc, "eh_abort_cnt");
1658 stats->custom[0].value = conn->eh_abort_cnt;
1659 stats->custom_length = 1;
1660 }
1661
1662
1663
1664
1665
1666
1667
1668
1669 static struct bnx2i_hba *bnx2i_check_route(struct sockaddr *dst_addr)
1670 {
1671 struct sockaddr_in *desti = (struct sockaddr_in *) dst_addr;
1672 struct bnx2i_hba *hba;
1673 struct cnic_dev *cnic = NULL;
1674
1675 hba = get_adapter_list_head();
1676 if (hba && hba->cnic)
1677 cnic = hba->cnic->cm_select_dev(desti, CNIC_ULP_ISCSI);
1678 if (!cnic) {
1679 printk(KERN_ALERT "bnx2i: no route,"
1680 "can't connect using cnic\n");
1681 goto no_nx2_route;
1682 }
1683 hba = bnx2i_find_hba_for_cnic(cnic);
1684 if (!hba)
1685 goto no_nx2_route;
1686
1687 if (bnx2i_adapter_ready(hba)) {
1688 printk(KERN_ALERT "bnx2i: check route, hba not found\n");
1689 goto no_nx2_route;
1690 }
1691 if (hba->netdev->mtu > hba->mtu_supported) {
1692 printk(KERN_ALERT "bnx2i: %s network i/f mtu is set to %d\n",
1693 hba->netdev->name, hba->netdev->mtu);
1694 printk(KERN_ALERT "bnx2i: iSCSI HBA can support mtu of %d\n",
1695 hba->mtu_supported);
1696 goto no_nx2_route;
1697 }
1698 return hba;
1699 no_nx2_route:
1700 return NULL;
1701 }
1702
1703
1704
1705
1706
1707
1708
1709
1710
1711 static int bnx2i_tear_down_conn(struct bnx2i_hba *hba,
1712 struct bnx2i_endpoint *ep)
1713 {
1714 if (test_bit(BNX2I_CNIC_REGISTERED, &hba->reg_with_cnic) && ep->cm_sk)
1715 hba->cnic->cm_destroy(ep->cm_sk);
1716
1717 if (test_bit(BNX2I_NX2_DEV_57710, &hba->cnic_dev_type) &&
1718 ep->state == EP_STATE_DISCONN_TIMEDOUT) {
1719 if (ep->conn && ep->conn->cls_conn &&
1720 ep->conn->cls_conn->dd_data) {
1721 struct iscsi_conn *conn = ep->conn->cls_conn->dd_data;
1722
1723
1724 set_bit(ISCSI_CONN_FLAG_SUSPEND_RX, &conn->flags);
1725 }
1726
1727
1728
1729
1730 printk(KERN_ALERT "bnx2i (%s): - WARN - CONN_DISCON timed out, "
1731 "please submit GRC Dump, NW/PCIe trace, "
1732 "driver msgs to developers for analysis\n",
1733 hba->netdev->name);
1734 }
1735
1736 ep->state = EP_STATE_CLEANUP_START;
1737 timer_setup(&ep->ofld_timer, bnx2i_ep_ofld_timer, 0);
1738 ep->ofld_timer.expires = hba->conn_ctx_destroy_tmo + jiffies;
1739 add_timer(&ep->ofld_timer);
1740
1741 bnx2i_ep_destroy_list_add(hba, ep);
1742
1743
1744 if (bnx2i_send_conn_destroy(hba, ep))
1745 ep->state = EP_STATE_CLEANUP_CMPL;
1746
1747 wait_event_interruptible(ep->ofld_wait,
1748 (ep->state != EP_STATE_CLEANUP_START));
1749
1750 if (signal_pending(current))
1751 flush_signals(current);
1752 del_timer_sync(&ep->ofld_timer);
1753
1754 bnx2i_ep_destroy_list_del(hba, ep);
1755
1756 if (ep->state != EP_STATE_CLEANUP_CMPL)
1757
1758 printk(KERN_ALERT "bnx2i - conn destroy failed\n");
1759
1760 return 0;
1761 }
1762
1763
1764
1765
1766
1767
1768
1769
1770
1771
1772
1773
1774
1775
1776 static struct iscsi_endpoint *bnx2i_ep_connect(struct Scsi_Host *shost,
1777 struct sockaddr *dst_addr,
1778 int non_blocking)
1779 {
1780 u32 iscsi_cid = BNX2I_CID_RESERVED;
1781 struct sockaddr_in *desti = (struct sockaddr_in *) dst_addr;
1782 struct sockaddr_in6 *desti6;
1783 struct bnx2i_endpoint *bnx2i_ep;
1784 struct bnx2i_hba *hba;
1785 struct cnic_dev *cnic;
1786 struct cnic_sockaddr saddr;
1787 struct iscsi_endpoint *ep;
1788 int rc = 0;
1789
1790 if (shost) {
1791
1792 hba = iscsi_host_priv(shost);
1793 } else
1794
1795
1796
1797
1798 hba = bnx2i_check_route(dst_addr);
1799
1800 if (!hba) {
1801 rc = -EINVAL;
1802 goto nohba;
1803 }
1804 mutex_lock(&hba->net_dev_lock);
1805
1806 if (bnx2i_adapter_ready(hba) || !hba->cid_que.cid_free_cnt) {
1807 rc = -EPERM;
1808 goto check_busy;
1809 }
1810 cnic = hba->cnic;
1811 ep = bnx2i_alloc_ep(hba);
1812 if (!ep) {
1813 rc = -ENOMEM;
1814 goto check_busy;
1815 }
1816 bnx2i_ep = ep->dd_data;
1817
1818 atomic_set(&bnx2i_ep->num_active_cmds, 0);
1819 iscsi_cid = bnx2i_alloc_iscsi_cid(hba);
1820 if (iscsi_cid == -1) {
1821 printk(KERN_ALERT "bnx2i (%s): alloc_ep - unable to allocate "
1822 "iscsi cid\n", hba->netdev->name);
1823 rc = -ENOMEM;
1824 bnx2i_free_ep(ep);
1825 goto check_busy;
1826 }
1827 bnx2i_ep->hba_age = hba->age;
1828
1829 rc = bnx2i_alloc_qp_resc(hba, bnx2i_ep);
1830 if (rc != 0) {
1831 printk(KERN_ALERT "bnx2i (%s): ep_conn - alloc QP resc error"
1832 "\n", hba->netdev->name);
1833 rc = -ENOMEM;
1834 goto qp_resc_err;
1835 }
1836
1837 bnx2i_ep->ep_iscsi_cid = (u16)iscsi_cid;
1838 bnx2i_ep->state = EP_STATE_OFLD_START;
1839 bnx2i_ep_ofld_list_add(hba, bnx2i_ep);
1840
1841 timer_setup(&bnx2i_ep->ofld_timer, bnx2i_ep_ofld_timer, 0);
1842 bnx2i_ep->ofld_timer.expires = 2 * HZ + jiffies;
1843 add_timer(&bnx2i_ep->ofld_timer);
1844
1845 if (bnx2i_send_conn_ofld_req(hba, bnx2i_ep)) {
1846 if (bnx2i_ep->state == EP_STATE_OFLD_FAILED_CID_BUSY) {
1847 printk(KERN_ALERT "bnx2i (%s): iscsi cid %d is busy\n",
1848 hba->netdev->name, bnx2i_ep->ep_iscsi_cid);
1849 rc = -EBUSY;
1850 } else
1851 rc = -ENOSPC;
1852 printk(KERN_ALERT "bnx2i (%s): unable to send conn offld kwqe"
1853 "\n", hba->netdev->name);
1854 bnx2i_ep_ofld_list_del(hba, bnx2i_ep);
1855 goto conn_failed;
1856 }
1857
1858
1859 wait_event_interruptible(bnx2i_ep->ofld_wait,
1860 bnx2i_ep->state != EP_STATE_OFLD_START);
1861
1862 if (signal_pending(current))
1863 flush_signals(current);
1864 del_timer_sync(&bnx2i_ep->ofld_timer);
1865
1866 bnx2i_ep_ofld_list_del(hba, bnx2i_ep);
1867
1868 if (bnx2i_ep->state != EP_STATE_OFLD_COMPL) {
1869 if (bnx2i_ep->state == EP_STATE_OFLD_FAILED_CID_BUSY) {
1870 printk(KERN_ALERT "bnx2i (%s): iscsi cid %d is busy\n",
1871 hba->netdev->name, bnx2i_ep->ep_iscsi_cid);
1872 rc = -EBUSY;
1873 } else
1874 rc = -ENOSPC;
1875 goto conn_failed;
1876 }
1877
1878 rc = cnic->cm_create(cnic, CNIC_ULP_ISCSI, bnx2i_ep->ep_cid,
1879 iscsi_cid, &bnx2i_ep->cm_sk, bnx2i_ep);
1880 if (rc) {
1881 rc = -EINVAL;
1882
1883 goto release_ep;
1884 }
1885
1886 bnx2i_ep->cm_sk->rcv_buf = 256 * 1024;
1887 bnx2i_ep->cm_sk->snd_buf = 256 * 1024;
1888 clear_bit(SK_TCP_TIMESTAMP, &bnx2i_ep->cm_sk->tcp_flags);
1889
1890 memset(&saddr, 0, sizeof(saddr));
1891 if (dst_addr->sa_family == AF_INET) {
1892 desti = (struct sockaddr_in *) dst_addr;
1893 saddr.remote.v4 = *desti;
1894 saddr.local.v4.sin_family = desti->sin_family;
1895 } else if (dst_addr->sa_family == AF_INET6) {
1896 desti6 = (struct sockaddr_in6 *) dst_addr;
1897 saddr.remote.v6 = *desti6;
1898 saddr.local.v6.sin6_family = desti6->sin6_family;
1899 }
1900
1901 bnx2i_ep->timestamp = jiffies;
1902 bnx2i_ep->state = EP_STATE_CONNECT_START;
1903 if (!test_bit(BNX2I_CNIC_REGISTERED, &hba->reg_with_cnic)) {
1904 rc = -EINVAL;
1905 goto conn_failed;
1906 } else
1907 rc = cnic->cm_connect(bnx2i_ep->cm_sk, &saddr);
1908 if (rc)
1909 goto release_ep;
1910
1911 bnx2i_ep_active_list_add(hba, bnx2i_ep);
1912
1913 rc = bnx2i_map_ep_dbell_regs(bnx2i_ep);
1914 if (rc)
1915 goto del_active_ep;
1916
1917 mutex_unlock(&hba->net_dev_lock);
1918 return ep;
1919
1920 del_active_ep:
1921 bnx2i_ep_active_list_del(hba, bnx2i_ep);
1922 release_ep:
1923 if (bnx2i_tear_down_conn(hba, bnx2i_ep)) {
1924 mutex_unlock(&hba->net_dev_lock);
1925 return ERR_PTR(rc);
1926 }
1927 conn_failed:
1928 bnx2i_free_qp_resc(hba, bnx2i_ep);
1929 qp_resc_err:
1930 bnx2i_free_ep(ep);
1931 check_busy:
1932 mutex_unlock(&hba->net_dev_lock);
1933 nohba:
1934 return ERR_PTR(rc);
1935 }
1936
1937
1938
1939
1940
1941
1942
1943
1944
1945 static int bnx2i_ep_poll(struct iscsi_endpoint *ep, int timeout_ms)
1946 {
1947 struct bnx2i_endpoint *bnx2i_ep;
1948 int rc = 0;
1949
1950 bnx2i_ep = ep->dd_data;
1951 if ((bnx2i_ep->state == EP_STATE_IDLE) ||
1952 (bnx2i_ep->state == EP_STATE_CONNECT_FAILED) ||
1953 (bnx2i_ep->state == EP_STATE_OFLD_FAILED))
1954 return -1;
1955 if (bnx2i_ep->state == EP_STATE_CONNECT_COMPL)
1956 return 1;
1957
1958 rc = wait_event_interruptible_timeout(bnx2i_ep->ofld_wait,
1959 ((bnx2i_ep->state ==
1960 EP_STATE_OFLD_FAILED) ||
1961 (bnx2i_ep->state ==
1962 EP_STATE_CONNECT_FAILED) ||
1963 (bnx2i_ep->state ==
1964 EP_STATE_CONNECT_COMPL)),
1965 msecs_to_jiffies(timeout_ms));
1966 if (bnx2i_ep->state == EP_STATE_OFLD_FAILED)
1967 rc = -1;
1968
1969 if (rc > 0)
1970 return 1;
1971 else if (!rc)
1972 return 0;
1973 else
1974 return rc;
1975 }
1976
1977
1978
1979
1980
1981
1982
1983
1984 static int bnx2i_ep_tcp_conn_active(struct bnx2i_endpoint *bnx2i_ep)
1985 {
1986 int ret;
1987 int cnic_dev_10g = 0;
1988
1989 if (test_bit(BNX2I_NX2_DEV_57710, &bnx2i_ep->hba->cnic_dev_type))
1990 cnic_dev_10g = 1;
1991
1992 switch (bnx2i_ep->state) {
1993 case EP_STATE_CLEANUP_FAILED:
1994 case EP_STATE_OFLD_FAILED:
1995 case EP_STATE_DISCONN_TIMEDOUT:
1996 ret = 0;
1997 break;
1998 case EP_STATE_CONNECT_START:
1999 case EP_STATE_CONNECT_FAILED:
2000 case EP_STATE_CONNECT_COMPL:
2001 case EP_STATE_ULP_UPDATE_START:
2002 case EP_STATE_ULP_UPDATE_COMPL:
2003 case EP_STATE_TCP_FIN_RCVD:
2004 case EP_STATE_LOGOUT_SENT:
2005 case EP_STATE_LOGOUT_RESP_RCVD:
2006 case EP_STATE_ULP_UPDATE_FAILED:
2007 ret = 1;
2008 break;
2009 case EP_STATE_TCP_RST_RCVD:
2010 if (cnic_dev_10g)
2011 ret = 0;
2012 else
2013 ret = 1;
2014 break;
2015 default:
2016 ret = 0;
2017 }
2018
2019 return ret;
2020 }
2021
2022
2023
2024
2025
2026
2027
2028
2029 int bnx2i_hw_ep_disconnect(struct bnx2i_endpoint *bnx2i_ep)
2030 {
2031 struct bnx2i_hba *hba = bnx2i_ep->hba;
2032 struct cnic_dev *cnic;
2033 struct iscsi_session *session = NULL;
2034 struct iscsi_conn *conn = NULL;
2035 int ret = 0;
2036 int close = 0;
2037 int close_ret = 0;
2038
2039 if (!hba)
2040 return 0;
2041
2042 cnic = hba->cnic;
2043 if (!cnic)
2044 return 0;
2045
2046 if (bnx2i_ep->state == EP_STATE_IDLE ||
2047 bnx2i_ep->state == EP_STATE_DISCONN_TIMEDOUT)
2048 return 0;
2049
2050 if (!bnx2i_ep_tcp_conn_active(bnx2i_ep))
2051 goto destroy_conn;
2052
2053 if (bnx2i_ep->conn) {
2054 conn = bnx2i_ep->conn->cls_conn->dd_data;
2055 session = conn->session;
2056 }
2057
2058 timer_setup(&bnx2i_ep->ofld_timer, bnx2i_ep_ofld_timer, 0);
2059 bnx2i_ep->ofld_timer.expires = hba->conn_teardown_tmo + jiffies;
2060 add_timer(&bnx2i_ep->ofld_timer);
2061
2062 if (!test_bit(BNX2I_CNIC_REGISTERED, &hba->reg_with_cnic))
2063 goto out;
2064
2065 if (session) {
2066 spin_lock_bh(&session->frwd_lock);
2067 if (bnx2i_ep->state != EP_STATE_TCP_FIN_RCVD) {
2068 if (session->state == ISCSI_STATE_LOGGING_OUT) {
2069 if (bnx2i_ep->state == EP_STATE_LOGOUT_SENT) {
2070
2071 printk(KERN_ALERT "bnx2i (%s): WARNING"
2072 " logout response was not "
2073 "received!\n",
2074 bnx2i_ep->hba->netdev->name);
2075 } else if (bnx2i_ep->state ==
2076 EP_STATE_LOGOUT_RESP_RCVD)
2077 close = 1;
2078 }
2079 } else
2080 close = 1;
2081
2082 spin_unlock_bh(&session->frwd_lock);
2083 }
2084
2085 bnx2i_ep->state = EP_STATE_DISCONN_START;
2086
2087 if (close)
2088 close_ret = cnic->cm_close(bnx2i_ep->cm_sk);
2089 else
2090 close_ret = cnic->cm_abort(bnx2i_ep->cm_sk);
2091
2092 if (close_ret)
2093 printk(KERN_ALERT "bnx2i (%s): close/abort(%d) returned %d\n",
2094 bnx2i_ep->hba->netdev->name, close, close_ret);
2095 else
2096
2097 wait_event_interruptible(bnx2i_ep->ofld_wait,
2098 ((bnx2i_ep->state != EP_STATE_DISCONN_START)
2099 && (bnx2i_ep->state != EP_STATE_TCP_FIN_RCVD)));
2100
2101 if (signal_pending(current))
2102 flush_signals(current);
2103 del_timer_sync(&bnx2i_ep->ofld_timer);
2104
2105 destroy_conn:
2106 bnx2i_ep_active_list_del(hba, bnx2i_ep);
2107 if (bnx2i_tear_down_conn(hba, bnx2i_ep))
2108 return -EINVAL;
2109 out:
2110 bnx2i_ep->state = EP_STATE_IDLE;
2111 return ret;
2112 }
2113
2114
2115
2116
2117
2118
2119
2120
2121 static void bnx2i_ep_disconnect(struct iscsi_endpoint *ep)
2122 {
2123 struct bnx2i_endpoint *bnx2i_ep;
2124 struct bnx2i_conn *bnx2i_conn = NULL;
2125 struct bnx2i_hba *hba;
2126
2127 bnx2i_ep = ep->dd_data;
2128
2129
2130
2131
2132
2133 while ((bnx2i_ep->state == EP_STATE_CONNECT_START) &&
2134 !time_after(jiffies, bnx2i_ep->timestamp + (12 * HZ)))
2135 msleep(250);
2136
2137 if (bnx2i_ep->conn)
2138 bnx2i_conn = bnx2i_ep->conn;
2139 hba = bnx2i_ep->hba;
2140
2141 mutex_lock(&hba->net_dev_lock);
2142
2143 if (bnx2i_ep->state == EP_STATE_DISCONN_TIMEDOUT)
2144 goto out;
2145
2146 if (bnx2i_ep->state == EP_STATE_IDLE)
2147 goto free_resc;
2148
2149 if (!test_bit(ADAPTER_STATE_UP, &hba->adapter_state) ||
2150 (bnx2i_ep->hba_age != hba->age)) {
2151 bnx2i_ep_active_list_del(hba, bnx2i_ep);
2152 goto free_resc;
2153 }
2154
2155
2156 if (bnx2i_hw_ep_disconnect(bnx2i_ep)) {
2157 mutex_unlock(&hba->net_dev_lock);
2158 return;
2159 }
2160 free_resc:
2161 bnx2i_free_qp_resc(hba, bnx2i_ep);
2162
2163 if (bnx2i_conn)
2164 bnx2i_conn->ep = NULL;
2165
2166 bnx2i_free_ep(ep);
2167 out:
2168 mutex_unlock(&hba->net_dev_lock);
2169
2170 wake_up_interruptible(&hba->eh_wait);
2171 }
2172
2173
2174
2175
2176
2177
2178
2179 static int bnx2i_nl_set_path(struct Scsi_Host *shost, struct iscsi_path *params)
2180 {
2181 struct bnx2i_hba *hba = iscsi_host_priv(shost);
2182 char *buf = (char *) params;
2183 u16 len = sizeof(*params);
2184
2185
2186 hba->cnic->iscsi_nl_msg_recv(hba->cnic, ISCSI_UEVENT_PATH_UPDATE, buf,
2187 len);
2188
2189 return 0;
2190 }
2191
2192 static umode_t bnx2i_attr_is_visible(int param_type, int param)
2193 {
2194 switch (param_type) {
2195 case ISCSI_HOST_PARAM:
2196 switch (param) {
2197 case ISCSI_HOST_PARAM_NETDEV_NAME:
2198 case ISCSI_HOST_PARAM_HWADDRESS:
2199 case ISCSI_HOST_PARAM_IPADDRESS:
2200 return S_IRUGO;
2201 default:
2202 return 0;
2203 }
2204 case ISCSI_PARAM:
2205 switch (param) {
2206 case ISCSI_PARAM_MAX_RECV_DLENGTH:
2207 case ISCSI_PARAM_MAX_XMIT_DLENGTH:
2208 case ISCSI_PARAM_HDRDGST_EN:
2209 case ISCSI_PARAM_DATADGST_EN:
2210 case ISCSI_PARAM_CONN_ADDRESS:
2211 case ISCSI_PARAM_CONN_PORT:
2212 case ISCSI_PARAM_EXP_STATSN:
2213 case ISCSI_PARAM_PERSISTENT_ADDRESS:
2214 case ISCSI_PARAM_PERSISTENT_PORT:
2215 case ISCSI_PARAM_PING_TMO:
2216 case ISCSI_PARAM_RECV_TMO:
2217 case ISCSI_PARAM_INITIAL_R2T_EN:
2218 case ISCSI_PARAM_MAX_R2T:
2219 case ISCSI_PARAM_IMM_DATA_EN:
2220 case ISCSI_PARAM_FIRST_BURST:
2221 case ISCSI_PARAM_MAX_BURST:
2222 case ISCSI_PARAM_PDU_INORDER_EN:
2223 case ISCSI_PARAM_DATASEQ_INORDER_EN:
2224 case ISCSI_PARAM_ERL:
2225 case ISCSI_PARAM_TARGET_NAME:
2226 case ISCSI_PARAM_TPGT:
2227 case ISCSI_PARAM_USERNAME:
2228 case ISCSI_PARAM_PASSWORD:
2229 case ISCSI_PARAM_USERNAME_IN:
2230 case ISCSI_PARAM_PASSWORD_IN:
2231 case ISCSI_PARAM_FAST_ABORT:
2232 case ISCSI_PARAM_ABORT_TMO:
2233 case ISCSI_PARAM_LU_RESET_TMO:
2234 case ISCSI_PARAM_TGT_RESET_TMO:
2235 case ISCSI_PARAM_IFACE_NAME:
2236 case ISCSI_PARAM_INITIATOR_NAME:
2237 case ISCSI_PARAM_BOOT_ROOT:
2238 case ISCSI_PARAM_BOOT_NIC:
2239 case ISCSI_PARAM_BOOT_TARGET:
2240 return S_IRUGO;
2241 default:
2242 return 0;
2243 }
2244 }
2245
2246 return 0;
2247 }
2248
2249
2250
2251
2252
2253 static struct scsi_host_template bnx2i_host_template = {
2254 .module = THIS_MODULE,
2255 .name = "QLogic Offload iSCSI Initiator",
2256 .proc_name = "bnx2i",
2257 .queuecommand = iscsi_queuecommand,
2258 .eh_timed_out = iscsi_eh_cmd_timed_out,
2259 .eh_abort_handler = iscsi_eh_abort,
2260 .eh_device_reset_handler = iscsi_eh_device_reset,
2261 .eh_target_reset_handler = iscsi_eh_recover_target,
2262 .change_queue_depth = scsi_change_queue_depth,
2263 .target_alloc = iscsi_target_alloc,
2264 .can_queue = 2048,
2265 .max_sectors = 127,
2266 .cmd_per_lun = 128,
2267 .this_id = -1,
2268 .sg_tablesize = ISCSI_MAX_BDS_PER_CMD,
2269 .shost_groups = bnx2i_dev_groups,
2270 .track_queue_depth = 1,
2271 .cmd_size = sizeof(struct iscsi_cmd),
2272 };
2273
2274 struct iscsi_transport bnx2i_iscsi_transport = {
2275 .owner = THIS_MODULE,
2276 .name = "bnx2i",
2277 .caps = CAP_RECOVERY_L0 | CAP_HDRDGST |
2278 CAP_MULTI_R2T | CAP_DATADGST |
2279 CAP_DATA_PATH_OFFLOAD |
2280 CAP_TEXT_NEGO,
2281 .create_session = bnx2i_session_create,
2282 .destroy_session = bnx2i_session_destroy,
2283 .create_conn = bnx2i_conn_create,
2284 .bind_conn = bnx2i_conn_bind,
2285 .unbind_conn = iscsi_conn_unbind,
2286 .destroy_conn = bnx2i_conn_destroy,
2287 .attr_is_visible = bnx2i_attr_is_visible,
2288 .set_param = iscsi_set_param,
2289 .get_conn_param = iscsi_conn_get_param,
2290 .get_session_param = iscsi_session_get_param,
2291 .get_host_param = bnx2i_host_get_param,
2292 .start_conn = bnx2i_conn_start,
2293 .stop_conn = iscsi_conn_stop,
2294 .send_pdu = iscsi_conn_send_pdu,
2295 .xmit_task = bnx2i_task_xmit,
2296 .get_stats = bnx2i_conn_get_stats,
2297
2298 .get_ep_param = bnx2i_ep_get_param,
2299 .ep_connect = bnx2i_ep_connect,
2300 .ep_poll = bnx2i_ep_poll,
2301 .ep_disconnect = bnx2i_ep_disconnect,
2302 .set_path = bnx2i_nl_set_path,
2303
2304 .session_recovery_timedout = iscsi_session_recovery_timedout,
2305 .cleanup_task = bnx2i_cleanup_task,
2306 };