dc6311dd2a
bnx2fc_queuecommand(): when allocating a new io_req, the tgt_lock spinlock must be locked before calling bnx2fc_cmd_alloc(). The spinlock should also be kept locked until bnx2fc_post_io_req() has been completed. If not, a kernel thread may call bnx2fc_process_cq_compl() that extracts the newly allocated io_req from hba->cmd_mgr->cmds and destroys it while it is still being used by bnx2fc_post_io_req(). BUG: unable to handle kernel NULL pointer dereference at 000000000000004c IP: [<ffffffffa03130da>] bnx2fc_init_task+0x6a/0x230 [bnx2fc] PGD 0 Oops: 0000 [#1] SMP last sysfs file: /sys/devices/pci0000:00/0000:00:02.0/0000:04:00.3/net/eth3/type CPU 33 Modules linked in: autofs4 target_core_iblock target_core_file target_core_pscsi target_core_mod configfs bnx2fc cnic uio fcoe libfcoe libfc scsi_transport_fc 8021q garp scsi_tgt stp llc cpufreq_ondemand freq_table pcc_cpufreq ipt_REJECT nf_conntrack_ipv4 nf_defrag_ipv4 iptable_filter ip_tables ip6t_REJECT nf_conntrack_ipv6 nf_defrag_ipv6 xt_state nf_conntrack ip6table_filter ip6_tables ipv6 power_meter microcode iTCO_wdt iTCO_vendor_support hpilo hpwdt sg bnx2x libcrc32c mdio serio_raw lpc_ich mfd_core shpchp ext4 jbd2 mbcache sd_mod crc_t10dif hpsa video output dm_mirror dm_region_hash dm_log dm_mod [last unloaded: scsi_wait_scan] Pid: 7355, comm: bnx2fc_thread/3 Not tainted 2.6.32-431.el6.x86_64 #1 HP ProLiant BL460c Gen8 RIP: 0010:[<ffffffffa03130da>] [<ffffffffa03130da>] bnx2fc_init_task+0x6a/0x230 [bnx2fc] RSP: 0018:ffff8820b0da3b68 EFLAGS: 00010246 RAX: 0000000000000000 RBX: ffff882003801080 RCX: 0000000000000000 RDX: 0000000000000000 RSI: 0000000000000000 RDI: ffff882003801100 RBP: ffff8820b0da3bc8 R08: ffffffff8160d4e8 R09: 0000000000000040 R10: 0000000000000000 R11: 0000000000000000 R12: ffff88400e600e00 R13: ffff8840108fbe40 R14: ffff88200ffe5400 R15: 0000000000000000 FS: 0000000000000000(0000) GS:ffff8820b0da0000(0000) knlGS:0000000000000000 CS: 0010 DS: 0018 ES: 0018 CR0: 000000008005003b CR2: 000000000000004c CR3: 0000002010b67000 CR4: 00000000001407e0 DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000 DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400 Process bnx2fc_thread/3 (pid: 7355, threadinfo ffff88401f940000, task ffff884012f5f540) Stack: ffff8820b0da3bc8 ffffffff81527303 ffff884000000020 ffff8820b0da3bd8 <d> ffff8820b0da3b98 000000028138931a ffff88400f506078 ffff88400e600e00 <d> ffff88200ffe5400 ffff88200ffe5590 0000000000000021 0000000000000002 Call Trace: <IRQ> [<ffffffff81527303>] ? printk+0x41/0x46 [<ffffffffa03169bc>] bnx2fc_post_io_req+0x11c/0x440 [bnx2fc] [<ffffffff812825b9>] ? cpumask_next_and+0x29/0x50 [<ffffffff8137ffd0>] ? scsi_done+0x0/0x60 [<ffffffffa0316df7>] bnx2fc_queuecommand+0x117/0x140 [bnx2fc] [<ffffffff81380245>] scsi_dispatch_cmd+0xe5/0x310 [<ffffffff81388b9e>] scsi_request_fn+0x5ee/0x7a0 [<ffffffff812658f1>] __blk_run_queue+0x31/0x40 [<ffffffff81265a40>] blk_run_queue+0x30/0x50 [<ffffffff81387da6>] scsi_run_queue+0xc6/0x270 [<ffffffff81260f92>] ? elv_requeue_request+0x52/0xa0 [<ffffffff813897a0>] scsi_requeue_command+0x90/0xb0 [<ffffffff81389b84>] scsi_io_completion+0x154/0x6c0 [<ffffffff8137ff62>] scsi_finish_command+0xc2/0x130 [<ffffffff8138a255>] scsi_softirq_done+0x145/0x170 [<ffffffff8126e865>] blk_done_softirq+0x85/0xa0 [<ffffffff8107a8e1>] __do_softirq+0xc1/0x1e0 [<ffffffff8100c30c>] ? call_softirq+0x1c/0x30 [<ffffffff8100c30c>] call_softirq+0x1c/0x30 <EOI> [<ffffffff8100fa75>] ? do_softirq+0x65/0xa0 [<ffffffff8107a40a>] local_bh_enable_ip+0x9a/0xb0 [<ffffffff8152a4eb>] _spin_unlock_bh+0x1b/0x20 [<ffffffffa0313937>] bnx2fc_process_cq_compl+0x257/0x2b0 [bnx2fc] [<ffffffffa03114ea>] bnx2fc_percpu_io_thread+0xea/0x160 [bnx2fc] [<ffffffffa0311400>] ? bnx2fc_percpu_io_thread+0x0/0x160 [bnx2fc] [<ffffffff8109aef6>] kthread+0x96/0xa0 [<ffffffff8100c20a>] child_rip+0xa/0x20 [<ffffffff8109ae60>] ? kthread+0x0/0xa0 [<ffffffff8100c200>] ? child_rip+0x0/0x20 Code: 89 df 45 8b 7e 30 0f 85 75 01 00 00 89 d1 31 c0 c1 e9 03 83 e2 04 89 c9 f3 48 ab 74 06 c7 07 00 00 00 00 49 89 9c 24 88 01 00 00 <83> 7e 4c 01 b8 01 00 00 00 0f 84 e7 00 00 00 89 c2 0a 53 38 41 RIP [<ffffffffa03130da>] bnx2fc_init_task+0x6a/0x230 [bnx2fc] RSP <ffff8820b0da3b68> CR2: 000000000000004c Signed-off-by: Maurizio Lombardi <mlombard@redhat.com> Acked-by: Chad Dupuis <chad.dupuis@qlogic.com> Signed-off-by: Christoph Hellwig <hch@lst.de>
918 lines
25 KiB
C
918 lines
25 KiB
C
/*
|
|
* bnx2fc_els.c: QLogic NetXtreme II Linux FCoE offload driver.
|
|
* This file contains helper routines that handle ELS requests
|
|
* and responses.
|
|
*
|
|
* Copyright (c) 2008 - 2013 Broadcom Corporation
|
|
* Copyright (c) 2014, QLogic Corporation
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation.
|
|
*
|
|
* Written by: Bhanu Prakash Gollapudi (bprakash@broadcom.com)
|
|
*/
|
|
|
|
#include "bnx2fc.h"
|
|
|
|
static void bnx2fc_logo_resp(struct fc_seq *seq, struct fc_frame *fp,
|
|
void *arg);
|
|
static void bnx2fc_flogi_resp(struct fc_seq *seq, struct fc_frame *fp,
|
|
void *arg);
|
|
static int bnx2fc_initiate_els(struct bnx2fc_rport *tgt, unsigned int op,
|
|
void *data, u32 data_len,
|
|
void (*cb_func)(struct bnx2fc_els_cb_arg *cb_arg),
|
|
struct bnx2fc_els_cb_arg *cb_arg, u32 timer_msec);
|
|
|
|
static void bnx2fc_rrq_compl(struct bnx2fc_els_cb_arg *cb_arg)
|
|
{
|
|
struct bnx2fc_cmd *orig_io_req;
|
|
struct bnx2fc_cmd *rrq_req;
|
|
int rc = 0;
|
|
|
|
BUG_ON(!cb_arg);
|
|
rrq_req = cb_arg->io_req;
|
|
orig_io_req = cb_arg->aborted_io_req;
|
|
BUG_ON(!orig_io_req);
|
|
BNX2FC_ELS_DBG("rrq_compl: orig xid = 0x%x, rrq_xid = 0x%x\n",
|
|
orig_io_req->xid, rrq_req->xid);
|
|
|
|
kref_put(&orig_io_req->refcount, bnx2fc_cmd_release);
|
|
|
|
if (test_and_clear_bit(BNX2FC_FLAG_ELS_TIMEOUT, &rrq_req->req_flags)) {
|
|
/*
|
|
* els req is timed out. cleanup the IO with FW and
|
|
* drop the completion. Remove from active_cmd_queue.
|
|
*/
|
|
BNX2FC_ELS_DBG("rrq xid - 0x%x timed out, clean it up\n",
|
|
rrq_req->xid);
|
|
|
|
if (rrq_req->on_active_queue) {
|
|
list_del_init(&rrq_req->link);
|
|
rrq_req->on_active_queue = 0;
|
|
rc = bnx2fc_initiate_cleanup(rrq_req);
|
|
BUG_ON(rc);
|
|
}
|
|
}
|
|
kfree(cb_arg);
|
|
}
|
|
int bnx2fc_send_rrq(struct bnx2fc_cmd *aborted_io_req)
|
|
{
|
|
|
|
struct fc_els_rrq rrq;
|
|
struct bnx2fc_rport *tgt = aborted_io_req->tgt;
|
|
struct fc_lport *lport = tgt->rdata->local_port;
|
|
struct bnx2fc_els_cb_arg *cb_arg = NULL;
|
|
u32 sid = tgt->sid;
|
|
u32 r_a_tov = lport->r_a_tov;
|
|
unsigned long start = jiffies;
|
|
int rc;
|
|
|
|
BNX2FC_ELS_DBG("Sending RRQ orig_xid = 0x%x\n",
|
|
aborted_io_req->xid);
|
|
memset(&rrq, 0, sizeof(rrq));
|
|
|
|
cb_arg = kzalloc(sizeof(struct bnx2fc_els_cb_arg), GFP_NOIO);
|
|
if (!cb_arg) {
|
|
printk(KERN_ERR PFX "Unable to allocate cb_arg for RRQ\n");
|
|
rc = -ENOMEM;
|
|
goto rrq_err;
|
|
}
|
|
|
|
cb_arg->aborted_io_req = aborted_io_req;
|
|
|
|
rrq.rrq_cmd = ELS_RRQ;
|
|
hton24(rrq.rrq_s_id, sid);
|
|
rrq.rrq_ox_id = htons(aborted_io_req->xid);
|
|
rrq.rrq_rx_id = htons(aborted_io_req->task->rxwr_txrd.var_ctx.rx_id);
|
|
|
|
retry_rrq:
|
|
rc = bnx2fc_initiate_els(tgt, ELS_RRQ, &rrq, sizeof(rrq),
|
|
bnx2fc_rrq_compl, cb_arg,
|
|
r_a_tov);
|
|
if (rc == -ENOMEM) {
|
|
if (time_after(jiffies, start + (10 * HZ))) {
|
|
BNX2FC_ELS_DBG("rrq Failed\n");
|
|
rc = FAILED;
|
|
goto rrq_err;
|
|
}
|
|
msleep(20);
|
|
goto retry_rrq;
|
|
}
|
|
rrq_err:
|
|
if (rc) {
|
|
BNX2FC_ELS_DBG("RRQ failed - release orig io req 0x%x\n",
|
|
aborted_io_req->xid);
|
|
kfree(cb_arg);
|
|
spin_lock_bh(&tgt->tgt_lock);
|
|
kref_put(&aborted_io_req->refcount, bnx2fc_cmd_release);
|
|
spin_unlock_bh(&tgt->tgt_lock);
|
|
}
|
|
return rc;
|
|
}
|
|
|
|
static void bnx2fc_l2_els_compl(struct bnx2fc_els_cb_arg *cb_arg)
|
|
{
|
|
struct bnx2fc_cmd *els_req;
|
|
struct bnx2fc_rport *tgt;
|
|
struct bnx2fc_mp_req *mp_req;
|
|
struct fc_frame_header *fc_hdr;
|
|
unsigned char *buf;
|
|
void *resp_buf;
|
|
u32 resp_len, hdr_len;
|
|
u16 l2_oxid;
|
|
int frame_len;
|
|
int rc = 0;
|
|
|
|
l2_oxid = cb_arg->l2_oxid;
|
|
BNX2FC_ELS_DBG("ELS COMPL - l2_oxid = 0x%x\n", l2_oxid);
|
|
|
|
els_req = cb_arg->io_req;
|
|
if (test_and_clear_bit(BNX2FC_FLAG_ELS_TIMEOUT, &els_req->req_flags)) {
|
|
/*
|
|
* els req is timed out. cleanup the IO with FW and
|
|
* drop the completion. libfc will handle the els timeout
|
|
*/
|
|
if (els_req->on_active_queue) {
|
|
list_del_init(&els_req->link);
|
|
els_req->on_active_queue = 0;
|
|
rc = bnx2fc_initiate_cleanup(els_req);
|
|
BUG_ON(rc);
|
|
}
|
|
goto free_arg;
|
|
}
|
|
|
|
tgt = els_req->tgt;
|
|
mp_req = &(els_req->mp_req);
|
|
fc_hdr = &(mp_req->resp_fc_hdr);
|
|
resp_len = mp_req->resp_len;
|
|
resp_buf = mp_req->resp_buf;
|
|
|
|
buf = kzalloc(PAGE_SIZE, GFP_ATOMIC);
|
|
if (!buf) {
|
|
printk(KERN_ERR PFX "Unable to alloc mp buf\n");
|
|
goto free_arg;
|
|
}
|
|
hdr_len = sizeof(*fc_hdr);
|
|
if (hdr_len + resp_len > PAGE_SIZE) {
|
|
printk(KERN_ERR PFX "l2_els_compl: resp len is "
|
|
"beyond page size\n");
|
|
goto free_buf;
|
|
}
|
|
memcpy(buf, fc_hdr, hdr_len);
|
|
memcpy(buf + hdr_len, resp_buf, resp_len);
|
|
frame_len = hdr_len + resp_len;
|
|
|
|
bnx2fc_process_l2_frame_compl(tgt, buf, frame_len, l2_oxid);
|
|
|
|
free_buf:
|
|
kfree(buf);
|
|
free_arg:
|
|
kfree(cb_arg);
|
|
}
|
|
|
|
int bnx2fc_send_adisc(struct bnx2fc_rport *tgt, struct fc_frame *fp)
|
|
{
|
|
struct fc_els_adisc *adisc;
|
|
struct fc_frame_header *fh;
|
|
struct bnx2fc_els_cb_arg *cb_arg;
|
|
struct fc_lport *lport = tgt->rdata->local_port;
|
|
u32 r_a_tov = lport->r_a_tov;
|
|
int rc;
|
|
|
|
fh = fc_frame_header_get(fp);
|
|
cb_arg = kzalloc(sizeof(struct bnx2fc_els_cb_arg), GFP_ATOMIC);
|
|
if (!cb_arg) {
|
|
printk(KERN_ERR PFX "Unable to allocate cb_arg for ADISC\n");
|
|
return -ENOMEM;
|
|
}
|
|
|
|
cb_arg->l2_oxid = ntohs(fh->fh_ox_id);
|
|
|
|
BNX2FC_ELS_DBG("send ADISC: l2_oxid = 0x%x\n", cb_arg->l2_oxid);
|
|
adisc = fc_frame_payload_get(fp, sizeof(*adisc));
|
|
/* adisc is initialized by libfc */
|
|
rc = bnx2fc_initiate_els(tgt, ELS_ADISC, adisc, sizeof(*adisc),
|
|
bnx2fc_l2_els_compl, cb_arg, 2 * r_a_tov);
|
|
if (rc)
|
|
kfree(cb_arg);
|
|
return rc;
|
|
}
|
|
|
|
int bnx2fc_send_logo(struct bnx2fc_rport *tgt, struct fc_frame *fp)
|
|
{
|
|
struct fc_els_logo *logo;
|
|
struct fc_frame_header *fh;
|
|
struct bnx2fc_els_cb_arg *cb_arg;
|
|
struct fc_lport *lport = tgt->rdata->local_port;
|
|
u32 r_a_tov = lport->r_a_tov;
|
|
int rc;
|
|
|
|
fh = fc_frame_header_get(fp);
|
|
cb_arg = kzalloc(sizeof(struct bnx2fc_els_cb_arg), GFP_ATOMIC);
|
|
if (!cb_arg) {
|
|
printk(KERN_ERR PFX "Unable to allocate cb_arg for LOGO\n");
|
|
return -ENOMEM;
|
|
}
|
|
|
|
cb_arg->l2_oxid = ntohs(fh->fh_ox_id);
|
|
|
|
BNX2FC_ELS_DBG("Send LOGO: l2_oxid = 0x%x\n", cb_arg->l2_oxid);
|
|
logo = fc_frame_payload_get(fp, sizeof(*logo));
|
|
/* logo is initialized by libfc */
|
|
rc = bnx2fc_initiate_els(tgt, ELS_LOGO, logo, sizeof(*logo),
|
|
bnx2fc_l2_els_compl, cb_arg, 2 * r_a_tov);
|
|
if (rc)
|
|
kfree(cb_arg);
|
|
return rc;
|
|
}
|
|
|
|
int bnx2fc_send_rls(struct bnx2fc_rport *tgt, struct fc_frame *fp)
|
|
{
|
|
struct fc_els_rls *rls;
|
|
struct fc_frame_header *fh;
|
|
struct bnx2fc_els_cb_arg *cb_arg;
|
|
struct fc_lport *lport = tgt->rdata->local_port;
|
|
u32 r_a_tov = lport->r_a_tov;
|
|
int rc;
|
|
|
|
fh = fc_frame_header_get(fp);
|
|
cb_arg = kzalloc(sizeof(struct bnx2fc_els_cb_arg), GFP_ATOMIC);
|
|
if (!cb_arg) {
|
|
printk(KERN_ERR PFX "Unable to allocate cb_arg for LOGO\n");
|
|
return -ENOMEM;
|
|
}
|
|
|
|
cb_arg->l2_oxid = ntohs(fh->fh_ox_id);
|
|
|
|
rls = fc_frame_payload_get(fp, sizeof(*rls));
|
|
/* rls is initialized by libfc */
|
|
rc = bnx2fc_initiate_els(tgt, ELS_RLS, rls, sizeof(*rls),
|
|
bnx2fc_l2_els_compl, cb_arg, 2 * r_a_tov);
|
|
if (rc)
|
|
kfree(cb_arg);
|
|
return rc;
|
|
}
|
|
|
|
void bnx2fc_srr_compl(struct bnx2fc_els_cb_arg *cb_arg)
|
|
{
|
|
struct bnx2fc_mp_req *mp_req;
|
|
struct fc_frame_header *fc_hdr, *fh;
|
|
struct bnx2fc_cmd *srr_req;
|
|
struct bnx2fc_cmd *orig_io_req;
|
|
struct fc_frame *fp;
|
|
unsigned char *buf;
|
|
void *resp_buf;
|
|
u32 resp_len, hdr_len;
|
|
u8 opcode;
|
|
int rc = 0;
|
|
|
|
orig_io_req = cb_arg->aborted_io_req;
|
|
srr_req = cb_arg->io_req;
|
|
if (test_and_clear_bit(BNX2FC_FLAG_ELS_TIMEOUT, &srr_req->req_flags)) {
|
|
/* SRR timedout */
|
|
BNX2FC_IO_DBG(srr_req, "srr timed out, abort "
|
|
"orig_io - 0x%x\n",
|
|
orig_io_req->xid);
|
|
rc = bnx2fc_initiate_abts(srr_req);
|
|
if (rc != SUCCESS) {
|
|
BNX2FC_IO_DBG(srr_req, "srr_compl: initiate_abts "
|
|
"failed. issue cleanup\n");
|
|
bnx2fc_initiate_cleanup(srr_req);
|
|
}
|
|
if (test_bit(BNX2FC_FLAG_IO_COMPL, &orig_io_req->req_flags) ||
|
|
test_bit(BNX2FC_FLAG_ISSUE_ABTS, &orig_io_req->req_flags)) {
|
|
BNX2FC_IO_DBG(srr_req, "srr_compl:xid 0x%x flags = %lx",
|
|
orig_io_req->xid, orig_io_req->req_flags);
|
|
goto srr_compl_done;
|
|
}
|
|
orig_io_req->srr_retry++;
|
|
if (orig_io_req->srr_retry <= SRR_RETRY_COUNT) {
|
|
struct bnx2fc_rport *tgt = orig_io_req->tgt;
|
|
spin_unlock_bh(&tgt->tgt_lock);
|
|
rc = bnx2fc_send_srr(orig_io_req,
|
|
orig_io_req->srr_offset,
|
|
orig_io_req->srr_rctl);
|
|
spin_lock_bh(&tgt->tgt_lock);
|
|
if (!rc)
|
|
goto srr_compl_done;
|
|
}
|
|
|
|
rc = bnx2fc_initiate_abts(orig_io_req);
|
|
if (rc != SUCCESS) {
|
|
BNX2FC_IO_DBG(srr_req, "srr_compl: initiate_abts "
|
|
"failed xid = 0x%x. issue cleanup\n",
|
|
orig_io_req->xid);
|
|
bnx2fc_initiate_cleanup(orig_io_req);
|
|
}
|
|
goto srr_compl_done;
|
|
}
|
|
if (test_bit(BNX2FC_FLAG_IO_COMPL, &orig_io_req->req_flags) ||
|
|
test_bit(BNX2FC_FLAG_ISSUE_ABTS, &orig_io_req->req_flags)) {
|
|
BNX2FC_IO_DBG(srr_req, "srr_compl:xid - 0x%x flags = %lx",
|
|
orig_io_req->xid, orig_io_req->req_flags);
|
|
goto srr_compl_done;
|
|
}
|
|
mp_req = &(srr_req->mp_req);
|
|
fc_hdr = &(mp_req->resp_fc_hdr);
|
|
resp_len = mp_req->resp_len;
|
|
resp_buf = mp_req->resp_buf;
|
|
|
|
hdr_len = sizeof(*fc_hdr);
|
|
buf = kzalloc(PAGE_SIZE, GFP_ATOMIC);
|
|
if (!buf) {
|
|
printk(KERN_ERR PFX "srr buf: mem alloc failure\n");
|
|
goto srr_compl_done;
|
|
}
|
|
memcpy(buf, fc_hdr, hdr_len);
|
|
memcpy(buf + hdr_len, resp_buf, resp_len);
|
|
|
|
fp = fc_frame_alloc(NULL, resp_len);
|
|
if (!fp) {
|
|
printk(KERN_ERR PFX "fc_frame_alloc failure\n");
|
|
goto free_buf;
|
|
}
|
|
|
|
fh = (struct fc_frame_header *) fc_frame_header_get(fp);
|
|
/* Copy FC Frame header and payload into the frame */
|
|
memcpy(fh, buf, hdr_len + resp_len);
|
|
|
|
opcode = fc_frame_payload_op(fp);
|
|
switch (opcode) {
|
|
case ELS_LS_ACC:
|
|
BNX2FC_IO_DBG(srr_req, "SRR success\n");
|
|
break;
|
|
case ELS_LS_RJT:
|
|
BNX2FC_IO_DBG(srr_req, "SRR rejected\n");
|
|
rc = bnx2fc_initiate_abts(orig_io_req);
|
|
if (rc != SUCCESS) {
|
|
BNX2FC_IO_DBG(srr_req, "srr_compl: initiate_abts "
|
|
"failed xid = 0x%x. issue cleanup\n",
|
|
orig_io_req->xid);
|
|
bnx2fc_initiate_cleanup(orig_io_req);
|
|
}
|
|
break;
|
|
default:
|
|
BNX2FC_IO_DBG(srr_req, "srr compl - invalid opcode = %d\n",
|
|
opcode);
|
|
break;
|
|
}
|
|
fc_frame_free(fp);
|
|
free_buf:
|
|
kfree(buf);
|
|
srr_compl_done:
|
|
kref_put(&orig_io_req->refcount, bnx2fc_cmd_release);
|
|
}
|
|
|
|
void bnx2fc_rec_compl(struct bnx2fc_els_cb_arg *cb_arg)
|
|
{
|
|
struct bnx2fc_cmd *orig_io_req, *new_io_req;
|
|
struct bnx2fc_cmd *rec_req;
|
|
struct bnx2fc_mp_req *mp_req;
|
|
struct fc_frame_header *fc_hdr, *fh;
|
|
struct fc_els_ls_rjt *rjt;
|
|
struct fc_els_rec_acc *acc;
|
|
struct bnx2fc_rport *tgt;
|
|
struct fcoe_err_report_entry *err_entry;
|
|
struct scsi_cmnd *sc_cmd;
|
|
enum fc_rctl r_ctl;
|
|
unsigned char *buf;
|
|
void *resp_buf;
|
|
struct fc_frame *fp;
|
|
u8 opcode;
|
|
u32 offset;
|
|
u32 e_stat;
|
|
u32 resp_len, hdr_len;
|
|
int rc = 0;
|
|
bool send_seq_clnp = false;
|
|
bool abort_io = false;
|
|
|
|
BNX2FC_MISC_DBG("Entered rec_compl callback\n");
|
|
rec_req = cb_arg->io_req;
|
|
orig_io_req = cb_arg->aborted_io_req;
|
|
BNX2FC_IO_DBG(rec_req, "rec_compl: orig xid = 0x%x", orig_io_req->xid);
|
|
tgt = orig_io_req->tgt;
|
|
|
|
/* Handle REC timeout case */
|
|
if (test_and_clear_bit(BNX2FC_FLAG_ELS_TIMEOUT, &rec_req->req_flags)) {
|
|
BNX2FC_IO_DBG(rec_req, "timed out, abort "
|
|
"orig_io - 0x%x\n",
|
|
orig_io_req->xid);
|
|
/* els req is timed out. send abts for els */
|
|
rc = bnx2fc_initiate_abts(rec_req);
|
|
if (rc != SUCCESS) {
|
|
BNX2FC_IO_DBG(rec_req, "rec_compl: initiate_abts "
|
|
"failed. issue cleanup\n");
|
|
bnx2fc_initiate_cleanup(rec_req);
|
|
}
|
|
orig_io_req->rec_retry++;
|
|
/* REC timedout. send ABTS to the orig IO req */
|
|
if (orig_io_req->rec_retry <= REC_RETRY_COUNT) {
|
|
spin_unlock_bh(&tgt->tgt_lock);
|
|
rc = bnx2fc_send_rec(orig_io_req);
|
|
spin_lock_bh(&tgt->tgt_lock);
|
|
if (!rc)
|
|
goto rec_compl_done;
|
|
}
|
|
rc = bnx2fc_initiate_abts(orig_io_req);
|
|
if (rc != SUCCESS) {
|
|
BNX2FC_IO_DBG(rec_req, "rec_compl: initiate_abts "
|
|
"failed xid = 0x%x. issue cleanup\n",
|
|
orig_io_req->xid);
|
|
bnx2fc_initiate_cleanup(orig_io_req);
|
|
}
|
|
goto rec_compl_done;
|
|
}
|
|
|
|
if (test_bit(BNX2FC_FLAG_IO_COMPL, &orig_io_req->req_flags)) {
|
|
BNX2FC_IO_DBG(rec_req, "completed"
|
|
"orig_io - 0x%x\n",
|
|
orig_io_req->xid);
|
|
goto rec_compl_done;
|
|
}
|
|
if (test_bit(BNX2FC_FLAG_ISSUE_ABTS, &orig_io_req->req_flags)) {
|
|
BNX2FC_IO_DBG(rec_req, "abts in prog "
|
|
"orig_io - 0x%x\n",
|
|
orig_io_req->xid);
|
|
goto rec_compl_done;
|
|
}
|
|
|
|
mp_req = &(rec_req->mp_req);
|
|
fc_hdr = &(mp_req->resp_fc_hdr);
|
|
resp_len = mp_req->resp_len;
|
|
acc = resp_buf = mp_req->resp_buf;
|
|
|
|
hdr_len = sizeof(*fc_hdr);
|
|
|
|
buf = kzalloc(PAGE_SIZE, GFP_ATOMIC);
|
|
if (!buf) {
|
|
printk(KERN_ERR PFX "rec buf: mem alloc failure\n");
|
|
goto rec_compl_done;
|
|
}
|
|
memcpy(buf, fc_hdr, hdr_len);
|
|
memcpy(buf + hdr_len, resp_buf, resp_len);
|
|
|
|
fp = fc_frame_alloc(NULL, resp_len);
|
|
if (!fp) {
|
|
printk(KERN_ERR PFX "fc_frame_alloc failure\n");
|
|
goto free_buf;
|
|
}
|
|
|
|
fh = (struct fc_frame_header *) fc_frame_header_get(fp);
|
|
/* Copy FC Frame header and payload into the frame */
|
|
memcpy(fh, buf, hdr_len + resp_len);
|
|
|
|
opcode = fc_frame_payload_op(fp);
|
|
if (opcode == ELS_LS_RJT) {
|
|
BNX2FC_IO_DBG(rec_req, "opcode is RJT\n");
|
|
rjt = fc_frame_payload_get(fp, sizeof(*rjt));
|
|
if ((rjt->er_reason == ELS_RJT_LOGIC ||
|
|
rjt->er_reason == ELS_RJT_UNAB) &&
|
|
rjt->er_explan == ELS_EXPL_OXID_RXID) {
|
|
BNX2FC_IO_DBG(rec_req, "handle CMD LOST case\n");
|
|
new_io_req = bnx2fc_cmd_alloc(tgt);
|
|
if (!new_io_req)
|
|
goto abort_io;
|
|
new_io_req->sc_cmd = orig_io_req->sc_cmd;
|
|
/* cleanup orig_io_req that is with the FW */
|
|
set_bit(BNX2FC_FLAG_CMD_LOST,
|
|
&orig_io_req->req_flags);
|
|
bnx2fc_initiate_cleanup(orig_io_req);
|
|
/* Post a new IO req with the same sc_cmd */
|
|
BNX2FC_IO_DBG(rec_req, "Post IO request again\n");
|
|
rc = bnx2fc_post_io_req(tgt, new_io_req);
|
|
if (!rc)
|
|
goto free_frame;
|
|
BNX2FC_IO_DBG(rec_req, "REC: io post err\n");
|
|
}
|
|
abort_io:
|
|
rc = bnx2fc_initiate_abts(orig_io_req);
|
|
if (rc != SUCCESS) {
|
|
BNX2FC_IO_DBG(rec_req, "rec_compl: initiate_abts "
|
|
"failed. issue cleanup\n");
|
|
bnx2fc_initiate_cleanup(orig_io_req);
|
|
}
|
|
} else if (opcode == ELS_LS_ACC) {
|
|
/* REVISIT: Check if the exchange is already aborted */
|
|
offset = ntohl(acc->reca_fc4value);
|
|
e_stat = ntohl(acc->reca_e_stat);
|
|
if (e_stat & ESB_ST_SEQ_INIT) {
|
|
BNX2FC_IO_DBG(rec_req, "target has the seq init\n");
|
|
goto free_frame;
|
|
}
|
|
BNX2FC_IO_DBG(rec_req, "e_stat = 0x%x, offset = 0x%x\n",
|
|
e_stat, offset);
|
|
/* Seq initiative is with us */
|
|
err_entry = (struct fcoe_err_report_entry *)
|
|
&orig_io_req->err_entry;
|
|
sc_cmd = orig_io_req->sc_cmd;
|
|
if (sc_cmd->sc_data_direction == DMA_TO_DEVICE) {
|
|
/* SCSI WRITE command */
|
|
if (offset == orig_io_req->data_xfer_len) {
|
|
BNX2FC_IO_DBG(rec_req, "WRITE - resp lost\n");
|
|
/* FCP_RSP lost */
|
|
r_ctl = FC_RCTL_DD_CMD_STATUS;
|
|
offset = 0;
|
|
} else {
|
|
/* start transmitting from offset */
|
|
BNX2FC_IO_DBG(rec_req, "XFER_RDY/DATA lost\n");
|
|
send_seq_clnp = true;
|
|
r_ctl = FC_RCTL_DD_DATA_DESC;
|
|
if (bnx2fc_initiate_seq_cleanup(orig_io_req,
|
|
offset, r_ctl))
|
|
abort_io = true;
|
|
/* XFER_RDY */
|
|
}
|
|
} else {
|
|
/* SCSI READ command */
|
|
if (err_entry->data.rx_buf_off ==
|
|
orig_io_req->data_xfer_len) {
|
|
/* FCP_RSP lost */
|
|
BNX2FC_IO_DBG(rec_req, "READ - resp lost\n");
|
|
r_ctl = FC_RCTL_DD_CMD_STATUS;
|
|
offset = 0;
|
|
} else {
|
|
/* request retransmission from this offset */
|
|
send_seq_clnp = true;
|
|
offset = err_entry->data.rx_buf_off;
|
|
BNX2FC_IO_DBG(rec_req, "RD DATA lost\n");
|
|
/* FCP_DATA lost */
|
|
r_ctl = FC_RCTL_DD_SOL_DATA;
|
|
if (bnx2fc_initiate_seq_cleanup(orig_io_req,
|
|
offset, r_ctl))
|
|
abort_io = true;
|
|
}
|
|
}
|
|
if (abort_io) {
|
|
rc = bnx2fc_initiate_abts(orig_io_req);
|
|
if (rc != SUCCESS) {
|
|
BNX2FC_IO_DBG(rec_req, "rec_compl:initiate_abts"
|
|
" failed. issue cleanup\n");
|
|
bnx2fc_initiate_cleanup(orig_io_req);
|
|
}
|
|
} else if (!send_seq_clnp) {
|
|
BNX2FC_IO_DBG(rec_req, "Send SRR - FCP_RSP\n");
|
|
spin_unlock_bh(&tgt->tgt_lock);
|
|
rc = bnx2fc_send_srr(orig_io_req, offset, r_ctl);
|
|
spin_lock_bh(&tgt->tgt_lock);
|
|
|
|
if (rc) {
|
|
BNX2FC_IO_DBG(rec_req, "Unable to send SRR"
|
|
" IO will abort\n");
|
|
}
|
|
}
|
|
}
|
|
free_frame:
|
|
fc_frame_free(fp);
|
|
free_buf:
|
|
kfree(buf);
|
|
rec_compl_done:
|
|
kref_put(&orig_io_req->refcount, bnx2fc_cmd_release);
|
|
kfree(cb_arg);
|
|
}
|
|
|
|
int bnx2fc_send_rec(struct bnx2fc_cmd *orig_io_req)
|
|
{
|
|
struct fc_els_rec rec;
|
|
struct bnx2fc_rport *tgt = orig_io_req->tgt;
|
|
struct fc_lport *lport = tgt->rdata->local_port;
|
|
struct bnx2fc_els_cb_arg *cb_arg = NULL;
|
|
u32 sid = tgt->sid;
|
|
u32 r_a_tov = lport->r_a_tov;
|
|
int rc;
|
|
|
|
BNX2FC_IO_DBG(orig_io_req, "Sending REC\n");
|
|
memset(&rec, 0, sizeof(rec));
|
|
|
|
cb_arg = kzalloc(sizeof(struct bnx2fc_els_cb_arg), GFP_ATOMIC);
|
|
if (!cb_arg) {
|
|
printk(KERN_ERR PFX "Unable to allocate cb_arg for REC\n");
|
|
rc = -ENOMEM;
|
|
goto rec_err;
|
|
}
|
|
kref_get(&orig_io_req->refcount);
|
|
|
|
cb_arg->aborted_io_req = orig_io_req;
|
|
|
|
rec.rec_cmd = ELS_REC;
|
|
hton24(rec.rec_s_id, sid);
|
|
rec.rec_ox_id = htons(orig_io_req->xid);
|
|
rec.rec_rx_id = htons(orig_io_req->task->rxwr_txrd.var_ctx.rx_id);
|
|
|
|
rc = bnx2fc_initiate_els(tgt, ELS_REC, &rec, sizeof(rec),
|
|
bnx2fc_rec_compl, cb_arg,
|
|
r_a_tov);
|
|
rec_err:
|
|
if (rc) {
|
|
BNX2FC_IO_DBG(orig_io_req, "REC failed - release\n");
|
|
spin_lock_bh(&tgt->tgt_lock);
|
|
kref_put(&orig_io_req->refcount, bnx2fc_cmd_release);
|
|
spin_unlock_bh(&tgt->tgt_lock);
|
|
kfree(cb_arg);
|
|
}
|
|
return rc;
|
|
}
|
|
|
|
int bnx2fc_send_srr(struct bnx2fc_cmd *orig_io_req, u32 offset, u8 r_ctl)
|
|
{
|
|
struct fcp_srr srr;
|
|
struct bnx2fc_rport *tgt = orig_io_req->tgt;
|
|
struct fc_lport *lport = tgt->rdata->local_port;
|
|
struct bnx2fc_els_cb_arg *cb_arg = NULL;
|
|
u32 r_a_tov = lport->r_a_tov;
|
|
int rc;
|
|
|
|
BNX2FC_IO_DBG(orig_io_req, "Sending SRR\n");
|
|
memset(&srr, 0, sizeof(srr));
|
|
|
|
cb_arg = kzalloc(sizeof(struct bnx2fc_els_cb_arg), GFP_ATOMIC);
|
|
if (!cb_arg) {
|
|
printk(KERN_ERR PFX "Unable to allocate cb_arg for SRR\n");
|
|
rc = -ENOMEM;
|
|
goto srr_err;
|
|
}
|
|
kref_get(&orig_io_req->refcount);
|
|
|
|
cb_arg->aborted_io_req = orig_io_req;
|
|
|
|
srr.srr_op = ELS_SRR;
|
|
srr.srr_ox_id = htons(orig_io_req->xid);
|
|
srr.srr_rx_id = htons(orig_io_req->task->rxwr_txrd.var_ctx.rx_id);
|
|
srr.srr_rel_off = htonl(offset);
|
|
srr.srr_r_ctl = r_ctl;
|
|
orig_io_req->srr_offset = offset;
|
|
orig_io_req->srr_rctl = r_ctl;
|
|
|
|
rc = bnx2fc_initiate_els(tgt, ELS_SRR, &srr, sizeof(srr),
|
|
bnx2fc_srr_compl, cb_arg,
|
|
r_a_tov);
|
|
srr_err:
|
|
if (rc) {
|
|
BNX2FC_IO_DBG(orig_io_req, "SRR failed - release\n");
|
|
spin_lock_bh(&tgt->tgt_lock);
|
|
kref_put(&orig_io_req->refcount, bnx2fc_cmd_release);
|
|
spin_unlock_bh(&tgt->tgt_lock);
|
|
kfree(cb_arg);
|
|
} else
|
|
set_bit(BNX2FC_FLAG_SRR_SENT, &orig_io_req->req_flags);
|
|
|
|
return rc;
|
|
}
|
|
|
|
static int bnx2fc_initiate_els(struct bnx2fc_rport *tgt, unsigned int op,
|
|
void *data, u32 data_len,
|
|
void (*cb_func)(struct bnx2fc_els_cb_arg *cb_arg),
|
|
struct bnx2fc_els_cb_arg *cb_arg, u32 timer_msec)
|
|
{
|
|
struct fcoe_port *port = tgt->port;
|
|
struct bnx2fc_interface *interface = port->priv;
|
|
struct fc_rport *rport = tgt->rport;
|
|
struct fc_lport *lport = port->lport;
|
|
struct bnx2fc_cmd *els_req;
|
|
struct bnx2fc_mp_req *mp_req;
|
|
struct fc_frame_header *fc_hdr;
|
|
struct fcoe_task_ctx_entry *task;
|
|
struct fcoe_task_ctx_entry *task_page;
|
|
int rc = 0;
|
|
int task_idx, index;
|
|
u32 did, sid;
|
|
u16 xid;
|
|
|
|
rc = fc_remote_port_chkready(rport);
|
|
if (rc) {
|
|
printk(KERN_ERR PFX "els 0x%x: rport not ready\n", op);
|
|
rc = -EINVAL;
|
|
goto els_err;
|
|
}
|
|
if (lport->state != LPORT_ST_READY || !(lport->link_up)) {
|
|
printk(KERN_ERR PFX "els 0x%x: link is not ready\n", op);
|
|
rc = -EINVAL;
|
|
goto els_err;
|
|
}
|
|
if (!(test_bit(BNX2FC_FLAG_SESSION_READY, &tgt->flags)) ||
|
|
(test_bit(BNX2FC_FLAG_EXPL_LOGO, &tgt->flags))) {
|
|
printk(KERN_ERR PFX "els 0x%x: tgt not ready\n", op);
|
|
rc = -EINVAL;
|
|
goto els_err;
|
|
}
|
|
els_req = bnx2fc_elstm_alloc(tgt, BNX2FC_ELS);
|
|
if (!els_req) {
|
|
rc = -ENOMEM;
|
|
goto els_err;
|
|
}
|
|
|
|
els_req->sc_cmd = NULL;
|
|
els_req->port = port;
|
|
els_req->tgt = tgt;
|
|
els_req->cb_func = cb_func;
|
|
cb_arg->io_req = els_req;
|
|
els_req->cb_arg = cb_arg;
|
|
|
|
mp_req = (struct bnx2fc_mp_req *)&(els_req->mp_req);
|
|
rc = bnx2fc_init_mp_req(els_req);
|
|
if (rc == FAILED) {
|
|
printk(KERN_ERR PFX "ELS MP request init failed\n");
|
|
spin_lock_bh(&tgt->tgt_lock);
|
|
kref_put(&els_req->refcount, bnx2fc_cmd_release);
|
|
spin_unlock_bh(&tgt->tgt_lock);
|
|
rc = -ENOMEM;
|
|
goto els_err;
|
|
} else {
|
|
/* rc SUCCESS */
|
|
rc = 0;
|
|
}
|
|
|
|
/* Set the data_xfer_len to the size of ELS payload */
|
|
mp_req->req_len = data_len;
|
|
els_req->data_xfer_len = mp_req->req_len;
|
|
|
|
/* Fill ELS Payload */
|
|
if ((op >= ELS_LS_RJT) && (op <= ELS_AUTH_ELS)) {
|
|
memcpy(mp_req->req_buf, data, data_len);
|
|
} else {
|
|
printk(KERN_ERR PFX "Invalid ELS op 0x%x\n", op);
|
|
els_req->cb_func = NULL;
|
|
els_req->cb_arg = NULL;
|
|
spin_lock_bh(&tgt->tgt_lock);
|
|
kref_put(&els_req->refcount, bnx2fc_cmd_release);
|
|
spin_unlock_bh(&tgt->tgt_lock);
|
|
rc = -EINVAL;
|
|
}
|
|
|
|
if (rc)
|
|
goto els_err;
|
|
|
|
/* Fill FC header */
|
|
fc_hdr = &(mp_req->req_fc_hdr);
|
|
|
|
did = tgt->rport->port_id;
|
|
sid = tgt->sid;
|
|
|
|
if (op == ELS_SRR)
|
|
__fc_fill_fc_hdr(fc_hdr, FC_RCTL_ELS4_REQ, did, sid,
|
|
FC_TYPE_FCP, FC_FC_FIRST_SEQ |
|
|
FC_FC_END_SEQ | FC_FC_SEQ_INIT, 0);
|
|
else
|
|
__fc_fill_fc_hdr(fc_hdr, FC_RCTL_ELS_REQ, did, sid,
|
|
FC_TYPE_ELS, FC_FC_FIRST_SEQ |
|
|
FC_FC_END_SEQ | FC_FC_SEQ_INIT, 0);
|
|
|
|
/* Obtain exchange id */
|
|
xid = els_req->xid;
|
|
task_idx = xid/BNX2FC_TASKS_PER_PAGE;
|
|
index = xid % BNX2FC_TASKS_PER_PAGE;
|
|
|
|
/* Initialize task context for this IO request */
|
|
task_page = (struct fcoe_task_ctx_entry *)
|
|
interface->hba->task_ctx[task_idx];
|
|
task = &(task_page[index]);
|
|
bnx2fc_init_mp_task(els_req, task);
|
|
|
|
spin_lock_bh(&tgt->tgt_lock);
|
|
|
|
if (!test_bit(BNX2FC_FLAG_SESSION_READY, &tgt->flags)) {
|
|
printk(KERN_ERR PFX "initiate_els.. session not ready\n");
|
|
els_req->cb_func = NULL;
|
|
els_req->cb_arg = NULL;
|
|
kref_put(&els_req->refcount, bnx2fc_cmd_release);
|
|
spin_unlock_bh(&tgt->tgt_lock);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (timer_msec)
|
|
bnx2fc_cmd_timer_set(els_req, timer_msec);
|
|
bnx2fc_add_2_sq(tgt, xid);
|
|
|
|
els_req->on_active_queue = 1;
|
|
list_add_tail(&els_req->link, &tgt->els_queue);
|
|
|
|
/* Ring doorbell */
|
|
bnx2fc_ring_doorbell(tgt);
|
|
spin_unlock_bh(&tgt->tgt_lock);
|
|
|
|
els_err:
|
|
return rc;
|
|
}
|
|
|
|
void bnx2fc_process_els_compl(struct bnx2fc_cmd *els_req,
|
|
struct fcoe_task_ctx_entry *task, u8 num_rq)
|
|
{
|
|
struct bnx2fc_mp_req *mp_req;
|
|
struct fc_frame_header *fc_hdr;
|
|
u64 *hdr;
|
|
u64 *temp_hdr;
|
|
|
|
BNX2FC_ELS_DBG("Entered process_els_compl xid = 0x%x"
|
|
"cmd_type = %d\n", els_req->xid, els_req->cmd_type);
|
|
|
|
if (test_and_set_bit(BNX2FC_FLAG_ELS_DONE,
|
|
&els_req->req_flags)) {
|
|
BNX2FC_ELS_DBG("Timer context finished processing this "
|
|
"els - 0x%x\n", els_req->xid);
|
|
/* This IO doesn't receive cleanup completion */
|
|
kref_put(&els_req->refcount, bnx2fc_cmd_release);
|
|
return;
|
|
}
|
|
|
|
/* Cancel the timeout_work, as we received the response */
|
|
if (cancel_delayed_work(&els_req->timeout_work))
|
|
kref_put(&els_req->refcount,
|
|
bnx2fc_cmd_release); /* drop timer hold */
|
|
|
|
if (els_req->on_active_queue) {
|
|
list_del_init(&els_req->link);
|
|
els_req->on_active_queue = 0;
|
|
}
|
|
|
|
mp_req = &(els_req->mp_req);
|
|
fc_hdr = &(mp_req->resp_fc_hdr);
|
|
|
|
hdr = (u64 *)fc_hdr;
|
|
temp_hdr = (u64 *)
|
|
&task->rxwr_only.union_ctx.comp_info.mp_rsp.fc_hdr;
|
|
hdr[0] = cpu_to_be64(temp_hdr[0]);
|
|
hdr[1] = cpu_to_be64(temp_hdr[1]);
|
|
hdr[2] = cpu_to_be64(temp_hdr[2]);
|
|
|
|
mp_req->resp_len =
|
|
task->rxwr_only.union_ctx.comp_info.mp_rsp.mp_payload_len;
|
|
|
|
/* Parse ELS response */
|
|
if ((els_req->cb_func) && (els_req->cb_arg)) {
|
|
els_req->cb_func(els_req->cb_arg);
|
|
els_req->cb_arg = NULL;
|
|
}
|
|
|
|
kref_put(&els_req->refcount, bnx2fc_cmd_release);
|
|
}
|
|
|
|
static void bnx2fc_flogi_resp(struct fc_seq *seq, struct fc_frame *fp,
|
|
void *arg)
|
|
{
|
|
struct fcoe_ctlr *fip = arg;
|
|
struct fc_exch *exch = fc_seq_exch(seq);
|
|
struct fc_lport *lport = exch->lp;
|
|
u8 *mac;
|
|
u8 op;
|
|
|
|
if (IS_ERR(fp))
|
|
goto done;
|
|
|
|
mac = fr_cb(fp)->granted_mac;
|
|
if (is_zero_ether_addr(mac)) {
|
|
op = fc_frame_payload_op(fp);
|
|
if (lport->vport) {
|
|
if (op == ELS_LS_RJT) {
|
|
printk(KERN_ERR PFX "bnx2fc_flogi_resp is LS_RJT\n");
|
|
fc_vport_terminate(lport->vport);
|
|
fc_frame_free(fp);
|
|
return;
|
|
}
|
|
}
|
|
fcoe_ctlr_recv_flogi(fip, lport, fp);
|
|
}
|
|
if (!is_zero_ether_addr(mac))
|
|
fip->update_mac(lport, mac);
|
|
done:
|
|
fc_lport_flogi_resp(seq, fp, lport);
|
|
}
|
|
|
|
static void bnx2fc_logo_resp(struct fc_seq *seq, struct fc_frame *fp,
|
|
void *arg)
|
|
{
|
|
struct fcoe_ctlr *fip = arg;
|
|
struct fc_exch *exch = fc_seq_exch(seq);
|
|
struct fc_lport *lport = exch->lp;
|
|
static u8 zero_mac[ETH_ALEN] = { 0 };
|
|
|
|
if (!IS_ERR(fp))
|
|
fip->update_mac(lport, zero_mac);
|
|
fc_lport_logo_resp(seq, fp, lport);
|
|
}
|
|
|
|
struct fc_seq *bnx2fc_elsct_send(struct fc_lport *lport, u32 did,
|
|
struct fc_frame *fp, unsigned int op,
|
|
void (*resp)(struct fc_seq *,
|
|
struct fc_frame *,
|
|
void *),
|
|
void *arg, u32 timeout)
|
|
{
|
|
struct fcoe_port *port = lport_priv(lport);
|
|
struct bnx2fc_interface *interface = port->priv;
|
|
struct fcoe_ctlr *fip = bnx2fc_to_ctlr(interface);
|
|
struct fc_frame_header *fh = fc_frame_header_get(fp);
|
|
|
|
switch (op) {
|
|
case ELS_FLOGI:
|
|
case ELS_FDISC:
|
|
return fc_elsct_send(lport, did, fp, op, bnx2fc_flogi_resp,
|
|
fip, timeout);
|
|
case ELS_LOGO:
|
|
/* only hook onto fabric logouts, not port logouts */
|
|
if (ntoh24(fh->fh_d_id) != FC_FID_FLOGI)
|
|
break;
|
|
return fc_elsct_send(lport, did, fp, op, bnx2fc_logo_resp,
|
|
fip, timeout);
|
|
}
|
|
return fc_elsct_send(lport, did, fp, op, resp, arg, timeout);
|
|
}
|