Add the rt linux 4.1.3-rt3 as base
[kvmfornfv.git] / kernel / drivers / scsi / bnx2fc / bnx2fc_els.c
diff --git a/kernel/drivers/scsi/bnx2fc/bnx2fc_els.c b/kernel/drivers/scsi/bnx2fc/bnx2fc_els.c
new file mode 100644 (file)
index 0000000..ef355c1
--- /dev/null
@@ -0,0 +1,918 @@
+/*
+ * bnx2fc_els.c: QLogic NetXtreme II Linux FCoE offload driver.
+ * This file contains helper routines that handle ELS requests
+ * and responses.
+ *
+ * Copyright (c) 2008 - 2013 Broadcom Corporation
+ * Copyright (c) 2014, QLogic Corporation
+ *
+ * This program is free software; you can redistribute it and/or modify
+ * it under the terms of the GNU General Public License as published by
+ * the Free Software Foundation.
+ *
+ * Written by: Bhanu Prakash Gollapudi (bprakash@broadcom.com)
+ */
+
+#include "bnx2fc.h"
+
+static void bnx2fc_logo_resp(struct fc_seq *seq, struct fc_frame *fp,
+                            void *arg);
+static void bnx2fc_flogi_resp(struct fc_seq *seq, struct fc_frame *fp,
+                             void *arg);
+static int bnx2fc_initiate_els(struct bnx2fc_rport *tgt, unsigned int op,
+                       void *data, u32 data_len,
+                       void (*cb_func)(struct bnx2fc_els_cb_arg *cb_arg),
+                       struct bnx2fc_els_cb_arg *cb_arg, u32 timer_msec);
+
+static void bnx2fc_rrq_compl(struct bnx2fc_els_cb_arg *cb_arg)
+{
+       struct bnx2fc_cmd *orig_io_req;
+       struct bnx2fc_cmd *rrq_req;
+       int rc = 0;
+
+       BUG_ON(!cb_arg);
+       rrq_req = cb_arg->io_req;
+       orig_io_req = cb_arg->aborted_io_req;
+       BUG_ON(!orig_io_req);
+       BNX2FC_ELS_DBG("rrq_compl: orig xid = 0x%x, rrq_xid = 0x%x\n",
+                  orig_io_req->xid, rrq_req->xid);
+
+       kref_put(&orig_io_req->refcount, bnx2fc_cmd_release);
+
+       if (test_and_clear_bit(BNX2FC_FLAG_ELS_TIMEOUT, &rrq_req->req_flags)) {
+               /*
+                * els req is timed out. cleanup the IO with FW and
+                * drop the completion. Remove from active_cmd_queue.
+                */
+               BNX2FC_ELS_DBG("rrq xid - 0x%x timed out, clean it up\n",
+                          rrq_req->xid);
+
+               if (rrq_req->on_active_queue) {
+                       list_del_init(&rrq_req->link);
+                       rrq_req->on_active_queue = 0;
+                       rc = bnx2fc_initiate_cleanup(rrq_req);
+                       BUG_ON(rc);
+               }
+       }
+       kfree(cb_arg);
+}
+int bnx2fc_send_rrq(struct bnx2fc_cmd *aborted_io_req)
+{
+
+       struct fc_els_rrq rrq;
+       struct bnx2fc_rport *tgt = aborted_io_req->tgt;
+       struct fc_lport *lport = tgt->rdata->local_port;
+       struct bnx2fc_els_cb_arg *cb_arg = NULL;
+       u32 sid = tgt->sid;
+       u32 r_a_tov = lport->r_a_tov;
+       unsigned long start = jiffies;
+       int rc;
+
+       BNX2FC_ELS_DBG("Sending RRQ orig_xid = 0x%x\n",
+                  aborted_io_req->xid);
+       memset(&rrq, 0, sizeof(rrq));
+
+       cb_arg = kzalloc(sizeof(struct bnx2fc_els_cb_arg), GFP_NOIO);
+       if (!cb_arg) {
+               printk(KERN_ERR PFX "Unable to allocate cb_arg for RRQ\n");
+               rc = -ENOMEM;
+               goto rrq_err;
+       }
+
+       cb_arg->aborted_io_req = aborted_io_req;
+
+       rrq.rrq_cmd = ELS_RRQ;
+       hton24(rrq.rrq_s_id, sid);
+       rrq.rrq_ox_id = htons(aborted_io_req->xid);
+       rrq.rrq_rx_id = htons(aborted_io_req->task->rxwr_txrd.var_ctx.rx_id);
+
+retry_rrq:
+       rc = bnx2fc_initiate_els(tgt, ELS_RRQ, &rrq, sizeof(rrq),
+                                bnx2fc_rrq_compl, cb_arg,
+                                r_a_tov);
+       if (rc == -ENOMEM) {
+               if (time_after(jiffies, start + (10 * HZ))) {
+                       BNX2FC_ELS_DBG("rrq Failed\n");
+                       rc = FAILED;
+                       goto rrq_err;
+               }
+               msleep(20);
+               goto retry_rrq;
+       }
+rrq_err:
+       if (rc) {
+               BNX2FC_ELS_DBG("RRQ failed - release orig io req 0x%x\n",
+                       aborted_io_req->xid);
+               kfree(cb_arg);
+               spin_lock_bh(&tgt->tgt_lock);
+               kref_put(&aborted_io_req->refcount, bnx2fc_cmd_release);
+               spin_unlock_bh(&tgt->tgt_lock);
+       }
+       return rc;
+}
+
+static void bnx2fc_l2_els_compl(struct bnx2fc_els_cb_arg *cb_arg)
+{
+       struct bnx2fc_cmd *els_req;
+       struct bnx2fc_rport *tgt;
+       struct bnx2fc_mp_req *mp_req;
+       struct fc_frame_header *fc_hdr;
+       unsigned char *buf;
+       void *resp_buf;
+       u32 resp_len, hdr_len;
+       u16 l2_oxid;
+       int frame_len;
+       int rc = 0;
+
+       l2_oxid = cb_arg->l2_oxid;
+       BNX2FC_ELS_DBG("ELS COMPL - l2_oxid = 0x%x\n", l2_oxid);
+
+       els_req = cb_arg->io_req;
+       if (test_and_clear_bit(BNX2FC_FLAG_ELS_TIMEOUT, &els_req->req_flags)) {
+               /*
+                * els req is timed out. cleanup the IO with FW and
+                * drop the completion. libfc will handle the els timeout
+                */
+               if (els_req->on_active_queue) {
+                       list_del_init(&els_req->link);
+                       els_req->on_active_queue = 0;
+                       rc = bnx2fc_initiate_cleanup(els_req);
+                       BUG_ON(rc);
+               }
+               goto free_arg;
+       }
+
+       tgt = els_req->tgt;
+       mp_req = &(els_req->mp_req);
+       fc_hdr = &(mp_req->resp_fc_hdr);
+       resp_len = mp_req->resp_len;
+       resp_buf = mp_req->resp_buf;
+
+       buf = kzalloc(PAGE_SIZE, GFP_ATOMIC);
+       if (!buf) {
+               printk(KERN_ERR PFX "Unable to alloc mp buf\n");
+               goto free_arg;
+       }
+       hdr_len = sizeof(*fc_hdr);
+       if (hdr_len + resp_len > PAGE_SIZE) {
+               printk(KERN_ERR PFX "l2_els_compl: resp len is "
+                                   "beyond page size\n");
+               goto free_buf;
+       }
+       memcpy(buf, fc_hdr, hdr_len);
+       memcpy(buf + hdr_len, resp_buf, resp_len);
+       frame_len = hdr_len + resp_len;
+
+       bnx2fc_process_l2_frame_compl(tgt, buf, frame_len, l2_oxid);
+
+free_buf:
+       kfree(buf);
+free_arg:
+       kfree(cb_arg);
+}
+
+int bnx2fc_send_adisc(struct bnx2fc_rport *tgt, struct fc_frame *fp)
+{
+       struct fc_els_adisc *adisc;
+       struct fc_frame_header *fh;
+       struct bnx2fc_els_cb_arg *cb_arg;
+       struct fc_lport *lport = tgt->rdata->local_port;
+       u32 r_a_tov = lport->r_a_tov;
+       int rc;
+
+       fh = fc_frame_header_get(fp);
+       cb_arg = kzalloc(sizeof(struct bnx2fc_els_cb_arg), GFP_ATOMIC);
+       if (!cb_arg) {
+               printk(KERN_ERR PFX "Unable to allocate cb_arg for ADISC\n");
+               return -ENOMEM;
+       }
+
+       cb_arg->l2_oxid = ntohs(fh->fh_ox_id);
+
+       BNX2FC_ELS_DBG("send ADISC: l2_oxid = 0x%x\n", cb_arg->l2_oxid);
+       adisc = fc_frame_payload_get(fp, sizeof(*adisc));
+       /* adisc is initialized by libfc */
+       rc = bnx2fc_initiate_els(tgt, ELS_ADISC, adisc, sizeof(*adisc),
+                                bnx2fc_l2_els_compl, cb_arg, 2 * r_a_tov);
+       if (rc)
+               kfree(cb_arg);
+       return rc;
+}
+
+int bnx2fc_send_logo(struct bnx2fc_rport *tgt, struct fc_frame *fp)
+{
+       struct fc_els_logo *logo;
+       struct fc_frame_header *fh;
+       struct bnx2fc_els_cb_arg *cb_arg;
+       struct fc_lport *lport = tgt->rdata->local_port;
+       u32 r_a_tov = lport->r_a_tov;
+       int rc;
+
+       fh = fc_frame_header_get(fp);
+       cb_arg = kzalloc(sizeof(struct bnx2fc_els_cb_arg), GFP_ATOMIC);
+       if (!cb_arg) {
+               printk(KERN_ERR PFX "Unable to allocate cb_arg for LOGO\n");
+               return -ENOMEM;
+       }
+
+       cb_arg->l2_oxid = ntohs(fh->fh_ox_id);
+
+       BNX2FC_ELS_DBG("Send LOGO: l2_oxid = 0x%x\n", cb_arg->l2_oxid);
+       logo = fc_frame_payload_get(fp, sizeof(*logo));
+       /* logo is initialized by libfc */
+       rc = bnx2fc_initiate_els(tgt, ELS_LOGO, logo, sizeof(*logo),
+                                bnx2fc_l2_els_compl, cb_arg, 2 * r_a_tov);
+       if (rc)
+               kfree(cb_arg);
+       return rc;
+}
+
+int bnx2fc_send_rls(struct bnx2fc_rport *tgt, struct fc_frame *fp)
+{
+       struct fc_els_rls *rls;
+       struct fc_frame_header *fh;
+       struct bnx2fc_els_cb_arg *cb_arg;
+       struct fc_lport *lport = tgt->rdata->local_port;
+       u32 r_a_tov = lport->r_a_tov;
+       int rc;
+
+       fh = fc_frame_header_get(fp);
+       cb_arg = kzalloc(sizeof(struct bnx2fc_els_cb_arg), GFP_ATOMIC);
+       if (!cb_arg) {
+               printk(KERN_ERR PFX "Unable to allocate cb_arg for LOGO\n");
+               return -ENOMEM;
+       }
+
+       cb_arg->l2_oxid = ntohs(fh->fh_ox_id);
+
+       rls = fc_frame_payload_get(fp, sizeof(*rls));
+       /* rls is initialized by libfc */
+       rc = bnx2fc_initiate_els(tgt, ELS_RLS, rls, sizeof(*rls),
+                                 bnx2fc_l2_els_compl, cb_arg, 2 * r_a_tov);
+       if (rc)
+               kfree(cb_arg);
+       return rc;
+}
+
+void bnx2fc_srr_compl(struct bnx2fc_els_cb_arg *cb_arg)
+{
+       struct bnx2fc_mp_req *mp_req;
+       struct fc_frame_header *fc_hdr, *fh;
+       struct bnx2fc_cmd *srr_req;
+       struct bnx2fc_cmd *orig_io_req;
+       struct fc_frame *fp;
+       unsigned char *buf;
+       void *resp_buf;
+       u32 resp_len, hdr_len;
+       u8 opcode;
+       int rc = 0;
+
+       orig_io_req = cb_arg->aborted_io_req;
+       srr_req = cb_arg->io_req;
+       if (test_and_clear_bit(BNX2FC_FLAG_ELS_TIMEOUT, &srr_req->req_flags)) {
+               /* SRR timedout */
+               BNX2FC_IO_DBG(srr_req, "srr timed out, abort "
+                      "orig_io - 0x%x\n",
+                       orig_io_req->xid);
+               rc = bnx2fc_initiate_abts(srr_req);
+               if (rc != SUCCESS) {
+                       BNX2FC_IO_DBG(srr_req, "srr_compl: initiate_abts "
+                               "failed. issue cleanup\n");
+                       bnx2fc_initiate_cleanup(srr_req);
+               }
+               if (test_bit(BNX2FC_FLAG_IO_COMPL, &orig_io_req->req_flags) ||
+                   test_bit(BNX2FC_FLAG_ISSUE_ABTS, &orig_io_req->req_flags)) {
+                       BNX2FC_IO_DBG(srr_req, "srr_compl:xid 0x%x flags = %lx",
+                                     orig_io_req->xid, orig_io_req->req_flags);
+                       goto srr_compl_done;
+               }
+               orig_io_req->srr_retry++;
+               if (orig_io_req->srr_retry <= SRR_RETRY_COUNT) {
+                       struct bnx2fc_rport *tgt = orig_io_req->tgt;
+                       spin_unlock_bh(&tgt->tgt_lock);
+                       rc = bnx2fc_send_srr(orig_io_req,
+                                            orig_io_req->srr_offset,
+                                            orig_io_req->srr_rctl);
+                       spin_lock_bh(&tgt->tgt_lock);
+                       if (!rc)
+                               goto srr_compl_done;
+               }
+
+               rc = bnx2fc_initiate_abts(orig_io_req);
+               if (rc != SUCCESS) {
+                       BNX2FC_IO_DBG(srr_req, "srr_compl: initiate_abts "
+                               "failed xid = 0x%x. issue cleanup\n",
+                               orig_io_req->xid);
+                       bnx2fc_initiate_cleanup(orig_io_req);
+               }
+               goto srr_compl_done;
+       }
+       if (test_bit(BNX2FC_FLAG_IO_COMPL, &orig_io_req->req_flags) ||
+           test_bit(BNX2FC_FLAG_ISSUE_ABTS, &orig_io_req->req_flags)) {
+               BNX2FC_IO_DBG(srr_req, "srr_compl:xid - 0x%x flags = %lx",
+                             orig_io_req->xid, orig_io_req->req_flags);
+               goto srr_compl_done;
+       }
+       mp_req = &(srr_req->mp_req);
+       fc_hdr = &(mp_req->resp_fc_hdr);
+       resp_len = mp_req->resp_len;
+       resp_buf = mp_req->resp_buf;
+
+       hdr_len = sizeof(*fc_hdr);
+       buf = kzalloc(PAGE_SIZE, GFP_ATOMIC);
+       if (!buf) {
+               printk(KERN_ERR PFX "srr buf: mem alloc failure\n");
+               goto srr_compl_done;
+       }
+       memcpy(buf, fc_hdr, hdr_len);
+       memcpy(buf + hdr_len, resp_buf, resp_len);
+
+       fp = fc_frame_alloc(NULL, resp_len);
+       if (!fp) {
+               printk(KERN_ERR PFX "fc_frame_alloc failure\n");
+               goto free_buf;
+       }
+
+       fh = (struct fc_frame_header *) fc_frame_header_get(fp);
+       /* Copy FC Frame header and payload into the frame */
+       memcpy(fh, buf, hdr_len + resp_len);
+
+       opcode = fc_frame_payload_op(fp);
+       switch (opcode) {
+       case ELS_LS_ACC:
+               BNX2FC_IO_DBG(srr_req, "SRR success\n");
+               break;
+       case ELS_LS_RJT:
+               BNX2FC_IO_DBG(srr_req, "SRR rejected\n");
+               rc = bnx2fc_initiate_abts(orig_io_req);
+               if (rc != SUCCESS) {
+                       BNX2FC_IO_DBG(srr_req, "srr_compl: initiate_abts "
+                               "failed xid = 0x%x. issue cleanup\n",
+                               orig_io_req->xid);
+                       bnx2fc_initiate_cleanup(orig_io_req);
+               }
+               break;
+       default:
+               BNX2FC_IO_DBG(srr_req, "srr compl - invalid opcode = %d\n",
+                       opcode);
+               break;
+       }
+       fc_frame_free(fp);
+free_buf:
+       kfree(buf);
+srr_compl_done:
+       kref_put(&orig_io_req->refcount, bnx2fc_cmd_release);
+}
+
+void bnx2fc_rec_compl(struct bnx2fc_els_cb_arg *cb_arg)
+{
+       struct bnx2fc_cmd *orig_io_req, *new_io_req;
+       struct bnx2fc_cmd *rec_req;
+       struct bnx2fc_mp_req *mp_req;
+       struct fc_frame_header *fc_hdr, *fh;
+       struct fc_els_ls_rjt *rjt;
+       struct fc_els_rec_acc *acc;
+       struct bnx2fc_rport *tgt;
+       struct fcoe_err_report_entry *err_entry;
+       struct scsi_cmnd *sc_cmd;
+       enum fc_rctl r_ctl;
+       unsigned char *buf;
+       void *resp_buf;
+       struct fc_frame *fp;
+       u8 opcode;
+       u32 offset;
+       u32 e_stat;
+       u32 resp_len, hdr_len;
+       int rc = 0;
+       bool send_seq_clnp = false;
+       bool abort_io = false;
+
+       BNX2FC_MISC_DBG("Entered rec_compl callback\n");
+       rec_req = cb_arg->io_req;
+       orig_io_req = cb_arg->aborted_io_req;
+       BNX2FC_IO_DBG(rec_req, "rec_compl: orig xid = 0x%x", orig_io_req->xid);
+       tgt = orig_io_req->tgt;
+
+       /* Handle REC timeout case */
+       if (test_and_clear_bit(BNX2FC_FLAG_ELS_TIMEOUT, &rec_req->req_flags)) {
+               BNX2FC_IO_DBG(rec_req, "timed out, abort "
+                      "orig_io - 0x%x\n",
+                       orig_io_req->xid);
+               /* els req is timed out. send abts for els */
+               rc = bnx2fc_initiate_abts(rec_req);
+               if (rc != SUCCESS) {
+                       BNX2FC_IO_DBG(rec_req, "rec_compl: initiate_abts "
+                               "failed. issue cleanup\n");
+                       bnx2fc_initiate_cleanup(rec_req);
+               }
+               orig_io_req->rec_retry++;
+               /* REC timedout. send ABTS to the orig IO req */
+               if (orig_io_req->rec_retry <= REC_RETRY_COUNT) {
+                       spin_unlock_bh(&tgt->tgt_lock);
+                       rc = bnx2fc_send_rec(orig_io_req);
+                       spin_lock_bh(&tgt->tgt_lock);
+                       if (!rc)
+                               goto rec_compl_done;
+               }
+               rc = bnx2fc_initiate_abts(orig_io_req);
+               if (rc != SUCCESS) {
+                       BNX2FC_IO_DBG(rec_req, "rec_compl: initiate_abts "
+                               "failed xid = 0x%x. issue cleanup\n",
+                               orig_io_req->xid);
+                       bnx2fc_initiate_cleanup(orig_io_req);
+               }
+               goto rec_compl_done;
+       }
+
+       if (test_bit(BNX2FC_FLAG_IO_COMPL, &orig_io_req->req_flags)) {
+               BNX2FC_IO_DBG(rec_req, "completed"
+                      "orig_io - 0x%x\n",
+                       orig_io_req->xid);
+               goto rec_compl_done;
+       }
+       if (test_bit(BNX2FC_FLAG_ISSUE_ABTS, &orig_io_req->req_flags)) {
+               BNX2FC_IO_DBG(rec_req, "abts in prog "
+                      "orig_io - 0x%x\n",
+                       orig_io_req->xid);
+               goto rec_compl_done;
+       }
+
+       mp_req = &(rec_req->mp_req);
+       fc_hdr = &(mp_req->resp_fc_hdr);
+       resp_len = mp_req->resp_len;
+       acc = resp_buf = mp_req->resp_buf;
+
+       hdr_len = sizeof(*fc_hdr);
+
+       buf = kzalloc(PAGE_SIZE, GFP_ATOMIC);
+       if (!buf) {
+               printk(KERN_ERR PFX "rec buf: mem alloc failure\n");
+               goto rec_compl_done;
+       }
+       memcpy(buf, fc_hdr, hdr_len);
+       memcpy(buf + hdr_len, resp_buf, resp_len);
+
+       fp = fc_frame_alloc(NULL, resp_len);
+       if (!fp) {
+               printk(KERN_ERR PFX "fc_frame_alloc failure\n");
+               goto free_buf;
+       }
+
+       fh = (struct fc_frame_header *) fc_frame_header_get(fp);
+       /* Copy FC Frame header and payload into the frame */
+       memcpy(fh, buf, hdr_len + resp_len);
+
+       opcode = fc_frame_payload_op(fp);
+       if (opcode == ELS_LS_RJT) {
+               BNX2FC_IO_DBG(rec_req, "opcode is RJT\n");
+               rjt = fc_frame_payload_get(fp, sizeof(*rjt));
+               if ((rjt->er_reason == ELS_RJT_LOGIC ||
+                   rjt->er_reason == ELS_RJT_UNAB) &&
+                   rjt->er_explan == ELS_EXPL_OXID_RXID) {
+                       BNX2FC_IO_DBG(rec_req, "handle CMD LOST case\n");
+                       new_io_req = bnx2fc_cmd_alloc(tgt);
+                       if (!new_io_req)
+                               goto abort_io;
+                       new_io_req->sc_cmd = orig_io_req->sc_cmd;
+                       /* cleanup orig_io_req that is with the FW */
+                       set_bit(BNX2FC_FLAG_CMD_LOST,
+                               &orig_io_req->req_flags);
+                       bnx2fc_initiate_cleanup(orig_io_req);
+                       /* Post a new IO req with the same sc_cmd */
+                       BNX2FC_IO_DBG(rec_req, "Post IO request again\n");
+                       rc = bnx2fc_post_io_req(tgt, new_io_req);
+                       if (!rc)
+                               goto free_frame;
+                       BNX2FC_IO_DBG(rec_req, "REC: io post err\n");
+               }
+abort_io:
+               rc = bnx2fc_initiate_abts(orig_io_req);
+               if (rc != SUCCESS) {
+                       BNX2FC_IO_DBG(rec_req, "rec_compl: initiate_abts "
+                               "failed. issue cleanup\n");
+                       bnx2fc_initiate_cleanup(orig_io_req);
+               }
+       } else if (opcode == ELS_LS_ACC) {
+               /* REVISIT: Check if the exchange is already aborted */
+               offset = ntohl(acc->reca_fc4value);
+               e_stat = ntohl(acc->reca_e_stat);
+               if (e_stat & ESB_ST_SEQ_INIT)  {
+                       BNX2FC_IO_DBG(rec_req, "target has the seq init\n");
+                       goto free_frame;
+               }
+               BNX2FC_IO_DBG(rec_req, "e_stat = 0x%x, offset = 0x%x\n",
+                       e_stat, offset);
+               /* Seq initiative is with us */
+               err_entry = (struct fcoe_err_report_entry *)
+                            &orig_io_req->err_entry;
+               sc_cmd = orig_io_req->sc_cmd;
+               if (sc_cmd->sc_data_direction == DMA_TO_DEVICE) {
+                       /* SCSI WRITE command */
+                       if (offset == orig_io_req->data_xfer_len) {
+                               BNX2FC_IO_DBG(rec_req, "WRITE - resp lost\n");
+                               /* FCP_RSP lost */
+                               r_ctl = FC_RCTL_DD_CMD_STATUS;
+                               offset = 0;
+                       } else  {
+                               /* start transmitting from offset */
+                               BNX2FC_IO_DBG(rec_req, "XFER_RDY/DATA lost\n");
+                               send_seq_clnp = true;
+                               r_ctl = FC_RCTL_DD_DATA_DESC;
+                               if (bnx2fc_initiate_seq_cleanup(orig_io_req,
+                                                               offset, r_ctl))
+                                       abort_io = true;
+                               /* XFER_RDY */
+                       }
+               } else {
+                       /* SCSI READ command */
+                       if (err_entry->data.rx_buf_off ==
+                                       orig_io_req->data_xfer_len) {
+                               /* FCP_RSP lost */
+                               BNX2FC_IO_DBG(rec_req, "READ - resp lost\n");
+                               r_ctl = FC_RCTL_DD_CMD_STATUS;
+                               offset = 0;
+                       } else  {
+                               /* request retransmission from this offset */
+                               send_seq_clnp = true;
+                               offset = err_entry->data.rx_buf_off;
+                               BNX2FC_IO_DBG(rec_req, "RD DATA lost\n");
+                               /* FCP_DATA lost */
+                               r_ctl = FC_RCTL_DD_SOL_DATA;
+                               if (bnx2fc_initiate_seq_cleanup(orig_io_req,
+                                                               offset, r_ctl))
+                                       abort_io = true;
+                       }
+               }
+               if (abort_io) {
+                       rc = bnx2fc_initiate_abts(orig_io_req);
+                       if (rc != SUCCESS) {
+                               BNX2FC_IO_DBG(rec_req, "rec_compl:initiate_abts"
+                                             " failed. issue cleanup\n");
+                               bnx2fc_initiate_cleanup(orig_io_req);
+                       }
+               } else if (!send_seq_clnp) {
+                       BNX2FC_IO_DBG(rec_req, "Send SRR - FCP_RSP\n");
+                       spin_unlock_bh(&tgt->tgt_lock);
+                       rc = bnx2fc_send_srr(orig_io_req, offset, r_ctl);
+                       spin_lock_bh(&tgt->tgt_lock);
+
+                       if (rc) {
+                               BNX2FC_IO_DBG(rec_req, "Unable to send SRR"
+                                       " IO will abort\n");
+                       }
+               }
+       }
+free_frame:
+       fc_frame_free(fp);
+free_buf:
+       kfree(buf);
+rec_compl_done:
+       kref_put(&orig_io_req->refcount, bnx2fc_cmd_release);
+       kfree(cb_arg);
+}
+
+int bnx2fc_send_rec(struct bnx2fc_cmd *orig_io_req)
+{
+       struct fc_els_rec rec;
+       struct bnx2fc_rport *tgt = orig_io_req->tgt;
+       struct fc_lport *lport = tgt->rdata->local_port;
+       struct bnx2fc_els_cb_arg *cb_arg = NULL;
+       u32 sid = tgt->sid;
+       u32 r_a_tov = lport->r_a_tov;
+       int rc;
+
+       BNX2FC_IO_DBG(orig_io_req, "Sending REC\n");
+       memset(&rec, 0, sizeof(rec));
+
+       cb_arg = kzalloc(sizeof(struct bnx2fc_els_cb_arg), GFP_ATOMIC);
+       if (!cb_arg) {
+               printk(KERN_ERR PFX "Unable to allocate cb_arg for REC\n");
+               rc = -ENOMEM;
+               goto rec_err;
+       }
+       kref_get(&orig_io_req->refcount);
+
+       cb_arg->aborted_io_req = orig_io_req;
+
+       rec.rec_cmd = ELS_REC;
+       hton24(rec.rec_s_id, sid);
+       rec.rec_ox_id = htons(orig_io_req->xid);
+       rec.rec_rx_id = htons(orig_io_req->task->rxwr_txrd.var_ctx.rx_id);
+
+       rc = bnx2fc_initiate_els(tgt, ELS_REC, &rec, sizeof(rec),
+                                bnx2fc_rec_compl, cb_arg,
+                                r_a_tov);
+rec_err:
+       if (rc) {
+               BNX2FC_IO_DBG(orig_io_req, "REC failed - release\n");
+               spin_lock_bh(&tgt->tgt_lock);
+               kref_put(&orig_io_req->refcount, bnx2fc_cmd_release);
+               spin_unlock_bh(&tgt->tgt_lock);
+               kfree(cb_arg);
+       }
+       return rc;
+}
+
+int bnx2fc_send_srr(struct bnx2fc_cmd *orig_io_req, u32 offset, u8 r_ctl)
+{
+       struct fcp_srr srr;
+       struct bnx2fc_rport *tgt = orig_io_req->tgt;
+       struct fc_lport *lport = tgt->rdata->local_port;
+       struct bnx2fc_els_cb_arg *cb_arg = NULL;
+       u32 r_a_tov = lport->r_a_tov;
+       int rc;
+
+       BNX2FC_IO_DBG(orig_io_req, "Sending SRR\n");
+       memset(&srr, 0, sizeof(srr));
+
+       cb_arg = kzalloc(sizeof(struct bnx2fc_els_cb_arg), GFP_ATOMIC);
+       if (!cb_arg) {
+               printk(KERN_ERR PFX "Unable to allocate cb_arg for SRR\n");
+               rc = -ENOMEM;
+               goto srr_err;
+       }
+       kref_get(&orig_io_req->refcount);
+
+       cb_arg->aborted_io_req = orig_io_req;
+
+       srr.srr_op = ELS_SRR;
+       srr.srr_ox_id = htons(orig_io_req->xid);
+       srr.srr_rx_id = htons(orig_io_req->task->rxwr_txrd.var_ctx.rx_id);
+       srr.srr_rel_off = htonl(offset);
+       srr.srr_r_ctl = r_ctl;
+       orig_io_req->srr_offset = offset;
+       orig_io_req->srr_rctl = r_ctl;
+
+       rc = bnx2fc_initiate_els(tgt, ELS_SRR, &srr, sizeof(srr),
+                                bnx2fc_srr_compl, cb_arg,
+                                r_a_tov);
+srr_err:
+       if (rc) {
+               BNX2FC_IO_DBG(orig_io_req, "SRR failed - release\n");
+               spin_lock_bh(&tgt->tgt_lock);
+               kref_put(&orig_io_req->refcount, bnx2fc_cmd_release);
+               spin_unlock_bh(&tgt->tgt_lock);
+               kfree(cb_arg);
+       } else
+               set_bit(BNX2FC_FLAG_SRR_SENT, &orig_io_req->req_flags);
+
+       return rc;
+}
+
+static int bnx2fc_initiate_els(struct bnx2fc_rport *tgt, unsigned int op,
+                       void *data, u32 data_len,
+                       void (*cb_func)(struct bnx2fc_els_cb_arg *cb_arg),
+                       struct bnx2fc_els_cb_arg *cb_arg, u32 timer_msec)
+{
+       struct fcoe_port *port = tgt->port;
+       struct bnx2fc_interface *interface = port->priv;
+       struct fc_rport *rport = tgt->rport;
+       struct fc_lport *lport = port->lport;
+       struct bnx2fc_cmd *els_req;
+       struct bnx2fc_mp_req *mp_req;
+       struct fc_frame_header *fc_hdr;
+       struct fcoe_task_ctx_entry *task;
+       struct fcoe_task_ctx_entry *task_page;
+       int rc = 0;
+       int task_idx, index;
+       u32 did, sid;
+       u16 xid;
+
+       rc = fc_remote_port_chkready(rport);
+       if (rc) {
+               printk(KERN_ERR PFX "els 0x%x: rport not ready\n", op);
+               rc = -EINVAL;
+               goto els_err;
+       }
+       if (lport->state != LPORT_ST_READY || !(lport->link_up)) {
+               printk(KERN_ERR PFX "els 0x%x: link is not ready\n", op);
+               rc = -EINVAL;
+               goto els_err;
+       }
+       if (!(test_bit(BNX2FC_FLAG_SESSION_READY, &tgt->flags)) ||
+            (test_bit(BNX2FC_FLAG_EXPL_LOGO, &tgt->flags))) {
+               printk(KERN_ERR PFX "els 0x%x: tgt not ready\n", op);
+               rc = -EINVAL;
+               goto els_err;
+       }
+       els_req = bnx2fc_elstm_alloc(tgt, BNX2FC_ELS);
+       if (!els_req) {
+               rc = -ENOMEM;
+               goto els_err;
+       }
+
+       els_req->sc_cmd = NULL;
+       els_req->port = port;
+       els_req->tgt = tgt;
+       els_req->cb_func = cb_func;
+       cb_arg->io_req = els_req;
+       els_req->cb_arg = cb_arg;
+
+       mp_req = (struct bnx2fc_mp_req *)&(els_req->mp_req);
+       rc = bnx2fc_init_mp_req(els_req);
+       if (rc == FAILED) {
+               printk(KERN_ERR PFX "ELS MP request init failed\n");
+               spin_lock_bh(&tgt->tgt_lock);
+               kref_put(&els_req->refcount, bnx2fc_cmd_release);
+               spin_unlock_bh(&tgt->tgt_lock);
+               rc = -ENOMEM;
+               goto els_err;
+       } else {
+               /* rc SUCCESS */
+               rc = 0;
+       }
+
+       /* Set the data_xfer_len to the size of ELS payload */
+       mp_req->req_len = data_len;
+       els_req->data_xfer_len = mp_req->req_len;
+
+       /* Fill ELS Payload */
+       if ((op >= ELS_LS_RJT) && (op <= ELS_AUTH_ELS)) {
+               memcpy(mp_req->req_buf, data, data_len);
+       } else {
+               printk(KERN_ERR PFX "Invalid ELS op 0x%x\n", op);
+               els_req->cb_func = NULL;
+               els_req->cb_arg = NULL;
+               spin_lock_bh(&tgt->tgt_lock);
+               kref_put(&els_req->refcount, bnx2fc_cmd_release);
+               spin_unlock_bh(&tgt->tgt_lock);
+               rc = -EINVAL;
+       }
+
+       if (rc)
+               goto els_err;
+
+       /* Fill FC header */
+       fc_hdr = &(mp_req->req_fc_hdr);
+
+       did = tgt->rport->port_id;
+       sid = tgt->sid;
+
+       if (op == ELS_SRR)
+               __fc_fill_fc_hdr(fc_hdr, FC_RCTL_ELS4_REQ, did, sid,
+                                  FC_TYPE_FCP, FC_FC_FIRST_SEQ |
+                                  FC_FC_END_SEQ | FC_FC_SEQ_INIT, 0);
+       else
+               __fc_fill_fc_hdr(fc_hdr, FC_RCTL_ELS_REQ, did, sid,
+                                  FC_TYPE_ELS, FC_FC_FIRST_SEQ |
+                                  FC_FC_END_SEQ | FC_FC_SEQ_INIT, 0);
+
+       /* Obtain exchange id */
+       xid = els_req->xid;
+       task_idx = xid/BNX2FC_TASKS_PER_PAGE;
+       index = xid % BNX2FC_TASKS_PER_PAGE;
+
+       /* Initialize task context for this IO request */
+       task_page = (struct fcoe_task_ctx_entry *)
+                       interface->hba->task_ctx[task_idx];
+       task = &(task_page[index]);
+       bnx2fc_init_mp_task(els_req, task);
+
+       spin_lock_bh(&tgt->tgt_lock);
+
+       if (!test_bit(BNX2FC_FLAG_SESSION_READY, &tgt->flags)) {
+               printk(KERN_ERR PFX "initiate_els.. session not ready\n");
+               els_req->cb_func = NULL;
+               els_req->cb_arg = NULL;
+               kref_put(&els_req->refcount, bnx2fc_cmd_release);
+               spin_unlock_bh(&tgt->tgt_lock);
+               return -EINVAL;
+       }
+
+       if (timer_msec)
+               bnx2fc_cmd_timer_set(els_req, timer_msec);
+       bnx2fc_add_2_sq(tgt, xid);
+
+       els_req->on_active_queue = 1;
+       list_add_tail(&els_req->link, &tgt->els_queue);
+
+       /* Ring doorbell */
+       bnx2fc_ring_doorbell(tgt);
+       spin_unlock_bh(&tgt->tgt_lock);
+
+els_err:
+       return rc;
+}
+
+void bnx2fc_process_els_compl(struct bnx2fc_cmd *els_req,
+                             struct fcoe_task_ctx_entry *task, u8 num_rq)
+{
+       struct bnx2fc_mp_req *mp_req;
+       struct fc_frame_header *fc_hdr;
+       u64 *hdr;
+       u64 *temp_hdr;
+
+       BNX2FC_ELS_DBG("Entered process_els_compl xid = 0x%x"
+                       "cmd_type = %d\n", els_req->xid, els_req->cmd_type);
+
+       if (test_and_set_bit(BNX2FC_FLAG_ELS_DONE,
+                            &els_req->req_flags)) {
+               BNX2FC_ELS_DBG("Timer context finished processing this "
+                          "els - 0x%x\n", els_req->xid);
+               /* This IO doesn't receive cleanup completion */
+               kref_put(&els_req->refcount, bnx2fc_cmd_release);
+               return;
+       }
+
+       /* Cancel the timeout_work, as we received the response */
+       if (cancel_delayed_work(&els_req->timeout_work))
+               kref_put(&els_req->refcount,
+                        bnx2fc_cmd_release); /* drop timer hold */
+
+       if (els_req->on_active_queue) {
+               list_del_init(&els_req->link);
+               els_req->on_active_queue = 0;
+       }
+
+       mp_req = &(els_req->mp_req);
+       fc_hdr = &(mp_req->resp_fc_hdr);
+
+       hdr = (u64 *)fc_hdr;
+       temp_hdr = (u64 *)
+               &task->rxwr_only.union_ctx.comp_info.mp_rsp.fc_hdr;
+       hdr[0] = cpu_to_be64(temp_hdr[0]);
+       hdr[1] = cpu_to_be64(temp_hdr[1]);
+       hdr[2] = cpu_to_be64(temp_hdr[2]);
+
+       mp_req->resp_len =
+               task->rxwr_only.union_ctx.comp_info.mp_rsp.mp_payload_len;
+
+       /* Parse ELS response */
+       if ((els_req->cb_func) && (els_req->cb_arg)) {
+               els_req->cb_func(els_req->cb_arg);
+               els_req->cb_arg = NULL;
+       }
+
+       kref_put(&els_req->refcount, bnx2fc_cmd_release);
+}
+
+static void bnx2fc_flogi_resp(struct fc_seq *seq, struct fc_frame *fp,
+                             void *arg)
+{
+       struct fcoe_ctlr *fip = arg;
+       struct fc_exch *exch = fc_seq_exch(seq);
+       struct fc_lport *lport = exch->lp;
+       u8 *mac;
+       u8 op;
+
+       if (IS_ERR(fp))
+               goto done;
+
+       mac = fr_cb(fp)->granted_mac;
+       if (is_zero_ether_addr(mac)) {
+               op = fc_frame_payload_op(fp);
+               if (lport->vport) {
+                       if (op == ELS_LS_RJT) {
+                               printk(KERN_ERR PFX "bnx2fc_flogi_resp is LS_RJT\n");
+                               fc_vport_terminate(lport->vport);
+                               fc_frame_free(fp);
+                               return;
+                       }
+               }
+               fcoe_ctlr_recv_flogi(fip, lport, fp);
+       }
+       if (!is_zero_ether_addr(mac))
+               fip->update_mac(lport, mac);
+done:
+       fc_lport_flogi_resp(seq, fp, lport);
+}
+
+static void bnx2fc_logo_resp(struct fc_seq *seq, struct fc_frame *fp,
+                            void *arg)
+{
+       struct fcoe_ctlr *fip = arg;
+       struct fc_exch *exch = fc_seq_exch(seq);
+       struct fc_lport *lport = exch->lp;
+       static u8 zero_mac[ETH_ALEN] = { 0 };
+
+       if (!IS_ERR(fp))
+               fip->update_mac(lport, zero_mac);
+       fc_lport_logo_resp(seq, fp, lport);
+}
+
+struct fc_seq *bnx2fc_elsct_send(struct fc_lport *lport, u32 did,
+                                     struct fc_frame *fp, unsigned int op,
+                                     void (*resp)(struct fc_seq *,
+                                                  struct fc_frame *,
+                                                  void *),
+                                     void *arg, u32 timeout)
+{
+       struct fcoe_port *port = lport_priv(lport);
+       struct bnx2fc_interface *interface = port->priv;
+       struct fcoe_ctlr *fip = bnx2fc_to_ctlr(interface);
+       struct fc_frame_header *fh = fc_frame_header_get(fp);
+
+       switch (op) {
+       case ELS_FLOGI:
+       case ELS_FDISC:
+               return fc_elsct_send(lport, did, fp, op, bnx2fc_flogi_resp,
+                                    fip, timeout);
+       case ELS_LOGO:
+               /* only hook onto fabric logouts, not port logouts */
+               if (ntoh24(fh->fh_d_id) != FC_FID_FLOGI)
+                       break;
+               return fc_elsct_send(lport, did, fp, op, bnx2fc_logo_resp,
+                                    fip, timeout);
+       }
+       return fc_elsct_send(lport, did, fp, op, resp, arg, timeout);
+}