1603 lines
40 KiB
C
1603 lines
40 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Copyright (C) 2021 Broadcom. All Rights Reserved. The term
|
|
* “Broadcom” refers to Broadcom Inc. and/or its subsidiaries.
|
|
*/
|
|
|
|
/*
|
|
* device_sm Node State Machine: Remote Device States
|
|
*/
|
|
|
|
#include "efc.h"
|
|
#include "efc_device.h"
|
|
#include "efc_fabric.h"
|
|
|
|
void
|
|
efc_d_send_prli_rsp(struct efc_node *node, u16 ox_id)
|
|
{
|
|
int rc = EFC_SCSI_CALL_COMPLETE;
|
|
struct efc *efc = node->efc;
|
|
|
|
node->ls_acc_oxid = ox_id;
|
|
node->send_ls_acc = EFC_NODE_SEND_LS_ACC_PRLI;
|
|
|
|
/*
|
|
* Wait for backend session registration
|
|
* to complete before sending PRLI resp
|
|
*/
|
|
|
|
if (node->init) {
|
|
efc_log_info(efc, "[%s] found(initiator) WWPN:%s WWNN:%s\n",
|
|
node->display_name, node->wwpn, node->wwnn);
|
|
if (node->nport->enable_tgt)
|
|
rc = efc->tt.scsi_new_node(efc, node);
|
|
}
|
|
|
|
if (rc < 0)
|
|
efc_node_post_event(node, EFC_EVT_NODE_SESS_REG_FAIL, NULL);
|
|
|
|
if (rc == EFC_SCSI_CALL_COMPLETE)
|
|
efc_node_post_event(node, EFC_EVT_NODE_SESS_REG_OK, NULL);
|
|
}
|
|
|
|
static void
|
|
__efc_d_common(const char *funcname, struct efc_sm_ctx *ctx,
|
|
enum efc_sm_event evt, void *arg)
|
|
{
|
|
struct efc_node *node = NULL;
|
|
struct efc *efc = NULL;
|
|
|
|
node = ctx->app;
|
|
efc = node->efc;
|
|
|
|
switch (evt) {
|
|
/* Handle shutdown events */
|
|
case EFC_EVT_SHUTDOWN:
|
|
efc_log_debug(efc, "[%s] %-20s %-20s\n", node->display_name,
|
|
funcname, efc_sm_event_name(evt));
|
|
node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT;
|
|
efc_node_transition(node, __efc_d_initiate_shutdown, NULL);
|
|
break;
|
|
case EFC_EVT_SHUTDOWN_EXPLICIT_LOGO:
|
|
efc_log_debug(efc, "[%s] %-20s %-20s\n",
|
|
node->display_name, funcname,
|
|
efc_sm_event_name(evt));
|
|
node->shutdown_reason = EFC_NODE_SHUTDOWN_EXPLICIT_LOGO;
|
|
efc_node_transition(node, __efc_d_initiate_shutdown, NULL);
|
|
break;
|
|
case EFC_EVT_SHUTDOWN_IMPLICIT_LOGO:
|
|
efc_log_debug(efc, "[%s] %-20s %-20s\n", node->display_name,
|
|
funcname, efc_sm_event_name(evt));
|
|
node->shutdown_reason = EFC_NODE_SHUTDOWN_IMPLICIT_LOGO;
|
|
efc_node_transition(node, __efc_d_initiate_shutdown, NULL);
|
|
break;
|
|
|
|
default:
|
|
/* call default event handler common to all nodes */
|
|
__efc_node_common(funcname, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
static void
|
|
__efc_d_wait_del_node(struct efc_sm_ctx *ctx,
|
|
enum efc_sm_event evt, void *arg)
|
|
{
|
|
struct efc_node *node = ctx->app;
|
|
|
|
efc_node_evt_set(ctx, evt, __func__);
|
|
|
|
/*
|
|
* State is entered when a node sends a delete initiator/target call
|
|
* to the target-server/initiator-client and needs to wait for that
|
|
* work to complete.
|
|
*/
|
|
node_sm_trace();
|
|
|
|
switch (evt) {
|
|
case EFC_EVT_ENTER:
|
|
efc_node_hold_frames(node);
|
|
fallthrough;
|
|
|
|
case EFC_EVT_NODE_ACTIVE_IO_LIST_EMPTY:
|
|
case EFC_EVT_ALL_CHILD_NODES_FREE:
|
|
/* These are expected events. */
|
|
break;
|
|
|
|
case EFC_EVT_NODE_DEL_INI_COMPLETE:
|
|
case EFC_EVT_NODE_DEL_TGT_COMPLETE:
|
|
/*
|
|
* node has either been detached or is in the process
|
|
* of being detached,
|
|
* call common node's initiate cleanup function
|
|
*/
|
|
efc_node_initiate_cleanup(node);
|
|
break;
|
|
|
|
case EFC_EVT_EXIT:
|
|
efc_node_accept_frames(node);
|
|
break;
|
|
|
|
case EFC_EVT_SRRS_ELS_REQ_FAIL:
|
|
/* Can happen as ELS IO IO's complete */
|
|
WARN_ON(!node->els_req_cnt);
|
|
node->els_req_cnt--;
|
|
break;
|
|
|
|
/* ignore shutdown events as we're already in shutdown path */
|
|
case EFC_EVT_SHUTDOWN:
|
|
/* have default shutdown event take precedence */
|
|
node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT;
|
|
fallthrough;
|
|
|
|
case EFC_EVT_SHUTDOWN_EXPLICIT_LOGO:
|
|
case EFC_EVT_SHUTDOWN_IMPLICIT_LOGO:
|
|
node_printf(node, "%s received\n", efc_sm_event_name(evt));
|
|
break;
|
|
case EFC_EVT_DOMAIN_ATTACH_OK:
|
|
/* don't care about domain_attach_ok */
|
|
break;
|
|
default:
|
|
__efc_d_common(__func__, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
static void
|
|
__efc_d_wait_del_ini_tgt(struct efc_sm_ctx *ctx,
|
|
enum efc_sm_event evt, void *arg)
|
|
{
|
|
struct efc_node *node = ctx->app;
|
|
|
|
efc_node_evt_set(ctx, evt, __func__);
|
|
|
|
node_sm_trace();
|
|
|
|
switch (evt) {
|
|
case EFC_EVT_ENTER:
|
|
efc_node_hold_frames(node);
|
|
fallthrough;
|
|
|
|
case EFC_EVT_NODE_ACTIVE_IO_LIST_EMPTY:
|
|
case EFC_EVT_ALL_CHILD_NODES_FREE:
|
|
/* These are expected events. */
|
|
break;
|
|
|
|
case EFC_EVT_NODE_DEL_INI_COMPLETE:
|
|
case EFC_EVT_NODE_DEL_TGT_COMPLETE:
|
|
efc_node_transition(node, __efc_d_wait_del_node, NULL);
|
|
break;
|
|
|
|
case EFC_EVT_EXIT:
|
|
efc_node_accept_frames(node);
|
|
break;
|
|
|
|
case EFC_EVT_SRRS_ELS_REQ_FAIL:
|
|
/* Can happen as ELS IO IO's complete */
|
|
WARN_ON(!node->els_req_cnt);
|
|
node->els_req_cnt--;
|
|
break;
|
|
|
|
/* ignore shutdown events as we're already in shutdown path */
|
|
case EFC_EVT_SHUTDOWN:
|
|
/* have default shutdown event take precedence */
|
|
node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT;
|
|
fallthrough;
|
|
|
|
case EFC_EVT_SHUTDOWN_EXPLICIT_LOGO:
|
|
case EFC_EVT_SHUTDOWN_IMPLICIT_LOGO:
|
|
node_printf(node, "%s received\n", efc_sm_event_name(evt));
|
|
break;
|
|
case EFC_EVT_DOMAIN_ATTACH_OK:
|
|
/* don't care about domain_attach_ok */
|
|
break;
|
|
default:
|
|
__efc_d_common(__func__, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
void
|
|
__efc_d_initiate_shutdown(struct efc_sm_ctx *ctx,
|
|
enum efc_sm_event evt, void *arg)
|
|
{
|
|
struct efc_node *node = ctx->app;
|
|
struct efc *efc = node->efc;
|
|
|
|
efc_node_evt_set(ctx, evt, __func__);
|
|
|
|
node_sm_trace();
|
|
|
|
switch (evt) {
|
|
case EFC_EVT_ENTER: {
|
|
int rc = EFC_SCSI_CALL_COMPLETE;
|
|
|
|
/* assume no wait needed */
|
|
node->els_io_enabled = false;
|
|
|
|
/* make necessary delete upcall(s) */
|
|
if (node->init && !node->targ) {
|
|
efc_log_info(node->efc,
|
|
"[%s] delete (initiator) WWPN %s WWNN %s\n",
|
|
node->display_name,
|
|
node->wwpn, node->wwnn);
|
|
efc_node_transition(node,
|
|
__efc_d_wait_del_node,
|
|
NULL);
|
|
if (node->nport->enable_tgt)
|
|
rc = efc->tt.scsi_del_node(efc, node,
|
|
EFC_SCSI_INITIATOR_DELETED);
|
|
|
|
if (rc == EFC_SCSI_CALL_COMPLETE || rc < 0)
|
|
efc_node_post_event(node,
|
|
EFC_EVT_NODE_DEL_INI_COMPLETE, NULL);
|
|
|
|
} else if (node->targ && !node->init) {
|
|
efc_log_info(node->efc,
|
|
"[%s] delete (target) WWPN %s WWNN %s\n",
|
|
node->display_name,
|
|
node->wwpn, node->wwnn);
|
|
efc_node_transition(node,
|
|
__efc_d_wait_del_node,
|
|
NULL);
|
|
if (node->nport->enable_ini)
|
|
rc = efc->tt.scsi_del_node(efc, node,
|
|
EFC_SCSI_TARGET_DELETED);
|
|
|
|
if (rc == EFC_SCSI_CALL_COMPLETE)
|
|
efc_node_post_event(node,
|
|
EFC_EVT_NODE_DEL_TGT_COMPLETE, NULL);
|
|
|
|
} else if (node->init && node->targ) {
|
|
efc_log_info(node->efc,
|
|
"[%s] delete (I+T) WWPN %s WWNN %s\n",
|
|
node->display_name, node->wwpn, node->wwnn);
|
|
efc_node_transition(node, __efc_d_wait_del_ini_tgt,
|
|
NULL);
|
|
if (node->nport->enable_tgt)
|
|
rc = efc->tt.scsi_del_node(efc, node,
|
|
EFC_SCSI_INITIATOR_DELETED);
|
|
|
|
if (rc == EFC_SCSI_CALL_COMPLETE)
|
|
efc_node_post_event(node,
|
|
EFC_EVT_NODE_DEL_INI_COMPLETE, NULL);
|
|
/* assume no wait needed */
|
|
rc = EFC_SCSI_CALL_COMPLETE;
|
|
if (node->nport->enable_ini)
|
|
rc = efc->tt.scsi_del_node(efc, node,
|
|
EFC_SCSI_TARGET_DELETED);
|
|
|
|
if (rc == EFC_SCSI_CALL_COMPLETE)
|
|
efc_node_post_event(node,
|
|
EFC_EVT_NODE_DEL_TGT_COMPLETE, NULL);
|
|
}
|
|
|
|
/* we've initiated the upcalls as needed, now kick off the node
|
|
* detach to precipitate the aborting of outstanding exchanges
|
|
* associated with said node
|
|
*
|
|
* Beware: if we've made upcall(s), we've already transitioned
|
|
* to a new state by the time we execute this.
|
|
* consider doing this before the upcalls?
|
|
*/
|
|
if (node->attached) {
|
|
/* issue hw node free; don't care if succeeds right
|
|
* away or sometime later, will check node->attached
|
|
* later in shutdown process
|
|
*/
|
|
rc = efc_cmd_node_detach(efc, &node->rnode);
|
|
if (rc < 0)
|
|
node_printf(node,
|
|
"Failed freeing HW node, rc=%d\n",
|
|
rc);
|
|
}
|
|
|
|
/* if neither initiator nor target, proceed to cleanup */
|
|
if (!node->init && !node->targ) {
|
|
/*
|
|
* node has either been detached or is in
|
|
* the process of being detached,
|
|
* call common node's initiate cleanup function
|
|
*/
|
|
efc_node_initiate_cleanup(node);
|
|
}
|
|
break;
|
|
}
|
|
case EFC_EVT_ALL_CHILD_NODES_FREE:
|
|
/* Ignore, this can happen if an ELS is
|
|
* aborted while in a delay/retry state
|
|
*/
|
|
break;
|
|
default:
|
|
__efc_d_common(__func__, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
void
|
|
__efc_d_wait_loop(struct efc_sm_ctx *ctx,
|
|
enum efc_sm_event evt, void *arg)
|
|
{
|
|
struct efc_node *node = ctx->app;
|
|
|
|
efc_node_evt_set(ctx, evt, __func__);
|
|
|
|
node_sm_trace();
|
|
|
|
switch (evt) {
|
|
case EFC_EVT_ENTER:
|
|
efc_node_hold_frames(node);
|
|
break;
|
|
|
|
case EFC_EVT_EXIT:
|
|
efc_node_accept_frames(node);
|
|
break;
|
|
|
|
case EFC_EVT_DOMAIN_ATTACH_OK: {
|
|
/* send PLOGI automatically if initiator */
|
|
efc_node_init_device(node, true);
|
|
break;
|
|
}
|
|
default:
|
|
__efc_d_common(__func__, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
void
|
|
efc_send_ls_acc_after_attach(struct efc_node *node,
|
|
struct fc_frame_header *hdr,
|
|
enum efc_node_send_ls_acc ls)
|
|
{
|
|
u16 ox_id = be16_to_cpu(hdr->fh_ox_id);
|
|
|
|
/* Save the OX_ID for sending LS_ACC sometime later */
|
|
WARN_ON(node->send_ls_acc != EFC_NODE_SEND_LS_ACC_NONE);
|
|
|
|
node->ls_acc_oxid = ox_id;
|
|
node->send_ls_acc = ls;
|
|
node->ls_acc_did = ntoh24(hdr->fh_d_id);
|
|
}
|
|
|
|
void
|
|
efc_process_prli_payload(struct efc_node *node, void *prli)
|
|
{
|
|
struct {
|
|
struct fc_els_prli prli;
|
|
struct fc_els_spp sp;
|
|
} *pp;
|
|
|
|
pp = prli;
|
|
node->init = (pp->sp.spp_flags & FCP_SPPF_INIT_FCN) != 0;
|
|
node->targ = (pp->sp.spp_flags & FCP_SPPF_TARG_FCN) != 0;
|
|
}
|
|
|
|
void
|
|
__efc_d_wait_plogi_acc_cmpl(struct efc_sm_ctx *ctx,
|
|
enum efc_sm_event evt, void *arg)
|
|
{
|
|
struct efc_node *node = ctx->app;
|
|
|
|
efc_node_evt_set(ctx, evt, __func__);
|
|
|
|
node_sm_trace();
|
|
|
|
switch (evt) {
|
|
case EFC_EVT_ENTER:
|
|
efc_node_hold_frames(node);
|
|
break;
|
|
|
|
case EFC_EVT_EXIT:
|
|
efc_node_accept_frames(node);
|
|
break;
|
|
|
|
case EFC_EVT_SRRS_ELS_CMPL_FAIL:
|
|
WARN_ON(!node->els_cmpl_cnt);
|
|
node->els_cmpl_cnt--;
|
|
node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT;
|
|
efc_node_transition(node, __efc_d_initiate_shutdown, NULL);
|
|
break;
|
|
|
|
case EFC_EVT_SRRS_ELS_CMPL_OK: /* PLOGI ACC completions */
|
|
WARN_ON(!node->els_cmpl_cnt);
|
|
node->els_cmpl_cnt--;
|
|
efc_node_transition(node, __efc_d_port_logged_in, NULL);
|
|
break;
|
|
|
|
default:
|
|
__efc_d_common(__func__, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
void
|
|
__efc_d_wait_logo_rsp(struct efc_sm_ctx *ctx,
|
|
enum efc_sm_event evt, void *arg)
|
|
{
|
|
struct efc_node *node = ctx->app;
|
|
|
|
efc_node_evt_set(ctx, evt, __func__);
|
|
|
|
node_sm_trace();
|
|
|
|
switch (evt) {
|
|
case EFC_EVT_ENTER:
|
|
efc_node_hold_frames(node);
|
|
break;
|
|
|
|
case EFC_EVT_EXIT:
|
|
efc_node_accept_frames(node);
|
|
break;
|
|
|
|
case EFC_EVT_SRRS_ELS_REQ_OK:
|
|
case EFC_EVT_SRRS_ELS_REQ_RJT:
|
|
case EFC_EVT_SRRS_ELS_REQ_FAIL:
|
|
/* LOGO response received, sent shutdown */
|
|
if (efc_node_check_els_req(ctx, evt, arg, ELS_LOGO,
|
|
__efc_d_common, __func__))
|
|
return;
|
|
|
|
WARN_ON(!node->els_req_cnt);
|
|
node->els_req_cnt--;
|
|
node_printf(node,
|
|
"LOGO sent (evt=%s), shutdown node\n",
|
|
efc_sm_event_name(evt));
|
|
/* sm: / post explicit logout */
|
|
efc_node_post_event(node, EFC_EVT_SHUTDOWN_EXPLICIT_LOGO,
|
|
NULL);
|
|
break;
|
|
|
|
default:
|
|
__efc_d_common(__func__, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
void
|
|
efc_node_init_device(struct efc_node *node, bool send_plogi)
|
|
{
|
|
node->send_plogi = send_plogi;
|
|
if ((node->efc->nodedb_mask & EFC_NODEDB_PAUSE_NEW_NODES) &&
|
|
(node->rnode.fc_id != FC_FID_DOM_MGR)) {
|
|
node->nodedb_state = __efc_d_init;
|
|
efc_node_transition(node, __efc_node_paused, NULL);
|
|
} else {
|
|
efc_node_transition(node, __efc_d_init, NULL);
|
|
}
|
|
}
|
|
|
|
static void
|
|
efc_d_check_plogi_topology(struct efc_node *node, u32 d_id)
|
|
{
|
|
switch (node->nport->topology) {
|
|
case EFC_NPORT_TOPO_P2P:
|
|
/* we're not attached and nport is p2p,
|
|
* need to attach
|
|
*/
|
|
efc_domain_attach(node->nport->domain, d_id);
|
|
efc_node_transition(node, __efc_d_wait_domain_attach, NULL);
|
|
break;
|
|
case EFC_NPORT_TOPO_FABRIC:
|
|
/* we're not attached and nport is fabric, domain
|
|
* attach should have already been requested as part
|
|
* of the fabric state machine, wait for it
|
|
*/
|
|
efc_node_transition(node, __efc_d_wait_domain_attach, NULL);
|
|
break;
|
|
case EFC_NPORT_TOPO_UNKNOWN:
|
|
/* Two possibilities:
|
|
* 1. received a PLOGI before our FLOGI has completed
|
|
* (possible since completion comes in on another
|
|
* CQ), thus we don't know what we're connected to
|
|
* yet; transition to a state to wait for the
|
|
* fabric node to tell us;
|
|
* 2. PLOGI received before link went down and we
|
|
* haven't performed domain attach yet.
|
|
* Note: we cannot distinguish between 1. and 2.
|
|
* so have to assume PLOGI
|
|
* was received after link back up.
|
|
*/
|
|
node_printf(node, "received PLOGI, unknown topology did=0x%x\n",
|
|
d_id);
|
|
efc_node_transition(node, __efc_d_wait_topology_notify, NULL);
|
|
break;
|
|
default:
|
|
node_printf(node, "received PLOGI, unexpected topology %d\n",
|
|
node->nport->topology);
|
|
}
|
|
}
|
|
|
|
void
|
|
__efc_d_init(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg)
|
|
{
|
|
struct efc_node_cb *cbdata = arg;
|
|
struct efc_node *node = ctx->app;
|
|
|
|
efc_node_evt_set(ctx, evt, __func__);
|
|
|
|
node_sm_trace();
|
|
|
|
/*
|
|
* This state is entered when a node is instantiated,
|
|
* either having been discovered from a name services query,
|
|
* or having received a PLOGI/FLOGI.
|
|
*/
|
|
switch (evt) {
|
|
case EFC_EVT_ENTER:
|
|
if (!node->send_plogi)
|
|
break;
|
|
/* only send if we have initiator capability,
|
|
* and domain is attached
|
|
*/
|
|
if (node->nport->enable_ini &&
|
|
node->nport->domain->attached) {
|
|
efc_send_plogi(node);
|
|
|
|
efc_node_transition(node, __efc_d_wait_plogi_rsp, NULL);
|
|
} else {
|
|
node_printf(node, "not sending plogi nport.ini=%d,",
|
|
node->nport->enable_ini);
|
|
node_printf(node, "domain attached=%d\n",
|
|
node->nport->domain->attached);
|
|
}
|
|
break;
|
|
case EFC_EVT_PLOGI_RCVD: {
|
|
/* T, or I+T */
|
|
struct fc_frame_header *hdr = cbdata->header->dma.virt;
|
|
int rc;
|
|
|
|
efc_node_save_sparms(node, cbdata->payload->dma.virt);
|
|
efc_send_ls_acc_after_attach(node,
|
|
cbdata->header->dma.virt,
|
|
EFC_NODE_SEND_LS_ACC_PLOGI);
|
|
|
|
/* domain not attached; several possibilities: */
|
|
if (!node->nport->domain->attached) {
|
|
efc_d_check_plogi_topology(node, ntoh24(hdr->fh_d_id));
|
|
break;
|
|
}
|
|
|
|
/* domain already attached */
|
|
rc = efc_node_attach(node);
|
|
efc_node_transition(node, __efc_d_wait_node_attach, NULL);
|
|
if (rc < 0)
|
|
efc_node_post_event(node, EFC_EVT_NODE_ATTACH_FAIL, NULL);
|
|
|
|
break;
|
|
}
|
|
|
|
case EFC_EVT_FDISC_RCVD: {
|
|
__efc_d_common(__func__, ctx, evt, arg);
|
|
break;
|
|
}
|
|
|
|
case EFC_EVT_FLOGI_RCVD: {
|
|
struct fc_frame_header *hdr = cbdata->header->dma.virt;
|
|
u32 d_id = ntoh24(hdr->fh_d_id);
|
|
|
|
/* sm: / save sparams, send FLOGI acc */
|
|
memcpy(node->nport->domain->flogi_service_params,
|
|
cbdata->payload->dma.virt,
|
|
sizeof(struct fc_els_flogi));
|
|
|
|
/* send FC LS_ACC response, override s_id */
|
|
efc_fabric_set_topology(node, EFC_NPORT_TOPO_P2P);
|
|
|
|
efc_send_flogi_p2p_acc(node, be16_to_cpu(hdr->fh_ox_id), d_id);
|
|
|
|
if (efc_p2p_setup(node->nport)) {
|
|
node_printf(node, "p2p failed, shutting down node\n");
|
|
efc_node_post_event(node, EFC_EVT_SHUTDOWN, NULL);
|
|
break;
|
|
}
|
|
|
|
efc_node_transition(node, __efc_p2p_wait_flogi_acc_cmpl, NULL);
|
|
break;
|
|
}
|
|
|
|
case EFC_EVT_LOGO_RCVD: {
|
|
struct fc_frame_header *hdr = cbdata->header->dma.virt;
|
|
|
|
if (!node->nport->domain->attached) {
|
|
/* most likely a frame left over from before a link
|
|
* down; drop and
|
|
* shut node down w/ "explicit logout" so pending
|
|
* frames are processed
|
|
*/
|
|
node_printf(node, "%s domain not attached, dropping\n",
|
|
efc_sm_event_name(evt));
|
|
efc_node_post_event(node,
|
|
EFC_EVT_SHUTDOWN_EXPLICIT_LOGO, NULL);
|
|
break;
|
|
}
|
|
|
|
efc_send_logo_acc(node, be16_to_cpu(hdr->fh_ox_id));
|
|
efc_node_transition(node, __efc_d_wait_logo_acc_cmpl, NULL);
|
|
break;
|
|
}
|
|
|
|
case EFC_EVT_PRLI_RCVD:
|
|
case EFC_EVT_PRLO_RCVD:
|
|
case EFC_EVT_PDISC_RCVD:
|
|
case EFC_EVT_ADISC_RCVD:
|
|
case EFC_EVT_RSCN_RCVD: {
|
|
struct fc_frame_header *hdr = cbdata->header->dma.virt;
|
|
|
|
if (!node->nport->domain->attached) {
|
|
/* most likely a frame left over from before a link
|
|
* down; drop and shut node down w/ "explicit logout"
|
|
* so pending frames are processed
|
|
*/
|
|
node_printf(node, "%s domain not attached, dropping\n",
|
|
efc_sm_event_name(evt));
|
|
|
|
efc_node_post_event(node,
|
|
EFC_EVT_SHUTDOWN_EXPLICIT_LOGO,
|
|
NULL);
|
|
break;
|
|
}
|
|
node_printf(node, "%s received, sending reject\n",
|
|
efc_sm_event_name(evt));
|
|
|
|
efc_send_ls_rjt(node, be16_to_cpu(hdr->fh_ox_id),
|
|
ELS_RJT_UNAB, ELS_EXPL_PLOGI_REQD, 0);
|
|
|
|
break;
|
|
}
|
|
|
|
case EFC_EVT_FCP_CMD_RCVD: {
|
|
/* note: problem, we're now expecting an ELS REQ completion
|
|
* from both the LOGO and PLOGI
|
|
*/
|
|
if (!node->nport->domain->attached) {
|
|
/* most likely a frame left over from before a
|
|
* link down; drop and
|
|
* shut node down w/ "explicit logout" so pending
|
|
* frames are processed
|
|
*/
|
|
node_printf(node, "%s domain not attached, dropping\n",
|
|
efc_sm_event_name(evt));
|
|
efc_node_post_event(node,
|
|
EFC_EVT_SHUTDOWN_EXPLICIT_LOGO,
|
|
NULL);
|
|
break;
|
|
}
|
|
|
|
/* Send LOGO */
|
|
node_printf(node, "FCP_CMND received, send LOGO\n");
|
|
if (efc_send_logo(node)) {
|
|
/*
|
|
* failed to send LOGO, go ahead and cleanup node
|
|
* anyways
|
|
*/
|
|
node_printf(node, "Failed to send LOGO\n");
|
|
efc_node_post_event(node,
|
|
EFC_EVT_SHUTDOWN_EXPLICIT_LOGO,
|
|
NULL);
|
|
} else {
|
|
/* sent LOGO, wait for response */
|
|
efc_node_transition(node,
|
|
__efc_d_wait_logo_rsp, NULL);
|
|
}
|
|
break;
|
|
}
|
|
case EFC_EVT_DOMAIN_ATTACH_OK:
|
|
/* don't care about domain_attach_ok */
|
|
break;
|
|
|
|
default:
|
|
__efc_d_common(__func__, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
void
|
|
__efc_d_wait_plogi_rsp(struct efc_sm_ctx *ctx,
|
|
enum efc_sm_event evt, void *arg)
|
|
{
|
|
int rc;
|
|
struct efc_node_cb *cbdata = arg;
|
|
struct efc_node *node = ctx->app;
|
|
|
|
efc_node_evt_set(ctx, evt, __func__);
|
|
|
|
node_sm_trace();
|
|
|
|
switch (evt) {
|
|
case EFC_EVT_PLOGI_RCVD: {
|
|
/* T, or I+T */
|
|
/* received PLOGI with svc parms, go ahead and attach node
|
|
* when PLOGI that was sent ultimately completes, it'll be a
|
|
* no-op
|
|
*
|
|
* If there is an outstanding PLOGI sent, can we set a flag
|
|
* to indicate that we don't want to retry it if it times out?
|
|
*/
|
|
efc_node_save_sparms(node, cbdata->payload->dma.virt);
|
|
efc_send_ls_acc_after_attach(node,
|
|
cbdata->header->dma.virt,
|
|
EFC_NODE_SEND_LS_ACC_PLOGI);
|
|
/* sm: domain->attached / efc_node_attach */
|
|
rc = efc_node_attach(node);
|
|
efc_node_transition(node, __efc_d_wait_node_attach, NULL);
|
|
if (rc < 0)
|
|
efc_node_post_event(node,
|
|
EFC_EVT_NODE_ATTACH_FAIL, NULL);
|
|
|
|
break;
|
|
}
|
|
|
|
case EFC_EVT_PRLI_RCVD:
|
|
/* I, or I+T */
|
|
/* sent PLOGI and before completion was seen, received the
|
|
* PRLI from the remote node (WCQEs and RCQEs come in on
|
|
* different queues and order of processing cannot be assumed)
|
|
* Save OXID so PRLI can be sent after the attach and continue
|
|
* to wait for PLOGI response
|
|
*/
|
|
efc_process_prli_payload(node, cbdata->payload->dma.virt);
|
|
efc_send_ls_acc_after_attach(node,
|
|
cbdata->header->dma.virt,
|
|
EFC_NODE_SEND_LS_ACC_PRLI);
|
|
efc_node_transition(node, __efc_d_wait_plogi_rsp_recvd_prli,
|
|
NULL);
|
|
break;
|
|
|
|
case EFC_EVT_LOGO_RCVD: /* why don't we do a shutdown here?? */
|
|
case EFC_EVT_PRLO_RCVD:
|
|
case EFC_EVT_PDISC_RCVD:
|
|
case EFC_EVT_FDISC_RCVD:
|
|
case EFC_EVT_ADISC_RCVD:
|
|
case EFC_EVT_RSCN_RCVD:
|
|
case EFC_EVT_SCR_RCVD: {
|
|
struct fc_frame_header *hdr = cbdata->header->dma.virt;
|
|
|
|
node_printf(node, "%s received, sending reject\n",
|
|
efc_sm_event_name(evt));
|
|
|
|
efc_send_ls_rjt(node, be16_to_cpu(hdr->fh_ox_id),
|
|
ELS_RJT_UNAB, ELS_EXPL_PLOGI_REQD, 0);
|
|
|
|
break;
|
|
}
|
|
|
|
case EFC_EVT_SRRS_ELS_REQ_OK: /* PLOGI response received */
|
|
/* Completion from PLOGI sent */
|
|
if (efc_node_check_els_req(ctx, evt, arg, ELS_PLOGI,
|
|
__efc_d_common, __func__))
|
|
return;
|
|
|
|
WARN_ON(!node->els_req_cnt);
|
|
node->els_req_cnt--;
|
|
/* sm: / save sparams, efc_node_attach */
|
|
efc_node_save_sparms(node, cbdata->els_rsp.virt);
|
|
rc = efc_node_attach(node);
|
|
efc_node_transition(node, __efc_d_wait_node_attach, NULL);
|
|
if (rc < 0)
|
|
efc_node_post_event(node,
|
|
EFC_EVT_NODE_ATTACH_FAIL, NULL);
|
|
|
|
break;
|
|
|
|
case EFC_EVT_SRRS_ELS_REQ_FAIL: /* PLOGI response received */
|
|
/* PLOGI failed, shutdown the node */
|
|
if (efc_node_check_els_req(ctx, evt, arg, ELS_PLOGI,
|
|
__efc_d_common, __func__))
|
|
return;
|
|
|
|
WARN_ON(!node->els_req_cnt);
|
|
node->els_req_cnt--;
|
|
efc_node_post_event(node, EFC_EVT_SHUTDOWN, NULL);
|
|
break;
|
|
|
|
case EFC_EVT_SRRS_ELS_REQ_RJT:
|
|
/* Our PLOGI was rejected, this is ok in some cases */
|
|
if (efc_node_check_els_req(ctx, evt, arg, ELS_PLOGI,
|
|
__efc_d_common, __func__))
|
|
return;
|
|
|
|
WARN_ON(!node->els_req_cnt);
|
|
node->els_req_cnt--;
|
|
break;
|
|
|
|
case EFC_EVT_FCP_CMD_RCVD: {
|
|
/* not logged in yet and outstanding PLOGI so don't send LOGO,
|
|
* just drop
|
|
*/
|
|
node_printf(node, "FCP_CMND received, drop\n");
|
|
break;
|
|
}
|
|
|
|
default:
|
|
__efc_d_common(__func__, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
void
|
|
__efc_d_wait_plogi_rsp_recvd_prli(struct efc_sm_ctx *ctx,
|
|
enum efc_sm_event evt, void *arg)
|
|
{
|
|
int rc;
|
|
struct efc_node_cb *cbdata = arg;
|
|
struct efc_node *node = ctx->app;
|
|
|
|
efc_node_evt_set(ctx, evt, __func__);
|
|
|
|
node_sm_trace();
|
|
|
|
switch (evt) {
|
|
case EFC_EVT_ENTER:
|
|
/*
|
|
* Since we've received a PRLI, we have a port login and will
|
|
* just need to wait for the PLOGI response to do the node
|
|
* attach and then we can send the LS_ACC for the PRLI. If,
|
|
* during this time, we receive FCP_CMNDs (which is possible
|
|
* since we've already sent a PRLI and our peer may have
|
|
* accepted). At this time, we are not waiting on any other
|
|
* unsolicited frames to continue with the login process. Thus,
|
|
* it will not hurt to hold frames here.
|
|
*/
|
|
efc_node_hold_frames(node);
|
|
break;
|
|
|
|
case EFC_EVT_EXIT:
|
|
efc_node_accept_frames(node);
|
|
break;
|
|
|
|
case EFC_EVT_SRRS_ELS_REQ_OK: /* PLOGI response received */
|
|
/* Completion from PLOGI sent */
|
|
if (efc_node_check_els_req(ctx, evt, arg, ELS_PLOGI,
|
|
__efc_d_common, __func__))
|
|
return;
|
|
|
|
WARN_ON(!node->els_req_cnt);
|
|
node->els_req_cnt--;
|
|
/* sm: / save sparams, efc_node_attach */
|
|
efc_node_save_sparms(node, cbdata->els_rsp.virt);
|
|
rc = efc_node_attach(node);
|
|
efc_node_transition(node, __efc_d_wait_node_attach, NULL);
|
|
if (rc < 0)
|
|
efc_node_post_event(node, EFC_EVT_NODE_ATTACH_FAIL,
|
|
NULL);
|
|
|
|
break;
|
|
|
|
case EFC_EVT_SRRS_ELS_REQ_FAIL: /* PLOGI response received */
|
|
case EFC_EVT_SRRS_ELS_REQ_RJT:
|
|
/* PLOGI failed, shutdown the node */
|
|
if (efc_node_check_els_req(ctx, evt, arg, ELS_PLOGI,
|
|
__efc_d_common, __func__))
|
|
return;
|
|
|
|
WARN_ON(!node->els_req_cnt);
|
|
node->els_req_cnt--;
|
|
efc_node_post_event(node, EFC_EVT_SHUTDOWN, NULL);
|
|
break;
|
|
|
|
default:
|
|
__efc_d_common(__func__, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
void
|
|
__efc_d_wait_domain_attach(struct efc_sm_ctx *ctx,
|
|
enum efc_sm_event evt, void *arg)
|
|
{
|
|
int rc;
|
|
struct efc_node *node = ctx->app;
|
|
|
|
efc_node_evt_set(ctx, evt, __func__);
|
|
|
|
node_sm_trace();
|
|
|
|
switch (evt) {
|
|
case EFC_EVT_ENTER:
|
|
efc_node_hold_frames(node);
|
|
break;
|
|
|
|
case EFC_EVT_EXIT:
|
|
efc_node_accept_frames(node);
|
|
break;
|
|
|
|
case EFC_EVT_DOMAIN_ATTACH_OK:
|
|
WARN_ON(!node->nport->domain->attached);
|
|
/* sm: / efc_node_attach */
|
|
rc = efc_node_attach(node);
|
|
efc_node_transition(node, __efc_d_wait_node_attach, NULL);
|
|
if (rc < 0)
|
|
efc_node_post_event(node, EFC_EVT_NODE_ATTACH_FAIL,
|
|
NULL);
|
|
|
|
break;
|
|
|
|
default:
|
|
__efc_d_common(__func__, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
void
|
|
__efc_d_wait_topology_notify(struct efc_sm_ctx *ctx,
|
|
enum efc_sm_event evt, void *arg)
|
|
{
|
|
int rc;
|
|
struct efc_node *node = ctx->app;
|
|
|
|
efc_node_evt_set(ctx, evt, __func__);
|
|
|
|
node_sm_trace();
|
|
|
|
switch (evt) {
|
|
case EFC_EVT_ENTER:
|
|
efc_node_hold_frames(node);
|
|
break;
|
|
|
|
case EFC_EVT_EXIT:
|
|
efc_node_accept_frames(node);
|
|
break;
|
|
|
|
case EFC_EVT_NPORT_TOPOLOGY_NOTIFY: {
|
|
enum efc_nport_topology *topology = arg;
|
|
|
|
WARN_ON(node->nport->domain->attached);
|
|
|
|
WARN_ON(node->send_ls_acc != EFC_NODE_SEND_LS_ACC_PLOGI);
|
|
|
|
node_printf(node, "topology notification, topology=%d\n",
|
|
*topology);
|
|
|
|
/* At the time the PLOGI was received, the topology was unknown,
|
|
* so we didn't know which node would perform the domain attach:
|
|
* 1. The node from which the PLOGI was sent (p2p) or
|
|
* 2. The node to which the FLOGI was sent (fabric).
|
|
*/
|
|
if (*topology == EFC_NPORT_TOPO_P2P) {
|
|
/* if this is p2p, need to attach to the domain using
|
|
* the d_id from the PLOGI received
|
|
*/
|
|
efc_domain_attach(node->nport->domain,
|
|
node->ls_acc_did);
|
|
}
|
|
/* else, if this is fabric, the domain attach
|
|
* should be performed by the fabric node (node sending FLOGI);
|
|
* just wait for attach to complete
|
|
*/
|
|
|
|
efc_node_transition(node, __efc_d_wait_domain_attach, NULL);
|
|
break;
|
|
}
|
|
case EFC_EVT_DOMAIN_ATTACH_OK:
|
|
WARN_ON(!node->nport->domain->attached);
|
|
node_printf(node, "domain attach ok\n");
|
|
/* sm: / efc_node_attach */
|
|
rc = efc_node_attach(node);
|
|
efc_node_transition(node, __efc_d_wait_node_attach, NULL);
|
|
if (rc < 0)
|
|
efc_node_post_event(node,
|
|
EFC_EVT_NODE_ATTACH_FAIL, NULL);
|
|
|
|
break;
|
|
|
|
default:
|
|
__efc_d_common(__func__, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
void
|
|
__efc_d_wait_node_attach(struct efc_sm_ctx *ctx,
|
|
enum efc_sm_event evt, void *arg)
|
|
{
|
|
struct efc_node *node = ctx->app;
|
|
|
|
efc_node_evt_set(ctx, evt, __func__);
|
|
|
|
node_sm_trace();
|
|
|
|
switch (evt) {
|
|
case EFC_EVT_ENTER:
|
|
efc_node_hold_frames(node);
|
|
break;
|
|
|
|
case EFC_EVT_EXIT:
|
|
efc_node_accept_frames(node);
|
|
break;
|
|
|
|
case EFC_EVT_NODE_ATTACH_OK:
|
|
node->attached = true;
|
|
switch (node->send_ls_acc) {
|
|
case EFC_NODE_SEND_LS_ACC_PLOGI: {
|
|
/* sm: send_plogi_acc is set / send PLOGI acc */
|
|
/* Normal case for T, or I+T */
|
|
efc_send_plogi_acc(node, node->ls_acc_oxid);
|
|
efc_node_transition(node, __efc_d_wait_plogi_acc_cmpl,
|
|
NULL);
|
|
node->send_ls_acc = EFC_NODE_SEND_LS_ACC_NONE;
|
|
node->ls_acc_io = NULL;
|
|
break;
|
|
}
|
|
case EFC_NODE_SEND_LS_ACC_PRLI: {
|
|
efc_d_send_prli_rsp(node, node->ls_acc_oxid);
|
|
node->send_ls_acc = EFC_NODE_SEND_LS_ACC_NONE;
|
|
node->ls_acc_io = NULL;
|
|
break;
|
|
}
|
|
case EFC_NODE_SEND_LS_ACC_NONE:
|
|
default:
|
|
/* Normal case for I */
|
|
/* sm: send_plogi_acc is not set / send PLOGI acc */
|
|
efc_node_transition(node,
|
|
__efc_d_port_logged_in, NULL);
|
|
break;
|
|
}
|
|
break;
|
|
|
|
case EFC_EVT_NODE_ATTACH_FAIL:
|
|
/* node attach failed, shutdown the node */
|
|
node->attached = false;
|
|
node_printf(node, "node attach failed\n");
|
|
node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT;
|
|
efc_node_transition(node, __efc_d_initiate_shutdown, NULL);
|
|
break;
|
|
|
|
/* Handle shutdown events */
|
|
case EFC_EVT_SHUTDOWN:
|
|
node_printf(node, "%s received\n", efc_sm_event_name(evt));
|
|
node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT;
|
|
efc_node_transition(node, __efc_d_wait_attach_evt_shutdown,
|
|
NULL);
|
|
break;
|
|
case EFC_EVT_SHUTDOWN_EXPLICIT_LOGO:
|
|
node_printf(node, "%s received\n", efc_sm_event_name(evt));
|
|
node->shutdown_reason = EFC_NODE_SHUTDOWN_EXPLICIT_LOGO;
|
|
efc_node_transition(node, __efc_d_wait_attach_evt_shutdown,
|
|
NULL);
|
|
break;
|
|
case EFC_EVT_SHUTDOWN_IMPLICIT_LOGO:
|
|
node_printf(node, "%s received\n", efc_sm_event_name(evt));
|
|
node->shutdown_reason = EFC_NODE_SHUTDOWN_IMPLICIT_LOGO;
|
|
efc_node_transition(node,
|
|
__efc_d_wait_attach_evt_shutdown, NULL);
|
|
break;
|
|
default:
|
|
__efc_d_common(__func__, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
void
|
|
__efc_d_wait_attach_evt_shutdown(struct efc_sm_ctx *ctx,
|
|
enum efc_sm_event evt, void *arg)
|
|
{
|
|
struct efc_node *node = ctx->app;
|
|
|
|
efc_node_evt_set(ctx, evt, __func__);
|
|
|
|
node_sm_trace();
|
|
|
|
switch (evt) {
|
|
case EFC_EVT_ENTER:
|
|
efc_node_hold_frames(node);
|
|
break;
|
|
|
|
case EFC_EVT_EXIT:
|
|
efc_node_accept_frames(node);
|
|
break;
|
|
|
|
/* wait for any of these attach events and then shutdown */
|
|
case EFC_EVT_NODE_ATTACH_OK:
|
|
node->attached = true;
|
|
node_printf(node, "Attach evt=%s, proceed to shutdown\n",
|
|
efc_sm_event_name(evt));
|
|
efc_node_transition(node, __efc_d_initiate_shutdown, NULL);
|
|
break;
|
|
|
|
case EFC_EVT_NODE_ATTACH_FAIL:
|
|
/* node attach failed, shutdown the node */
|
|
node->attached = false;
|
|
node_printf(node, "Attach evt=%s, proceed to shutdown\n",
|
|
efc_sm_event_name(evt));
|
|
efc_node_transition(node, __efc_d_initiate_shutdown, NULL);
|
|
break;
|
|
|
|
/* ignore shutdown events as we're already in shutdown path */
|
|
case EFC_EVT_SHUTDOWN:
|
|
/* have default shutdown event take precedence */
|
|
node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT;
|
|
fallthrough;
|
|
|
|
case EFC_EVT_SHUTDOWN_EXPLICIT_LOGO:
|
|
case EFC_EVT_SHUTDOWN_IMPLICIT_LOGO:
|
|
node_printf(node, "%s received\n", efc_sm_event_name(evt));
|
|
break;
|
|
|
|
default:
|
|
__efc_d_common(__func__, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
void
|
|
__efc_d_port_logged_in(struct efc_sm_ctx *ctx,
|
|
enum efc_sm_event evt, void *arg)
|
|
{
|
|
struct efc_node_cb *cbdata = arg;
|
|
struct efc_node *node = ctx->app;
|
|
|
|
efc_node_evt_set(ctx, evt, __func__);
|
|
|
|
node_sm_trace();
|
|
|
|
switch (evt) {
|
|
case EFC_EVT_ENTER:
|
|
/* Normal case for I or I+T */
|
|
if (node->nport->enable_ini &&
|
|
!(node->rnode.fc_id != FC_FID_DOM_MGR)) {
|
|
/* sm: if enable_ini / send PRLI */
|
|
efc_send_prli(node);
|
|
/* can now expect ELS_REQ_OK/FAIL/RJT */
|
|
}
|
|
break;
|
|
|
|
case EFC_EVT_FCP_CMD_RCVD: {
|
|
break;
|
|
}
|
|
|
|
case EFC_EVT_PRLI_RCVD: {
|
|
/* Normal case for T or I+T */
|
|
struct fc_frame_header *hdr = cbdata->header->dma.virt;
|
|
struct {
|
|
struct fc_els_prli prli;
|
|
struct fc_els_spp sp;
|
|
} *pp;
|
|
|
|
pp = cbdata->payload->dma.virt;
|
|
if (pp->sp.spp_type != FC_TYPE_FCP) {
|
|
/*Only FCP is supported*/
|
|
efc_send_ls_rjt(node, be16_to_cpu(hdr->fh_ox_id),
|
|
ELS_RJT_UNAB, ELS_EXPL_UNSUPR, 0);
|
|
break;
|
|
}
|
|
|
|
efc_process_prli_payload(node, cbdata->payload->dma.virt);
|
|
efc_d_send_prli_rsp(node, be16_to_cpu(hdr->fh_ox_id));
|
|
break;
|
|
}
|
|
|
|
case EFC_EVT_NODE_SESS_REG_OK:
|
|
if (node->send_ls_acc == EFC_NODE_SEND_LS_ACC_PRLI)
|
|
efc_send_prli_acc(node, node->ls_acc_oxid);
|
|
|
|
node->send_ls_acc = EFC_NODE_SEND_LS_ACC_NONE;
|
|
efc_node_transition(node, __efc_d_device_ready, NULL);
|
|
break;
|
|
|
|
case EFC_EVT_NODE_SESS_REG_FAIL:
|
|
efc_send_ls_rjt(node, node->ls_acc_oxid, ELS_RJT_UNAB,
|
|
ELS_EXPL_UNSUPR, 0);
|
|
node->send_ls_acc = EFC_NODE_SEND_LS_ACC_NONE;
|
|
break;
|
|
|
|
case EFC_EVT_SRRS_ELS_REQ_OK: { /* PRLI response */
|
|
/* Normal case for I or I+T */
|
|
if (efc_node_check_els_req(ctx, evt, arg, ELS_PRLI,
|
|
__efc_d_common, __func__))
|
|
return;
|
|
|
|
WARN_ON(!node->els_req_cnt);
|
|
node->els_req_cnt--;
|
|
/* sm: / process PRLI payload */
|
|
efc_process_prli_payload(node, cbdata->els_rsp.virt);
|
|
efc_node_transition(node, __efc_d_device_ready, NULL);
|
|
break;
|
|
}
|
|
|
|
case EFC_EVT_SRRS_ELS_REQ_FAIL: { /* PRLI response failed */
|
|
/* I, I+T, assume some link failure, shutdown node */
|
|
if (efc_node_check_els_req(ctx, evt, arg, ELS_PRLI,
|
|
__efc_d_common, __func__))
|
|
return;
|
|
|
|
WARN_ON(!node->els_req_cnt);
|
|
node->els_req_cnt--;
|
|
efc_node_post_event(node, EFC_EVT_SHUTDOWN, NULL);
|
|
break;
|
|
}
|
|
|
|
case EFC_EVT_SRRS_ELS_REQ_RJT: {
|
|
/* PRLI rejected by remote
|
|
* Normal for I, I+T (connected to an I)
|
|
* Node doesn't want to be a target, stay here and wait for a
|
|
* PRLI from the remote node
|
|
* if it really wants to connect to us as target
|
|
*/
|
|
if (efc_node_check_els_req(ctx, evt, arg, ELS_PRLI,
|
|
__efc_d_common, __func__))
|
|
return;
|
|
|
|
WARN_ON(!node->els_req_cnt);
|
|
node->els_req_cnt--;
|
|
break;
|
|
}
|
|
|
|
case EFC_EVT_SRRS_ELS_CMPL_OK: {
|
|
/* Normal T, I+T, target-server rejected the process login */
|
|
/* This would be received only in the case where we sent
|
|
* LS_RJT for the PRLI, so
|
|
* do nothing. (note: as T only we could shutdown the node)
|
|
*/
|
|
WARN_ON(!node->els_cmpl_cnt);
|
|
node->els_cmpl_cnt--;
|
|
break;
|
|
}
|
|
|
|
case EFC_EVT_PLOGI_RCVD: {
|
|
/*sm: / save sparams, set send_plogi_acc,
|
|
*post implicit logout
|
|
* Save plogi parameters
|
|
*/
|
|
efc_node_save_sparms(node, cbdata->payload->dma.virt);
|
|
efc_send_ls_acc_after_attach(node,
|
|
cbdata->header->dma.virt,
|
|
EFC_NODE_SEND_LS_ACC_PLOGI);
|
|
|
|
/* Restart node attach with new service parameters,
|
|
* and send ACC
|
|
*/
|
|
efc_node_post_event(node, EFC_EVT_SHUTDOWN_IMPLICIT_LOGO,
|
|
NULL);
|
|
break;
|
|
}
|
|
|
|
case EFC_EVT_LOGO_RCVD: {
|
|
/* I, T, I+T */
|
|
struct fc_frame_header *hdr = cbdata->header->dma.virt;
|
|
|
|
node_printf(node, "%s received attached=%d\n",
|
|
efc_sm_event_name(evt),
|
|
node->attached);
|
|
/* sm: / send LOGO acc */
|
|
efc_send_logo_acc(node, be16_to_cpu(hdr->fh_ox_id));
|
|
efc_node_transition(node, __efc_d_wait_logo_acc_cmpl, NULL);
|
|
break;
|
|
}
|
|
|
|
default:
|
|
__efc_d_common(__func__, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
void
|
|
__efc_d_wait_logo_acc_cmpl(struct efc_sm_ctx *ctx,
|
|
enum efc_sm_event evt, void *arg)
|
|
{
|
|
struct efc_node *node = ctx->app;
|
|
|
|
efc_node_evt_set(ctx, evt, __func__);
|
|
|
|
node_sm_trace();
|
|
|
|
switch (evt) {
|
|
case EFC_EVT_ENTER:
|
|
efc_node_hold_frames(node);
|
|
break;
|
|
|
|
case EFC_EVT_EXIT:
|
|
efc_node_accept_frames(node);
|
|
break;
|
|
|
|
case EFC_EVT_SRRS_ELS_CMPL_OK:
|
|
case EFC_EVT_SRRS_ELS_CMPL_FAIL:
|
|
/* sm: / post explicit logout */
|
|
WARN_ON(!node->els_cmpl_cnt);
|
|
node->els_cmpl_cnt--;
|
|
efc_node_post_event(node,
|
|
EFC_EVT_SHUTDOWN_EXPLICIT_LOGO, NULL);
|
|
break;
|
|
default:
|
|
__efc_d_common(__func__, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
void
|
|
__efc_d_device_ready(struct efc_sm_ctx *ctx,
|
|
enum efc_sm_event evt, void *arg)
|
|
{
|
|
struct efc_node_cb *cbdata = arg;
|
|
struct efc_node *node = ctx->app;
|
|
struct efc *efc = node->efc;
|
|
|
|
efc_node_evt_set(ctx, evt, __func__);
|
|
|
|
if (evt != EFC_EVT_FCP_CMD_RCVD)
|
|
node_sm_trace();
|
|
|
|
switch (evt) {
|
|
case EFC_EVT_ENTER:
|
|
node->fcp_enabled = true;
|
|
if (node->targ) {
|
|
efc_log_info(efc,
|
|
"[%s] found (target) WWPN %s WWNN %s\n",
|
|
node->display_name,
|
|
node->wwpn, node->wwnn);
|
|
if (node->nport->enable_ini)
|
|
efc->tt.scsi_new_node(efc, node);
|
|
}
|
|
break;
|
|
|
|
case EFC_EVT_EXIT:
|
|
node->fcp_enabled = false;
|
|
break;
|
|
|
|
case EFC_EVT_PLOGI_RCVD: {
|
|
/* sm: / save sparams, set send_plogi_acc, post implicit
|
|
* logout
|
|
* Save plogi parameters
|
|
*/
|
|
efc_node_save_sparms(node, cbdata->payload->dma.virt);
|
|
efc_send_ls_acc_after_attach(node,
|
|
cbdata->header->dma.virt,
|
|
EFC_NODE_SEND_LS_ACC_PLOGI);
|
|
|
|
/*
|
|
* Restart node attach with new service parameters,
|
|
* and send ACC
|
|
*/
|
|
efc_node_post_event(node,
|
|
EFC_EVT_SHUTDOWN_IMPLICIT_LOGO, NULL);
|
|
break;
|
|
}
|
|
|
|
case EFC_EVT_PRLI_RCVD: {
|
|
/* T, I+T: remote initiator is slow to get started */
|
|
struct fc_frame_header *hdr = cbdata->header->dma.virt;
|
|
struct {
|
|
struct fc_els_prli prli;
|
|
struct fc_els_spp sp;
|
|
} *pp;
|
|
|
|
pp = cbdata->payload->dma.virt;
|
|
if (pp->sp.spp_type != FC_TYPE_FCP) {
|
|
/*Only FCP is supported*/
|
|
efc_send_ls_rjt(node, be16_to_cpu(hdr->fh_ox_id),
|
|
ELS_RJT_UNAB, ELS_EXPL_UNSUPR, 0);
|
|
break;
|
|
}
|
|
|
|
efc_process_prli_payload(node, cbdata->payload->dma.virt);
|
|
efc_send_prli_acc(node, be16_to_cpu(hdr->fh_ox_id));
|
|
break;
|
|
}
|
|
|
|
case EFC_EVT_PRLO_RCVD: {
|
|
struct fc_frame_header *hdr = cbdata->header->dma.virt;
|
|
/* sm: / send PRLO acc */
|
|
efc_send_prlo_acc(node, be16_to_cpu(hdr->fh_ox_id));
|
|
/* need implicit logout? */
|
|
break;
|
|
}
|
|
|
|
case EFC_EVT_LOGO_RCVD: {
|
|
struct fc_frame_header *hdr = cbdata->header->dma.virt;
|
|
|
|
node_printf(node, "%s received attached=%d\n",
|
|
efc_sm_event_name(evt), node->attached);
|
|
/* sm: / send LOGO acc */
|
|
efc_send_logo_acc(node, be16_to_cpu(hdr->fh_ox_id));
|
|
efc_node_transition(node, __efc_d_wait_logo_acc_cmpl, NULL);
|
|
break;
|
|
}
|
|
|
|
case EFC_EVT_ADISC_RCVD: {
|
|
struct fc_frame_header *hdr = cbdata->header->dma.virt;
|
|
/* sm: / send ADISC acc */
|
|
efc_send_adisc_acc(node, be16_to_cpu(hdr->fh_ox_id));
|
|
break;
|
|
}
|
|
|
|
case EFC_EVT_ABTS_RCVD:
|
|
/* sm: / process ABTS */
|
|
efc_log_err(efc, "Unexpected event:%s\n",
|
|
efc_sm_event_name(evt));
|
|
break;
|
|
|
|
case EFC_EVT_NODE_ACTIVE_IO_LIST_EMPTY:
|
|
break;
|
|
|
|
case EFC_EVT_NODE_REFOUND:
|
|
break;
|
|
|
|
case EFC_EVT_NODE_MISSING:
|
|
if (node->nport->enable_rscn)
|
|
efc_node_transition(node, __efc_d_device_gone, NULL);
|
|
|
|
break;
|
|
|
|
case EFC_EVT_SRRS_ELS_CMPL_OK:
|
|
/* T, or I+T, PRLI accept completed ok */
|
|
WARN_ON(!node->els_cmpl_cnt);
|
|
node->els_cmpl_cnt--;
|
|
break;
|
|
|
|
case EFC_EVT_SRRS_ELS_CMPL_FAIL:
|
|
/* T, or I+T, PRLI accept failed to complete */
|
|
WARN_ON(!node->els_cmpl_cnt);
|
|
node->els_cmpl_cnt--;
|
|
node_printf(node, "Failed to send PRLI LS_ACC\n");
|
|
break;
|
|
|
|
default:
|
|
__efc_d_common(__func__, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
void
|
|
__efc_d_device_gone(struct efc_sm_ctx *ctx,
|
|
enum efc_sm_event evt, void *arg)
|
|
{
|
|
struct efc_node_cb *cbdata = arg;
|
|
struct efc_node *node = ctx->app;
|
|
struct efc *efc = node->efc;
|
|
|
|
efc_node_evt_set(ctx, evt, __func__);
|
|
|
|
node_sm_trace();
|
|
|
|
switch (evt) {
|
|
case EFC_EVT_ENTER: {
|
|
int rc = EFC_SCSI_CALL_COMPLETE;
|
|
int rc_2 = EFC_SCSI_CALL_COMPLETE;
|
|
static const char * const labels[] = {
|
|
"none", "initiator", "target", "initiator+target"
|
|
};
|
|
|
|
efc_log_info(efc, "[%s] missing (%s) WWPN %s WWNN %s\n",
|
|
node->display_name,
|
|
labels[(node->targ << 1) | (node->init)],
|
|
node->wwpn, node->wwnn);
|
|
|
|
switch (efc_node_get_enable(node)) {
|
|
case EFC_NODE_ENABLE_T_TO_T:
|
|
case EFC_NODE_ENABLE_I_TO_T:
|
|
case EFC_NODE_ENABLE_IT_TO_T:
|
|
rc = efc->tt.scsi_del_node(efc, node,
|
|
EFC_SCSI_TARGET_MISSING);
|
|
break;
|
|
|
|
case EFC_NODE_ENABLE_T_TO_I:
|
|
case EFC_NODE_ENABLE_I_TO_I:
|
|
case EFC_NODE_ENABLE_IT_TO_I:
|
|
rc = efc->tt.scsi_del_node(efc, node,
|
|
EFC_SCSI_INITIATOR_MISSING);
|
|
break;
|
|
|
|
case EFC_NODE_ENABLE_T_TO_IT:
|
|
rc = efc->tt.scsi_del_node(efc, node,
|
|
EFC_SCSI_INITIATOR_MISSING);
|
|
break;
|
|
|
|
case EFC_NODE_ENABLE_I_TO_IT:
|
|
rc = efc->tt.scsi_del_node(efc, node,
|
|
EFC_SCSI_TARGET_MISSING);
|
|
break;
|
|
|
|
case EFC_NODE_ENABLE_IT_TO_IT:
|
|
rc = efc->tt.scsi_del_node(efc, node,
|
|
EFC_SCSI_INITIATOR_MISSING);
|
|
rc_2 = efc->tt.scsi_del_node(efc, node,
|
|
EFC_SCSI_TARGET_MISSING);
|
|
break;
|
|
|
|
default:
|
|
rc = EFC_SCSI_CALL_COMPLETE;
|
|
break;
|
|
}
|
|
|
|
if (rc == EFC_SCSI_CALL_COMPLETE &&
|
|
rc_2 == EFC_SCSI_CALL_COMPLETE)
|
|
efc_node_post_event(node, EFC_EVT_SHUTDOWN, NULL);
|
|
|
|
break;
|
|
}
|
|
case EFC_EVT_NODE_REFOUND:
|
|
/* two approaches, reauthenticate with PLOGI/PRLI, or ADISC */
|
|
|
|
/* reauthenticate with PLOGI/PRLI */
|
|
/* efc_node_transition(node, __efc_d_discovered, NULL); */
|
|
|
|
/* reauthenticate with ADISC */
|
|
/* sm: / send ADISC */
|
|
efc_send_adisc(node);
|
|
efc_node_transition(node, __efc_d_wait_adisc_rsp, NULL);
|
|
break;
|
|
|
|
case EFC_EVT_PLOGI_RCVD: {
|
|
/* sm: / save sparams, set send_plogi_acc, post implicit
|
|
* logout
|
|
* Save plogi parameters
|
|
*/
|
|
efc_node_save_sparms(node, cbdata->payload->dma.virt);
|
|
efc_send_ls_acc_after_attach(node,
|
|
cbdata->header->dma.virt,
|
|
EFC_NODE_SEND_LS_ACC_PLOGI);
|
|
|
|
/*
|
|
* Restart node attach with new service parameters, and send
|
|
* ACC
|
|
*/
|
|
efc_node_post_event(node, EFC_EVT_SHUTDOWN_IMPLICIT_LOGO,
|
|
NULL);
|
|
break;
|
|
}
|
|
|
|
case EFC_EVT_FCP_CMD_RCVD: {
|
|
/* most likely a stale frame (received prior to link down),
|
|
* if attempt to send LOGO, will probably timeout and eat
|
|
* up 20s; thus, drop FCP_CMND
|
|
*/
|
|
node_printf(node, "FCP_CMND received, drop\n");
|
|
break;
|
|
}
|
|
case EFC_EVT_LOGO_RCVD: {
|
|
/* I, T, I+T */
|
|
struct fc_frame_header *hdr = cbdata->header->dma.virt;
|
|
|
|
node_printf(node, "%s received attached=%d\n",
|
|
efc_sm_event_name(evt), node->attached);
|
|
/* sm: / send LOGO acc */
|
|
efc_send_logo_acc(node, be16_to_cpu(hdr->fh_ox_id));
|
|
efc_node_transition(node, __efc_d_wait_logo_acc_cmpl, NULL);
|
|
break;
|
|
}
|
|
default:
|
|
__efc_d_common(__func__, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
void
|
|
__efc_d_wait_adisc_rsp(struct efc_sm_ctx *ctx,
|
|
enum efc_sm_event evt, void *arg)
|
|
{
|
|
struct efc_node_cb *cbdata = arg;
|
|
struct efc_node *node = ctx->app;
|
|
|
|
efc_node_evt_set(ctx, evt, __func__);
|
|
|
|
node_sm_trace();
|
|
|
|
switch (evt) {
|
|
case EFC_EVT_SRRS_ELS_REQ_OK:
|
|
if (efc_node_check_els_req(ctx, evt, arg, ELS_ADISC,
|
|
__efc_d_common, __func__))
|
|
return;
|
|
|
|
WARN_ON(!node->els_req_cnt);
|
|
node->els_req_cnt--;
|
|
efc_node_transition(node, __efc_d_device_ready, NULL);
|
|
break;
|
|
|
|
case EFC_EVT_SRRS_ELS_REQ_RJT:
|
|
/* received an LS_RJT, in this case, send shutdown
|
|
* (explicit logo) event which will unregister the node,
|
|
* and start over with PLOGI
|
|
*/
|
|
if (efc_node_check_els_req(ctx, evt, arg, ELS_ADISC,
|
|
__efc_d_common, __func__))
|
|
return;
|
|
|
|
WARN_ON(!node->els_req_cnt);
|
|
node->els_req_cnt--;
|
|
/* sm: / post explicit logout */
|
|
efc_node_post_event(node,
|
|
EFC_EVT_SHUTDOWN_EXPLICIT_LOGO,
|
|
NULL);
|
|
break;
|
|
|
|
case EFC_EVT_LOGO_RCVD: {
|
|
/* In this case, we have the equivalent of an LS_RJT for
|
|
* the ADISC, so we need to abort the ADISC, and re-login
|
|
* with PLOGI
|
|
*/
|
|
/* sm: / request abort, send LOGO acc */
|
|
struct fc_frame_header *hdr = cbdata->header->dma.virt;
|
|
|
|
node_printf(node, "%s received attached=%d\n",
|
|
efc_sm_event_name(evt), node->attached);
|
|
|
|
efc_send_logo_acc(node, be16_to_cpu(hdr->fh_ox_id));
|
|
efc_node_transition(node, __efc_d_wait_logo_acc_cmpl, NULL);
|
|
break;
|
|
}
|
|
default:
|
|
__efc_d_common(__func__, ctx, evt, arg);
|
|
}
|
|
}
|