// SPDX-License-Identifier: GPL-2.0 /* * Copyright (C) 2021 Broadcom. All Rights Reserved. The term * “Broadcom” refers to Broadcom Inc. and/or its subsidiaries. */ /* * device_sm Node State Machine: Remote Device States */ #include "efc.h" #include "efc_device.h" #include "efc_fabric.h" void efc_d_send_prli_rsp(struct efc_node *node, u16 ox_id) { int rc = EFC_SCSI_CALL_COMPLETE; struct efc *efc = node->efc; node->ls_acc_oxid = ox_id; node->send_ls_acc = EFC_NODE_SEND_LS_ACC_PRLI; /* * Wait for backend session registration * to complete before sending PRLI resp */ if (node->init) { efc_log_info(efc, "[%s] found(initiator) WWPN:%s WWNN:%s\n", node->display_name, node->wwpn, node->wwnn); if (node->nport->enable_tgt) rc = efc->tt.scsi_new_node(efc, node); } if (rc < 0) efc_node_post_event(node, EFC_EVT_NODE_SESS_REG_FAIL, NULL); if (rc == EFC_SCSI_CALL_COMPLETE) efc_node_post_event(node, EFC_EVT_NODE_SESS_REG_OK, NULL); } static void __efc_d_common(const char *funcname, struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = NULL; struct efc *efc = NULL; node = ctx->app; efc = node->efc; switch (evt) { /* Handle shutdown events */ case EFC_EVT_SHUTDOWN: efc_log_debug(efc, "[%s] %-20s %-20s\n", node->display_name, funcname, efc_sm_event_name(evt)); node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT; efc_node_transition(node, __efc_d_initiate_shutdown, NULL); break; case EFC_EVT_SHUTDOWN_EXPLICIT_LOGO: efc_log_debug(efc, "[%s] %-20s %-20s\n", node->display_name, funcname, efc_sm_event_name(evt)); node->shutdown_reason = EFC_NODE_SHUTDOWN_EXPLICIT_LOGO; efc_node_transition(node, __efc_d_initiate_shutdown, NULL); break; case EFC_EVT_SHUTDOWN_IMPLICIT_LOGO: efc_log_debug(efc, "[%s] %-20s %-20s\n", node->display_name, funcname, efc_sm_event_name(evt)); node->shutdown_reason = EFC_NODE_SHUTDOWN_IMPLICIT_LOGO; efc_node_transition(node, __efc_d_initiate_shutdown, NULL); break; default: /* call default event handler common to all nodes */ __efc_node_common(funcname, ctx, evt, arg); } } static void __efc_d_wait_del_node(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); /* * State is entered when a node sends a delete initiator/target call * to the target-server/initiator-client and needs to wait for that * work to complete. */ node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: efc_node_hold_frames(node); fallthrough; case EFC_EVT_NODE_ACTIVE_IO_LIST_EMPTY: case EFC_EVT_ALL_CHILD_NODES_FREE: /* These are expected events. */ break; case EFC_EVT_NODE_DEL_INI_COMPLETE: case EFC_EVT_NODE_DEL_TGT_COMPLETE: /* * node has either been detached or is in the process * of being detached, * call common node's initiate cleanup function */ efc_node_initiate_cleanup(node); break; case EFC_EVT_EXIT: efc_node_accept_frames(node); break; case EFC_EVT_SRRS_ELS_REQ_FAIL: /* Can happen as ELS IO IO's complete */ WARN_ON(!node->els_req_cnt); node->els_req_cnt--; break; /* ignore shutdown events as we're already in shutdown path */ case EFC_EVT_SHUTDOWN: /* have default shutdown event take precedence */ node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT; fallthrough; case EFC_EVT_SHUTDOWN_EXPLICIT_LOGO: case EFC_EVT_SHUTDOWN_IMPLICIT_LOGO: node_printf(node, "%s received\n", efc_sm_event_name(evt)); break; case EFC_EVT_DOMAIN_ATTACH_OK: /* don't care about domain_attach_ok */ break; default: __efc_d_common(__func__, ctx, evt, arg); } } static void __efc_d_wait_del_ini_tgt(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: efc_node_hold_frames(node); fallthrough; case EFC_EVT_NODE_ACTIVE_IO_LIST_EMPTY: case EFC_EVT_ALL_CHILD_NODES_FREE: /* These are expected events. */ break; case EFC_EVT_NODE_DEL_INI_COMPLETE: case EFC_EVT_NODE_DEL_TGT_COMPLETE: efc_node_transition(node, __efc_d_wait_del_node, NULL); break; case EFC_EVT_EXIT: efc_node_accept_frames(node); break; case EFC_EVT_SRRS_ELS_REQ_FAIL: /* Can happen as ELS IO IO's complete */ WARN_ON(!node->els_req_cnt); node->els_req_cnt--; break; /* ignore shutdown events as we're already in shutdown path */ case EFC_EVT_SHUTDOWN: /* have default shutdown event take precedence */ node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT; fallthrough; case EFC_EVT_SHUTDOWN_EXPLICIT_LOGO: case EFC_EVT_SHUTDOWN_IMPLICIT_LOGO: node_printf(node, "%s received\n", efc_sm_event_name(evt)); break; case EFC_EVT_DOMAIN_ATTACH_OK: /* don't care about domain_attach_ok */ break; default: __efc_d_common(__func__, ctx, evt, arg); } } void __efc_d_initiate_shutdown(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = ctx->app; struct efc *efc = node->efc; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: { int rc = EFC_SCSI_CALL_COMPLETE; /* assume no wait needed */ node->els_io_enabled = false; /* make necessary delete upcall(s) */ if (node->init && !node->targ) { efc_log_info(node->efc, "[%s] delete (initiator) WWPN %s WWNN %s\n", node->display_name, node->wwpn, node->wwnn); efc_node_transition(node, __efc_d_wait_del_node, NULL); if (node->nport->enable_tgt) rc = efc->tt.scsi_del_node(efc, node, EFC_SCSI_INITIATOR_DELETED); if (rc == EFC_SCSI_CALL_COMPLETE || rc < 0) efc_node_post_event(node, EFC_EVT_NODE_DEL_INI_COMPLETE, NULL); } else if (node->targ && !node->init) { efc_log_info(node->efc, "[%s] delete (target) WWPN %s WWNN %s\n", node->display_name, node->wwpn, node->wwnn); efc_node_transition(node, __efc_d_wait_del_node, NULL); if (node->nport->enable_ini) rc = efc->tt.scsi_del_node(efc, node, EFC_SCSI_TARGET_DELETED); if (rc == EFC_SCSI_CALL_COMPLETE) efc_node_post_event(node, EFC_EVT_NODE_DEL_TGT_COMPLETE, NULL); } else if (node->init && node->targ) { efc_log_info(node->efc, "[%s] delete (I+T) WWPN %s WWNN %s\n", node->display_name, node->wwpn, node->wwnn); efc_node_transition(node, __efc_d_wait_del_ini_tgt, NULL); if (node->nport->enable_tgt) rc = efc->tt.scsi_del_node(efc, node, EFC_SCSI_INITIATOR_DELETED); if (rc == EFC_SCSI_CALL_COMPLETE) efc_node_post_event(node, EFC_EVT_NODE_DEL_INI_COMPLETE, NULL); /* assume no wait needed */ rc = EFC_SCSI_CALL_COMPLETE; if (node->nport->enable_ini) rc = efc->tt.scsi_del_node(efc, node, EFC_SCSI_TARGET_DELETED); if (rc == EFC_SCSI_CALL_COMPLETE) efc_node_post_event(node, EFC_EVT_NODE_DEL_TGT_COMPLETE, NULL); } /* we've initiated the upcalls as needed, now kick off the node * detach to precipitate the aborting of outstanding exchanges * associated with said node * * Beware: if we've made upcall(s), we've already transitioned * to a new state by the time we execute this. * consider doing this before the upcalls? */ if (node->attached) { /* issue hw node free; don't care if succeeds right * away or sometime later, will check node->attached * later in shutdown process */ rc = efc_cmd_node_detach(efc, &node->rnode); if (rc < 0) node_printf(node, "Failed freeing HW node, rc=%d\n", rc); } /* if neither initiator nor target, proceed to cleanup */ if (!node->init && !node->targ) { /* * node has either been detached or is in * the process of being detached, * call common node's initiate cleanup function */ efc_node_initiate_cleanup(node); } break; } case EFC_EVT_ALL_CHILD_NODES_FREE: /* Ignore, this can happen if an ELS is * aborted while in a delay/retry state */ break; default: __efc_d_common(__func__, ctx, evt, arg); } } void __efc_d_wait_loop(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: efc_node_hold_frames(node); break; case EFC_EVT_EXIT: efc_node_accept_frames(node); break; case EFC_EVT_DOMAIN_ATTACH_OK: { /* send PLOGI automatically if initiator */ efc_node_init_device(node, true); break; } default: __efc_d_common(__func__, ctx, evt, arg); } } void efc_send_ls_acc_after_attach(struct efc_node *node, struct fc_frame_header *hdr, enum efc_node_send_ls_acc ls) { u16 ox_id = be16_to_cpu(hdr->fh_ox_id); /* Save the OX_ID for sending LS_ACC sometime later */ WARN_ON(node->send_ls_acc != EFC_NODE_SEND_LS_ACC_NONE); node->ls_acc_oxid = ox_id; node->send_ls_acc = ls; node->ls_acc_did = ntoh24(hdr->fh_d_id); } void efc_process_prli_payload(struct efc_node *node, void *prli) { struct { struct fc_els_prli prli; struct fc_els_spp sp; } *pp; pp = prli; node->init = (pp->sp.spp_flags & FCP_SPPF_INIT_FCN) != 0; node->targ = (pp->sp.spp_flags & FCP_SPPF_TARG_FCN) != 0; } void __efc_d_wait_plogi_acc_cmpl(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: efc_node_hold_frames(node); break; case EFC_EVT_EXIT: efc_node_accept_frames(node); break; case EFC_EVT_SRRS_ELS_CMPL_FAIL: WARN_ON(!node->els_cmpl_cnt); node->els_cmpl_cnt--; node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT; efc_node_transition(node, __efc_d_initiate_shutdown, NULL); break; case EFC_EVT_SRRS_ELS_CMPL_OK: /* PLOGI ACC completions */ WARN_ON(!node->els_cmpl_cnt); node->els_cmpl_cnt--; efc_node_transition(node, __efc_d_port_logged_in, NULL); break; default: __efc_d_common(__func__, ctx, evt, arg); } } void __efc_d_wait_logo_rsp(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: efc_node_hold_frames(node); break; case EFC_EVT_EXIT: efc_node_accept_frames(node); break; case EFC_EVT_SRRS_ELS_REQ_OK: case EFC_EVT_SRRS_ELS_REQ_RJT: case EFC_EVT_SRRS_ELS_REQ_FAIL: /* LOGO response received, sent shutdown */ if (efc_node_check_els_req(ctx, evt, arg, ELS_LOGO, __efc_d_common, __func__)) return; WARN_ON(!node->els_req_cnt); node->els_req_cnt--; node_printf(node, "LOGO sent (evt=%s), shutdown node\n", efc_sm_event_name(evt)); /* sm: / post explicit logout */ efc_node_post_event(node, EFC_EVT_SHUTDOWN_EXPLICIT_LOGO, NULL); break; default: __efc_d_common(__func__, ctx, evt, arg); } } void efc_node_init_device(struct efc_node *node, bool send_plogi) { node->send_plogi = send_plogi; if ((node->efc->nodedb_mask & EFC_NODEDB_PAUSE_NEW_NODES) && (node->rnode.fc_id != FC_FID_DOM_MGR)) { node->nodedb_state = __efc_d_init; efc_node_transition(node, __efc_node_paused, NULL); } else { efc_node_transition(node, __efc_d_init, NULL); } } static void efc_d_check_plogi_topology(struct efc_node *node, u32 d_id) { switch (node->nport->topology) { case EFC_NPORT_TOPO_P2P: /* we're not attached and nport is p2p, * need to attach */ efc_domain_attach(node->nport->domain, d_id); efc_node_transition(node, __efc_d_wait_domain_attach, NULL); break; case EFC_NPORT_TOPO_FABRIC: /* we're not attached and nport is fabric, domain * attach should have already been requested as part * of the fabric state machine, wait for it */ efc_node_transition(node, __efc_d_wait_domain_attach, NULL); break; case EFC_NPORT_TOPO_UNKNOWN: /* Two possibilities: * 1. received a PLOGI before our FLOGI has completed * (possible since completion comes in on another * CQ), thus we don't know what we're connected to * yet; transition to a state to wait for the * fabric node to tell us; * 2. PLOGI received before link went down and we * haven't performed domain attach yet. * Note: we cannot distinguish between 1. and 2. * so have to assume PLOGI * was received after link back up. */ node_printf(node, "received PLOGI, unknown topology did=0x%x\n", d_id); efc_node_transition(node, __efc_d_wait_topology_notify, NULL); break; default: node_printf(node, "received PLOGI, unexpected topology %d\n", node->nport->topology); } } void __efc_d_init(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node_cb *cbdata = arg; struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); /* * This state is entered when a node is instantiated, * either having been discovered from a name services query, * or having received a PLOGI/FLOGI. */ switch (evt) { case EFC_EVT_ENTER: if (!node->send_plogi) break; /* only send if we have initiator capability, * and domain is attached */ if (node->nport->enable_ini && node->nport->domain->attached) { efc_send_plogi(node); efc_node_transition(node, __efc_d_wait_plogi_rsp, NULL); } else { node_printf(node, "not sending plogi nport.ini=%d,", node->nport->enable_ini); node_printf(node, "domain attached=%d\n", node->nport->domain->attached); } break; case EFC_EVT_PLOGI_RCVD: { /* T, or I+T */ struct fc_frame_header *hdr = cbdata->header->dma.virt; int rc; efc_node_save_sparms(node, cbdata->payload->dma.virt); efc_send_ls_acc_after_attach(node, cbdata->header->dma.virt, EFC_NODE_SEND_LS_ACC_PLOGI); /* domain not attached; several possibilities: */ if (!node->nport->domain->attached) { efc_d_check_plogi_topology(node, ntoh24(hdr->fh_d_id)); break; } /* domain already attached */ rc = efc_node_attach(node); efc_node_transition(node, __efc_d_wait_node_attach, NULL); if (rc < 0) efc_node_post_event(node, EFC_EVT_NODE_ATTACH_FAIL, NULL); break; } case EFC_EVT_FDISC_RCVD: { __efc_d_common(__func__, ctx, evt, arg); break; } case EFC_EVT_FLOGI_RCVD: { struct fc_frame_header *hdr = cbdata->header->dma.virt; u32 d_id = ntoh24(hdr->fh_d_id); /* sm: / save sparams, send FLOGI acc */ memcpy(node->nport->domain->flogi_service_params, cbdata->payload->dma.virt, sizeof(struct fc_els_flogi)); /* send FC LS_ACC response, override s_id */ efc_fabric_set_topology(node, EFC_NPORT_TOPO_P2P); efc_send_flogi_p2p_acc(node, be16_to_cpu(hdr->fh_ox_id), d_id); if (efc_p2p_setup(node->nport)) { node_printf(node, "p2p failed, shutting down node\n"); efc_node_post_event(node, EFC_EVT_SHUTDOWN, NULL); break; } efc_node_transition(node, __efc_p2p_wait_flogi_acc_cmpl, NULL); break; } case EFC_EVT_LOGO_RCVD: { struct fc_frame_header *hdr = cbdata->header->dma.virt; if (!node->nport->domain->attached) { /* most likely a frame left over from before a link * down; drop and * shut node down w/ "explicit logout" so pending * frames are processed */ node_printf(node, "%s domain not attached, dropping\n", efc_sm_event_name(evt)); efc_node_post_event(node, EFC_EVT_SHUTDOWN_EXPLICIT_LOGO, NULL); break; } efc_send_logo_acc(node, be16_to_cpu(hdr->fh_ox_id)); efc_node_transition(node, __efc_d_wait_logo_acc_cmpl, NULL); break; } case EFC_EVT_PRLI_RCVD: case EFC_EVT_PRLO_RCVD: case EFC_EVT_PDISC_RCVD: case EFC_EVT_ADISC_RCVD: case EFC_EVT_RSCN_RCVD: { struct fc_frame_header *hdr = cbdata->header->dma.virt; if (!node->nport->domain->attached) { /* most likely a frame left over from before a link * down; drop and shut node down w/ "explicit logout" * so pending frames are processed */ node_printf(node, "%s domain not attached, dropping\n", efc_sm_event_name(evt)); efc_node_post_event(node, EFC_EVT_SHUTDOWN_EXPLICIT_LOGO, NULL); break; } node_printf(node, "%s received, sending reject\n", efc_sm_event_name(evt)); efc_send_ls_rjt(node, be16_to_cpu(hdr->fh_ox_id), ELS_RJT_UNAB, ELS_EXPL_PLOGI_REQD, 0); break; } case EFC_EVT_FCP_CMD_RCVD: { /* note: problem, we're now expecting an ELS REQ completion * from both the LOGO and PLOGI */ if (!node->nport->domain->attached) { /* most likely a frame left over from before a * link down; drop and * shut node down w/ "explicit logout" so pending * frames are processed */ node_printf(node, "%s domain not attached, dropping\n", efc_sm_event_name(evt)); efc_node_post_event(node, EFC_EVT_SHUTDOWN_EXPLICIT_LOGO, NULL); break; } /* Send LOGO */ node_printf(node, "FCP_CMND received, send LOGO\n"); if (efc_send_logo(node)) { /* * failed to send LOGO, go ahead and cleanup node * anyways */ node_printf(node, "Failed to send LOGO\n"); efc_node_post_event(node, EFC_EVT_SHUTDOWN_EXPLICIT_LOGO, NULL); } else { /* sent LOGO, wait for response */ efc_node_transition(node, __efc_d_wait_logo_rsp, NULL); } break; } case EFC_EVT_DOMAIN_ATTACH_OK: /* don't care about domain_attach_ok */ break; default: __efc_d_common(__func__, ctx, evt, arg); } } void __efc_d_wait_plogi_rsp(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { int rc; struct efc_node_cb *cbdata = arg; struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_PLOGI_RCVD: { /* T, or I+T */ /* received PLOGI with svc parms, go ahead and attach node * when PLOGI that was sent ultimately completes, it'll be a * no-op * * If there is an outstanding PLOGI sent, can we set a flag * to indicate that we don't want to retry it if it times out? */ efc_node_save_sparms(node, cbdata->payload->dma.virt); efc_send_ls_acc_after_attach(node, cbdata->header->dma.virt, EFC_NODE_SEND_LS_ACC_PLOGI); /* sm: domain->attached / efc_node_attach */ rc = efc_node_attach(node); efc_node_transition(node, __efc_d_wait_node_attach, NULL); if (rc < 0) efc_node_post_event(node, EFC_EVT_NODE_ATTACH_FAIL, NULL); break; } case EFC_EVT_PRLI_RCVD: /* I, or I+T */ /* sent PLOGI and before completion was seen, received the * PRLI from the remote node (WCQEs and RCQEs come in on * different queues and order of processing cannot be assumed) * Save OXID so PRLI can be sent after the attach and continue * to wait for PLOGI response */ efc_process_prli_payload(node, cbdata->payload->dma.virt); efc_send_ls_acc_after_attach(node, cbdata->header->dma.virt, EFC_NODE_SEND_LS_ACC_PRLI); efc_node_transition(node, __efc_d_wait_plogi_rsp_recvd_prli, NULL); break; case EFC_EVT_LOGO_RCVD: /* why don't we do a shutdown here?? */ case EFC_EVT_PRLO_RCVD: case EFC_EVT_PDISC_RCVD: case EFC_EVT_FDISC_RCVD: case EFC_EVT_ADISC_RCVD: case EFC_EVT_RSCN_RCVD: case EFC_EVT_SCR_RCVD: { struct fc_frame_header *hdr = cbdata->header->dma.virt; node_printf(node, "%s received, sending reject\n", efc_sm_event_name(evt)); efc_send_ls_rjt(node, be16_to_cpu(hdr->fh_ox_id), ELS_RJT_UNAB, ELS_EXPL_PLOGI_REQD, 0); break; } case EFC_EVT_SRRS_ELS_REQ_OK: /* PLOGI response received */ /* Completion from PLOGI sent */ if (efc_node_check_els_req(ctx, evt, arg, ELS_PLOGI, __efc_d_common, __func__)) return; WARN_ON(!node->els_req_cnt); node->els_req_cnt--; /* sm: / save sparams, efc_node_attach */ efc_node_save_sparms(node, cbdata->els_rsp.virt); rc = efc_node_attach(node); efc_node_transition(node, __efc_d_wait_node_attach, NULL); if (rc < 0) efc_node_post_event(node, EFC_EVT_NODE_ATTACH_FAIL, NULL); break; case EFC_EVT_SRRS_ELS_REQ_FAIL: /* PLOGI response received */ /* PLOGI failed, shutdown the node */ if (efc_node_check_els_req(ctx, evt, arg, ELS_PLOGI, __efc_d_common, __func__)) return; WARN_ON(!node->els_req_cnt); node->els_req_cnt--; efc_node_post_event(node, EFC_EVT_SHUTDOWN, NULL); break; case EFC_EVT_SRRS_ELS_REQ_RJT: /* Our PLOGI was rejected, this is ok in some cases */ if (efc_node_check_els_req(ctx, evt, arg, ELS_PLOGI, __efc_d_common, __func__)) return; WARN_ON(!node->els_req_cnt); node->els_req_cnt--; break; case EFC_EVT_FCP_CMD_RCVD: { /* not logged in yet and outstanding PLOGI so don't send LOGO, * just drop */ node_printf(node, "FCP_CMND received, drop\n"); break; } default: __efc_d_common(__func__, ctx, evt, arg); } } void __efc_d_wait_plogi_rsp_recvd_prli(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { int rc; struct efc_node_cb *cbdata = arg; struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: /* * Since we've received a PRLI, we have a port login and will * just need to wait for the PLOGI response to do the node * attach and then we can send the LS_ACC for the PRLI. If, * during this time, we receive FCP_CMNDs (which is possible * since we've already sent a PRLI and our peer may have * accepted). At this time, we are not waiting on any other * unsolicited frames to continue with the login process. Thus, * it will not hurt to hold frames here. */ efc_node_hold_frames(node); break; case EFC_EVT_EXIT: efc_node_accept_frames(node); break; case EFC_EVT_SRRS_ELS_REQ_OK: /* PLOGI response received */ /* Completion from PLOGI sent */ if (efc_node_check_els_req(ctx, evt, arg, ELS_PLOGI, __efc_d_common, __func__)) return; WARN_ON(!node->els_req_cnt); node->els_req_cnt--; /* sm: / save sparams, efc_node_attach */ efc_node_save_sparms(node, cbdata->els_rsp.virt); rc = efc_node_attach(node); efc_node_transition(node, __efc_d_wait_node_attach, NULL); if (rc < 0) efc_node_post_event(node, EFC_EVT_NODE_ATTACH_FAIL, NULL); break; case EFC_EVT_SRRS_ELS_REQ_FAIL: /* PLOGI response received */ case EFC_EVT_SRRS_ELS_REQ_RJT: /* PLOGI failed, shutdown the node */ if (efc_node_check_els_req(ctx, evt, arg, ELS_PLOGI, __efc_d_common, __func__)) return; WARN_ON(!node->els_req_cnt); node->els_req_cnt--; efc_node_post_event(node, EFC_EVT_SHUTDOWN, NULL); break; default: __efc_d_common(__func__, ctx, evt, arg); } } void __efc_d_wait_domain_attach(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { int rc; struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: efc_node_hold_frames(node); break; case EFC_EVT_EXIT: efc_node_accept_frames(node); break; case EFC_EVT_DOMAIN_ATTACH_OK: WARN_ON(!node->nport->domain->attached); /* sm: / efc_node_attach */ rc = efc_node_attach(node); efc_node_transition(node, __efc_d_wait_node_attach, NULL); if (rc < 0) efc_node_post_event(node, EFC_EVT_NODE_ATTACH_FAIL, NULL); break; default: __efc_d_common(__func__, ctx, evt, arg); } } void __efc_d_wait_topology_notify(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { int rc; struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: efc_node_hold_frames(node); break; case EFC_EVT_EXIT: efc_node_accept_frames(node); break; case EFC_EVT_NPORT_TOPOLOGY_NOTIFY: { enum efc_nport_topology *topology = arg; WARN_ON(node->nport->domain->attached); WARN_ON(node->send_ls_acc != EFC_NODE_SEND_LS_ACC_PLOGI); node_printf(node, "topology notification, topology=%d\n", *topology); /* At the time the PLOGI was received, the topology was unknown, * so we didn't know which node would perform the domain attach: * 1. The node from which the PLOGI was sent (p2p) or * 2. The node to which the FLOGI was sent (fabric). */ if (*topology == EFC_NPORT_TOPO_P2P) { /* if this is p2p, need to attach to the domain using * the d_id from the PLOGI received */ efc_domain_attach(node->nport->domain, node->ls_acc_did); } /* else, if this is fabric, the domain attach * should be performed by the fabric node (node sending FLOGI); * just wait for attach to complete */ efc_node_transition(node, __efc_d_wait_domain_attach, NULL); break; } case EFC_EVT_DOMAIN_ATTACH_OK: WARN_ON(!node->nport->domain->attached); node_printf(node, "domain attach ok\n"); /* sm: / efc_node_attach */ rc = efc_node_attach(node); efc_node_transition(node, __efc_d_wait_node_attach, NULL); if (rc < 0) efc_node_post_event(node, EFC_EVT_NODE_ATTACH_FAIL, NULL); break; default: __efc_d_common(__func__, ctx, evt, arg); } } void __efc_d_wait_node_attach(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: efc_node_hold_frames(node); break; case EFC_EVT_EXIT: efc_node_accept_frames(node); break; case EFC_EVT_NODE_ATTACH_OK: node->attached = true; switch (node->send_ls_acc) { case EFC_NODE_SEND_LS_ACC_PLOGI: { /* sm: send_plogi_acc is set / send PLOGI acc */ /* Normal case for T, or I+T */ efc_send_plogi_acc(node, node->ls_acc_oxid); efc_node_transition(node, __efc_d_wait_plogi_acc_cmpl, NULL); node->send_ls_acc = EFC_NODE_SEND_LS_ACC_NONE; node->ls_acc_io = NULL; break; } case EFC_NODE_SEND_LS_ACC_PRLI: { efc_d_send_prli_rsp(node, node->ls_acc_oxid); node->send_ls_acc = EFC_NODE_SEND_LS_ACC_NONE; node->ls_acc_io = NULL; break; } case EFC_NODE_SEND_LS_ACC_NONE: default: /* Normal case for I */ /* sm: send_plogi_acc is not set / send PLOGI acc */ efc_node_transition(node, __efc_d_port_logged_in, NULL); break; } break; case EFC_EVT_NODE_ATTACH_FAIL: /* node attach failed, shutdown the node */ node->attached = false; node_printf(node, "node attach failed\n"); node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT; efc_node_transition(node, __efc_d_initiate_shutdown, NULL); break; /* Handle shutdown events */ case EFC_EVT_SHUTDOWN: node_printf(node, "%s received\n", efc_sm_event_name(evt)); node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT; efc_node_transition(node, __efc_d_wait_attach_evt_shutdown, NULL); break; case EFC_EVT_SHUTDOWN_EXPLICIT_LOGO: node_printf(node, "%s received\n", efc_sm_event_name(evt)); node->shutdown_reason = EFC_NODE_SHUTDOWN_EXPLICIT_LOGO; efc_node_transition(node, __efc_d_wait_attach_evt_shutdown, NULL); break; case EFC_EVT_SHUTDOWN_IMPLICIT_LOGO: node_printf(node, "%s received\n", efc_sm_event_name(evt)); node->shutdown_reason = EFC_NODE_SHUTDOWN_IMPLICIT_LOGO; efc_node_transition(node, __efc_d_wait_attach_evt_shutdown, NULL); break; default: __efc_d_common(__func__, ctx, evt, arg); } } void __efc_d_wait_attach_evt_shutdown(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: efc_node_hold_frames(node); break; case EFC_EVT_EXIT: efc_node_accept_frames(node); break; /* wait for any of these attach events and then shutdown */ case EFC_EVT_NODE_ATTACH_OK: node->attached = true; node_printf(node, "Attach evt=%s, proceed to shutdown\n", efc_sm_event_name(evt)); efc_node_transition(node, __efc_d_initiate_shutdown, NULL); break; case EFC_EVT_NODE_ATTACH_FAIL: /* node attach failed, shutdown the node */ node->attached = false; node_printf(node, "Attach evt=%s, proceed to shutdown\n", efc_sm_event_name(evt)); efc_node_transition(node, __efc_d_initiate_shutdown, NULL); break; /* ignore shutdown events as we're already in shutdown path */ case EFC_EVT_SHUTDOWN: /* have default shutdown event take precedence */ node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT; fallthrough; case EFC_EVT_SHUTDOWN_EXPLICIT_LOGO: case EFC_EVT_SHUTDOWN_IMPLICIT_LOGO: node_printf(node, "%s received\n", efc_sm_event_name(evt)); break; default: __efc_d_common(__func__, ctx, evt, arg); } } void __efc_d_port_logged_in(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node_cb *cbdata = arg; struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: /* Normal case for I or I+T */ if (node->nport->enable_ini && !(node->rnode.fc_id != FC_FID_DOM_MGR)) { /* sm: if enable_ini / send PRLI */ efc_send_prli(node); /* can now expect ELS_REQ_OK/FAIL/RJT */ } break; case EFC_EVT_FCP_CMD_RCVD: { break; } case EFC_EVT_PRLI_RCVD: { /* Normal case for T or I+T */ struct fc_frame_header *hdr = cbdata->header->dma.virt; struct { struct fc_els_prli prli; struct fc_els_spp sp; } *pp; pp = cbdata->payload->dma.virt; if (pp->sp.spp_type != FC_TYPE_FCP) { /*Only FCP is supported*/ efc_send_ls_rjt(node, be16_to_cpu(hdr->fh_ox_id), ELS_RJT_UNAB, ELS_EXPL_UNSUPR, 0); break; } efc_process_prli_payload(node, cbdata->payload->dma.virt); efc_d_send_prli_rsp(node, be16_to_cpu(hdr->fh_ox_id)); break; } case EFC_EVT_NODE_SESS_REG_OK: if (node->send_ls_acc == EFC_NODE_SEND_LS_ACC_PRLI) efc_send_prli_acc(node, node->ls_acc_oxid); node->send_ls_acc = EFC_NODE_SEND_LS_ACC_NONE; efc_node_transition(node, __efc_d_device_ready, NULL); break; case EFC_EVT_NODE_SESS_REG_FAIL: efc_send_ls_rjt(node, node->ls_acc_oxid, ELS_RJT_UNAB, ELS_EXPL_UNSUPR, 0); node->send_ls_acc = EFC_NODE_SEND_LS_ACC_NONE; break; case EFC_EVT_SRRS_ELS_REQ_OK: { /* PRLI response */ /* Normal case for I or I+T */ if (efc_node_check_els_req(ctx, evt, arg, ELS_PRLI, __efc_d_common, __func__)) return; WARN_ON(!node->els_req_cnt); node->els_req_cnt--; /* sm: / process PRLI payload */ efc_process_prli_payload(node, cbdata->els_rsp.virt); efc_node_transition(node, __efc_d_device_ready, NULL); break; } case EFC_EVT_SRRS_ELS_REQ_FAIL: { /* PRLI response failed */ /* I, I+T, assume some link failure, shutdown node */ if (efc_node_check_els_req(ctx, evt, arg, ELS_PRLI, __efc_d_common, __func__)) return; WARN_ON(!node->els_req_cnt); node->els_req_cnt--; efc_node_post_event(node, EFC_EVT_SHUTDOWN, NULL); break; } case EFC_EVT_SRRS_ELS_REQ_RJT: { /* PRLI rejected by remote * Normal for I, I+T (connected to an I) * Node doesn't want to be a target, stay here and wait for a * PRLI from the remote node * if it really wants to connect to us as target */ if (efc_node_check_els_req(ctx, evt, arg, ELS_PRLI, __efc_d_common, __func__)) return; WARN_ON(!node->els_req_cnt); node->els_req_cnt--; break; } case EFC_EVT_SRRS_ELS_CMPL_OK: { /* Normal T, I+T, target-server rejected the process login */ /* This would be received only in the case where we sent * LS_RJT for the PRLI, so * do nothing. (note: as T only we could shutdown the node) */ WARN_ON(!node->els_cmpl_cnt); node->els_cmpl_cnt--; break; } case EFC_EVT_PLOGI_RCVD: { /*sm: / save sparams, set send_plogi_acc, *post implicit logout * Save plogi parameters */ efc_node_save_sparms(node, cbdata->payload->dma.virt); efc_send_ls_acc_after_attach(node, cbdata->header->dma.virt, EFC_NODE_SEND_LS_ACC_PLOGI); /* Restart node attach with new service parameters, * and send ACC */ efc_node_post_event(node, EFC_EVT_SHUTDOWN_IMPLICIT_LOGO, NULL); break; } case EFC_EVT_LOGO_RCVD: { /* I, T, I+T */ struct fc_frame_header *hdr = cbdata->header->dma.virt; node_printf(node, "%s received attached=%d\n", efc_sm_event_name(evt), node->attached); /* sm: / send LOGO acc */ efc_send_logo_acc(node, be16_to_cpu(hdr->fh_ox_id)); efc_node_transition(node, __efc_d_wait_logo_acc_cmpl, NULL); break; } default: __efc_d_common(__func__, ctx, evt, arg); } } void __efc_d_wait_logo_acc_cmpl(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: efc_node_hold_frames(node); break; case EFC_EVT_EXIT: efc_node_accept_frames(node); break; case EFC_EVT_SRRS_ELS_CMPL_OK: case EFC_EVT_SRRS_ELS_CMPL_FAIL: /* sm: / post explicit logout */ WARN_ON(!node->els_cmpl_cnt); node->els_cmpl_cnt--; efc_node_post_event(node, EFC_EVT_SHUTDOWN_EXPLICIT_LOGO, NULL); break; default: __efc_d_common(__func__, ctx, evt, arg); } } void __efc_d_device_ready(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node_cb *cbdata = arg; struct efc_node *node = ctx->app; struct efc *efc = node->efc; efc_node_evt_set(ctx, evt, __func__); if (evt != EFC_EVT_FCP_CMD_RCVD) node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: node->fcp_enabled = true; if (node->targ) { efc_log_info(efc, "[%s] found (target) WWPN %s WWNN %s\n", node->display_name, node->wwpn, node->wwnn); if (node->nport->enable_ini) efc->tt.scsi_new_node(efc, node); } break; case EFC_EVT_EXIT: node->fcp_enabled = false; break; case EFC_EVT_PLOGI_RCVD: { /* sm: / save sparams, set send_plogi_acc, post implicit * logout * Save plogi parameters */ efc_node_save_sparms(node, cbdata->payload->dma.virt); efc_send_ls_acc_after_attach(node, cbdata->header->dma.virt, EFC_NODE_SEND_LS_ACC_PLOGI); /* * Restart node attach with new service parameters, * and send ACC */ efc_node_post_event(node, EFC_EVT_SHUTDOWN_IMPLICIT_LOGO, NULL); break; } case EFC_EVT_PRLI_RCVD: { /* T, I+T: remote initiator is slow to get started */ struct fc_frame_header *hdr = cbdata->header->dma.virt; struct { struct fc_els_prli prli; struct fc_els_spp sp; } *pp; pp = cbdata->payload->dma.virt; if (pp->sp.spp_type != FC_TYPE_FCP) { /*Only FCP is supported*/ efc_send_ls_rjt(node, be16_to_cpu(hdr->fh_ox_id), ELS_RJT_UNAB, ELS_EXPL_UNSUPR, 0); break; } efc_process_prli_payload(node, cbdata->payload->dma.virt); efc_send_prli_acc(node, be16_to_cpu(hdr->fh_ox_id)); break; } case EFC_EVT_PRLO_RCVD: { struct fc_frame_header *hdr = cbdata->header->dma.virt; /* sm: / send PRLO acc */ efc_send_prlo_acc(node, be16_to_cpu(hdr->fh_ox_id)); /* need implicit logout? */ break; } case EFC_EVT_LOGO_RCVD: { struct fc_frame_header *hdr = cbdata->header->dma.virt; node_printf(node, "%s received attached=%d\n", efc_sm_event_name(evt), node->attached); /* sm: / send LOGO acc */ efc_send_logo_acc(node, be16_to_cpu(hdr->fh_ox_id)); efc_node_transition(node, __efc_d_wait_logo_acc_cmpl, NULL); break; } case EFC_EVT_ADISC_RCVD: { struct fc_frame_header *hdr = cbdata->header->dma.virt; /* sm: / send ADISC acc */ efc_send_adisc_acc(node, be16_to_cpu(hdr->fh_ox_id)); break; } case EFC_EVT_ABTS_RCVD: /* sm: / process ABTS */ efc_log_err(efc, "Unexpected event:%s\n", efc_sm_event_name(evt)); break; case EFC_EVT_NODE_ACTIVE_IO_LIST_EMPTY: break; case EFC_EVT_NODE_REFOUND: break; case EFC_EVT_NODE_MISSING: if (node->nport->enable_rscn) efc_node_transition(node, __efc_d_device_gone, NULL); break; case EFC_EVT_SRRS_ELS_CMPL_OK: /* T, or I+T, PRLI accept completed ok */ WARN_ON(!node->els_cmpl_cnt); node->els_cmpl_cnt--; break; case EFC_EVT_SRRS_ELS_CMPL_FAIL: /* T, or I+T, PRLI accept failed to complete */ WARN_ON(!node->els_cmpl_cnt); node->els_cmpl_cnt--; node_printf(node, "Failed to send PRLI LS_ACC\n"); break; default: __efc_d_common(__func__, ctx, evt, arg); } } void __efc_d_device_gone(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node_cb *cbdata = arg; struct efc_node *node = ctx->app; struct efc *efc = node->efc; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: { int rc = EFC_SCSI_CALL_COMPLETE; int rc_2 = EFC_SCSI_CALL_COMPLETE; static const char * const labels[] = { "none", "initiator", "target", "initiator+target" }; efc_log_info(efc, "[%s] missing (%s) WWPN %s WWNN %s\n", node->display_name, labels[(node->targ << 1) | (node->init)], node->wwpn, node->wwnn); switch (efc_node_get_enable(node)) { case EFC_NODE_ENABLE_T_TO_T: case EFC_NODE_ENABLE_I_TO_T: case EFC_NODE_ENABLE_IT_TO_T: rc = efc->tt.scsi_del_node(efc, node, EFC_SCSI_TARGET_MISSING); break; case EFC_NODE_ENABLE_T_TO_I: case EFC_NODE_ENABLE_I_TO_I: case EFC_NODE_ENABLE_IT_TO_I: rc = efc->tt.scsi_del_node(efc, node, EFC_SCSI_INITIATOR_MISSING); break; case EFC_NODE_ENABLE_T_TO_IT: rc = efc->tt.scsi_del_node(efc, node, EFC_SCSI_INITIATOR_MISSING); break; case EFC_NODE_ENABLE_I_TO_IT: rc = efc->tt.scsi_del_node(efc, node, EFC_SCSI_TARGET_MISSING); break; case EFC_NODE_ENABLE_IT_TO_IT: rc = efc->tt.scsi_del_node(efc, node, EFC_SCSI_INITIATOR_MISSING); rc_2 = efc->tt.scsi_del_node(efc, node, EFC_SCSI_TARGET_MISSING); break; default: rc = EFC_SCSI_CALL_COMPLETE; break; } if (rc == EFC_SCSI_CALL_COMPLETE && rc_2 == EFC_SCSI_CALL_COMPLETE) efc_node_post_event(node, EFC_EVT_SHUTDOWN, NULL); break; } case EFC_EVT_NODE_REFOUND: /* two approaches, reauthenticate with PLOGI/PRLI, or ADISC */ /* reauthenticate with PLOGI/PRLI */ /* efc_node_transition(node, __efc_d_discovered, NULL); */ /* reauthenticate with ADISC */ /* sm: / send ADISC */ efc_send_adisc(node); efc_node_transition(node, __efc_d_wait_adisc_rsp, NULL); break; case EFC_EVT_PLOGI_RCVD: { /* sm: / save sparams, set send_plogi_acc, post implicit * logout * Save plogi parameters */ efc_node_save_sparms(node, cbdata->payload->dma.virt); efc_send_ls_acc_after_attach(node, cbdata->header->dma.virt, EFC_NODE_SEND_LS_ACC_PLOGI); /* * Restart node attach with new service parameters, and send * ACC */ efc_node_post_event(node, EFC_EVT_SHUTDOWN_IMPLICIT_LOGO, NULL); break; } case EFC_EVT_FCP_CMD_RCVD: { /* most likely a stale frame (received prior to link down), * if attempt to send LOGO, will probably timeout and eat * up 20s; thus, drop FCP_CMND */ node_printf(node, "FCP_CMND received, drop\n"); break; } case EFC_EVT_LOGO_RCVD: { /* I, T, I+T */ struct fc_frame_header *hdr = cbdata->header->dma.virt; node_printf(node, "%s received attached=%d\n", efc_sm_event_name(evt), node->attached); /* sm: / send LOGO acc */ efc_send_logo_acc(node, be16_to_cpu(hdr->fh_ox_id)); efc_node_transition(node, __efc_d_wait_logo_acc_cmpl, NULL); break; } default: __efc_d_common(__func__, ctx, evt, arg); } } void __efc_d_wait_adisc_rsp(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node_cb *cbdata = arg; struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_SRRS_ELS_REQ_OK: if (efc_node_check_els_req(ctx, evt, arg, ELS_ADISC, __efc_d_common, __func__)) return; WARN_ON(!node->els_req_cnt); node->els_req_cnt--; efc_node_transition(node, __efc_d_device_ready, NULL); break; case EFC_EVT_SRRS_ELS_REQ_RJT: /* received an LS_RJT, in this case, send shutdown * (explicit logo) event which will unregister the node, * and start over with PLOGI */ if (efc_node_check_els_req(ctx, evt, arg, ELS_ADISC, __efc_d_common, __func__)) return; WARN_ON(!node->els_req_cnt); node->els_req_cnt--; /* sm: / post explicit logout */ efc_node_post_event(node, EFC_EVT_SHUTDOWN_EXPLICIT_LOGO, NULL); break; case EFC_EVT_LOGO_RCVD: { /* In this case, we have the equivalent of an LS_RJT for * the ADISC, so we need to abort the ADISC, and re-login * with PLOGI */ /* sm: / request abort, send LOGO acc */ struct fc_frame_header *hdr = cbdata->header->dma.virt; node_printf(node, "%s received attached=%d\n", efc_sm_event_name(evt), node->attached); efc_send_logo_acc(node, be16_to_cpu(hdr->fh_ox_id)); efc_node_transition(node, __efc_d_wait_logo_acc_cmpl, NULL); break; } default: __efc_d_common(__func__, ctx, evt, arg); } }