// SPDX-License-Identifier: GPL-2.0-only /* * Copyright 2008 Cisco Systems, Inc. All rights reserved. * Copyright 2007 Nuova Systems, Inc. All rights reserved. */ #include #include "fnic.h" #include "fdls_fc.h" #include "fnic_fdls.h" #include #include #include #define FC_FC4_TYPE_SCSI 0x08 #define PORT_SPEED_BIT_8 8 #define PORT_SPEED_BIT_9 9 #define PORT_SPEED_BIT_14 14 #define PORT_SPEED_BIT_15 15 /* FNIC FDMI Register HBA Macros */ #define FNIC_FDMI_NUM_PORTS 1 #define FNIC_FDMI_NUM_HBA_ATTRS 9 #define FNIC_FDMI_TYPE_NODE_NAME 0X1 #define FNIC_FDMI_TYPE_MANUFACTURER 0X2 #define FNIC_FDMI_MANUFACTURER "Cisco Systems" #define FNIC_FDMI_TYPE_SERIAL_NUMBER 0X3 #define FNIC_FDMI_TYPE_MODEL 0X4 #define FNIC_FDMI_TYPE_MODEL_DES 0X5 #define FNIC_FDMI_MODEL_DESCRIPTION "Cisco Virtual Interface Card" #define FNIC_FDMI_TYPE_HARDWARE_VERSION 0X6 #define FNIC_FDMI_TYPE_DRIVER_VERSION 0X7 #define FNIC_FDMI_TYPE_ROM_VERSION 0X8 #define FNIC_FDMI_TYPE_FIRMWARE_VERSION 0X9 #define FNIC_FDMI_NN_LEN 8 #define FNIC_FDMI_MANU_LEN 20 #define FNIC_FDMI_SERIAL_LEN 16 #define FNIC_FDMI_MODEL_LEN 12 #define FNIC_FDMI_MODEL_DES_LEN 56 #define FNIC_FDMI_HW_VER_LEN 16 #define FNIC_FDMI_DR_VER_LEN 28 #define FNIC_FDMI_ROM_VER_LEN 8 #define FNIC_FDMI_FW_VER_LEN 16 /* FNIC FDMI Register PA Macros */ #define FNIC_FDMI_TYPE_FC4_TYPES 0X1 #define FNIC_FDMI_TYPE_SUPPORTED_SPEEDS 0X2 #define FNIC_FDMI_TYPE_CURRENT_SPEED 0X3 #define FNIC_FDMI_TYPE_MAX_FRAME_SIZE 0X4 #define FNIC_FDMI_TYPE_OS_NAME 0X5 #define FNIC_FDMI_TYPE_HOST_NAME 0X6 #define FNIC_FDMI_NUM_PORT_ATTRS 6 #define FNIC_FDMI_FC4_LEN 32 #define FNIC_FDMI_SUPP_SPEED_LEN 4 #define FNIC_FDMI_CUR_SPEED_LEN 4 #define FNIC_FDMI_MFS_LEN 4 #define FNIC_FDMI_MFS 0x800 #define FNIC_FDMI_OS_NAME_LEN 16 #define FNIC_FDMI_HN_LEN 24 #define FDLS_FDMI_PLOGI_PENDING 0x1 #define FDLS_FDMI_REG_HBA_PENDING 0x2 #define FDLS_FDMI_RPA_PENDING 0x4 #define FDLS_FDMI_ABORT_PENDING 0x8 #define FDLS_FDMI_MAX_RETRY 3 #define RETRIES_EXHAUSTED(iport) \ (iport->fabric.retry_counter == FABRIC_LOGO_MAX_RETRY) #define FNIC_TPORT_MAX_NEXUS_RESTART (8) #define SCHEDULE_OXID_FREE_RETRY_TIME (300) /* Private Functions */ static void fdls_fdmi_register_hba(struct fnic_iport_s *iport); static void fdls_fdmi_register_pa(struct fnic_iport_s *iport); static void fdls_send_rpn_id(struct fnic_iport_s *iport); static void fdls_process_flogi_rsp(struct fnic_iport_s *iport, struct fc_frame_header *fchdr, void *rx_frame); static void fnic_fdls_start_plogi(struct fnic_iport_s *iport); static void fnic_fdls_start_flogi(struct fnic_iport_s *iport); static struct fnic_tport_s *fdls_create_tport(struct fnic_iport_s *iport, uint32_t fcid, uint64_t wwpn); static void fdls_target_restart_nexus(struct fnic_tport_s *tport); static void fdls_start_tport_timer(struct fnic_iport_s *iport, struct fnic_tport_s *tport, int timeout); static void fdls_tport_timer_callback(struct timer_list *t); static void fdls_send_fdmi_plogi(struct fnic_iport_s *iport); static void fdls_start_fabric_timer(struct fnic_iport_s *iport, int timeout); static void fdls_init_plogi_frame(uint8_t *frame, struct fnic_iport_s *iport); static void fdls_init_els_acc_frame(uint8_t *frame, struct fnic_iport_s *iport); static void fdls_init_els_rjt_frame(uint8_t *frame, struct fnic_iport_s *iport); static void fdls_init_logo_frame(uint8_t *frame, struct fnic_iport_s *iport); static void fdls_init_fabric_abts_frame(uint8_t *frame, struct fnic_iport_s *iport); uint8_t *fdls_alloc_frame(struct fnic_iport_s *iport) { struct fnic *fnic = iport->fnic; uint8_t *frame = NULL; frame = mempool_alloc(fnic->frame_pool, GFP_ATOMIC); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame"); return NULL; } memset(frame, 0, FNIC_FCOE_FRAME_MAXSZ); return frame; } /** * fdls_alloc_oxid - Allocate an oxid from the bitmap based oxid pool * @iport: Handle to iport instance * @oxid_frame_type: Type of frame to allocate * @active_oxid: the oxid which is in use * * Called with fnic lock held */ uint16_t fdls_alloc_oxid(struct fnic_iport_s *iport, int oxid_frame_type, uint16_t *active_oxid) { struct fnic *fnic = iport->fnic; struct fnic_oxid_pool_s *oxid_pool = &iport->oxid_pool; int idx; uint16_t oxid; lockdep_assert_held(&fnic->fnic_lock); /* * Allocate next available oxid from bitmap */ idx = find_next_zero_bit(oxid_pool->bitmap, FNIC_OXID_POOL_SZ, oxid_pool->next_idx); if (idx == FNIC_OXID_POOL_SZ) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Alloc oxid: all oxid slots are busy iport state:%d\n", iport->state); return FNIC_UNASSIGNED_OXID; } WARN_ON(test_and_set_bit(idx, oxid_pool->bitmap)); oxid_pool->next_idx = (idx + 1) % FNIC_OXID_POOL_SZ; /* cycle through the bitmap */ oxid = FNIC_OXID_ENCODE(idx, oxid_frame_type); *active_oxid = oxid; FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "alloc oxid: 0x%x, iport state: %d\n", oxid, iport->state); return oxid; } /** * fdls_free_oxid_idx - Free the oxid using the idx * @iport: Handle to iport instance * @oxid_idx: The index to free * * Free the oxid immediately and make it available for new requests * Called with fnic lock held */ static void fdls_free_oxid_idx(struct fnic_iport_s *iport, uint16_t oxid_idx) { struct fnic *fnic = iport->fnic; struct fnic_oxid_pool_s *oxid_pool = &iport->oxid_pool; lockdep_assert_held(&fnic->fnic_lock); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "free oxid idx: 0x%x\n", oxid_idx); WARN_ON(!test_and_clear_bit(oxid_idx, oxid_pool->bitmap)); } /** * fdls_reclaim_oxid_handler - Callback handler for delayed_oxid_work * @work: Handle to work_struct * * Scheduled when an oxid is to be freed later * After freeing expired oxid(s), the handler schedules * another callback with the remaining time * of next unexpired entry in the reclaim list. */ void fdls_reclaim_oxid_handler(struct work_struct *work) { struct fnic_oxid_pool_s *oxid_pool = container_of(work, struct fnic_oxid_pool_s, oxid_reclaim_work.work); struct fnic_iport_s *iport = container_of(oxid_pool, struct fnic_iport_s, oxid_pool); struct fnic *fnic = iport->fnic; struct reclaim_entry_s *reclaim_entry, *next; unsigned long delay_j, cur_jiffies; FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Reclaim oxid callback\n"); spin_lock_irqsave(&fnic->fnic_lock, fnic->lock_flags); /* Though the work was scheduled for one entry, * walk through and free the expired entries which might have been scheduled * at around the same time as the first entry */ list_for_each_entry_safe(reclaim_entry, next, &(oxid_pool->oxid_reclaim_list), links) { /* The list is always maintained in the order of expiry time */ cur_jiffies = jiffies; if (time_before(cur_jiffies, reclaim_entry->expires)) break; list_del(&reclaim_entry->links); fdls_free_oxid_idx(iport, reclaim_entry->oxid_idx); kfree(reclaim_entry); } /* schedule to free up the next entry */ if (!list_empty(&oxid_pool->oxid_reclaim_list)) { reclaim_entry = list_first_entry(&oxid_pool->oxid_reclaim_list, struct reclaim_entry_s, links); delay_j = reclaim_entry->expires - cur_jiffies; schedule_delayed_work(&oxid_pool->oxid_reclaim_work, delay_j); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Scheduling next callback at:%ld jiffies\n", delay_j); } spin_unlock_irqrestore(&fnic->fnic_lock, fnic->lock_flags); } /** * fdls_free_oxid - Helper function to free the oxid * @iport: Handle to iport instance * @oxid: oxid to free * @active_oxid: the oxid which is in use * * Called with fnic lock held */ void fdls_free_oxid(struct fnic_iport_s *iport, uint16_t oxid, uint16_t *active_oxid) { fdls_free_oxid_idx(iport, FNIC_OXID_IDX(oxid)); *active_oxid = FNIC_UNASSIGNED_OXID; } /** * fdls_schedule_oxid_free - Schedule oxid to be freed later * @iport: Handle to iport instance * @active_oxid: the oxid which is in use * * Gets called in a rare case scenario when both a command * (fdls or target discovery) timed out and the following ABTS * timed out as well, without a link change. * * Called with fnic lock held */ void fdls_schedule_oxid_free(struct fnic_iport_s *iport, uint16_t *active_oxid) { struct fnic *fnic = iport->fnic; struct fnic_oxid_pool_s *oxid_pool = &iport->oxid_pool; struct reclaim_entry_s *reclaim_entry; unsigned long delay_j = msecs_to_jiffies(OXID_RECLAIM_TOV(iport)); int oxid_idx = FNIC_OXID_IDX(*active_oxid); lockdep_assert_held(&fnic->fnic_lock); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Schedule oxid free. oxid: 0x%x\n", *active_oxid); *active_oxid = FNIC_UNASSIGNED_OXID; reclaim_entry = (struct reclaim_entry_s *) kzalloc(sizeof(struct reclaim_entry_s), GFP_ATOMIC); if (!reclaim_entry) { FNIC_FCS_DBG(KERN_WARNING, fnic->host, fnic->fnic_num, "Failed to allocate memory for reclaim struct for oxid idx: %d\n", oxid_idx); /* Retry the scheduling */ WARN_ON(test_and_set_bit(oxid_idx, oxid_pool->pending_schedule_free)); schedule_delayed_work(&oxid_pool->schedule_oxid_free_retry, 0); return; } reclaim_entry->oxid_idx = oxid_idx; reclaim_entry->expires = round_jiffies(jiffies + delay_j); list_add_tail(&reclaim_entry->links, &oxid_pool->oxid_reclaim_list); schedule_delayed_work(&oxid_pool->oxid_reclaim_work, delay_j); } /** * fdls_schedule_oxid_free_retry_work - Thread to schedule the * oxid to be freed later * * @work: Handle to the work struct */ void fdls_schedule_oxid_free_retry_work(struct work_struct *work) { struct fnic_oxid_pool_s *oxid_pool = container_of(work, struct fnic_oxid_pool_s, schedule_oxid_free_retry.work); struct fnic_iport_s *iport = container_of(oxid_pool, struct fnic_iport_s, oxid_pool); struct fnic *fnic = iport->fnic; struct reclaim_entry_s *reclaim_entry; unsigned long delay_j = msecs_to_jiffies(OXID_RECLAIM_TOV(iport)); int idx; spin_lock_irqsave(&fnic->fnic_lock, fnic->lock_flags); for_each_set_bit(idx, oxid_pool->pending_schedule_free, FNIC_OXID_POOL_SZ) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Schedule oxid free. oxid idx: %d\n", idx); spin_unlock_irqrestore(&fnic->fnic_lock, fnic->lock_flags); reclaim_entry = (struct reclaim_entry_s *) kzalloc(sizeof(struct reclaim_entry_s), GFP_KERNEL); spin_lock_irqsave(&fnic->fnic_lock, fnic->lock_flags); if (!reclaim_entry) { FNIC_FCS_DBG(KERN_WARNING, fnic->host, fnic->fnic_num, "Failed to allocate memory for reclaim struct for oxid idx: 0x%x\n", idx); schedule_delayed_work(&oxid_pool->schedule_oxid_free_retry, msecs_to_jiffies(SCHEDULE_OXID_FREE_RETRY_TIME)); spin_unlock_irqrestore(&fnic->fnic_lock, fnic->lock_flags); return; } if (test_and_clear_bit(idx, oxid_pool->pending_schedule_free)) { reclaim_entry->oxid_idx = idx; reclaim_entry->expires = round_jiffies(jiffies + delay_j); list_add_tail(&reclaim_entry->links, &oxid_pool->oxid_reclaim_list); schedule_delayed_work(&oxid_pool->oxid_reclaim_work, delay_j); } else { /* unlikely scenario, free the allocated memory and continue */ kfree(reclaim_entry); } } spin_unlock_irqrestore(&fnic->fnic_lock, fnic->lock_flags); } static bool fdls_is_oxid_fabric_req(uint16_t oxid) { int oxid_frame_type = FNIC_FRAME_TYPE(oxid); switch (oxid_frame_type) { case FNIC_FRAME_TYPE_FABRIC_FLOGI: case FNIC_FRAME_TYPE_FABRIC_PLOGI: case FNIC_FRAME_TYPE_FABRIC_RPN: case FNIC_FRAME_TYPE_FABRIC_RFT: case FNIC_FRAME_TYPE_FABRIC_RFF: case FNIC_FRAME_TYPE_FABRIC_GPN_FT: case FNIC_FRAME_TYPE_FABRIC_LOGO: break; default: return false; } return true; } static bool fdls_is_oxid_fdmi_req(uint16_t oxid) { int oxid_frame_type = FNIC_FRAME_TYPE(oxid); switch (oxid_frame_type) { case FNIC_FRAME_TYPE_FDMI_PLOGI: case FNIC_FRAME_TYPE_FDMI_RHBA: case FNIC_FRAME_TYPE_FDMI_RPA: break; default: return false; } return true; } static bool fdls_is_oxid_tgt_req(uint16_t oxid) { int oxid_frame_type = FNIC_FRAME_TYPE(oxid); switch (oxid_frame_type) { case FNIC_FRAME_TYPE_TGT_PLOGI: case FNIC_FRAME_TYPE_TGT_PRLI: case FNIC_FRAME_TYPE_TGT_ADISC: case FNIC_FRAME_TYPE_TGT_LOGO: break; default: return false; } return true; } static void fdls_reset_oxid_pool(struct fnic_iport_s *iport) { struct fnic_oxid_pool_s *oxid_pool = &iport->oxid_pool; oxid_pool->next_idx = 0; } void fnic_del_fabric_timer_sync(struct fnic *fnic) { fnic->iport.fabric.del_timer_inprogress = 1; spin_unlock_irqrestore(&fnic->fnic_lock, fnic->lock_flags); del_timer_sync(&fnic->iport.fabric.retry_timer); spin_lock_irqsave(&fnic->fnic_lock, fnic->lock_flags); fnic->iport.fabric.del_timer_inprogress = 0; } void fnic_del_tport_timer_sync(struct fnic *fnic, struct fnic_tport_s *tport) { tport->del_timer_inprogress = 1; spin_unlock_irqrestore(&fnic->fnic_lock, fnic->lock_flags); del_timer_sync(&tport->retry_timer); spin_lock_irqsave(&fnic->fnic_lock, fnic->lock_flags); tport->del_timer_inprogress = 0; } static void fdls_start_fabric_timer(struct fnic_iport_s *iport, int timeout) { u64 fabric_tov; struct fnic *fnic = iport->fnic; if (iport->fabric.timer_pending) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "iport fcid: 0x%x: Canceling fabric disc timer\n", iport->fcid); fnic_del_fabric_timer_sync(fnic); iport->fabric.timer_pending = 0; } if (!(iport->fabric.flags & FNIC_FDLS_FABRIC_ABORT_ISSUED)) iport->fabric.retry_counter++; fabric_tov = jiffies + msecs_to_jiffies(timeout); mod_timer(&iport->fabric.retry_timer, round_jiffies(fabric_tov)); iport->fabric.timer_pending = 1; FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "fabric timer is %d ", timeout); } static void fdls_start_tport_timer(struct fnic_iport_s *iport, struct fnic_tport_s *tport, int timeout) { u64 fabric_tov; struct fnic *fnic = iport->fnic; if (tport->timer_pending) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "tport fcid 0x%x: Canceling disc timer\n", tport->fcid); fnic_del_tport_timer_sync(fnic, tport); tport->timer_pending = 0; } if (!(tport->flags & FNIC_FDLS_TGT_ABORT_ISSUED)) tport->retry_counter++; fabric_tov = jiffies + msecs_to_jiffies(timeout); mod_timer(&tport->retry_timer, round_jiffies(fabric_tov)); tport->timer_pending = 1; } void fdls_init_plogi_frame(uint8_t *frame, struct fnic_iport_s *iport) { struct fc_std_flogi *pplogi; uint8_t s_id[3]; pplogi = (struct fc_std_flogi *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); *pplogi = (struct fc_std_flogi) { .fchdr = {.fh_r_ctl = FC_RCTL_ELS_REQ, .fh_d_id = {0xFF, 0xFF, 0xFC}, .fh_type = FC_TYPE_ELS, .fh_f_ctl = {FNIC_ELS_REQ_FCTL, 0, 0}, .fh_rx_id = cpu_to_be16(FNIC_UNASSIGNED_RXID)}, .els = { .fl_cmd = ELS_PLOGI, .fl_csp = {.sp_hi_ver = FNIC_FC_PH_VER_HI, .sp_lo_ver = FNIC_FC_PH_VER_LO, .sp_bb_cred = cpu_to_be16(FNIC_FC_B2B_CREDIT), .sp_features = cpu_to_be16(FC_SP_FT_CIRO), .sp_bb_data = cpu_to_be16(FNIC_FC_B2B_RDF_SZ), .sp_tot_seq = cpu_to_be16(FNIC_FC_CONCUR_SEQS), .sp_rel_off = cpu_to_be16(FNIC_FC_RO_INFO), .sp_e_d_tov = cpu_to_be32(FC_DEF_E_D_TOV)}, .fl_cssp[2].cp_class = cpu_to_be16(FC_CPC_VALID | FC_CPC_SEQ), .fl_cssp[2].cp_rdfs = cpu_to_be16(0x800), .fl_cssp[2].cp_con_seq = cpu_to_be16(0xFF), .fl_cssp[2].cp_open_seq = 1} }; FNIC_STD_SET_NPORT_NAME(&pplogi->els.fl_wwpn, iport->wwpn); FNIC_STD_SET_NODE_NAME(&pplogi->els.fl_wwnn, iport->wwnn); FNIC_LOGI_SET_RDF_SIZE(pplogi->els, iport->max_payload_size); hton24(s_id, iport->fcid); FNIC_STD_SET_S_ID(pplogi->fchdr, s_id); } static void fdls_init_els_acc_frame(uint8_t *frame, struct fnic_iport_s *iport) { struct fc_std_els_acc_rsp *pels_acc; uint8_t s_id[3]; pels_acc = (struct fc_std_els_acc_rsp *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); *pels_acc = (struct fc_std_els_acc_rsp) { .fchdr = {.fh_r_ctl = FC_RCTL_ELS_REP, .fh_type = FC_TYPE_ELS, .fh_f_ctl = {FNIC_ELS_REP_FCTL, 0, 0}}, .acc.la_cmd = ELS_LS_ACC, }; hton24(s_id, iport->fcid); FNIC_STD_SET_S_ID(pels_acc->fchdr, s_id); FNIC_STD_SET_RX_ID(pels_acc->fchdr, FNIC_UNASSIGNED_RXID); } static void fdls_init_els_rjt_frame(uint8_t *frame, struct fnic_iport_s *iport) { struct fc_std_els_rjt_rsp *pels_rjt; pels_rjt = (struct fc_std_els_rjt_rsp *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); *pels_rjt = (struct fc_std_els_rjt_rsp) { .fchdr = {.fh_r_ctl = FC_RCTL_ELS_REP, .fh_type = FC_TYPE_ELS, .fh_f_ctl = {FNIC_ELS_REP_FCTL, 0, 0}}, .rej.er_cmd = ELS_LS_RJT, }; FNIC_STD_SET_RX_ID(pels_rjt->fchdr, FNIC_UNASSIGNED_RXID); } static void fdls_init_logo_frame(uint8_t *frame, struct fnic_iport_s *iport) { struct fc_std_logo *plogo; uint8_t s_id[3]; plogo = (struct fc_std_logo *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); *plogo = (struct fc_std_logo) { .fchdr = {.fh_r_ctl = FC_RCTL_ELS_REQ, .fh_type = FC_TYPE_ELS, .fh_f_ctl = {FNIC_ELS_REQ_FCTL, 0, 0}}, .els.fl_cmd = ELS_LOGO, }; hton24(s_id, iport->fcid); FNIC_STD_SET_S_ID(plogo->fchdr, s_id); memcpy(plogo->els.fl_n_port_id, s_id, 3); FNIC_STD_SET_NPORT_NAME(&plogo->els.fl_n_port_wwn, iport->wwpn); } static void fdls_init_fabric_abts_frame(uint8_t *frame, struct fnic_iport_s *iport) { struct fc_frame_header *pfabric_abts; pfabric_abts = (struct fc_frame_header *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); *pfabric_abts = (struct fc_frame_header) { .fh_r_ctl = FC_RCTL_BA_ABTS, /* ABTS */ .fh_s_id = {0x00, 0x00, 0x00}, .fh_cs_ctl = 0x00, .fh_type = FC_TYPE_BLS, .fh_f_ctl = {FNIC_REQ_ABTS_FCTL, 0, 0}, .fh_seq_id = 0x00, .fh_df_ctl = 0x00, .fh_seq_cnt = 0x0000, .fh_rx_id = cpu_to_be16(FNIC_UNASSIGNED_RXID), .fh_parm_offset = 0x00000000, /* bit:0 = 0 Abort a exchange */ }; } static void fdls_send_rscn_resp(struct fnic_iport_s *iport, struct fc_frame_header *rscn_fchdr) { uint8_t *frame; struct fc_std_els_acc_rsp *pels_acc; struct fnic *fnic = iport->fnic; uint16_t oxid; uint16_t frame_size = FNIC_ETH_FCOE_HDRS_OFFSET + sizeof(struct fc_std_els_acc_rsp); frame = fdls_alloc_frame(iport); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame to send RSCN response"); return; } pels_acc = (struct fc_std_els_acc_rsp *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); fdls_init_els_acc_frame(frame, iport); FNIC_STD_SET_D_ID(pels_acc->fchdr, rscn_fchdr->fh_s_id); oxid = FNIC_STD_GET_OX_ID(rscn_fchdr); FNIC_STD_SET_OX_ID(pels_acc->fchdr, oxid); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS send RSCN response with oxid: 0x%x", iport->fcid, oxid); fnic_send_fcoe_frame(iport, frame, frame_size); } static void fdls_send_logo_resp(struct fnic_iport_s *iport, struct fc_frame_header *req_fchdr) { uint8_t *frame; struct fc_std_els_acc_rsp *plogo_resp; struct fnic *fnic = iport->fnic; uint16_t oxid; uint16_t frame_size = FNIC_ETH_FCOE_HDRS_OFFSET + sizeof(struct fc_std_els_acc_rsp); frame = fdls_alloc_frame(iport); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame to send LOGO response"); return; } plogo_resp = (struct fc_std_els_acc_rsp *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); fdls_init_els_acc_frame(frame, iport); FNIC_STD_SET_D_ID(plogo_resp->fchdr, req_fchdr->fh_s_id); oxid = FNIC_STD_GET_OX_ID(req_fchdr); FNIC_STD_SET_OX_ID(plogo_resp->fchdr, oxid); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS send LOGO response with oxid: 0x%x", iport->fcid, oxid); fnic_send_fcoe_frame(iport, frame, frame_size); } void fdls_send_tport_abts(struct fnic_iport_s *iport, struct fnic_tport_s *tport) { uint8_t *frame; uint8_t s_id[3]; uint8_t d_id[3]; struct fnic *fnic = iport->fnic; struct fc_frame_header *ptport_abts; uint16_t frame_size = FNIC_ETH_FCOE_HDRS_OFFSET + sizeof(struct fc_frame_header); frame = fdls_alloc_frame(iport); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame to send tport ABTS"); return; } ptport_abts = (struct fc_frame_header *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); *ptport_abts = (struct fc_frame_header) { .fh_r_ctl = FC_RCTL_BA_ABTS, /* ABTS */ .fh_cs_ctl = 0x00, .fh_type = FC_TYPE_BLS, .fh_f_ctl = {FNIC_REQ_ABTS_FCTL, 0, 0}, .fh_seq_id = 0x00, .fh_df_ctl = 0x00, .fh_seq_cnt = 0x0000, .fh_rx_id = cpu_to_be16(FNIC_UNASSIGNED_RXID), .fh_parm_offset = 0x00000000, /* bit:0 = 0 Abort a exchange */ }; hton24(s_id, iport->fcid); hton24(d_id, tport->fcid); FNIC_STD_SET_S_ID(*ptport_abts, s_id); FNIC_STD_SET_D_ID(*ptport_abts, d_id); tport->flags |= FNIC_FDLS_TGT_ABORT_ISSUED; FNIC_STD_SET_OX_ID(*ptport_abts, tport->active_oxid); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS send tport abts: tport->state: %d ", iport->fcid, tport->state); fnic_send_fcoe_frame(iport, frame, frame_size); /* Even if fnic_send_fcoe_frame() fails we want to retry after timeout */ fdls_start_tport_timer(iport, tport, 2 * iport->e_d_tov); } static void fdls_send_fabric_abts(struct fnic_iport_s *iport) { uint8_t *frame; uint8_t s_id[3]; uint8_t d_id[3]; struct fnic *fnic = iport->fnic; struct fc_frame_header *pfabric_abts; uint16_t oxid; uint16_t frame_size = FNIC_ETH_FCOE_HDRS_OFFSET + sizeof(struct fc_frame_header); frame = fdls_alloc_frame(iport); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame to send fabric ABTS"); return; } pfabric_abts = (struct fc_frame_header *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); fdls_init_fabric_abts_frame(frame, iport); hton24(s_id, iport->fcid); switch (iport->fabric.state) { case FDLS_STATE_FABRIC_LOGO: hton24(d_id, FC_FID_FLOGI); FNIC_STD_SET_D_ID(*pfabric_abts, d_id); break; case FDLS_STATE_FABRIC_FLOGI: hton24(d_id, FC_FID_FLOGI); FNIC_STD_SET_D_ID(*pfabric_abts, d_id); break; case FDLS_STATE_FABRIC_PLOGI: FNIC_STD_SET_S_ID(*pfabric_abts, s_id); hton24(d_id, FC_FID_DIR_SERV); FNIC_STD_SET_D_ID(*pfabric_abts, d_id); break; case FDLS_STATE_RPN_ID: FNIC_STD_SET_S_ID(*pfabric_abts, s_id); hton24(d_id, FC_FID_DIR_SERV); FNIC_STD_SET_D_ID(*pfabric_abts, d_id); break; case FDLS_STATE_SCR: FNIC_STD_SET_S_ID(*pfabric_abts, s_id); hton24(d_id, FC_FID_FCTRL); FNIC_STD_SET_D_ID(*pfabric_abts, d_id); break; case FDLS_STATE_REGISTER_FC4_TYPES: FNIC_STD_SET_S_ID(*pfabric_abts, s_id); hton24(d_id, FC_FID_DIR_SERV); FNIC_STD_SET_D_ID(*pfabric_abts, d_id); break; case FDLS_STATE_REGISTER_FC4_FEATURES: FNIC_STD_SET_S_ID(*pfabric_abts, s_id); hton24(d_id, FC_FID_DIR_SERV); FNIC_STD_SET_D_ID(*pfabric_abts, d_id); break; case FDLS_STATE_GPN_FT: FNIC_STD_SET_S_ID(*pfabric_abts, s_id); hton24(d_id, FC_FID_DIR_SERV); FNIC_STD_SET_D_ID(*pfabric_abts, d_id); break; default: return; } oxid = iport->active_oxid_fabric_req; FNIC_STD_SET_OX_ID(*pfabric_abts, oxid); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS send fabric abts. iport->fabric.state: %d oxid: 0x%x", iport->fcid, iport->fabric.state, oxid); iport->fabric.flags |= FNIC_FDLS_FABRIC_ABORT_ISSUED; fnic_send_fcoe_frame(iport, frame, frame_size); /* Even if fnic_send_fcoe_frame() fails we want to retry after timeout */ fdls_start_fabric_timer(iport, 2 * iport->e_d_tov); iport->fabric.timer_pending = 1; } static void fdls_send_fdmi_abts(struct fnic_iport_s *iport) { uint8_t *frame; uint8_t d_id[3]; struct fnic *fnic = iport->fnic; struct fc_frame_header *pfabric_abts; unsigned long fdmi_tov; uint16_t oxid; uint16_t frame_size = FNIC_ETH_FCOE_HDRS_OFFSET + sizeof(struct fc_frame_header); frame = fdls_alloc_frame(iport); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame to send FDMI ABTS"); return; } pfabric_abts = (struct fc_frame_header *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); fdls_init_fabric_abts_frame(frame, iport); hton24(d_id, FC_FID_MGMT_SERV); FNIC_STD_SET_D_ID(*pfabric_abts, d_id); if (iport->fabric.fdmi_pending & FDLS_FDMI_PLOGI_PENDING) { oxid = iport->active_oxid_fdmi_plogi; FNIC_STD_SET_OX_ID(*pfabric_abts, oxid); fnic_send_fcoe_frame(iport, frame, frame_size); } else { if (iport->fabric.fdmi_pending & FDLS_FDMI_REG_HBA_PENDING) { oxid = iport->active_oxid_fdmi_rhba; FNIC_STD_SET_OX_ID(*pfabric_abts, oxid); fnic_send_fcoe_frame(iport, frame, frame_size); } if (iport->fabric.fdmi_pending & FDLS_FDMI_RPA_PENDING) { oxid = iport->active_oxid_fdmi_rpa; FNIC_STD_SET_OX_ID(*pfabric_abts, oxid); fnic_send_fcoe_frame(iport, frame, frame_size); } } fdmi_tov = jiffies + msecs_to_jiffies(2 * iport->e_d_tov); mod_timer(&iport->fabric.fdmi_timer, round_jiffies(fdmi_tov)); iport->fabric.fdmi_pending |= FDLS_FDMI_ABORT_PENDING; } static void fdls_send_fabric_flogi(struct fnic_iport_s *iport) { uint8_t *frame; struct fc_std_flogi *pflogi; struct fnic *fnic = iport->fnic; uint16_t oxid; uint16_t frame_size = FNIC_ETH_FCOE_HDRS_OFFSET + sizeof(struct fc_std_flogi); frame = fdls_alloc_frame(iport); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame to send FLOGI"); iport->fabric.flags |= FNIC_FDLS_RETRY_FRAME; goto err_out; } pflogi = (struct fc_std_flogi *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); *pflogi = (struct fc_std_flogi) { .fchdr = {.fh_r_ctl = FC_RCTL_ELS_REQ, .fh_d_id = {0xFF, 0xFF, 0xFE}, .fh_type = FC_TYPE_ELS, .fh_f_ctl = {FNIC_ELS_REQ_FCTL, 0, 0}, .fh_rx_id = cpu_to_be16(FNIC_UNASSIGNED_RXID)}, .els.fl_cmd = ELS_FLOGI, .els.fl_csp = {.sp_hi_ver = FNIC_FC_PH_VER_HI, .sp_lo_ver = FNIC_FC_PH_VER_LO, .sp_bb_cred = cpu_to_be16(FNIC_FC_B2B_CREDIT), .sp_bb_data = cpu_to_be16(FNIC_FC_B2B_RDF_SZ)}, .els.fl_cssp[2].cp_class = cpu_to_be16(FC_CPC_VALID | FC_CPC_SEQ) }; FNIC_STD_SET_NPORT_NAME(&pflogi->els.fl_wwpn, iport->wwpn); FNIC_STD_SET_NODE_NAME(&pflogi->els.fl_wwnn, iport->wwnn); FNIC_LOGI_SET_RDF_SIZE(pflogi->els, iport->max_payload_size); FNIC_LOGI_SET_R_A_TOV(pflogi->els, iport->r_a_tov); FNIC_LOGI_SET_E_D_TOV(pflogi->els, iport->e_d_tov); oxid = fdls_alloc_oxid(iport, FNIC_FRAME_TYPE_FABRIC_FLOGI, &iport->active_oxid_fabric_req); if (oxid == FNIC_UNASSIGNED_OXID) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: Failed to allocate OXID to send FLOGI", iport->fcid); mempool_free(frame, fnic->frame_pool); iport->fabric.flags |= FNIC_FDLS_RETRY_FRAME; goto err_out; } FNIC_STD_SET_OX_ID(pflogi->fchdr, oxid); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS send fabric FLOGI with oxid: 0x%x", iport->fcid, oxid); fnic_send_fcoe_frame(iport, frame, frame_size); atomic64_inc(&iport->iport_stats.fabric_flogi_sent); err_out: /* Even if fnic_send_fcoe_frame() fails we want to retry after timeout */ fdls_start_fabric_timer(iport, 2 * iport->e_d_tov); } static void fdls_send_fabric_plogi(struct fnic_iport_s *iport) { uint8_t *frame; struct fc_std_flogi *pplogi; struct fnic *fnic = iport->fnic; uint16_t oxid; uint16_t frame_size = FNIC_ETH_FCOE_HDRS_OFFSET + sizeof(struct fc_std_flogi); frame = fdls_alloc_frame(iport); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame to send PLOGI"); iport->fabric.flags |= FNIC_FDLS_RETRY_FRAME; goto err_out; } pplogi = (struct fc_std_flogi *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); fdls_init_plogi_frame(frame, iport); oxid = fdls_alloc_oxid(iport, FNIC_FRAME_TYPE_FABRIC_PLOGI, &iport->active_oxid_fabric_req); if (oxid == FNIC_UNASSIGNED_OXID) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: Failed to allocate OXID to send fabric PLOGI", iport->fcid); mempool_free(frame, fnic->frame_pool); iport->fabric.flags |= FNIC_FDLS_RETRY_FRAME; goto err_out; } FNIC_STD_SET_OX_ID(pplogi->fchdr, oxid); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS send fabric PLOGI with oxid: 0x%x", iport->fcid, oxid); fnic_send_fcoe_frame(iport, frame, frame_size); atomic64_inc(&iport->iport_stats.fabric_plogi_sent); err_out: /* Even if fnic_send_fcoe_frame() fails we want to retry after timeout */ fdls_start_fabric_timer(iport, 2 * iport->e_d_tov); } static void fdls_send_fdmi_plogi(struct fnic_iport_s *iport) { uint8_t *frame; struct fc_std_flogi *pplogi; struct fnic *fnic = iport->fnic; uint16_t oxid; uint16_t frame_size = FNIC_ETH_FCOE_HDRS_OFFSET + sizeof(struct fc_std_flogi); uint8_t d_id[3]; u64 fdmi_tov; frame = fdls_alloc_frame(iport); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame to send FDMI PLOGI"); goto err_out; } pplogi = (struct fc_std_flogi *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); fdls_init_plogi_frame(frame, iport); oxid = fdls_alloc_oxid(iport, FNIC_FRAME_TYPE_FDMI_PLOGI, &iport->active_oxid_fdmi_plogi); if (oxid == FNIC_UNASSIGNED_OXID) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: Failed to allocate OXID to send FDMI PLOGI", iport->fcid); mempool_free(frame, fnic->frame_pool); goto err_out; } FNIC_STD_SET_OX_ID(pplogi->fchdr, oxid); hton24(d_id, FC_FID_MGMT_SERV); FNIC_STD_SET_D_ID(pplogi->fchdr, d_id); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS send FDMI PLOGI with oxid: 0x%x", iport->fcid, oxid); fnic_send_fcoe_frame(iport, frame, frame_size); err_out: fdmi_tov = jiffies + msecs_to_jiffies(2 * iport->e_d_tov); mod_timer(&iport->fabric.fdmi_timer, round_jiffies(fdmi_tov)); iport->fabric.fdmi_pending = FDLS_FDMI_PLOGI_PENDING; } static void fdls_send_rpn_id(struct fnic_iport_s *iport) { uint8_t *frame; struct fc_std_rpn_id *prpn_id; struct fnic *fnic = iport->fnic; uint16_t oxid; uint16_t frame_size = FNIC_ETH_FCOE_HDRS_OFFSET + sizeof(struct fc_std_rpn_id); uint8_t fcid[3]; frame = fdls_alloc_frame(iport); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame to send RPN_ID"); iport->fabric.flags |= FNIC_FDLS_RETRY_FRAME; goto err_out; } prpn_id = (struct fc_std_rpn_id *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); *prpn_id = (struct fc_std_rpn_id) { .fchdr = {.fh_r_ctl = FC_RCTL_DD_UNSOL_CTL, .fh_d_id = {0xFF, 0xFF, 0xFC}, .fh_type = FC_TYPE_CT, .fh_f_ctl = {FNIC_ELS_REQ_FCTL, 0, 0}, .fh_rx_id = cpu_to_be16(FNIC_UNASSIGNED_RXID)}, .fc_std_ct_hdr = {.ct_rev = FC_CT_REV, .ct_fs_type = FC_FST_DIR, .ct_fs_subtype = FC_NS_SUBTYPE, .ct_cmd = cpu_to_be16(FC_NS_RPN_ID)} }; hton24(fcid, iport->fcid); FNIC_STD_SET_S_ID(prpn_id->fchdr, fcid); FNIC_STD_SET_PORT_ID(prpn_id->rpn_id, fcid); FNIC_STD_SET_PORT_NAME(prpn_id->rpn_id, iport->wwpn); oxid = fdls_alloc_oxid(iport, FNIC_FRAME_TYPE_FABRIC_RPN, &iport->active_oxid_fabric_req); if (oxid == FNIC_UNASSIGNED_OXID) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: Failed to allocate OXID to send RPN_ID", iport->fcid); mempool_free(frame, fnic->frame_pool); iport->fabric.flags |= FNIC_FDLS_RETRY_FRAME; goto err_out; } FNIC_STD_SET_OX_ID(prpn_id->fchdr, oxid); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS send RPN ID with oxid: 0x%x", iport->fcid, oxid); fnic_send_fcoe_frame(iport, frame, frame_size); err_out: /* Even if fnic_send_fcoe_frame() fails we want to retry after timeout */ fdls_start_fabric_timer(iport, 2 * iport->e_d_tov); } static void fdls_send_scr(struct fnic_iport_s *iport) { uint8_t *frame; struct fc_std_scr *pscr; struct fnic *fnic = iport->fnic; uint16_t oxid; uint16_t frame_size = FNIC_ETH_FCOE_HDRS_OFFSET + sizeof(struct fc_std_scr); uint8_t fcid[3]; frame = fdls_alloc_frame(iport); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame to send SCR"); iport->fabric.flags |= FNIC_FDLS_RETRY_FRAME; goto err_out; } pscr = (struct fc_std_scr *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); *pscr = (struct fc_std_scr) { .fchdr = {.fh_r_ctl = FC_RCTL_ELS_REQ, .fh_d_id = {0xFF, 0xFF, 0xFD}, .fh_type = FC_TYPE_ELS, .fh_f_ctl = {FNIC_ELS_REQ_FCTL, 0, 0}, .fh_rx_id = cpu_to_be16(FNIC_UNASSIGNED_RXID)}, .scr = {.scr_cmd = ELS_SCR, .scr_reg_func = ELS_SCRF_FULL} }; hton24(fcid, iport->fcid); FNIC_STD_SET_S_ID(pscr->fchdr, fcid); oxid = fdls_alloc_oxid(iport, FNIC_FRAME_TYPE_FABRIC_SCR, &iport->active_oxid_fabric_req); if (oxid == FNIC_UNASSIGNED_OXID) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: Failed to allocate OXID to send SCR", iport->fcid); mempool_free(frame, fnic->frame_pool); iport->fabric.flags |= FNIC_FDLS_RETRY_FRAME; goto err_out; } FNIC_STD_SET_OX_ID(pscr->fchdr, oxid); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS send SCR with oxid: 0x%x", iport->fcid, oxid); fnic_send_fcoe_frame(iport, frame, frame_size); atomic64_inc(&iport->iport_stats.fabric_scr_sent); err_out: /* Even if fnic_send_fcoe_frame() fails we want to retry after timeout */ fdls_start_fabric_timer(iport, 2 * iport->e_d_tov); } static void fdls_send_gpn_ft(struct fnic_iport_s *iport, int fdls_state) { uint8_t *frame; struct fc_std_gpn_ft *pgpn_ft; struct fnic *fnic = iport->fnic; uint16_t oxid; uint16_t frame_size = FNIC_ETH_FCOE_HDRS_OFFSET + sizeof(struct fc_std_gpn_ft); uint8_t fcid[3]; frame = fdls_alloc_frame(iport); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame to send GPN FT"); iport->fabric.flags |= FNIC_FDLS_RETRY_FRAME; goto err_out; } pgpn_ft = (struct fc_std_gpn_ft *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); *pgpn_ft = (struct fc_std_gpn_ft) { .fchdr = {.fh_r_ctl = FC_RCTL_DD_UNSOL_CTL, .fh_d_id = {0xFF, 0xFF, 0xFC}, .fh_type = FC_TYPE_CT, .fh_f_ctl = {FNIC_ELS_REQ_FCTL, 0, 0}, .fh_rx_id = cpu_to_be16(FNIC_UNASSIGNED_RXID)}, .fc_std_ct_hdr = {.ct_rev = FC_CT_REV, .ct_fs_type = FC_FST_DIR, .ct_fs_subtype = FC_NS_SUBTYPE, .ct_cmd = cpu_to_be16(FC_NS_GPN_FT)}, .gpn_ft.fn_fc4_type = 0x08 }; hton24(fcid, iport->fcid); FNIC_STD_SET_S_ID(pgpn_ft->fchdr, fcid); oxid = fdls_alloc_oxid(iport, FNIC_FRAME_TYPE_FABRIC_GPN_FT, &iport->active_oxid_fabric_req); if (oxid == FNIC_UNASSIGNED_OXID) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: Failed to allocate OXID to send GPN FT", iport->fcid); mempool_free(frame, fnic->frame_pool); iport->fabric.flags |= FNIC_FDLS_RETRY_FRAME; goto err_out; } FNIC_STD_SET_OX_ID(pgpn_ft->fchdr, oxid); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS send GPN FT with oxid: 0x%x", iport->fcid, oxid); fnic_send_fcoe_frame(iport, frame, frame_size); err_out: /* Even if fnic_send_fcoe_frame() fails we want to retry after timeout */ fdls_start_fabric_timer(iport, 2 * iport->e_d_tov); fdls_set_state((&iport->fabric), fdls_state); } static void fdls_send_tgt_adisc(struct fnic_iport_s *iport, struct fnic_tport_s *tport) { uint8_t *frame; struct fc_std_els_adisc *padisc; uint8_t s_id[3]; uint8_t d_id[3]; uint16_t oxid; struct fnic *fnic = iport->fnic; uint16_t frame_size = FNIC_ETH_FCOE_HDRS_OFFSET + sizeof(struct fc_std_els_adisc); frame = fdls_alloc_frame(iport); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame to send TGT ADISC"); tport->flags |= FNIC_FDLS_RETRY_FRAME; goto err_out; } padisc = (struct fc_std_els_adisc *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); hton24(s_id, iport->fcid); hton24(d_id, tport->fcid); memcpy(padisc->els.adisc_port_id, s_id, 3); FNIC_STD_SET_S_ID(padisc->fchdr, s_id); FNIC_STD_SET_D_ID(padisc->fchdr, d_id); FNIC_STD_SET_F_CTL(padisc->fchdr, FNIC_ELS_REQ_FCTL << 16); FNIC_STD_SET_R_CTL(padisc->fchdr, FC_RCTL_ELS_REQ); FNIC_STD_SET_TYPE(padisc->fchdr, FC_TYPE_ELS); oxid = fdls_alloc_oxid(iport, FNIC_FRAME_TYPE_TGT_ADISC, &tport->active_oxid); if (oxid == FNIC_UNASSIGNED_OXID) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: Failed to allocate OXID to send TGT ADISC", iport->fcid); mempool_free(frame, fnic->frame_pool); tport->flags |= FNIC_FDLS_RETRY_FRAME; goto err_out; } FNIC_STD_SET_OX_ID(padisc->fchdr, oxid); FNIC_STD_SET_RX_ID(padisc->fchdr, FNIC_UNASSIGNED_RXID); tport->flags &= ~FNIC_FDLS_TGT_ABORT_ISSUED; FNIC_STD_SET_NPORT_NAME(&padisc->els.adisc_wwpn, iport->wwpn); FNIC_STD_SET_NODE_NAME(&padisc->els.adisc_wwnn, iport->wwnn); padisc->els.adisc_cmd = ELS_ADISC; FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS send ADISC to tgt fcid: 0x%x", iport->fcid, tport->fcid); atomic64_inc(&iport->iport_stats.tport_adisc_sent); fnic_send_fcoe_frame(iport, frame, frame_size); err_out: /* Even if fnic_send_fcoe_frame() fails we want to retry after timeout */ fdls_start_tport_timer(iport, tport, 2 * iport->e_d_tov); } bool fdls_delete_tport(struct fnic_iport_s *iport, struct fnic_tport_s *tport) { struct fnic_tport_event_s *tport_del_evt; struct fnic *fnic = iport->fnic; if ((tport->state == FDLS_TGT_STATE_OFFLINING) || (tport->state == FDLS_TGT_STATE_OFFLINE)) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "tport fcid 0x%x: tport state is offlining/offline\n", tport->fcid); return false; } fdls_set_tport_state(tport, FDLS_TGT_STATE_OFFLINING); /* * By setting this flag, the tport will not be seen in a look-up * in an RSCN. Even if we move to multithreaded model, this tport * will be destroyed and a new RSCN will have to create a new one */ tport->flags |= FNIC_FDLS_TPORT_TERMINATING; if (tport->timer_pending) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "tport fcid 0x%x: Canceling disc timer\n", tport->fcid); fnic_del_tport_timer_sync(fnic, tport); tport->timer_pending = 0; } spin_unlock_irqrestore(&fnic->fnic_lock, fnic->lock_flags); fnic_rport_exch_reset(iport->fnic, tport->fcid); spin_lock_irqsave(&fnic->fnic_lock, fnic->lock_flags); if (tport->flags & FNIC_FDLS_SCSI_REGISTERED) { tport_del_evt = kzalloc(sizeof(struct fnic_tport_event_s), GFP_ATOMIC); if (!tport_del_evt) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Failed to allocate memory for tport fcid: 0x%0x\n", tport->fcid); return false; } tport_del_evt->event = TGT_EV_RPORT_DEL; tport_del_evt->arg1 = (void *) tport; list_add_tail(&tport_del_evt->links, &fnic->tport_event_list); queue_work(fnic_event_queue, &fnic->tport_work); } else { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "tport 0x%x not reg with scsi_transport. Freeing locally", tport->fcid); list_del(&tport->links); kfree(tport); } return true; } static void fdls_send_tgt_plogi(struct fnic_iport_s *iport, struct fnic_tport_s *tport) { uint8_t *frame; struct fc_std_flogi *pplogi; struct fnic *fnic = iport->fnic; uint16_t oxid; uint16_t frame_size = FNIC_ETH_FCOE_HDRS_OFFSET + sizeof(struct fc_std_flogi); uint8_t d_id[3]; uint32_t timeout; frame = fdls_alloc_frame(iport); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame to send TGT PLOGI"); tport->flags |= FNIC_FDLS_RETRY_FRAME; goto err_out; } pplogi = (struct fc_std_flogi *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); fdls_init_plogi_frame(frame, iport); oxid = fdls_alloc_oxid(iport, FNIC_FRAME_TYPE_TGT_PLOGI, &tport->active_oxid); if (oxid == FNIC_UNASSIGNED_OXID) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: Failed to allocate oxid to send PLOGI to fcid: 0x%x", iport->fcid, tport->fcid); mempool_free(frame, fnic->frame_pool); tport->flags |= FNIC_FDLS_RETRY_FRAME; goto err_out; } FNIC_STD_SET_OX_ID(pplogi->fchdr, oxid); tport->flags &= ~FNIC_FDLS_TGT_ABORT_ISSUED; hton24(d_id, tport->fcid); FNIC_STD_SET_D_ID(pplogi->fchdr, d_id); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS send tgt PLOGI to tgt: 0x%x with oxid: 0x%x", iport->fcid, tport->fcid, oxid); fnic_send_fcoe_frame(iport, frame, frame_size); atomic64_inc(&iport->iport_stats.tport_plogi_sent); err_out: timeout = max(2 * iport->e_d_tov, iport->plogi_timeout); /* Even if fnic_send_fcoe_frame() fails we want to retry after timeout */ fdls_start_tport_timer(iport, tport, timeout); } static uint16_t fnic_fc_plogi_rsp_rdf(struct fnic_iport_s *iport, struct fc_std_flogi *plogi_rsp) { uint16_t b2b_rdf_size = be16_to_cpu(FNIC_LOGI_RDF_SIZE(plogi_rsp->els)); uint16_t spc3_rdf_size = be16_to_cpu(plogi_rsp->els.fl_cssp[2].cp_rdfs) & FNIC_FC_C3_RDF; struct fnic *fnic = iport->fnic; FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "MFS: b2b_rdf_size: 0x%x spc3_rdf_size: 0x%x", b2b_rdf_size, spc3_rdf_size); return min(b2b_rdf_size, spc3_rdf_size); } static void fdls_send_register_fc4_types(struct fnic_iport_s *iport) { uint8_t *frame; struct fc_std_rft_id *prft_id; struct fnic *fnic = iport->fnic; uint16_t oxid; uint16_t frame_size = FNIC_ETH_FCOE_HDRS_OFFSET + sizeof(struct fc_std_rft_id); uint8_t fcid[3]; frame = fdls_alloc_frame(iport); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame to send RFT"); return; } prft_id = (struct fc_std_rft_id *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); *prft_id = (struct fc_std_rft_id) { .fchdr = {.fh_r_ctl = FC_RCTL_DD_UNSOL_CTL, .fh_d_id = {0xFF, 0xFF, 0xFC}, .fh_type = FC_TYPE_CT, .fh_f_ctl = {FNIC_ELS_REQ_FCTL, 0, 0}, .fh_rx_id = cpu_to_be16(FNIC_UNASSIGNED_RXID)}, .fc_std_ct_hdr = {.ct_rev = FC_CT_REV, .ct_fs_type = FC_FST_DIR, .ct_fs_subtype = FC_NS_SUBTYPE, .ct_cmd = cpu_to_be16(FC_NS_RFT_ID)} }; hton24(fcid, iport->fcid); FNIC_STD_SET_S_ID(prft_id->fchdr, fcid); FNIC_STD_SET_PORT_ID(prft_id->rft_id, fcid); oxid = fdls_alloc_oxid(iport, FNIC_FRAME_TYPE_FABRIC_RFT, &iport->active_oxid_fabric_req); if (oxid == FNIC_UNASSIGNED_OXID) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: Failed to allocate OXID to send RFT", iport->fcid); mempool_free(frame, fnic->frame_pool); return; } FNIC_STD_SET_OX_ID(prft_id->fchdr, oxid); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS send RFT with oxid: 0x%x", iport->fcid, oxid); prft_id->rft_id.fr_fts.ff_type_map[0] = cpu_to_be32(1 << FC_TYPE_FCP); prft_id->rft_id.fr_fts.ff_type_map[1] = cpu_to_be32(1 << (FC_TYPE_CT % FC_NS_BPW)); fnic_send_fcoe_frame(iport, frame, frame_size); /* Even if fnic_send_fcoe_frame() fails we want to retry after timeout */ fdls_start_fabric_timer(iport, 2 * iport->e_d_tov); } static void fdls_send_register_fc4_features(struct fnic_iport_s *iport) { uint8_t *frame; struct fc_std_rff_id *prff_id; struct fnic *fnic = iport->fnic; uint16_t oxid; uint16_t frame_size = FNIC_ETH_FCOE_HDRS_OFFSET + sizeof(struct fc_std_rff_id); uint8_t fcid[3]; frame = fdls_alloc_frame(iport); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame to send RFF"); return; } prff_id = (struct fc_std_rff_id *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); *prff_id = (struct fc_std_rff_id) { .fchdr = {.fh_r_ctl = FC_RCTL_DD_UNSOL_CTL, .fh_d_id = {0xFF, 0xFF, 0xFC}, .fh_type = FC_TYPE_CT, .fh_f_ctl = {FNIC_ELS_REQ_FCTL, 0, 0}, .fh_rx_id = cpu_to_be16(FNIC_UNASSIGNED_RXID)}, .fc_std_ct_hdr = {.ct_rev = FC_CT_REV, .ct_fs_type = FC_FST_DIR, .ct_fs_subtype = FC_NS_SUBTYPE, .ct_cmd = cpu_to_be16(FC_NS_RFF_ID)}, .rff_id.fr_feat = 0x2, .rff_id.fr_type = FC_TYPE_FCP }; hton24(fcid, iport->fcid); FNIC_STD_SET_S_ID(prff_id->fchdr, fcid); FNIC_STD_SET_PORT_ID(prff_id->rff_id, fcid); oxid = fdls_alloc_oxid(iport, FNIC_FRAME_TYPE_FABRIC_RFF, &iport->active_oxid_fabric_req); if (oxid == FNIC_UNASSIGNED_OXID) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: Failed to allocate OXID to send RFF", iport->fcid); mempool_free(frame, fnic->frame_pool); return; } FNIC_STD_SET_OX_ID(prff_id->fchdr, oxid); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS send RFF with oxid: 0x%x", iport->fcid, oxid); prff_id->rff_id.fr_type = FC_TYPE_FCP; fnic_send_fcoe_frame(iport, frame, frame_size); /* Even if fnic_send_fcoe_frame() fails we want to retry after timeout */ fdls_start_fabric_timer(iport, 2 * iport->e_d_tov); } static void fdls_send_tgt_prli(struct fnic_iport_s *iport, struct fnic_tport_s *tport) { uint8_t *frame; struct fc_std_els_prli *pprli; struct fnic *fnic = iport->fnic; uint16_t oxid; uint16_t frame_size = FNIC_ETH_FCOE_HDRS_OFFSET + sizeof(struct fc_std_els_prli); uint8_t s_id[3]; uint8_t d_id[3]; uint32_t timeout; frame = fdls_alloc_frame(iport); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame to send TGT PRLI"); tport->flags |= FNIC_FDLS_RETRY_FRAME; goto err_out; } pprli = (struct fc_std_els_prli *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); *pprli = (struct fc_std_els_prli) { .fchdr = {.fh_r_ctl = FC_RCTL_ELS_REQ, .fh_type = FC_TYPE_ELS, .fh_f_ctl = {FNIC_ELS_REQ_FCTL, 0, 0}, .fh_rx_id = cpu_to_be16(FNIC_UNASSIGNED_RXID)}, .els_prli = {.prli_cmd = ELS_PRLI, .prli_spp_len = 16, .prli_len = cpu_to_be16(0x14)}, .sp = {.spp_type = 0x08, .spp_flags = 0x0020, .spp_params = cpu_to_be32(0xA2)} }; oxid = fdls_alloc_oxid(iport, FNIC_FRAME_TYPE_TGT_PRLI, &tport->active_oxid); if (oxid == FNIC_UNASSIGNED_OXID) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "0x%x: Failed to allocate OXID to send TGT PRLI to 0x%x", iport->fcid, tport->fcid); mempool_free(frame, fnic->frame_pool); tport->flags |= FNIC_FDLS_RETRY_FRAME; goto err_out; } tport->flags &= ~FNIC_FDLS_TGT_ABORT_ISSUED; hton24(s_id, iport->fcid); hton24(d_id, tport->fcid); FNIC_STD_SET_OX_ID(pprli->fchdr, oxid); FNIC_STD_SET_S_ID(pprli->fchdr, s_id); FNIC_STD_SET_D_ID(pprli->fchdr, d_id); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS send PRLI to tgt: 0x%x with oxid: 0x%x", iport->fcid, tport->fcid, oxid); fnic_send_fcoe_frame(iport, frame, frame_size); atomic64_inc(&iport->iport_stats.tport_prli_sent); err_out: timeout = max(2 * iport->e_d_tov, iport->plogi_timeout); /* Even if fnic_send_fcoe_frame() fails we want to retry after timeout */ fdls_start_tport_timer(iport, tport, timeout); } /** * fdls_send_fabric_logo - Send flogo to the fcf * @iport: Handle to fnic iport * * This function does not change or check the fabric state. * It the caller's responsibility to set the appropriate iport fabric * state when this is called. Normally it is FDLS_STATE_FABRIC_LOGO. * Currently this assumes to be called with fnic lock held. */ void fdls_send_fabric_logo(struct fnic_iport_s *iport) { uint8_t *frame; struct fc_std_logo *plogo; struct fnic *fnic = iport->fnic; uint8_t d_id[3]; uint16_t oxid; uint16_t frame_size = FNIC_ETH_FCOE_HDRS_OFFSET + sizeof(struct fc_std_logo); frame = fdls_alloc_frame(iport); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame to send fabric LOGO"); return; } plogo = (struct fc_std_logo *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); fdls_init_logo_frame(frame, iport); oxid = fdls_alloc_oxid(iport, FNIC_FRAME_TYPE_FABRIC_LOGO, &iport->active_oxid_fabric_req); if (oxid == FNIC_UNASSIGNED_OXID) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: Failed to allocate OXID to send fabric LOGO", iport->fcid); mempool_free(frame, fnic->frame_pool); return; } FNIC_STD_SET_OX_ID(plogo->fchdr, oxid); hton24(d_id, FC_FID_FLOGI); FNIC_STD_SET_D_ID(plogo->fchdr, d_id); iport->fabric.flags &= ~FNIC_FDLS_FABRIC_ABORT_ISSUED; FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS send fabric LOGO with oxid: 0x%x", iport->fcid, oxid); fnic_send_fcoe_frame(iport, frame, frame_size); fdls_start_fabric_timer(iport, 2 * iport->e_d_tov); } /** * fdls_tgt_logout - Send plogo to the remote port * @iport: Handle to fnic iport * @tport: Handle to remote port * * This function does not change or check the fabric/tport state. * It the caller's responsibility to set the appropriate tport/fabric * state when this is called. Normally that is fdls_tgt_state_plogo. * This could be used to send plogo to nameserver process * also not just target processes */ void fdls_tgt_logout(struct fnic_iport_s *iport, struct fnic_tport_s *tport) { uint8_t *frame; struct fc_std_logo *plogo; struct fnic *fnic = iport->fnic; uint8_t d_id[3]; uint16_t oxid; uint16_t frame_size = FNIC_ETH_FCOE_HDRS_OFFSET + sizeof(struct fc_std_logo); frame = fdls_alloc_frame(iport); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame to send fabric LOGO"); return; } plogo = (struct fc_std_logo *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); fdls_init_logo_frame(frame, iport); oxid = fdls_alloc_oxid(iport, FNIC_FRAME_TYPE_TGT_LOGO, &tport->active_oxid); if (oxid == FNIC_UNASSIGNED_OXID) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: Failed to allocate OXID to send tgt LOGO", iport->fcid); mempool_free(frame, fnic->frame_pool); return; } FNIC_STD_SET_OX_ID(plogo->fchdr, oxid); hton24(d_id, tport->fcid); FNIC_STD_SET_D_ID(plogo->fchdr, d_id); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS send tgt LOGO with oxid: 0x%x", iport->fcid, oxid); fnic_send_fcoe_frame(iport, frame, frame_size); atomic64_inc(&iport->iport_stats.tport_logo_sent); } static void fdls_tgt_discovery_start(struct fnic_iport_s *iport) { struct fnic_tport_s *tport, *next; u32 old_link_down_cnt = iport->fnic->link_down_cnt; struct fnic *fnic = iport->fnic; FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: Starting FDLS target discovery", iport->fcid); list_for_each_entry_safe(tport, next, &iport->tport_list, links) { if ((old_link_down_cnt != iport->fnic->link_down_cnt) || (iport->state != FNIC_IPORT_STATE_READY)) { break; } /* if we marked the tport as deleted due to GPN_FT * We should not send ADISC anymore */ if ((tport->state == FDLS_TGT_STATE_OFFLINING) || (tport->state == FDLS_TGT_STATE_OFFLINE)) continue; /* For tports which have received RSCN */ if (tport->flags & FNIC_FDLS_TPORT_SEND_ADISC) { tport->retry_counter = 0; fdls_set_tport_state(tport, FDLS_TGT_STATE_ADISC); tport->flags &= ~FNIC_FDLS_TPORT_SEND_ADISC; fdls_send_tgt_adisc(iport, tport); continue; } if (fdls_get_tport_state(tport) != FDLS_TGT_STATE_INIT) { /* Not a new port, skip */ continue; } tport->retry_counter = 0; fdls_set_tport_state(tport, FDLS_TGT_STATE_PLOGI); fdls_send_tgt_plogi(iport, tport); } fdls_set_state((&iport->fabric), FDLS_STATE_TGT_DISCOVERY); } /* * Function to restart the IT nexus if we received any out of * sequence PLOGI/PRLI response from the target. * The memory for the new tport structure is allocated * inside fdls_create_tport and added to the iport's tport list. * This will get freed later during tport_offline/linkdown * or module unload. The new_tport pointer will go out of scope * safely since the memory it is * pointing to it will be freed later */ static void fdls_target_restart_nexus(struct fnic_tport_s *tport) { struct fnic_iport_s *iport = tport->iport; struct fnic_tport_s *new_tport = NULL; uint32_t fcid; uint64_t wwpn; int nexus_restart_count; struct fnic *fnic = iport->fnic; bool retval = true; FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "tport fcid: 0x%x state: %d restart_count: %d", tport->fcid, tport->state, tport->nexus_restart_count); fcid = tport->fcid; wwpn = tport->wwpn; nexus_restart_count = tport->nexus_restart_count; retval = fdls_delete_tport(iport, tport); if (retval != true) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Error deleting tport: 0x%x", fcid); return; } if (nexus_restart_count >= FNIC_TPORT_MAX_NEXUS_RESTART) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Exceeded nexus restart retries tport: 0x%x", fcid); return; } /* * Allocate memory for the new tport and add it to * iport's tport list. * This memory will be freed during tport_offline/linkdown * or module unload. The pointer new_tport is safe to go * out of scope when this function returns, since the memory * it is pointing to is guaranteed to be freed later * as mentioned above. */ new_tport = fdls_create_tport(iport, fcid, wwpn); if (!new_tport) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Error creating new tport: 0x%x", fcid); return; } new_tport->nexus_restart_count = nexus_restart_count + 1; fdls_send_tgt_plogi(iport, new_tport); fdls_set_tport_state(new_tport, FDLS_TGT_STATE_PLOGI); } struct fnic_tport_s *fnic_find_tport_by_fcid(struct fnic_iport_s *iport, uint32_t fcid) { struct fnic_tport_s *tport, *next; list_for_each_entry_safe(tport, next, &(iport->tport_list), links) { if ((tport->fcid == fcid) && !(tport->flags & FNIC_FDLS_TPORT_TERMINATING)) return tport; } return NULL; } static struct fnic_tport_s *fdls_create_tport(struct fnic_iport_s *iport, uint32_t fcid, uint64_t wwpn) { struct fnic_tport_s *tport; struct fnic *fnic = iport->fnic; FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "FDLS create tport: fcid: 0x%x wwpn: 0x%llx", fcid, wwpn); tport = kzalloc(sizeof(struct fnic_tport_s), GFP_ATOMIC); if (!tport) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Memory allocation failure while creating tport: 0x%x\n", fcid); return NULL; } tport->max_payload_size = FNIC_FCOE_MAX_FRAME_SZ; tport->r_a_tov = FC_DEF_R_A_TOV; tport->e_d_tov = FC_DEF_E_D_TOV; tport->fcid = fcid; tport->wwpn = wwpn; tport->iport = iport; FNIC_FCS_DBG(KERN_DEBUG, fnic->host, fnic->fnic_num, "Need to setup tport timer callback"); timer_setup(&tport->retry_timer, fdls_tport_timer_callback, 0); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Added tport 0x%x", tport->fcid); fdls_set_tport_state(tport, FDLS_TGT_STATE_INIT); list_add_tail(&tport->links, &iport->tport_list); atomic_set(&tport->in_flight, 0); return tport; } struct fnic_tport_s *fnic_find_tport_by_wwpn(struct fnic_iport_s *iport, uint64_t wwpn) { struct fnic_tport_s *tport, *next; list_for_each_entry_safe(tport, next, &(iport->tport_list), links) { if ((tport->wwpn == wwpn) && !(tport->flags & FNIC_FDLS_TPORT_TERMINATING)) return tport; } return NULL; } static void fnic_fdmi_attr_set(void *attr_start, u16 type, u16 len, void *data, u32 *off) { u16 size = len + FC_FDMI_ATTR_ENTRY_HEADER_LEN; struct fc_fdmi_attr_entry *fdmi_attr = (struct fc_fdmi_attr_entry *) ((u8 *)attr_start + *off); put_unaligned_be16(type, &fdmi_attr->type); put_unaligned_be16(size, &fdmi_attr->len); memcpy(fdmi_attr->value, data, len); *off += size; } static void fdls_fdmi_register_hba(struct fnic_iport_s *iport) { uint8_t *frame; struct fc_std_fdmi_rhba *prhba; struct fc_fdmi_attr_entry *fdmi_attr; uint8_t fcid[3]; int err; struct fnic *fnic = iport->fnic; struct vnic_devcmd_fw_info *fw_info = NULL; uint16_t oxid; u32 attr_off_bytes, len; u8 data[64]; uint16_t frame_size = FNIC_ETH_FCOE_HDRS_OFFSET; frame = fdls_alloc_frame(iport); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame to send FDMI RHBA"); return; } prhba = (struct fc_std_fdmi_rhba *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); *prhba = (struct fc_std_fdmi_rhba) { .fchdr = { .fh_r_ctl = FC_RCTL_DD_UNSOL_CTL, .fh_d_id = {0xFF, 0XFF, 0XFA}, .fh_type = FC_TYPE_CT, .fh_f_ctl = {FNIC_ELS_REQ_FCTL, 0, 0}, .fh_rx_id = cpu_to_be16(FNIC_UNASSIGNED_RXID) }, .fc_std_ct_hdr = { .ct_rev = FC_CT_REV, .ct_fs_type = FC_FST_MGMT, .ct_fs_subtype = FC_FDMI_SUBTYPE, .ct_cmd = cpu_to_be16(FC_FDMI_RHBA) }, }; hton24(fcid, iport->fcid); FNIC_STD_SET_S_ID(prhba->fchdr, fcid); oxid = fdls_alloc_oxid(iport, FNIC_FRAME_TYPE_FDMI_RHBA, &iport->active_oxid_fdmi_rhba); if (oxid == FNIC_UNASSIGNED_OXID) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: Failed to allocate OXID to send FDMI RHBA", iport->fcid); mempool_free(frame, fnic->frame_pool); return; } FNIC_STD_SET_OX_ID(prhba->fchdr, oxid); put_unaligned_be64(iport->wwpn, &prhba->rhba.hbaid.id); put_unaligned_be32(FNIC_FDMI_NUM_PORTS, &prhba->rhba.port.numport); put_unaligned_be64(iport->wwpn, &prhba->rhba.port.port[0].portname); put_unaligned_be32(FNIC_FDMI_NUM_HBA_ATTRS, &prhba->rhba.hba_attrs.numattrs); fdmi_attr = prhba->rhba.hba_attrs.attr; attr_off_bytes = 0; put_unaligned_be64(iport->wwnn, data); fnic_fdmi_attr_set(fdmi_attr, FNIC_FDMI_TYPE_NODE_NAME, FNIC_FDMI_NN_LEN, data, &attr_off_bytes); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "NN set, off=%d", attr_off_bytes); strscpy_pad(data, FNIC_FDMI_MANUFACTURER, FNIC_FDMI_MANU_LEN); fnic_fdmi_attr_set(fdmi_attr, FNIC_FDMI_TYPE_MANUFACTURER, FNIC_FDMI_MANU_LEN, data, &attr_off_bytes); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "MFG set <%s>, off=%d", data, attr_off_bytes); err = vnic_dev_fw_info(fnic->vdev, &fw_info); if (!err) { strscpy_pad(data, fw_info->hw_serial_number, FNIC_FDMI_SERIAL_LEN); fnic_fdmi_attr_set(fdmi_attr, FNIC_FDMI_TYPE_SERIAL_NUMBER, FNIC_FDMI_SERIAL_LEN, data, &attr_off_bytes); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "SERIAL set <%s>, off=%d", data, attr_off_bytes); } if (fnic->subsys_desc_len >= FNIC_FDMI_MODEL_LEN) fnic->subsys_desc_len = FNIC_FDMI_MODEL_LEN - 1; strscpy_pad(data, fnic->subsys_desc, FNIC_FDMI_MODEL_LEN); data[FNIC_FDMI_MODEL_LEN - 1] = 0; fnic_fdmi_attr_set(fdmi_attr, FNIC_FDMI_TYPE_MODEL, FNIC_FDMI_MODEL_LEN, data, &attr_off_bytes); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "MODEL set <%s>, off=%d", data, attr_off_bytes); strscpy_pad(data, FNIC_FDMI_MODEL_DESCRIPTION, FNIC_FDMI_MODEL_DES_LEN); fnic_fdmi_attr_set(fdmi_attr, FNIC_FDMI_TYPE_MODEL_DES, FNIC_FDMI_MODEL_DES_LEN, data, &attr_off_bytes); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "MODEL_DESC set <%s>, off=%d", data, attr_off_bytes); if (!err) { strscpy_pad(data, fw_info->hw_version, FNIC_FDMI_HW_VER_LEN); fnic_fdmi_attr_set(fdmi_attr, FNIC_FDMI_TYPE_HARDWARE_VERSION, FNIC_FDMI_HW_VER_LEN, data, &attr_off_bytes); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "HW_VER set <%s>, off=%d", data, attr_off_bytes); } strscpy_pad(data, DRV_VERSION, FNIC_FDMI_DR_VER_LEN); fnic_fdmi_attr_set(fdmi_attr, FNIC_FDMI_TYPE_DRIVER_VERSION, FNIC_FDMI_DR_VER_LEN, data, &attr_off_bytes); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "DRV_VER set <%s>, off=%d", data, attr_off_bytes); strscpy_pad(data, "N/A", FNIC_FDMI_ROM_VER_LEN); fnic_fdmi_attr_set(fdmi_attr, FNIC_FDMI_TYPE_ROM_VERSION, FNIC_FDMI_ROM_VER_LEN, data, &attr_off_bytes); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "ROM_VER set <%s>, off=%d", data, attr_off_bytes); if (!err) { strscpy_pad(data, fw_info->fw_version, FNIC_FDMI_FW_VER_LEN); fnic_fdmi_attr_set(fdmi_attr, FNIC_FDMI_TYPE_FIRMWARE_VERSION, FNIC_FDMI_FW_VER_LEN, data, &attr_off_bytes); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "FW_VER set <%s>, off=%d", data, attr_off_bytes); } len = sizeof(struct fc_std_fdmi_rhba) + attr_off_bytes; frame_size += len; FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS send FDMI RHBA with oxid: 0x%x fs: %d", iport->fcid, oxid, frame_size); fnic_send_fcoe_frame(iport, frame, frame_size); iport->fabric.fdmi_pending |= FDLS_FDMI_REG_HBA_PENDING; } static void fdls_fdmi_register_pa(struct fnic_iport_s *iport) { uint8_t *frame; struct fc_std_fdmi_rpa *prpa; struct fc_fdmi_attr_entry *fdmi_attr; uint8_t fcid[3]; struct fnic *fnic = iport->fnic; u32 port_speed_bm; u32 port_speed = vnic_dev_port_speed(fnic->vdev); uint16_t oxid; u32 attr_off_bytes, len; u8 tmp_data[16], data[64]; uint16_t frame_size = FNIC_ETH_FCOE_HDRS_OFFSET; frame = fdls_alloc_frame(iport); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame to send FDMI RPA"); return; } prpa = (struct fc_std_fdmi_rpa *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); *prpa = (struct fc_std_fdmi_rpa) { .fchdr = { .fh_r_ctl = FC_RCTL_DD_UNSOL_CTL, .fh_d_id = {0xFF, 0xFF, 0xFA}, .fh_type = FC_TYPE_CT, .fh_f_ctl = {FNIC_ELS_REQ_FCTL, 0, 0}, .fh_rx_id = cpu_to_be16(FNIC_UNASSIGNED_RXID) }, .fc_std_ct_hdr = { .ct_rev = FC_CT_REV, .ct_fs_type = FC_FST_MGMT, .ct_fs_subtype = FC_FDMI_SUBTYPE, .ct_cmd = cpu_to_be16(FC_FDMI_RPA) }, }; hton24(fcid, iport->fcid); FNIC_STD_SET_S_ID(prpa->fchdr, fcid); oxid = fdls_alloc_oxid(iport, FNIC_FRAME_TYPE_FDMI_RPA, &iport->active_oxid_fdmi_rpa); if (oxid == FNIC_UNASSIGNED_OXID) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: Failed to allocate OXID to send FDMI RPA", iport->fcid); mempool_free(frame, fnic->frame_pool); return; } FNIC_STD_SET_OX_ID(prpa->fchdr, oxid); put_unaligned_be64(iport->wwpn, &prpa->rpa.port.portname); put_unaligned_be32(FNIC_FDMI_NUM_PORT_ATTRS, &prpa->rpa.hba_attrs.numattrs); /* MDS does not support GIGE speed. * Bit shift standard definitions from scsi_transport_fc.h to * match FC spec. */ switch (port_speed) { case DCEM_PORTSPEED_10G: case DCEM_PORTSPEED_20G: /* There is no bit for 20G */ port_speed_bm = FC_PORTSPEED_10GBIT << PORT_SPEED_BIT_14; break; case DCEM_PORTSPEED_25G: port_speed_bm = FC_PORTSPEED_25GBIT << PORT_SPEED_BIT_8; break; case DCEM_PORTSPEED_40G: case DCEM_PORTSPEED_4x10G: port_speed_bm = FC_PORTSPEED_40GBIT << PORT_SPEED_BIT_9; break; case DCEM_PORTSPEED_100G: port_speed_bm = FC_PORTSPEED_100GBIT << PORT_SPEED_BIT_8; break; default: port_speed_bm = FC_PORTSPEED_1GBIT << PORT_SPEED_BIT_15; break; } attr_off_bytes = 0; fdmi_attr = prpa->rpa.hba_attrs.attr; put_unaligned_be64(iport->wwnn, data); memset(data, 0, FNIC_FDMI_FC4_LEN); data[2] = 1; fnic_fdmi_attr_set(fdmi_attr, FNIC_FDMI_TYPE_FC4_TYPES, FNIC_FDMI_FC4_LEN, data, &attr_off_bytes); put_unaligned_be32(port_speed_bm, data); fnic_fdmi_attr_set(fdmi_attr, FNIC_FDMI_TYPE_SUPPORTED_SPEEDS, FNIC_FDMI_SUPP_SPEED_LEN, data, &attr_off_bytes); put_unaligned_be32(port_speed_bm, data); fnic_fdmi_attr_set(fdmi_attr, FNIC_FDMI_TYPE_CURRENT_SPEED, FNIC_FDMI_CUR_SPEED_LEN, data, &attr_off_bytes); put_unaligned_be32(FNIC_FDMI_MFS, data); fnic_fdmi_attr_set(fdmi_attr, FNIC_FDMI_TYPE_MAX_FRAME_SIZE, FNIC_FDMI_MFS_LEN, data, &attr_off_bytes); snprintf(tmp_data, FNIC_FDMI_OS_NAME_LEN - 1, "host%d", fnic->host->host_no); strscpy_pad(data, tmp_data, FNIC_FDMI_OS_NAME_LEN); data[FNIC_FDMI_OS_NAME_LEN - 1] = 0; fnic_fdmi_attr_set(fdmi_attr, FNIC_FDMI_TYPE_OS_NAME, FNIC_FDMI_OS_NAME_LEN, data, &attr_off_bytes); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "OS name set <%s>, off=%d", data, attr_off_bytes); sprintf(fc_host_system_hostname(fnic->host), "%s", utsname()->nodename); strscpy_pad(data, fc_host_system_hostname(fnic->host), FNIC_FDMI_HN_LEN); data[FNIC_FDMI_HN_LEN - 1] = 0; fnic_fdmi_attr_set(fdmi_attr, FNIC_FDMI_TYPE_HOST_NAME, FNIC_FDMI_HN_LEN, data, &attr_off_bytes); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Host name set <%s>, off=%d", data, attr_off_bytes); len = sizeof(struct fc_std_fdmi_rpa) + attr_off_bytes; frame_size += len; FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS send FDMI RPA with oxid: 0x%x fs: %d", iport->fcid, oxid, frame_size); fnic_send_fcoe_frame(iport, frame, frame_size); iport->fabric.fdmi_pending |= FDLS_FDMI_RPA_PENDING; } void fdls_fabric_timer_callback(struct timer_list *t) { struct fnic_fdls_fabric_s *fabric = from_timer(fabric, t, retry_timer); struct fnic_iport_s *iport = container_of(fabric, struct fnic_iport_s, fabric); struct fnic *fnic = iport->fnic; unsigned long flags; FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "tp: %d fab state: %d fab retry counter: %d max_flogi_retries: %d", iport->fabric.timer_pending, iport->fabric.state, iport->fabric.retry_counter, iport->max_flogi_retries); spin_lock_irqsave(&fnic->fnic_lock, flags); if (!iport->fabric.timer_pending) { spin_unlock_irqrestore(&fnic->fnic_lock, flags); return; } if (iport->fabric.del_timer_inprogress) { iport->fabric.del_timer_inprogress = 0; spin_unlock_irqrestore(&fnic->fnic_lock, flags); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "fabric_del_timer inprogress(%d). Skip timer cb", iport->fabric.del_timer_inprogress); return; } iport->fabric.timer_pending = 0; /* The fabric state indicates which frames have time out, and we retry */ switch (iport->fabric.state) { case FDLS_STATE_FABRIC_FLOGI: /* Flogi received a LS_RJT with busy we retry from here */ if ((iport->fabric.flags & FNIC_FDLS_RETRY_FRAME) && (iport->fabric.retry_counter < iport->max_flogi_retries)) { iport->fabric.flags &= ~FNIC_FDLS_RETRY_FRAME; fdls_send_fabric_flogi(iport); } else if (!(iport->fabric.flags & FNIC_FDLS_FABRIC_ABORT_ISSUED)) { /* Flogi has time out 2*ed_tov send abts */ fdls_send_fabric_abts(iport); } else { /* ABTS has timed out * Mark the OXID to be freed after 2 * r_a_tov and retry the req */ fdls_schedule_oxid_free(iport, &iport->active_oxid_fabric_req); if (iport->fabric.retry_counter < iport->max_flogi_retries) { iport->fabric.flags &= ~FNIC_FDLS_FABRIC_ABORT_ISSUED; fdls_send_fabric_flogi(iport); } else FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Exceeded max FLOGI retries"); } break; case FDLS_STATE_FABRIC_PLOGI: /* Plogi received a LS_RJT with busy we retry from here */ if ((iport->fabric.flags & FNIC_FDLS_RETRY_FRAME) && (iport->fabric.retry_counter < iport->max_plogi_retries)) { iport->fabric.flags &= ~FNIC_FDLS_RETRY_FRAME; fdls_send_fabric_plogi(iport); } else if (!(iport->fabric.flags & FNIC_FDLS_FABRIC_ABORT_ISSUED)) { /* Plogi has timed out 2*ed_tov send abts */ fdls_send_fabric_abts(iport); } else { /* ABTS has timed out * Mark the OXID to be freed after 2 * r_a_tov and retry the req */ fdls_schedule_oxid_free(iport, &iport->active_oxid_fabric_req); if (iport->fabric.retry_counter < iport->max_plogi_retries) { iport->fabric.flags &= ~FNIC_FDLS_FABRIC_ABORT_ISSUED; fdls_send_fabric_plogi(iport); } else FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Exceeded max PLOGI retries"); } break; case FDLS_STATE_RPN_ID: /* Rpn_id received a LS_RJT with busy we retry from here */ if ((iport->fabric.flags & FNIC_FDLS_RETRY_FRAME) && (iport->fabric.retry_counter < FDLS_RETRY_COUNT)) { iport->fabric.flags &= ~FNIC_FDLS_RETRY_FRAME; fdls_send_rpn_id(iport); } else if (!(iport->fabric.flags & FNIC_FDLS_FABRIC_ABORT_ISSUED)) /* RPN has timed out. Send abts */ fdls_send_fabric_abts(iport); else { /* ABTS has timed out */ fdls_schedule_oxid_free(iport, &iport->active_oxid_fabric_req); fnic_fdls_start_plogi(iport); /* go back to fabric Plogi */ } break; case FDLS_STATE_SCR: /* scr received a LS_RJT with busy we retry from here */ if ((iport->fabric.flags & FNIC_FDLS_RETRY_FRAME) && (iport->fabric.retry_counter < FDLS_RETRY_COUNT)) { iport->fabric.flags &= ~FNIC_FDLS_RETRY_FRAME; fdls_send_scr(iport); } else if (!(iport->fabric.flags & FNIC_FDLS_FABRIC_ABORT_ISSUED)) /* scr has timed out. Send abts */ fdls_send_fabric_abts(iport); else { /* ABTS has timed out */ fdls_schedule_oxid_free(iport, &iport->active_oxid_fabric_req); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "ABTS timed out. Starting PLOGI: %p", iport); fnic_fdls_start_plogi(iport); } break; case FDLS_STATE_REGISTER_FC4_TYPES: /* scr received a LS_RJT with busy we retry from here */ if ((iport->fabric.flags & FNIC_FDLS_RETRY_FRAME) && (iport->fabric.retry_counter < FDLS_RETRY_COUNT)) { iport->fabric.flags &= ~FNIC_FDLS_RETRY_FRAME; fdls_send_register_fc4_types(iport); } else if (!(iport->fabric.flags & FNIC_FDLS_FABRIC_ABORT_ISSUED)) { /* RFT_ID timed out send abts */ fdls_send_fabric_abts(iport); } else { /* ABTS has timed out */ fdls_schedule_oxid_free(iport, &iport->active_oxid_fabric_req); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "ABTS timed out. Starting PLOGI: %p", iport); fnic_fdls_start_plogi(iport); /* go back to fabric Plogi */ } break; case FDLS_STATE_REGISTER_FC4_FEATURES: /* scr received a LS_RJT with busy we retry from here */ if ((iport->fabric.flags & FNIC_FDLS_RETRY_FRAME) && (iport->fabric.retry_counter < FDLS_RETRY_COUNT)) { iport->fabric.flags &= ~FNIC_FDLS_RETRY_FRAME; fdls_send_register_fc4_features(iport); } else if (!(iport->fabric.flags & FNIC_FDLS_FABRIC_ABORT_ISSUED)) /* SCR has timed out. Send abts */ fdls_send_fabric_abts(iport); else { /* ABTS has timed out */ fdls_schedule_oxid_free(iport, &iport->active_oxid_fabric_req); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "ABTS timed out. Starting PLOGI %p", iport); fnic_fdls_start_plogi(iport); /* go back to fabric Plogi */ } break; case FDLS_STATE_RSCN_GPN_FT: case FDLS_STATE_SEND_GPNFT: case FDLS_STATE_GPN_FT: /* GPN_FT received a LS_RJT with busy we retry from here */ if ((iport->fabric.flags & FNIC_FDLS_RETRY_FRAME) && (iport->fabric.retry_counter < FDLS_RETRY_COUNT)) { iport->fabric.flags &= ~FNIC_FDLS_RETRY_FRAME; fdls_send_gpn_ft(iport, iport->fabric.state); } else if (!(iport->fabric.flags & FNIC_FDLS_FABRIC_ABORT_ISSUED)) { /* gpn_ft has timed out. Send abts */ fdls_send_fabric_abts(iport); } else { /* ABTS has timed out */ fdls_schedule_oxid_free(iport, &iport->active_oxid_fabric_req); if (iport->fabric.retry_counter < FDLS_RETRY_COUNT) { fdls_send_gpn_ft(iport, iport->fabric.state); } else { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "ABTS timeout for fabric GPN_FT. Check name server: %p", iport); } } break; default: break; } spin_unlock_irqrestore(&fnic->fnic_lock, flags); } void fdls_fdmi_timer_callback(struct timer_list *t) { struct fnic_fdls_fabric_s *fabric = from_timer(fabric, t, fdmi_timer); struct fnic_iport_s *iport = container_of(fabric, struct fnic_iport_s, fabric); struct fnic *fnic = iport->fnic; unsigned long flags; spin_lock_irqsave(&fnic->fnic_lock, flags); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "fdmi timer callback : 0x%x\n", iport->fabric.fdmi_pending); if (!iport->fabric.fdmi_pending) { /* timer expired after fdmi responses received. */ spin_unlock_irqrestore(&fnic->fnic_lock, flags); return; } FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "fdmi timer callback : 0x%x\n", iport->fabric.fdmi_pending); /* if not abort pending, send an abort */ if (!(iport->fabric.fdmi_pending & FDLS_FDMI_ABORT_PENDING)) { fdls_send_fdmi_abts(iport); spin_unlock_irqrestore(&fnic->fnic_lock, flags); return; } FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "fdmi timer callback : 0x%x\n", iport->fabric.fdmi_pending); /* ABTS pending for an active fdmi request that is pending. * That means FDMI ABTS timed out * Schedule to free the OXID after 2*r_a_tov and proceed */ if (iport->fabric.fdmi_pending & FDLS_FDMI_PLOGI_PENDING) { fdls_schedule_oxid_free(iport, &iport->active_oxid_fdmi_plogi); } else { if (iport->fabric.fdmi_pending & FDLS_FDMI_REG_HBA_PENDING) fdls_schedule_oxid_free(iport, &iport->active_oxid_fdmi_rhba); if (iport->fabric.fdmi_pending & FDLS_FDMI_RPA_PENDING) fdls_schedule_oxid_free(iport, &iport->active_oxid_fdmi_rpa); } FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "fdmi timer callback : 0x%x\n", iport->fabric.fdmi_pending); iport->fabric.fdmi_pending = 0; /* If max retries not exhaused, start over from fdmi plogi */ if (iport->fabric.fdmi_retry < FDLS_FDMI_MAX_RETRY) { iport->fabric.fdmi_retry++; FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "retry fdmi timer %d", iport->fabric.fdmi_retry); fdls_send_fdmi_plogi(iport); } FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "fdmi timer callback : 0x%x\n", iport->fabric.fdmi_pending); spin_unlock_irqrestore(&fnic->fnic_lock, flags); } static void fdls_send_delete_tport_msg(struct fnic_tport_s *tport) { struct fnic_iport_s *iport = (struct fnic_iport_s *) tport->iport; struct fnic *fnic = iport->fnic; struct fnic_tport_event_s *tport_del_evt; tport_del_evt = kzalloc(sizeof(struct fnic_tport_event_s), GFP_ATOMIC); if (!tport_del_evt) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Failed to allocate memory for tport event fcid: 0x%x", tport->fcid); return; } tport_del_evt->event = TGT_EV_TPORT_DELETE; tport_del_evt->arg1 = (void *) tport; list_add_tail(&tport_del_evt->links, &fnic->tport_event_list); queue_work(fnic_event_queue, &fnic->tport_work); } static void fdls_tport_timer_callback(struct timer_list *t) { struct fnic_tport_s *tport = from_timer(tport, t, retry_timer); struct fnic_iport_s *iport = (struct fnic_iport_s *) tport->iport; struct fnic *fnic = iport->fnic; uint16_t oxid; unsigned long flags; spin_lock_irqsave(&fnic->fnic_lock, flags); if (!tport->timer_pending) { spin_unlock_irqrestore(&fnic->fnic_lock, flags); return; } if (iport->state != FNIC_IPORT_STATE_READY) { spin_unlock_irqrestore(&fnic->fnic_lock, flags); return; } if (tport->del_timer_inprogress) { tport->del_timer_inprogress = 0; spin_unlock_irqrestore(&fnic->fnic_lock, flags); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "tport_del_timer inprogress. Skip timer cb tport fcid: 0x%x\n", tport->fcid); return; } FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "tport fcid: 0x%x timer pending: %d state: %d retry counter: %d", tport->fcid, tport->timer_pending, tport->state, tport->retry_counter); tport->timer_pending = 0; oxid = tport->active_oxid; /* We retry plogi/prli/adisc frames depending on the tport state */ switch (tport->state) { case FDLS_TGT_STATE_PLOGI: /* PLOGI frame received a LS_RJT with busy, we retry from here */ if ((tport->flags & FNIC_FDLS_RETRY_FRAME) && (tport->retry_counter < iport->max_plogi_retries)) { tport->flags &= ~FNIC_FDLS_RETRY_FRAME; fdls_send_tgt_plogi(iport, tport); } else if (!(tport->flags & FNIC_FDLS_TGT_ABORT_ISSUED)) { /* Plogi frame has timed out, send abts */ fdls_send_tport_abts(iport, tport); } else if (tport->retry_counter < iport->max_plogi_retries) { /* * ABTS has timed out */ fdls_schedule_oxid_free(iport, &tport->active_oxid); fdls_send_tgt_plogi(iport, tport); } else { /* exceeded plogi retry count */ fdls_schedule_oxid_free(iport, &tport->active_oxid); fdls_send_delete_tport_msg(tport); } break; case FDLS_TGT_STATE_PRLI: /* PRLI received a LS_RJT with busy , hence we retry from here */ if ((tport->flags & FNIC_FDLS_RETRY_FRAME) && (tport->retry_counter < FDLS_RETRY_COUNT)) { tport->flags &= ~FNIC_FDLS_RETRY_FRAME; fdls_send_tgt_prli(iport, tport); } else if (!(tport->flags & FNIC_FDLS_TGT_ABORT_ISSUED)) { /* PRLI has time out, send abts */ fdls_send_tport_abts(iport, tport); } else { /* ABTS has timed out for prli, we go back to PLOGI */ fdls_schedule_oxid_free(iport, &tport->active_oxid); fdls_send_tgt_plogi(iport, tport); fdls_set_tport_state(tport, FDLS_TGT_STATE_PLOGI); } break; case FDLS_TGT_STATE_ADISC: /* ADISC timed out send an ABTS */ if (!(tport->flags & FNIC_FDLS_TGT_ABORT_ISSUED)) { fdls_send_tport_abts(iport, tport); } else if ((tport->flags & FNIC_FDLS_TGT_ABORT_ISSUED) && (tport->retry_counter < FDLS_RETRY_COUNT)) { /* * ABTS has timed out */ fdls_schedule_oxid_free(iport, &tport->active_oxid); fdls_send_tgt_adisc(iport, tport); } else { /* exceeded retry count */ fdls_schedule_oxid_free(iport, &tport->active_oxid); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "ADISC not responding. Deleting target port: 0x%x", tport->fcid); fdls_send_delete_tport_msg(tport); } break; default: FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "oxid: 0x%x Unknown tport state: 0x%x", oxid, tport->state); break; } spin_unlock_irqrestore(&fnic->fnic_lock, flags); } static void fnic_fdls_start_flogi(struct fnic_iport_s *iport) { iport->fabric.retry_counter = 0; fdls_send_fabric_flogi(iport); fdls_set_state((&iport->fabric), FDLS_STATE_FABRIC_FLOGI); iport->fabric.flags = 0; } static void fnic_fdls_start_plogi(struct fnic_iport_s *iport) { iport->fabric.retry_counter = 0; fdls_send_fabric_plogi(iport); fdls_set_state((&iport->fabric), FDLS_STATE_FABRIC_PLOGI); iport->fabric.flags &= ~FNIC_FDLS_FABRIC_ABORT_ISSUED; if ((fnic_fdmi_support == 1) && (!(iport->flags & FNIC_FDMI_ACTIVE))) { /* we can do FDMI at the same time */ iport->fabric.fdmi_retry = 0; timer_setup(&iport->fabric.fdmi_timer, fdls_fdmi_timer_callback, 0); fdls_send_fdmi_plogi(iport); iport->flags |= FNIC_FDMI_ACTIVE; } } static void fdls_process_tgt_adisc_rsp(struct fnic_iport_s *iport, struct fc_frame_header *fchdr) { uint32_t tgt_fcid; struct fnic_tport_s *tport; uint8_t *fcid; uint64_t frame_wwnn; uint64_t frame_wwpn; uint16_t oxid; struct fc_std_els_adisc *adisc_rsp = (struct fc_std_els_adisc *)fchdr; struct fc_std_els_rjt_rsp *els_rjt = (struct fc_std_els_rjt_rsp *)fchdr; struct fnic *fnic = iport->fnic; fcid = FNIC_STD_GET_S_ID(fchdr); tgt_fcid = ntoh24(fcid); tport = fnic_find_tport_by_fcid(iport, tgt_fcid); if (!tport) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Tgt ADISC response tport not found: 0x%x", tgt_fcid); return; } if ((iport->state != FNIC_IPORT_STATE_READY) || (tport->state != FDLS_TGT_STATE_ADISC) || (tport->flags & FNIC_FDLS_TGT_ABORT_ISSUED)) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Dropping this ADISC response"); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "iport state: %d tport state: %d Is abort issued on PRLI? %d", iport->state, tport->state, (tport->flags & FNIC_FDLS_TGT_ABORT_ISSUED)); return; } if (FNIC_STD_GET_OX_ID(fchdr) != tport->active_oxid) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Dropping frame from target: 0x%x", tgt_fcid); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Reason: Stale ADISC/Aborted ADISC/OOO frame delivery"); return; } oxid = FNIC_STD_GET_OX_ID(fchdr); fdls_free_oxid(iport, oxid, &tport->active_oxid); switch (adisc_rsp->els.adisc_cmd) { case ELS_LS_ACC: atomic64_inc(&iport->iport_stats.tport_adisc_ls_accepts); if (tport->timer_pending) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "tport 0x%p Canceling fabric disc timer\n", tport); fnic_del_tport_timer_sync(fnic, tport); } tport->timer_pending = 0; tport->retry_counter = 0; frame_wwnn = get_unaligned_be64(&adisc_rsp->els.adisc_wwnn); frame_wwpn = get_unaligned_be64(&adisc_rsp->els.adisc_wwpn); if ((frame_wwnn == tport->wwnn) && (frame_wwpn == tport->wwpn)) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "ADISC accepted from target: 0x%x. Target logged in", tgt_fcid); fdls_set_tport_state(tport, FDLS_TGT_STATE_READY); } else { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Error mismatch frame: ADISC"); } break; case ELS_LS_RJT: atomic64_inc(&iport->iport_stats.tport_adisc_ls_rejects); if (((els_rjt->rej.er_reason == ELS_RJT_BUSY) || (els_rjt->rej.er_reason == ELS_RJT_UNAB)) && (tport->retry_counter < FDLS_RETRY_COUNT)) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "ADISC ret ELS_LS_RJT BUSY. Retry from timer routine: 0x%x", tgt_fcid); /* Retry ADISC again from the timer routine. */ tport->flags |= FNIC_FDLS_RETRY_FRAME; } else { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "ADISC returned ELS_LS_RJT from target: 0x%x", tgt_fcid); fdls_delete_tport(iport, tport); } break; } } static void fdls_process_tgt_plogi_rsp(struct fnic_iport_s *iport, struct fc_frame_header *fchdr) { uint32_t tgt_fcid; struct fnic_tport_s *tport; uint8_t *fcid; uint16_t oxid; struct fc_std_flogi *plogi_rsp = (struct fc_std_flogi *)fchdr; struct fc_std_els_rjt_rsp *els_rjt = (struct fc_std_els_rjt_rsp *)fchdr; uint16_t max_payload_size; struct fnic *fnic = iport->fnic; fcid = FNIC_STD_GET_S_ID(fchdr); tgt_fcid = ntoh24(fcid); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "FDLS processing target PLOGI response: tgt_fcid: 0x%x", tgt_fcid); tport = fnic_find_tport_by_fcid(iport, tgt_fcid); if (!tport) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "tport not found: 0x%x", tgt_fcid); return; } if ((iport->state != FNIC_IPORT_STATE_READY) || (tport->flags & FNIC_FDLS_TGT_ABORT_ISSUED)) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Dropping frame! iport state: %d tport state: %d", iport->state, tport->state); return; } if (tport->state != FDLS_TGT_STATE_PLOGI) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "PLOGI rsp recvd in wrong state. Drop the frame and restart nexus"); fdls_target_restart_nexus(tport); return; } if (FNIC_STD_GET_OX_ID(fchdr) != tport->active_oxid) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "PLOGI response from target: 0x%x. Dropping frame", tgt_fcid); return; } oxid = FNIC_STD_GET_OX_ID(fchdr); fdls_free_oxid(iport, oxid, &tport->active_oxid); switch (plogi_rsp->els.fl_cmd) { case ELS_LS_ACC: atomic64_inc(&iport->iport_stats.tport_plogi_ls_accepts); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "PLOGI accepted by target: 0x%x", tgt_fcid); break; case ELS_LS_RJT: atomic64_inc(&iport->iport_stats.tport_plogi_ls_rejects); if (((els_rjt->rej.er_reason == ELS_RJT_BUSY) || (els_rjt->rej.er_reason == ELS_RJT_UNAB)) && (tport->retry_counter < iport->max_plogi_retries)) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "PLOGI ret ELS_LS_RJT BUSY. Retry from timer routine: 0x%x", tgt_fcid); /* Retry plogi again from the timer routine. */ tport->flags |= FNIC_FDLS_RETRY_FRAME; return; } FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "PLOGI returned ELS_LS_RJT from target: 0x%x", tgt_fcid); fdls_delete_tport(iport, tport); return; default: atomic64_inc(&iport->iport_stats.tport_plogi_misc_rejects); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "PLOGI not accepted from target fcid: 0x%x", tgt_fcid); return; } FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Found the PLOGI target: 0x%x and state: %d", (unsigned int) tgt_fcid, tport->state); if (tport->timer_pending) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "tport fcid 0x%x: Canceling disc timer\n", tport->fcid); fnic_del_tport_timer_sync(fnic, tport); } tport->timer_pending = 0; tport->wwpn = get_unaligned_be64(&FNIC_LOGI_PORT_NAME(plogi_rsp->els)); tport->wwnn = get_unaligned_be64(&FNIC_LOGI_NODE_NAME(plogi_rsp->els)); /* Learn the Service Params */ /* Max frame size - choose the lowest */ max_payload_size = fnic_fc_plogi_rsp_rdf(iport, plogi_rsp); tport->max_payload_size = min(max_payload_size, iport->max_payload_size); if (tport->max_payload_size < FNIC_MIN_DATA_FIELD_SIZE) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "MFS: tport max frame size below spec bounds: %d", tport->max_payload_size); tport->max_payload_size = FNIC_MIN_DATA_FIELD_SIZE; } FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "MAX frame size: %u iport max_payload_size: %d tport mfs: %d", max_payload_size, iport->max_payload_size, tport->max_payload_size); tport->max_concur_seqs = FNIC_FC_PLOGI_RSP_CONCUR_SEQ(plogi_rsp); tport->retry_counter = 0; fdls_set_tport_state(tport, FDLS_TGT_STATE_PRLI); fdls_send_tgt_prli(iport, tport); } static void fdls_process_tgt_prli_rsp(struct fnic_iport_s *iport, struct fc_frame_header *fchdr) { uint32_t tgt_fcid; struct fnic_tport_s *tport; uint8_t *fcid; uint16_t oxid; struct fc_std_els_prli *prli_rsp = (struct fc_std_els_prli *)fchdr; struct fc_std_els_rjt_rsp *els_rjt = (struct fc_std_els_rjt_rsp *)fchdr; struct fnic_tport_event_s *tport_add_evt; struct fnic *fnic = iport->fnic; bool mismatched_tgt = false; fcid = FNIC_STD_GET_S_ID(fchdr); tgt_fcid = ntoh24(fcid); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "FDLS process tgt PRLI response: 0x%x", tgt_fcid); tport = fnic_find_tport_by_fcid(iport, tgt_fcid); if (!tport) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "tport not found: 0x%x", tgt_fcid); /* Handle or just drop? */ return; } if ((iport->state != FNIC_IPORT_STATE_READY) || (tport->flags & FNIC_FDLS_TGT_ABORT_ISSUED)) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Dropping frame! iport st: %d tport st: %d tport fcid: 0x%x", iport->state, tport->state, tport->fcid); return; } if (tport->state != FDLS_TGT_STATE_PRLI) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "PRLI rsp recvd in wrong state. Drop frame. Restarting nexus"); fdls_target_restart_nexus(tport); return; } if (FNIC_STD_GET_OX_ID(fchdr) != tport->active_oxid) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Dropping PRLI response from target: 0x%x ", tgt_fcid); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Reason: Stale PRLI response/Aborted PDISC/OOO frame delivery"); return; } oxid = FNIC_STD_GET_OX_ID(fchdr); fdls_free_oxid(iport, oxid, &tport->active_oxid); switch (prli_rsp->els_prli.prli_cmd) { case ELS_LS_ACC: atomic64_inc(&iport->iport_stats.tport_prli_ls_accepts); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "PRLI accepted from target: 0x%x", tgt_fcid); if (prli_rsp->sp.spp_type != FC_FC4_TYPE_SCSI) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "mismatched target zoned with FC SCSI initiator: 0x%x", tgt_fcid); mismatched_tgt = true; } if (mismatched_tgt) { fdls_tgt_logout(iport, tport); fdls_delete_tport(iport, tport); return; } break; case ELS_LS_RJT: atomic64_inc(&iport->iport_stats.tport_prli_ls_rejects); if (((els_rjt->rej.er_reason == ELS_RJT_BUSY) || (els_rjt->rej.er_reason == ELS_RJT_UNAB)) && (tport->retry_counter < FDLS_RETRY_COUNT)) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "PRLI ret ELS_LS_RJT BUSY. Retry from timer routine: 0x%x", tgt_fcid); /*Retry Plogi again from the timer routine. */ tport->flags |= FNIC_FDLS_RETRY_FRAME; return; } FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "PRLI returned ELS_LS_RJT from target: 0x%x", tgt_fcid); fdls_tgt_logout(iport, tport); fdls_delete_tport(iport, tport); return; default: atomic64_inc(&iport->iport_stats.tport_prli_misc_rejects); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "PRLI not accepted from target: 0x%x", tgt_fcid); return; } FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Found the PRLI target: 0x%x and state: %d", (unsigned int) tgt_fcid, tport->state); if (tport->timer_pending) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "tport fcid 0x%x: Canceling disc timer\n", tport->fcid); fnic_del_tport_timer_sync(fnic, tport); } tport->timer_pending = 0; /* Learn Service Params */ tport->fcp_csp = be32_to_cpu(prli_rsp->sp.spp_params); tport->retry_counter = 0; if (tport->fcp_csp & FCP_SPPF_RETRY) tport->tgt_flags |= FNIC_FC_RP_FLAGS_RETRY; /* Check if the device plays Target Mode Function */ if (!(tport->fcp_csp & FCP_PRLI_FUNC_TARGET)) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Remote port(0x%x): no target support. Deleting it\n", tgt_fcid); fdls_tgt_logout(iport, tport); fdls_delete_tport(iport, tport); return; } fdls_set_tport_state(tport, FDLS_TGT_STATE_READY); /* Inform the driver about new target added */ tport_add_evt = kzalloc(sizeof(struct fnic_tport_event_s), GFP_ATOMIC); if (!tport_add_evt) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "tport event memory allocation failure: 0x%0x\n", tport->fcid); return; } tport_add_evt->event = TGT_EV_RPORT_ADD; tport_add_evt->arg1 = (void *) tport; FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "iport fcid: 0x%x add tport event fcid: 0x%x\n", tport->fcid, iport->fcid); list_add_tail(&tport_add_evt->links, &fnic->tport_event_list); queue_work(fnic_event_queue, &fnic->tport_work); } static void fdls_process_rff_id_rsp(struct fnic_iport_s *iport, struct fc_frame_header *fchdr) { struct fnic *fnic = iport->fnic; struct fnic_fdls_fabric_s *fdls = &iport->fabric; struct fc_std_rff_id *rff_rsp = (struct fc_std_rff_id *) fchdr; uint16_t rsp; uint8_t reason_code; uint16_t oxid = FNIC_STD_GET_OX_ID(fchdr); if (fdls_get_state(fdls) != FDLS_STATE_REGISTER_FC4_FEATURES) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "RFF_ID resp recvd in state(%d). Dropping.", fdls_get_state(fdls)); return; } if (iport->active_oxid_fabric_req != oxid) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Incorrect OXID in response. state: %d, oxid recvd: 0x%x, active oxid: 0x%x\n", fdls_get_state(fdls), oxid, iport->active_oxid_fabric_req); return; } rsp = FNIC_STD_GET_FC_CT_CMD((&rff_rsp->fc_std_ct_hdr)); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS process RFF ID response: 0x%04x", iport->fcid, (uint32_t) rsp); fdls_free_oxid(iport, oxid, &iport->active_oxid_fabric_req); switch (rsp) { case FC_FS_ACC: if (iport->fabric.timer_pending) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Canceling fabric disc timer %p\n", iport); fnic_del_fabric_timer_sync(fnic); } iport->fabric.timer_pending = 0; fdls->retry_counter = 0; fdls_set_state((&iport->fabric), FDLS_STATE_SCR); fdls_send_scr(iport); break; case FC_FS_RJT: reason_code = rff_rsp->fc_std_ct_hdr.ct_reason; if (((reason_code == FC_FS_RJT_BSY) || (reason_code == FC_FS_RJT_UNABL)) && (fdls->retry_counter < FDLS_RETRY_COUNT)) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "RFF_ID ret ELS_LS_RJT BUSY. Retry from timer routine %p", iport); /* Retry again from the timer routine */ fdls->flags |= FNIC_FDLS_RETRY_FRAME; } else { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "RFF_ID returned ELS_LS_RJT. Halting discovery %p", iport); if (iport->fabric.timer_pending) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Canceling fabric disc timer %p\n", iport); fnic_del_fabric_timer_sync(fnic); } fdls->timer_pending = 0; fdls->retry_counter = 0; } break; default: break; } } static void fdls_process_rft_id_rsp(struct fnic_iport_s *iport, struct fc_frame_header *fchdr) { struct fnic_fdls_fabric_s *fdls = &iport->fabric; struct fc_std_rft_id *rft_rsp = (struct fc_std_rft_id *) fchdr; uint16_t rsp; uint8_t reason_code; struct fnic *fnic = iport->fnic; uint16_t oxid = FNIC_STD_GET_OX_ID(fchdr); if (fdls_get_state(fdls) != FDLS_STATE_REGISTER_FC4_TYPES) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "RFT_ID resp recvd in state(%d). Dropping.", fdls_get_state(fdls)); return; } if (iport->active_oxid_fabric_req != oxid) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Incorrect OXID in response. state: %d, oxid recvd: 0x%x, active oxid: 0x%x\n", fdls_get_state(fdls), oxid, iport->active_oxid_fabric_req); return; } rsp = FNIC_STD_GET_FC_CT_CMD((&rft_rsp->fc_std_ct_hdr)); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS process RFT ID response: 0x%04x", iport->fcid, (uint32_t) rsp); fdls_free_oxid(iport, oxid, &iport->active_oxid_fabric_req); switch (rsp) { case FC_FS_ACC: if (iport->fabric.timer_pending) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Canceling fabric disc timer %p\n", iport); fnic_del_fabric_timer_sync(fnic); } iport->fabric.timer_pending = 0; fdls->retry_counter = 0; fdls_send_register_fc4_features(iport); fdls_set_state((&iport->fabric), FDLS_STATE_REGISTER_FC4_FEATURES); break; case FC_FS_RJT: reason_code = rft_rsp->fc_std_ct_hdr.ct_reason; if (((reason_code == FC_FS_RJT_BSY) || (reason_code == FC_FS_RJT_UNABL)) && (fdls->retry_counter < FDLS_RETRY_COUNT)) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: RFT_ID ret ELS_LS_RJT BUSY. Retry from timer routine", iport->fcid); /* Retry again from the timer routine */ fdls->flags |= FNIC_FDLS_RETRY_FRAME; } else { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: RFT_ID REJ. Halting discovery reason %d expl %d", iport->fcid, reason_code, rft_rsp->fc_std_ct_hdr.ct_explan); if (iport->fabric.timer_pending) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Canceling fabric disc timer %p\n", iport); fnic_del_fabric_timer_sync(fnic); } fdls->timer_pending = 0; fdls->retry_counter = 0; } break; default: break; } } static void fdls_process_rpn_id_rsp(struct fnic_iport_s *iport, struct fc_frame_header *fchdr) { struct fnic_fdls_fabric_s *fdls = &iport->fabric; struct fc_std_rpn_id *rpn_rsp = (struct fc_std_rpn_id *) fchdr; uint16_t rsp; uint8_t reason_code; struct fnic *fnic = iport->fnic; uint16_t oxid = FNIC_STD_GET_OX_ID(fchdr); if (fdls_get_state(fdls) != FDLS_STATE_RPN_ID) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "RPN_ID resp recvd in state(%d). Dropping.", fdls_get_state(fdls)); return; } if (iport->active_oxid_fabric_req != oxid) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Incorrect OXID in response. state: %d, oxid recvd: 0x%x, active oxid: 0x%x\n", fdls_get_state(fdls), oxid, iport->active_oxid_fabric_req); return; } rsp = FNIC_STD_GET_FC_CT_CMD((&rpn_rsp->fc_std_ct_hdr)); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS process RPN ID response: 0x%04x", iport->fcid, (uint32_t) rsp); fdls_free_oxid(iport, oxid, &iport->active_oxid_fabric_req); switch (rsp) { case FC_FS_ACC: if (iport->fabric.timer_pending) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Canceling fabric disc timer %p\n", iport); fnic_del_fabric_timer_sync(fnic); } iport->fabric.timer_pending = 0; fdls->retry_counter = 0; fdls_send_register_fc4_types(iport); fdls_set_state((&iport->fabric), FDLS_STATE_REGISTER_FC4_TYPES); break; case FC_FS_RJT: reason_code = rpn_rsp->fc_std_ct_hdr.ct_reason; if (((reason_code == FC_FS_RJT_BSY) || (reason_code == FC_FS_RJT_UNABL)) && (fdls->retry_counter < FDLS_RETRY_COUNT)) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "RPN_ID returned REJ BUSY. Retry from timer routine %p", iport); /* Retry again from the timer routine */ fdls->flags |= FNIC_FDLS_RETRY_FRAME; } else { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "RPN_ID ELS_LS_RJT. Halting discovery %p", iport); if (iport->fabric.timer_pending) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Canceling fabric disc timer %p\n", iport); fnic_del_fabric_timer_sync(fnic); } fdls->timer_pending = 0; fdls->retry_counter = 0; } break; default: break; } } static void fdls_process_scr_rsp(struct fnic_iport_s *iport, struct fc_frame_header *fchdr) { struct fnic_fdls_fabric_s *fdls = &iport->fabric; struct fc_std_scr *scr_rsp = (struct fc_std_scr *) fchdr; struct fc_std_els_rjt_rsp *els_rjt = (struct fc_std_els_rjt_rsp *) fchdr; struct fnic *fnic = iport->fnic; uint16_t oxid = FNIC_STD_GET_OX_ID(fchdr); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "FDLS process SCR response: 0x%04x", (uint32_t) scr_rsp->scr.scr_cmd); if (fdls_get_state(fdls) != FDLS_STATE_SCR) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "SCR resp recvd in state(%d). Dropping.", fdls_get_state(fdls)); return; } if (iport->active_oxid_fabric_req != oxid) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Incorrect OXID in response. state: %d, oxid recvd: 0x%x, active oxid: 0x%x\n", fdls_get_state(fdls), oxid, iport->active_oxid_fabric_req); } fdls_free_oxid(iport, oxid, &iport->active_oxid_fabric_req); switch (scr_rsp->scr.scr_cmd) { case ELS_LS_ACC: atomic64_inc(&iport->iport_stats.fabric_scr_ls_accepts); if (iport->fabric.timer_pending) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Canceling fabric disc timer %p\n", iport); fnic_del_fabric_timer_sync(fnic); } iport->fabric.timer_pending = 0; iport->fabric.retry_counter = 0; fdls_send_gpn_ft(iport, FDLS_STATE_GPN_FT); break; case ELS_LS_RJT: atomic64_inc(&iport->iport_stats.fabric_scr_ls_rejects); if (((els_rjt->rej.er_reason == ELS_RJT_BUSY) || (els_rjt->rej.er_reason == ELS_RJT_UNAB)) && (fdls->retry_counter < FDLS_RETRY_COUNT)) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "SCR ELS_LS_RJT BUSY. Retry from timer routine %p", iport); /* Retry again from the timer routine */ fdls->flags |= FNIC_FDLS_RETRY_FRAME; } else { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "SCR returned ELS_LS_RJT. Halting discovery %p", iport); if (iport->fabric.timer_pending) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Canceling fabric disc timer %p\n", iport); fnic_del_fabric_timer_sync(fnic); } fdls->timer_pending = 0; fdls->retry_counter = 0; } break; default: atomic64_inc(&iport->iport_stats.fabric_scr_misc_rejects); break; } } static void fdls_process_gpn_ft_tgt_list(struct fnic_iport_s *iport, struct fc_frame_header *fchdr, int len) { struct fc_gpn_ft_rsp_iu *gpn_ft_tgt; struct fnic_tport_s *tport, *next; uint32_t fcid; uint64_t wwpn; int rem_len = len; u32 old_link_down_cnt = iport->fnic->link_down_cnt; struct fnic *fnic = iport->fnic; FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS process GPN_FT tgt list", iport->fcid); gpn_ft_tgt = (struct fc_gpn_ft_rsp_iu *)((uint8_t *) fchdr + sizeof(struct fc_frame_header) + sizeof(struct fc_ct_hdr)); len -= sizeof(struct fc_frame_header) + sizeof(struct fc_ct_hdr); while (rem_len > 0) { fcid = ntoh24(gpn_ft_tgt->fcid); wwpn = be64_to_cpu(gpn_ft_tgt->wwpn); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "tport: 0x%x: ctrl:0x%x", fcid, gpn_ft_tgt->ctrl); if (fcid == iport->fcid) { if (gpn_ft_tgt->ctrl & FC_NS_FID_LAST) break; gpn_ft_tgt++; rem_len -= sizeof(struct fc_gpn_ft_rsp_iu); continue; } tport = fnic_find_tport_by_wwpn(iport, wwpn); if (!tport) { /* * New port registered with the switch or first time query */ tport = fdls_create_tport(iport, fcid, wwpn); if (!tport) return; } /* * check if this was an existing tport with same fcid * but whose wwpn has changed now ,then remove it and * create a new one */ if (tport->fcid != fcid) { fdls_delete_tport(iport, tport); tport = fdls_create_tport(iport, fcid, wwpn); if (!tport) return; } /* * If this GPN_FT rsp is after RSCN then mark the tports which * matches with the new GPN_FT list, if some tport is not * found in GPN_FT we went to delete that tport later. */ if (fdls_get_state((&iport->fabric)) == FDLS_STATE_RSCN_GPN_FT) tport->flags |= FNIC_FDLS_TPORT_IN_GPN_FT_LIST; if (gpn_ft_tgt->ctrl & FC_NS_FID_LAST) break; gpn_ft_tgt++; rem_len -= sizeof(struct fc_gpn_ft_rsp_iu); } if (rem_len <= 0) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "GPN_FT response: malformed/corrupt frame rxlen: %d remlen: %d", len, rem_len); } /*remove those ports which was not listed in GPN_FT */ if (fdls_get_state((&iport->fabric)) == FDLS_STATE_RSCN_GPN_FT) { list_for_each_entry_safe(tport, next, &iport->tport_list, links) { if (!(tport->flags & FNIC_FDLS_TPORT_IN_GPN_FT_LIST)) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Remove port: 0x%x not found in GPN_FT list", tport->fcid); fdls_delete_tport(iport, tport); } else { tport->flags &= ~FNIC_FDLS_TPORT_IN_GPN_FT_LIST; } if ((old_link_down_cnt != iport->fnic->link_down_cnt) || (iport->state != FNIC_IPORT_STATE_READY)) { return; } } } } static void fdls_process_gpn_ft_rsp(struct fnic_iport_s *iport, struct fc_frame_header *fchdr, int len) { struct fnic_fdls_fabric_s *fdls = &iport->fabric; struct fc_std_gpn_ft *gpn_ft_rsp = (struct fc_std_gpn_ft *) fchdr; uint16_t rsp; uint8_t reason_code; int count = 0; struct fnic_tport_s *tport, *next; u32 old_link_down_cnt = iport->fnic->link_down_cnt; struct fnic *fnic = iport->fnic; uint16_t oxid = FNIC_STD_GET_OX_ID(fchdr); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "FDLS process GPN_FT response: iport state: %d len: %d", iport->state, len); /* * GPNFT response :- * FDLS_STATE_GPN_FT : GPNFT send after SCR state * during fabric discovery(FNIC_IPORT_STATE_FABRIC_DISC) * FDLS_STATE_RSCN_GPN_FT : GPNFT send in response to RSCN * FDLS_STATE_SEND_GPNFT : GPNFT send after deleting a Target, * e.g. after receiving Target LOGO * FDLS_STATE_TGT_DISCOVERY :Target discovery is currently in progress * from previous GPNFT response,a new GPNFT response has come. */ if (!(((iport->state == FNIC_IPORT_STATE_FABRIC_DISC) && (fdls_get_state(fdls) == FDLS_STATE_GPN_FT)) || ((iport->state == FNIC_IPORT_STATE_READY) && ((fdls_get_state(fdls) == FDLS_STATE_RSCN_GPN_FT) || (fdls_get_state(fdls) == FDLS_STATE_SEND_GPNFT) || (fdls_get_state(fdls) == FDLS_STATE_TGT_DISCOVERY))))) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "GPNFT resp recvd in fab state(%d) iport_state(%d). Dropping.", fdls_get_state(fdls), iport->state); return; } if (iport->active_oxid_fabric_req != oxid) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Incorrect OXID in response. state: %d, oxid recvd: 0x%x, active oxid: 0x%x\n", fdls_get_state(fdls), oxid, iport->active_oxid_fabric_req); } fdls_free_oxid(iport, oxid, &iport->active_oxid_fabric_req); iport->state = FNIC_IPORT_STATE_READY; rsp = FNIC_STD_GET_FC_CT_CMD((&gpn_ft_rsp->fc_std_ct_hdr)); switch (rsp) { case FC_FS_ACC: FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: GPNFT_RSP accept", iport->fcid); if (iport->fabric.timer_pending) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: Canceling fabric disc timer\n", iport->fcid); fnic_del_fabric_timer_sync(fnic); } iport->fabric.timer_pending = 0; iport->fabric.retry_counter = 0; fdls_process_gpn_ft_tgt_list(iport, fchdr, len); /* * iport state can change only if link down event happened * We don't need to undo fdls_process_gpn_ft_tgt_list, * that will be taken care in next link up event */ if (iport->state != FNIC_IPORT_STATE_READY) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Halting target discovery: fab st: %d iport st: %d ", fdls_get_state(fdls), iport->state); break; } fdls_tgt_discovery_start(iport); break; case FC_FS_RJT: reason_code = gpn_ft_rsp->fc_std_ct_hdr.ct_reason; FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: GPNFT_RSP Reject reason: %d", iport->fcid, reason_code); if (((reason_code == FC_FS_RJT_BSY) || (reason_code == FC_FS_RJT_UNABL)) && (fdls->retry_counter < FDLS_RETRY_COUNT)) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: GPNFT_RSP ret REJ/BSY. Retry from timer routine", iport->fcid); /* Retry again from the timer routine */ fdls->flags |= FNIC_FDLS_RETRY_FRAME; } else { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: GPNFT_RSP reject", iport->fcid); if (iport->fabric.timer_pending) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: Canceling fabric disc timer\n", iport->fcid); fnic_del_fabric_timer_sync(fnic); } iport->fabric.timer_pending = 0; iport->fabric.retry_counter = 0; /* * If GPN_FT ls_rjt then we should delete * all existing tports */ count = 0; list_for_each_entry_safe(tport, next, &iport->tport_list, links) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "GPN_FT_REJECT: Remove port: 0x%x", tport->fcid); fdls_delete_tport(iport, tport); if ((old_link_down_cnt != iport->fnic->link_down_cnt) || (iport->state != FNIC_IPORT_STATE_READY)) { return; } count++; } FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "GPN_FT_REJECT: Removed (0x%x) ports", count); } break; default: break; } } /** * fdls_process_fabric_logo_rsp - Handle an flogo response from the fcf * @iport: Handle to fnic iport * @fchdr: Incoming frame */ static void fdls_process_fabric_logo_rsp(struct fnic_iport_s *iport, struct fc_frame_header *fchdr) { struct fc_std_flogi *flogo_rsp = (struct fc_std_flogi *) fchdr; struct fnic_fdls_fabric_s *fdls = &iport->fabric; struct fnic *fnic = iport->fnic; uint16_t oxid = FNIC_STD_GET_OX_ID(fchdr); if (iport->active_oxid_fabric_req != oxid) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Incorrect OXID in response. state: %d, oxid recvd: 0x%x, active oxid: 0x%x\n", fdls_get_state(fdls), oxid, iport->active_oxid_fabric_req); } fdls_free_oxid(iport, oxid, &iport->active_oxid_fabric_req); switch (flogo_rsp->els.fl_cmd) { case ELS_LS_ACC: if (iport->fabric.state != FDLS_STATE_FABRIC_LOGO) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Flogo response. Fabric not in LOGO state. Dropping! %p", iport); return; } iport->fabric.state = FDLS_STATE_FLOGO_DONE; iport->state = FNIC_IPORT_STATE_LINK_WAIT; if (iport->fabric.timer_pending) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "iport 0x%p Canceling fabric disc timer\n", iport); fnic_del_fabric_timer_sync(fnic); } iport->fabric.timer_pending = 0; FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Flogo response from Fabric for did: 0x%x", ntoh24(fchdr->fh_d_id)); return; case ELS_LS_RJT: FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Flogo response from Fabric for did: 0x%x returned ELS_LS_RJT", ntoh24(fchdr->fh_d_id)); return; default: FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "FLOGO response not accepted or rejected: 0x%x", flogo_rsp->els.fl_cmd); } } static void fdls_process_flogi_rsp(struct fnic_iport_s *iport, struct fc_frame_header *fchdr, void *rx_frame) { struct fnic_fdls_fabric_s *fabric = &iport->fabric; struct fc_std_flogi *flogi_rsp = (struct fc_std_flogi *) fchdr; uint8_t *fcid; uint16_t rdf_size; uint8_t fcmac[6] = { 0x0E, 0XFC, 0x00, 0x00, 0x00, 0x00 }; struct fnic *fnic = iport->fnic; uint16_t oxid = FNIC_STD_GET_OX_ID(fchdr); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS processing FLOGI response", iport->fcid); if (fdls_get_state(fabric) != FDLS_STATE_FABRIC_FLOGI) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "FLOGI response received in state (%d). Dropping frame", fdls_get_state(fabric)); return; } if (iport->active_oxid_fabric_req != oxid) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Incorrect OXID in response. state: %d, oxid recvd: 0x%x, active oxid: 0x%x\n", fdls_get_state(fabric), oxid, iport->active_oxid_fabric_req); return; } fdls_free_oxid(iport, oxid, &iport->active_oxid_fabric_req); switch (flogi_rsp->els.fl_cmd) { case ELS_LS_ACC: atomic64_inc(&iport->iport_stats.fabric_flogi_ls_accepts); if (iport->fabric.timer_pending) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "iport fcid: 0x%x Canceling fabric disc timer\n", iport->fcid); fnic_del_fabric_timer_sync(fnic); } iport->fabric.timer_pending = 0; iport->fabric.retry_counter = 0; fcid = FNIC_STD_GET_D_ID(fchdr); iport->fcid = ntoh24(fcid); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FLOGI response accepted", iport->fcid); /* Learn the Service Params */ rdf_size = be16_to_cpu(FNIC_LOGI_RDF_SIZE(flogi_rsp->els)); if ((rdf_size >= FNIC_MIN_DATA_FIELD_SIZE) && (rdf_size < FNIC_FC_MAX_PAYLOAD_LEN)) iport->max_payload_size = min(rdf_size, iport->max_payload_size); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "max_payload_size from fabric: %u set: %d", rdf_size, iport->max_payload_size); iport->r_a_tov = be32_to_cpu(FNIC_LOGI_R_A_TOV(flogi_rsp->els)); iport->e_d_tov = be32_to_cpu(FNIC_LOGI_E_D_TOV(flogi_rsp->els)); if (FNIC_LOGI_FEATURES(flogi_rsp->els) & FNIC_FC_EDTOV_NSEC) iport->e_d_tov = iport->e_d_tov / FNIC_NSEC_TO_MSEC; FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "From fabric: R_A_TOV: %d E_D_TOV: %d", iport->r_a_tov, iport->e_d_tov); fc_host_fabric_name(iport->fnic->host) = get_unaligned_be64(&FNIC_LOGI_NODE_NAME(flogi_rsp->els)); fc_host_port_id(iport->fnic->host) = iport->fcid; fnic_fdls_learn_fcoe_macs(iport, rx_frame, fcid); if (fnic_fdls_register_portid(iport, iport->fcid, rx_frame) != 0) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FLOGI registration failed", iport->fcid); break; } memcpy(&fcmac[3], fcid, 3); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Adding vNIC device MAC addr: %02x:%02x:%02x:%02x:%02x:%02x", fcmac[0], fcmac[1], fcmac[2], fcmac[3], fcmac[4], fcmac[5]); vnic_dev_add_addr(iport->fnic->vdev, fcmac); if (fdls_get_state(fabric) == FDLS_STATE_FABRIC_FLOGI) { fnic_fdls_start_plogi(iport); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "FLOGI response received. Starting PLOGI"); } else { /* From FDLS_STATE_FABRIC_FLOGI state fabric can only go to * FDLS_STATE_LINKDOWN * state, hence we don't have to worry about undoing: * the fnic_fdls_register_portid and vnic_dev_add_addr */ FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "FLOGI response received in state (%d). Dropping frame", fdls_get_state(fabric)); } break; case ELS_LS_RJT: atomic64_inc(&iport->iport_stats.fabric_flogi_ls_rejects); if (fabric->retry_counter < iport->max_flogi_retries) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "FLOGI returned ELS_LS_RJT BUSY. Retry from timer routine %p", iport); /* Retry Flogi again from the timer routine. */ fabric->flags |= FNIC_FDLS_RETRY_FRAME; } else { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "FLOGI returned ELS_LS_RJT. Halting discovery %p", iport); if (iport->fabric.timer_pending) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "iport 0x%p Canceling fabric disc timer\n", iport); fnic_del_fabric_timer_sync(fnic); } fabric->timer_pending = 0; fabric->retry_counter = 0; } break; default: FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "FLOGI response not accepted: 0x%x", flogi_rsp->els.fl_cmd); atomic64_inc(&iport->iport_stats.fabric_flogi_misc_rejects); break; } } static void fdls_process_fabric_plogi_rsp(struct fnic_iport_s *iport, struct fc_frame_header *fchdr) { struct fc_std_flogi *plogi_rsp = (struct fc_std_flogi *) fchdr; struct fc_std_els_rjt_rsp *els_rjt = (struct fc_std_els_rjt_rsp *) fchdr; struct fnic_fdls_fabric_s *fdls = &iport->fabric; struct fnic *fnic = iport->fnic; uint16_t oxid = FNIC_STD_GET_OX_ID(fchdr); if (fdls_get_state((&iport->fabric)) != FDLS_STATE_FABRIC_PLOGI) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Fabric PLOGI response received in state (%d). Dropping frame", fdls_get_state(&iport->fabric)); return; } if (iport->active_oxid_fabric_req != oxid) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Incorrect OXID in response. state: %d, oxid recvd: 0x%x, active oxid: 0x%x\n", fdls_get_state(fdls), oxid, iport->active_oxid_fabric_req); return; } fdls_free_oxid(iport, oxid, &iport->active_oxid_fabric_req); switch (plogi_rsp->els.fl_cmd) { case ELS_LS_ACC: atomic64_inc(&iport->iport_stats.fabric_plogi_ls_accepts); if (iport->fabric.timer_pending) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "iport fcid: 0x%x fabric PLOGI response: Accepted\n", iport->fcid); fnic_del_fabric_timer_sync(fnic); } iport->fabric.timer_pending = 0; iport->fabric.retry_counter = 0; fdls_set_state(&iport->fabric, FDLS_STATE_RPN_ID); fdls_send_rpn_id(iport); break; case ELS_LS_RJT: atomic64_inc(&iport->iport_stats.fabric_plogi_ls_rejects); if (((els_rjt->rej.er_reason == ELS_RJT_BUSY) || (els_rjt->rej.er_reason == ELS_RJT_UNAB)) && (iport->fabric.retry_counter < iport->max_plogi_retries)) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: Fabric PLOGI ELS_LS_RJT BUSY. Retry from timer routine", iport->fcid); } else { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: Fabric PLOGI ELS_LS_RJT. Halting discovery", iport->fcid); if (iport->fabric.timer_pending) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "iport fcid: 0x%x Canceling fabric disc timer\n", iport->fcid); fnic_del_fabric_timer_sync(fnic); } iport->fabric.timer_pending = 0; iport->fabric.retry_counter = 0; return; } break; default: FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "PLOGI response not accepted: 0x%x", plogi_rsp->els.fl_cmd); atomic64_inc(&iport->iport_stats.fabric_plogi_misc_rejects); break; } } static void fdls_process_fdmi_plogi_rsp(struct fnic_iport_s *iport, struct fc_frame_header *fchdr) { struct fc_std_flogi *plogi_rsp = (struct fc_std_flogi *)fchdr; struct fc_std_els_rjt_rsp *els_rjt = (struct fc_std_els_rjt_rsp *)fchdr; struct fnic_fdls_fabric_s *fdls = &iport->fabric; struct fnic *fnic = iport->fnic; u64 fdmi_tov; uint16_t oxid = FNIC_STD_GET_OX_ID(fchdr); if (iport->active_oxid_fdmi_plogi != oxid) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Incorrect OXID in response. state: %d, oxid recvd: 0x%x, active oxid: 0x%x\n", fdls_get_state(fdls), oxid, iport->active_oxid_fdmi_plogi); return; } iport->fabric.fdmi_pending &= ~FDLS_FDMI_PLOGI_PENDING; fdls_free_oxid(iport, oxid, &iport->active_oxid_fdmi_plogi); if (ntoh24(fchdr->fh_s_id) == FC_FID_MGMT_SERV) { del_timer_sync(&iport->fabric.fdmi_timer); iport->fabric.fdmi_pending = 0; switch (plogi_rsp->els.fl_cmd) { case ELS_LS_ACC: FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "FDLS process fdmi PLOGI response status: ELS_LS_ACC\n"); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Sending fdmi registration for port 0x%x\n", iport->fcid); fdls_fdmi_register_hba(iport); fdls_fdmi_register_pa(iport); fdmi_tov = jiffies + msecs_to_jiffies(5000); mod_timer(&iport->fabric.fdmi_timer, round_jiffies(fdmi_tov)); break; case ELS_LS_RJT: FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Fabric FDMI PLOGI returned ELS_LS_RJT reason: 0x%x", els_rjt->rej.er_reason); if (((els_rjt->rej.er_reason == ELS_RJT_BUSY) || (els_rjt->rej.er_reason == ELS_RJT_UNAB)) && (iport->fabric.fdmi_retry < 7)) { iport->fabric.fdmi_retry++; fdls_send_fdmi_plogi(iport); } break; default: break; } } } static void fdls_process_fdmi_reg_ack(struct fnic_iport_s *iport, struct fc_frame_header *fchdr, int rsp_type) { struct fnic *fnic = iport->fnic; uint16_t oxid; if (!iport->fabric.fdmi_pending) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Received FDMI ack while not waiting: 0x%x\n", FNIC_STD_GET_OX_ID(fchdr)); return; } oxid = FNIC_STD_GET_OX_ID(fchdr); if ((iport->active_oxid_fdmi_rhba != oxid) && (iport->active_oxid_fdmi_rpa != oxid)) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Incorrect OXID in response. oxid recvd: 0x%x, active oxids(rhba,rpa): 0x%x, 0x%x\n", oxid, iport->active_oxid_fdmi_rhba, iport->active_oxid_fdmi_rpa); return; } if (FNIC_FRAME_TYPE(oxid) == FNIC_FRAME_TYPE_FDMI_RHBA) { iport->fabric.fdmi_pending &= ~FDLS_FDMI_REG_HBA_PENDING; fdls_free_oxid(iport, oxid, &iport->active_oxid_fdmi_rhba); } else { iport->fabric.fdmi_pending &= ~FDLS_FDMI_RPA_PENDING; fdls_free_oxid(iport, oxid, &iport->active_oxid_fdmi_rpa); } FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "iport fcid: 0x%x: Received FDMI registration ack\n", iport->fcid); if (!iport->fabric.fdmi_pending) { del_timer_sync(&iport->fabric.fdmi_timer); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "iport fcid: 0x%x: Canceling FDMI timer\n", iport->fcid); } } static void fdls_process_fdmi_abts_rsp(struct fnic_iport_s *iport, struct fc_frame_header *fchdr) { uint32_t s_id; struct fnic *fnic = iport->fnic; uint16_t oxid; s_id = ntoh24(FNIC_STD_GET_S_ID(fchdr)); if (!(s_id != FC_FID_MGMT_SERV)) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Received abts rsp with invalid SID: 0x%x. Dropping frame", s_id); return; } oxid = FNIC_STD_GET_OX_ID(fchdr); switch (FNIC_FRAME_TYPE(oxid)) { case FNIC_FRAME_TYPE_FDMI_PLOGI: fdls_free_oxid(iport, oxid, &iport->active_oxid_fdmi_plogi); break; case FNIC_FRAME_TYPE_FDMI_RHBA: fdls_free_oxid(iport, oxid, &iport->active_oxid_fdmi_rhba); break; case FNIC_FRAME_TYPE_FDMI_RPA: fdls_free_oxid(iport, oxid, &iport->active_oxid_fdmi_rpa); break; default: FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Received abts rsp with invalid oxid: 0x%x. Dropping frame", oxid); break; } del_timer_sync(&iport->fabric.fdmi_timer); iport->fabric.fdmi_pending &= ~FDLS_FDMI_ABORT_PENDING; fdls_send_fdmi_plogi(iport); } static void fdls_process_fabric_abts_rsp(struct fnic_iport_s *iport, struct fc_frame_header *fchdr) { uint32_t s_id; struct fc_std_abts_ba_acc *ba_acc = (struct fc_std_abts_ba_acc *)fchdr; struct fc_std_abts_ba_rjt *ba_rjt; uint32_t fabric_state = iport->fabric.state; struct fnic *fnic = iport->fnic; int frame_type; uint16_t oxid; s_id = ntoh24(fchdr->fh_s_id); ba_rjt = (struct fc_std_abts_ba_rjt *) fchdr; if (!((s_id == FC_FID_DIR_SERV) || (s_id == FC_FID_FLOGI) || (s_id == FC_FID_FCTRL))) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Received abts rsp with invalid SID: 0x%x. Dropping frame", s_id); return; } oxid = FNIC_STD_GET_OX_ID(fchdr); if (iport->active_oxid_fabric_req != oxid) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Received abts rsp with invalid oxid: 0x%x. Dropping frame", oxid); return; } if (iport->fabric.timer_pending) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Canceling fabric disc timer %p\n", iport); fnic_del_fabric_timer_sync(fnic); } iport->fabric.timer_pending = 0; iport->fabric.flags &= ~FNIC_FDLS_FABRIC_ABORT_ISSUED; if (fchdr->fh_r_ctl == FC_RCTL_BA_ACC) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Received abts rsp BA_ACC for fabric_state: %d OX_ID: 0x%x", fabric_state, be16_to_cpu(ba_acc->acc.ba_ox_id)); } else if (fchdr->fh_r_ctl == FC_RCTL_BA_RJT) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "BA_RJT fs: %d OX_ID: 0x%x rc: 0x%x rce: 0x%x", fabric_state, FNIC_STD_GET_OX_ID(&ba_rjt->fchdr), ba_rjt->rjt.br_reason, ba_rjt->rjt.br_explan); } frame_type = FNIC_FRAME_TYPE(oxid); fdls_free_oxid(iport, oxid, &iport->active_oxid_fabric_req); /* currently error handling/retry logic is same for ABTS BA_ACC & BA_RJT */ switch (frame_type) { case FNIC_FRAME_TYPE_FABRIC_FLOGI: if (iport->fabric.retry_counter < iport->max_flogi_retries) fdls_send_fabric_flogi(iport); else FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Exceeded max FLOGI retries"); break; case FNIC_FRAME_TYPE_FABRIC_LOGO: if (iport->fabric.retry_counter < FABRIC_LOGO_MAX_RETRY) fdls_send_fabric_logo(iport); break; case FNIC_FRAME_TYPE_FABRIC_PLOGI: if (iport->fabric.retry_counter < iport->max_plogi_retries) fdls_send_fabric_plogi(iport); else FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Exceeded max PLOGI retries"); break; case FNIC_FRAME_TYPE_FABRIC_RPN: if (iport->fabric.retry_counter < FDLS_RETRY_COUNT) fdls_send_rpn_id(iport); else /* go back to fabric Plogi */ fnic_fdls_start_plogi(iport); break; case FNIC_FRAME_TYPE_FABRIC_SCR: if (iport->fabric.retry_counter < FDLS_RETRY_COUNT) fdls_send_scr(iport); else { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "SCR exhausted retries. Start fabric PLOGI %p", iport); fnic_fdls_start_plogi(iport); /* go back to fabric Plogi */ } break; case FNIC_FRAME_TYPE_FABRIC_RFT: if (iport->fabric.retry_counter < FDLS_RETRY_COUNT) fdls_send_register_fc4_types(iport); else { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "RFT exhausted retries. Start fabric PLOGI %p", iport); fnic_fdls_start_plogi(iport); /* go back to fabric Plogi */ } break; case FNIC_FRAME_TYPE_FABRIC_RFF: if (iport->fabric.retry_counter < FDLS_RETRY_COUNT) fdls_send_register_fc4_features(iport); else { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "RFF exhausted retries. Start fabric PLOGI %p", iport); fnic_fdls_start_plogi(iport); /* go back to fabric Plogi */ } break; case FNIC_FRAME_TYPE_FABRIC_GPN_FT: if (iport->fabric.retry_counter <= FDLS_RETRY_COUNT) fdls_send_gpn_ft(iport, fabric_state); else FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "GPN FT exhausted retries. Start fabric PLOGI %p", iport); break; default: /* * We should not be here since we already validated rx oxid with * our active_oxid_fabric_req */ FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Invalid OXID/active oxid 0x%x\n", oxid); WARN_ON(true); return; } } static void fdls_process_abts_req(struct fnic_iport_s *iport, struct fc_frame_header *fchdr) { uint8_t *frame; struct fc_std_abts_ba_acc *pba_acc; uint32_t nport_id; uint16_t oxid = FNIC_STD_GET_OX_ID(fchdr); struct fnic_tport_s *tport; struct fnic *fnic = iport->fnic; uint16_t frame_size = FNIC_ETH_FCOE_HDRS_OFFSET + sizeof(struct fc_std_abts_ba_acc); nport_id = ntoh24(fchdr->fh_s_id); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Received abort from SID 0x%8x", nport_id); tport = fnic_find_tport_by_fcid(iport, nport_id); if (tport) { if (tport->active_oxid == oxid) { tport->flags |= FNIC_FDLS_TGT_ABORT_ISSUED; fdls_free_oxid(iport, oxid, &tport->active_oxid); } } frame = fdls_alloc_frame(iport); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "0x%x: Failed to allocate frame to send response for ABTS req", iport->fcid); return; } pba_acc = (struct fc_std_abts_ba_acc *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); *pba_acc = (struct fc_std_abts_ba_acc) { .fchdr = {.fh_r_ctl = FC_RCTL_BA_ACC, .fh_f_ctl = {FNIC_FCP_RSP_FCTL, 0, 0}}, .acc = {.ba_low_seq_cnt = 0, .ba_high_seq_cnt = cpu_to_be16(0xFFFF)} }; FNIC_STD_SET_S_ID(pba_acc->fchdr, fchdr->fh_d_id); FNIC_STD_SET_D_ID(pba_acc->fchdr, fchdr->fh_s_id); FNIC_STD_SET_OX_ID(pba_acc->fchdr, FNIC_STD_GET_OX_ID(fchdr)); FNIC_STD_SET_RX_ID(pba_acc->fchdr, FNIC_STD_GET_RX_ID(fchdr)); pba_acc->acc.ba_rx_id = cpu_to_be16(FNIC_STD_GET_RX_ID(fchdr)); pba_acc->acc.ba_ox_id = cpu_to_be16(FNIC_STD_GET_OX_ID(fchdr)); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS send BA ACC with oxid: 0x%x", iport->fcid, oxid); fnic_send_fcoe_frame(iport, frame, frame_size); } static void fdls_process_unsupported_els_req(struct fnic_iport_s *iport, struct fc_frame_header *fchdr) { uint8_t *frame; struct fc_std_els_rjt_rsp *pls_rsp; uint16_t oxid; uint32_t d_id = ntoh24(fchdr->fh_d_id); struct fnic *fnic = iport->fnic; uint16_t frame_size = FNIC_ETH_FCOE_HDRS_OFFSET + sizeof(struct fc_std_els_rjt_rsp); if (iport->fcid != d_id) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Dropping unsupported ELS with illegal frame bits 0x%x\n", d_id); atomic64_inc(&iport->iport_stats.unsupported_frames_dropped); return; } if ((iport->state != FNIC_IPORT_STATE_READY) && (iport->state != FNIC_IPORT_STATE_FABRIC_DISC)) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Dropping unsupported ELS request in iport state: %d", iport->state); atomic64_inc(&iport->iport_stats.unsupported_frames_dropped); return; } frame = fdls_alloc_frame(iport); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame to send response to unsupported ELS request"); return; } pls_rsp = (struct fc_std_els_rjt_rsp *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); fdls_init_els_rjt_frame(frame, iport); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: Process unsupported ELS request from SID: 0x%x", iport->fcid, ntoh24(fchdr->fh_s_id)); /* We don't support this ELS request, send a reject */ pls_rsp->rej.er_reason = 0x0B; pls_rsp->rej.er_explan = 0x0; pls_rsp->rej.er_vendor = 0x0; FNIC_STD_SET_S_ID(pls_rsp->fchdr, fchdr->fh_d_id); FNIC_STD_SET_D_ID(pls_rsp->fchdr, fchdr->fh_s_id); oxid = FNIC_STD_GET_OX_ID(fchdr); FNIC_STD_SET_OX_ID(pls_rsp->fchdr, oxid); fnic_send_fcoe_frame(iport, frame, frame_size); } static void fdls_process_rls_req(struct fnic_iport_s *iport, struct fc_frame_header *fchdr) { uint8_t *frame; struct fc_std_rls_acc *prls_acc_rsp; uint16_t oxid; struct fnic *fnic = iport->fnic; uint16_t frame_size = FNIC_ETH_FCOE_HDRS_OFFSET + sizeof(struct fc_std_rls_acc); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Process RLS request %d", iport->fnic->fnic_num); if ((iport->state != FNIC_IPORT_STATE_READY) && (iport->state != FNIC_IPORT_STATE_FABRIC_DISC)) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Received RLS req in iport state: %d. Dropping the frame.", iport->state); return; } frame = fdls_alloc_frame(iport); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame to send RLS accept"); return; } prls_acc_rsp = (struct fc_std_rls_acc *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); FNIC_STD_SET_S_ID(prls_acc_rsp->fchdr, fchdr->fh_d_id); FNIC_STD_SET_D_ID(prls_acc_rsp->fchdr, fchdr->fh_s_id); oxid = FNIC_STD_GET_OX_ID(fchdr); FNIC_STD_SET_OX_ID(prls_acc_rsp->fchdr, oxid); FNIC_STD_SET_RX_ID(prls_acc_rsp->fchdr, FNIC_UNASSIGNED_RXID); FNIC_STD_SET_F_CTL(prls_acc_rsp->fchdr, FNIC_ELS_REP_FCTL << 16); FNIC_STD_SET_R_CTL(prls_acc_rsp->fchdr, FC_RCTL_ELS_REP); FNIC_STD_SET_TYPE(prls_acc_rsp->fchdr, FC_TYPE_ELS); prls_acc_rsp->els.rls_cmd = ELS_LS_ACC; prls_acc_rsp->els.rls_lesb.lesb_link_fail = cpu_to_be32(iport->fnic->link_down_cnt); fnic_send_fcoe_frame(iport, frame, frame_size); } static void fdls_process_els_req(struct fnic_iport_s *iport, struct fc_frame_header *fchdr, uint32_t len) { uint8_t *frame; struct fc_std_els_acc_rsp *pels_acc; uint16_t oxid; uint8_t *fc_payload; uint8_t type; struct fnic *fnic = iport->fnic; uint16_t frame_size = FNIC_ETH_FCOE_HDRS_OFFSET; fc_payload = (uint8_t *) fchdr + sizeof(struct fc_frame_header); type = *fc_payload; if ((iport->state != FNIC_IPORT_STATE_READY) && (iport->state != FNIC_IPORT_STATE_FABRIC_DISC)) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Dropping ELS frame type: 0x%x in iport state: %d", type, iport->state); return; } switch (type) { case ELS_ECHO: FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "sending LS_ACC for ECHO request %d\n", iport->fnic->fnic_num); break; case ELS_RRQ: FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "sending LS_ACC for RRQ request %d\n", iport->fnic->fnic_num); break; default: FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "sending LS_ACC for 0x%x ELS frame\n", type); break; } frame = fdls_alloc_frame(iport); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame to send ELS response for 0x%x", type); return; } if (type == ELS_ECHO) { /* Brocade sends a longer payload, copy all frame back */ memcpy(frame, fchdr, len); } pels_acc = (struct fc_std_els_acc_rsp *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); fdls_init_els_acc_frame(frame, iport); FNIC_STD_SET_D_ID(pels_acc->fchdr, fchdr->fh_s_id); oxid = FNIC_STD_GET_OX_ID(fchdr); FNIC_STD_SET_OX_ID(pels_acc->fchdr, oxid); if (type == ELS_ECHO) frame_size += len; else frame_size += sizeof(struct fc_std_els_acc_rsp); fnic_send_fcoe_frame(iport, frame, frame_size); } static void fdls_process_tgt_abts_rsp(struct fnic_iport_s *iport, struct fc_frame_header *fchdr) { uint32_t s_id; struct fnic_tport_s *tport; uint32_t tport_state; struct fc_std_abts_ba_acc *ba_acc; struct fc_std_abts_ba_rjt *ba_rjt; uint16_t oxid; struct fnic *fnic = iport->fnic; int frame_type; s_id = ntoh24(fchdr->fh_s_id); ba_acc = (struct fc_std_abts_ba_acc *)fchdr; ba_rjt = (struct fc_std_abts_ba_rjt *)fchdr; tport = fnic_find_tport_by_fcid(iport, s_id); if (!tport) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Received tgt abts rsp with invalid SID: 0x%x", s_id); return; } if (tport->timer_pending) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "tport 0x%p Canceling fabric disc timer\n", tport); fnic_del_tport_timer_sync(fnic, tport); } if (iport->state != FNIC_IPORT_STATE_READY) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Received tgt abts rsp in iport state(%d). Dropping.", iport->state); return; } tport->timer_pending = 0; tport->flags &= ~FNIC_FDLS_TGT_ABORT_ISSUED; tport_state = tport->state; oxid = FNIC_STD_GET_OX_ID(fchdr); /*This abort rsp is for ADISC */ frame_type = FNIC_FRAME_TYPE(oxid); switch (frame_type) { case FNIC_FRAME_TYPE_TGT_ADISC: if (fchdr->fh_r_ctl == FC_RCTL_BA_ACC) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "OX_ID: 0x%x tgt_fcid: 0x%x rcvd tgt adisc abts resp BA_ACC", be16_to_cpu(ba_acc->acc.ba_ox_id), tport->fcid); } else if (fchdr->fh_r_ctl == FC_RCTL_BA_RJT) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "ADISC BA_RJT rcvd tport_fcid: 0x%x tport_state: %d ", tport->fcid, tport_state); FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "reason code: 0x%x reason code explanation:0x%x ", ba_rjt->rjt.br_reason, ba_rjt->rjt.br_explan); } if ((tport->retry_counter < FDLS_RETRY_COUNT) && (fchdr->fh_r_ctl == FC_RCTL_BA_ACC)) { fdls_free_oxid(iport, oxid, &tport->active_oxid); fdls_send_tgt_adisc(iport, tport); return; } fdls_free_oxid(iport, oxid, &tport->active_oxid); FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "ADISC not responding. Deleting target port: 0x%x", tport->fcid); fdls_delete_tport(iport, tport); /* Restart discovery of targets */ if ((iport->state == FNIC_IPORT_STATE_READY) && (iport->fabric.state != FDLS_STATE_SEND_GPNFT) && (iport->fabric.state != FDLS_STATE_RSCN_GPN_FT)) { fdls_send_gpn_ft(iport, FDLS_STATE_SEND_GPNFT); } break; case FNIC_FRAME_TYPE_TGT_PLOGI: if (fchdr->fh_r_ctl == FC_RCTL_BA_ACC) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Received tgt PLOGI abts response BA_ACC tgt_fcid: 0x%x", tport->fcid); } else if (fchdr->fh_r_ctl == FC_RCTL_BA_RJT) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "PLOGI BA_RJT received for tport_fcid: 0x%x OX_ID: 0x%x", tport->fcid, FNIC_STD_GET_OX_ID(fchdr)); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "reason code: 0x%x reason code explanation: 0x%x", ba_rjt->rjt.br_reason, ba_rjt->rjt.br_explan); } if ((tport->retry_counter < iport->max_plogi_retries) && (fchdr->fh_r_ctl == FC_RCTL_BA_ACC)) { fdls_free_oxid(iport, oxid, &tport->active_oxid); fdls_send_tgt_plogi(iport, tport); return; } fdls_free_oxid(iport, oxid, &tport->active_oxid); fdls_delete_tport(iport, tport); /* Restart discovery of targets */ if ((iport->state == FNIC_IPORT_STATE_READY) && (iport->fabric.state != FDLS_STATE_SEND_GPNFT) && (iport->fabric.state != FDLS_STATE_RSCN_GPN_FT)) { fdls_send_gpn_ft(iport, FDLS_STATE_SEND_GPNFT); } break; case FNIC_FRAME_TYPE_TGT_PRLI: if (fchdr->fh_r_ctl == FC_RCTL_BA_ACC) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: Received tgt PRLI abts response BA_ACC", tport->fcid); } else if (fchdr->fh_r_ctl == FC_RCTL_BA_RJT) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "PRLI BA_RJT received for tport_fcid: 0x%x OX_ID: 0x%x ", tport->fcid, FNIC_STD_GET_OX_ID(fchdr)); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "reason code: 0x%x reason code explanation: 0x%x", ba_rjt->rjt.br_reason, ba_rjt->rjt.br_explan); } if ((tport->retry_counter < FDLS_RETRY_COUNT) && (fchdr->fh_r_ctl == FC_RCTL_BA_ACC)) { fdls_free_oxid(iport, oxid, &tport->active_oxid); fdls_send_tgt_prli(iport, tport); return; } fdls_free_oxid(iport, oxid, &tport->active_oxid); fdls_send_tgt_plogi(iport, tport); /* go back to plogi */ fdls_set_tport_state(tport, FDLS_TGT_STATE_PLOGI); break; default: FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Received ABTS response for unknown frame %p", iport); break; } } static void fdls_process_plogi_req(struct fnic_iport_s *iport, struct fc_frame_header *fchdr) { uint8_t *frame; struct fc_std_els_rjt_rsp *pplogi_rsp; uint16_t oxid; uint32_t d_id = ntoh24(fchdr->fh_d_id); struct fnic *fnic = iport->fnic; uint16_t frame_size = FNIC_ETH_FCOE_HDRS_OFFSET + sizeof(struct fc_std_els_rjt_rsp); if (iport->fcid != d_id) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Received PLOGI with illegal frame bits. Dropping frame from 0x%x", d_id); return; } if (iport->state != FNIC_IPORT_STATE_READY) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Received PLOGI request in iport state: %d Dropping frame", iport->state); return; } frame = fdls_alloc_frame(iport); if (frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame to send response to PLOGI request"); return; } pplogi_rsp = (struct fc_std_els_rjt_rsp *) (frame + FNIC_ETH_FCOE_HDRS_OFFSET); fdls_init_els_rjt_frame(frame, iport); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: Process PLOGI request from SID: 0x%x", iport->fcid, ntoh24(fchdr->fh_s_id)); /* We don't support PLOGI request, send a reject */ pplogi_rsp->rej.er_reason = 0x0B; pplogi_rsp->rej.er_explan = 0x0; pplogi_rsp->rej.er_vendor = 0x0; FNIC_STD_SET_S_ID(pplogi_rsp->fchdr, fchdr->fh_d_id); FNIC_STD_SET_D_ID(pplogi_rsp->fchdr, fchdr->fh_s_id); oxid = FNIC_STD_GET_OX_ID(fchdr); FNIC_STD_SET_OX_ID(pplogi_rsp->fchdr, oxid); fnic_send_fcoe_frame(iport, frame, frame_size); } static void fdls_process_logo_req(struct fnic_iport_s *iport, struct fc_frame_header *fchdr) { struct fc_std_logo *logo = (struct fc_std_logo *)fchdr; uint32_t nport_id; uint64_t nport_name; struct fnic_tport_s *tport; struct fnic *fnic = iport->fnic; uint16_t oxid; nport_id = ntoh24(logo->els.fl_n_port_id); nport_name = be64_to_cpu(logo->els.fl_n_port_wwn); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Process LOGO request from fcid: 0x%x", nport_id); if (iport->state != FNIC_IPORT_STATE_READY) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Dropping LOGO req from 0x%x in iport state: %d", nport_id, iport->state); return; } tport = fnic_find_tport_by_fcid(iport, nport_id); if (!tport) { /* We are not logged in with the nport, log and drop... */ FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Received LOGO from an nport not logged in: 0x%x(0x%llx)", nport_id, nport_name); return; } if (tport->fcid != nport_id) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Received LOGO with invalid target port fcid: 0x%x(0x%llx)", nport_id, nport_name); return; } if (tport->timer_pending) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "tport fcid 0x%x: Canceling disc timer\n", tport->fcid); fnic_del_tport_timer_sync(fnic, tport); tport->timer_pending = 0; } /* got a logo in response to adisc to a target which has logged out */ if (tport->state == FDLS_TGT_STATE_ADISC) { tport->retry_counter = 0; oxid = tport->active_oxid; fdls_free_oxid(iport, oxid, &tport->active_oxid); fdls_delete_tport(iport, tport); fdls_send_logo_resp(iport, &logo->fchdr); if ((iport->state == FNIC_IPORT_STATE_READY) && (fdls_get_state(&iport->fabric) != FDLS_STATE_SEND_GPNFT) && (fdls_get_state(&iport->fabric) != FDLS_STATE_RSCN_GPN_FT)) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Sending GPNFT in response to LOGO from Target:0x%x", nport_id); fdls_send_gpn_ft(iport, FDLS_STATE_SEND_GPNFT); return; } } else { fdls_delete_tport(iport, tport); } if (iport->state == FNIC_IPORT_STATE_READY) { fdls_send_logo_resp(iport, &logo->fchdr); if ((fdls_get_state(&iport->fabric) != FDLS_STATE_SEND_GPNFT) && (fdls_get_state(&iport->fabric) != FDLS_STATE_RSCN_GPN_FT)) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Sending GPNFT in response to LOGO from Target:0x%x", nport_id); fdls_send_gpn_ft(iport, FDLS_STATE_SEND_GPNFT); } } } static void fdls_process_rscn(struct fnic_iport_s *iport, struct fc_frame_header *fchdr) { struct fc_std_rscn *rscn; struct fc_els_rscn_page *rscn_port = NULL; int num_ports; struct fnic_tport_s *tport, *next; uint32_t nport_id; uint8_t fcid[3]; int newports = 0; struct fnic_fdls_fabric_s *fdls = &iport->fabric; struct fnic *fnic = iport->fnic; int rscn_type = NOT_PC_RSCN; uint32_t sid = ntoh24(fchdr->fh_s_id); unsigned long reset_fnic_list_lock_flags = 0; uint16_t rscn_payload_len; atomic64_inc(&iport->iport_stats.num_rscns); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "FDLS process RSCN %p", iport); if (iport->state != FNIC_IPORT_STATE_READY) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "FDLS RSCN received in state(%d). Dropping", fdls_get_state(fdls)); return; } rscn = (struct fc_std_rscn *)fchdr; rscn_payload_len = be16_to_cpu(rscn->els.rscn_plen); /* frame validation */ if ((rscn_payload_len % 4 != 0) || (rscn_payload_len < 8) || (rscn_payload_len > 1024) || (rscn->els.rscn_page_len != 4)) { num_ports = 0; if ((rscn_payload_len == 0xFFFF) && (sid == FC_FID_FCTRL)) { rscn_type = PC_RSCN; FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "pcrscn: PCRSCN received. sid: 0x%x payload len: 0x%x", sid, rscn_payload_len); } else { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "RSCN payload_len: 0x%x page_len: 0x%x", rscn_payload_len, rscn->els.rscn_page_len); /* if this happens then we need to send ADISC to all the tports. */ list_for_each_entry_safe(tport, next, &iport->tport_list, links) { if (tport->state == FDLS_TGT_STATE_READY) tport->flags |= FNIC_FDLS_TPORT_SEND_ADISC; FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "RSCN for port id: 0x%x", tport->fcid); } } /* end else */ } else { num_ports = (rscn_payload_len - 4) / rscn->els.rscn_page_len; rscn_port = (struct fc_els_rscn_page *)(rscn + 1); } FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "RSCN received for num_ports: %d payload_len: %d page_len: %d ", num_ports, rscn_payload_len, rscn->els.rscn_page_len); /* * RSCN have at least one Port_ID page , but may not have any port_id * in it. If no port_id is specified in the Port_ID page , we send * ADISC to all the tports */ while (num_ports) { memcpy(fcid, rscn_port->rscn_fid, 3); nport_id = ntoh24(fcid); rscn_port++; num_ports--; /* if this happens then we need to send ADISC to all the tports. */ if (nport_id == 0) { list_for_each_entry_safe(tport, next, &iport->tport_list, links) { if (tport->state == FDLS_TGT_STATE_READY) tport->flags |= FNIC_FDLS_TPORT_SEND_ADISC; FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "RSCN for port id: 0x%x", tport->fcid); } break; } tport = fnic_find_tport_by_fcid(iport, nport_id); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "RSCN port id list: 0x%x", nport_id); if (!tport) { newports++; continue; } if (tport->state == FDLS_TGT_STATE_READY) tport->flags |= FNIC_FDLS_TPORT_SEND_ADISC; } if (pc_rscn_handling_feature_flag == PC_RSCN_HANDLING_FEATURE_ON && rscn_type == PC_RSCN && fnic->role == FNIC_ROLE_FCP_INITIATOR) { if (fnic->pc_rscn_handling_status == PC_RSCN_HANDLING_IN_PROGRESS) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "PCRSCN handling already in progress. Skip host reset: %d", iport->fnic->fnic_num); return; } FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Processing PCRSCN. Queuing fnic for host reset: %d", iport->fnic->fnic_num); fnic->pc_rscn_handling_status = PC_RSCN_HANDLING_IN_PROGRESS; spin_unlock_irqrestore(&fnic->fnic_lock, fnic->lock_flags); spin_lock_irqsave(&reset_fnic_list_lock, reset_fnic_list_lock_flags); list_add_tail(&fnic->links, &reset_fnic_list); spin_unlock_irqrestore(&reset_fnic_list_lock, reset_fnic_list_lock_flags); queue_work(reset_fnic_work_queue, &reset_fnic_work); spin_lock_irqsave(&fnic->fnic_lock, fnic->lock_flags); } else { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "FDLS process RSCN sending GPN_FT: newports: %d", newports); fdls_send_gpn_ft(iport, FDLS_STATE_RSCN_GPN_FT); fdls_send_rscn_resp(iport, fchdr); } } void fnic_fdls_disc_start(struct fnic_iport_s *iport) { struct fnic *fnic = iport->fnic; fc_host_fabric_name(iport->fnic->host) = 0; fc_host_post_event(iport->fnic->host, fc_get_event_number(), FCH_EVT_LIPRESET, 0); if (!iport->usefip) { if (iport->flags & FNIC_FIRST_LINK_UP) { spin_unlock_irqrestore(&fnic->fnic_lock, fnic->lock_flags); fnic_scsi_fcpio_reset(iport->fnic); spin_lock_irqsave(&fnic->fnic_lock, fnic->lock_flags); iport->flags &= ~FNIC_FIRST_LINK_UP; } fnic_fdls_start_flogi(iport); } else fnic_fdls_start_plogi(iport); } static void fdls_process_adisc_req(struct fnic_iport_s *iport, struct fc_frame_header *fchdr) { struct fc_std_els_adisc *padisc_acc; struct fc_std_els_adisc *adisc_req = (struct fc_std_els_adisc *)fchdr; uint64_t frame_wwnn; uint64_t frame_wwpn; uint32_t tgt_fcid; struct fnic_tport_s *tport; uint8_t *fcid; uint8_t *rjt_frame; uint8_t *acc_frame; struct fc_std_els_rjt_rsp *prjts_rsp; uint16_t oxid; struct fnic *fnic = iport->fnic; uint16_t rjt_frame_size = FNIC_ETH_FCOE_HDRS_OFFSET + sizeof(struct fc_std_els_rjt_rsp); uint16_t acc_frame_size = FNIC_ETH_FCOE_HDRS_OFFSET + sizeof(struct fc_std_els_adisc); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Process ADISC request %d", iport->fnic->fnic_num); fcid = FNIC_STD_GET_S_ID(fchdr); tgt_fcid = ntoh24(fcid); tport = fnic_find_tport_by_fcid(iport, tgt_fcid); if (!tport) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "tport for fcid: 0x%x not found. Dropping ADISC req.", tgt_fcid); return; } if (iport->state != FNIC_IPORT_STATE_READY) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Dropping ADISC req from fcid: 0x%x in iport state: %d", tgt_fcid, iport->state); return; } frame_wwnn = be64_to_cpu(adisc_req->els.adisc_wwnn); frame_wwpn = be64_to_cpu(adisc_req->els.adisc_wwpn); if ((frame_wwnn != tport->wwnn) || (frame_wwpn != tport->wwpn)) { /* send reject */ FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "ADISC req from fcid: 0x%x mismatch wwpn: 0x%llx wwnn: 0x%llx", tgt_fcid, frame_wwpn, frame_wwnn); FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "local tport wwpn: 0x%llx wwnn: 0x%llx. Sending RJT", tport->wwpn, tport->wwnn); rjt_frame = fdls_alloc_frame(iport); if (rjt_frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate rjt_frame to send response to ADISC request"); return; } prjts_rsp = (struct fc_std_els_rjt_rsp *) (rjt_frame + FNIC_ETH_FCOE_HDRS_OFFSET); fdls_init_els_rjt_frame(rjt_frame, iport); prjts_rsp->rej.er_reason = 0x03; /* logical error */ prjts_rsp->rej.er_explan = 0x1E; /* N_port login required */ prjts_rsp->rej.er_vendor = 0x0; FNIC_STD_SET_S_ID(prjts_rsp->fchdr, fchdr->fh_d_id); FNIC_STD_SET_D_ID(prjts_rsp->fchdr, fchdr->fh_s_id); oxid = FNIC_STD_GET_OX_ID(fchdr); FNIC_STD_SET_OX_ID(prjts_rsp->fchdr, oxid); fnic_send_fcoe_frame(iport, rjt_frame, rjt_frame_size); return; } acc_frame = fdls_alloc_frame(iport); if (acc_frame == NULL) { FNIC_FCS_DBG(KERN_ERR, fnic->host, fnic->fnic_num, "Failed to allocate frame to send ADISC accept"); return; } padisc_acc = (struct fc_std_els_adisc *) (acc_frame + FNIC_ETH_FCOE_HDRS_OFFSET); FNIC_STD_SET_S_ID(padisc_acc->fchdr, fchdr->fh_d_id); FNIC_STD_SET_D_ID(padisc_acc->fchdr, fchdr->fh_s_id); FNIC_STD_SET_F_CTL(padisc_acc->fchdr, FNIC_ELS_REP_FCTL << 16); FNIC_STD_SET_R_CTL(padisc_acc->fchdr, FC_RCTL_ELS_REP); FNIC_STD_SET_TYPE(padisc_acc->fchdr, FC_TYPE_ELS); oxid = FNIC_STD_GET_OX_ID(fchdr); FNIC_STD_SET_OX_ID(padisc_acc->fchdr, oxid); FNIC_STD_SET_RX_ID(padisc_acc->fchdr, FNIC_UNASSIGNED_RXID); padisc_acc->els.adisc_cmd = ELS_LS_ACC; FNIC_STD_SET_NPORT_NAME(&padisc_acc->els.adisc_wwpn, iport->wwpn); FNIC_STD_SET_NODE_NAME(&padisc_acc->els.adisc_wwnn, iport->wwnn); memcpy(padisc_acc->els.adisc_port_id, fchdr->fh_d_id, 3); fnic_send_fcoe_frame(iport, acc_frame, acc_frame_size); } /* * Performs a validation for all FCOE frames and return the frame type */ int fnic_fdls_validate_and_get_frame_type(struct fnic_iport_s *iport, struct fc_frame_header *fchdr) { uint8_t type; uint8_t *fc_payload; uint16_t oxid; uint32_t s_id; uint32_t d_id; struct fnic *fnic = iport->fnic; struct fnic_fdls_fabric_s *fabric = &iport->fabric; int oxid_frame_type; oxid = FNIC_STD_GET_OX_ID(fchdr); fc_payload = (uint8_t *) fchdr + sizeof(struct fc_frame_header); type = *fc_payload; s_id = ntoh24(fchdr->fh_s_id); d_id = ntoh24(fchdr->fh_d_id); /* some common validation */ if (fdls_get_state(fabric) > FDLS_STATE_FABRIC_FLOGI) { if ((iport->fcid != d_id) || (!FNIC_FC_FRAME_CS_CTL(fchdr))) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "invalid frame received. Dropping frame"); return -1; } } /* BLS ABTS response */ if ((fchdr->fh_r_ctl == FC_RCTL_BA_ACC) || (fchdr->fh_r_ctl == FC_RCTL_BA_RJT)) { if (!(FNIC_FC_FRAME_TYPE_BLS(fchdr))) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Received ABTS invalid frame. Dropping frame"); return -1; } if (fdls_is_oxid_fabric_req(oxid)) { if (!(iport->fabric.flags & FNIC_FDLS_FABRIC_ABORT_ISSUED)) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Received unexpected ABTS RSP(oxid:0x%x) from 0x%x. Dropping frame", oxid, s_id); return -1; } return FNIC_FABRIC_BLS_ABTS_RSP; } else if (fdls_is_oxid_fdmi_req(oxid)) { return FNIC_FDMI_BLS_ABTS_RSP; } else if (fdls_is_oxid_tgt_req(oxid)) { return FNIC_TPORT_BLS_ABTS_RSP; } FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Received ABTS rsp with unknown oxid(0x%x) from 0x%x. Dropping frame", oxid, s_id); return -1; } /* BLS ABTS Req */ if ((fchdr->fh_r_ctl == FC_RCTL_BA_ABTS) && (FNIC_FC_FRAME_TYPE_BLS(fchdr))) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Receiving Abort Request from s_id: 0x%x", s_id); return FNIC_BLS_ABTS_REQ; } /* unsolicited requests frames */ if (FNIC_FC_FRAME_UNSOLICITED(fchdr)) { switch (type) { case ELS_LOGO: if ((!FNIC_FC_FRAME_FCTL_FIRST_LAST_SEQINIT(fchdr)) || (!FNIC_FC_FRAME_UNSOLICITED(fchdr)) || (!FNIC_FC_FRAME_TYPE_ELS(fchdr))) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Received LOGO invalid frame. Dropping frame"); return -1; } return FNIC_ELS_LOGO_REQ; case ELS_RSCN: if ((!FNIC_FC_FRAME_FCTL_FIRST_LAST_SEQINIT(fchdr)) || (!FNIC_FC_FRAME_TYPE_ELS(fchdr)) || (!FNIC_FC_FRAME_UNSOLICITED(fchdr))) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Received RSCN invalid FCTL. Dropping frame"); return -1; } if (s_id != FC_FID_FCTRL) FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Received RSCN from target FCTL: 0x%x type: 0x%x s_id: 0x%x.", fchdr->fh_f_ctl[0], fchdr->fh_type, s_id); return FNIC_ELS_RSCN_REQ; case ELS_PLOGI: return FNIC_ELS_PLOGI_REQ; case ELS_ECHO: return FNIC_ELS_ECHO_REQ; case ELS_ADISC: return FNIC_ELS_ADISC; case ELS_RLS: return FNIC_ELS_RLS; case ELS_RRQ: return FNIC_ELS_RRQ; default: FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Unsupported frame (type:0x%02x) from fcid: 0x%x", type, s_id); return FNIC_ELS_UNSUPPORTED_REQ; } } /* solicited response from fabric or target */ oxid_frame_type = FNIC_FRAME_TYPE(oxid); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "oxid frame code: 0x%x, oxid: 0x%x\n", oxid_frame_type, oxid); switch (oxid_frame_type) { case FNIC_FRAME_TYPE_FABRIC_FLOGI: if (type == ELS_LS_ACC) { if ((s_id != FC_FID_FLOGI) || (!FNIC_FC_FRAME_TYPE_ELS(fchdr))) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Received unknown frame. Dropping frame"); return -1; } } return FNIC_FABRIC_FLOGI_RSP; case FNIC_FRAME_TYPE_FABRIC_PLOGI: if (type == ELS_LS_ACC) { if ((s_id != FC_FID_DIR_SERV) || (!FNIC_FC_FRAME_TYPE_ELS(fchdr))) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Received unknown frame. Dropping frame"); return -1; } } return FNIC_FABRIC_PLOGI_RSP; case FNIC_FRAME_TYPE_FABRIC_SCR: if (type == ELS_LS_ACC) { if ((s_id != FC_FID_FCTRL) || (!FNIC_FC_FRAME_TYPE_ELS(fchdr))) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Received unknown frame. Dropping frame"); return -1; } } return FNIC_FABRIC_SCR_RSP; case FNIC_FRAME_TYPE_FABRIC_RPN: if ((s_id != FC_FID_DIR_SERV) || (!FNIC_FC_FRAME_TYPE_FC_GS(fchdr))) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Received unknown frame. Dropping frame"); return -1; } return FNIC_FABRIC_RPN_RSP; case FNIC_FRAME_TYPE_FABRIC_RFT: if ((s_id != FC_FID_DIR_SERV) || (!FNIC_FC_FRAME_TYPE_FC_GS(fchdr))) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Received unknown frame. Dropping frame"); return -1; } return FNIC_FABRIC_RFT_RSP; case FNIC_FRAME_TYPE_FABRIC_RFF: if ((s_id != FC_FID_DIR_SERV) || (!FNIC_FC_FRAME_TYPE_FC_GS(fchdr))) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Received unknown frame. Dropping frame"); return -1; } return FNIC_FABRIC_RFF_RSP; case FNIC_FRAME_TYPE_FABRIC_GPN_FT: if ((s_id != FC_FID_DIR_SERV) || (!FNIC_FC_FRAME_TYPE_FC_GS(fchdr))) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Received unknown frame. Dropping frame"); return -1; } return FNIC_FABRIC_GPN_FT_RSP; case FNIC_FRAME_TYPE_FABRIC_LOGO: return FNIC_FABRIC_LOGO_RSP; case FNIC_FRAME_TYPE_FDMI_PLOGI: return FNIC_FDMI_PLOGI_RSP; case FNIC_FRAME_TYPE_FDMI_RHBA: return FNIC_FDMI_REG_HBA_RSP; case FNIC_FRAME_TYPE_FDMI_RPA: return FNIC_FDMI_RPA_RSP; case FNIC_FRAME_TYPE_TGT_PLOGI: return FNIC_TPORT_PLOGI_RSP; case FNIC_FRAME_TYPE_TGT_PRLI: return FNIC_TPORT_PRLI_RSP; case FNIC_FRAME_TYPE_TGT_ADISC: return FNIC_TPORT_ADISC_RSP; case FNIC_FRAME_TYPE_TGT_LOGO: if (!FNIC_FC_FRAME_TYPE_ELS(fchdr)) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Dropping Unknown frame in tport solicited exchange range type: 0x%x.", fchdr->fh_type); return -1; } return FNIC_TPORT_LOGO_RSP; default: /* Drop the Rx frame and log/stats it */ FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Solicited response: unknown OXID: 0x%x", oxid); return -1; } return -1; } void fnic_fdls_recv_frame(struct fnic_iport_s *iport, void *rx_frame, int len, int fchdr_offset) { struct fc_frame_header *fchdr; uint32_t s_id = 0; uint32_t d_id = 0; struct fnic *fnic = iport->fnic; int frame_type; fchdr = (struct fc_frame_header *) ((uint8_t *) rx_frame + fchdr_offset); s_id = ntoh24(fchdr->fh_s_id); d_id = ntoh24(fchdr->fh_d_id); fnic_debug_dump_fc_frame(fnic, fchdr, len, "Incoming"); frame_type = fnic_fdls_validate_and_get_frame_type(iport, fchdr); /*if we are in flogo drop everything else */ if (iport->fabric.state == FDLS_STATE_FABRIC_LOGO && frame_type != FNIC_FABRIC_LOGO_RSP) return; switch (frame_type) { case FNIC_FABRIC_FLOGI_RSP: fdls_process_flogi_rsp(iport, fchdr, rx_frame); break; case FNIC_FABRIC_PLOGI_RSP: fdls_process_fabric_plogi_rsp(iport, fchdr); break; case FNIC_FDMI_PLOGI_RSP: fdls_process_fdmi_plogi_rsp(iport, fchdr); break; case FNIC_FABRIC_RPN_RSP: fdls_process_rpn_id_rsp(iport, fchdr); break; case FNIC_FABRIC_RFT_RSP: fdls_process_rft_id_rsp(iport, fchdr); break; case FNIC_FABRIC_RFF_RSP: fdls_process_rff_id_rsp(iport, fchdr); break; case FNIC_FABRIC_SCR_RSP: fdls_process_scr_rsp(iport, fchdr); break; case FNIC_FABRIC_GPN_FT_RSP: fdls_process_gpn_ft_rsp(iport, fchdr, len); break; case FNIC_TPORT_PLOGI_RSP: fdls_process_tgt_plogi_rsp(iport, fchdr); break; case FNIC_TPORT_PRLI_RSP: fdls_process_tgt_prli_rsp(iport, fchdr); break; case FNIC_TPORT_ADISC_RSP: fdls_process_tgt_adisc_rsp(iport, fchdr); break; case FNIC_TPORT_BLS_ABTS_RSP: fdls_process_tgt_abts_rsp(iport, fchdr); break; case FNIC_TPORT_LOGO_RSP: /* Logo response from tgt which we have deleted */ FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Logo response from tgt: 0x%x", ntoh24(fchdr->fh_s_id)); break; case FNIC_FABRIC_LOGO_RSP: fdls_process_fabric_logo_rsp(iport, fchdr); break; case FNIC_FABRIC_BLS_ABTS_RSP: fdls_process_fabric_abts_rsp(iport, fchdr); break; case FNIC_FDMI_BLS_ABTS_RSP: fdls_process_fdmi_abts_rsp(iport, fchdr); break; case FNIC_BLS_ABTS_REQ: fdls_process_abts_req(iport, fchdr); break; case FNIC_ELS_UNSUPPORTED_REQ: fdls_process_unsupported_els_req(iport, fchdr); break; case FNIC_ELS_PLOGI_REQ: fdls_process_plogi_req(iport, fchdr); break; case FNIC_ELS_RSCN_REQ: fdls_process_rscn(iport, fchdr); break; case FNIC_ELS_LOGO_REQ: fdls_process_logo_req(iport, fchdr); break; case FNIC_ELS_RRQ: case FNIC_ELS_ECHO_REQ: fdls_process_els_req(iport, fchdr, len); break; case FNIC_ELS_ADISC: fdls_process_adisc_req(iport, fchdr); break; case FNIC_ELS_RLS: fdls_process_rls_req(iport, fchdr); break; case FNIC_FDMI_REG_HBA_RSP: case FNIC_FDMI_RPA_RSP: fdls_process_fdmi_reg_ack(iport, fchdr, frame_type); break; default: FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "s_id: 0x%x d_did: 0x%x", s_id, d_id); FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "Received unknown FCoE frame of len: %d. Dropping frame", len); break; } } void fnic_fdls_disc_init(struct fnic_iport_s *iport) { fdls_reset_oxid_pool(iport); fdls_set_state((&iport->fabric), FDLS_STATE_INIT); } void fnic_fdls_link_down(struct fnic_iport_s *iport) { struct fnic_tport_s *tport, *next; struct fnic *fnic = iport->fnic; FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS processing link down", iport->fcid); fdls_set_state((&iport->fabric), FDLS_STATE_LINKDOWN); iport->fabric.flags = 0; spin_unlock_irqrestore(&fnic->fnic_lock, fnic->lock_flags); fnic_scsi_fcpio_reset(iport->fnic); spin_lock_irqsave(&fnic->fnic_lock, fnic->lock_flags); list_for_each_entry_safe(tport, next, &iport->tport_list, links) { FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "removing rport: 0x%x", tport->fcid); fdls_delete_tport(iport, tport); } if ((fnic_fdmi_support == 1) && (iport->fabric.fdmi_pending > 0)) { del_timer_sync(&iport->fabric.fdmi_timer); iport->fabric.fdmi_pending = 0; } FNIC_FCS_DBG(KERN_INFO, fnic->host, fnic->fnic_num, "0x%x: FDLS finish processing link down", iport->fcid); }