blob: ba17a8f740a95d43b67d5f39a517a4d73180744b [file] [log] [blame]
dea31012005-04-17 16:05:31 -05001/*******************************************************************
2 * This file is part of the Emulex Linux Device Driver for *
James.Smart@Emulex.Comc44ce172005-06-25 10:34:39 -04003 * Fibre Channel Host Bus Adapters. *
James Smart67073c62021-03-01 09:18:21 -08004 * Copyright (C) 2017-2021 Broadcom. All Rights Reserved. The term *
James Smart3e21d1c2018-05-04 20:37:59 -07005 * “Broadcom” refers to Broadcom Inc. and/or its subsidiaries. *
James Smart50611572016-03-31 14:12:34 -07006 * Copyright (C) 2004-2016 Emulex. All rights reserved. *
James.Smart@Emulex.Comc44ce172005-06-25 10:34:39 -04007 * EMULEX and SLI are trademarks of Emulex. *
James Smartd080abe2017-02-12 13:52:39 -08008 * www.broadcom.com *
James.Smart@Emulex.Comc44ce172005-06-25 10:34:39 -04009 * Portions Copyright (C) 2004-2005 Christoph Hellwig *
dea31012005-04-17 16:05:31 -050010 * *
11 * This program is free software; you can redistribute it and/or *
James.Smart@Emulex.Comc44ce172005-06-25 10:34:39 -040012 * modify it under the terms of version 2 of the GNU General *
13 * Public License as published by the Free Software Foundation. *
14 * This program is distributed in the hope that it will be useful. *
15 * ALL EXPRESS OR IMPLIED CONDITIONS, REPRESENTATIONS AND *
16 * WARRANTIES, INCLUDING ANY IMPLIED WARRANTY OF MERCHANTABILITY, *
17 * FITNESS FOR A PARTICULAR PURPOSE, OR NON-INFRINGEMENT, ARE *
18 * DISCLAIMED, EXCEPT TO THE EXTENT THAT SUCH DISCLAIMERS ARE HELD *
19 * TO BE LEGALLY INVALID. See the GNU General Public License for *
20 * more details, a copy of which can be found in the file COPYING *
21 * included with this package. *
dea31012005-04-17 16:05:31 -050022 *******************************************************************/
23
dea31012005-04-17 16:05:31 -050024#include <linux/blkdev.h>
25#include <linux/delay.h>
26#include <linux/dma-mapping.h>
27#include <linux/idr.h>
28#include <linux/interrupt.h>
Paul Gortmakeracf3368f2011-05-27 09:47:43 -040029#include <linux/module.h>
dea31012005-04-17 16:05:31 -050030#include <linux/kthread.h>
31#include <linux/pci.h>
32#include <linux/spinlock.h>
James Smart92d7f7b2007-06-17 19:56:38 -050033#include <linux/ctype.h>
James Smart0d878412009-10-02 15:16:56 -040034#include <linux/aer.h>
Tejun Heo5a0e3ad2010-03-24 17:04:11 +090035#include <linux/slab.h>
James Smart52d52442011-05-24 11:42:45 -040036#include <linux/firmware.h>
James Smart3ef6d242012-01-18 16:23:48 -050037#include <linux/miscdevice.h>
James Smart7bb03bb2013-04-17 20:19:16 -040038#include <linux/percpu.h>
James Smart895427b2017-02-12 13:52:30 -080039#include <linux/msi.h>
James Smart6a828b02019-01-28 11:14:31 -080040#include <linux/irq.h>
Maurizio Lombardi286871a2017-08-23 16:55:48 -070041#include <linux/bitops.h>
James Smart31f06d22019-08-14 16:56:31 -070042#include <linux/crash_dump.h>
James Smartdcaa2132019-11-04 16:57:06 -080043#include <linux/cpu.h>
James Smart93a4d6f2019-11-04 16:57:05 -080044#include <linux/cpuhotplug.h>
dea31012005-04-17 16:05:31 -050045
James.Smart@Emulex.Com91886522005-08-10 15:03:09 -040046#include <scsi/scsi.h>
dea31012005-04-17 16:05:31 -050047#include <scsi/scsi_device.h>
48#include <scsi/scsi_host.h>
49#include <scsi/scsi_transport_fc.h>
James Smart86c67372017-04-21 16:05:04 -070050#include <scsi/scsi_tcq.h>
51#include <scsi/fc/fc_fs.h>
52
James Smartda0436e2009-05-22 14:51:39 -040053#include "lpfc_hw4.h"
dea31012005-04-17 16:05:31 -050054#include "lpfc_hw.h"
55#include "lpfc_sli.h"
James Smartda0436e2009-05-22 14:51:39 -040056#include "lpfc_sli4.h"
James Smartea2151b2008-09-07 11:52:10 -040057#include "lpfc_nl.h"
dea31012005-04-17 16:05:31 -050058#include "lpfc_disc.h"
dea31012005-04-17 16:05:31 -050059#include "lpfc.h"
James Smart895427b2017-02-12 13:52:30 -080060#include "lpfc_scsi.h"
61#include "lpfc_nvme.h"
dea31012005-04-17 16:05:31 -050062#include "lpfc_logmsg.h"
63#include "lpfc_crtn.h"
James Smart92d7f7b2007-06-17 19:56:38 -050064#include "lpfc_vport.h"
dea31012005-04-17 16:05:31 -050065#include "lpfc_version.h"
James Smart12f44452016-07-06 12:36:03 -070066#include "lpfc_ids.h"
dea31012005-04-17 16:05:31 -050067
James Smart93a4d6f2019-11-04 16:57:05 -080068static enum cpuhp_state lpfc_cpuhp_state;
James Smart7bb03bb2013-04-17 20:19:16 -040069/* Used when mapping IRQ vectors in a driver centric manner */
YueHaibingd7b761b2019-05-31 23:28:41 +080070static uint32_t lpfc_present_cpu;
James Smarta5b141a2021-09-27 11:35:18 -070071static bool lpfc_pldv_detect;
James Smart7bb03bb2013-04-17 20:19:16 -040072
James Smart93a4d6f2019-11-04 16:57:05 -080073static void __lpfc_cpuhp_remove(struct lpfc_hba *phba);
74static void lpfc_cpuhp_remove(struct lpfc_hba *phba);
75static void lpfc_cpuhp_add(struct lpfc_hba *phba);
dea31012005-04-17 16:05:31 -050076static void lpfc_get_hba_model_desc(struct lpfc_hba *, uint8_t *, uint8_t *);
77static int lpfc_post_rcv_buf(struct lpfc_hba *);
James Smart5350d872011-10-10 21:33:49 -040078static int lpfc_sli4_queue_verify(struct lpfc_hba *);
James Smartda0436e2009-05-22 14:51:39 -040079static int lpfc_create_bootstrap_mbox(struct lpfc_hba *);
80static int lpfc_setup_endian_order(struct lpfc_hba *);
James Smartda0436e2009-05-22 14:51:39 -040081static void lpfc_destroy_bootstrap_mbox(struct lpfc_hba *);
James Smart8a9d2e82012-05-09 21:16:12 -040082static void lpfc_free_els_sgl_list(struct lpfc_hba *);
James Smartf358dd02017-02-12 13:52:34 -080083static void lpfc_free_nvmet_sgl_list(struct lpfc_hba *);
James Smart8a9d2e82012-05-09 21:16:12 -040084static void lpfc_init_sgl_list(struct lpfc_hba *);
James Smartda0436e2009-05-22 14:51:39 -040085static int lpfc_init_active_sgl_array(struct lpfc_hba *);
86static void lpfc_free_active_sgl(struct lpfc_hba *);
87static int lpfc_hba_down_post_s3(struct lpfc_hba *phba);
88static int lpfc_hba_down_post_s4(struct lpfc_hba *phba);
89static int lpfc_sli4_cq_event_pool_create(struct lpfc_hba *);
90static void lpfc_sli4_cq_event_pool_destroy(struct lpfc_hba *);
91static void lpfc_sli4_cq_event_release_all(struct lpfc_hba *);
James Smart618a5232012-06-12 13:54:36 -040092static void lpfc_sli4_disable_intr(struct lpfc_hba *);
93static uint32_t lpfc_sli4_enable_intr(struct lpfc_hba *, uint32_t);
James Smart1ba981f2014-02-20 09:56:45 -050094static void lpfc_sli4_oas_verify(struct lpfc_hba *phba);
James Smart6a828b02019-01-28 11:14:31 -080095static uint16_t lpfc_find_cpu_handle(struct lpfc_hba *, uint16_t, int);
James Smartaa6ff302019-05-21 17:49:09 -070096static void lpfc_setup_bg(struct lpfc_hba *, struct Scsi_Host *);
James Smart02243832021-08-16 09:28:54 -070097static int lpfc_sli4_cgn_parm_chg_evt(struct lpfc_hba *);
dea31012005-04-17 16:05:31 -050098
99static struct scsi_transport_template *lpfc_transport_template = NULL;
James Smart92d7f7b2007-06-17 19:56:38 -0500100static struct scsi_transport_template *lpfc_vport_transport_template = NULL;
dea31012005-04-17 16:05:31 -0500101static DEFINE_IDR(lpfc_hba_index);
James Smartf358dd02017-02-12 13:52:34 -0800102#define LPFC_NVMET_BUF_POST 254
Gaurav Srivastava5e633302021-06-08 10:05:49 +0530103static int lpfc_vmid_res_alloc(struct lpfc_hba *phba, struct lpfc_vport *vport);
dea31012005-04-17 16:05:31 -0500104
James Smarte59058c2008-08-24 21:49:00 -0400105/**
James Smart3621a712009-04-06 18:47:14 -0400106 * lpfc_config_port_prep - Perform lpfc initialization prior to config port
James Smarte59058c2008-08-24 21:49:00 -0400107 * @phba: pointer to lpfc hba data structure.
108 *
109 * This routine will do LPFC initialization prior to issuing the CONFIG_PORT
110 * mailbox command. It retrieves the revision information from the HBA and
111 * collects the Vital Product Data (VPD) about the HBA for preparing the
112 * configuration of the HBA.
113 *
114 * Return codes:
115 * 0 - success.
116 * -ERESTART - requests the SLI layer to reset the HBA and try again.
117 * Any other value - indicates an error.
118 **/
dea31012005-04-17 16:05:31 -0500119int
James Smart2e0fef82007-06-17 19:56:36 -0500120lpfc_config_port_prep(struct lpfc_hba *phba)
dea31012005-04-17 16:05:31 -0500121{
122 lpfc_vpd_t *vp = &phba->vpd;
123 int i = 0, rc;
124 LPFC_MBOXQ_t *pmb;
125 MAILBOX_t *mb;
126 char *lpfc_vpd_data = NULL;
127 uint16_t offset = 0;
128 static char licensed[56] =
129 "key unlock for use with gnu public licensed code only\0";
James Smart65a29c12006-07-06 15:50:50 -0400130 static int init_key = 1;
dea31012005-04-17 16:05:31 -0500131
132 pmb = mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL);
133 if (!pmb) {
James Smart2e0fef82007-06-17 19:56:36 -0500134 phba->link_state = LPFC_HBA_ERROR;
dea31012005-04-17 16:05:31 -0500135 return -ENOMEM;
136 }
137
James Smart04c68492009-05-22 14:52:52 -0400138 mb = &pmb->u.mb;
James Smart2e0fef82007-06-17 19:56:36 -0500139 phba->link_state = LPFC_INIT_MBX_CMDS;
dea31012005-04-17 16:05:31 -0500140
141 if (lpfc_is_LC_HBA(phba->pcidev->device)) {
James Smart65a29c12006-07-06 15:50:50 -0400142 if (init_key) {
143 uint32_t *ptext = (uint32_t *) licensed;
dea31012005-04-17 16:05:31 -0500144
James Smart65a29c12006-07-06 15:50:50 -0400145 for (i = 0; i < 56; i += sizeof (uint32_t), ptext++)
146 *ptext = cpu_to_be32(*ptext);
147 init_key = 0;
148 }
dea31012005-04-17 16:05:31 -0500149
150 lpfc_read_nv(phba, pmb);
151 memset((char*)mb->un.varRDnvp.rsvd3, 0,
152 sizeof (mb->un.varRDnvp.rsvd3));
153 memcpy((char*)mb->un.varRDnvp.rsvd3, licensed,
154 sizeof (licensed));
155
156 rc = lpfc_sli_issue_mbox(phba, pmb, MBX_POLL);
157
158 if (rc != MBX_SUCCESS) {
Dick Kennedy372c1872020-06-30 14:50:00 -0700159 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smarte8b62012007-08-02 11:10:09 -0400160 "0324 Config Port initialization "
dea31012005-04-17 16:05:31 -0500161 "error, mbxCmd x%x READ_NVPARM, "
162 "mbxStatus x%x\n",
dea31012005-04-17 16:05:31 -0500163 mb->mbxCommand, mb->mbxStatus);
164 mempool_free(pmb, phba->mbox_mem_pool);
165 return -ERESTART;
166 }
167 memcpy(phba->wwnn, (char *)mb->un.varRDnvp.nodename,
James Smart2e0fef82007-06-17 19:56:36 -0500168 sizeof(phba->wwnn));
169 memcpy(phba->wwpn, (char *)mb->un.varRDnvp.portname,
170 sizeof(phba->wwpn));
dea31012005-04-17 16:05:31 -0500171 }
172
Martin Wilckdfb75132018-11-12 09:58:37 +0100173 /*
174 * Clear all option bits except LPFC_SLI3_BG_ENABLED,
175 * which was already set in lpfc_get_cfgparam()
176 */
177 phba->sli3_options &= (uint32_t)LPFC_SLI3_BG_ENABLED;
James Smart92d7f7b2007-06-17 19:56:38 -0500178
dea31012005-04-17 16:05:31 -0500179 /* Setup and issue mailbox READ REV command */
180 lpfc_read_rev(phba, pmb);
181 rc = lpfc_sli_issue_mbox(phba, pmb, MBX_POLL);
182 if (rc != MBX_SUCCESS) {
Dick Kennedy372c1872020-06-30 14:50:00 -0700183 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smarte8b62012007-08-02 11:10:09 -0400184 "0439 Adapter failed to init, mbxCmd x%x "
dea31012005-04-17 16:05:31 -0500185 "READ_REV, mbxStatus x%x\n",
dea31012005-04-17 16:05:31 -0500186 mb->mbxCommand, mb->mbxStatus);
187 mempool_free( pmb, phba->mbox_mem_pool);
188 return -ERESTART;
189 }
190
James Smart92d7f7b2007-06-17 19:56:38 -0500191
James.Smart@Emulex.Com1de933f2005-11-28 11:41:15 -0500192 /*
193 * The value of rr must be 1 since the driver set the cv field to 1.
194 * This setting requires the FW to set all revision fields.
dea31012005-04-17 16:05:31 -0500195 */
James.Smart@Emulex.Com1de933f2005-11-28 11:41:15 -0500196 if (mb->un.varRdRev.rr == 0) {
dea31012005-04-17 16:05:31 -0500197 vp->rev.rBit = 0;
Dick Kennedy372c1872020-06-30 14:50:00 -0700198 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smarte8b62012007-08-02 11:10:09 -0400199 "0440 Adapter failed to init, READ_REV has "
200 "missing revision information.\n");
dea31012005-04-17 16:05:31 -0500201 mempool_free(pmb, phba->mbox_mem_pool);
202 return -ERESTART;
dea31012005-04-17 16:05:31 -0500203 }
204
James Smart495a7142008-06-14 22:52:59 -0400205 if (phba->sli_rev == 3 && !mb->un.varRdRev.v3rsp) {
206 mempool_free(pmb, phba->mbox_mem_pool);
James Smarted957682007-06-17 19:56:37 -0500207 return -EINVAL;
James Smart495a7142008-06-14 22:52:59 -0400208 }
James Smarted957682007-06-17 19:56:37 -0500209
dea31012005-04-17 16:05:31 -0500210 /* Save information as VPD data */
James.Smart@Emulex.Com1de933f2005-11-28 11:41:15 -0500211 vp->rev.rBit = 1;
James Smart92d7f7b2007-06-17 19:56:38 -0500212 memcpy(&vp->sli3Feat, &mb->un.varRdRev.sli3Feat, sizeof(uint32_t));
James.Smart@Emulex.Com1de933f2005-11-28 11:41:15 -0500213 vp->rev.sli1FwRev = mb->un.varRdRev.sli1FwRev;
214 memcpy(vp->rev.sli1FwName, (char*) mb->un.varRdRev.sli1FwName, 16);
215 vp->rev.sli2FwRev = mb->un.varRdRev.sli2FwRev;
216 memcpy(vp->rev.sli2FwName, (char *) mb->un.varRdRev.sli2FwName, 16);
dea31012005-04-17 16:05:31 -0500217 vp->rev.biuRev = mb->un.varRdRev.biuRev;
218 vp->rev.smRev = mb->un.varRdRev.smRev;
219 vp->rev.smFwRev = mb->un.varRdRev.un.smFwRev;
220 vp->rev.endecRev = mb->un.varRdRev.endecRev;
221 vp->rev.fcphHigh = mb->un.varRdRev.fcphHigh;
222 vp->rev.fcphLow = mb->un.varRdRev.fcphLow;
223 vp->rev.feaLevelHigh = mb->un.varRdRev.feaLevelHigh;
224 vp->rev.feaLevelLow = mb->un.varRdRev.feaLevelLow;
225 vp->rev.postKernRev = mb->un.varRdRev.postKernRev;
226 vp->rev.opFwRev = mb->un.varRdRev.opFwRev;
227
James Smart92d7f7b2007-06-17 19:56:38 -0500228 /* If the sli feature level is less then 9, we must
229 * tear down all RPIs and VPIs on link down if NPIV
230 * is enabled.
231 */
232 if (vp->rev.feaLevelHigh < 9)
233 phba->sli3_options |= LPFC_SLI3_VPORT_TEARDOWN;
234
dea31012005-04-17 16:05:31 -0500235 if (lpfc_is_LC_HBA(phba->pcidev->device))
236 memcpy(phba->RandomData, (char *)&mb->un.varWords[24],
237 sizeof (phba->RandomData));
238
dea31012005-04-17 16:05:31 -0500239 /* Get adapter VPD information */
dea31012005-04-17 16:05:31 -0500240 lpfc_vpd_data = kmalloc(DMP_VPD_SIZE, GFP_KERNEL);
241 if (!lpfc_vpd_data)
James Smartd7c255b2008-08-24 21:50:00 -0400242 goto out_free_mbox;
dea31012005-04-17 16:05:31 -0500243 do {
James Smarta0c87cb2009-07-19 10:01:10 -0400244 lpfc_dump_mem(phba, pmb, offset, DMP_REGION_VPD);
dea31012005-04-17 16:05:31 -0500245 rc = lpfc_sli_issue_mbox(phba, pmb, MBX_POLL);
246
247 if (rc != MBX_SUCCESS) {
248 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
James Smarte8b62012007-08-02 11:10:09 -0400249 "0441 VPD not present on adapter, "
dea31012005-04-17 16:05:31 -0500250 "mbxCmd x%x DUMP VPD, mbxStatus x%x\n",
dea31012005-04-17 16:05:31 -0500251 mb->mbxCommand, mb->mbxStatus);
Jamie Wellnitz74b72a52006-02-28 22:33:04 -0500252 mb->un.varDmp.word_cnt = 0;
dea31012005-04-17 16:05:31 -0500253 }
James Smart04c68492009-05-22 14:52:52 -0400254 /* dump mem may return a zero when finished or we got a
255 * mailbox error, either way we are done.
256 */
257 if (mb->un.varDmp.word_cnt == 0)
258 break;
Dick Kennedyd91e3ab2020-06-30 14:49:52 -0700259
James Smarte4ec1022021-04-21 16:45:11 -0700260 if (mb->un.varDmp.word_cnt > DMP_VPD_SIZE - offset)
261 mb->un.varDmp.word_cnt = DMP_VPD_SIZE - offset;
James Smartd7c255b2008-08-24 21:50:00 -0400262 lpfc_sli_pcimem_bcopy(((uint8_t *)mb) + DMP_RSP_OFFSET,
James Smarte4ec1022021-04-21 16:45:11 -0700263 lpfc_vpd_data + offset,
264 mb->un.varDmp.word_cnt);
265 offset += mb->un.varDmp.word_cnt;
266 } while (mb->un.varDmp.word_cnt && offset < DMP_VPD_SIZE);
Dick Kennedyd91e3ab2020-06-30 14:49:52 -0700267
Jamie Wellnitz74b72a52006-02-28 22:33:04 -0500268 lpfc_parse_vpd(phba, lpfc_vpd_data, offset);
dea31012005-04-17 16:05:31 -0500269
270 kfree(lpfc_vpd_data);
dea31012005-04-17 16:05:31 -0500271out_free_mbox:
272 mempool_free(pmb, phba->mbox_mem_pool);
273 return 0;
274}
275
James Smarte59058c2008-08-24 21:49:00 -0400276/**
James Smart3621a712009-04-06 18:47:14 -0400277 * lpfc_config_async_cmpl - Completion handler for config async event mbox cmd
James Smarte59058c2008-08-24 21:49:00 -0400278 * @phba: pointer to lpfc hba data structure.
279 * @pmboxq: pointer to the driver internal queue element for mailbox command.
280 *
281 * This is the completion handler for driver's configuring asynchronous event
282 * mailbox command to the device. If the mailbox command returns successfully,
283 * it will set internal async event support flag to 1; otherwise, it will
284 * set internal async event support flag to 0.
285 **/
James Smart57127f12007-10-27 13:37:05 -0400286static void
287lpfc_config_async_cmpl(struct lpfc_hba * phba, LPFC_MBOXQ_t * pmboxq)
288{
James Smart04c68492009-05-22 14:52:52 -0400289 if (pmboxq->u.mb.mbxStatus == MBX_SUCCESS)
James Smart57127f12007-10-27 13:37:05 -0400290 phba->temp_sensor_support = 1;
291 else
292 phba->temp_sensor_support = 0;
293 mempool_free(pmboxq, phba->mbox_mem_pool);
294 return;
295}
296
James Smarte59058c2008-08-24 21:49:00 -0400297/**
James Smart3621a712009-04-06 18:47:14 -0400298 * lpfc_dump_wakeup_param_cmpl - dump memory mailbox command completion handler
James Smart97207482008-12-04 22:39:19 -0500299 * @phba: pointer to lpfc hba data structure.
300 * @pmboxq: pointer to the driver internal queue element for mailbox command.
301 *
302 * This is the completion handler for dump mailbox command for getting
303 * wake up parameters. When this command complete, the response contain
304 * Option rom version of the HBA. This function translate the version number
305 * into a human readable string and store it in OptionROMVersion.
306 **/
307static void
308lpfc_dump_wakeup_param_cmpl(struct lpfc_hba *phba, LPFC_MBOXQ_t *pmboxq)
309{
310 struct prog_id *prg;
311 uint32_t prog_id_word;
312 char dist = ' ';
313 /* character array used for decoding dist type. */
314 char dist_char[] = "nabx";
315
James Smart04c68492009-05-22 14:52:52 -0400316 if (pmboxq->u.mb.mbxStatus != MBX_SUCCESS) {
James Smart9f1e1b52008-12-04 22:39:40 -0500317 mempool_free(pmboxq, phba->mbox_mem_pool);
James Smart97207482008-12-04 22:39:19 -0500318 return;
James Smart9f1e1b52008-12-04 22:39:40 -0500319 }
James Smart97207482008-12-04 22:39:19 -0500320
321 prg = (struct prog_id *) &prog_id_word;
322
323 /* word 7 contain option rom version */
James Smart04c68492009-05-22 14:52:52 -0400324 prog_id_word = pmboxq->u.mb.un.varWords[7];
James Smart97207482008-12-04 22:39:19 -0500325
326 /* Decode the Option rom version word to a readable string */
327 if (prg->dist < 4)
328 dist = dist_char[prg->dist];
329
330 if ((prg->dist == 3) && (prg->num == 0))
James Smarta2fc4aef2014-09-03 12:57:55 -0400331 snprintf(phba->OptionROMVersion, 32, "%d.%d%d",
James Smart97207482008-12-04 22:39:19 -0500332 prg->ver, prg->rev, prg->lev);
333 else
James Smarta2fc4aef2014-09-03 12:57:55 -0400334 snprintf(phba->OptionROMVersion, 32, "%d.%d%d%c%d",
James Smart97207482008-12-04 22:39:19 -0500335 prg->ver, prg->rev, prg->lev,
336 dist, prg->num);
James Smart9f1e1b52008-12-04 22:39:40 -0500337 mempool_free(pmboxq, phba->mbox_mem_pool);
James Smart97207482008-12-04 22:39:19 -0500338 return;
339}
340
341/**
James Smart05580562011-05-24 11:40:48 -0400342 * lpfc_update_vport_wwn - Updates the fc_nodename, fc_portname,
343 * cfg_soft_wwnn, cfg_soft_wwpn
344 * @vport: pointer to lpfc vport data structure.
345 *
346 *
347 * Return codes
348 * None.
349 **/
350void
351lpfc_update_vport_wwn(struct lpfc_vport *vport)
352{
James Smartaeb3c812017-04-21 16:05:02 -0700353 uint8_t vvvl = vport->fc_sparam.cmn.valid_vendor_ver_level;
354 u32 *fawwpn_key = (u32 *)&vport->fc_sparam.un.vendorVersion[0];
355
James Smart05580562011-05-24 11:40:48 -0400356 /* If the soft name exists then update it using the service params */
357 if (vport->phba->cfg_soft_wwnn)
358 u64_to_wwn(vport->phba->cfg_soft_wwnn,
359 vport->fc_sparam.nodeName.u.wwn);
360 if (vport->phba->cfg_soft_wwpn)
361 u64_to_wwn(vport->phba->cfg_soft_wwpn,
362 vport->fc_sparam.portName.u.wwn);
363
364 /*
365 * If the name is empty or there exists a soft name
366 * then copy the service params name, otherwise use the fc name
367 */
368 if (vport->fc_nodename.u.wwn[0] == 0 || vport->phba->cfg_soft_wwnn)
369 memcpy(&vport->fc_nodename, &vport->fc_sparam.nodeName,
370 sizeof(struct lpfc_name));
371 else
372 memcpy(&vport->fc_sparam.nodeName, &vport->fc_nodename,
373 sizeof(struct lpfc_name));
374
James Smartaeb3c812017-04-21 16:05:02 -0700375 /*
376 * If the port name has changed, then set the Param changes flag
377 * to unreg the login
378 */
379 if (vport->fc_portname.u.wwn[0] != 0 &&
380 memcmp(&vport->fc_portname, &vport->fc_sparam.portName,
381 sizeof(struct lpfc_name)))
382 vport->vport_flag |= FAWWPN_PARAM_CHG;
383
384 if (vport->fc_portname.u.wwn[0] == 0 ||
385 vport->phba->cfg_soft_wwpn ||
386 (vvvl == 1 && cpu_to_be32(*fawwpn_key) == FAPWWN_KEY_VENDOR) ||
387 vport->vport_flag & FAWWPN_SET) {
James Smart05580562011-05-24 11:40:48 -0400388 memcpy(&vport->fc_portname, &vport->fc_sparam.portName,
389 sizeof(struct lpfc_name));
James Smartaeb3c812017-04-21 16:05:02 -0700390 vport->vport_flag &= ~FAWWPN_SET;
391 if (vvvl == 1 && cpu_to_be32(*fawwpn_key) == FAPWWN_KEY_VENDOR)
392 vport->vport_flag |= FAWWPN_SET;
393 }
James Smart05580562011-05-24 11:40:48 -0400394 else
395 memcpy(&vport->fc_sparam.portName, &vport->fc_portname,
396 sizeof(struct lpfc_name));
397}
398
399/**
James Smart3621a712009-04-06 18:47:14 -0400400 * lpfc_config_port_post - Perform lpfc initialization after config port
James Smarte59058c2008-08-24 21:49:00 -0400401 * @phba: pointer to lpfc hba data structure.
402 *
403 * This routine will do LPFC initialization after the CONFIG_PORT mailbox
404 * command call. It performs all internal resource and state setups on the
405 * port: post IOCB buffers, enable appropriate host interrupt attentions,
406 * ELS ring timers, etc.
407 *
408 * Return codes
409 * 0 - success.
410 * Any other value - error.
411 **/
dea31012005-04-17 16:05:31 -0500412int
James Smart2e0fef82007-06-17 19:56:36 -0500413lpfc_config_port_post(struct lpfc_hba *phba)
dea31012005-04-17 16:05:31 -0500414{
James Smart2e0fef82007-06-17 19:56:36 -0500415 struct lpfc_vport *vport = phba->pport;
James Smarta257bf92009-04-06 18:48:10 -0400416 struct Scsi_Host *shost = lpfc_shost_from_vport(vport);
dea31012005-04-17 16:05:31 -0500417 LPFC_MBOXQ_t *pmb;
418 MAILBOX_t *mb;
419 struct lpfc_dmabuf *mp;
420 struct lpfc_sli *psli = &phba->sli;
421 uint32_t status, timeout;
James Smart2e0fef82007-06-17 19:56:36 -0500422 int i, j;
423 int rc;
dea31012005-04-17 16:05:31 -0500424
James Smart7af67052007-10-27 13:38:11 -0400425 spin_lock_irq(&phba->hbalock);
426 /*
427 * If the Config port completed correctly the HBA is not
428 * over heated any more.
429 */
430 if (phba->over_temp_state == HBA_OVER_TEMP)
431 phba->over_temp_state = HBA_NORMAL_TEMP;
432 spin_unlock_irq(&phba->hbalock);
433
dea31012005-04-17 16:05:31 -0500434 pmb = mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL);
435 if (!pmb) {
James Smart2e0fef82007-06-17 19:56:36 -0500436 phba->link_state = LPFC_HBA_ERROR;
dea31012005-04-17 16:05:31 -0500437 return -ENOMEM;
438 }
James Smart04c68492009-05-22 14:52:52 -0400439 mb = &pmb->u.mb;
dea31012005-04-17 16:05:31 -0500440
dea31012005-04-17 16:05:31 -0500441 /* Get login parameters for NID. */
James Smart9f1177a2010-02-26 14:12:57 -0500442 rc = lpfc_read_sparam(phba, pmb, 0);
443 if (rc) {
444 mempool_free(pmb, phba->mbox_mem_pool);
445 return -ENOMEM;
446 }
447
James Smarted957682007-06-17 19:56:37 -0500448 pmb->vport = vport;
dea31012005-04-17 16:05:31 -0500449 if (lpfc_sli_issue_mbox(phba, pmb, MBX_POLL) != MBX_SUCCESS) {
Dick Kennedy372c1872020-06-30 14:50:00 -0700450 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smarte8b62012007-08-02 11:10:09 -0400451 "0448 Adapter failed init, mbxCmd x%x "
dea31012005-04-17 16:05:31 -0500452 "READ_SPARM mbxStatus x%x\n",
dea31012005-04-17 16:05:31 -0500453 mb->mbxCommand, mb->mbxStatus);
James Smart2e0fef82007-06-17 19:56:36 -0500454 phba->link_state = LPFC_HBA_ERROR;
James Smart3e1f0712018-11-29 16:09:29 -0800455 mp = (struct lpfc_dmabuf *)pmb->ctx_buf;
James Smart9f1177a2010-02-26 14:12:57 -0500456 mempool_free(pmb, phba->mbox_mem_pool);
dea31012005-04-17 16:05:31 -0500457 lpfc_mbuf_free(phba, mp->virt, mp->phys);
458 kfree(mp);
459 return -EIO;
460 }
461
James Smart3e1f0712018-11-29 16:09:29 -0800462 mp = (struct lpfc_dmabuf *)pmb->ctx_buf;
dea31012005-04-17 16:05:31 -0500463
James Smart2e0fef82007-06-17 19:56:36 -0500464 memcpy(&vport->fc_sparam, mp->virt, sizeof (struct serv_parm));
dea31012005-04-17 16:05:31 -0500465 lpfc_mbuf_free(phba, mp->virt, mp->phys);
466 kfree(mp);
James Smart3e1f0712018-11-29 16:09:29 -0800467 pmb->ctx_buf = NULL;
James Smart05580562011-05-24 11:40:48 -0400468 lpfc_update_vport_wwn(vport);
James Smarta257bf92009-04-06 18:48:10 -0400469
470 /* Update the fc_host data structures with new wwn. */
471 fc_host_node_name(shost) = wwn_to_u64(vport->fc_nodename.u.wwn);
472 fc_host_port_name(shost) = wwn_to_u64(vport->fc_portname.u.wwn);
James Smart21e9a0a2009-05-22 14:53:21 -0400473 fc_host_max_npiv_vports(shost) = phba->max_vpi;
James Smarta257bf92009-04-06 18:48:10 -0400474
dea31012005-04-17 16:05:31 -0500475 /* If no serial number in VPD data, use low 6 bytes of WWNN */
476 /* This should be consolidated into parse_vpd ? - mr */
477 if (phba->SerialNumber[0] == 0) {
478 uint8_t *outptr;
479
James Smart2e0fef82007-06-17 19:56:36 -0500480 outptr = &vport->fc_nodename.u.s.IEEE[0];
dea31012005-04-17 16:05:31 -0500481 for (i = 0; i < 12; i++) {
482 status = *outptr++;
483 j = ((status & 0xf0) >> 4);
484 if (j <= 9)
485 phba->SerialNumber[i] =
486 (char)((uint8_t) 0x30 + (uint8_t) j);
487 else
488 phba->SerialNumber[i] =
489 (char)((uint8_t) 0x61 + (uint8_t) (j - 10));
490 i++;
491 j = (status & 0xf);
492 if (j <= 9)
493 phba->SerialNumber[i] =
494 (char)((uint8_t) 0x30 + (uint8_t) j);
495 else
496 phba->SerialNumber[i] =
497 (char)((uint8_t) 0x61 + (uint8_t) (j - 10));
498 }
499 }
500
dea31012005-04-17 16:05:31 -0500501 lpfc_read_config(phba, pmb);
James Smarted957682007-06-17 19:56:37 -0500502 pmb->vport = vport;
dea31012005-04-17 16:05:31 -0500503 if (lpfc_sli_issue_mbox(phba, pmb, MBX_POLL) != MBX_SUCCESS) {
Dick Kennedy372c1872020-06-30 14:50:00 -0700504 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smarte8b62012007-08-02 11:10:09 -0400505 "0453 Adapter failed to init, mbxCmd x%x "
dea31012005-04-17 16:05:31 -0500506 "READ_CONFIG, mbxStatus x%x\n",
dea31012005-04-17 16:05:31 -0500507 mb->mbxCommand, mb->mbxStatus);
James Smart2e0fef82007-06-17 19:56:36 -0500508 phba->link_state = LPFC_HBA_ERROR;
dea31012005-04-17 16:05:31 -0500509 mempool_free( pmb, phba->mbox_mem_pool);
510 return -EIO;
511 }
512
James Smarta0c87cb2009-07-19 10:01:10 -0400513 /* Check if the port is disabled */
514 lpfc_sli_read_link_ste(phba);
515
dea31012005-04-17 16:05:31 -0500516 /* Reset the DFT_HBA_Q_DEPTH to the max xri */
James Smartf6770e72020-01-27 16:23:09 -0800517 if (phba->cfg_hba_queue_depth > mb->un.varRdConfig.max_xri) {
James Smart572709e2013-07-15 18:32:43 -0400518 lpfc_printf_log(phba, KERN_WARNING, LOG_INIT,
519 "3359 HBA queue depth changed from %d to %d\n",
James Smartf6770e72020-01-27 16:23:09 -0800520 phba->cfg_hba_queue_depth,
521 mb->un.varRdConfig.max_xri);
522 phba->cfg_hba_queue_depth = mb->un.varRdConfig.max_xri;
James Smart572709e2013-07-15 18:32:43 -0400523 }
dea31012005-04-17 16:05:31 -0500524
525 phba->lmt = mb->un.varRdConfig.lmt;
Jamie Wellnitz74b72a52006-02-28 22:33:04 -0500526
527 /* Get the default values for Model Name and Description */
528 lpfc_get_hba_model_desc(phba, phba->ModelName, phba->ModelDesc);
529
James Smart2e0fef82007-06-17 19:56:36 -0500530 phba->link_state = LPFC_LINK_DOWN;
dea31012005-04-17 16:05:31 -0500531
James Smart0b727fe2007-10-27 13:37:25 -0400532 /* Only process IOCBs on ELS ring till hba_state is READY */
James Smart895427b2017-02-12 13:52:30 -0800533 if (psli->sli3_ring[LPFC_EXTRA_RING].sli.sli3.cmdringaddr)
534 psli->sli3_ring[LPFC_EXTRA_RING].flag |= LPFC_STOP_IOCB_EVENT;
535 if (psli->sli3_ring[LPFC_FCP_RING].sli.sli3.cmdringaddr)
536 psli->sli3_ring[LPFC_FCP_RING].flag |= LPFC_STOP_IOCB_EVENT;
dea31012005-04-17 16:05:31 -0500537
538 /* Post receive buffers for desired rings */
James Smarted957682007-06-17 19:56:37 -0500539 if (phba->sli_rev != 3)
540 lpfc_post_rcv_buf(phba);
dea31012005-04-17 16:05:31 -0500541
James Smart93996272008-08-24 21:50:30 -0400542 /*
543 * Configure HBA MSI-X attention conditions to messages if MSI-X mode
544 */
545 if (phba->intr_type == MSIX) {
546 rc = lpfc_config_msi(phba, pmb);
547 if (rc) {
548 mempool_free(pmb, phba->mbox_mem_pool);
549 return -EIO;
550 }
551 rc = lpfc_sli_issue_mbox(phba, pmb, MBX_POLL);
552 if (rc != MBX_SUCCESS) {
Dick Kennedy372c1872020-06-30 14:50:00 -0700553 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart93996272008-08-24 21:50:30 -0400554 "0352 Config MSI mailbox command "
555 "failed, mbxCmd x%x, mbxStatus x%x\n",
James Smart04c68492009-05-22 14:52:52 -0400556 pmb->u.mb.mbxCommand,
557 pmb->u.mb.mbxStatus);
James Smart93996272008-08-24 21:50:30 -0400558 mempool_free(pmb, phba->mbox_mem_pool);
559 return -EIO;
560 }
561 }
562
James Smart04c68492009-05-22 14:52:52 -0400563 spin_lock_irq(&phba->hbalock);
James Smart93996272008-08-24 21:50:30 -0400564 /* Initialize ERATT handling flag */
565 phba->hba_flag &= ~HBA_ERATT_HANDLED;
566
dea31012005-04-17 16:05:31 -0500567 /* Enable appropriate host interrupts */
James Smart9940b972011-03-11 16:06:12 -0500568 if (lpfc_readl(phba->HCregaddr, &status)) {
569 spin_unlock_irq(&phba->hbalock);
570 return -EIO;
571 }
dea31012005-04-17 16:05:31 -0500572 status |= HC_MBINT_ENA | HC_ERINT_ENA | HC_LAINT_ENA;
573 if (psli->num_rings > 0)
574 status |= HC_R0INT_ENA;
575 if (psli->num_rings > 1)
576 status |= HC_R1INT_ENA;
577 if (psli->num_rings > 2)
578 status |= HC_R2INT_ENA;
579 if (psli->num_rings > 3)
580 status |= HC_R3INT_ENA;
581
James.Smart@Emulex.Com875fbdf2005-11-29 16:32:13 -0500582 if ((phba->cfg_poll & ENABLE_FCP_RING_POLLING) &&
583 (phba->cfg_poll & DISABLE_FCP_RING_INT))
James Smart93996272008-08-24 21:50:30 -0400584 status &= ~(HC_R0INT_ENA);
James.Smart@Emulex.Com875fbdf2005-11-29 16:32:13 -0500585
dea31012005-04-17 16:05:31 -0500586 writel(status, phba->HCregaddr);
587 readl(phba->HCregaddr); /* flush */
James Smart2e0fef82007-06-17 19:56:36 -0500588 spin_unlock_irq(&phba->hbalock);
dea31012005-04-17 16:05:31 -0500589
James Smart93996272008-08-24 21:50:30 -0400590 /* Set up ring-0 (ELS) timer */
591 timeout = phba->fc_ratov * 2;
James Smart256ec0d2013-04-17 20:14:58 -0400592 mod_timer(&vport->els_tmofunc,
593 jiffies + msecs_to_jiffies(1000 * timeout));
James Smart93996272008-08-24 21:50:30 -0400594 /* Set up heart beat (HB) timer */
James Smart256ec0d2013-04-17 20:14:58 -0400595 mod_timer(&phba->hb_tmofunc,
596 jiffies + msecs_to_jiffies(1000 * LPFC_HB_MBOX_INTERVAL));
James Smarta22d73b2021-01-04 10:02:38 -0800597 phba->hba_flag &= ~(HBA_HBEAT_INP | HBA_HBEAT_TMO);
James Smart858c9f62007-06-17 19:56:39 -0500598 phba->last_completion_time = jiffies;
James Smart93996272008-08-24 21:50:30 -0400599 /* Set up error attention (ERATT) polling timer */
James Smart256ec0d2013-04-17 20:14:58 -0400600 mod_timer(&phba->eratt_poll,
James Smart65791f12016-07-06 12:35:56 -0700601 jiffies + msecs_to_jiffies(1000 * phba->eratt_poll_interval));
dea31012005-04-17 16:05:31 -0500602
James Smarta0c87cb2009-07-19 10:01:10 -0400603 if (phba->hba_flag & LINK_DISABLED) {
Dick Kennedy372c1872020-06-30 14:50:00 -0700604 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
605 "2598 Adapter Link is disabled.\n");
James Smarta0c87cb2009-07-19 10:01:10 -0400606 lpfc_down_link(phba, pmb);
607 pmb->mbox_cmpl = lpfc_sli_def_mbox_cmpl;
608 rc = lpfc_sli_issue_mbox(phba, pmb, MBX_NOWAIT);
609 if ((rc != MBX_SUCCESS) && (rc != MBX_BUSY)) {
Dick Kennedy372c1872020-06-30 14:50:00 -0700610 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
611 "2599 Adapter failed to issue DOWN_LINK"
612 " mbox command rc 0x%x\n", rc);
James Smarta0c87cb2009-07-19 10:01:10 -0400613
614 mempool_free(pmb, phba->mbox_mem_pool);
615 return -EIO;
616 }
James Smarte40a02c2010-02-26 14:13:54 -0500617 } else if (phba->cfg_suppress_link_up == LPFC_INITIALIZE_LINK) {
James Smart026abb82011-12-13 13:20:45 -0500618 mempool_free(pmb, phba->mbox_mem_pool);
619 rc = phba->lpfc_hba_init_link(phba, MBX_NOWAIT);
620 if (rc)
621 return rc;
dea31012005-04-17 16:05:31 -0500622 }
623 /* MBOX buffer will be freed in mbox compl */
James Smart57127f12007-10-27 13:37:05 -0400624 pmb = mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL);
James Smart9f1177a2010-02-26 14:12:57 -0500625 if (!pmb) {
626 phba->link_state = LPFC_HBA_ERROR;
627 return -ENOMEM;
628 }
629
James Smart57127f12007-10-27 13:37:05 -0400630 lpfc_config_async(phba, pmb, LPFC_ELS_RING);
631 pmb->mbox_cmpl = lpfc_config_async_cmpl;
632 pmb->vport = phba->pport;
633 rc = lpfc_sli_issue_mbox(phba, pmb, MBX_NOWAIT);
dea31012005-04-17 16:05:31 -0500634
James Smart57127f12007-10-27 13:37:05 -0400635 if ((rc != MBX_BUSY) && (rc != MBX_SUCCESS)) {
Dick Kennedy372c1872020-06-30 14:50:00 -0700636 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart57127f12007-10-27 13:37:05 -0400637 "0456 Adapter failed to issue "
James Smarte4e74272009-07-19 10:01:38 -0400638 "ASYNCEVT_ENABLE mbox status x%x\n",
James Smart57127f12007-10-27 13:37:05 -0400639 rc);
640 mempool_free(pmb, phba->mbox_mem_pool);
641 }
James Smart97207482008-12-04 22:39:19 -0500642
643 /* Get Option rom version */
644 pmb = mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL);
James Smart9f1177a2010-02-26 14:12:57 -0500645 if (!pmb) {
646 phba->link_state = LPFC_HBA_ERROR;
647 return -ENOMEM;
648 }
649
James Smart97207482008-12-04 22:39:19 -0500650 lpfc_dump_wakeup_param(phba, pmb);
651 pmb->mbox_cmpl = lpfc_dump_wakeup_param_cmpl;
652 pmb->vport = phba->pport;
653 rc = lpfc_sli_issue_mbox(phba, pmb, MBX_NOWAIT);
654
655 if ((rc != MBX_BUSY) && (rc != MBX_SUCCESS)) {
Dick Kennedy372c1872020-06-30 14:50:00 -0700656 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
657 "0435 Adapter failed "
James Smarte4e74272009-07-19 10:01:38 -0400658 "to get Option ROM version status x%x\n", rc);
James Smart97207482008-12-04 22:39:19 -0500659 mempool_free(pmb, phba->mbox_mem_pool);
660 }
661
James Smartd7c255b2008-08-24 21:50:00 -0400662 return 0;
James Smartce8b3ce2006-07-06 15:50:36 -0400663}
664
James Smarte59058c2008-08-24 21:49:00 -0400665/**
James Smart7a1dda92021-10-20 14:14:12 -0700666 * lpfc_sli4_refresh_params - update driver copy of params.
667 * @phba: Pointer to HBA context object.
668 *
669 * This is called to refresh driver copy of dynamic fields from the
670 * common_get_sli4_parameters descriptor.
671 **/
672int
673lpfc_sli4_refresh_params(struct lpfc_hba *phba)
674{
675 LPFC_MBOXQ_t *mboxq;
676 struct lpfc_mqe *mqe;
677 struct lpfc_sli4_parameters *mbx_sli4_parameters;
678 int length, rc;
679
680 mboxq = (LPFC_MBOXQ_t *)mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL);
681 if (!mboxq)
682 return -ENOMEM;
683
684 mqe = &mboxq->u.mqe;
685 /* Read the port's SLI4 Config Parameters */
686 length = (sizeof(struct lpfc_mbx_get_sli4_parameters) -
687 sizeof(struct lpfc_sli4_cfg_mhdr));
688 lpfc_sli4_config(phba, mboxq, LPFC_MBOX_SUBSYSTEM_COMMON,
689 LPFC_MBOX_OPCODE_GET_SLI4_PARAMETERS,
690 length, LPFC_SLI4_MBX_EMBED);
691
692 rc = lpfc_sli_issue_mbox(phba, mboxq, MBX_POLL);
693 if (unlikely(rc)) {
694 mempool_free(mboxq, phba->mbox_mem_pool);
695 return rc;
696 }
697 mbx_sli4_parameters = &mqe->un.get_sli4_parameters.sli4_parameters;
698 phba->sli4_hba.pc_sli4_params.mi_ver =
699 bf_get(cfg_mi_ver, mbx_sli4_parameters);
700 phba->sli4_hba.pc_sli4_params.cmf =
701 bf_get(cfg_cmf, mbx_sli4_parameters);
702 phba->sli4_hba.pc_sli4_params.pls =
703 bf_get(cfg_pvl, mbx_sli4_parameters);
704
705 mempool_free(mboxq, phba->mbox_mem_pool);
706 return rc;
707}
708
709/**
James Smart84d1b002010-02-12 14:42:33 -0500710 * lpfc_hba_init_link - Initialize the FC link
711 * @phba: pointer to lpfc hba data structure.
James Smart6e7288d2010-06-07 15:23:35 -0400712 * @flag: mailbox command issue mode - either MBX_POLL or MBX_NOWAIT
James Smart84d1b002010-02-12 14:42:33 -0500713 *
714 * This routine will issue the INIT_LINK mailbox command call.
715 * It is available to other drivers through the lpfc_hba data
716 * structure for use as a delayed link up mechanism with the
717 * module parameter lpfc_suppress_link_up.
718 *
719 * Return code
720 * 0 - success
721 * Any other value - error
722 **/
Rashika Kheriae399b222014-09-03 12:55:28 -0400723static int
James Smart6e7288d2010-06-07 15:23:35 -0400724lpfc_hba_init_link(struct lpfc_hba *phba, uint32_t flag)
James Smart84d1b002010-02-12 14:42:33 -0500725{
James Smart1b511972011-12-13 13:23:09 -0500726 return lpfc_hba_init_link_fc_topology(phba, phba->cfg_topology, flag);
727}
728
729/**
730 * lpfc_hba_init_link_fc_topology - Initialize FC link with desired topology
731 * @phba: pointer to lpfc hba data structure.
732 * @fc_topology: desired fc topology.
733 * @flag: mailbox command issue mode - either MBX_POLL or MBX_NOWAIT
734 *
735 * This routine will issue the INIT_LINK mailbox command call.
736 * It is available to other drivers through the lpfc_hba data
737 * structure for use as a delayed link up mechanism with the
738 * module parameter lpfc_suppress_link_up.
739 *
740 * Return code
741 * 0 - success
742 * Any other value - error
743 **/
744int
745lpfc_hba_init_link_fc_topology(struct lpfc_hba *phba, uint32_t fc_topology,
746 uint32_t flag)
747{
James Smart84d1b002010-02-12 14:42:33 -0500748 struct lpfc_vport *vport = phba->pport;
749 LPFC_MBOXQ_t *pmb;
750 MAILBOX_t *mb;
751 int rc;
752
753 pmb = mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL);
754 if (!pmb) {
755 phba->link_state = LPFC_HBA_ERROR;
756 return -ENOMEM;
757 }
758 mb = &pmb->u.mb;
759 pmb->vport = vport;
760
James Smart026abb82011-12-13 13:20:45 -0500761 if ((phba->cfg_link_speed > LPFC_USER_LINK_SPEED_MAX) ||
762 ((phba->cfg_link_speed == LPFC_USER_LINK_SPEED_1G) &&
763 !(phba->lmt & LMT_1Gb)) ||
764 ((phba->cfg_link_speed == LPFC_USER_LINK_SPEED_2G) &&
765 !(phba->lmt & LMT_2Gb)) ||
766 ((phba->cfg_link_speed == LPFC_USER_LINK_SPEED_4G) &&
767 !(phba->lmt & LMT_4Gb)) ||
768 ((phba->cfg_link_speed == LPFC_USER_LINK_SPEED_8G) &&
769 !(phba->lmt & LMT_8Gb)) ||
770 ((phba->cfg_link_speed == LPFC_USER_LINK_SPEED_10G) &&
771 !(phba->lmt & LMT_10Gb)) ||
772 ((phba->cfg_link_speed == LPFC_USER_LINK_SPEED_16G) &&
James Smartd38dd522015-08-31 16:48:17 -0400773 !(phba->lmt & LMT_16Gb)) ||
774 ((phba->cfg_link_speed == LPFC_USER_LINK_SPEED_32G) &&
James Smartfbd8a6b2018-02-22 08:18:45 -0800775 !(phba->lmt & LMT_32Gb)) ||
776 ((phba->cfg_link_speed == LPFC_USER_LINK_SPEED_64G) &&
777 !(phba->lmt & LMT_64Gb))) {
James Smart026abb82011-12-13 13:20:45 -0500778 /* Reset link speed to auto */
Dick Kennedy372c1872020-06-30 14:50:00 -0700779 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
780 "1302 Invalid speed for this board:%d "
781 "Reset link speed to auto.\n",
782 phba->cfg_link_speed);
James Smart026abb82011-12-13 13:20:45 -0500783 phba->cfg_link_speed = LPFC_USER_LINK_SPEED_AUTO;
784 }
James Smart1b511972011-12-13 13:23:09 -0500785 lpfc_init_link(phba, pmb, fc_topology, phba->cfg_link_speed);
James Smart84d1b002010-02-12 14:42:33 -0500786 pmb->mbox_cmpl = lpfc_sli_def_mbox_cmpl;
James Smart1b511972011-12-13 13:23:09 -0500787 if (phba->sli_rev < LPFC_SLI_REV4)
788 lpfc_set_loopback_flag(phba);
James Smart6e7288d2010-06-07 15:23:35 -0400789 rc = lpfc_sli_issue_mbox(phba, pmb, flag);
James Smart76a95d72010-11-20 23:11:48 -0500790 if ((rc != MBX_BUSY) && (rc != MBX_SUCCESS)) {
Dick Kennedy372c1872020-06-30 14:50:00 -0700791 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
792 "0498 Adapter failed to init, mbxCmd x%x "
793 "INIT_LINK, mbxStatus x%x\n",
794 mb->mbxCommand, mb->mbxStatus);
James Smart76a95d72010-11-20 23:11:48 -0500795 if (phba->sli_rev <= LPFC_SLI_REV3) {
796 /* Clear all interrupt enable conditions */
797 writel(0, phba->HCregaddr);
798 readl(phba->HCregaddr); /* flush */
799 /* Clear all pending interrupts */
800 writel(0xffffffff, phba->HAregaddr);
801 readl(phba->HAregaddr); /* flush */
802 }
James Smart84d1b002010-02-12 14:42:33 -0500803 phba->link_state = LPFC_HBA_ERROR;
James Smart6e7288d2010-06-07 15:23:35 -0400804 if (rc != MBX_BUSY || flag == MBX_POLL)
James Smart84d1b002010-02-12 14:42:33 -0500805 mempool_free(pmb, phba->mbox_mem_pool);
806 return -EIO;
807 }
James Smarte40a02c2010-02-26 14:13:54 -0500808 phba->cfg_suppress_link_up = LPFC_INITIALIZE_LINK;
James Smart6e7288d2010-06-07 15:23:35 -0400809 if (flag == MBX_POLL)
810 mempool_free(pmb, phba->mbox_mem_pool);
James Smart84d1b002010-02-12 14:42:33 -0500811
812 return 0;
813}
814
815/**
816 * lpfc_hba_down_link - this routine downs the FC link
James Smart6e7288d2010-06-07 15:23:35 -0400817 * @phba: pointer to lpfc hba data structure.
818 * @flag: mailbox command issue mode - either MBX_POLL or MBX_NOWAIT
James Smart84d1b002010-02-12 14:42:33 -0500819 *
820 * This routine will issue the DOWN_LINK mailbox command call.
821 * It is available to other drivers through the lpfc_hba data
822 * structure for use to stop the link.
823 *
824 * Return code
825 * 0 - success
826 * Any other value - error
827 **/
Rashika Kheriae399b222014-09-03 12:55:28 -0400828static int
James Smart6e7288d2010-06-07 15:23:35 -0400829lpfc_hba_down_link(struct lpfc_hba *phba, uint32_t flag)
James Smart84d1b002010-02-12 14:42:33 -0500830{
831 LPFC_MBOXQ_t *pmb;
832 int rc;
833
834 pmb = mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL);
835 if (!pmb) {
836 phba->link_state = LPFC_HBA_ERROR;
837 return -ENOMEM;
838 }
839
Dick Kennedy372c1872020-06-30 14:50:00 -0700840 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
841 "0491 Adapter Link is disabled.\n");
James Smart84d1b002010-02-12 14:42:33 -0500842 lpfc_down_link(phba, pmb);
843 pmb->mbox_cmpl = lpfc_sli_def_mbox_cmpl;
James Smart6e7288d2010-06-07 15:23:35 -0400844 rc = lpfc_sli_issue_mbox(phba, pmb, flag);
James Smart84d1b002010-02-12 14:42:33 -0500845 if ((rc != MBX_SUCCESS) && (rc != MBX_BUSY)) {
Dick Kennedy372c1872020-06-30 14:50:00 -0700846 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
847 "2522 Adapter failed to issue DOWN_LINK"
848 " mbox command rc 0x%x\n", rc);
James Smart84d1b002010-02-12 14:42:33 -0500849
850 mempool_free(pmb, phba->mbox_mem_pool);
851 return -EIO;
852 }
James Smart6e7288d2010-06-07 15:23:35 -0400853 if (flag == MBX_POLL)
854 mempool_free(pmb, phba->mbox_mem_pool);
855
James Smart84d1b002010-02-12 14:42:33 -0500856 return 0;
857}
858
859/**
James Smart3621a712009-04-06 18:47:14 -0400860 * lpfc_hba_down_prep - Perform lpfc uninitialization prior to HBA reset
James Smarte59058c2008-08-24 21:49:00 -0400861 * @phba: pointer to lpfc HBA data structure.
862 *
863 * This routine will do LPFC uninitialization before the HBA is reset when
864 * bringing down the SLI Layer.
865 *
866 * Return codes
867 * 0 - success.
868 * Any other value - error.
869 **/
dea31012005-04-17 16:05:31 -0500870int
James Smart2e0fef82007-06-17 19:56:36 -0500871lpfc_hba_down_prep(struct lpfc_hba *phba)
dea31012005-04-17 16:05:31 -0500872{
James Smart1b32f6a2008-02-08 18:49:39 -0500873 struct lpfc_vport **vports;
874 int i;
James Smart3772a992009-05-22 14:50:54 -0400875
876 if (phba->sli_rev <= LPFC_SLI_REV3) {
877 /* Disable interrupts */
878 writel(0, phba->HCregaddr);
879 readl(phba->HCregaddr); /* flush */
880 }
dea31012005-04-17 16:05:31 -0500881
James Smart1b32f6a2008-02-08 18:49:39 -0500882 if (phba->pport->load_flag & FC_UNLOADING)
883 lpfc_cleanup_discovery_resources(phba->pport);
884 else {
885 vports = lpfc_create_vport_work_array(phba);
886 if (vports != NULL)
James Smart3772a992009-05-22 14:50:54 -0400887 for (i = 0; i <= phba->max_vports &&
888 vports[i] != NULL; i++)
James Smart1b32f6a2008-02-08 18:49:39 -0500889 lpfc_cleanup_discovery_resources(vports[i]);
890 lpfc_destroy_vport_work_array(phba, vports);
James Smart7f5f3d02008-02-08 18:50:14 -0500891 }
892 return 0;
dea31012005-04-17 16:05:31 -0500893}
894
James Smarte59058c2008-08-24 21:49:00 -0400895/**
James Smart68e814f2014-05-21 08:04:59 -0400896 * lpfc_sli4_free_sp_events - Cleanup sp_queue_events to free
897 * rspiocb which got deferred
898 *
899 * @phba: pointer to lpfc HBA data structure.
900 *
901 * This routine will cleanup completed slow path events after HBA is reset
902 * when bringing down the SLI Layer.
903 *
904 *
905 * Return codes
906 * void.
907 **/
908static void
909lpfc_sli4_free_sp_events(struct lpfc_hba *phba)
910{
911 struct lpfc_iocbq *rspiocbq;
912 struct hbq_dmabuf *dmabuf;
913 struct lpfc_cq_event *cq_event;
914
915 spin_lock_irq(&phba->hbalock);
916 phba->hba_flag &= ~HBA_SP_QUEUE_EVT;
917 spin_unlock_irq(&phba->hbalock);
918
919 while (!list_empty(&phba->sli4_hba.sp_queue_event)) {
920 /* Get the response iocb from the head of work queue */
921 spin_lock_irq(&phba->hbalock);
922 list_remove_head(&phba->sli4_hba.sp_queue_event,
923 cq_event, struct lpfc_cq_event, list);
924 spin_unlock_irq(&phba->hbalock);
925
926 switch (bf_get(lpfc_wcqe_c_code, &cq_event->cqe.wcqe_cmpl)) {
927 case CQE_CODE_COMPL_WQE:
928 rspiocbq = container_of(cq_event, struct lpfc_iocbq,
929 cq_event);
930 lpfc_sli_release_iocbq(phba, rspiocbq);
931 break;
932 case CQE_CODE_RECEIVE:
933 case CQE_CODE_RECEIVE_V1:
934 dmabuf = container_of(cq_event, struct hbq_dmabuf,
935 cq_event);
936 lpfc_in_buf_free(phba, &dmabuf->dbuf);
937 }
938 }
939}
940
941/**
James Smartbcece5f2014-04-04 13:51:34 -0400942 * lpfc_hba_free_post_buf - Perform lpfc uninitialization after HBA reset
943 * @phba: pointer to lpfc HBA data structure.
944 *
945 * This routine will cleanup posted ELS buffers after the HBA is reset
946 * when bringing down the SLI Layer.
947 *
948 *
949 * Return codes
950 * void.
951 **/
952static void
953lpfc_hba_free_post_buf(struct lpfc_hba *phba)
954{
955 struct lpfc_sli *psli = &phba->sli;
956 struct lpfc_sli_ring *pring;
957 struct lpfc_dmabuf *mp, *next_mp;
James Smart07eab622014-04-04 13:51:44 -0400958 LIST_HEAD(buflist);
959 int count;
James Smartbcece5f2014-04-04 13:51:34 -0400960
961 if (phba->sli3_options & LPFC_SLI3_HBQ_ENABLED)
962 lpfc_sli_hbqbuf_free_all(phba);
963 else {
964 /* Cleanup preposted buffers on the ELS ring */
James Smart895427b2017-02-12 13:52:30 -0800965 pring = &psli->sli3_ring[LPFC_ELS_RING];
James Smart07eab622014-04-04 13:51:44 -0400966 spin_lock_irq(&phba->hbalock);
967 list_splice_init(&pring->postbufq, &buflist);
968 spin_unlock_irq(&phba->hbalock);
969
970 count = 0;
971 list_for_each_entry_safe(mp, next_mp, &buflist, list) {
James Smartbcece5f2014-04-04 13:51:34 -0400972 list_del(&mp->list);
James Smart07eab622014-04-04 13:51:44 -0400973 count++;
James Smartbcece5f2014-04-04 13:51:34 -0400974 lpfc_mbuf_free(phba, mp->virt, mp->phys);
975 kfree(mp);
976 }
James Smart07eab622014-04-04 13:51:44 -0400977
978 spin_lock_irq(&phba->hbalock);
979 pring->postbufq_cnt -= count;
James Smartbcece5f2014-04-04 13:51:34 -0400980 spin_unlock_irq(&phba->hbalock);
981 }
982}
983
984/**
985 * lpfc_hba_clean_txcmplq - Perform lpfc uninitialization after HBA reset
986 * @phba: pointer to lpfc HBA data structure.
987 *
988 * This routine will cleanup the txcmplq after the HBA is reset when bringing
989 * down the SLI Layer.
990 *
991 * Return codes
992 * void
993 **/
994static void
995lpfc_hba_clean_txcmplq(struct lpfc_hba *phba)
996{
997 struct lpfc_sli *psli = &phba->sli;
James Smart895427b2017-02-12 13:52:30 -0800998 struct lpfc_queue *qp = NULL;
James Smartbcece5f2014-04-04 13:51:34 -0400999 struct lpfc_sli_ring *pring;
1000 LIST_HEAD(completions);
1001 int i;
James Smartc1dd9112018-01-30 15:58:57 -08001002 struct lpfc_iocbq *piocb, *next_iocb;
James Smartbcece5f2014-04-04 13:51:34 -04001003
James Smart895427b2017-02-12 13:52:30 -08001004 if (phba->sli_rev != LPFC_SLI_REV4) {
1005 for (i = 0; i < psli->num_rings; i++) {
1006 pring = &psli->sli3_ring[i];
James Smartbcece5f2014-04-04 13:51:34 -04001007 spin_lock_irq(&phba->hbalock);
James Smart895427b2017-02-12 13:52:30 -08001008 /* At this point in time the HBA is either reset or DOA
1009 * Nothing should be on txcmplq as it will
1010 * NEVER complete.
1011 */
1012 list_splice_init(&pring->txcmplq, &completions);
1013 pring->txcmplq_cnt = 0;
James Smartbcece5f2014-04-04 13:51:34 -04001014 spin_unlock_irq(&phba->hbalock);
1015
James Smart895427b2017-02-12 13:52:30 -08001016 lpfc_sli_abort_iocb_ring(phba, pring);
1017 }
James Smartbcece5f2014-04-04 13:51:34 -04001018 /* Cancel all the IOCBs from the completions list */
James Smart895427b2017-02-12 13:52:30 -08001019 lpfc_sli_cancel_iocbs(phba, &completions,
1020 IOSTAT_LOCAL_REJECT, IOERR_SLI_ABORTED);
1021 return;
1022 }
1023 list_for_each_entry(qp, &phba->sli4_hba.lpfc_wq_list, wq_list) {
1024 pring = qp->pring;
1025 if (!pring)
1026 continue;
1027 spin_lock_irq(&pring->ring_lock);
James Smartc1dd9112018-01-30 15:58:57 -08001028 list_for_each_entry_safe(piocb, next_iocb,
1029 &pring->txcmplq, list)
1030 piocb->iocb_flag &= ~LPFC_IO_ON_TXCMPLQ;
James Smart895427b2017-02-12 13:52:30 -08001031 list_splice_init(&pring->txcmplq, &completions);
1032 pring->txcmplq_cnt = 0;
1033 spin_unlock_irq(&pring->ring_lock);
James Smartbcece5f2014-04-04 13:51:34 -04001034 lpfc_sli_abort_iocb_ring(phba, pring);
1035 }
James Smart895427b2017-02-12 13:52:30 -08001036 /* Cancel all the IOCBs from the completions list */
1037 lpfc_sli_cancel_iocbs(phba, &completions,
1038 IOSTAT_LOCAL_REJECT, IOERR_SLI_ABORTED);
James Smartbcece5f2014-04-04 13:51:34 -04001039}
1040
1041/**
James Smart3772a992009-05-22 14:50:54 -04001042 * lpfc_hba_down_post_s3 - Perform lpfc uninitialization after HBA reset
James Smarte59058c2008-08-24 21:49:00 -04001043 * @phba: pointer to lpfc HBA data structure.
1044 *
1045 * This routine will do uninitialization after the HBA is reset when bring
1046 * down the SLI Layer.
1047 *
1048 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -02001049 * 0 - success.
James Smarte59058c2008-08-24 21:49:00 -04001050 * Any other value - error.
1051 **/
James Smart3772a992009-05-22 14:50:54 -04001052static int
1053lpfc_hba_down_post_s3(struct lpfc_hba *phba)
Jamie Wellnitz41415862006-02-28 19:25:27 -05001054{
James Smartbcece5f2014-04-04 13:51:34 -04001055 lpfc_hba_free_post_buf(phba);
1056 lpfc_hba_clean_txcmplq(phba);
Jamie Wellnitz41415862006-02-28 19:25:27 -05001057 return 0;
1058}
James Smart5af5eee2010-10-22 11:06:38 -04001059
James Smartda0436e2009-05-22 14:51:39 -04001060/**
1061 * lpfc_hba_down_post_s4 - Perform lpfc uninitialization after HBA reset
1062 * @phba: pointer to lpfc HBA data structure.
1063 *
1064 * This routine will do uninitialization after the HBA is reset when bring
1065 * down the SLI Layer.
1066 *
1067 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -02001068 * 0 - success.
James Smartda0436e2009-05-22 14:51:39 -04001069 * Any other value - error.
1070 **/
1071static int
1072lpfc_hba_down_post_s4(struct lpfc_hba *phba)
1073{
James Smartc4908502019-01-28 11:14:28 -08001074 struct lpfc_io_buf *psb, *psb_next;
James Smart7cacae22020-03-31 09:50:03 -07001075 struct lpfc_async_xchg_ctx *ctxp, *ctxp_next;
James Smart5e5b5112019-01-28 11:14:22 -08001076 struct lpfc_sli4_hdw_queue *qp;
James Smartda0436e2009-05-22 14:51:39 -04001077 LIST_HEAD(aborts);
James Smart895427b2017-02-12 13:52:30 -08001078 LIST_HEAD(nvme_aborts);
James Smart86c67372017-04-21 16:05:04 -07001079 LIST_HEAD(nvmet_aborts);
James Smart0f65ff62010-02-26 14:14:23 -05001080 struct lpfc_sglq *sglq_entry = NULL;
James Smart5e5b5112019-01-28 11:14:22 -08001081 int cnt, idx;
James Smart0f65ff62010-02-26 14:14:23 -05001082
James Smart895427b2017-02-12 13:52:30 -08001083
1084 lpfc_sli_hbqbuf_free_all(phba);
James Smartbcece5f2014-04-04 13:51:34 -04001085 lpfc_hba_clean_txcmplq(phba);
1086
James Smartda0436e2009-05-22 14:51:39 -04001087 /* At this point in time the HBA is either reset or DOA. Either
1088 * way, nothing should be on lpfc_abts_els_sgl_list, it needs to be
James Smart895427b2017-02-12 13:52:30 -08001089 * on the lpfc_els_sgl_list so that it can either be freed if the
James Smartda0436e2009-05-22 14:51:39 -04001090 * driver is unloading or reposted if the driver is restarting
1091 * the port.
1092 */
James Smarta7892412021-04-11 18:31:15 -07001093
James Smart895427b2017-02-12 13:52:30 -08001094 /* sgl_list_lock required because worker thread uses this
James Smartda0436e2009-05-22 14:51:39 -04001095 * list.
1096 */
James Smarta7892412021-04-11 18:31:15 -07001097 spin_lock_irq(&phba->sli4_hba.sgl_list_lock);
James Smart0f65ff62010-02-26 14:14:23 -05001098 list_for_each_entry(sglq_entry,
1099 &phba->sli4_hba.lpfc_abts_els_sgl_list, list)
1100 sglq_entry->state = SGL_FREED;
1101
James Smartda0436e2009-05-22 14:51:39 -04001102 list_splice_init(&phba->sli4_hba.lpfc_abts_els_sgl_list,
James Smart895427b2017-02-12 13:52:30 -08001103 &phba->sli4_hba.lpfc_els_sgl_list);
1104
James Smartf358dd02017-02-12 13:52:34 -08001105
James Smarta7892412021-04-11 18:31:15 -07001106 spin_unlock_irq(&phba->sli4_hba.sgl_list_lock);
James Smart5e5b5112019-01-28 11:14:22 -08001107
1108 /* abts_xxxx_buf_list_lock required because worker thread uses this
James Smartda0436e2009-05-22 14:51:39 -04001109 * list.
1110 */
James Smarta7892412021-04-11 18:31:15 -07001111 spin_lock_irq(&phba->hbalock);
James Smart5e5b5112019-01-28 11:14:22 -08001112 cnt = 0;
1113 for (idx = 0; idx < phba->cfg_hdw_queue; idx++) {
1114 qp = &phba->sli4_hba.hdwq[idx];
1115
James Smartc00f62e2019-08-14 16:57:11 -07001116 spin_lock(&qp->abts_io_buf_list_lock);
1117 list_splice_init(&qp->lpfc_abts_io_buf_list,
James Smart895427b2017-02-12 13:52:30 -08001118 &aborts);
James Smart895427b2017-02-12 13:52:30 -08001119
James Smart0794d602019-01-28 11:14:19 -08001120 list_for_each_entry_safe(psb, psb_next, &aborts, list) {
James Smart86c67372017-04-21 16:05:04 -07001121 psb->pCmd = NULL;
1122 psb->status = IOSTAT_SUCCESS;
James Smartcf1a1d32017-12-08 17:18:03 -08001123 cnt++;
James Smart86c67372017-04-21 16:05:04 -07001124 }
James Smart5e5b5112019-01-28 11:14:22 -08001125 spin_lock(&qp->io_buf_list_put_lock);
1126 list_splice_init(&aborts, &qp->lpfc_io_buf_list_put);
1127 qp->put_io_bufs += qp->abts_scsi_io_bufs;
James Smartc00f62e2019-08-14 16:57:11 -07001128 qp->put_io_bufs += qp->abts_nvme_io_bufs;
James Smart5e5b5112019-01-28 11:14:22 -08001129 qp->abts_scsi_io_bufs = 0;
James Smartc00f62e2019-08-14 16:57:11 -07001130 qp->abts_nvme_io_bufs = 0;
James Smart5e5b5112019-01-28 11:14:22 -08001131 spin_unlock(&qp->io_buf_list_put_lock);
James Smartc00f62e2019-08-14 16:57:11 -07001132 spin_unlock(&qp->abts_io_buf_list_lock);
James Smart5e5b5112019-01-28 11:14:22 -08001133 }
James Smart731eedc2019-03-12 16:30:12 -07001134 spin_unlock_irq(&phba->hbalock);
James Smart5e5b5112019-01-28 11:14:22 -08001135
1136 if (phba->cfg_enable_fc4_type & LPFC_ENABLE_NVME) {
James Smart731eedc2019-03-12 16:30:12 -07001137 spin_lock_irq(&phba->sli4_hba.abts_nvmet_buf_list_lock);
James Smart5e5b5112019-01-28 11:14:22 -08001138 list_splice_init(&phba->sli4_hba.lpfc_abts_nvmet_ctx_list,
1139 &nvmet_aborts);
James Smart731eedc2019-03-12 16:30:12 -07001140 spin_unlock_irq(&phba->sli4_hba.abts_nvmet_buf_list_lock);
James Smart86c67372017-04-21 16:05:04 -07001141 list_for_each_entry_safe(ctxp, ctxp_next, &nvmet_aborts, list) {
James Smart7b7f5512020-03-31 09:50:04 -07001142 ctxp->flag &= ~(LPFC_NVME_XBUSY | LPFC_NVME_ABORT_OP);
James Smart6c621a22017-05-15 15:20:45 -07001143 lpfc_nvmet_ctxbuf_post(phba, ctxp->ctxbuf);
James Smart86c67372017-04-21 16:05:04 -07001144 }
James Smart895427b2017-02-12 13:52:30 -08001145 }
James Smart895427b2017-02-12 13:52:30 -08001146
James Smart68e814f2014-05-21 08:04:59 -04001147 lpfc_sli4_free_sp_events(phba);
James Smart5e5b5112019-01-28 11:14:22 -08001148 return cnt;
James Smartda0436e2009-05-22 14:51:39 -04001149}
1150
1151/**
1152 * lpfc_hba_down_post - Wrapper func for hba down post routine
1153 * @phba: pointer to lpfc HBA data structure.
1154 *
1155 * This routine wraps the actual SLI3 or SLI4 routine for performing
1156 * uninitialization after the HBA is reset when bring down the SLI Layer.
1157 *
1158 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -02001159 * 0 - success.
James Smartda0436e2009-05-22 14:51:39 -04001160 * Any other value - error.
1161 **/
1162int
1163lpfc_hba_down_post(struct lpfc_hba *phba)
1164{
1165 return (*phba->lpfc_hba_down_post)(phba);
1166}
Jamie Wellnitz41415862006-02-28 19:25:27 -05001167
James Smarte59058c2008-08-24 21:49:00 -04001168/**
James Smart3621a712009-04-06 18:47:14 -04001169 * lpfc_hb_timeout - The HBA-timer timeout handler
Lee Jonesfe614ac2020-07-23 13:24:22 +01001170 * @t: timer context used to obtain the pointer to lpfc hba data structure.
James Smarte59058c2008-08-24 21:49:00 -04001171 *
1172 * This is the HBA-timer timeout handler registered to the lpfc driver. When
1173 * this timer fires, a HBA timeout event shall be posted to the lpfc driver
1174 * work-port-events bitmap and the worker thread is notified. This timeout
1175 * event will be used by the worker thread to invoke the actual timeout
1176 * handler routine, lpfc_hb_timeout_handler. Any periodical operations will
1177 * be performed in the timeout handler and the HBA timeout event bit shall
1178 * be cleared by the worker thread after it has taken the event bitmap out.
1179 **/
Adrian Bunka6ababd2007-11-05 18:07:33 +01001180static void
Kees Cookf22eb4d2017-09-06 20:24:26 -07001181lpfc_hb_timeout(struct timer_list *t)
James Smart858c9f62007-06-17 19:56:39 -05001182{
1183 struct lpfc_hba *phba;
James Smart5e9d9b82008-06-14 22:52:53 -04001184 uint32_t tmo_posted;
James Smart858c9f62007-06-17 19:56:39 -05001185 unsigned long iflag;
1186
Kees Cookf22eb4d2017-09-06 20:24:26 -07001187 phba = from_timer(phba, t, hb_tmofunc);
James Smart93996272008-08-24 21:50:30 -04001188
1189 /* Check for heart beat timeout conditions */
James Smart858c9f62007-06-17 19:56:39 -05001190 spin_lock_irqsave(&phba->pport->work_port_lock, iflag);
James Smart5e9d9b82008-06-14 22:52:53 -04001191 tmo_posted = phba->pport->work_port_events & WORKER_HB_TMO;
1192 if (!tmo_posted)
James Smart858c9f62007-06-17 19:56:39 -05001193 phba->pport->work_port_events |= WORKER_HB_TMO;
1194 spin_unlock_irqrestore(&phba->pport->work_port_lock, iflag);
1195
James Smart93996272008-08-24 21:50:30 -04001196 /* Tell the worker thread there is work to do */
James Smart5e9d9b82008-06-14 22:52:53 -04001197 if (!tmo_posted)
1198 lpfc_worker_wake_up(phba);
James Smart858c9f62007-06-17 19:56:39 -05001199 return;
1200}
1201
James Smarte59058c2008-08-24 21:49:00 -04001202/**
James Smart19ca7602010-11-20 23:11:55 -05001203 * lpfc_rrq_timeout - The RRQ-timer timeout handler
Lee Jonesfe614ac2020-07-23 13:24:22 +01001204 * @t: timer context used to obtain the pointer to lpfc hba data structure.
James Smart19ca7602010-11-20 23:11:55 -05001205 *
1206 * This is the RRQ-timer timeout handler registered to the lpfc driver. When
1207 * this timer fires, a RRQ timeout event shall be posted to the lpfc driver
1208 * work-port-events bitmap and the worker thread is notified. This timeout
1209 * event will be used by the worker thread to invoke the actual timeout
1210 * handler routine, lpfc_rrq_handler. Any periodical operations will
1211 * be performed in the timeout handler and the RRQ timeout event bit shall
1212 * be cleared by the worker thread after it has taken the event bitmap out.
1213 **/
1214static void
Kees Cookf22eb4d2017-09-06 20:24:26 -07001215lpfc_rrq_timeout(struct timer_list *t)
James Smart19ca7602010-11-20 23:11:55 -05001216{
1217 struct lpfc_hba *phba;
James Smart19ca7602010-11-20 23:11:55 -05001218 unsigned long iflag;
1219
Kees Cookf22eb4d2017-09-06 20:24:26 -07001220 phba = from_timer(phba, t, rrq_tmr);
James Smart19ca7602010-11-20 23:11:55 -05001221 spin_lock_irqsave(&phba->pport->work_port_lock, iflag);
James Smart06918ac2014-02-20 09:57:57 -05001222 if (!(phba->pport->load_flag & FC_UNLOADING))
1223 phba->hba_flag |= HBA_RRQ_ACTIVE;
1224 else
1225 phba->hba_flag &= ~HBA_RRQ_ACTIVE;
James Smart19ca7602010-11-20 23:11:55 -05001226 spin_unlock_irqrestore(&phba->pport->work_port_lock, iflag);
James Smart06918ac2014-02-20 09:57:57 -05001227
1228 if (!(phba->pport->load_flag & FC_UNLOADING))
1229 lpfc_worker_wake_up(phba);
James Smart19ca7602010-11-20 23:11:55 -05001230}
1231
1232/**
James Smart3621a712009-04-06 18:47:14 -04001233 * lpfc_hb_mbox_cmpl - The lpfc heart-beat mailbox command callback function
James Smarte59058c2008-08-24 21:49:00 -04001234 * @phba: pointer to lpfc hba data structure.
1235 * @pmboxq: pointer to the driver internal queue element for mailbox command.
1236 *
1237 * This is the callback function to the lpfc heart-beat mailbox command.
1238 * If configured, the lpfc driver issues the heart-beat mailbox command to
1239 * the HBA every LPFC_HB_MBOX_INTERVAL (current 5) seconds. At the time the
1240 * heart-beat mailbox command is issued, the driver shall set up heart-beat
1241 * timeout timer to LPFC_HB_MBOX_TIMEOUT (current 30) seconds and marks
1242 * heart-beat outstanding state. Once the mailbox command comes back and
1243 * no error conditions detected, the heart-beat mailbox command timer is
1244 * reset to LPFC_HB_MBOX_INTERVAL seconds and the heart-beat outstanding
1245 * state is cleared for the next heart-beat. If the timer expired with the
1246 * heart-beat outstanding state set, the driver will put the HBA offline.
1247 **/
James Smart858c9f62007-06-17 19:56:39 -05001248static void
1249lpfc_hb_mbox_cmpl(struct lpfc_hba * phba, LPFC_MBOXQ_t * pmboxq)
1250{
1251 unsigned long drvr_flag;
1252
1253 spin_lock_irqsave(&phba->hbalock, drvr_flag);
James Smarta22d73b2021-01-04 10:02:38 -08001254 phba->hba_flag &= ~(HBA_HBEAT_INP | HBA_HBEAT_TMO);
James Smart858c9f62007-06-17 19:56:39 -05001255 spin_unlock_irqrestore(&phba->hbalock, drvr_flag);
1256
James Smarta22d73b2021-01-04 10:02:38 -08001257 /* Check and reset heart-beat timer if necessary */
James Smart858c9f62007-06-17 19:56:39 -05001258 mempool_free(pmboxq, phba->mbox_mem_pool);
1259 if (!(phba->pport->fc_flag & FC_OFFLINE_MODE) &&
1260 !(phba->link_state == LPFC_HBA_ERROR) &&
James Smart51ef4c22007-08-02 11:10:31 -04001261 !(phba->pport->load_flag & FC_UNLOADING))
James Smart858c9f62007-06-17 19:56:39 -05001262 mod_timer(&phba->hb_tmofunc,
James Smart256ec0d2013-04-17 20:14:58 -04001263 jiffies +
1264 msecs_to_jiffies(1000 * LPFC_HB_MBOX_INTERVAL));
James Smart858c9f62007-06-17 19:56:39 -05001265 return;
1266}
1267
Lee Jonesfe614ac2020-07-23 13:24:22 +01001268/*
Dick Kennedy317aeb82020-06-30 14:49:59 -07001269 * lpfc_idle_stat_delay_work - idle_stat tracking
1270 *
1271 * This routine tracks per-cq idle_stat and determines polling decisions.
1272 *
1273 * Return codes:
1274 * None
1275 **/
1276static void
1277lpfc_idle_stat_delay_work(struct work_struct *work)
1278{
1279 struct lpfc_hba *phba = container_of(to_delayed_work(work),
1280 struct lpfc_hba,
1281 idle_stat_delay_work);
1282 struct lpfc_queue *cq;
1283 struct lpfc_sli4_hdw_queue *hdwq;
1284 struct lpfc_idle_stat *idle_stat;
1285 u32 i, idle_percent;
1286 u64 wall, wall_idle, diff_wall, diff_idle, busy_time;
1287
1288 if (phba->pport->load_flag & FC_UNLOADING)
1289 return;
1290
1291 if (phba->link_state == LPFC_HBA_ERROR ||
James Smart9064aeb2021-08-16 09:28:50 -07001292 phba->pport->fc_flag & FC_OFFLINE_MODE ||
1293 phba->cmf_active_mode != LPFC_CFG_OFF)
Dick Kennedy317aeb82020-06-30 14:49:59 -07001294 goto requeue;
1295
1296 for_each_present_cpu(i) {
1297 hdwq = &phba->sli4_hba.hdwq[phba->sli4_hba.cpu_map[i].hdwq];
1298 cq = hdwq->io_cq;
1299
1300 /* Skip if we've already handled this cq's primary CPU */
1301 if (cq->chann != i)
1302 continue;
1303
1304 idle_stat = &phba->sli4_hba.idle_stat[i];
1305
1306 /* get_cpu_idle_time returns values as running counters. Thus,
1307 * to know the amount for this period, the prior counter values
1308 * need to be subtracted from the current counter values.
1309 * From there, the idle time stat can be calculated as a
1310 * percentage of 100 - the sum of the other consumption times.
1311 */
1312 wall_idle = get_cpu_idle_time(i, &wall, 1);
1313 diff_idle = wall_idle - idle_stat->prev_idle;
1314 diff_wall = wall - idle_stat->prev_wall;
1315
1316 if (diff_wall <= diff_idle)
1317 busy_time = 0;
1318 else
1319 busy_time = diff_wall - diff_idle;
1320
1321 idle_percent = div64_u64(100 * busy_time, diff_wall);
1322 idle_percent = 100 - idle_percent;
1323
1324 if (idle_percent < 15)
1325 cq->poll_mode = LPFC_QUEUE_WORK;
1326 else
1327 cq->poll_mode = LPFC_IRQ_POLL;
1328
1329 idle_stat->prev_idle = wall_idle;
1330 idle_stat->prev_wall = wall;
1331 }
1332
1333requeue:
1334 schedule_delayed_work(&phba->idle_stat_delay_work,
1335 msecs_to_jiffies(LPFC_IDLE_STAT_DELAY));
1336}
1337
James Smart32517fc2019-01-28 11:14:33 -08001338static void
1339lpfc_hb_eq_delay_work(struct work_struct *work)
1340{
1341 struct lpfc_hba *phba = container_of(to_delayed_work(work),
1342 struct lpfc_hba, eq_delay_work);
1343 struct lpfc_eq_intr_info *eqi, *eqi_new;
1344 struct lpfc_queue *eq, *eq_next;
James Smart8156d372019-10-18 14:18:26 -07001345 unsigned char *ena_delay = NULL;
James Smart32517fc2019-01-28 11:14:33 -08001346 uint32_t usdelay;
1347 int i;
1348
1349 if (!phba->cfg_auto_imax || phba->pport->load_flag & FC_UNLOADING)
1350 return;
1351
1352 if (phba->link_state == LPFC_HBA_ERROR ||
1353 phba->pport->fc_flag & FC_OFFLINE_MODE)
1354 goto requeue;
1355
James Smart8156d372019-10-18 14:18:26 -07001356 ena_delay = kcalloc(phba->sli4_hba.num_possible_cpu, sizeof(*ena_delay),
1357 GFP_KERNEL);
1358 if (!ena_delay)
James Smart32517fc2019-01-28 11:14:33 -08001359 goto requeue;
1360
James Smart8156d372019-10-18 14:18:26 -07001361 for (i = 0; i < phba->cfg_irq_chann; i++) {
1362 /* Get the EQ corresponding to the IRQ vector */
1363 eq = phba->sli4_hba.hba_eq_hdl[i].eq;
1364 if (!eq)
1365 continue;
1366 if (eq->q_mode || eq->q_flag & HBA_EQ_DELAY_CHK) {
1367 eq->q_flag &= ~HBA_EQ_DELAY_CHK;
1368 ena_delay[eq->last_cpu] = 1;
James Smart8d34a592019-08-14 16:56:35 -07001369 }
James Smart8156d372019-10-18 14:18:26 -07001370 }
James Smart32517fc2019-01-28 11:14:33 -08001371
1372 for_each_present_cpu(i) {
James Smart32517fc2019-01-28 11:14:33 -08001373 eqi = per_cpu_ptr(phba->sli4_hba.eq_info, i);
James Smart8156d372019-10-18 14:18:26 -07001374 if (ena_delay[i]) {
1375 usdelay = (eqi->icnt >> 10) * LPFC_EQ_DELAY_STEP;
1376 if (usdelay > LPFC_MAX_AUTO_EQ_DELAY)
1377 usdelay = LPFC_MAX_AUTO_EQ_DELAY;
1378 } else {
1379 usdelay = 0;
James Smart8d34a592019-08-14 16:56:35 -07001380 }
James Smart32517fc2019-01-28 11:14:33 -08001381
James Smart32517fc2019-01-28 11:14:33 -08001382 eqi->icnt = 0;
1383
1384 list_for_each_entry_safe(eq, eq_next, &eqi->list, cpu_list) {
James Smart8156d372019-10-18 14:18:26 -07001385 if (unlikely(eq->last_cpu != i)) {
James Smart32517fc2019-01-28 11:14:33 -08001386 eqi_new = per_cpu_ptr(phba->sli4_hba.eq_info,
1387 eq->last_cpu);
1388 list_move_tail(&eq->cpu_list, &eqi_new->list);
1389 continue;
1390 }
1391 if (usdelay != eq->q_mode)
1392 lpfc_modify_hba_eq_delay(phba, eq->hdwq, 1,
1393 usdelay);
1394 }
1395 }
1396
James Smart8156d372019-10-18 14:18:26 -07001397 kfree(ena_delay);
James Smart32517fc2019-01-28 11:14:33 -08001398
1399requeue:
1400 queue_delayed_work(phba->wq, &phba->eq_delay_work,
1401 msecs_to_jiffies(LPFC_EQ_DELAY_MSECS));
1402}
1403
James Smarte59058c2008-08-24 21:49:00 -04001404/**
James Smartc4908502019-01-28 11:14:28 -08001405 * lpfc_hb_mxp_handler - Multi-XRI pools handler to adjust XRI distribution
1406 * @phba: pointer to lpfc hba data structure.
1407 *
1408 * For each heartbeat, this routine does some heuristic methods to adjust
1409 * XRI distribution. The goal is to fully utilize free XRIs.
1410 **/
1411static void lpfc_hb_mxp_handler(struct lpfc_hba *phba)
1412{
1413 u32 i;
1414 u32 hwq_count;
1415
1416 hwq_count = phba->cfg_hdw_queue;
1417 for (i = 0; i < hwq_count; i++) {
1418 /* Adjust XRIs in private pool */
1419 lpfc_adjust_pvt_pool_count(phba, i);
1420
1421 /* Adjust high watermark */
1422 lpfc_adjust_high_watermark(phba, i);
1423
1424#ifdef LPFC_MXP_STAT
1425 /* Snapshot pbl, pvt and busy count */
1426 lpfc_snapshot_mxp(phba, i);
1427#endif
1428 }
1429}
1430
James Smarte59058c2008-08-24 21:49:00 -04001431/**
James Smarta22d73b2021-01-04 10:02:38 -08001432 * lpfc_issue_hb_mbox - Issues heart-beat mailbox command
1433 * @phba: pointer to lpfc hba data structure.
1434 *
1435 * If a HB mbox is not already in progrees, this routine will allocate
1436 * a LPFC_MBOXQ_t, populate it with a MBX_HEARTBEAT (0x31) command,
1437 * and issue it. The HBA_HBEAT_INP flag means the command is in progress.
1438 **/
1439int
1440lpfc_issue_hb_mbox(struct lpfc_hba *phba)
1441{
1442 LPFC_MBOXQ_t *pmboxq;
1443 int retval;
1444
1445 /* Is a Heartbeat mbox already in progress */
1446 if (phba->hba_flag & HBA_HBEAT_INP)
1447 return 0;
1448
1449 pmboxq = mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL);
1450 if (!pmboxq)
1451 return -ENOMEM;
1452
1453 lpfc_heart_beat(phba, pmboxq);
1454 pmboxq->mbox_cmpl = lpfc_hb_mbox_cmpl;
1455 pmboxq->vport = phba->pport;
1456 retval = lpfc_sli_issue_mbox(phba, pmboxq, MBX_NOWAIT);
1457
1458 if (retval != MBX_BUSY && retval != MBX_SUCCESS) {
1459 mempool_free(pmboxq, phba->mbox_mem_pool);
1460 return -ENXIO;
1461 }
1462 phba->hba_flag |= HBA_HBEAT_INP;
1463
1464 return 0;
1465}
1466
1467/**
1468 * lpfc_issue_hb_tmo - Signals heartbeat timer to issue mbox command
1469 * @phba: pointer to lpfc hba data structure.
1470 *
1471 * The heartbeat timer (every 5 sec) will fire. If the HBA_HBEAT_TMO
1472 * flag is set, it will force a MBX_HEARTBEAT mbox command, regardless
1473 * of the value of lpfc_enable_hba_heartbeat.
1474 * If lpfc_enable_hba_heartbeat is set, the timeout routine will always
1475 * try to issue a MBX_HEARTBEAT mbox command.
1476 **/
1477void
1478lpfc_issue_hb_tmo(struct lpfc_hba *phba)
1479{
1480 if (phba->cfg_enable_hba_heartbeat)
1481 return;
1482 phba->hba_flag |= HBA_HBEAT_TMO;
1483}
1484
1485/**
James Smart3621a712009-04-06 18:47:14 -04001486 * lpfc_hb_timeout_handler - The HBA-timer timeout handler
James Smarte59058c2008-08-24 21:49:00 -04001487 * @phba: pointer to lpfc hba data structure.
1488 *
1489 * This is the actual HBA-timer timeout handler to be invoked by the worker
1490 * thread whenever the HBA timer fired and HBA-timeout event posted. This
1491 * handler performs any periodic operations needed for the device. If such
1492 * periodic event has already been attended to either in the interrupt handler
1493 * or by processing slow-ring or fast-ring events within the HBA-timer
1494 * timeout window (LPFC_HB_MBOX_INTERVAL), this handler just simply resets
1495 * the timer for the next timeout period. If lpfc heart-beat mailbox command
1496 * is configured and there is no heart-beat mailbox command outstanding, a
1497 * heart-beat mailbox is issued and timer set properly. Otherwise, if there
1498 * has been a heart-beat mailbox command outstanding, the HBA shall be put
1499 * to offline.
1500 **/
James Smart858c9f62007-06-17 19:56:39 -05001501void
1502lpfc_hb_timeout_handler(struct lpfc_hba *phba)
1503{
James Smart45ed1192009-10-02 15:17:02 -04001504 struct lpfc_vport **vports;
James Smart0ff10d42008-01-11 01:52:36 -05001505 struct lpfc_dmabuf *buf_ptr;
James Smarta22d73b2021-01-04 10:02:38 -08001506 int retval = 0;
1507 int i, tmo;
James Smart858c9f62007-06-17 19:56:39 -05001508 struct lpfc_sli *psli = &phba->sli;
James Smart0ff10d42008-01-11 01:52:36 -05001509 LIST_HEAD(completions);
James Smart858c9f62007-06-17 19:56:39 -05001510
James Smartc4908502019-01-28 11:14:28 -08001511 if (phba->cfg_xri_rebalancing) {
1512 /* Multi-XRI pools handler */
1513 lpfc_hb_mxp_handler(phba);
1514 }
James Smart858c9f62007-06-17 19:56:39 -05001515
James Smart45ed1192009-10-02 15:17:02 -04001516 vports = lpfc_create_vport_work_array(phba);
1517 if (vports != NULL)
James Smart4258e982015-12-16 18:11:58 -05001518 for (i = 0; i <= phba->max_vports && vports[i] != NULL; i++) {
James Smart45ed1192009-10-02 15:17:02 -04001519 lpfc_rcv_seq_check_edtov(vports[i]);
James Smarte3ba04c2019-12-18 15:58:02 -08001520 lpfc_fdmi_change_check(vports[i]);
James Smart4258e982015-12-16 18:11:58 -05001521 }
James Smart45ed1192009-10-02 15:17:02 -04001522 lpfc_destroy_vport_work_array(phba, vports);
1523
James Smart858c9f62007-06-17 19:56:39 -05001524 if ((phba->link_state == LPFC_HBA_ERROR) ||
James Smart51ef4c22007-08-02 11:10:31 -04001525 (phba->pport->load_flag & FC_UNLOADING) ||
James Smart858c9f62007-06-17 19:56:39 -05001526 (phba->pport->fc_flag & FC_OFFLINE_MODE))
1527 return;
1528
James Smart0ff10d42008-01-11 01:52:36 -05001529 if (phba->elsbuf_cnt &&
1530 (phba->elsbuf_cnt == phba->elsbuf_prev_cnt)) {
1531 spin_lock_irq(&phba->hbalock);
1532 list_splice_init(&phba->elsbuf, &completions);
1533 phba->elsbuf_cnt = 0;
1534 phba->elsbuf_prev_cnt = 0;
1535 spin_unlock_irq(&phba->hbalock);
1536
1537 while (!list_empty(&completions)) {
1538 list_remove_head(&completions, buf_ptr,
1539 struct lpfc_dmabuf, list);
1540 lpfc_mbuf_free(phba, buf_ptr->virt, buf_ptr->phys);
1541 kfree(buf_ptr);
1542 }
1543 }
1544 phba->elsbuf_prev_cnt = phba->elsbuf_cnt;
1545
James Smart858c9f62007-06-17 19:56:39 -05001546 /* If there is no heart beat outstanding, issue a heartbeat command */
James Smart13815c82008-01-11 01:52:48 -05001547 if (phba->cfg_enable_hba_heartbeat) {
James Smarta22d73b2021-01-04 10:02:38 -08001548 /* If IOs are completing, no need to issue a MBX_HEARTBEAT */
1549 spin_lock_irq(&phba->pport->work_port_lock);
1550 if (time_after(phba->last_completion_time +
1551 msecs_to_jiffies(1000 * LPFC_HB_MBOX_INTERVAL),
1552 jiffies)) {
1553 spin_unlock_irq(&phba->pport->work_port_lock);
1554 if (phba->hba_flag & HBA_HBEAT_INP)
1555 tmo = (1000 * LPFC_HB_MBOX_TIMEOUT);
1556 else
1557 tmo = (1000 * LPFC_HB_MBOX_INTERVAL);
1558 goto out;
1559 }
1560 spin_unlock_irq(&phba->pport->work_port_lock);
1561
1562 /* Check if a MBX_HEARTBEAT is already in progress */
1563 if (phba->hba_flag & HBA_HBEAT_INP) {
1564 /*
1565 * If heart beat timeout called with HBA_HBEAT_INP set
1566 * we need to give the hb mailbox cmd a chance to
1567 * complete or TMO.
1568 */
1569 lpfc_printf_log(phba, KERN_WARNING, LOG_INIT,
1570 "0459 Adapter heartbeat still outstanding: "
1571 "last compl time was %d ms.\n",
1572 jiffies_to_msecs(jiffies
1573 - phba->last_completion_time));
1574 tmo = (1000 * LPFC_HB_MBOX_TIMEOUT);
1575 } else {
James Smartbc739052010-08-04 16:11:18 -04001576 if ((!(psli->sli_flag & LPFC_SLI_MBOX_ACTIVE)) &&
1577 (list_empty(&psli->mboxq))) {
James Smart13815c82008-01-11 01:52:48 -05001578
James Smarta22d73b2021-01-04 10:02:38 -08001579 retval = lpfc_issue_hb_mbox(phba);
1580 if (retval) {
1581 tmo = (1000 * LPFC_HB_MBOX_INTERVAL);
1582 goto out;
James Smartbc739052010-08-04 16:11:18 -04001583 }
1584 phba->skipped_hb = 0;
James Smartbc739052010-08-04 16:11:18 -04001585 } else if (time_before_eq(phba->last_completion_time,
1586 phba->skipped_hb)) {
1587 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
1588 "2857 Last completion time not "
1589 " updated in %d ms\n",
1590 jiffies_to_msecs(jiffies
1591 - phba->last_completion_time));
1592 } else
1593 phba->skipped_hb = jiffies;
1594
James Smarta22d73b2021-01-04 10:02:38 -08001595 tmo = (1000 * LPFC_HB_MBOX_TIMEOUT);
1596 goto out;
James Smart858c9f62007-06-17 19:56:39 -05001597 }
James Smart4258e982015-12-16 18:11:58 -05001598 } else {
James Smarta22d73b2021-01-04 10:02:38 -08001599 /* Check to see if we want to force a MBX_HEARTBEAT */
1600 if (phba->hba_flag & HBA_HBEAT_TMO) {
1601 retval = lpfc_issue_hb_mbox(phba);
1602 if (retval)
1603 tmo = (1000 * LPFC_HB_MBOX_INTERVAL);
1604 else
1605 tmo = (1000 * LPFC_HB_MBOX_TIMEOUT);
1606 goto out;
1607 }
1608 tmo = (1000 * LPFC_HB_MBOX_INTERVAL);
James Smart858c9f62007-06-17 19:56:39 -05001609 }
James Smarta22d73b2021-01-04 10:02:38 -08001610out:
1611 mod_timer(&phba->hb_tmofunc, jiffies + msecs_to_jiffies(tmo));
James Smart858c9f62007-06-17 19:56:39 -05001612}
1613
James Smarte59058c2008-08-24 21:49:00 -04001614/**
James Smart3621a712009-04-06 18:47:14 -04001615 * lpfc_offline_eratt - Bring lpfc offline on hardware error attention
James Smarte59058c2008-08-24 21:49:00 -04001616 * @phba: pointer to lpfc hba data structure.
1617 *
1618 * This routine is called to bring the HBA offline when HBA hardware error
1619 * other than Port Error 6 has been detected.
1620 **/
James Smart09372822008-01-11 01:52:54 -05001621static void
1622lpfc_offline_eratt(struct lpfc_hba *phba)
1623{
1624 struct lpfc_sli *psli = &phba->sli;
1625
1626 spin_lock_irq(&phba->hbalock);
James Smartf4b4c682009-05-22 14:53:12 -04001627 psli->sli_flag &= ~LPFC_SLI_ACTIVE;
James Smart09372822008-01-11 01:52:54 -05001628 spin_unlock_irq(&phba->hbalock);
James Smart618a5232012-06-12 13:54:36 -04001629 lpfc_offline_prep(phba, LPFC_MBX_NO_WAIT);
James Smart09372822008-01-11 01:52:54 -05001630
1631 lpfc_offline(phba);
1632 lpfc_reset_barrier(phba);
James Smartf4b4c682009-05-22 14:53:12 -04001633 spin_lock_irq(&phba->hbalock);
James Smart09372822008-01-11 01:52:54 -05001634 lpfc_sli_brdreset(phba);
James Smartf4b4c682009-05-22 14:53:12 -04001635 spin_unlock_irq(&phba->hbalock);
James Smart09372822008-01-11 01:52:54 -05001636 lpfc_hba_down_post(phba);
1637 lpfc_sli_brdready(phba, HS_MBRDY);
1638 lpfc_unblock_mgmt_io(phba);
1639 phba->link_state = LPFC_HBA_ERROR;
1640 return;
1641}
1642
James Smarte59058c2008-08-24 21:49:00 -04001643/**
James Smartda0436e2009-05-22 14:51:39 -04001644 * lpfc_sli4_offline_eratt - Bring lpfc offline on SLI4 hardware error attention
1645 * @phba: pointer to lpfc hba data structure.
1646 *
1647 * This routine is called to bring a SLI4 HBA offline when HBA hardware error
1648 * other than Port Error 6 has been detected.
1649 **/
James Smarta88dbb62013-04-17 20:18:39 -04001650void
James Smartda0436e2009-05-22 14:51:39 -04001651lpfc_sli4_offline_eratt(struct lpfc_hba *phba)
1652{
James Smart946727d2015-04-07 15:07:09 -04001653 spin_lock_irq(&phba->hbalock);
James Smart25ac2c972021-09-10 16:31:54 -07001654 if (phba->link_state == LPFC_HBA_ERROR &&
1655 phba->hba_flag & HBA_PCI_ERR) {
1656 spin_unlock_irq(&phba->hbalock);
1657 return;
1658 }
James Smart946727d2015-04-07 15:07:09 -04001659 phba->link_state = LPFC_HBA_ERROR;
1660 spin_unlock_irq(&phba->hbalock);
1661
James Smart618a5232012-06-12 13:54:36 -04001662 lpfc_offline_prep(phba, LPFC_MBX_NO_WAIT);
James Smartc00f62e2019-08-14 16:57:11 -07001663 lpfc_sli_flush_io_rings(phba);
James Smartda0436e2009-05-22 14:51:39 -04001664 lpfc_offline(phba);
James Smartda0436e2009-05-22 14:51:39 -04001665 lpfc_hba_down_post(phba);
James Smartda0436e2009-05-22 14:51:39 -04001666 lpfc_unblock_mgmt_io(phba);
James Smartda0436e2009-05-22 14:51:39 -04001667}
1668
1669/**
James Smarta257bf92009-04-06 18:48:10 -04001670 * lpfc_handle_deferred_eratt - The HBA hardware deferred error handler
1671 * @phba: pointer to lpfc hba data structure.
1672 *
1673 * This routine is invoked to handle the deferred HBA hardware error
1674 * conditions. This type of error is indicated by HBA by setting ER1
1675 * and another ER bit in the host status register. The driver will
1676 * wait until the ER1 bit clears before handling the error condition.
1677 **/
1678static void
1679lpfc_handle_deferred_eratt(struct lpfc_hba *phba)
1680{
1681 uint32_t old_host_status = phba->work_hs;
James Smarta257bf92009-04-06 18:48:10 -04001682 struct lpfc_sli *psli = &phba->sli;
1683
James Smartf4b4c682009-05-22 14:53:12 -04001684 /* If the pci channel is offline, ignore possible errors,
1685 * since we cannot communicate with the pci card anyway.
1686 */
1687 if (pci_channel_offline(phba->pcidev)) {
1688 spin_lock_irq(&phba->hbalock);
1689 phba->hba_flag &= ~DEFER_ERATT;
1690 spin_unlock_irq(&phba->hbalock);
1691 return;
1692 }
1693
Dick Kennedy372c1872020-06-30 14:50:00 -07001694 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
1695 "0479 Deferred Adapter Hardware Error "
1696 "Data: x%x x%x x%x\n",
1697 phba->work_hs, phba->work_status[0],
1698 phba->work_status[1]);
James Smarta257bf92009-04-06 18:48:10 -04001699
1700 spin_lock_irq(&phba->hbalock);
James Smartf4b4c682009-05-22 14:53:12 -04001701 psli->sli_flag &= ~LPFC_SLI_ACTIVE;
James Smarta257bf92009-04-06 18:48:10 -04001702 spin_unlock_irq(&phba->hbalock);
1703
1704
1705 /*
1706 * Firmware stops when it triggred erratt. That could cause the I/Os
1707 * dropped by the firmware. Error iocb (I/O) on txcmplq and let the
1708 * SCSI layer retry it after re-establishing link.
1709 */
James Smartdb55fba2014-04-04 13:52:02 -04001710 lpfc_sli_abort_fcp_rings(phba);
James Smarta257bf92009-04-06 18:48:10 -04001711
1712 /*
1713 * There was a firmware error. Take the hba offline and then
1714 * attempt to restart it.
1715 */
James Smart618a5232012-06-12 13:54:36 -04001716 lpfc_offline_prep(phba, LPFC_MBX_WAIT);
James Smarta257bf92009-04-06 18:48:10 -04001717 lpfc_offline(phba);
1718
1719 /* Wait for the ER1 bit to clear.*/
1720 while (phba->work_hs & HS_FFER1) {
1721 msleep(100);
James Smart9940b972011-03-11 16:06:12 -05001722 if (lpfc_readl(phba->HSregaddr, &phba->work_hs)) {
1723 phba->work_hs = UNPLUG_ERR ;
1724 break;
1725 }
James Smarta257bf92009-04-06 18:48:10 -04001726 /* If driver is unloading let the worker thread continue */
1727 if (phba->pport->load_flag & FC_UNLOADING) {
1728 phba->work_hs = 0;
1729 break;
1730 }
1731 }
1732
1733 /*
1734 * This is to ptrotect against a race condition in which
1735 * first write to the host attention register clear the
1736 * host status register.
1737 */
1738 if ((!phba->work_hs) && (!(phba->pport->load_flag & FC_UNLOADING)))
1739 phba->work_hs = old_host_status & ~HS_FFER1;
1740
James Smart3772a992009-05-22 14:50:54 -04001741 spin_lock_irq(&phba->hbalock);
James Smarta257bf92009-04-06 18:48:10 -04001742 phba->hba_flag &= ~DEFER_ERATT;
James Smart3772a992009-05-22 14:50:54 -04001743 spin_unlock_irq(&phba->hbalock);
James Smarta257bf92009-04-06 18:48:10 -04001744 phba->work_status[0] = readl(phba->MBslimaddr + 0xa8);
1745 phba->work_status[1] = readl(phba->MBslimaddr + 0xac);
1746}
1747
James Smart3772a992009-05-22 14:50:54 -04001748static void
1749lpfc_board_errevt_to_mgmt(struct lpfc_hba *phba)
1750{
1751 struct lpfc_board_event_header board_event;
1752 struct Scsi_Host *shost;
1753
1754 board_event.event_type = FC_REG_BOARD_EVENT;
1755 board_event.subcategory = LPFC_EVENT_PORTINTERR;
1756 shost = lpfc_shost_from_vport(phba->pport);
1757 fc_host_post_vendor_event(shost, fc_get_event_number(),
1758 sizeof(board_event),
1759 (char *) &board_event,
1760 LPFC_NL_VENDOR_ID);
1761}
1762
James Smarta257bf92009-04-06 18:48:10 -04001763/**
James Smart3772a992009-05-22 14:50:54 -04001764 * lpfc_handle_eratt_s3 - The SLI3 HBA hardware error handler
James Smarte59058c2008-08-24 21:49:00 -04001765 * @phba: pointer to lpfc hba data structure.
1766 *
1767 * This routine is invoked to handle the following HBA hardware error
1768 * conditions:
1769 * 1 - HBA error attention interrupt
1770 * 2 - DMA ring index out of range
1771 * 3 - Mailbox command came back as unknown
1772 **/
James Smart3772a992009-05-22 14:50:54 -04001773static void
1774lpfc_handle_eratt_s3(struct lpfc_hba *phba)
dea31012005-04-17 16:05:31 -05001775{
James Smart2e0fef82007-06-17 19:56:36 -05001776 struct lpfc_vport *vport = phba->pport;
James Smart2e0fef82007-06-17 19:56:36 -05001777 struct lpfc_sli *psli = &phba->sli;
James Smartd2873e42006-08-18 17:46:43 -04001778 uint32_t event_data;
James Smart57127f12007-10-27 13:37:05 -04001779 unsigned long temperature;
1780 struct temp_event temp_event_data;
James Smart92d7f7b2007-06-17 19:56:38 -05001781 struct Scsi_Host *shost;
James Smart2e0fef82007-06-17 19:56:36 -05001782
Linas Vepstas8d63f372007-02-14 14:28:36 -06001783 /* If the pci channel is offline, ignore possible errors,
James Smart3772a992009-05-22 14:50:54 -04001784 * since we cannot communicate with the pci card anyway.
1785 */
1786 if (pci_channel_offline(phba->pcidev)) {
1787 spin_lock_irq(&phba->hbalock);
1788 phba->hba_flag &= ~DEFER_ERATT;
1789 spin_unlock_irq(&phba->hbalock);
Linas Vepstas8d63f372007-02-14 14:28:36 -06001790 return;
James Smart3772a992009-05-22 14:50:54 -04001791 }
1792
James Smart13815c82008-01-11 01:52:48 -05001793 /* If resets are disabled then leave the HBA alone and return */
1794 if (!phba->cfg_enable_hba_reset)
1795 return;
dea31012005-04-17 16:05:31 -05001796
James Smartea2151b2008-09-07 11:52:10 -04001797 /* Send an internal error event to mgmt application */
James Smart3772a992009-05-22 14:50:54 -04001798 lpfc_board_errevt_to_mgmt(phba);
James Smartea2151b2008-09-07 11:52:10 -04001799
James Smarta257bf92009-04-06 18:48:10 -04001800 if (phba->hba_flag & DEFER_ERATT)
1801 lpfc_handle_deferred_eratt(phba);
1802
James Smartdcf2a4e2010-09-29 11:18:53 -04001803 if ((phba->work_hs & HS_FFER6) || (phba->work_hs & HS_FFER8)) {
1804 if (phba->work_hs & HS_FFER6)
1805 /* Re-establishing Link */
1806 lpfc_printf_log(phba, KERN_INFO, LOG_LINK_EVENT,
1807 "1301 Re-establishing Link "
1808 "Data: x%x x%x x%x\n",
1809 phba->work_hs, phba->work_status[0],
1810 phba->work_status[1]);
1811 if (phba->work_hs & HS_FFER8)
1812 /* Device Zeroization */
1813 lpfc_printf_log(phba, KERN_INFO, LOG_LINK_EVENT,
1814 "2861 Host Authentication device "
1815 "zeroization Data:x%x x%x x%x\n",
1816 phba->work_hs, phba->work_status[0],
1817 phba->work_status[1]);
James Smart58da1ff2008-04-07 10:15:56 -04001818
James Smart92d7f7b2007-06-17 19:56:38 -05001819 spin_lock_irq(&phba->hbalock);
James Smartf4b4c682009-05-22 14:53:12 -04001820 psli->sli_flag &= ~LPFC_SLI_ACTIVE;
James Smart92d7f7b2007-06-17 19:56:38 -05001821 spin_unlock_irq(&phba->hbalock);
dea31012005-04-17 16:05:31 -05001822
1823 /*
1824 * Firmware stops when it triggled erratt with HS_FFER6.
1825 * That could cause the I/Os dropped by the firmware.
1826 * Error iocb (I/O) on txcmplq and let the SCSI layer
1827 * retry it after re-establishing link.
1828 */
James Smartdb55fba2014-04-04 13:52:02 -04001829 lpfc_sli_abort_fcp_rings(phba);
dea31012005-04-17 16:05:31 -05001830
dea31012005-04-17 16:05:31 -05001831 /*
1832 * There was a firmware error. Take the hba offline and then
1833 * attempt to restart it.
1834 */
James Smart618a5232012-06-12 13:54:36 -04001835 lpfc_offline_prep(phba, LPFC_MBX_NO_WAIT);
dea31012005-04-17 16:05:31 -05001836 lpfc_offline(phba);
Jamie Wellnitz41415862006-02-28 19:25:27 -05001837 lpfc_sli_brdrestart(phba);
dea31012005-04-17 16:05:31 -05001838 if (lpfc_online(phba) == 0) { /* Initialize the HBA */
James Smart46fa3112007-04-25 09:51:45 -04001839 lpfc_unblock_mgmt_io(phba);
dea31012005-04-17 16:05:31 -05001840 return;
1841 }
James Smart46fa3112007-04-25 09:51:45 -04001842 lpfc_unblock_mgmt_io(phba);
James Smart57127f12007-10-27 13:37:05 -04001843 } else if (phba->work_hs & HS_CRIT_TEMP) {
1844 temperature = readl(phba->MBslimaddr + TEMPERATURE_OFFSET);
1845 temp_event_data.event_type = FC_REG_TEMPERATURE_EVENT;
1846 temp_event_data.event_code = LPFC_CRIT_TEMP;
1847 temp_event_data.data = (uint32_t)temperature;
1848
Dick Kennedy372c1872020-06-30 14:50:00 -07001849 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartd7c255b2008-08-24 21:50:00 -04001850 "0406 Adapter maximum temperature exceeded "
James Smart57127f12007-10-27 13:37:05 -04001851 "(%ld), taking this port offline "
1852 "Data: x%x x%x x%x\n",
1853 temperature, phba->work_hs,
1854 phba->work_status[0], phba->work_status[1]);
1855
1856 shost = lpfc_shost_from_vport(phba->pport);
1857 fc_host_post_vendor_event(shost, fc_get_event_number(),
1858 sizeof(temp_event_data),
1859 (char *) &temp_event_data,
1860 SCSI_NL_VID_TYPE_PCI
1861 | PCI_VENDOR_ID_EMULEX);
1862
James Smart7af67052007-10-27 13:38:11 -04001863 spin_lock_irq(&phba->hbalock);
James Smart7af67052007-10-27 13:38:11 -04001864 phba->over_temp_state = HBA_OVER_TEMP;
1865 spin_unlock_irq(&phba->hbalock);
James Smart09372822008-01-11 01:52:54 -05001866 lpfc_offline_eratt(phba);
James Smart57127f12007-10-27 13:37:05 -04001867
dea31012005-04-17 16:05:31 -05001868 } else {
1869 /* The if clause above forces this code path when the status
James Smart93996272008-08-24 21:50:30 -04001870 * failure is a value other than FFER6. Do not call the offline
1871 * twice. This is the adapter hardware error path.
dea31012005-04-17 16:05:31 -05001872 */
Dick Kennedy372c1872020-06-30 14:50:00 -07001873 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smarte8b62012007-08-02 11:10:09 -04001874 "0457 Adapter Hardware Error "
dea31012005-04-17 16:05:31 -05001875 "Data: x%x x%x x%x\n",
James Smarte8b62012007-08-02 11:10:09 -04001876 phba->work_hs,
dea31012005-04-17 16:05:31 -05001877 phba->work_status[0], phba->work_status[1]);
1878
James Smartd2873e42006-08-18 17:46:43 -04001879 event_data = FC_REG_DUMP_EVENT;
James Smart92d7f7b2007-06-17 19:56:38 -05001880 shost = lpfc_shost_from_vport(vport);
James Smart2e0fef82007-06-17 19:56:36 -05001881 fc_host_post_vendor_event(shost, fc_get_event_number(),
James Smartd2873e42006-08-18 17:46:43 -04001882 sizeof(event_data), (char *) &event_data,
1883 SCSI_NL_VID_TYPE_PCI | PCI_VENDOR_ID_EMULEX);
1884
James Smart09372822008-01-11 01:52:54 -05001885 lpfc_offline_eratt(phba);
dea31012005-04-17 16:05:31 -05001886 }
James Smart93996272008-08-24 21:50:30 -04001887 return;
dea31012005-04-17 16:05:31 -05001888}
1889
James Smarte59058c2008-08-24 21:49:00 -04001890/**
James Smart618a5232012-06-12 13:54:36 -04001891 * lpfc_sli4_port_sta_fn_reset - The SLI4 function reset due to port status reg
1892 * @phba: pointer to lpfc hba data structure.
1893 * @mbx_action: flag for mailbox shutdown action.
Lee Jonesfe614ac2020-07-23 13:24:22 +01001894 * @en_rn_msg: send reset/port recovery message.
James Smart618a5232012-06-12 13:54:36 -04001895 * This routine is invoked to perform an SLI4 port PCI function reset in
1896 * response to port status register polling attention. It waits for port
1897 * status register (ERR, RDY, RN) bits before proceeding with function reset.
1898 * During this process, interrupt vectors are freed and later requested
1899 * for handling possible port resource change.
1900 **/
1901static int
James Smarte10b2022014-02-20 09:57:43 -05001902lpfc_sli4_port_sta_fn_reset(struct lpfc_hba *phba, int mbx_action,
1903 bool en_rn_msg)
James Smart618a5232012-06-12 13:54:36 -04001904{
1905 int rc;
1906 uint32_t intr_mode;
James Smarta9978e32021-07-07 11:43:43 -07001907 LPFC_MBOXQ_t *mboxq;
James Smart618a5232012-06-12 13:54:36 -04001908
James Smart27d6ac02018-02-22 08:18:42 -08001909 if (bf_get(lpfc_sli_intf_if_type, &phba->sli4_hba.sli_intf) >=
James Smart65791f12016-07-06 12:35:56 -07001910 LPFC_SLI_INTF_IF_TYPE_2) {
1911 /*
1912 * On error status condition, driver need to wait for port
1913 * ready before performing reset.
1914 */
1915 rc = lpfc_sli4_pdev_status_reg_wait(phba);
James Smart0e916ee2016-07-06 12:36:06 -07001916 if (rc)
James Smart65791f12016-07-06 12:35:56 -07001917 return rc;
James Smart618a5232012-06-12 13:54:36 -04001918 }
James Smart0e916ee2016-07-06 12:36:06 -07001919
James Smart65791f12016-07-06 12:35:56 -07001920 /* need reset: attempt for port recovery */
1921 if (en_rn_msg)
James Smart0b3ad322021-01-04 10:02:39 -08001922 lpfc_printf_log(phba, KERN_ERR, LOG_SLI,
James Smart65791f12016-07-06 12:35:56 -07001923 "2887 Reset Needed: Attempting Port "
1924 "Recovery...\n");
James Smart3ba62162021-01-04 10:02:31 -08001925
1926 /* If we are no wait, the HBA has been reset and is not
James Smarta9978e32021-07-07 11:43:43 -07001927 * functional, thus we should clear
1928 * (LPFC_SLI_ACTIVE | LPFC_SLI_MBOX_ACTIVE) flags.
James Smart3ba62162021-01-04 10:02:31 -08001929 */
1930 if (mbx_action == LPFC_MBX_NO_WAIT) {
1931 spin_lock_irq(&phba->hbalock);
1932 phba->sli.sli_flag &= ~LPFC_SLI_ACTIVE;
James Smarta9978e32021-07-07 11:43:43 -07001933 if (phba->sli.mbox_active) {
1934 mboxq = phba->sli.mbox_active;
1935 mboxq->u.mb.mbxStatus = MBX_NOT_FINISHED;
1936 __lpfc_mbox_cmpl_put(phba, mboxq);
1937 phba->sli.sli_flag &= ~LPFC_SLI_MBOX_ACTIVE;
1938 phba->sli.mbox_active = NULL;
1939 }
James Smart3ba62162021-01-04 10:02:31 -08001940 spin_unlock_irq(&phba->hbalock);
1941 }
1942
James Smart65791f12016-07-06 12:35:56 -07001943 lpfc_offline_prep(phba, mbx_action);
James Smartc00f62e2019-08-14 16:57:11 -07001944 lpfc_sli_flush_io_rings(phba);
James Smart65791f12016-07-06 12:35:56 -07001945 lpfc_offline(phba);
1946 /* release interrupt for possible resource change */
1947 lpfc_sli4_disable_intr(phba);
James Smart5a9eeff2018-11-29 16:09:32 -08001948 rc = lpfc_sli_brdrestart(phba);
1949 if (rc) {
Dick Kennedy372c1872020-06-30 14:50:00 -07001950 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart5a9eeff2018-11-29 16:09:32 -08001951 "6309 Failed to restart board\n");
1952 return rc;
1953 }
James Smart65791f12016-07-06 12:35:56 -07001954 /* request and enable interrupt */
1955 intr_mode = lpfc_sli4_enable_intr(phba, phba->intr_mode);
1956 if (intr_mode == LPFC_INTR_ERROR) {
Dick Kennedy372c1872020-06-30 14:50:00 -07001957 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart65791f12016-07-06 12:35:56 -07001958 "3175 Failed to enable interrupt\n");
1959 return -EIO;
1960 }
1961 phba->intr_mode = intr_mode;
1962 rc = lpfc_online(phba);
1963 if (rc == 0)
1964 lpfc_unblock_mgmt_io(phba);
1965
James Smart618a5232012-06-12 13:54:36 -04001966 return rc;
1967}
1968
1969/**
James Smartda0436e2009-05-22 14:51:39 -04001970 * lpfc_handle_eratt_s4 - The SLI4 HBA hardware error handler
1971 * @phba: pointer to lpfc hba data structure.
1972 *
1973 * This routine is invoked to handle the SLI4 HBA hardware error attention
1974 * conditions.
1975 **/
1976static void
1977lpfc_handle_eratt_s4(struct lpfc_hba *phba)
1978{
1979 struct lpfc_vport *vport = phba->pport;
1980 uint32_t event_data;
1981 struct Scsi_Host *shost;
James Smart2fcee4b2010-12-15 17:57:46 -05001982 uint32_t if_type;
James Smart2e90f4b2011-12-13 13:22:37 -05001983 struct lpfc_register portstat_reg = {0};
1984 uint32_t reg_err1, reg_err2;
1985 uint32_t uerrlo_reg, uemasklo_reg;
James Smart65791f12016-07-06 12:35:56 -07001986 uint32_t smphr_port_status = 0, pci_rd_rc1, pci_rd_rc2;
James Smarte10b2022014-02-20 09:57:43 -05001987 bool en_rn_msg = true;
James Smart946727d2015-04-07 15:07:09 -04001988 struct temp_event temp_event_data;
James Smart65791f12016-07-06 12:35:56 -07001989 struct lpfc_register portsmphr_reg;
1990 int rc, i;
James Smartda0436e2009-05-22 14:51:39 -04001991
1992 /* If the pci channel is offline, ignore possible errors, since
1993 * we cannot communicate with the pci card anyway.
1994 */
James Smart32a93102019-03-12 16:30:13 -07001995 if (pci_channel_offline(phba->pcidev)) {
Dick Kennedy372c1872020-06-30 14:50:00 -07001996 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart32a93102019-03-12 16:30:13 -07001997 "3166 pci channel is offline\n");
James Smartda0436e2009-05-22 14:51:39 -04001998 return;
James Smart32a93102019-03-12 16:30:13 -07001999 }
James Smartda0436e2009-05-22 14:51:39 -04002000
James Smart65791f12016-07-06 12:35:56 -07002001 memset(&portsmphr_reg, 0, sizeof(portsmphr_reg));
James Smart2fcee4b2010-12-15 17:57:46 -05002002 if_type = bf_get(lpfc_sli_intf_if_type, &phba->sli4_hba.sli_intf);
2003 switch (if_type) {
2004 case LPFC_SLI_INTF_IF_TYPE_0:
James Smart2e90f4b2011-12-13 13:22:37 -05002005 pci_rd_rc1 = lpfc_readl(
2006 phba->sli4_hba.u.if_type0.UERRLOregaddr,
2007 &uerrlo_reg);
2008 pci_rd_rc2 = lpfc_readl(
2009 phba->sli4_hba.u.if_type0.UEMASKLOregaddr,
2010 &uemasklo_reg);
2011 /* consider PCI bus read error as pci_channel_offline */
2012 if (pci_rd_rc1 == -EIO && pci_rd_rc2 == -EIO)
2013 return;
James Smart65791f12016-07-06 12:35:56 -07002014 if (!(phba->hba_flag & HBA_RECOVERABLE_UE)) {
2015 lpfc_sli4_offline_eratt(phba);
2016 return;
2017 }
Dick Kennedy372c1872020-06-30 14:50:00 -07002018 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart65791f12016-07-06 12:35:56 -07002019 "7623 Checking UE recoverable");
2020
2021 for (i = 0; i < phba->sli4_hba.ue_to_sr / 1000; i++) {
2022 if (lpfc_readl(phba->sli4_hba.PSMPHRregaddr,
2023 &portsmphr_reg.word0))
2024 continue;
2025
2026 smphr_port_status = bf_get(lpfc_port_smphr_port_status,
2027 &portsmphr_reg);
2028 if ((smphr_port_status & LPFC_PORT_SEM_MASK) ==
2029 LPFC_PORT_SEM_UE_RECOVERABLE)
2030 break;
2031 /*Sleep for 1Sec, before checking SEMAPHORE */
2032 msleep(1000);
2033 }
2034
Dick Kennedy372c1872020-06-30 14:50:00 -07002035 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart65791f12016-07-06 12:35:56 -07002036 "4827 smphr_port_status x%x : Waited %dSec",
2037 smphr_port_status, i);
2038
2039 /* Recoverable UE, reset the HBA device */
2040 if ((smphr_port_status & LPFC_PORT_SEM_MASK) ==
2041 LPFC_PORT_SEM_UE_RECOVERABLE) {
2042 for (i = 0; i < 20; i++) {
2043 msleep(1000);
2044 if (!lpfc_readl(phba->sli4_hba.PSMPHRregaddr,
2045 &portsmphr_reg.word0) &&
2046 (LPFC_POST_STAGE_PORT_READY ==
2047 bf_get(lpfc_port_smphr_port_status,
2048 &portsmphr_reg))) {
2049 rc = lpfc_sli4_port_sta_fn_reset(phba,
2050 LPFC_MBX_NO_WAIT, en_rn_msg);
2051 if (rc == 0)
2052 return;
Dick Kennedy372c1872020-06-30 14:50:00 -07002053 lpfc_printf_log(phba, KERN_ERR,
2054 LOG_TRACE_EVENT,
James Smart65791f12016-07-06 12:35:56 -07002055 "4215 Failed to recover UE");
2056 break;
2057 }
2058 }
2059 }
Dick Kennedy372c1872020-06-30 14:50:00 -07002060 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart65791f12016-07-06 12:35:56 -07002061 "7624 Firmware not ready: Failing UE recovery,"
2062 " waited %dSec", i);
James Smart8c24a4f2019-08-14 16:56:53 -07002063 phba->link_state = LPFC_HBA_ERROR;
James Smart2fcee4b2010-12-15 17:57:46 -05002064 break;
James Smart946727d2015-04-07 15:07:09 -04002065
James Smart2fcee4b2010-12-15 17:57:46 -05002066 case LPFC_SLI_INTF_IF_TYPE_2:
James Smart27d6ac02018-02-22 08:18:42 -08002067 case LPFC_SLI_INTF_IF_TYPE_6:
James Smart2e90f4b2011-12-13 13:22:37 -05002068 pci_rd_rc1 = lpfc_readl(
2069 phba->sli4_hba.u.if_type2.STATUSregaddr,
2070 &portstat_reg.word0);
2071 /* consider PCI bus read error as pci_channel_offline */
James Smart6b5151f2012-01-18 16:24:06 -05002072 if (pci_rd_rc1 == -EIO) {
Dick Kennedy372c1872020-06-30 14:50:00 -07002073 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart6b5151f2012-01-18 16:24:06 -05002074 "3151 PCI bus read access failure: x%x\n",
2075 readl(phba->sli4_hba.u.if_type2.STATUSregaddr));
James Smart32a93102019-03-12 16:30:13 -07002076 lpfc_sli4_offline_eratt(phba);
James Smart2e90f4b2011-12-13 13:22:37 -05002077 return;
James Smart6b5151f2012-01-18 16:24:06 -05002078 }
James Smart2e90f4b2011-12-13 13:22:37 -05002079 reg_err1 = readl(phba->sli4_hba.u.if_type2.ERR1regaddr);
2080 reg_err2 = readl(phba->sli4_hba.u.if_type2.ERR2regaddr);
James Smart2fcee4b2010-12-15 17:57:46 -05002081 if (bf_get(lpfc_sliport_status_oti, &portstat_reg)) {
Dick Kennedy372c1872020-06-30 14:50:00 -07002082 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
2083 "2889 Port Overtemperature event, "
2084 "taking port offline Data: x%x x%x\n",
2085 reg_err1, reg_err2);
James Smart946727d2015-04-07 15:07:09 -04002086
James Smart310429e2016-07-06 12:35:54 -07002087 phba->sfp_alarm |= LPFC_TRANSGRESSION_HIGH_TEMPERATURE;
James Smart946727d2015-04-07 15:07:09 -04002088 temp_event_data.event_type = FC_REG_TEMPERATURE_EVENT;
2089 temp_event_data.event_code = LPFC_CRIT_TEMP;
2090 temp_event_data.data = 0xFFFFFFFF;
2091
2092 shost = lpfc_shost_from_vport(phba->pport);
2093 fc_host_post_vendor_event(shost, fc_get_event_number(),
2094 sizeof(temp_event_data),
2095 (char *)&temp_event_data,
2096 SCSI_NL_VID_TYPE_PCI
2097 | PCI_VENDOR_ID_EMULEX);
2098
James Smart2fcee4b2010-12-15 17:57:46 -05002099 spin_lock_irq(&phba->hbalock);
2100 phba->over_temp_state = HBA_OVER_TEMP;
2101 spin_unlock_irq(&phba->hbalock);
2102 lpfc_sli4_offline_eratt(phba);
James Smart946727d2015-04-07 15:07:09 -04002103 return;
James Smart2fcee4b2010-12-15 17:57:46 -05002104 }
James Smart2e90f4b2011-12-13 13:22:37 -05002105 if (reg_err1 == SLIPORT_ERR1_REG_ERR_CODE_2 &&
James Smarte10b2022014-02-20 09:57:43 -05002106 reg_err2 == SLIPORT_ERR2_REG_FW_RESTART) {
Dick Kennedy372c1872020-06-30 14:50:00 -07002107 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smarte10b2022014-02-20 09:57:43 -05002108 "3143 Port Down: Firmware Update "
2109 "Detected\n");
2110 en_rn_msg = false;
2111 } else if (reg_err1 == SLIPORT_ERR1_REG_ERR_CODE_2 &&
James Smart2e90f4b2011-12-13 13:22:37 -05002112 reg_err2 == SLIPORT_ERR2_REG_FORCED_DUMP)
Dick Kennedy372c1872020-06-30 14:50:00 -07002113 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart2e90f4b2011-12-13 13:22:37 -05002114 "3144 Port Down: Debug Dump\n");
2115 else if (reg_err1 == SLIPORT_ERR1_REG_ERR_CODE_2 &&
2116 reg_err2 == SLIPORT_ERR2_REG_FUNC_PROVISON)
Dick Kennedy372c1872020-06-30 14:50:00 -07002117 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart2e90f4b2011-12-13 13:22:37 -05002118 "3145 Port Down: Provisioning\n");
James Smart618a5232012-06-12 13:54:36 -04002119
James Smart946727d2015-04-07 15:07:09 -04002120 /* If resets are disabled then leave the HBA alone and return */
2121 if (!phba->cfg_enable_hba_reset)
2122 return;
2123
James Smart618a5232012-06-12 13:54:36 -04002124 /* Check port status register for function reset */
James Smarte10b2022014-02-20 09:57:43 -05002125 rc = lpfc_sli4_port_sta_fn_reset(phba, LPFC_MBX_NO_WAIT,
2126 en_rn_msg);
James Smart618a5232012-06-12 13:54:36 -04002127 if (rc == 0) {
2128 /* don't report event on forced debug dump */
2129 if (reg_err1 == SLIPORT_ERR1_REG_ERR_CODE_2 &&
2130 reg_err2 == SLIPORT_ERR2_REG_FORCED_DUMP)
2131 return;
2132 else
2133 break;
James Smart2fcee4b2010-12-15 17:57:46 -05002134 }
James Smart618a5232012-06-12 13:54:36 -04002135 /* fall through for not able to recover */
Dick Kennedy372c1872020-06-30 14:50:00 -07002136 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart8c24a4f2019-08-14 16:56:53 -07002137 "3152 Unrecoverable error\n");
2138 phba->link_state = LPFC_HBA_ERROR;
James Smart2fcee4b2010-12-15 17:57:46 -05002139 break;
2140 case LPFC_SLI_INTF_IF_TYPE_1:
2141 default:
2142 break;
2143 }
James Smart2e90f4b2011-12-13 13:22:37 -05002144 lpfc_printf_log(phba, KERN_WARNING, LOG_INIT,
2145 "3123 Report dump event to upper layer\n");
2146 /* Send an internal error event to mgmt application */
2147 lpfc_board_errevt_to_mgmt(phba);
2148
2149 event_data = FC_REG_DUMP_EVENT;
2150 shost = lpfc_shost_from_vport(vport);
2151 fc_host_post_vendor_event(shost, fc_get_event_number(),
2152 sizeof(event_data), (char *) &event_data,
2153 SCSI_NL_VID_TYPE_PCI | PCI_VENDOR_ID_EMULEX);
James Smartda0436e2009-05-22 14:51:39 -04002154}
2155
2156/**
2157 * lpfc_handle_eratt - Wrapper func for handling hba error attention
2158 * @phba: pointer to lpfc HBA data structure.
2159 *
2160 * This routine wraps the actual SLI3 or SLI4 hba error attention handling
2161 * routine from the API jump table function pointer from the lpfc_hba struct.
2162 *
2163 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -02002164 * 0 - success.
James Smartda0436e2009-05-22 14:51:39 -04002165 * Any other value - error.
2166 **/
2167void
2168lpfc_handle_eratt(struct lpfc_hba *phba)
2169{
2170 (*phba->lpfc_handle_eratt)(phba);
2171}
2172
2173/**
James Smart3621a712009-04-06 18:47:14 -04002174 * lpfc_handle_latt - The HBA link event handler
James Smarte59058c2008-08-24 21:49:00 -04002175 * @phba: pointer to lpfc hba data structure.
2176 *
2177 * This routine is invoked from the worker thread to handle a HBA host
James Smart895427b2017-02-12 13:52:30 -08002178 * attention link event. SLI3 only.
James Smarte59058c2008-08-24 21:49:00 -04002179 **/
dea31012005-04-17 16:05:31 -05002180void
James Smart2e0fef82007-06-17 19:56:36 -05002181lpfc_handle_latt(struct lpfc_hba *phba)
dea31012005-04-17 16:05:31 -05002182{
James Smart2e0fef82007-06-17 19:56:36 -05002183 struct lpfc_vport *vport = phba->pport;
2184 struct lpfc_sli *psli = &phba->sli;
dea31012005-04-17 16:05:31 -05002185 LPFC_MBOXQ_t *pmb;
2186 volatile uint32_t control;
2187 struct lpfc_dmabuf *mp;
James Smart09372822008-01-11 01:52:54 -05002188 int rc = 0;
dea31012005-04-17 16:05:31 -05002189
2190 pmb = (LPFC_MBOXQ_t *)mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL);
James Smart09372822008-01-11 01:52:54 -05002191 if (!pmb) {
2192 rc = 1;
dea31012005-04-17 16:05:31 -05002193 goto lpfc_handle_latt_err_exit;
James Smart09372822008-01-11 01:52:54 -05002194 }
dea31012005-04-17 16:05:31 -05002195
2196 mp = kmalloc(sizeof(struct lpfc_dmabuf), GFP_KERNEL);
James Smart09372822008-01-11 01:52:54 -05002197 if (!mp) {
2198 rc = 2;
dea31012005-04-17 16:05:31 -05002199 goto lpfc_handle_latt_free_pmb;
James Smart09372822008-01-11 01:52:54 -05002200 }
dea31012005-04-17 16:05:31 -05002201
2202 mp->virt = lpfc_mbuf_alloc(phba, 0, &mp->phys);
James Smart09372822008-01-11 01:52:54 -05002203 if (!mp->virt) {
2204 rc = 3;
dea31012005-04-17 16:05:31 -05002205 goto lpfc_handle_latt_free_mp;
James Smart09372822008-01-11 01:52:54 -05002206 }
dea31012005-04-17 16:05:31 -05002207
James.Smart@Emulex.Com6281bfe2005-11-28 11:41:33 -05002208 /* Cleanup any outstanding ELS commands */
James Smart549e55c2007-08-02 11:09:51 -04002209 lpfc_els_flush_all_cmd(phba);
dea31012005-04-17 16:05:31 -05002210
2211 psli->slistat.link_event++;
James Smart76a95d72010-11-20 23:11:48 -05002212 lpfc_read_topology(phba, pmb, mp);
2213 pmb->mbox_cmpl = lpfc_mbx_cmpl_read_topology;
James Smart2e0fef82007-06-17 19:56:36 -05002214 pmb->vport = vport;
James Smart0d2b6b82008-06-14 22:52:47 -04002215 /* Block ELS IOCBs until we have processed this mbox command */
James Smart895427b2017-02-12 13:52:30 -08002216 phba->sli.sli3_ring[LPFC_ELS_RING].flag |= LPFC_STOP_IOCB_EVENT;
James Smart0b727fe2007-10-27 13:37:25 -04002217 rc = lpfc_sli_issue_mbox (phba, pmb, MBX_NOWAIT);
James Smart09372822008-01-11 01:52:54 -05002218 if (rc == MBX_NOT_FINISHED) {
2219 rc = 4;
James Smart14691152006-12-02 13:34:28 -05002220 goto lpfc_handle_latt_free_mbuf;
James Smart09372822008-01-11 01:52:54 -05002221 }
dea31012005-04-17 16:05:31 -05002222
2223 /* Clear Link Attention in HA REG */
James Smart2e0fef82007-06-17 19:56:36 -05002224 spin_lock_irq(&phba->hbalock);
dea31012005-04-17 16:05:31 -05002225 writel(HA_LATT, phba->HAregaddr);
2226 readl(phba->HAregaddr); /* flush */
James Smart2e0fef82007-06-17 19:56:36 -05002227 spin_unlock_irq(&phba->hbalock);
dea31012005-04-17 16:05:31 -05002228
2229 return;
2230
James Smart14691152006-12-02 13:34:28 -05002231lpfc_handle_latt_free_mbuf:
James Smart895427b2017-02-12 13:52:30 -08002232 phba->sli.sli3_ring[LPFC_ELS_RING].flag &= ~LPFC_STOP_IOCB_EVENT;
James Smart14691152006-12-02 13:34:28 -05002233 lpfc_mbuf_free(phba, mp->virt, mp->phys);
dea31012005-04-17 16:05:31 -05002234lpfc_handle_latt_free_mp:
2235 kfree(mp);
2236lpfc_handle_latt_free_pmb:
James Smart1dcb58e2007-04-25 09:51:30 -04002237 mempool_free(pmb, phba->mbox_mem_pool);
dea31012005-04-17 16:05:31 -05002238lpfc_handle_latt_err_exit:
2239 /* Enable Link attention interrupts */
James Smart2e0fef82007-06-17 19:56:36 -05002240 spin_lock_irq(&phba->hbalock);
dea31012005-04-17 16:05:31 -05002241 psli->sli_flag |= LPFC_PROCESS_LA;
2242 control = readl(phba->HCregaddr);
2243 control |= HC_LAINT_ENA;
2244 writel(control, phba->HCregaddr);
2245 readl(phba->HCregaddr); /* flush */
2246
2247 /* Clear Link Attention in HA REG */
2248 writel(HA_LATT, phba->HAregaddr);
2249 readl(phba->HAregaddr); /* flush */
James Smart2e0fef82007-06-17 19:56:36 -05002250 spin_unlock_irq(&phba->hbalock);
dea31012005-04-17 16:05:31 -05002251 lpfc_linkdown(phba);
James Smart2e0fef82007-06-17 19:56:36 -05002252 phba->link_state = LPFC_HBA_ERROR;
dea31012005-04-17 16:05:31 -05002253
Dick Kennedy372c1872020-06-30 14:50:00 -07002254 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
2255 "0300 LATT: Cannot issue READ_LA: Data:%d\n", rc);
dea31012005-04-17 16:05:31 -05002256
2257 return;
2258}
2259
James Smarte59058c2008-08-24 21:49:00 -04002260/**
James Smart3621a712009-04-06 18:47:14 -04002261 * lpfc_parse_vpd - Parse VPD (Vital Product Data)
James Smarte59058c2008-08-24 21:49:00 -04002262 * @phba: pointer to lpfc hba data structure.
2263 * @vpd: pointer to the vital product data.
2264 * @len: length of the vital product data in bytes.
2265 *
2266 * This routine parses the Vital Product Data (VPD). The VPD is treated as
2267 * an array of characters. In this routine, the ModelName, ProgramType, and
2268 * ModelDesc, etc. fields of the phba data structure will be populated.
2269 *
2270 * Return codes
2271 * 0 - pointer to the VPD passed in is NULL
2272 * 1 - success
2273 **/
James Smart3772a992009-05-22 14:50:54 -04002274int
James Smart2e0fef82007-06-17 19:56:36 -05002275lpfc_parse_vpd(struct lpfc_hba *phba, uint8_t *vpd, int len)
dea31012005-04-17 16:05:31 -05002276{
2277 uint8_t lenlo, lenhi;
Anton Blanchard07da60c2007-03-21 08:41:47 -05002278 int Length;
dea31012005-04-17 16:05:31 -05002279 int i, j;
2280 int finished = 0;
2281 int index = 0;
2282
2283 if (!vpd)
2284 return 0;
2285
2286 /* Vital Product */
James Smarted957682007-06-17 19:56:37 -05002287 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
James Smarte8b62012007-08-02 11:10:09 -04002288 "0455 Vital Product Data: x%x x%x x%x x%x\n",
dea31012005-04-17 16:05:31 -05002289 (uint32_t) vpd[0], (uint32_t) vpd[1], (uint32_t) vpd[2],
2290 (uint32_t) vpd[3]);
Jamie Wellnitz74b72a52006-02-28 22:33:04 -05002291 while (!finished && (index < (len - 4))) {
dea31012005-04-17 16:05:31 -05002292 switch (vpd[index]) {
2293 case 0x82:
Jamie Wellnitz74b72a52006-02-28 22:33:04 -05002294 case 0x91:
dea31012005-04-17 16:05:31 -05002295 index += 1;
2296 lenlo = vpd[index];
2297 index += 1;
2298 lenhi = vpd[index];
2299 index += 1;
2300 i = ((((unsigned short)lenhi) << 8) + lenlo);
2301 index += i;
2302 break;
2303 case 0x90:
2304 index += 1;
2305 lenlo = vpd[index];
2306 index += 1;
2307 lenhi = vpd[index];
2308 index += 1;
2309 Length = ((((unsigned short)lenhi) << 8) + lenlo);
Jamie Wellnitz74b72a52006-02-28 22:33:04 -05002310 if (Length > len - index)
2311 Length = len - index;
dea31012005-04-17 16:05:31 -05002312 while (Length > 0) {
2313 /* Look for Serial Number */
2314 if ((vpd[index] == 'S') && (vpd[index+1] == 'N')) {
2315 index += 2;
2316 i = vpd[index];
2317 index += 1;
2318 j = 0;
2319 Length -= (3+i);
2320 while(i--) {
2321 phba->SerialNumber[j++] = vpd[index++];
2322 if (j == 31)
2323 break;
2324 }
2325 phba->SerialNumber[j] = 0;
2326 continue;
2327 }
2328 else if ((vpd[index] == 'V') && (vpd[index+1] == '1')) {
2329 phba->vpd_flag |= VPD_MODEL_DESC;
2330 index += 2;
2331 i = vpd[index];
2332 index += 1;
2333 j = 0;
2334 Length -= (3+i);
2335 while(i--) {
2336 phba->ModelDesc[j++] = vpd[index++];
2337 if (j == 255)
2338 break;
2339 }
2340 phba->ModelDesc[j] = 0;
2341 continue;
2342 }
2343 else if ((vpd[index] == 'V') && (vpd[index+1] == '2')) {
2344 phba->vpd_flag |= VPD_MODEL_NAME;
2345 index += 2;
2346 i = vpd[index];
2347 index += 1;
2348 j = 0;
2349 Length -= (3+i);
2350 while(i--) {
2351 phba->ModelName[j++] = vpd[index++];
2352 if (j == 79)
2353 break;
2354 }
2355 phba->ModelName[j] = 0;
2356 continue;
2357 }
2358 else if ((vpd[index] == 'V') && (vpd[index+1] == '3')) {
2359 phba->vpd_flag |= VPD_PROGRAM_TYPE;
2360 index += 2;
2361 i = vpd[index];
2362 index += 1;
2363 j = 0;
2364 Length -= (3+i);
2365 while(i--) {
2366 phba->ProgramType[j++] = vpd[index++];
2367 if (j == 255)
2368 break;
2369 }
2370 phba->ProgramType[j] = 0;
2371 continue;
2372 }
2373 else if ((vpd[index] == 'V') && (vpd[index+1] == '4')) {
2374 phba->vpd_flag |= VPD_PORT;
2375 index += 2;
2376 i = vpd[index];
2377 index += 1;
2378 j = 0;
2379 Length -= (3+i);
2380 while(i--) {
James Smartcd1c8302011-10-10 21:33:25 -04002381 if ((phba->sli_rev == LPFC_SLI_REV4) &&
2382 (phba->sli4_hba.pport_name_sta ==
2383 LPFC_SLI4_PPNAME_GET)) {
2384 j++;
2385 index++;
2386 } else
2387 phba->Port[j++] = vpd[index++];
2388 if (j == 19)
2389 break;
dea31012005-04-17 16:05:31 -05002390 }
James Smartcd1c8302011-10-10 21:33:25 -04002391 if ((phba->sli_rev != LPFC_SLI_REV4) ||
2392 (phba->sli4_hba.pport_name_sta ==
2393 LPFC_SLI4_PPNAME_NON))
2394 phba->Port[j] = 0;
dea31012005-04-17 16:05:31 -05002395 continue;
2396 }
2397 else {
2398 index += 2;
2399 i = vpd[index];
2400 index += 1;
2401 index += i;
2402 Length -= (3 + i);
2403 }
2404 }
2405 finished = 0;
2406 break;
2407 case 0x78:
2408 finished = 1;
2409 break;
2410 default:
2411 index ++;
2412 break;
2413 }
Jamie Wellnitz74b72a52006-02-28 22:33:04 -05002414 }
dea31012005-04-17 16:05:31 -05002415
2416 return(1);
2417}
2418
James Smarte59058c2008-08-24 21:49:00 -04002419/**
James Smart3621a712009-04-06 18:47:14 -04002420 * lpfc_get_hba_model_desc - Retrieve HBA device model name and description
James Smarte59058c2008-08-24 21:49:00 -04002421 * @phba: pointer to lpfc hba data structure.
2422 * @mdp: pointer to the data structure to hold the derived model name.
2423 * @descp: pointer to the data structure to hold the derived description.
2424 *
2425 * This routine retrieves HBA's description based on its registered PCI device
2426 * ID. The @descp passed into this function points to an array of 256 chars. It
2427 * shall be returned with the model name, maximum speed, and the host bus type.
2428 * The @mdp passed into this function points to an array of 80 chars. When the
2429 * function returns, the @mdp will be filled with the model name.
2430 **/
dea31012005-04-17 16:05:31 -05002431static void
James Smart2e0fef82007-06-17 19:56:36 -05002432lpfc_get_hba_model_desc(struct lpfc_hba *phba, uint8_t *mdp, uint8_t *descp)
dea31012005-04-17 16:05:31 -05002433{
2434 lpfc_vpd_t *vp;
James.Smart@Emulex.Comfefcb2b2005-11-28 15:08:56 -05002435 uint16_t dev_id = phba->pcidev->device;
Jamie Wellnitz74b72a52006-02-28 22:33:04 -05002436 int max_speed;
James Smart84774a42008-08-24 21:50:06 -04002437 int GE = 0;
James Smartda0436e2009-05-22 14:51:39 -04002438 int oneConnect = 0; /* default is not a oneConnect */
Jamie Wellnitz74b72a52006-02-28 22:33:04 -05002439 struct {
James Smarta747c9c2009-11-18 15:41:10 -05002440 char *name;
2441 char *bus;
2442 char *function;
2443 } m = {"<Unknown>", "", ""};
Jamie Wellnitz74b72a52006-02-28 22:33:04 -05002444
2445 if (mdp && mdp[0] != '\0'
2446 && descp && descp[0] != '\0')
2447 return;
2448
James Smartfbd8a6b2018-02-22 08:18:45 -08002449 if (phba->lmt & LMT_64Gb)
2450 max_speed = 64;
2451 else if (phba->lmt & LMT_32Gb)
James Smartd38dd522015-08-31 16:48:17 -04002452 max_speed = 32;
2453 else if (phba->lmt & LMT_16Gb)
James Smartc0c11512011-05-24 11:41:34 -04002454 max_speed = 16;
2455 else if (phba->lmt & LMT_10Gb)
Jamie Wellnitz74b72a52006-02-28 22:33:04 -05002456 max_speed = 10;
2457 else if (phba->lmt & LMT_8Gb)
2458 max_speed = 8;
2459 else if (phba->lmt & LMT_4Gb)
2460 max_speed = 4;
2461 else if (phba->lmt & LMT_2Gb)
2462 max_speed = 2;
James Smart4169d862012-09-29 11:32:09 -04002463 else if (phba->lmt & LMT_1Gb)
Jamie Wellnitz74b72a52006-02-28 22:33:04 -05002464 max_speed = 1;
James Smart4169d862012-09-29 11:32:09 -04002465 else
2466 max_speed = 0;
dea31012005-04-17 16:05:31 -05002467
2468 vp = &phba->vpd;
dea31012005-04-17 16:05:31 -05002469
James.Smart@Emulex.Come4adb202005-11-28 11:42:12 -05002470 switch (dev_id) {
James.Smart@Emulex.Com06325e72005-06-25 10:34:22 -04002471 case PCI_DEVICE_ID_FIREFLY:
James Smart12222f42014-05-21 08:05:19 -04002472 m = (typeof(m)){"LP6000", "PCI",
2473 "Obsolete, Unsupported Fibre Channel Adapter"};
James.Smart@Emulex.Com06325e72005-06-25 10:34:22 -04002474 break;
dea31012005-04-17 16:05:31 -05002475 case PCI_DEVICE_ID_SUPERFLY:
2476 if (vp->rev.biuRev >= 1 && vp->rev.biuRev <= 3)
James Smart12222f42014-05-21 08:05:19 -04002477 m = (typeof(m)){"LP7000", "PCI", ""};
dea31012005-04-17 16:05:31 -05002478 else
James Smart12222f42014-05-21 08:05:19 -04002479 m = (typeof(m)){"LP7000E", "PCI", ""};
2480 m.function = "Obsolete, Unsupported Fibre Channel Adapter";
dea31012005-04-17 16:05:31 -05002481 break;
2482 case PCI_DEVICE_ID_DRAGONFLY:
James Smarta747c9c2009-11-18 15:41:10 -05002483 m = (typeof(m)){"LP8000", "PCI",
James Smart12222f42014-05-21 08:05:19 -04002484 "Obsolete, Unsupported Fibre Channel Adapter"};
dea31012005-04-17 16:05:31 -05002485 break;
2486 case PCI_DEVICE_ID_CENTAUR:
2487 if (FC_JEDEC_ID(vp->rev.biuRev) == CENTAUR_2G_JEDEC_ID)
James Smart12222f42014-05-21 08:05:19 -04002488 m = (typeof(m)){"LP9002", "PCI", ""};
dea31012005-04-17 16:05:31 -05002489 else
James Smart12222f42014-05-21 08:05:19 -04002490 m = (typeof(m)){"LP9000", "PCI", ""};
2491 m.function = "Obsolete, Unsupported Fibre Channel Adapter";
dea31012005-04-17 16:05:31 -05002492 break;
2493 case PCI_DEVICE_ID_RFLY:
James Smarta747c9c2009-11-18 15:41:10 -05002494 m = (typeof(m)){"LP952", "PCI",
James Smart12222f42014-05-21 08:05:19 -04002495 "Obsolete, Unsupported Fibre Channel Adapter"};
dea31012005-04-17 16:05:31 -05002496 break;
2497 case PCI_DEVICE_ID_PEGASUS:
James Smarta747c9c2009-11-18 15:41:10 -05002498 m = (typeof(m)){"LP9802", "PCI-X",
James Smart12222f42014-05-21 08:05:19 -04002499 "Obsolete, Unsupported Fibre Channel Adapter"};
dea31012005-04-17 16:05:31 -05002500 break;
2501 case PCI_DEVICE_ID_THOR:
James Smarta747c9c2009-11-18 15:41:10 -05002502 m = (typeof(m)){"LP10000", "PCI-X",
James Smart12222f42014-05-21 08:05:19 -04002503 "Obsolete, Unsupported Fibre Channel Adapter"};
dea31012005-04-17 16:05:31 -05002504 break;
2505 case PCI_DEVICE_ID_VIPER:
James Smarta747c9c2009-11-18 15:41:10 -05002506 m = (typeof(m)){"LPX1000", "PCI-X",
James Smart12222f42014-05-21 08:05:19 -04002507 "Obsolete, Unsupported Fibre Channel Adapter"};
dea31012005-04-17 16:05:31 -05002508 break;
2509 case PCI_DEVICE_ID_PFLY:
James Smarta747c9c2009-11-18 15:41:10 -05002510 m = (typeof(m)){"LP982", "PCI-X",
James Smart12222f42014-05-21 08:05:19 -04002511 "Obsolete, Unsupported Fibre Channel Adapter"};
dea31012005-04-17 16:05:31 -05002512 break;
2513 case PCI_DEVICE_ID_TFLY:
James Smarta747c9c2009-11-18 15:41:10 -05002514 m = (typeof(m)){"LP1050", "PCI-X",
James Smart12222f42014-05-21 08:05:19 -04002515 "Obsolete, Unsupported Fibre Channel Adapter"};
dea31012005-04-17 16:05:31 -05002516 break;
2517 case PCI_DEVICE_ID_HELIOS:
James Smarta747c9c2009-11-18 15:41:10 -05002518 m = (typeof(m)){"LP11000", "PCI-X2",
James Smart12222f42014-05-21 08:05:19 -04002519 "Obsolete, Unsupported Fibre Channel Adapter"};
dea31012005-04-17 16:05:31 -05002520 break;
James.Smart@Emulex.Come4adb202005-11-28 11:42:12 -05002521 case PCI_DEVICE_ID_HELIOS_SCSP:
James Smarta747c9c2009-11-18 15:41:10 -05002522 m = (typeof(m)){"LP11000-SP", "PCI-X2",
James Smart12222f42014-05-21 08:05:19 -04002523 "Obsolete, Unsupported Fibre Channel Adapter"};
James.Smart@Emulex.Come4adb202005-11-28 11:42:12 -05002524 break;
2525 case PCI_DEVICE_ID_HELIOS_DCSP:
James Smarta747c9c2009-11-18 15:41:10 -05002526 m = (typeof(m)){"LP11002-SP", "PCI-X2",
James Smart12222f42014-05-21 08:05:19 -04002527 "Obsolete, Unsupported Fibre Channel Adapter"};
James.Smart@Emulex.Come4adb202005-11-28 11:42:12 -05002528 break;
2529 case PCI_DEVICE_ID_NEPTUNE:
James Smart12222f42014-05-21 08:05:19 -04002530 m = (typeof(m)){"LPe1000", "PCIe",
2531 "Obsolete, Unsupported Fibre Channel Adapter"};
James.Smart@Emulex.Come4adb202005-11-28 11:42:12 -05002532 break;
2533 case PCI_DEVICE_ID_NEPTUNE_SCSP:
James Smart12222f42014-05-21 08:05:19 -04002534 m = (typeof(m)){"LPe1000-SP", "PCIe",
2535 "Obsolete, Unsupported Fibre Channel Adapter"};
James.Smart@Emulex.Come4adb202005-11-28 11:42:12 -05002536 break;
2537 case PCI_DEVICE_ID_NEPTUNE_DCSP:
James Smart12222f42014-05-21 08:05:19 -04002538 m = (typeof(m)){"LPe1002-SP", "PCIe",
2539 "Obsolete, Unsupported Fibre Channel Adapter"};
James.Smart@Emulex.Come4adb202005-11-28 11:42:12 -05002540 break;
dea31012005-04-17 16:05:31 -05002541 case PCI_DEVICE_ID_BMID:
James Smarta747c9c2009-11-18 15:41:10 -05002542 m = (typeof(m)){"LP1150", "PCI-X2", "Fibre Channel Adapter"};
dea31012005-04-17 16:05:31 -05002543 break;
2544 case PCI_DEVICE_ID_BSMB:
James Smart12222f42014-05-21 08:05:19 -04002545 m = (typeof(m)){"LP111", "PCI-X2",
2546 "Obsolete, Unsupported Fibre Channel Adapter"};
dea31012005-04-17 16:05:31 -05002547 break;
2548 case PCI_DEVICE_ID_ZEPHYR:
James Smarta747c9c2009-11-18 15:41:10 -05002549 m = (typeof(m)){"LPe11000", "PCIe", "Fibre Channel Adapter"};
dea31012005-04-17 16:05:31 -05002550 break;
James.Smart@Emulex.Come4adb202005-11-28 11:42:12 -05002551 case PCI_DEVICE_ID_ZEPHYR_SCSP:
James Smarta747c9c2009-11-18 15:41:10 -05002552 m = (typeof(m)){"LPe11000", "PCIe", "Fibre Channel Adapter"};
James.Smart@Emulex.Come4adb202005-11-28 11:42:12 -05002553 break;
2554 case PCI_DEVICE_ID_ZEPHYR_DCSP:
James Smarta747c9c2009-11-18 15:41:10 -05002555 m = (typeof(m)){"LP2105", "PCIe", "FCoE Adapter"};
James Smarta257bf92009-04-06 18:48:10 -04002556 GE = 1;
James.Smart@Emulex.Come4adb202005-11-28 11:42:12 -05002557 break;
dea31012005-04-17 16:05:31 -05002558 case PCI_DEVICE_ID_ZMID:
James Smarta747c9c2009-11-18 15:41:10 -05002559 m = (typeof(m)){"LPe1150", "PCIe", "Fibre Channel Adapter"};
dea31012005-04-17 16:05:31 -05002560 break;
2561 case PCI_DEVICE_ID_ZSMB:
James Smarta747c9c2009-11-18 15:41:10 -05002562 m = (typeof(m)){"LPe111", "PCIe", "Fibre Channel Adapter"};
dea31012005-04-17 16:05:31 -05002563 break;
2564 case PCI_DEVICE_ID_LP101:
James Smart12222f42014-05-21 08:05:19 -04002565 m = (typeof(m)){"LP101", "PCI-X",
2566 "Obsolete, Unsupported Fibre Channel Adapter"};
dea31012005-04-17 16:05:31 -05002567 break;
2568 case PCI_DEVICE_ID_LP10000S:
James Smart12222f42014-05-21 08:05:19 -04002569 m = (typeof(m)){"LP10000-S", "PCI",
2570 "Obsolete, Unsupported Fibre Channel Adapter"};
James.Smart@Emulex.Com06325e72005-06-25 10:34:22 -04002571 break;
James.Smart@Emulex.Come4adb202005-11-28 11:42:12 -05002572 case PCI_DEVICE_ID_LP11000S:
James Smart12222f42014-05-21 08:05:19 -04002573 m = (typeof(m)){"LP11000-S", "PCI-X2",
2574 "Obsolete, Unsupported Fibre Channel Adapter"};
James Smart18a3b592006-12-02 13:35:08 -05002575 break;
James.Smart@Emulex.Come4adb202005-11-28 11:42:12 -05002576 case PCI_DEVICE_ID_LPE11000S:
James Smart12222f42014-05-21 08:05:19 -04002577 m = (typeof(m)){"LPe11000-S", "PCIe",
2578 "Obsolete, Unsupported Fibre Channel Adapter"};
James.Smart@Emulex.Com5cc36b32005-11-28 11:42:19 -05002579 break;
James Smartb87eab32007-04-25 09:53:28 -04002580 case PCI_DEVICE_ID_SAT:
James Smarta747c9c2009-11-18 15:41:10 -05002581 m = (typeof(m)){"LPe12000", "PCIe", "Fibre Channel Adapter"};
James Smartb87eab32007-04-25 09:53:28 -04002582 break;
2583 case PCI_DEVICE_ID_SAT_MID:
James Smarta747c9c2009-11-18 15:41:10 -05002584 m = (typeof(m)){"LPe1250", "PCIe", "Fibre Channel Adapter"};
James Smartb87eab32007-04-25 09:53:28 -04002585 break;
2586 case PCI_DEVICE_ID_SAT_SMB:
James Smarta747c9c2009-11-18 15:41:10 -05002587 m = (typeof(m)){"LPe121", "PCIe", "Fibre Channel Adapter"};
James Smartb87eab32007-04-25 09:53:28 -04002588 break;
2589 case PCI_DEVICE_ID_SAT_DCSP:
James Smarta747c9c2009-11-18 15:41:10 -05002590 m = (typeof(m)){"LPe12002-SP", "PCIe", "Fibre Channel Adapter"};
James Smartb87eab32007-04-25 09:53:28 -04002591 break;
2592 case PCI_DEVICE_ID_SAT_SCSP:
James Smarta747c9c2009-11-18 15:41:10 -05002593 m = (typeof(m)){"LPe12000-SP", "PCIe", "Fibre Channel Adapter"};
James Smartb87eab32007-04-25 09:53:28 -04002594 break;
2595 case PCI_DEVICE_ID_SAT_S:
James Smarta747c9c2009-11-18 15:41:10 -05002596 m = (typeof(m)){"LPe12000-S", "PCIe", "Fibre Channel Adapter"};
James Smartb87eab32007-04-25 09:53:28 -04002597 break;
James Smart84774a42008-08-24 21:50:06 -04002598 case PCI_DEVICE_ID_HORNET:
James Smart12222f42014-05-21 08:05:19 -04002599 m = (typeof(m)){"LP21000", "PCIe",
2600 "Obsolete, Unsupported FCoE Adapter"};
James Smart84774a42008-08-24 21:50:06 -04002601 GE = 1;
2602 break;
2603 case PCI_DEVICE_ID_PROTEUS_VF:
James Smarta747c9c2009-11-18 15:41:10 -05002604 m = (typeof(m)){"LPev12000", "PCIe IOV",
James Smart12222f42014-05-21 08:05:19 -04002605 "Obsolete, Unsupported Fibre Channel Adapter"};
James Smart84774a42008-08-24 21:50:06 -04002606 break;
2607 case PCI_DEVICE_ID_PROTEUS_PF:
James Smarta747c9c2009-11-18 15:41:10 -05002608 m = (typeof(m)){"LPev12000", "PCIe IOV",
James Smart12222f42014-05-21 08:05:19 -04002609 "Obsolete, Unsupported Fibre Channel Adapter"};
James Smart84774a42008-08-24 21:50:06 -04002610 break;
2611 case PCI_DEVICE_ID_PROTEUS_S:
James Smarta747c9c2009-11-18 15:41:10 -05002612 m = (typeof(m)){"LPemv12002-S", "PCIe IOV",
James Smart12222f42014-05-21 08:05:19 -04002613 "Obsolete, Unsupported Fibre Channel Adapter"};
James Smart84774a42008-08-24 21:50:06 -04002614 break;
James Smartda0436e2009-05-22 14:51:39 -04002615 case PCI_DEVICE_ID_TIGERSHARK:
2616 oneConnect = 1;
James Smarta747c9c2009-11-18 15:41:10 -05002617 m = (typeof(m)){"OCe10100", "PCIe", "FCoE"};
James Smartda0436e2009-05-22 14:51:39 -04002618 break;
James Smarta747c9c2009-11-18 15:41:10 -05002619 case PCI_DEVICE_ID_TOMCAT:
James Smart6669f9b2009-10-02 15:16:45 -04002620 oneConnect = 1;
James Smarta747c9c2009-11-18 15:41:10 -05002621 m = (typeof(m)){"OCe11100", "PCIe", "FCoE"};
2622 break;
2623 case PCI_DEVICE_ID_FALCON:
2624 m = (typeof(m)){"LPSe12002-ML1-E", "PCIe",
2625 "EmulexSecure Fibre"};
James Smart6669f9b2009-10-02 15:16:45 -04002626 break;
James Smart98fc5dd2010-06-07 15:24:29 -04002627 case PCI_DEVICE_ID_BALIUS:
2628 m = (typeof(m)){"LPVe12002", "PCIe Shared I/O",
James Smart12222f42014-05-21 08:05:19 -04002629 "Obsolete, Unsupported Fibre Channel Adapter"};
James Smart98fc5dd2010-06-07 15:24:29 -04002630 break;
James Smart085c6472010-11-20 23:11:37 -05002631 case PCI_DEVICE_ID_LANCER_FC:
James Smartc0c11512011-05-24 11:41:34 -04002632 m = (typeof(m)){"LPe16000", "PCIe", "Fibre Channel Adapter"};
James Smart085c6472010-11-20 23:11:37 -05002633 break;
James Smart12222f42014-05-21 08:05:19 -04002634 case PCI_DEVICE_ID_LANCER_FC_VF:
2635 m = (typeof(m)){"LPe16000", "PCIe",
2636 "Obsolete, Unsupported Fibre Channel Adapter"};
2637 break;
James Smart085c6472010-11-20 23:11:37 -05002638 case PCI_DEVICE_ID_LANCER_FCOE:
2639 oneConnect = 1;
James Smart079b5c92011-08-21 21:48:49 -04002640 m = (typeof(m)){"OCe15100", "PCIe", "FCoE"};
James Smart085c6472010-11-20 23:11:37 -05002641 break;
James Smart12222f42014-05-21 08:05:19 -04002642 case PCI_DEVICE_ID_LANCER_FCOE_VF:
2643 oneConnect = 1;
2644 m = (typeof(m)){"OCe15100", "PCIe",
2645 "Obsolete, Unsupported FCoE"};
2646 break;
James Smartd38dd522015-08-31 16:48:17 -04002647 case PCI_DEVICE_ID_LANCER_G6_FC:
2648 m = (typeof(m)){"LPe32000", "PCIe", "Fibre Channel Adapter"};
2649 break;
James Smartc238b9b2018-02-22 08:18:44 -08002650 case PCI_DEVICE_ID_LANCER_G7_FC:
2651 m = (typeof(m)){"LPe36000", "PCIe", "Fibre Channel Adapter"};
2652 break;
James Smartf449a3d2021-07-22 15:17:16 -07002653 case PCI_DEVICE_ID_LANCER_G7P_FC:
2654 m = (typeof(m)){"LPe38000", "PCIe", "Fibre Channel Adapter"};
2655 break;
James Smartf8cafd32012-08-03 12:42:51 -04002656 case PCI_DEVICE_ID_SKYHAWK:
2657 case PCI_DEVICE_ID_SKYHAWK_VF:
2658 oneConnect = 1;
2659 m = (typeof(m)){"OCe14000", "PCIe", "FCoE"};
2660 break;
James.Smart@Emulex.Com5cc36b32005-11-28 11:42:19 -05002661 default:
James Smarta747c9c2009-11-18 15:41:10 -05002662 m = (typeof(m)){"Unknown", "", ""};
James.Smart@Emulex.Come4adb202005-11-28 11:42:12 -05002663 break;
dea31012005-04-17 16:05:31 -05002664 }
Jamie Wellnitz74b72a52006-02-28 22:33:04 -05002665
2666 if (mdp && mdp[0] == '\0')
2667 snprintf(mdp, 79,"%s", m.name);
James Smartc0c11512011-05-24 11:41:34 -04002668 /*
2669 * oneConnect hba requires special processing, they are all initiators
James Smartda0436e2009-05-22 14:51:39 -04002670 * and we put the port number on the end
2671 */
2672 if (descp && descp[0] == '\0') {
2673 if (oneConnect)
2674 snprintf(descp, 255,
James Smart4169d862012-09-29 11:32:09 -04002675 "Emulex OneConnect %s, %s Initiator %s",
James Smarta747c9c2009-11-18 15:41:10 -05002676 m.name, m.function,
James Smartda0436e2009-05-22 14:51:39 -04002677 phba->Port);
James Smart4169d862012-09-29 11:32:09 -04002678 else if (max_speed == 0)
2679 snprintf(descp, 255,
Sebastian Herbszt290237d2015-08-31 16:48:08 -04002680 "Emulex %s %s %s",
James Smart4169d862012-09-29 11:32:09 -04002681 m.name, m.bus, m.function);
James Smartda0436e2009-05-22 14:51:39 -04002682 else
2683 snprintf(descp, 255,
2684 "Emulex %s %d%s %s %s",
James Smarta747c9c2009-11-18 15:41:10 -05002685 m.name, max_speed, (GE) ? "GE" : "Gb",
2686 m.bus, m.function);
James Smartda0436e2009-05-22 14:51:39 -04002687 }
dea31012005-04-17 16:05:31 -05002688}
2689
James Smarte59058c2008-08-24 21:49:00 -04002690/**
James Smart3621a712009-04-06 18:47:14 -04002691 * lpfc_post_buffer - Post IOCB(s) with DMA buffer descriptor(s) to a IOCB ring
James Smarte59058c2008-08-24 21:49:00 -04002692 * @phba: pointer to lpfc hba data structure.
2693 * @pring: pointer to a IOCB ring.
2694 * @cnt: the number of IOCBs to be posted to the IOCB ring.
2695 *
2696 * This routine posts a given number of IOCBs with the associated DMA buffer
2697 * descriptors specified by the cnt argument to the given IOCB ring.
2698 *
2699 * Return codes
2700 * The number of IOCBs NOT able to be posted to the IOCB ring.
2701 **/
dea31012005-04-17 16:05:31 -05002702int
James Smart495a7142008-06-14 22:52:59 -04002703lpfc_post_buffer(struct lpfc_hba *phba, struct lpfc_sli_ring *pring, int cnt)
dea31012005-04-17 16:05:31 -05002704{
2705 IOCB_t *icmd;
James.Smart@Emulex.Com0bd4ca22005-10-28 20:30:02 -04002706 struct lpfc_iocbq *iocb;
dea31012005-04-17 16:05:31 -05002707 struct lpfc_dmabuf *mp1, *mp2;
2708
2709 cnt += pring->missbufcnt;
2710
2711 /* While there are buffers to post */
2712 while (cnt > 0) {
2713 /* Allocate buffer for command iocb */
James.Smart@Emulex.Com0bd4ca22005-10-28 20:30:02 -04002714 iocb = lpfc_sli_get_iocbq(phba);
dea31012005-04-17 16:05:31 -05002715 if (iocb == NULL) {
2716 pring->missbufcnt = cnt;
2717 return cnt;
2718 }
dea31012005-04-17 16:05:31 -05002719 icmd = &iocb->iocb;
2720
2721 /* 2 buffers can be posted per command */
2722 /* Allocate buffer to post */
2723 mp1 = kmalloc(sizeof (struct lpfc_dmabuf), GFP_KERNEL);
2724 if (mp1)
James Smart98c9ea52007-10-27 13:37:33 -04002725 mp1->virt = lpfc_mbuf_alloc(phba, MEM_PRI, &mp1->phys);
2726 if (!mp1 || !mp1->virt) {
Jesper Juhlc9475cb2005-11-07 01:01:26 -08002727 kfree(mp1);
James Bottomley604a3e32005-10-29 10:28:33 -05002728 lpfc_sli_release_iocbq(phba, iocb);
dea31012005-04-17 16:05:31 -05002729 pring->missbufcnt = cnt;
2730 return cnt;
2731 }
2732
2733 INIT_LIST_HEAD(&mp1->list);
2734 /* Allocate buffer to post */
2735 if (cnt > 1) {
2736 mp2 = kmalloc(sizeof (struct lpfc_dmabuf), GFP_KERNEL);
2737 if (mp2)
2738 mp2->virt = lpfc_mbuf_alloc(phba, MEM_PRI,
2739 &mp2->phys);
James Smart98c9ea52007-10-27 13:37:33 -04002740 if (!mp2 || !mp2->virt) {
Jesper Juhlc9475cb2005-11-07 01:01:26 -08002741 kfree(mp2);
dea31012005-04-17 16:05:31 -05002742 lpfc_mbuf_free(phba, mp1->virt, mp1->phys);
2743 kfree(mp1);
James Bottomley604a3e32005-10-29 10:28:33 -05002744 lpfc_sli_release_iocbq(phba, iocb);
dea31012005-04-17 16:05:31 -05002745 pring->missbufcnt = cnt;
2746 return cnt;
2747 }
2748
2749 INIT_LIST_HEAD(&mp2->list);
2750 } else {
2751 mp2 = NULL;
2752 }
2753
2754 icmd->un.cont64[0].addrHigh = putPaddrHigh(mp1->phys);
2755 icmd->un.cont64[0].addrLow = putPaddrLow(mp1->phys);
2756 icmd->un.cont64[0].tus.f.bdeSize = FCELSSIZE;
2757 icmd->ulpBdeCount = 1;
2758 cnt--;
2759 if (mp2) {
2760 icmd->un.cont64[1].addrHigh = putPaddrHigh(mp2->phys);
2761 icmd->un.cont64[1].addrLow = putPaddrLow(mp2->phys);
2762 icmd->un.cont64[1].tus.f.bdeSize = FCELSSIZE;
2763 cnt--;
2764 icmd->ulpBdeCount = 2;
2765 }
2766
2767 icmd->ulpCommand = CMD_QUE_RING_BUF64_CN;
2768 icmd->ulpLe = 1;
2769
James Smart3772a992009-05-22 14:50:54 -04002770 if (lpfc_sli_issue_iocb(phba, pring->ringno, iocb, 0) ==
2771 IOCB_ERROR) {
dea31012005-04-17 16:05:31 -05002772 lpfc_mbuf_free(phba, mp1->virt, mp1->phys);
2773 kfree(mp1);
2774 cnt++;
2775 if (mp2) {
2776 lpfc_mbuf_free(phba, mp2->virt, mp2->phys);
2777 kfree(mp2);
2778 cnt++;
2779 }
James Bottomley604a3e32005-10-29 10:28:33 -05002780 lpfc_sli_release_iocbq(phba, iocb);
dea31012005-04-17 16:05:31 -05002781 pring->missbufcnt = cnt;
dea31012005-04-17 16:05:31 -05002782 return cnt;
2783 }
dea31012005-04-17 16:05:31 -05002784 lpfc_sli_ringpostbuf_put(phba, pring, mp1);
James Smart92d7f7b2007-06-17 19:56:38 -05002785 if (mp2)
dea31012005-04-17 16:05:31 -05002786 lpfc_sli_ringpostbuf_put(phba, pring, mp2);
dea31012005-04-17 16:05:31 -05002787 }
2788 pring->missbufcnt = 0;
2789 return 0;
2790}
2791
James Smarte59058c2008-08-24 21:49:00 -04002792/**
James Smart3621a712009-04-06 18:47:14 -04002793 * lpfc_post_rcv_buf - Post the initial receive IOCB buffers to ELS ring
James Smarte59058c2008-08-24 21:49:00 -04002794 * @phba: pointer to lpfc hba data structure.
2795 *
2796 * This routine posts initial receive IOCB buffers to the ELS ring. The
2797 * current number of initial IOCB buffers specified by LPFC_BUF_RING0 is
James Smart895427b2017-02-12 13:52:30 -08002798 * set to 64 IOCBs. SLI3 only.
James Smarte59058c2008-08-24 21:49:00 -04002799 *
2800 * Return codes
2801 * 0 - success (currently always success)
2802 **/
dea31012005-04-17 16:05:31 -05002803static int
James Smart2e0fef82007-06-17 19:56:36 -05002804lpfc_post_rcv_buf(struct lpfc_hba *phba)
dea31012005-04-17 16:05:31 -05002805{
2806 struct lpfc_sli *psli = &phba->sli;
2807
2808 /* Ring 0, ELS / CT buffers */
James Smart895427b2017-02-12 13:52:30 -08002809 lpfc_post_buffer(phba, &psli->sli3_ring[LPFC_ELS_RING], LPFC_BUF_RING0);
dea31012005-04-17 16:05:31 -05002810 /* Ring 2 - FCP no buffers needed */
2811
2812 return 0;
2813}
2814
2815#define S(N,V) (((V)<<(N))|((V)>>(32-(N))))
2816
James Smarte59058c2008-08-24 21:49:00 -04002817/**
James Smart3621a712009-04-06 18:47:14 -04002818 * lpfc_sha_init - Set up initial array of hash table entries
James Smarte59058c2008-08-24 21:49:00 -04002819 * @HashResultPointer: pointer to an array as hash table.
2820 *
2821 * This routine sets up the initial values to the array of hash table entries
2822 * for the LC HBAs.
2823 **/
dea31012005-04-17 16:05:31 -05002824static void
2825lpfc_sha_init(uint32_t * HashResultPointer)
2826{
2827 HashResultPointer[0] = 0x67452301;
2828 HashResultPointer[1] = 0xEFCDAB89;
2829 HashResultPointer[2] = 0x98BADCFE;
2830 HashResultPointer[3] = 0x10325476;
2831 HashResultPointer[4] = 0xC3D2E1F0;
2832}
2833
James Smarte59058c2008-08-24 21:49:00 -04002834/**
James Smart3621a712009-04-06 18:47:14 -04002835 * lpfc_sha_iterate - Iterate initial hash table with the working hash table
James Smarte59058c2008-08-24 21:49:00 -04002836 * @HashResultPointer: pointer to an initial/result hash table.
2837 * @HashWorkingPointer: pointer to an working hash table.
2838 *
2839 * This routine iterates an initial hash table pointed by @HashResultPointer
2840 * with the values from the working hash table pointeed by @HashWorkingPointer.
2841 * The results are putting back to the initial hash table, returned through
2842 * the @HashResultPointer as the result hash table.
2843 **/
dea31012005-04-17 16:05:31 -05002844static void
2845lpfc_sha_iterate(uint32_t * HashResultPointer, uint32_t * HashWorkingPointer)
2846{
2847 int t;
2848 uint32_t TEMP;
2849 uint32_t A, B, C, D, E;
2850 t = 16;
2851 do {
2852 HashWorkingPointer[t] =
2853 S(1,
2854 HashWorkingPointer[t - 3] ^ HashWorkingPointer[t -
2855 8] ^
2856 HashWorkingPointer[t - 14] ^ HashWorkingPointer[t - 16]);
2857 } while (++t <= 79);
2858 t = 0;
2859 A = HashResultPointer[0];
2860 B = HashResultPointer[1];
2861 C = HashResultPointer[2];
2862 D = HashResultPointer[3];
2863 E = HashResultPointer[4];
2864
2865 do {
2866 if (t < 20) {
2867 TEMP = ((B & C) | ((~B) & D)) + 0x5A827999;
2868 } else if (t < 40) {
2869 TEMP = (B ^ C ^ D) + 0x6ED9EBA1;
2870 } else if (t < 60) {
2871 TEMP = ((B & C) | (B & D) | (C & D)) + 0x8F1BBCDC;
2872 } else {
2873 TEMP = (B ^ C ^ D) + 0xCA62C1D6;
2874 }
2875 TEMP += S(5, A) + E + HashWorkingPointer[t];
2876 E = D;
2877 D = C;
2878 C = S(30, B);
2879 B = A;
2880 A = TEMP;
2881 } while (++t <= 79);
2882
2883 HashResultPointer[0] += A;
2884 HashResultPointer[1] += B;
2885 HashResultPointer[2] += C;
2886 HashResultPointer[3] += D;
2887 HashResultPointer[4] += E;
2888
2889}
2890
James Smarte59058c2008-08-24 21:49:00 -04002891/**
James Smart3621a712009-04-06 18:47:14 -04002892 * lpfc_challenge_key - Create challenge key based on WWPN of the HBA
James Smarte59058c2008-08-24 21:49:00 -04002893 * @RandomChallenge: pointer to the entry of host challenge random number array.
2894 * @HashWorking: pointer to the entry of the working hash array.
2895 *
2896 * This routine calculates the working hash array referred by @HashWorking
2897 * from the challenge random numbers associated with the host, referred by
2898 * @RandomChallenge. The result is put into the entry of the working hash
2899 * array and returned by reference through @HashWorking.
2900 **/
dea31012005-04-17 16:05:31 -05002901static void
2902lpfc_challenge_key(uint32_t * RandomChallenge, uint32_t * HashWorking)
2903{
2904 *HashWorking = (*RandomChallenge ^ *HashWorking);
2905}
2906
James Smarte59058c2008-08-24 21:49:00 -04002907/**
James Smart3621a712009-04-06 18:47:14 -04002908 * lpfc_hba_init - Perform special handling for LC HBA initialization
James Smarte59058c2008-08-24 21:49:00 -04002909 * @phba: pointer to lpfc hba data structure.
2910 * @hbainit: pointer to an array of unsigned 32-bit integers.
2911 *
2912 * This routine performs the special handling for LC HBA initialization.
2913 **/
dea31012005-04-17 16:05:31 -05002914void
2915lpfc_hba_init(struct lpfc_hba *phba, uint32_t *hbainit)
2916{
2917 int t;
2918 uint32_t *HashWorking;
James Smart2e0fef82007-06-17 19:56:36 -05002919 uint32_t *pwwnn = (uint32_t *) phba->wwnn;
dea31012005-04-17 16:05:31 -05002920
Mariusz Kozlowskibbfbbbc2007-08-11 10:13:24 +02002921 HashWorking = kcalloc(80, sizeof(uint32_t), GFP_KERNEL);
dea31012005-04-17 16:05:31 -05002922 if (!HashWorking)
2923 return;
2924
dea31012005-04-17 16:05:31 -05002925 HashWorking[0] = HashWorking[78] = *pwwnn++;
2926 HashWorking[1] = HashWorking[79] = *pwwnn;
2927
2928 for (t = 0; t < 7; t++)
2929 lpfc_challenge_key(phba->RandomData + t, HashWorking + t);
2930
2931 lpfc_sha_init(hbainit);
2932 lpfc_sha_iterate(hbainit, HashWorking);
2933 kfree(HashWorking);
2934}
2935
James Smarte59058c2008-08-24 21:49:00 -04002936/**
James Smart3621a712009-04-06 18:47:14 -04002937 * lpfc_cleanup - Performs vport cleanups before deleting a vport
James Smarte59058c2008-08-24 21:49:00 -04002938 * @vport: pointer to a virtual N_Port data structure.
2939 *
2940 * This routine performs the necessary cleanups before deleting the @vport.
2941 * It invokes the discovery state machine to perform necessary state
2942 * transitions and to release the ndlps associated with the @vport. Note,
2943 * the physical port is treated as @vport 0.
2944 **/
James Smart87af33f2007-10-27 13:37:43 -04002945void
James Smart2e0fef82007-06-17 19:56:36 -05002946lpfc_cleanup(struct lpfc_vport *vport)
dea31012005-04-17 16:05:31 -05002947{
James Smart87af33f2007-10-27 13:37:43 -04002948 struct lpfc_hba *phba = vport->phba;
dea31012005-04-17 16:05:31 -05002949 struct lpfc_nodelist *ndlp, *next_ndlp;
James Smarta8adb832007-10-27 13:37:53 -04002950 int i = 0;
dea31012005-04-17 16:05:31 -05002951
James Smart87af33f2007-10-27 13:37:43 -04002952 if (phba->link_state > LPFC_LINK_DOWN)
2953 lpfc_port_link_failure(vport);
2954
Gaurav Srivastava5e633302021-06-08 10:05:49 +05302955 /* Clean up VMID resources */
2956 if (lpfc_is_vmid_enabled(phba))
2957 lpfc_vmid_vport_cleanup(vport);
2958
James Smart87af33f2007-10-27 13:37:43 -04002959 list_for_each_entry_safe(ndlp, next_ndlp, &vport->fc_nodes, nlp_listp) {
James Smart58da1ff2008-04-07 10:15:56 -04002960 if (vport->port_type != LPFC_PHYSICAL_PORT &&
2961 ndlp->nlp_DID == Fabric_DID) {
2962 /* Just free up ndlp with Fabric_DID for vports */
2963 lpfc_nlp_put(ndlp);
2964 continue;
2965 }
2966
James Smarta70e63e2020-11-15 11:26:38 -08002967 if (ndlp->nlp_DID == Fabric_Cntl_DID &&
2968 ndlp->nlp_state == NLP_STE_UNUSED_NODE) {
James Smarteff4a012012-01-18 16:25:25 -05002969 lpfc_nlp_put(ndlp);
2970 continue;
2971 }
2972
James Smarte9b11082020-11-15 11:26:33 -08002973 /* Fabric Ports not in UNMAPPED state are cleaned up in the
2974 * DEVICE_RM event.
2975 */
2976 if (ndlp->nlp_type & NLP_FABRIC &&
2977 ndlp->nlp_state == NLP_STE_UNMAPPED_NODE)
James Smart87af33f2007-10-27 13:37:43 -04002978 lpfc_disc_state_machine(vport, ndlp, NULL,
2979 NLP_EVT_DEVICE_RECOVERY);
James Smarte47c9092008-02-08 18:49:26 -05002980
James Smarte9b11082020-11-15 11:26:33 -08002981 if (!(ndlp->fc4_xpt_flags & (NVME_XPT_REGD|SCSI_XPT_REGD)))
2982 lpfc_disc_state_machine(vport, ndlp, NULL,
2983 NLP_EVT_DEVICE_RM);
James Smart87af33f2007-10-27 13:37:43 -04002984 }
2985
James Smarta8adb832007-10-27 13:37:53 -04002986 /* At this point, ALL ndlp's should be gone
2987 * because of the previous NLP_EVT_DEVICE_RM.
2988 * Lets wait for this to happen, if needed.
2989 */
James Smart87af33f2007-10-27 13:37:43 -04002990 while (!list_empty(&vport->fc_nodes)) {
James Smarta8adb832007-10-27 13:37:53 -04002991 if (i++ > 3000) {
Dick Kennedy372c1872020-06-30 14:50:00 -07002992 lpfc_printf_vlog(vport, KERN_ERR,
2993 LOG_TRACE_EVENT,
James Smarta8adb832007-10-27 13:37:53 -04002994 "0233 Nodelist not empty\n");
James Smarte47c9092008-02-08 18:49:26 -05002995 list_for_each_entry_safe(ndlp, next_ndlp,
2996 &vport->fc_nodes, nlp_listp) {
2997 lpfc_printf_vlog(ndlp->vport, KERN_ERR,
James Smarte9b11082020-11-15 11:26:33 -08002998 LOG_TRACE_EVENT,
2999 "0282 did:x%x ndlp:x%px "
3000 "refcnt:%d xflags x%x nflag x%x\n",
3001 ndlp->nlp_DID, (void *)ndlp,
3002 kref_read(&ndlp->kref),
3003 ndlp->fc4_xpt_flags,
3004 ndlp->nlp_flag);
James Smarte47c9092008-02-08 18:49:26 -05003005 }
James Smarta8adb832007-10-27 13:37:53 -04003006 break;
James Smart87af33f2007-10-27 13:37:43 -04003007 }
James Smarta8adb832007-10-27 13:37:53 -04003008
3009 /* Wait for any activity on ndlps to settle */
3010 msleep(10);
James Smart87af33f2007-10-27 13:37:43 -04003011 }
James Smart1151e3e2011-02-16 12:39:35 -05003012 lpfc_cleanup_vports_rrqs(vport, NULL);
dea31012005-04-17 16:05:31 -05003013}
3014
James Smarte59058c2008-08-24 21:49:00 -04003015/**
James Smart3621a712009-04-06 18:47:14 -04003016 * lpfc_stop_vport_timers - Stop all the timers associated with a vport
James Smarte59058c2008-08-24 21:49:00 -04003017 * @vport: pointer to a virtual N_Port data structure.
3018 *
3019 * This routine stops all the timers associated with a @vport. This function
3020 * is invoked before disabling or deleting a @vport. Note that the physical
3021 * port is treated as @vport 0.
3022 **/
James Smart92d7f7b2007-06-17 19:56:38 -05003023void
3024lpfc_stop_vport_timers(struct lpfc_vport *vport)
dea31012005-04-17 16:05:31 -05003025{
James Smart92d7f7b2007-06-17 19:56:38 -05003026 del_timer_sync(&vport->els_tmofunc);
James Smart92494142011-02-16 12:39:44 -05003027 del_timer_sync(&vport->delayed_disc_tmo);
James Smart92d7f7b2007-06-17 19:56:38 -05003028 lpfc_can_disctmo(vport);
3029 return;
dea31012005-04-17 16:05:31 -05003030}
3031
James Smarte59058c2008-08-24 21:49:00 -04003032/**
James Smartecfd03c2010-02-12 14:41:27 -05003033 * __lpfc_sli4_stop_fcf_redisc_wait_timer - Stop FCF rediscovery wait timer
3034 * @phba: pointer to lpfc hba data structure.
3035 *
3036 * This routine stops the SLI4 FCF rediscover wait timer if it's on. The
3037 * caller of this routine should already hold the host lock.
3038 **/
3039void
3040__lpfc_sli4_stop_fcf_redisc_wait_timer(struct lpfc_hba *phba)
3041{
James Smart5ac6b302010-10-22 11:05:36 -04003042 /* Clear pending FCF rediscovery wait flag */
3043 phba->fcf.fcf_flag &= ~FCF_REDISC_PEND;
3044
James Smartecfd03c2010-02-12 14:41:27 -05003045 /* Now, try to stop the timer */
3046 del_timer(&phba->fcf.redisc_wait);
3047}
3048
3049/**
3050 * lpfc_sli4_stop_fcf_redisc_wait_timer - Stop FCF rediscovery wait timer
3051 * @phba: pointer to lpfc hba data structure.
3052 *
3053 * This routine stops the SLI4 FCF rediscover wait timer if it's on. It
3054 * checks whether the FCF rediscovery wait timer is pending with the host
3055 * lock held before proceeding with disabling the timer and clearing the
3056 * wait timer pendig flag.
3057 **/
3058void
3059lpfc_sli4_stop_fcf_redisc_wait_timer(struct lpfc_hba *phba)
3060{
3061 spin_lock_irq(&phba->hbalock);
3062 if (!(phba->fcf.fcf_flag & FCF_REDISC_PEND)) {
3063 /* FCF rediscovery timer already fired or stopped */
3064 spin_unlock_irq(&phba->hbalock);
3065 return;
3066 }
3067 __lpfc_sli4_stop_fcf_redisc_wait_timer(phba);
James Smart5ac6b302010-10-22 11:05:36 -04003068 /* Clear failover in progress flags */
3069 phba->fcf.fcf_flag &= ~(FCF_DEAD_DISC | FCF_ACVL_DISC);
James Smartecfd03c2010-02-12 14:41:27 -05003070 spin_unlock_irq(&phba->hbalock);
3071}
3072
3073/**
James Smart02243832021-08-16 09:28:54 -07003074 * lpfc_cmf_stop - Stop CMF processing
3075 * @phba: pointer to lpfc hba data structure.
3076 *
3077 * This is called when the link goes down or if CMF mode is turned OFF.
3078 * It is also called when going offline or unloaded just before the
3079 * congestion info buffer is unregistered.
3080 **/
3081void
3082lpfc_cmf_stop(struct lpfc_hba *phba)
3083{
3084 int cpu;
3085 struct lpfc_cgn_stat *cgs;
3086
3087 /* We only do something if CMF is enabled */
3088 if (!phba->sli4_hba.pc_sli4_params.cmf)
3089 return;
3090
3091 lpfc_printf_log(phba, KERN_INFO, LOG_CGN_MGMT,
3092 "6221 Stop CMF / Cancel Timer\n");
3093
3094 /* Cancel the CMF timer */
3095 hrtimer_cancel(&phba->cmf_timer);
3096
3097 /* Zero CMF counters */
3098 atomic_set(&phba->cmf_busy, 0);
3099 for_each_present_cpu(cpu) {
3100 cgs = per_cpu_ptr(phba->cmf_stat, cpu);
3101 atomic64_set(&cgs->total_bytes, 0);
3102 atomic64_set(&cgs->rcv_bytes, 0);
3103 atomic_set(&cgs->rx_io_cnt, 0);
3104 atomic64_set(&cgs->rx_latency, 0);
3105 }
3106 atomic_set(&phba->cmf_bw_wait, 0);
3107
3108 /* Resume any blocked IO - Queue unblock on workqueue */
3109 queue_work(phba->wq, &phba->unblock_request_work);
3110}
3111
3112static inline uint64_t
3113lpfc_get_max_line_rate(struct lpfc_hba *phba)
3114{
3115 uint64_t rate = lpfc_sli_port_speed_get(phba);
3116
3117 return ((((unsigned long)rate) * 1024 * 1024) / 10);
3118}
3119
3120void
3121lpfc_cmf_signal_init(struct lpfc_hba *phba)
3122{
3123 lpfc_printf_log(phba, KERN_INFO, LOG_CGN_MGMT,
3124 "6223 Signal CMF init\n");
3125
3126 /* Use the new fc_linkspeed to recalculate */
3127 phba->cmf_interval_rate = LPFC_CMF_INTERVAL;
3128 phba->cmf_max_line_rate = lpfc_get_max_line_rate(phba);
3129 phba->cmf_link_byte_count = div_u64(phba->cmf_max_line_rate *
3130 phba->cmf_interval_rate, 1000);
3131 phba->cmf_max_bytes_per_interval = phba->cmf_link_byte_count;
3132
3133 /* This is a signal to firmware to sync up CMF BW with link speed */
3134 lpfc_issue_cmf_sync_wqe(phba, 0, 0);
3135}
3136
3137/**
3138 * lpfc_cmf_start - Start CMF processing
3139 * @phba: pointer to lpfc hba data structure.
3140 *
3141 * This is called when the link comes up or if CMF mode is turned OFF
3142 * to Monitor or Managed.
3143 **/
3144void
3145lpfc_cmf_start(struct lpfc_hba *phba)
3146{
3147 struct lpfc_cgn_stat *cgs;
3148 int cpu;
3149
3150 /* We only do something if CMF is enabled */
3151 if (!phba->sli4_hba.pc_sli4_params.cmf ||
3152 phba->cmf_active_mode == LPFC_CFG_OFF)
3153 return;
3154
3155 /* Reinitialize congestion buffer info */
3156 lpfc_init_congestion_buf(phba);
3157
3158 atomic_set(&phba->cgn_fabric_warn_cnt, 0);
3159 atomic_set(&phba->cgn_fabric_alarm_cnt, 0);
3160 atomic_set(&phba->cgn_sync_alarm_cnt, 0);
3161 atomic_set(&phba->cgn_sync_warn_cnt, 0);
3162
3163 atomic_set(&phba->cmf_busy, 0);
3164 for_each_present_cpu(cpu) {
3165 cgs = per_cpu_ptr(phba->cmf_stat, cpu);
3166 atomic64_set(&cgs->total_bytes, 0);
3167 atomic64_set(&cgs->rcv_bytes, 0);
3168 atomic_set(&cgs->rx_io_cnt, 0);
3169 atomic64_set(&cgs->rx_latency, 0);
3170 }
3171 phba->cmf_latency.tv_sec = 0;
3172 phba->cmf_latency.tv_nsec = 0;
3173
3174 lpfc_cmf_signal_init(phba);
3175
3176 lpfc_printf_log(phba, KERN_INFO, LOG_CGN_MGMT,
3177 "6222 Start CMF / Timer\n");
3178
3179 phba->cmf_timer_cnt = 0;
3180 hrtimer_start(&phba->cmf_timer,
3181 ktime_set(0, LPFC_CMF_INTERVAL * 1000000),
3182 HRTIMER_MODE_REL);
3183 /* Setup for latency check in IO cmpl routines */
3184 ktime_get_real_ts64(&phba->cmf_latency);
3185
3186 atomic_set(&phba->cmf_bw_wait, 0);
3187 atomic_set(&phba->cmf_stop_io, 0);
3188}
3189
3190/**
James Smart3772a992009-05-22 14:50:54 -04003191 * lpfc_stop_hba_timers - Stop all the timers associated with an HBA
James Smarte59058c2008-08-24 21:49:00 -04003192 * @phba: pointer to lpfc hba data structure.
3193 *
3194 * This routine stops all the timers associated with a HBA. This function is
3195 * invoked before either putting a HBA offline or unloading the driver.
3196 **/
James Smart3772a992009-05-22 14:50:54 -04003197void
3198lpfc_stop_hba_timers(struct lpfc_hba *phba)
dea31012005-04-17 16:05:31 -05003199{
James Smartcdb42be2019-01-28 11:14:21 -08003200 if (phba->pport)
3201 lpfc_stop_vport_timers(phba->pport);
James Smart32517fc2019-01-28 11:14:33 -08003202 cancel_delayed_work_sync(&phba->eq_delay_work);
Dick Kennedy317aeb82020-06-30 14:49:59 -07003203 cancel_delayed_work_sync(&phba->idle_stat_delay_work);
James Smart2e0fef82007-06-17 19:56:36 -05003204 del_timer_sync(&phba->sli.mbox_tmo);
James Smart92d7f7b2007-06-17 19:56:38 -05003205 del_timer_sync(&phba->fabric_block_timer);
James Smart93996272008-08-24 21:50:30 -04003206 del_timer_sync(&phba->eratt_poll);
James Smart3772a992009-05-22 14:50:54 -04003207 del_timer_sync(&phba->hb_tmofunc);
James Smart1151e3e2011-02-16 12:39:35 -05003208 if (phba->sli_rev == LPFC_SLI_REV4) {
3209 del_timer_sync(&phba->rrq_tmr);
3210 phba->hba_flag &= ~HBA_RRQ_ACTIVE;
3211 }
James Smarta22d73b2021-01-04 10:02:38 -08003212 phba->hba_flag &= ~(HBA_HBEAT_INP | HBA_HBEAT_TMO);
James Smart3772a992009-05-22 14:50:54 -04003213
3214 switch (phba->pci_dev_grp) {
3215 case LPFC_PCI_DEV_LP:
3216 /* Stop any LightPulse device specific driver timers */
3217 del_timer_sync(&phba->fcp_poll_timer);
3218 break;
3219 case LPFC_PCI_DEV_OC:
Paul Walmsleycc0e5f12019-07-11 20:52:33 -07003220 /* Stop any OneConnect device specific driver timers */
James Smartecfd03c2010-02-12 14:41:27 -05003221 lpfc_sli4_stop_fcf_redisc_wait_timer(phba);
James Smart3772a992009-05-22 14:50:54 -04003222 break;
3223 default:
Dick Kennedy372c1872020-06-30 14:50:00 -07003224 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart3772a992009-05-22 14:50:54 -04003225 "0297 Invalid device group (x%x)\n",
3226 phba->pci_dev_grp);
3227 break;
3228 }
James Smart2e0fef82007-06-17 19:56:36 -05003229 return;
dea31012005-04-17 16:05:31 -05003230}
3231
James Smarte59058c2008-08-24 21:49:00 -04003232/**
James Smart3621a712009-04-06 18:47:14 -04003233 * lpfc_block_mgmt_io - Mark a HBA's management interface as blocked
James Smarte59058c2008-08-24 21:49:00 -04003234 * @phba: pointer to lpfc hba data structure.
Lee Jonesfe614ac2020-07-23 13:24:22 +01003235 * @mbx_action: flag for mailbox no wait action.
James Smarte59058c2008-08-24 21:49:00 -04003236 *
3237 * This routine marks a HBA's management interface as blocked. Once the HBA's
3238 * management interface is marked as blocked, all the user space access to
3239 * the HBA, whether they are from sysfs interface or libdfc interface will
3240 * all be blocked. The HBA is set to block the management interface when the
3241 * driver prepares the HBA interface for online or offline.
3242 **/
Adrian Bunka6ababd2007-11-05 18:07:33 +01003243static void
James Smart618a5232012-06-12 13:54:36 -04003244lpfc_block_mgmt_io(struct lpfc_hba *phba, int mbx_action)
Adrian Bunka6ababd2007-11-05 18:07:33 +01003245{
3246 unsigned long iflag;
James Smart6e7288d2010-06-07 15:23:35 -04003247 uint8_t actcmd = MBX_HEARTBEAT;
3248 unsigned long timeout;
3249
Adrian Bunka6ababd2007-11-05 18:07:33 +01003250 spin_lock_irqsave(&phba->hbalock, iflag);
3251 phba->sli.sli_flag |= LPFC_BLOCK_MGMT_IO;
James Smart618a5232012-06-12 13:54:36 -04003252 spin_unlock_irqrestore(&phba->hbalock, iflag);
3253 if (mbx_action == LPFC_MBX_NO_WAIT)
3254 return;
3255 timeout = msecs_to_jiffies(LPFC_MBOX_TMO * 1000) + jiffies;
3256 spin_lock_irqsave(&phba->hbalock, iflag);
James Smarta183a152011-10-10 21:32:43 -04003257 if (phba->sli.mbox_active) {
James Smart6e7288d2010-06-07 15:23:35 -04003258 actcmd = phba->sli.mbox_active->u.mb.mbxCommand;
James Smarta183a152011-10-10 21:32:43 -04003259 /* Determine how long we might wait for the active mailbox
3260 * command to be gracefully completed by firmware.
3261 */
3262 timeout = msecs_to_jiffies(lpfc_mbox_tmo_val(phba,
3263 phba->sli.mbox_active) * 1000) + jiffies;
3264 }
Adrian Bunka6ababd2007-11-05 18:07:33 +01003265 spin_unlock_irqrestore(&phba->hbalock, iflag);
James Smarta183a152011-10-10 21:32:43 -04003266
James Smart6e7288d2010-06-07 15:23:35 -04003267 /* Wait for the outstnading mailbox command to complete */
3268 while (phba->sli.mbox_active) {
3269 /* Check active mailbox complete status every 2ms */
3270 msleep(2);
3271 if (time_after(jiffies, timeout)) {
Dick Kennedy372c1872020-06-30 14:50:00 -07003272 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
3273 "2813 Mgmt IO is Blocked %x "
3274 "- mbox cmd %x still active\n",
3275 phba->sli.sli_flag, actcmd);
James Smart6e7288d2010-06-07 15:23:35 -04003276 break;
3277 }
3278 }
Adrian Bunka6ababd2007-11-05 18:07:33 +01003279}
3280
James Smarte59058c2008-08-24 21:49:00 -04003281/**
James Smart6b5151f2012-01-18 16:24:06 -05003282 * lpfc_sli4_node_prep - Assign RPIs for active nodes.
3283 * @phba: pointer to lpfc hba data structure.
3284 *
3285 * Allocate RPIs for all active remote nodes. This is needed whenever
3286 * an SLI4 adapter is reset and the driver is not unloading. Its purpose
3287 * is to fixup the temporary rpi assignments.
3288 **/
3289void
3290lpfc_sli4_node_prep(struct lpfc_hba *phba)
3291{
3292 struct lpfc_nodelist *ndlp, *next_ndlp;
3293 struct lpfc_vport **vports;
James Smart9d3d3402017-04-21 16:05:00 -07003294 int i, rpi;
James Smart6b5151f2012-01-18 16:24:06 -05003295
3296 if (phba->sli_rev != LPFC_SLI_REV4)
3297 return;
3298
3299 vports = lpfc_create_vport_work_array(phba);
James Smart9d3d3402017-04-21 16:05:00 -07003300 if (vports == NULL)
3301 return;
James Smart6b5151f2012-01-18 16:24:06 -05003302
James Smart9d3d3402017-04-21 16:05:00 -07003303 for (i = 0; i <= phba->max_vports && vports[i] != NULL; i++) {
3304 if (vports[i]->load_flag & FC_UNLOADING)
3305 continue;
3306
3307 list_for_each_entry_safe(ndlp, next_ndlp,
3308 &vports[i]->fc_nodes,
3309 nlp_listp) {
James Smart9d3d3402017-04-21 16:05:00 -07003310 rpi = lpfc_sli4_alloc_rpi(phba);
3311 if (rpi == LPFC_RPI_ALLOC_ERROR) {
James Smart307e3382020-11-15 11:26:30 -08003312 /* TODO print log? */
James Smart9d3d3402017-04-21 16:05:00 -07003313 continue;
James Smart6b5151f2012-01-18 16:24:06 -05003314 }
James Smart9d3d3402017-04-21 16:05:00 -07003315 ndlp->nlp_rpi = rpi;
James Smart0f154222019-09-21 20:58:52 -07003316 lpfc_printf_vlog(ndlp->vport, KERN_INFO,
3317 LOG_NODE | LOG_DISCOVERY,
3318 "0009 Assign RPI x%x to ndlp x%px "
James Smart307e3382020-11-15 11:26:30 -08003319 "DID:x%06x flg:x%x\n",
James Smart0f154222019-09-21 20:58:52 -07003320 ndlp->nlp_rpi, ndlp, ndlp->nlp_DID,
James Smart307e3382020-11-15 11:26:30 -08003321 ndlp->nlp_flag);
James Smart6b5151f2012-01-18 16:24:06 -05003322 }
3323 }
3324 lpfc_destroy_vport_work_array(phba, vports);
3325}
3326
3327/**
James Smartc4908502019-01-28 11:14:28 -08003328 * lpfc_create_expedite_pool - create expedite pool
3329 * @phba: pointer to lpfc hba data structure.
3330 *
3331 * This routine moves a batch of XRIs from lpfc_io_buf_list_put of HWQ 0
3332 * to expedite pool. Mark them as expedite.
3333 **/
Bart Van Assche3999df72019-03-28 11:06:16 -07003334static void lpfc_create_expedite_pool(struct lpfc_hba *phba)
James Smartc4908502019-01-28 11:14:28 -08003335{
3336 struct lpfc_sli4_hdw_queue *qp;
3337 struct lpfc_io_buf *lpfc_ncmd;
3338 struct lpfc_io_buf *lpfc_ncmd_next;
3339 struct lpfc_epd_pool *epd_pool;
3340 unsigned long iflag;
3341
3342 epd_pool = &phba->epd_pool;
3343 qp = &phba->sli4_hba.hdwq[0];
3344
3345 spin_lock_init(&epd_pool->lock);
3346 spin_lock_irqsave(&qp->io_buf_list_put_lock, iflag);
3347 spin_lock(&epd_pool->lock);
3348 INIT_LIST_HEAD(&epd_pool->list);
3349 list_for_each_entry_safe(lpfc_ncmd, lpfc_ncmd_next,
3350 &qp->lpfc_io_buf_list_put, list) {
3351 list_move_tail(&lpfc_ncmd->list, &epd_pool->list);
3352 lpfc_ncmd->expedite = true;
3353 qp->put_io_bufs--;
3354 epd_pool->count++;
3355 if (epd_pool->count >= XRI_BATCH)
3356 break;
3357 }
3358 spin_unlock(&epd_pool->lock);
3359 spin_unlock_irqrestore(&qp->io_buf_list_put_lock, iflag);
3360}
3361
3362/**
3363 * lpfc_destroy_expedite_pool - destroy expedite pool
3364 * @phba: pointer to lpfc hba data structure.
3365 *
3366 * This routine returns XRIs from expedite pool to lpfc_io_buf_list_put
3367 * of HWQ 0. Clear the mark.
3368 **/
Bart Van Assche3999df72019-03-28 11:06:16 -07003369static void lpfc_destroy_expedite_pool(struct lpfc_hba *phba)
James Smartc4908502019-01-28 11:14:28 -08003370{
3371 struct lpfc_sli4_hdw_queue *qp;
3372 struct lpfc_io_buf *lpfc_ncmd;
3373 struct lpfc_io_buf *lpfc_ncmd_next;
3374 struct lpfc_epd_pool *epd_pool;
3375 unsigned long iflag;
3376
3377 epd_pool = &phba->epd_pool;
3378 qp = &phba->sli4_hba.hdwq[0];
3379
3380 spin_lock_irqsave(&qp->io_buf_list_put_lock, iflag);
3381 spin_lock(&epd_pool->lock);
3382 list_for_each_entry_safe(lpfc_ncmd, lpfc_ncmd_next,
3383 &epd_pool->list, list) {
3384 list_move_tail(&lpfc_ncmd->list,
3385 &qp->lpfc_io_buf_list_put);
3386 lpfc_ncmd->flags = false;
3387 qp->put_io_bufs++;
3388 epd_pool->count--;
3389 }
3390 spin_unlock(&epd_pool->lock);
3391 spin_unlock_irqrestore(&qp->io_buf_list_put_lock, iflag);
3392}
3393
3394/**
3395 * lpfc_create_multixri_pools - create multi-XRI pools
3396 * @phba: pointer to lpfc hba data structure.
3397 *
3398 * This routine initialize public, private per HWQ. Then, move XRIs from
3399 * lpfc_io_buf_list_put to public pool. High and low watermark are also
3400 * Initialized.
3401 **/
3402void lpfc_create_multixri_pools(struct lpfc_hba *phba)
3403{
3404 u32 i, j;
3405 u32 hwq_count;
3406 u32 count_per_hwq;
3407 struct lpfc_io_buf *lpfc_ncmd;
3408 struct lpfc_io_buf *lpfc_ncmd_next;
3409 unsigned long iflag;
3410 struct lpfc_sli4_hdw_queue *qp;
3411 struct lpfc_multixri_pool *multixri_pool;
3412 struct lpfc_pbl_pool *pbl_pool;
3413 struct lpfc_pvt_pool *pvt_pool;
3414
3415 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
3416 "1234 num_hdw_queue=%d num_present_cpu=%d common_xri_cnt=%d\n",
3417 phba->cfg_hdw_queue, phba->sli4_hba.num_present_cpu,
3418 phba->sli4_hba.io_xri_cnt);
3419
3420 if (phba->cfg_enable_fc4_type & LPFC_ENABLE_NVME)
3421 lpfc_create_expedite_pool(phba);
3422
3423 hwq_count = phba->cfg_hdw_queue;
3424 count_per_hwq = phba->sli4_hba.io_xri_cnt / hwq_count;
3425
3426 for (i = 0; i < hwq_count; i++) {
3427 multixri_pool = kzalloc(sizeof(*multixri_pool), GFP_KERNEL);
3428
3429 if (!multixri_pool) {
3430 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
3431 "1238 Failed to allocate memory for "
3432 "multixri_pool\n");
3433
3434 if (phba->cfg_enable_fc4_type & LPFC_ENABLE_NVME)
3435 lpfc_destroy_expedite_pool(phba);
3436
3437 j = 0;
3438 while (j < i) {
3439 qp = &phba->sli4_hba.hdwq[j];
3440 kfree(qp->p_multixri_pool);
3441 j++;
3442 }
3443 phba->cfg_xri_rebalancing = 0;
3444 return;
3445 }
3446
3447 qp = &phba->sli4_hba.hdwq[i];
3448 qp->p_multixri_pool = multixri_pool;
3449
3450 multixri_pool->xri_limit = count_per_hwq;
3451 multixri_pool->rrb_next_hwqid = i;
3452
3453 /* Deal with public free xri pool */
3454 pbl_pool = &multixri_pool->pbl_pool;
3455 spin_lock_init(&pbl_pool->lock);
3456 spin_lock_irqsave(&qp->io_buf_list_put_lock, iflag);
3457 spin_lock(&pbl_pool->lock);
3458 INIT_LIST_HEAD(&pbl_pool->list);
3459 list_for_each_entry_safe(lpfc_ncmd, lpfc_ncmd_next,
3460 &qp->lpfc_io_buf_list_put, list) {
3461 list_move_tail(&lpfc_ncmd->list, &pbl_pool->list);
3462 qp->put_io_bufs--;
3463 pbl_pool->count++;
3464 }
3465 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
3466 "1235 Moved %d buffers from PUT list over to pbl_pool[%d]\n",
3467 pbl_pool->count, i);
3468 spin_unlock(&pbl_pool->lock);
3469 spin_unlock_irqrestore(&qp->io_buf_list_put_lock, iflag);
3470
3471 /* Deal with private free xri pool */
3472 pvt_pool = &multixri_pool->pvt_pool;
3473 pvt_pool->high_watermark = multixri_pool->xri_limit / 2;
3474 pvt_pool->low_watermark = XRI_BATCH;
3475 spin_lock_init(&pvt_pool->lock);
3476 spin_lock_irqsave(&pvt_pool->lock, iflag);
3477 INIT_LIST_HEAD(&pvt_pool->list);
3478 pvt_pool->count = 0;
3479 spin_unlock_irqrestore(&pvt_pool->lock, iflag);
3480 }
3481}
3482
3483/**
3484 * lpfc_destroy_multixri_pools - destroy multi-XRI pools
3485 * @phba: pointer to lpfc hba data structure.
3486 *
3487 * This routine returns XRIs from public/private to lpfc_io_buf_list_put.
3488 **/
Bart Van Assche3999df72019-03-28 11:06:16 -07003489static void lpfc_destroy_multixri_pools(struct lpfc_hba *phba)
James Smartc4908502019-01-28 11:14:28 -08003490{
3491 u32 i;
3492 u32 hwq_count;
3493 struct lpfc_io_buf *lpfc_ncmd;
3494 struct lpfc_io_buf *lpfc_ncmd_next;
3495 unsigned long iflag;
3496 struct lpfc_sli4_hdw_queue *qp;
3497 struct lpfc_multixri_pool *multixri_pool;
3498 struct lpfc_pbl_pool *pbl_pool;
3499 struct lpfc_pvt_pool *pvt_pool;
3500
3501 if (phba->cfg_enable_fc4_type & LPFC_ENABLE_NVME)
3502 lpfc_destroy_expedite_pool(phba);
3503
James Smartc00f62e2019-08-14 16:57:11 -07003504 if (!(phba->pport->load_flag & FC_UNLOADING))
3505 lpfc_sli_flush_io_rings(phba);
James Smartc66a9192019-03-12 16:30:19 -07003506
James Smartc4908502019-01-28 11:14:28 -08003507 hwq_count = phba->cfg_hdw_queue;
3508
3509 for (i = 0; i < hwq_count; i++) {
3510 qp = &phba->sli4_hba.hdwq[i];
3511 multixri_pool = qp->p_multixri_pool;
3512 if (!multixri_pool)
3513 continue;
3514
3515 qp->p_multixri_pool = NULL;
3516
3517 spin_lock_irqsave(&qp->io_buf_list_put_lock, iflag);
3518
3519 /* Deal with public free xri pool */
3520 pbl_pool = &multixri_pool->pbl_pool;
3521 spin_lock(&pbl_pool->lock);
3522
3523 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
3524 "1236 Moving %d buffers from pbl_pool[%d] TO PUT list\n",
3525 pbl_pool->count, i);
3526
3527 list_for_each_entry_safe(lpfc_ncmd, lpfc_ncmd_next,
3528 &pbl_pool->list, list) {
3529 list_move_tail(&lpfc_ncmd->list,
3530 &qp->lpfc_io_buf_list_put);
3531 qp->put_io_bufs++;
3532 pbl_pool->count--;
3533 }
3534
3535 INIT_LIST_HEAD(&pbl_pool->list);
3536 pbl_pool->count = 0;
3537
3538 spin_unlock(&pbl_pool->lock);
3539
3540 /* Deal with private free xri pool */
3541 pvt_pool = &multixri_pool->pvt_pool;
3542 spin_lock(&pvt_pool->lock);
3543
3544 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
3545 "1237 Moving %d buffers from pvt_pool[%d] TO PUT list\n",
3546 pvt_pool->count, i);
3547
3548 list_for_each_entry_safe(lpfc_ncmd, lpfc_ncmd_next,
3549 &pvt_pool->list, list) {
3550 list_move_tail(&lpfc_ncmd->list,
3551 &qp->lpfc_io_buf_list_put);
3552 qp->put_io_bufs++;
3553 pvt_pool->count--;
3554 }
3555
3556 INIT_LIST_HEAD(&pvt_pool->list);
3557 pvt_pool->count = 0;
3558
3559 spin_unlock(&pvt_pool->lock);
3560 spin_unlock_irqrestore(&qp->io_buf_list_put_lock, iflag);
3561
3562 kfree(multixri_pool);
3563 }
3564}
3565
3566/**
James Smart3621a712009-04-06 18:47:14 -04003567 * lpfc_online - Initialize and bring a HBA online
James Smarte59058c2008-08-24 21:49:00 -04003568 * @phba: pointer to lpfc hba data structure.
3569 *
3570 * This routine initializes the HBA and brings a HBA online. During this
3571 * process, the management interface is blocked to prevent user space access
3572 * to the HBA interfering with the driver initialization.
3573 *
3574 * Return codes
3575 * 0 - successful
3576 * 1 - failed
3577 **/
dea31012005-04-17 16:05:31 -05003578int
James Smart2e0fef82007-06-17 19:56:36 -05003579lpfc_online(struct lpfc_hba *phba)
dea31012005-04-17 16:05:31 -05003580{
Julia Lawall372bd282008-12-16 16:15:08 +01003581 struct lpfc_vport *vport;
James Smart549e55c2007-08-02 11:09:51 -04003582 struct lpfc_vport **vports;
Dick Kennedya145fda2017-08-23 16:55:44 -07003583 int i, error = 0;
James Smart16a3a202013-04-17 20:14:38 -04003584 bool vpis_cleared = false;
James Smart2e0fef82007-06-17 19:56:36 -05003585
dea31012005-04-17 16:05:31 -05003586 if (!phba)
3587 return 0;
Julia Lawall372bd282008-12-16 16:15:08 +01003588 vport = phba->pport;
dea31012005-04-17 16:05:31 -05003589
James Smart2e0fef82007-06-17 19:56:36 -05003590 if (!(vport->fc_flag & FC_OFFLINE_MODE))
dea31012005-04-17 16:05:31 -05003591 return 0;
3592
James Smarted957682007-06-17 19:56:37 -05003593 lpfc_printf_log(phba, KERN_WARNING, LOG_INIT,
James Smarte8b62012007-08-02 11:10:09 -04003594 "0458 Bring Adapter online\n");
dea31012005-04-17 16:05:31 -05003595
James Smart618a5232012-06-12 13:54:36 -04003596 lpfc_block_mgmt_io(phba, LPFC_MBX_WAIT);
dea31012005-04-17 16:05:31 -05003597
James Smartda0436e2009-05-22 14:51:39 -04003598 if (phba->sli_rev == LPFC_SLI_REV4) {
3599 if (lpfc_sli4_hba_setup(phba)) { /* Initialize SLI4 HBA */
3600 lpfc_unblock_mgmt_io(phba);
3601 return 1;
3602 }
James Smart16a3a202013-04-17 20:14:38 -04003603 spin_lock_irq(&phba->hbalock);
3604 if (!phba->sli4_hba.max_cfg_param.vpi_used)
3605 vpis_cleared = true;
3606 spin_unlock_irq(&phba->hbalock);
Dick Kennedya145fda2017-08-23 16:55:44 -07003607
3608 /* Reestablish the local initiator port.
3609 * The offline process destroyed the previous lport.
3610 */
3611 if (phba->cfg_enable_fc4_type & LPFC_ENABLE_NVME &&
3612 !phba->nvmet_support) {
3613 error = lpfc_nvme_create_localport(phba->pport);
3614 if (error)
Dick Kennedy372c1872020-06-30 14:50:00 -07003615 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
Dick Kennedya145fda2017-08-23 16:55:44 -07003616 "6132 NVME restore reg failed "
3617 "on nvmei error x%x\n", error);
3618 }
James Smartda0436e2009-05-22 14:51:39 -04003619 } else {
James Smart895427b2017-02-12 13:52:30 -08003620 lpfc_sli_queue_init(phba);
James Smartda0436e2009-05-22 14:51:39 -04003621 if (lpfc_sli_hba_setup(phba)) { /* Initialize SLI2/SLI3 HBA */
3622 lpfc_unblock_mgmt_io(phba);
3623 return 1;
3624 }
James Smart46fa3112007-04-25 09:51:45 -04003625 }
dea31012005-04-17 16:05:31 -05003626
James Smart549e55c2007-08-02 11:09:51 -04003627 vports = lpfc_create_vport_work_array(phba);
Arnd Bergmannaeb66412016-03-14 15:29:44 +01003628 if (vports != NULL) {
James Smartda0436e2009-05-22 14:51:39 -04003629 for (i = 0; i <= phba->max_vports && vports[i] != NULL; i++) {
James Smart549e55c2007-08-02 11:09:51 -04003630 struct Scsi_Host *shost;
3631 shost = lpfc_shost_from_vport(vports[i]);
3632 spin_lock_irq(shost->host_lock);
3633 vports[i]->fc_flag &= ~FC_OFFLINE_MODE;
3634 if (phba->sli3_options & LPFC_SLI3_NPIV_ENABLED)
3635 vports[i]->fc_flag |= FC_VPORT_NEEDS_REG_VPI;
James Smart16a3a202013-04-17 20:14:38 -04003636 if (phba->sli_rev == LPFC_SLI_REV4) {
James Smart1c6834a2009-07-19 10:01:26 -04003637 vports[i]->fc_flag |= FC_VPORT_NEEDS_INIT_VPI;
James Smart16a3a202013-04-17 20:14:38 -04003638 if ((vpis_cleared) &&
3639 (vports[i]->port_type !=
3640 LPFC_PHYSICAL_PORT))
3641 vports[i]->vpi = 0;
3642 }
James Smart549e55c2007-08-02 11:09:51 -04003643 spin_unlock_irq(shost->host_lock);
3644 }
Arnd Bergmannaeb66412016-03-14 15:29:44 +01003645 }
3646 lpfc_destroy_vport_work_array(phba, vports);
dea31012005-04-17 16:05:31 -05003647
James Smartc4908502019-01-28 11:14:28 -08003648 if (phba->cfg_xri_rebalancing)
3649 lpfc_create_multixri_pools(phba);
3650
James Smart93a4d6f2019-11-04 16:57:05 -08003651 lpfc_cpuhp_add(phba);
3652
James Smart46fa3112007-04-25 09:51:45 -04003653 lpfc_unblock_mgmt_io(phba);
dea31012005-04-17 16:05:31 -05003654 return 0;
3655}
3656
James Smarte59058c2008-08-24 21:49:00 -04003657/**
James Smart3621a712009-04-06 18:47:14 -04003658 * lpfc_unblock_mgmt_io - Mark a HBA's management interface to be not blocked
James Smarte59058c2008-08-24 21:49:00 -04003659 * @phba: pointer to lpfc hba data structure.
3660 *
3661 * This routine marks a HBA's management interface as not blocked. Once the
3662 * HBA's management interface is marked as not blocked, all the user space
3663 * access to the HBA, whether they are from sysfs interface or libdfc
3664 * interface will be allowed. The HBA is set to block the management interface
3665 * when the driver prepares the HBA interface for online or offline and then
3666 * set to unblock the management interface afterwards.
3667 **/
James Smart46fa3112007-04-25 09:51:45 -04003668void
James Smart46fa3112007-04-25 09:51:45 -04003669lpfc_unblock_mgmt_io(struct lpfc_hba * phba)
3670{
3671 unsigned long iflag;
3672
James Smart2e0fef82007-06-17 19:56:36 -05003673 spin_lock_irqsave(&phba->hbalock, iflag);
3674 phba->sli.sli_flag &= ~LPFC_BLOCK_MGMT_IO;
3675 spin_unlock_irqrestore(&phba->hbalock, iflag);
James Smart46fa3112007-04-25 09:51:45 -04003676}
3677
James Smarte59058c2008-08-24 21:49:00 -04003678/**
James Smart3621a712009-04-06 18:47:14 -04003679 * lpfc_offline_prep - Prepare a HBA to be brought offline
James Smarte59058c2008-08-24 21:49:00 -04003680 * @phba: pointer to lpfc hba data structure.
Lee Jonesfe614ac2020-07-23 13:24:22 +01003681 * @mbx_action: flag for mailbox shutdown action.
James Smarte59058c2008-08-24 21:49:00 -04003682 *
3683 * This routine is invoked to prepare a HBA to be brought offline. It performs
3684 * unregistration login to all the nodes on all vports and flushes the mailbox
3685 * queue to make it ready to be brought offline.
3686 **/
James Smart46fa3112007-04-25 09:51:45 -04003687void
James Smart618a5232012-06-12 13:54:36 -04003688lpfc_offline_prep(struct lpfc_hba *phba, int mbx_action)
James Smart46fa3112007-04-25 09:51:45 -04003689{
James Smart2e0fef82007-06-17 19:56:36 -05003690 struct lpfc_vport *vport = phba->pport;
James Smart46fa3112007-04-25 09:51:45 -04003691 struct lpfc_nodelist *ndlp, *next_ndlp;
James Smart87af33f2007-10-27 13:37:43 -04003692 struct lpfc_vport **vports;
James Smart72100cc2010-02-12 14:43:01 -05003693 struct Scsi_Host *shost;
James Smart87af33f2007-10-27 13:37:43 -04003694 int i;
James Smart25ac2c972021-09-10 16:31:54 -07003695 int offline = 0;
dea31012005-04-17 16:05:31 -05003696
James Smart2e0fef82007-06-17 19:56:36 -05003697 if (vport->fc_flag & FC_OFFLINE_MODE)
James Smart46fa3112007-04-25 09:51:45 -04003698 return;
dea31012005-04-17 16:05:31 -05003699
James Smart618a5232012-06-12 13:54:36 -04003700 lpfc_block_mgmt_io(phba, mbx_action);
dea31012005-04-17 16:05:31 -05003701
3702 lpfc_linkdown(phba);
3703
James Smart25ac2c972021-09-10 16:31:54 -07003704 offline = pci_channel_offline(phba->pcidev);
3705
James Smart87af33f2007-10-27 13:37:43 -04003706 /* Issue an unreg_login to all nodes on all vports */
3707 vports = lpfc_create_vport_work_array(phba);
3708 if (vports != NULL) {
James Smartda0436e2009-05-22 14:51:39 -04003709 for (i = 0; i <= phba->max_vports && vports[i] != NULL; i++) {
James Smarta8adb832007-10-27 13:37:53 -04003710 if (vports[i]->load_flag & FC_UNLOADING)
3711 continue;
James Smart72100cc2010-02-12 14:43:01 -05003712 shost = lpfc_shost_from_vport(vports[i]);
3713 spin_lock_irq(shost->host_lock);
James Smartc8685952009-11-18 15:39:16 -05003714 vports[i]->vpi_state &= ~LPFC_VPI_REGISTERED;
James Smart695a8142010-01-26 23:08:03 -05003715 vports[i]->fc_flag |= FC_VPORT_NEEDS_REG_VPI;
3716 vports[i]->fc_flag &= ~FC_VFI_REGISTERED;
James Smart72100cc2010-02-12 14:43:01 -05003717 spin_unlock_irq(shost->host_lock);
James Smart695a8142010-01-26 23:08:03 -05003718
James Smart87af33f2007-10-27 13:37:43 -04003719 shost = lpfc_shost_from_vport(vports[i]);
3720 list_for_each_entry_safe(ndlp, next_ndlp,
3721 &vports[i]->fc_nodes,
3722 nlp_listp) {
James Smart0f154222019-09-21 20:58:52 -07003723
James Smartc6adba12020-11-15 11:26:34 -08003724 spin_lock_irq(&ndlp->lock);
James Smart87af33f2007-10-27 13:37:43 -04003725 ndlp->nlp_flag &= ~NLP_NPR_ADISC;
James Smartc6adba12020-11-15 11:26:34 -08003726 spin_unlock_irq(&ndlp->lock);
James Smartaffbe242021-07-07 11:43:42 -07003727
James Smart25ac2c972021-09-10 16:31:54 -07003728 if (offline) {
3729 spin_lock_irq(&ndlp->lock);
3730 ndlp->nlp_flag &= ~(NLP_UNREG_INP |
3731 NLP_RPI_REGISTERED);
3732 spin_unlock_irq(&ndlp->lock);
3733 } else {
3734 lpfc_unreg_rpi(vports[i], ndlp);
3735 }
James Smart6b5151f2012-01-18 16:24:06 -05003736 /*
3737 * Whenever an SLI4 port goes offline, free the
James Smart401ee0c2012-03-01 22:35:34 -05003738 * RPI. Get a new RPI when the adapter port
3739 * comes back online.
James Smart6b5151f2012-01-18 16:24:06 -05003740 */
James Smartbe6bb942015-04-07 15:07:22 -04003741 if (phba->sli_rev == LPFC_SLI_REV4) {
James Smarte9b11082020-11-15 11:26:33 -08003742 lpfc_printf_vlog(vports[i], KERN_INFO,
James Smart0f154222019-09-21 20:58:52 -07003743 LOG_NODE | LOG_DISCOVERY,
3744 "0011 Free RPI x%x on "
James Smartf1156122021-04-11 18:31:24 -07003745 "ndlp: x%px did x%x\n",
James Smart0f154222019-09-21 20:58:52 -07003746 ndlp->nlp_rpi, ndlp,
James Smart307e3382020-11-15 11:26:30 -08003747 ndlp->nlp_DID);
James Smart6b5151f2012-01-18 16:24:06 -05003748 lpfc_sli4_free_rpi(phba, ndlp->nlp_rpi);
James Smart0f154222019-09-21 20:58:52 -07003749 ndlp->nlp_rpi = LPFC_RPI_ALLOC_ERROR;
James Smartbe6bb942015-04-07 15:07:22 -04003750 }
James Smart307e3382020-11-15 11:26:30 -08003751
3752 if (ndlp->nlp_type & NLP_FABRIC) {
3753 lpfc_disc_state_machine(vports[i], ndlp,
3754 NULL, NLP_EVT_DEVICE_RECOVERY);
James Smarte9b11082020-11-15 11:26:33 -08003755
James Smartaf984c82021-10-20 14:14:16 -07003756 /* Don't remove the node unless the node
James Smarte9b11082020-11-15 11:26:33 -08003757 * has been unregistered with the
James Smartaf984c82021-10-20 14:14:16 -07003758 * transport, and we're not in recovery
3759 * before dev_loss_tmo triggered.
3760 * Otherwise, let dev_loss take care of
3761 * the node.
James Smarte9b11082020-11-15 11:26:33 -08003762 */
James Smartaf984c82021-10-20 14:14:16 -07003763 if (!(ndlp->save_flags &
3764 NLP_IN_RECOV_POST_DEV_LOSS) &&
3765 !(ndlp->fc4_xpt_flags &
James Smarte9b11082020-11-15 11:26:33 -08003766 (NVME_XPT_REGD | SCSI_XPT_REGD)))
3767 lpfc_disc_state_machine
3768 (vports[i], ndlp,
3769 NULL,
3770 NLP_EVT_DEVICE_RM);
James Smart307e3382020-11-15 11:26:30 -08003771 }
James Smart87af33f2007-10-27 13:37:43 -04003772 }
3773 }
3774 }
James Smart09372822008-01-11 01:52:54 -05003775 lpfc_destroy_vport_work_array(phba, vports);
dea31012005-04-17 16:05:31 -05003776
James Smart618a5232012-06-12 13:54:36 -04003777 lpfc_sli_mbox_sys_shutdown(phba, mbx_action);
Dick Kennedyf485c182017-09-29 17:34:34 -07003778
3779 if (phba->wq)
3780 flush_workqueue(phba->wq);
James Smart46fa3112007-04-25 09:51:45 -04003781}
3782
James Smarte59058c2008-08-24 21:49:00 -04003783/**
James Smart3621a712009-04-06 18:47:14 -04003784 * lpfc_offline - Bring a HBA offline
James Smarte59058c2008-08-24 21:49:00 -04003785 * @phba: pointer to lpfc hba data structure.
3786 *
3787 * This routine actually brings a HBA offline. It stops all the timers
3788 * associated with the HBA, brings down the SLI layer, and eventually
3789 * marks the HBA as in offline state for the upper layer protocol.
3790 **/
James Smart46fa3112007-04-25 09:51:45 -04003791void
James Smart2e0fef82007-06-17 19:56:36 -05003792lpfc_offline(struct lpfc_hba *phba)
James Smart46fa3112007-04-25 09:51:45 -04003793{
James Smart549e55c2007-08-02 11:09:51 -04003794 struct Scsi_Host *shost;
3795 struct lpfc_vport **vports;
3796 int i;
James Smart46fa3112007-04-25 09:51:45 -04003797
James Smart549e55c2007-08-02 11:09:51 -04003798 if (phba->pport->fc_flag & FC_OFFLINE_MODE)
James Smart46fa3112007-04-25 09:51:45 -04003799 return;
James Smart688a8862006-07-06 15:49:56 -04003800
James Smartda0436e2009-05-22 14:51:39 -04003801 /* stop port and all timers associated with this hba */
3802 lpfc_stop_port(phba);
Dick Kennedy4b40d022017-08-23 16:55:38 -07003803
3804 /* Tear down the local and target port registrations. The
3805 * nvme transports need to cleanup.
3806 */
3807 lpfc_nvmet_destroy_targetport(phba);
3808 lpfc_nvme_destroy_localport(phba->pport);
3809
James Smart51ef4c22007-08-02 11:10:31 -04003810 vports = lpfc_create_vport_work_array(phba);
3811 if (vports != NULL)
James Smartda0436e2009-05-22 14:51:39 -04003812 for (i = 0; i <= phba->max_vports && vports[i] != NULL; i++)
James Smart51ef4c22007-08-02 11:10:31 -04003813 lpfc_stop_vport_timers(vports[i]);
James Smart09372822008-01-11 01:52:54 -05003814 lpfc_destroy_vport_work_array(phba, vports);
James Smart92d7f7b2007-06-17 19:56:38 -05003815 lpfc_printf_log(phba, KERN_WARNING, LOG_INIT,
James Smarte8b62012007-08-02 11:10:09 -04003816 "0460 Bring Adapter offline\n");
dea31012005-04-17 16:05:31 -05003817 /* Bring down the SLI Layer and cleanup. The HBA is offline
3818 now. */
3819 lpfc_sli_hba_down(phba);
James Smart92d7f7b2007-06-17 19:56:38 -05003820 spin_lock_irq(&phba->hbalock);
James Smart7054a602007-04-25 09:52:34 -04003821 phba->work_ha = 0;
James Smart92d7f7b2007-06-17 19:56:38 -05003822 spin_unlock_irq(&phba->hbalock);
James Smart549e55c2007-08-02 11:09:51 -04003823 vports = lpfc_create_vport_work_array(phba);
3824 if (vports != NULL)
James Smartda0436e2009-05-22 14:51:39 -04003825 for (i = 0; i <= phba->max_vports && vports[i] != NULL; i++) {
James Smart549e55c2007-08-02 11:09:51 -04003826 shost = lpfc_shost_from_vport(vports[i]);
James Smart549e55c2007-08-02 11:09:51 -04003827 spin_lock_irq(shost->host_lock);
3828 vports[i]->work_port_events = 0;
3829 vports[i]->fc_flag |= FC_OFFLINE_MODE;
3830 spin_unlock_irq(shost->host_lock);
3831 }
James Smart09372822008-01-11 01:52:54 -05003832 lpfc_destroy_vport_work_array(phba, vports);
James Smartf0871ab2021-01-04 10:02:32 -08003833 /* If OFFLINE flag is clear (i.e. unloading), cpuhp removal is handled
3834 * in hba_unset
3835 */
3836 if (phba->pport->fc_flag & FC_OFFLINE_MODE)
3837 __lpfc_cpuhp_remove(phba);
James Smartc4908502019-01-28 11:14:28 -08003838
3839 if (phba->cfg_xri_rebalancing)
3840 lpfc_destroy_multixri_pools(phba);
dea31012005-04-17 16:05:31 -05003841}
3842
James Smarte59058c2008-08-24 21:49:00 -04003843/**
James Smart3621a712009-04-06 18:47:14 -04003844 * lpfc_scsi_free - Free all the SCSI buffers and IOCBs from driver lists
James Smarte59058c2008-08-24 21:49:00 -04003845 * @phba: pointer to lpfc hba data structure.
3846 *
3847 * This routine is to free all the SCSI buffers and IOCBs from the driver
3848 * list back to kernel. It is called from lpfc_pci_remove_one to free
3849 * the internal resources before the device is removed from the system.
James Smarte59058c2008-08-24 21:49:00 -04003850 **/
James Smart8a9d2e82012-05-09 21:16:12 -04003851static void
James Smart2e0fef82007-06-17 19:56:36 -05003852lpfc_scsi_free(struct lpfc_hba *phba)
dea31012005-04-17 16:05:31 -05003853{
James Smartc4908502019-01-28 11:14:28 -08003854 struct lpfc_io_buf *sb, *sb_next;
dea31012005-04-17 16:05:31 -05003855
James Smart895427b2017-02-12 13:52:30 -08003856 if (!(phba->cfg_enable_fc4_type & LPFC_ENABLE_FCP))
3857 return;
3858
James Smart2e0fef82007-06-17 19:56:36 -05003859 spin_lock_irq(&phba->hbalock);
James Smarta40fc5f2013-04-17 20:17:40 -04003860
dea31012005-04-17 16:05:31 -05003861 /* Release all the lpfc_scsi_bufs maintained by this host. */
James Smarta40fc5f2013-04-17 20:17:40 -04003862
3863 spin_lock(&phba->scsi_buf_list_put_lock);
3864 list_for_each_entry_safe(sb, sb_next, &phba->lpfc_scsi_buf_list_put,
3865 list) {
dea31012005-04-17 16:05:31 -05003866 list_del(&sb->list);
Romain Perier771db5c2017-07-06 10:13:05 +02003867 dma_pool_free(phba->lpfc_sg_dma_buf_pool, sb->data,
James Smart92d7f7b2007-06-17 19:56:38 -05003868 sb->dma_handle);
dea31012005-04-17 16:05:31 -05003869 kfree(sb);
3870 phba->total_scsi_bufs--;
3871 }
James Smarta40fc5f2013-04-17 20:17:40 -04003872 spin_unlock(&phba->scsi_buf_list_put_lock);
3873
3874 spin_lock(&phba->scsi_buf_list_get_lock);
3875 list_for_each_entry_safe(sb, sb_next, &phba->lpfc_scsi_buf_list_get,
3876 list) {
3877 list_del(&sb->list);
Romain Perier771db5c2017-07-06 10:13:05 +02003878 dma_pool_free(phba->lpfc_sg_dma_buf_pool, sb->data,
James Smarta40fc5f2013-04-17 20:17:40 -04003879 sb->dma_handle);
3880 kfree(sb);
3881 phba->total_scsi_bufs--;
3882 }
3883 spin_unlock(&phba->scsi_buf_list_get_lock);
James Smart2e0fef82007-06-17 19:56:36 -05003884 spin_unlock_irq(&phba->hbalock);
James Smart8a9d2e82012-05-09 21:16:12 -04003885}
James Smart0794d602019-01-28 11:14:19 -08003886
James Smart8a9d2e82012-05-09 21:16:12 -04003887/**
James Smart5e5b5112019-01-28 11:14:22 -08003888 * lpfc_io_free - Free all the IO buffers and IOCBs from driver lists
James Smart895427b2017-02-12 13:52:30 -08003889 * @phba: pointer to lpfc hba data structure.
3890 *
James Smart0794d602019-01-28 11:14:19 -08003891 * This routine is to free all the IO buffers and IOCBs from the driver
James Smart895427b2017-02-12 13:52:30 -08003892 * list back to kernel. It is called from lpfc_pci_remove_one to free
3893 * the internal resources before the device is removed from the system.
3894 **/
James Smartc4908502019-01-28 11:14:28 -08003895void
James Smart5e5b5112019-01-28 11:14:22 -08003896lpfc_io_free(struct lpfc_hba *phba)
James Smart895427b2017-02-12 13:52:30 -08003897{
James Smartc4908502019-01-28 11:14:28 -08003898 struct lpfc_io_buf *lpfc_ncmd, *lpfc_ncmd_next;
James Smart5e5b5112019-01-28 11:14:22 -08003899 struct lpfc_sli4_hdw_queue *qp;
3900 int idx;
James Smart895427b2017-02-12 13:52:30 -08003901
James Smart5e5b5112019-01-28 11:14:22 -08003902 for (idx = 0; idx < phba->cfg_hdw_queue; idx++) {
3903 qp = &phba->sli4_hba.hdwq[idx];
3904 /* Release all the lpfc_nvme_bufs maintained by this host. */
3905 spin_lock(&qp->io_buf_list_put_lock);
3906 list_for_each_entry_safe(lpfc_ncmd, lpfc_ncmd_next,
3907 &qp->lpfc_io_buf_list_put,
3908 list) {
3909 list_del(&lpfc_ncmd->list);
3910 qp->put_io_bufs--;
3911 dma_pool_free(phba->lpfc_sg_dma_buf_pool,
3912 lpfc_ncmd->data, lpfc_ncmd->dma_handle);
James Smartd79c9e92019-08-14 16:57:09 -07003913 if (phba->cfg_xpsgl && !phba->nvmet_support)
3914 lpfc_put_sgl_per_hdwq(phba, lpfc_ncmd);
3915 lpfc_put_cmd_rsp_buf_per_hdwq(phba, lpfc_ncmd);
James Smart5e5b5112019-01-28 11:14:22 -08003916 kfree(lpfc_ncmd);
3917 qp->total_io_bufs--;
3918 }
3919 spin_unlock(&qp->io_buf_list_put_lock);
James Smart895427b2017-02-12 13:52:30 -08003920
James Smart5e5b5112019-01-28 11:14:22 -08003921 spin_lock(&qp->io_buf_list_get_lock);
3922 list_for_each_entry_safe(lpfc_ncmd, lpfc_ncmd_next,
3923 &qp->lpfc_io_buf_list_get,
3924 list) {
3925 list_del(&lpfc_ncmd->list);
3926 qp->get_io_bufs--;
3927 dma_pool_free(phba->lpfc_sg_dma_buf_pool,
3928 lpfc_ncmd->data, lpfc_ncmd->dma_handle);
James Smartd79c9e92019-08-14 16:57:09 -07003929 if (phba->cfg_xpsgl && !phba->nvmet_support)
3930 lpfc_put_sgl_per_hdwq(phba, lpfc_ncmd);
3931 lpfc_put_cmd_rsp_buf_per_hdwq(phba, lpfc_ncmd);
James Smart5e5b5112019-01-28 11:14:22 -08003932 kfree(lpfc_ncmd);
3933 qp->total_io_bufs--;
3934 }
3935 spin_unlock(&qp->io_buf_list_get_lock);
James Smart895427b2017-02-12 13:52:30 -08003936 }
James Smart895427b2017-02-12 13:52:30 -08003937}
James Smart0794d602019-01-28 11:14:19 -08003938
James Smart895427b2017-02-12 13:52:30 -08003939/**
3940 * lpfc_sli4_els_sgl_update - update ELS xri-sgl sizing and mapping
James Smart8a9d2e82012-05-09 21:16:12 -04003941 * @phba: pointer to lpfc hba data structure.
3942 *
3943 * This routine first calculates the sizes of the current els and allocated
3944 * scsi sgl lists, and then goes through all sgls to updates the physical
3945 * XRIs assigned due to port function reset. During port initialization, the
3946 * current els and allocated scsi sgl lists are 0s.
3947 *
3948 * Return codes
3949 * 0 - successful (for now, it always returns 0)
3950 **/
3951int
James Smart895427b2017-02-12 13:52:30 -08003952lpfc_sli4_els_sgl_update(struct lpfc_hba *phba)
James Smart8a9d2e82012-05-09 21:16:12 -04003953{
3954 struct lpfc_sglq *sglq_entry = NULL, *sglq_entry_next = NULL;
James Smart895427b2017-02-12 13:52:30 -08003955 uint16_t i, lxri, xri_cnt, els_xri_cnt;
James Smart8a9d2e82012-05-09 21:16:12 -04003956 LIST_HEAD(els_sgl_list);
James Smart8a9d2e82012-05-09 21:16:12 -04003957 int rc;
3958
3959 /*
3960 * update on pci function's els xri-sgl list
3961 */
3962 els_xri_cnt = lpfc_sli4_get_els_iocb_cnt(phba);
James Smart895427b2017-02-12 13:52:30 -08003963
James Smart8a9d2e82012-05-09 21:16:12 -04003964 if (els_xri_cnt > phba->sli4_hba.els_xri_cnt) {
3965 /* els xri-sgl expanded */
3966 xri_cnt = els_xri_cnt - phba->sli4_hba.els_xri_cnt;
3967 lpfc_printf_log(phba, KERN_INFO, LOG_SLI,
3968 "3157 ELS xri-sgl count increased from "
3969 "%d to %d\n", phba->sli4_hba.els_xri_cnt,
3970 els_xri_cnt);
3971 /* allocate the additional els sgls */
3972 for (i = 0; i < xri_cnt; i++) {
3973 sglq_entry = kzalloc(sizeof(struct lpfc_sglq),
3974 GFP_KERNEL);
3975 if (sglq_entry == NULL) {
Dick Kennedy372c1872020-06-30 14:50:00 -07003976 lpfc_printf_log(phba, KERN_ERR,
3977 LOG_TRACE_EVENT,
James Smart8a9d2e82012-05-09 21:16:12 -04003978 "2562 Failure to allocate an "
3979 "ELS sgl entry:%d\n", i);
3980 rc = -ENOMEM;
3981 goto out_free_mem;
3982 }
3983 sglq_entry->buff_type = GEN_BUFF_TYPE;
3984 sglq_entry->virt = lpfc_mbuf_alloc(phba, 0,
3985 &sglq_entry->phys);
3986 if (sglq_entry->virt == NULL) {
3987 kfree(sglq_entry);
Dick Kennedy372c1872020-06-30 14:50:00 -07003988 lpfc_printf_log(phba, KERN_ERR,
3989 LOG_TRACE_EVENT,
James Smart8a9d2e82012-05-09 21:16:12 -04003990 "2563 Failure to allocate an "
3991 "ELS mbuf:%d\n", i);
3992 rc = -ENOMEM;
3993 goto out_free_mem;
3994 }
3995 sglq_entry->sgl = sglq_entry->virt;
3996 memset(sglq_entry->sgl, 0, LPFC_BPL_SIZE);
3997 sglq_entry->state = SGL_FREED;
3998 list_add_tail(&sglq_entry->list, &els_sgl_list);
3999 }
James Smarta7892412021-04-11 18:31:15 -07004000 spin_lock_irq(&phba->sli4_hba.sgl_list_lock);
James Smart895427b2017-02-12 13:52:30 -08004001 list_splice_init(&els_sgl_list,
4002 &phba->sli4_hba.lpfc_els_sgl_list);
James Smarta7892412021-04-11 18:31:15 -07004003 spin_unlock_irq(&phba->sli4_hba.sgl_list_lock);
James Smart8a9d2e82012-05-09 21:16:12 -04004004 } else if (els_xri_cnt < phba->sli4_hba.els_xri_cnt) {
4005 /* els xri-sgl shrinked */
4006 xri_cnt = phba->sli4_hba.els_xri_cnt - els_xri_cnt;
4007 lpfc_printf_log(phba, KERN_INFO, LOG_SLI,
4008 "3158 ELS xri-sgl count decreased from "
4009 "%d to %d\n", phba->sli4_hba.els_xri_cnt,
4010 els_xri_cnt);
James Smarta7892412021-04-11 18:31:15 -07004011 spin_lock_irq(&phba->sli4_hba.sgl_list_lock);
James Smart895427b2017-02-12 13:52:30 -08004012 list_splice_init(&phba->sli4_hba.lpfc_els_sgl_list,
4013 &els_sgl_list);
James Smart8a9d2e82012-05-09 21:16:12 -04004014 /* release extra els sgls from list */
4015 for (i = 0; i < xri_cnt; i++) {
4016 list_remove_head(&els_sgl_list,
4017 sglq_entry, struct lpfc_sglq, list);
4018 if (sglq_entry) {
James Smart895427b2017-02-12 13:52:30 -08004019 __lpfc_mbuf_free(phba, sglq_entry->virt,
4020 sglq_entry->phys);
James Smart8a9d2e82012-05-09 21:16:12 -04004021 kfree(sglq_entry);
4022 }
4023 }
James Smart895427b2017-02-12 13:52:30 -08004024 list_splice_init(&els_sgl_list,
4025 &phba->sli4_hba.lpfc_els_sgl_list);
James Smarta7892412021-04-11 18:31:15 -07004026 spin_unlock_irq(&phba->sli4_hba.sgl_list_lock);
James Smart8a9d2e82012-05-09 21:16:12 -04004027 } else
4028 lpfc_printf_log(phba, KERN_INFO, LOG_SLI,
4029 "3163 ELS xri-sgl count unchanged: %d\n",
4030 els_xri_cnt);
4031 phba->sli4_hba.els_xri_cnt = els_xri_cnt;
4032
4033 /* update xris to els sgls on the list */
4034 sglq_entry = NULL;
4035 sglq_entry_next = NULL;
4036 list_for_each_entry_safe(sglq_entry, sglq_entry_next,
James Smart895427b2017-02-12 13:52:30 -08004037 &phba->sli4_hba.lpfc_els_sgl_list, list) {
James Smart8a9d2e82012-05-09 21:16:12 -04004038 lxri = lpfc_sli4_next_xritag(phba);
4039 if (lxri == NO_XRI) {
Dick Kennedy372c1872020-06-30 14:50:00 -07004040 lpfc_printf_log(phba, KERN_ERR,
4041 LOG_TRACE_EVENT,
James Smart8a9d2e82012-05-09 21:16:12 -04004042 "2400 Failed to allocate xri for "
4043 "ELS sgl\n");
4044 rc = -ENOMEM;
4045 goto out_free_mem;
4046 }
4047 sglq_entry->sli4_lxritag = lxri;
4048 sglq_entry->sli4_xritag = phba->sli4_hba.xri_ids[lxri];
4049 }
James Smart895427b2017-02-12 13:52:30 -08004050 return 0;
4051
4052out_free_mem:
4053 lpfc_free_els_sgl_list(phba);
4054 return rc;
4055}
4056
4057/**
James Smartf358dd02017-02-12 13:52:34 -08004058 * lpfc_sli4_nvmet_sgl_update - update xri-sgl sizing and mapping
4059 * @phba: pointer to lpfc hba data structure.
4060 *
4061 * This routine first calculates the sizes of the current els and allocated
4062 * scsi sgl lists, and then goes through all sgls to updates the physical
4063 * XRIs assigned due to port function reset. During port initialization, the
4064 * current els and allocated scsi sgl lists are 0s.
4065 *
4066 * Return codes
4067 * 0 - successful (for now, it always returns 0)
4068 **/
4069int
4070lpfc_sli4_nvmet_sgl_update(struct lpfc_hba *phba)
4071{
4072 struct lpfc_sglq *sglq_entry = NULL, *sglq_entry_next = NULL;
4073 uint16_t i, lxri, xri_cnt, els_xri_cnt;
James Smart6c621a22017-05-15 15:20:45 -07004074 uint16_t nvmet_xri_cnt;
James Smartf358dd02017-02-12 13:52:34 -08004075 LIST_HEAD(nvmet_sgl_list);
4076 int rc;
4077
4078 /*
4079 * update on pci function's nvmet xri-sgl list
4080 */
4081 els_xri_cnt = lpfc_sli4_get_els_iocb_cnt(phba);
James Smart61f3d4b2017-05-15 15:20:41 -07004082
James Smart6c621a22017-05-15 15:20:45 -07004083 /* For NVMET, ALL remaining XRIs are dedicated for IO processing */
4084 nvmet_xri_cnt = phba->sli4_hba.max_cfg_param.max_xri - els_xri_cnt;
James Smartf358dd02017-02-12 13:52:34 -08004085 if (nvmet_xri_cnt > phba->sli4_hba.nvmet_xri_cnt) {
4086 /* els xri-sgl expanded */
4087 xri_cnt = nvmet_xri_cnt - phba->sli4_hba.nvmet_xri_cnt;
4088 lpfc_printf_log(phba, KERN_INFO, LOG_SLI,
4089 "6302 NVMET xri-sgl cnt grew from %d to %d\n",
4090 phba->sli4_hba.nvmet_xri_cnt, nvmet_xri_cnt);
4091 /* allocate the additional nvmet sgls */
4092 for (i = 0; i < xri_cnt; i++) {
4093 sglq_entry = kzalloc(sizeof(struct lpfc_sglq),
4094 GFP_KERNEL);
4095 if (sglq_entry == NULL) {
Dick Kennedy372c1872020-06-30 14:50:00 -07004096 lpfc_printf_log(phba, KERN_ERR,
4097 LOG_TRACE_EVENT,
James Smartf358dd02017-02-12 13:52:34 -08004098 "6303 Failure to allocate an "
4099 "NVMET sgl entry:%d\n", i);
4100 rc = -ENOMEM;
4101 goto out_free_mem;
4102 }
4103 sglq_entry->buff_type = NVMET_BUFF_TYPE;
4104 sglq_entry->virt = lpfc_nvmet_buf_alloc(phba, 0,
4105 &sglq_entry->phys);
4106 if (sglq_entry->virt == NULL) {
4107 kfree(sglq_entry);
Dick Kennedy372c1872020-06-30 14:50:00 -07004108 lpfc_printf_log(phba, KERN_ERR,
4109 LOG_TRACE_EVENT,
James Smartf358dd02017-02-12 13:52:34 -08004110 "6304 Failure to allocate an "
4111 "NVMET buf:%d\n", i);
4112 rc = -ENOMEM;
4113 goto out_free_mem;
4114 }
4115 sglq_entry->sgl = sglq_entry->virt;
4116 memset(sglq_entry->sgl, 0,
4117 phba->cfg_sg_dma_buf_size);
4118 sglq_entry->state = SGL_FREED;
4119 list_add_tail(&sglq_entry->list, &nvmet_sgl_list);
4120 }
4121 spin_lock_irq(&phba->hbalock);
4122 spin_lock(&phba->sli4_hba.sgl_list_lock);
4123 list_splice_init(&nvmet_sgl_list,
4124 &phba->sli4_hba.lpfc_nvmet_sgl_list);
4125 spin_unlock(&phba->sli4_hba.sgl_list_lock);
4126 spin_unlock_irq(&phba->hbalock);
4127 } else if (nvmet_xri_cnt < phba->sli4_hba.nvmet_xri_cnt) {
4128 /* nvmet xri-sgl shrunk */
4129 xri_cnt = phba->sli4_hba.nvmet_xri_cnt - nvmet_xri_cnt;
4130 lpfc_printf_log(phba, KERN_INFO, LOG_SLI,
4131 "6305 NVMET xri-sgl count decreased from "
4132 "%d to %d\n", phba->sli4_hba.nvmet_xri_cnt,
4133 nvmet_xri_cnt);
4134 spin_lock_irq(&phba->hbalock);
4135 spin_lock(&phba->sli4_hba.sgl_list_lock);
4136 list_splice_init(&phba->sli4_hba.lpfc_nvmet_sgl_list,
4137 &nvmet_sgl_list);
4138 /* release extra nvmet sgls from list */
4139 for (i = 0; i < xri_cnt; i++) {
4140 list_remove_head(&nvmet_sgl_list,
4141 sglq_entry, struct lpfc_sglq, list);
4142 if (sglq_entry) {
4143 lpfc_nvmet_buf_free(phba, sglq_entry->virt,
4144 sglq_entry->phys);
4145 kfree(sglq_entry);
4146 }
4147 }
4148 list_splice_init(&nvmet_sgl_list,
4149 &phba->sli4_hba.lpfc_nvmet_sgl_list);
4150 spin_unlock(&phba->sli4_hba.sgl_list_lock);
4151 spin_unlock_irq(&phba->hbalock);
4152 } else
4153 lpfc_printf_log(phba, KERN_INFO, LOG_SLI,
4154 "6306 NVMET xri-sgl count unchanged: %d\n",
4155 nvmet_xri_cnt);
4156 phba->sli4_hba.nvmet_xri_cnt = nvmet_xri_cnt;
4157
4158 /* update xris to nvmet sgls on the list */
4159 sglq_entry = NULL;
4160 sglq_entry_next = NULL;
4161 list_for_each_entry_safe(sglq_entry, sglq_entry_next,
4162 &phba->sli4_hba.lpfc_nvmet_sgl_list, list) {
4163 lxri = lpfc_sli4_next_xritag(phba);
4164 if (lxri == NO_XRI) {
Dick Kennedy372c1872020-06-30 14:50:00 -07004165 lpfc_printf_log(phba, KERN_ERR,
4166 LOG_TRACE_EVENT,
James Smartf358dd02017-02-12 13:52:34 -08004167 "6307 Failed to allocate xri for "
4168 "NVMET sgl\n");
4169 rc = -ENOMEM;
4170 goto out_free_mem;
4171 }
4172 sglq_entry->sli4_lxritag = lxri;
4173 sglq_entry->sli4_xritag = phba->sli4_hba.xri_ids[lxri];
4174 }
4175 return 0;
4176
4177out_free_mem:
4178 lpfc_free_nvmet_sgl_list(phba);
4179 return rc;
4180}
4181
James Smart5e5b5112019-01-28 11:14:22 -08004182int
4183lpfc_io_buf_flush(struct lpfc_hba *phba, struct list_head *cbuf)
4184{
4185 LIST_HEAD(blist);
4186 struct lpfc_sli4_hdw_queue *qp;
James Smartc4908502019-01-28 11:14:28 -08004187 struct lpfc_io_buf *lpfc_cmd;
4188 struct lpfc_io_buf *iobufp, *prev_iobufp;
James Smart5e5b5112019-01-28 11:14:22 -08004189 int idx, cnt, xri, inserted;
4190
4191 cnt = 0;
4192 for (idx = 0; idx < phba->cfg_hdw_queue; idx++) {
4193 qp = &phba->sli4_hba.hdwq[idx];
4194 spin_lock_irq(&qp->io_buf_list_get_lock);
4195 spin_lock(&qp->io_buf_list_put_lock);
4196
4197 /* Take everything off the get and put lists */
4198 list_splice_init(&qp->lpfc_io_buf_list_get, &blist);
4199 list_splice(&qp->lpfc_io_buf_list_put, &blist);
4200 INIT_LIST_HEAD(&qp->lpfc_io_buf_list_get);
4201 INIT_LIST_HEAD(&qp->lpfc_io_buf_list_put);
4202 cnt += qp->get_io_bufs + qp->put_io_bufs;
4203 qp->get_io_bufs = 0;
4204 qp->put_io_bufs = 0;
4205 qp->total_io_bufs = 0;
4206 spin_unlock(&qp->io_buf_list_put_lock);
4207 spin_unlock_irq(&qp->io_buf_list_get_lock);
4208 }
4209
4210 /*
4211 * Take IO buffers off blist and put on cbuf sorted by XRI.
4212 * This is because POST_SGL takes a sequential range of XRIs
4213 * to post to the firmware.
4214 */
4215 for (idx = 0; idx < cnt; idx++) {
James Smartc4908502019-01-28 11:14:28 -08004216 list_remove_head(&blist, lpfc_cmd, struct lpfc_io_buf, list);
James Smart5e5b5112019-01-28 11:14:22 -08004217 if (!lpfc_cmd)
4218 return cnt;
4219 if (idx == 0) {
4220 list_add_tail(&lpfc_cmd->list, cbuf);
4221 continue;
4222 }
4223 xri = lpfc_cmd->cur_iocbq.sli4_xritag;
4224 inserted = 0;
4225 prev_iobufp = NULL;
4226 list_for_each_entry(iobufp, cbuf, list) {
4227 if (xri < iobufp->cur_iocbq.sli4_xritag) {
4228 if (prev_iobufp)
4229 list_add(&lpfc_cmd->list,
4230 &prev_iobufp->list);
4231 else
4232 list_add(&lpfc_cmd->list, cbuf);
4233 inserted = 1;
4234 break;
4235 }
4236 prev_iobufp = iobufp;
4237 }
4238 if (!inserted)
4239 list_add_tail(&lpfc_cmd->list, cbuf);
4240 }
4241 return cnt;
4242}
4243
4244int
4245lpfc_io_buf_replenish(struct lpfc_hba *phba, struct list_head *cbuf)
4246{
4247 struct lpfc_sli4_hdw_queue *qp;
James Smartc4908502019-01-28 11:14:28 -08004248 struct lpfc_io_buf *lpfc_cmd;
James Smart5e5b5112019-01-28 11:14:22 -08004249 int idx, cnt;
4250
4251 qp = phba->sli4_hba.hdwq;
4252 cnt = 0;
4253 while (!list_empty(cbuf)) {
4254 for (idx = 0; idx < phba->cfg_hdw_queue; idx++) {
4255 list_remove_head(cbuf, lpfc_cmd,
James Smartc4908502019-01-28 11:14:28 -08004256 struct lpfc_io_buf, list);
James Smart5e5b5112019-01-28 11:14:22 -08004257 if (!lpfc_cmd)
4258 return cnt;
4259 cnt++;
4260 qp = &phba->sli4_hba.hdwq[idx];
James Smart1fbf9742019-01-28 11:14:26 -08004261 lpfc_cmd->hdwq_no = idx;
4262 lpfc_cmd->hdwq = qp;
James Smart5e5b5112019-01-28 11:14:22 -08004263 lpfc_cmd->cur_iocbq.wqe_cmpl = NULL;
4264 lpfc_cmd->cur_iocbq.iocb_cmpl = NULL;
4265 spin_lock(&qp->io_buf_list_put_lock);
4266 list_add_tail(&lpfc_cmd->list,
4267 &qp->lpfc_io_buf_list_put);
4268 qp->put_io_bufs++;
4269 qp->total_io_bufs++;
4270 spin_unlock(&qp->io_buf_list_put_lock);
4271 }
4272 }
4273 return cnt;
4274}
4275
James Smartf358dd02017-02-12 13:52:34 -08004276/**
James Smart5e5b5112019-01-28 11:14:22 -08004277 * lpfc_sli4_io_sgl_update - update xri-sgl sizing and mapping
James Smart895427b2017-02-12 13:52:30 -08004278 * @phba: pointer to lpfc hba data structure.
4279 *
4280 * This routine first calculates the sizes of the current els and allocated
4281 * scsi sgl lists, and then goes through all sgls to updates the physical
4282 * XRIs assigned due to port function reset. During port initialization, the
4283 * current els and allocated scsi sgl lists are 0s.
4284 *
4285 * Return codes
4286 * 0 - successful (for now, it always returns 0)
4287 **/
4288int
James Smart5e5b5112019-01-28 11:14:22 -08004289lpfc_sli4_io_sgl_update(struct lpfc_hba *phba)
James Smart895427b2017-02-12 13:52:30 -08004290{
James Smartc4908502019-01-28 11:14:28 -08004291 struct lpfc_io_buf *lpfc_ncmd = NULL, *lpfc_ncmd_next = NULL;
James Smart0794d602019-01-28 11:14:19 -08004292 uint16_t i, lxri, els_xri_cnt;
James Smart5e5b5112019-01-28 11:14:22 -08004293 uint16_t io_xri_cnt, io_xri_max;
4294 LIST_HEAD(io_sgl_list);
James Smart0794d602019-01-28 11:14:19 -08004295 int rc, cnt;
James Smart895427b2017-02-12 13:52:30 -08004296
4297 /*
James Smart0794d602019-01-28 11:14:19 -08004298 * update on pci function's allocated nvme xri-sgl list
James Smart895427b2017-02-12 13:52:30 -08004299 */
James Smart0794d602019-01-28 11:14:19 -08004300
4301 /* maximum number of xris available for nvme buffers */
James Smart895427b2017-02-12 13:52:30 -08004302 els_xri_cnt = lpfc_sli4_get_els_iocb_cnt(phba);
James Smart5e5b5112019-01-28 11:14:22 -08004303 io_xri_max = phba->sli4_hba.max_cfg_param.max_xri - els_xri_cnt;
4304 phba->sli4_hba.io_xri_max = io_xri_max;
James Smart8a9d2e82012-05-09 21:16:12 -04004305
James Smarte8c0a772017-04-21 16:04:49 -07004306 lpfc_printf_log(phba, KERN_INFO, LOG_SLI,
James Smart0794d602019-01-28 11:14:19 -08004307 "6074 Current allocated XRI sgl count:%d, "
4308 "maximum XRI count:%d\n",
James Smart5e5b5112019-01-28 11:14:22 -08004309 phba->sli4_hba.io_xri_cnt,
4310 phba->sli4_hba.io_xri_max);
James Smarte8c0a772017-04-21 16:04:49 -07004311
James Smart5e5b5112019-01-28 11:14:22 -08004312 cnt = lpfc_io_buf_flush(phba, &io_sgl_list);
James Smart0794d602019-01-28 11:14:19 -08004313
James Smart5e5b5112019-01-28 11:14:22 -08004314 if (phba->sli4_hba.io_xri_cnt > phba->sli4_hba.io_xri_max) {
James Smart0794d602019-01-28 11:14:19 -08004315 /* max nvme xri shrunk below the allocated nvme buffers */
James Smart5e5b5112019-01-28 11:14:22 -08004316 io_xri_cnt = phba->sli4_hba.io_xri_cnt -
4317 phba->sli4_hba.io_xri_max;
James Smart0794d602019-01-28 11:14:19 -08004318 /* release the extra allocated nvme buffers */
James Smart5e5b5112019-01-28 11:14:22 -08004319 for (i = 0; i < io_xri_cnt; i++) {
4320 list_remove_head(&io_sgl_list, lpfc_ncmd,
James Smartc4908502019-01-28 11:14:28 -08004321 struct lpfc_io_buf, list);
James Smart0794d602019-01-28 11:14:19 -08004322 if (lpfc_ncmd) {
Romain Perier771db5c2017-07-06 10:13:05 +02004323 dma_pool_free(phba->lpfc_sg_dma_buf_pool,
James Smart0794d602019-01-28 11:14:19 -08004324 lpfc_ncmd->data,
4325 lpfc_ncmd->dma_handle);
4326 kfree(lpfc_ncmd);
James Smarta2fc4aef2014-09-03 12:57:55 -04004327 }
James Smart8a9d2e82012-05-09 21:16:12 -04004328 }
James Smart5e5b5112019-01-28 11:14:22 -08004329 phba->sli4_hba.io_xri_cnt -= io_xri_cnt;
James Smart8a9d2e82012-05-09 21:16:12 -04004330 }
4331
James Smart0794d602019-01-28 11:14:19 -08004332 /* update xris associated to remaining allocated nvme buffers */
4333 lpfc_ncmd = NULL;
4334 lpfc_ncmd_next = NULL;
James Smart5e5b5112019-01-28 11:14:22 -08004335 phba->sli4_hba.io_xri_cnt = cnt;
James Smart0794d602019-01-28 11:14:19 -08004336 list_for_each_entry_safe(lpfc_ncmd, lpfc_ncmd_next,
James Smart5e5b5112019-01-28 11:14:22 -08004337 &io_sgl_list, list) {
James Smart8a9d2e82012-05-09 21:16:12 -04004338 lxri = lpfc_sli4_next_xritag(phba);
4339 if (lxri == NO_XRI) {
Dick Kennedy372c1872020-06-30 14:50:00 -07004340 lpfc_printf_log(phba, KERN_ERR,
4341 LOG_TRACE_EVENT,
James Smart0794d602019-01-28 11:14:19 -08004342 "6075 Failed to allocate xri for "
4343 "nvme buffer\n");
James Smart8a9d2e82012-05-09 21:16:12 -04004344 rc = -ENOMEM;
4345 goto out_free_mem;
4346 }
James Smart0794d602019-01-28 11:14:19 -08004347 lpfc_ncmd->cur_iocbq.sli4_lxritag = lxri;
4348 lpfc_ncmd->cur_iocbq.sli4_xritag = phba->sli4_hba.xri_ids[lxri];
James Smart8a9d2e82012-05-09 21:16:12 -04004349 }
James Smart5e5b5112019-01-28 11:14:22 -08004350 cnt = lpfc_io_buf_replenish(phba, &io_sgl_list);
dea31012005-04-17 16:05:31 -05004351 return 0;
James Smart8a9d2e82012-05-09 21:16:12 -04004352
4353out_free_mem:
James Smart5e5b5112019-01-28 11:14:22 -08004354 lpfc_io_free(phba);
James Smart8a9d2e82012-05-09 21:16:12 -04004355 return rc;
dea31012005-04-17 16:05:31 -05004356}
4357
James Smart0794d602019-01-28 11:14:19 -08004358/**
James Smart5e5b5112019-01-28 11:14:22 -08004359 * lpfc_new_io_buf - IO buffer allocator for HBA with SLI4 IF spec
Lee Jonesfe614ac2020-07-23 13:24:22 +01004360 * @phba: Pointer to lpfc hba data structure.
4361 * @num_to_alloc: The requested number of buffers to allocate.
James Smart0794d602019-01-28 11:14:19 -08004362 *
4363 * This routine allocates nvme buffers for device with SLI-4 interface spec,
4364 * the nvme buffer contains all the necessary information needed to initiate
4365 * an I/O. After allocating up to @num_to_allocate IO buffers and put
4366 * them on a list, it post them to the port by using SGL block post.
4367 *
4368 * Return codes:
James Smart5e5b5112019-01-28 11:14:22 -08004369 * int - number of IO buffers that were allocated and posted.
James Smart0794d602019-01-28 11:14:19 -08004370 * 0 = failure, less than num_to_alloc is a partial failure.
4371 **/
4372int
James Smart5e5b5112019-01-28 11:14:22 -08004373lpfc_new_io_buf(struct lpfc_hba *phba, int num_to_alloc)
James Smart0794d602019-01-28 11:14:19 -08004374{
James Smartc4908502019-01-28 11:14:28 -08004375 struct lpfc_io_buf *lpfc_ncmd;
James Smart0794d602019-01-28 11:14:19 -08004376 struct lpfc_iocbq *pwqeq;
4377 uint16_t iotag, lxri = 0;
4378 int bcnt, num_posted;
4379 LIST_HEAD(prep_nblist);
4380 LIST_HEAD(post_nblist);
4381 LIST_HEAD(nvme_nblist);
4382
James Smart5e5b5112019-01-28 11:14:22 -08004383 phba->sli4_hba.io_xri_cnt = 0;
James Smart0794d602019-01-28 11:14:19 -08004384 for (bcnt = 0; bcnt < num_to_alloc; bcnt++) {
James Smart7f9989b2019-08-27 14:27:46 -07004385 lpfc_ncmd = kzalloc(sizeof(*lpfc_ncmd), GFP_KERNEL);
James Smart0794d602019-01-28 11:14:19 -08004386 if (!lpfc_ncmd)
4387 break;
4388 /*
4389 * Get memory from the pci pool to map the virt space to
4390 * pci bus space for an I/O. The DMA buffer includes the
4391 * number of SGE's necessary to support the sg_tablesize.
4392 */
Thomas Meyera5c990e2019-05-29 22:21:36 +02004393 lpfc_ncmd->data = dma_pool_zalloc(phba->lpfc_sg_dma_buf_pool,
4394 GFP_KERNEL,
4395 &lpfc_ncmd->dma_handle);
James Smart0794d602019-01-28 11:14:19 -08004396 if (!lpfc_ncmd->data) {
4397 kfree(lpfc_ncmd);
4398 break;
4399 }
James Smart0794d602019-01-28 11:14:19 -08004400
James Smartd79c9e92019-08-14 16:57:09 -07004401 if (phba->cfg_xpsgl && !phba->nvmet_support) {
4402 INIT_LIST_HEAD(&lpfc_ncmd->dma_sgl_xtra_list);
4403 } else {
4404 /*
4405 * 4K Page alignment is CRITICAL to BlockGuard, double
4406 * check to be sure.
4407 */
4408 if ((phba->sli3_options & LPFC_SLI3_BG_ENABLED) &&
4409 (((unsigned long)(lpfc_ncmd->data) &
4410 (unsigned long)(SLI4_PAGE_SIZE - 1)) != 0)) {
Dick Kennedy372c1872020-06-30 14:50:00 -07004411 lpfc_printf_log(phba, KERN_ERR,
4412 LOG_TRACE_EVENT,
James Smartd79c9e92019-08-14 16:57:09 -07004413 "3369 Memory alignment err: "
4414 "addr=%lx\n",
4415 (unsigned long)lpfc_ncmd->data);
4416 dma_pool_free(phba->lpfc_sg_dma_buf_pool,
4417 lpfc_ncmd->data,
4418 lpfc_ncmd->dma_handle);
4419 kfree(lpfc_ncmd);
4420 break;
4421 }
James Smart0794d602019-01-28 11:14:19 -08004422 }
4423
James Smartd79c9e92019-08-14 16:57:09 -07004424 INIT_LIST_HEAD(&lpfc_ncmd->dma_cmd_rsp_list);
4425
James Smart0794d602019-01-28 11:14:19 -08004426 lxri = lpfc_sli4_next_xritag(phba);
4427 if (lxri == NO_XRI) {
4428 dma_pool_free(phba->lpfc_sg_dma_buf_pool,
4429 lpfc_ncmd->data, lpfc_ncmd->dma_handle);
4430 kfree(lpfc_ncmd);
4431 break;
4432 }
4433 pwqeq = &lpfc_ncmd->cur_iocbq;
4434
4435 /* Allocate iotag for lpfc_ncmd->cur_iocbq. */
4436 iotag = lpfc_sli_next_iotag(phba, pwqeq);
4437 if (iotag == 0) {
4438 dma_pool_free(phba->lpfc_sg_dma_buf_pool,
4439 lpfc_ncmd->data, lpfc_ncmd->dma_handle);
4440 kfree(lpfc_ncmd);
Dick Kennedy372c1872020-06-30 14:50:00 -07004441 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart0794d602019-01-28 11:14:19 -08004442 "6121 Failed to allocate IOTAG for"
4443 " XRI:0x%x\n", lxri);
4444 lpfc_sli4_free_xri(phba, lxri);
4445 break;
4446 }
4447 pwqeq->sli4_lxritag = lxri;
4448 pwqeq->sli4_xritag = phba->sli4_hba.xri_ids[lxri];
4449 pwqeq->context1 = lpfc_ncmd;
4450
4451 /* Initialize local short-hand pointers. */
4452 lpfc_ncmd->dma_sgl = lpfc_ncmd->data;
4453 lpfc_ncmd->dma_phys_sgl = lpfc_ncmd->dma_handle;
4454 lpfc_ncmd->cur_iocbq.context1 = lpfc_ncmd;
James Smartc2017262019-01-28 11:14:37 -08004455 spin_lock_init(&lpfc_ncmd->buf_lock);
James Smart0794d602019-01-28 11:14:19 -08004456
4457 /* add the nvme buffer to a post list */
4458 list_add_tail(&lpfc_ncmd->list, &post_nblist);
James Smart5e5b5112019-01-28 11:14:22 -08004459 phba->sli4_hba.io_xri_cnt++;
James Smart0794d602019-01-28 11:14:19 -08004460 }
4461 lpfc_printf_log(phba, KERN_INFO, LOG_NVME,
4462 "6114 Allocate %d out of %d requested new NVME "
4463 "buffers\n", bcnt, num_to_alloc);
4464
4465 /* post the list of nvme buffer sgls to port if available */
4466 if (!list_empty(&post_nblist))
James Smart5e5b5112019-01-28 11:14:22 -08004467 num_posted = lpfc_sli4_post_io_sgl_list(
James Smart0794d602019-01-28 11:14:19 -08004468 phba, &post_nblist, bcnt);
4469 else
4470 num_posted = 0;
4471
4472 return num_posted;
4473}
4474
James Smart96418b52017-03-04 09:30:31 -08004475static uint64_t
4476lpfc_get_wwpn(struct lpfc_hba *phba)
4477{
4478 uint64_t wwn;
4479 int rc;
4480 LPFC_MBOXQ_t *mboxq;
4481 MAILBOX_t *mb;
4482
James Smart96418b52017-03-04 09:30:31 -08004483 mboxq = (LPFC_MBOXQ_t *) mempool_alloc(phba->mbox_mem_pool,
4484 GFP_KERNEL);
4485 if (!mboxq)
4486 return (uint64_t)-1;
4487
4488 /* First get WWN of HBA instance */
4489 lpfc_read_nv(phba, mboxq);
4490 rc = lpfc_sli_issue_mbox(phba, mboxq, MBX_POLL);
4491 if (rc != MBX_SUCCESS) {
Dick Kennedy372c1872020-06-30 14:50:00 -07004492 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart96418b52017-03-04 09:30:31 -08004493 "6019 Mailbox failed , mbxCmd x%x "
4494 "READ_NV, mbxStatus x%x\n",
4495 bf_get(lpfc_mqe_command, &mboxq->u.mqe),
4496 bf_get(lpfc_mqe_status, &mboxq->u.mqe));
4497 mempool_free(mboxq, phba->mbox_mem_pool);
4498 return (uint64_t) -1;
4499 }
4500 mb = &mboxq->u.mb;
4501 memcpy(&wwn, (char *)mb->un.varRDnvp.portname, sizeof(uint64_t));
4502 /* wwn is WWPN of HBA instance */
4503 mempool_free(mboxq, phba->mbox_mem_pool);
4504 if (phba->sli_rev == LPFC_SLI_REV4)
4505 return be64_to_cpu(wwn);
4506 else
Maurizio Lombardi286871a2017-08-23 16:55:48 -07004507 return rol64(wwn, 32);
James Smart96418b52017-03-04 09:30:31 -08004508}
4509
James Smarte59058c2008-08-24 21:49:00 -04004510/**
Gaurav Srivastava5e633302021-06-08 10:05:49 +05304511 * lpfc_vmid_res_alloc - Allocates resources for VMID
4512 * @phba: pointer to lpfc hba data structure.
4513 * @vport: pointer to vport data structure
4514 *
4515 * This routine allocated the resources needed for the VMID.
4516 *
4517 * Return codes
4518 * 0 on Success
4519 * Non-0 on Failure
4520 */
4521static int
4522lpfc_vmid_res_alloc(struct lpfc_hba *phba, struct lpfc_vport *vport)
4523{
4524 /* VMID feature is supported only on SLI4 */
4525 if (phba->sli_rev == LPFC_SLI_REV3) {
4526 phba->cfg_vmid_app_header = 0;
4527 phba->cfg_vmid_priority_tagging = 0;
4528 }
4529
4530 if (lpfc_is_vmid_enabled(phba)) {
4531 vport->vmid =
4532 kcalloc(phba->cfg_max_vmid, sizeof(struct lpfc_vmid),
4533 GFP_KERNEL);
4534 if (!vport->vmid)
4535 return -ENOMEM;
4536
4537 rwlock_init(&vport->vmid_lock);
4538
4539 /* Set the VMID parameters for the vport */
4540 vport->vmid_priority_tagging = phba->cfg_vmid_priority_tagging;
4541 vport->vmid_inactivity_timeout =
4542 phba->cfg_vmid_inactivity_timeout;
4543 vport->max_vmid = phba->cfg_max_vmid;
4544 vport->cur_vmid_cnt = 0;
4545
4546 vport->vmid_priority_range = bitmap_zalloc
4547 (LPFC_VMID_MAX_PRIORITY_RANGE, GFP_KERNEL);
4548
4549 if (!vport->vmid_priority_range) {
4550 kfree(vport->vmid);
4551 return -ENOMEM;
4552 }
4553
4554 hash_init(vport->hash_table);
4555 }
4556 return 0;
4557}
4558
4559/**
James Smart3621a712009-04-06 18:47:14 -04004560 * lpfc_create_port - Create an FC port
James Smarte59058c2008-08-24 21:49:00 -04004561 * @phba: pointer to lpfc hba data structure.
4562 * @instance: a unique integer ID to this FC port.
4563 * @dev: pointer to the device data structure.
4564 *
4565 * This routine creates a FC port for the upper layer protocol. The FC port
4566 * can be created on top of either a physical port or a virtual port provided
4567 * by the HBA. This routine also allocates a SCSI host data structure (shost)
4568 * and associates the FC port created before adding the shost into the SCSI
4569 * layer.
4570 *
4571 * Return codes
4572 * @vport - pointer to the virtual N_Port data structure.
4573 * NULL - port create failed.
4574 **/
James Smart2e0fef82007-06-17 19:56:36 -05004575struct lpfc_vport *
James Smart3de2a652007-08-02 11:09:59 -04004576lpfc_create_port(struct lpfc_hba *phba, int instance, struct device *dev)
James Smart47a86172007-04-25 09:53:22 -04004577{
James Smart2e0fef82007-06-17 19:56:36 -05004578 struct lpfc_vport *vport;
James Smart895427b2017-02-12 13:52:30 -08004579 struct Scsi_Host *shost = NULL;
James Smartc90b4482020-03-22 11:12:56 -07004580 struct scsi_host_template *template;
James Smart2e0fef82007-06-17 19:56:36 -05004581 int error = 0;
James Smart96418b52017-03-04 09:30:31 -08004582 int i;
4583 uint64_t wwn;
4584 bool use_no_reset_hba = false;
James Smart56bc8022017-06-15 22:56:43 -07004585 int rc;
James Smart96418b52017-03-04 09:30:31 -08004586
James Smart56bc8022017-06-15 22:56:43 -07004587 if (lpfc_no_hba_reset_cnt) {
4588 if (phba->sli_rev < LPFC_SLI_REV4 &&
4589 dev == &phba->pcidev->dev) {
4590 /* Reset the port first */
4591 lpfc_sli_brdrestart(phba);
4592 rc = lpfc_sli_chipset_init(phba);
4593 if (rc)
4594 return NULL;
4595 }
4596 wwn = lpfc_get_wwpn(phba);
4597 }
James Smart96418b52017-03-04 09:30:31 -08004598
4599 for (i = 0; i < lpfc_no_hba_reset_cnt; i++) {
4600 if (wwn == lpfc_no_hba_reset[i]) {
Dick Kennedy372c1872020-06-30 14:50:00 -07004601 lpfc_printf_log(phba, KERN_ERR,
4602 LOG_TRACE_EVENT,
James Smart96418b52017-03-04 09:30:31 -08004603 "6020 Setting use_no_reset port=%llx\n",
4604 wwn);
4605 use_no_reset_hba = true;
4606 break;
4607 }
4608 }
James Smart47a86172007-04-25 09:53:22 -04004609
James Smartc90b4482020-03-22 11:12:56 -07004610 /* Seed template for SCSI host registration */
4611 if (dev == &phba->pcidev->dev) {
4612 template = &phba->port_template;
4613
4614 if (phba->cfg_enable_fc4_type & LPFC_ENABLE_FCP) {
4615 /* Seed physical port template */
4616 memcpy(template, &lpfc_template, sizeof(*template));
4617
James Smart7c30bb62020-10-20 13:27:16 -07004618 if (use_no_reset_hba)
James Smartc90b4482020-03-22 11:12:56 -07004619 /* template is for a no reset SCSI Host */
James Smartc90b4482020-03-22 11:12:56 -07004620 template->eh_host_reset_handler = NULL;
James Smartc90b4482020-03-22 11:12:56 -07004621
4622 /* Template for all vports this physical port creates */
4623 memcpy(&phba->vport_template, &lpfc_template,
4624 sizeof(*template));
Bart Van Assche08adfa72021-10-12 16:35:39 -07004625 phba->vport_template.shost_groups = lpfc_vport_groups;
James Smartc90b4482020-03-22 11:12:56 -07004626 phba->vport_template.eh_bus_reset_handler = NULL;
4627 phba->vport_template.eh_host_reset_handler = NULL;
4628 phba->vport_template.vendor_id = 0;
4629
4630 /* Initialize the host templates with updated value */
4631 if (phba->sli_rev == LPFC_SLI_REV4) {
4632 template->sg_tablesize = phba->cfg_scsi_seg_cnt;
4633 phba->vport_template.sg_tablesize =
4634 phba->cfg_scsi_seg_cnt;
4635 } else {
4636 template->sg_tablesize = phba->cfg_sg_seg_cnt;
4637 phba->vport_template.sg_tablesize =
4638 phba->cfg_sg_seg_cnt;
4639 }
4640
James Smart895427b2017-02-12 13:52:30 -08004641 } else {
James Smartc90b4482020-03-22 11:12:56 -07004642 /* NVMET is for physical port only */
4643 memcpy(template, &lpfc_template_nvme,
4644 sizeof(*template));
James Smart895427b2017-02-12 13:52:30 -08004645 }
James Smartc90b4482020-03-22 11:12:56 -07004646 } else {
4647 template = &phba->vport_template;
James Smartea4142f2015-04-07 15:07:13 -04004648 }
James Smartc90b4482020-03-22 11:12:56 -07004649
4650 shost = scsi_host_alloc(template, sizeof(struct lpfc_vport));
James Smart2e0fef82007-06-17 19:56:36 -05004651 if (!shost)
4652 goto out;
James Smart47a86172007-04-25 09:53:22 -04004653
James Smart2e0fef82007-06-17 19:56:36 -05004654 vport = (struct lpfc_vport *) shost->hostdata;
4655 vport->phba = phba;
James Smart2e0fef82007-06-17 19:56:36 -05004656 vport->load_flag |= FC_LOADING;
James Smart92d7f7b2007-06-17 19:56:38 -05004657 vport->fc_flag |= FC_VPORT_NEEDS_REG_VPI;
James Smart7f5f3d02008-02-08 18:50:14 -05004658 vport->fc_rscn_flush = 0;
James Smart3de2a652007-08-02 11:09:59 -04004659 lpfc_get_vport_cfgparam(vport);
James Smart895427b2017-02-12 13:52:30 -08004660
James Smartf6e84792019-01-28 11:14:38 -08004661 /* Adjust value in vport */
4662 vport->cfg_enable_fc4_type = phba->cfg_enable_fc4_type;
4663
James Smart2e0fef82007-06-17 19:56:36 -05004664 shost->unique_id = instance;
4665 shost->max_id = LPFC_MAX_TARGET;
James Smart3de2a652007-08-02 11:09:59 -04004666 shost->max_lun = vport->cfg_max_luns;
James Smart2e0fef82007-06-17 19:56:36 -05004667 shost->this_id = -1;
4668 shost->max_cmd_len = 16;
James Smart6a828b02019-01-28 11:14:31 -08004669
James Smartda0436e2009-05-22 14:51:39 -04004670 if (phba->sli_rev == LPFC_SLI_REV4) {
James Smart77ffd342019-08-15 19:36:49 -07004671 if (!phba->cfg_fcp_mq_threshold ||
4672 phba->cfg_fcp_mq_threshold > phba->cfg_hdw_queue)
4673 phba->cfg_fcp_mq_threshold = phba->cfg_hdw_queue;
4674
4675 shost->nr_hw_queues = min_t(int, 2 * num_possible_nodes(),
4676 phba->cfg_fcp_mq_threshold);
James Smart6a828b02019-01-28 11:14:31 -08004677
James Smart28baac72010-02-12 14:42:03 -05004678 shost->dma_boundary =
James Smartcb5172e2010-03-15 11:25:07 -04004679 phba->sli4_hba.pc_sli4_params.sge_supp_len-1;
James Smartd79c9e92019-08-14 16:57:09 -07004680
4681 if (phba->cfg_xpsgl && !phba->nvmet_support)
4682 shost->sg_tablesize = LPFC_MAX_SG_TABLESIZE;
4683 else
4684 shost->sg_tablesize = phba->cfg_scsi_seg_cnt;
James Smartace44e42019-01-28 11:14:27 -08004685 } else
4686 /* SLI-3 has a limited number of hardware queues (3),
4687 * thus there is only one for FCP processing.
4688 */
4689 shost->nr_hw_queues = 1;
James Smart81301a92008-12-04 22:39:46 -05004690
James Smart47a86172007-04-25 09:53:22 -04004691 /*
James Smart2e0fef82007-06-17 19:56:36 -05004692 * Set initial can_queue value since 0 is no longer supported and
4693 * scsi_add_host will fail. This will be adjusted later based on the
4694 * max xri value determined in hba setup.
James Smart47a86172007-04-25 09:53:22 -04004695 */
James Smart2e0fef82007-06-17 19:56:36 -05004696 shost->can_queue = phba->cfg_hba_queue_depth - 10;
James Smart3de2a652007-08-02 11:09:59 -04004697 if (dev != &phba->pcidev->dev) {
James Smart92d7f7b2007-06-17 19:56:38 -05004698 shost->transportt = lpfc_vport_transport_template;
4699 vport->port_type = LPFC_NPIV_PORT;
4700 } else {
4701 shost->transportt = lpfc_transport_template;
4702 vport->port_type = LPFC_PHYSICAL_PORT;
4703 }
James Smart47a86172007-04-25 09:53:22 -04004704
James Smartc90b4482020-03-22 11:12:56 -07004705 lpfc_printf_log(phba, KERN_INFO, LOG_INIT | LOG_FCP,
4706 "9081 CreatePort TMPLATE type %x TBLsize %d "
4707 "SEGcnt %d/%d\n",
4708 vport->port_type, shost->sg_tablesize,
4709 phba->cfg_scsi_seg_cnt, phba->cfg_sg_seg_cnt);
4710
Gaurav Srivastava5e633302021-06-08 10:05:49 +05304711 /* Allocate the resources for VMID */
4712 rc = lpfc_vmid_res_alloc(phba, vport);
4713
4714 if (rc)
4715 goto out;
4716
James Smart2e0fef82007-06-17 19:56:36 -05004717 /* Initialize all internally managed lists. */
4718 INIT_LIST_HEAD(&vport->fc_nodes);
James Smartda0436e2009-05-22 14:51:39 -04004719 INIT_LIST_HEAD(&vport->rcv_buffer_list);
James Smart2e0fef82007-06-17 19:56:36 -05004720 spin_lock_init(&vport->work_port_lock);
James Smart47a86172007-04-25 09:53:22 -04004721
Kees Cookf22eb4d2017-09-06 20:24:26 -07004722 timer_setup(&vport->fc_disctmo, lpfc_disc_timeout, 0);
James Smart47a86172007-04-25 09:53:22 -04004723
Kees Cookf22eb4d2017-09-06 20:24:26 -07004724 timer_setup(&vport->els_tmofunc, lpfc_els_timeout, 0);
James Smart92494142011-02-16 12:39:44 -05004725
Kees Cookf22eb4d2017-09-06 20:24:26 -07004726 timer_setup(&vport->delayed_disc_tmo, lpfc_delayed_disc_tmo, 0);
James Smart92494142011-02-16 12:39:44 -05004727
James Smartaa6ff302019-05-21 17:49:09 -07004728 if (phba->sli3_options & LPFC_SLI3_BG_ENABLED)
4729 lpfc_setup_bg(phba, shost);
4730
James Bottomleyd139b9b2009-11-05 13:33:12 -06004731 error = scsi_add_host_with_dma(shost, dev, &phba->pcidev->dev);
James Smart2e0fef82007-06-17 19:56:36 -05004732 if (error)
4733 goto out_put_shost;
James Smart47a86172007-04-25 09:53:22 -04004734
James Smart523128e2018-09-10 10:30:46 -07004735 spin_lock_irq(&phba->port_list_lock);
James Smart2e0fef82007-06-17 19:56:36 -05004736 list_add_tail(&vport->listentry, &phba->port_list);
James Smart523128e2018-09-10 10:30:46 -07004737 spin_unlock_irq(&phba->port_list_lock);
James Smart2e0fef82007-06-17 19:56:36 -05004738 return vport;
James Smart47a86172007-04-25 09:53:22 -04004739
James Smart2e0fef82007-06-17 19:56:36 -05004740out_put_shost:
Gaurav Srivastava5e633302021-06-08 10:05:49 +05304741 kfree(vport->vmid);
4742 bitmap_free(vport->vmid_priority_range);
James Smart2e0fef82007-06-17 19:56:36 -05004743 scsi_host_put(shost);
4744out:
4745 return NULL;
James Smart47a86172007-04-25 09:53:22 -04004746}
4747
James Smarte59058c2008-08-24 21:49:00 -04004748/**
James Smart3621a712009-04-06 18:47:14 -04004749 * destroy_port - destroy an FC port
James Smarte59058c2008-08-24 21:49:00 -04004750 * @vport: pointer to an lpfc virtual N_Port data structure.
4751 *
4752 * This routine destroys a FC port from the upper layer protocol. All the
4753 * resources associated with the port are released.
4754 **/
James Smart2e0fef82007-06-17 19:56:36 -05004755void
4756destroy_port(struct lpfc_vport *vport)
James Smart47a86172007-04-25 09:53:22 -04004757{
James Smart92d7f7b2007-06-17 19:56:38 -05004758 struct Scsi_Host *shost = lpfc_shost_from_vport(vport);
4759 struct lpfc_hba *phba = vport->phba;
James Smart47a86172007-04-25 09:53:22 -04004760
James Smart858c9f62007-06-17 19:56:39 -05004761 lpfc_debugfs_terminate(vport);
James Smart92d7f7b2007-06-17 19:56:38 -05004762 fc_remove_host(shost);
4763 scsi_remove_host(shost);
James Smart47a86172007-04-25 09:53:22 -04004764
James Smart523128e2018-09-10 10:30:46 -07004765 spin_lock_irq(&phba->port_list_lock);
James Smart92d7f7b2007-06-17 19:56:38 -05004766 list_del_init(&vport->listentry);
James Smart523128e2018-09-10 10:30:46 -07004767 spin_unlock_irq(&phba->port_list_lock);
James Smart47a86172007-04-25 09:53:22 -04004768
James Smart92d7f7b2007-06-17 19:56:38 -05004769 lpfc_cleanup(vport);
James Smart47a86172007-04-25 09:53:22 -04004770 return;
James Smart47a86172007-04-25 09:53:22 -04004771}
4772
James Smarte59058c2008-08-24 21:49:00 -04004773/**
James Smart3621a712009-04-06 18:47:14 -04004774 * lpfc_get_instance - Get a unique integer ID
James Smarte59058c2008-08-24 21:49:00 -04004775 *
4776 * This routine allocates a unique integer ID from lpfc_hba_index pool. It
4777 * uses the kernel idr facility to perform the task.
4778 *
4779 * Return codes:
4780 * instance - a unique integer ID allocated as the new instance.
4781 * -1 - lpfc get instance failed.
4782 **/
James Smart92d7f7b2007-06-17 19:56:38 -05004783int
4784lpfc_get_instance(void)
4785{
Tejun Heoab516032013-02-27 17:04:44 -08004786 int ret;
James Smart92d7f7b2007-06-17 19:56:38 -05004787
Tejun Heoab516032013-02-27 17:04:44 -08004788 ret = idr_alloc(&lpfc_hba_index, NULL, 0, 0, GFP_KERNEL);
4789 return ret < 0 ? -1 : ret;
James Smart92d7f7b2007-06-17 19:56:38 -05004790}
4791
James Smarte59058c2008-08-24 21:49:00 -04004792/**
James Smart3621a712009-04-06 18:47:14 -04004793 * lpfc_scan_finished - method for SCSI layer to detect whether scan is done
James Smarte59058c2008-08-24 21:49:00 -04004794 * @shost: pointer to SCSI host data structure.
4795 * @time: elapsed time of the scan in jiffies.
4796 *
4797 * This routine is called by the SCSI layer with a SCSI host to determine
4798 * whether the scan host is finished.
4799 *
4800 * Note: there is no scan_start function as adapter initialization will have
4801 * asynchronously kicked off the link initialization.
4802 *
4803 * Return codes
4804 * 0 - SCSI host scan is not over yet.
4805 * 1 - SCSI host scan is over.
4806 **/
James Smart47a86172007-04-25 09:53:22 -04004807int lpfc_scan_finished(struct Scsi_Host *shost, unsigned long time)
4808{
James Smart2e0fef82007-06-17 19:56:36 -05004809 struct lpfc_vport *vport = (struct lpfc_vport *) shost->hostdata;
4810 struct lpfc_hba *phba = vport->phba;
James Smart858c9f62007-06-17 19:56:39 -05004811 int stat = 0;
James Smart47a86172007-04-25 09:53:22 -04004812
James Smart858c9f62007-06-17 19:56:39 -05004813 spin_lock_irq(shost->host_lock);
4814
James Smart51ef4c22007-08-02 11:10:31 -04004815 if (vport->load_flag & FC_UNLOADING) {
James Smart858c9f62007-06-17 19:56:39 -05004816 stat = 1;
James Smart47a86172007-04-25 09:53:22 -04004817 goto finished;
James Smart858c9f62007-06-17 19:56:39 -05004818 }
James Smart256ec0d2013-04-17 20:14:58 -04004819 if (time >= msecs_to_jiffies(30 * 1000)) {
James Smart2e0fef82007-06-17 19:56:36 -05004820 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
James Smarte8b62012007-08-02 11:10:09 -04004821 "0461 Scanning longer than 30 "
4822 "seconds. Continuing initialization\n");
James Smart858c9f62007-06-17 19:56:39 -05004823 stat = 1;
James Smart47a86172007-04-25 09:53:22 -04004824 goto finished;
James Smart2e0fef82007-06-17 19:56:36 -05004825 }
James Smart256ec0d2013-04-17 20:14:58 -04004826 if (time >= msecs_to_jiffies(15 * 1000) &&
4827 phba->link_state <= LPFC_LINK_DOWN) {
James Smart2e0fef82007-06-17 19:56:36 -05004828 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
James Smarte8b62012007-08-02 11:10:09 -04004829 "0465 Link down longer than 15 "
4830 "seconds. Continuing initialization\n");
James Smart858c9f62007-06-17 19:56:39 -05004831 stat = 1;
James Smart2e0fef82007-06-17 19:56:36 -05004832 goto finished;
James Smart47a86172007-04-25 09:53:22 -04004833 }
4834
James Smart2e0fef82007-06-17 19:56:36 -05004835 if (vport->port_state != LPFC_VPORT_READY)
James Smart858c9f62007-06-17 19:56:39 -05004836 goto finished;
James Smart2e0fef82007-06-17 19:56:36 -05004837 if (vport->num_disc_nodes || vport->fc_prli_sent)
James Smart858c9f62007-06-17 19:56:39 -05004838 goto finished;
James Smart256ec0d2013-04-17 20:14:58 -04004839 if (vport->fc_map_cnt == 0 && time < msecs_to_jiffies(2 * 1000))
James Smart858c9f62007-06-17 19:56:39 -05004840 goto finished;
James Smart2e0fef82007-06-17 19:56:36 -05004841 if ((phba->sli.sli_flag & LPFC_SLI_MBOX_ACTIVE) != 0)
James Smart858c9f62007-06-17 19:56:39 -05004842 goto finished;
4843
4844 stat = 1;
James Smart47a86172007-04-25 09:53:22 -04004845
4846finished:
James Smart858c9f62007-06-17 19:56:39 -05004847 spin_unlock_irq(shost->host_lock);
4848 return stat;
James Smart92d7f7b2007-06-17 19:56:38 -05004849}
4850
Bart Van Assche3999df72019-03-28 11:06:16 -07004851static void lpfc_host_supported_speeds_set(struct Scsi_Host *shost)
James Smartcd713482018-10-23 13:41:01 -07004852{
4853 struct lpfc_vport *vport = (struct lpfc_vport *)shost->hostdata;
4854 struct lpfc_hba *phba = vport->phba;
4855
4856 fc_host_supported_speeds(shost) = 0;
Dick Kennedya1e4d3d2020-08-03 14:02:22 -07004857 /*
4858 * Avoid reporting supported link speed for FCoE as it can't be
4859 * controlled via FCoE.
4860 */
4861 if (phba->hba_flag & HBA_FCOE_MODE)
4862 return;
4863
James Smartbfc47782021-07-22 15:17:19 -07004864 if (phba->lmt & LMT_256Gb)
4865 fc_host_supported_speeds(shost) |= FC_PORTSPEED_256GBIT;
James Smart1dc5ec22018-10-23 13:41:11 -07004866 if (phba->lmt & LMT_128Gb)
4867 fc_host_supported_speeds(shost) |= FC_PORTSPEED_128GBIT;
James Smartcd713482018-10-23 13:41:01 -07004868 if (phba->lmt & LMT_64Gb)
4869 fc_host_supported_speeds(shost) |= FC_PORTSPEED_64GBIT;
4870 if (phba->lmt & LMT_32Gb)
4871 fc_host_supported_speeds(shost) |= FC_PORTSPEED_32GBIT;
4872 if (phba->lmt & LMT_16Gb)
4873 fc_host_supported_speeds(shost) |= FC_PORTSPEED_16GBIT;
4874 if (phba->lmt & LMT_10Gb)
4875 fc_host_supported_speeds(shost) |= FC_PORTSPEED_10GBIT;
4876 if (phba->lmt & LMT_8Gb)
4877 fc_host_supported_speeds(shost) |= FC_PORTSPEED_8GBIT;
4878 if (phba->lmt & LMT_4Gb)
4879 fc_host_supported_speeds(shost) |= FC_PORTSPEED_4GBIT;
4880 if (phba->lmt & LMT_2Gb)
4881 fc_host_supported_speeds(shost) |= FC_PORTSPEED_2GBIT;
4882 if (phba->lmt & LMT_1Gb)
4883 fc_host_supported_speeds(shost) |= FC_PORTSPEED_1GBIT;
4884}
4885
James Smarte59058c2008-08-24 21:49:00 -04004886/**
James Smart3621a712009-04-06 18:47:14 -04004887 * lpfc_host_attrib_init - Initialize SCSI host attributes on a FC port
James Smarte59058c2008-08-24 21:49:00 -04004888 * @shost: pointer to SCSI host data structure.
4889 *
4890 * This routine initializes a given SCSI host attributes on a FC port. The
4891 * SCSI host can be either on top of a physical port or a virtual port.
4892 **/
James Smart92d7f7b2007-06-17 19:56:38 -05004893void lpfc_host_attrib_init(struct Scsi_Host *shost)
4894{
4895 struct lpfc_vport *vport = (struct lpfc_vport *) shost->hostdata;
4896 struct lpfc_hba *phba = vport->phba;
James Smart47a86172007-04-25 09:53:22 -04004897 /*
James Smart2e0fef82007-06-17 19:56:36 -05004898 * Set fixed host attributes. Must done after lpfc_sli_hba_setup().
James Smart47a86172007-04-25 09:53:22 -04004899 */
4900
James Smart2e0fef82007-06-17 19:56:36 -05004901 fc_host_node_name(shost) = wwn_to_u64(vport->fc_nodename.u.wwn);
4902 fc_host_port_name(shost) = wwn_to_u64(vport->fc_portname.u.wwn);
James Smart47a86172007-04-25 09:53:22 -04004903 fc_host_supported_classes(shost) = FC_COS_CLASS3;
4904
4905 memset(fc_host_supported_fc4s(shost), 0,
James Smart2e0fef82007-06-17 19:56:36 -05004906 sizeof(fc_host_supported_fc4s(shost)));
James Smart47a86172007-04-25 09:53:22 -04004907 fc_host_supported_fc4s(shost)[2] = 1;
4908 fc_host_supported_fc4s(shost)[7] = 1;
4909
James Smart92d7f7b2007-06-17 19:56:38 -05004910 lpfc_vport_symbolic_node_name(vport, fc_host_symbolic_name(shost),
4911 sizeof fc_host_symbolic_name(shost));
James Smart47a86172007-04-25 09:53:22 -04004912
James Smartcd713482018-10-23 13:41:01 -07004913 lpfc_host_supported_speeds_set(shost);
James Smart47a86172007-04-25 09:53:22 -04004914
4915 fc_host_maxframe_size(shost) =
James Smart2e0fef82007-06-17 19:56:36 -05004916 (((uint32_t) vport->fc_sparam.cmn.bbRcvSizeMsb & 0x0F) << 8) |
4917 (uint32_t) vport->fc_sparam.cmn.bbRcvSizeLsb;
James Smart47a86172007-04-25 09:53:22 -04004918
Mike Christie0af5d702010-09-15 16:52:31 -05004919 fc_host_dev_loss_tmo(shost) = vport->cfg_devloss_tmo;
4920
James Smart47a86172007-04-25 09:53:22 -04004921 /* This value is also unchanging */
4922 memset(fc_host_active_fc4s(shost), 0,
James Smart2e0fef82007-06-17 19:56:36 -05004923 sizeof(fc_host_active_fc4s(shost)));
James Smart47a86172007-04-25 09:53:22 -04004924 fc_host_active_fc4s(shost)[2] = 1;
4925 fc_host_active_fc4s(shost)[7] = 1;
4926
James Smart92d7f7b2007-06-17 19:56:38 -05004927 fc_host_max_npiv_vports(shost) = phba->max_vpi;
James Smart47a86172007-04-25 09:53:22 -04004928 spin_lock_irq(shost->host_lock);
James Smart51ef4c22007-08-02 11:10:31 -04004929 vport->load_flag &= ~FC_LOADING;
James Smart47a86172007-04-25 09:53:22 -04004930 spin_unlock_irq(shost->host_lock);
James Smart47a86172007-04-25 09:53:22 -04004931}
dea31012005-04-17 16:05:31 -05004932
James Smarte59058c2008-08-24 21:49:00 -04004933/**
James Smartda0436e2009-05-22 14:51:39 -04004934 * lpfc_stop_port_s3 - Stop SLI3 device port
James Smarte59058c2008-08-24 21:49:00 -04004935 * @phba: pointer to lpfc hba data structure.
4936 *
James Smartda0436e2009-05-22 14:51:39 -04004937 * This routine is invoked to stop an SLI3 device port, it stops the device
4938 * from generating interrupts and stops the device driver's timers for the
4939 * device.
James Smarte59058c2008-08-24 21:49:00 -04004940 **/
James Smartdb2378e2008-02-08 18:49:51 -05004941static void
James Smartda0436e2009-05-22 14:51:39 -04004942lpfc_stop_port_s3(struct lpfc_hba *phba)
James Smartdb2378e2008-02-08 18:49:51 -05004943{
James Smartda0436e2009-05-22 14:51:39 -04004944 /* Clear all interrupt enable conditions */
4945 writel(0, phba->HCregaddr);
4946 readl(phba->HCregaddr); /* flush */
4947 /* Clear all pending interrupts */
4948 writel(0xffffffff, phba->HAregaddr);
4949 readl(phba->HAregaddr); /* flush */
James Smart93996272008-08-24 21:50:30 -04004950
James Smartda0436e2009-05-22 14:51:39 -04004951 /* Reset some HBA SLI setup states */
4952 lpfc_stop_hba_timers(phba);
4953 phba->pport->work_port_events = 0;
James Smartdb2378e2008-02-08 18:49:51 -05004954}
4955
James Smarte59058c2008-08-24 21:49:00 -04004956/**
James Smartda0436e2009-05-22 14:51:39 -04004957 * lpfc_stop_port_s4 - Stop SLI4 device port
James Smart5b75da22008-12-04 22:39:35 -05004958 * @phba: pointer to lpfc hba data structure.
4959 *
James Smartda0436e2009-05-22 14:51:39 -04004960 * This routine is invoked to stop an SLI4 device port, it stops the device
4961 * from generating interrupts and stops the device driver's timers for the
4962 * device.
4963 **/
4964static void
4965lpfc_stop_port_s4(struct lpfc_hba *phba)
4966{
4967 /* Reset some HBA SLI4 setup states */
4968 lpfc_stop_hba_timers(phba);
James Smartcdb42be2019-01-28 11:14:21 -08004969 if (phba->pport)
4970 phba->pport->work_port_events = 0;
James Smartda0436e2009-05-22 14:51:39 -04004971 phba->sli4_hba.intr_enable = 0;
James Smartda0436e2009-05-22 14:51:39 -04004972}
4973
4974/**
4975 * lpfc_stop_port - Wrapper function for stopping hba port
4976 * @phba: Pointer to HBA context object.
James Smart5b75da22008-12-04 22:39:35 -05004977 *
James Smartda0436e2009-05-22 14:51:39 -04004978 * This routine wraps the actual SLI3 or SLI4 hba stop port routine from
4979 * the API jump table function pointer from the lpfc_hba struct.
4980 **/
4981void
4982lpfc_stop_port(struct lpfc_hba *phba)
4983{
4984 phba->lpfc_stop_port(phba);
Dick Kennedyf485c182017-09-29 17:34:34 -07004985
4986 if (phba->wq)
4987 flush_workqueue(phba->wq);
James Smartda0436e2009-05-22 14:51:39 -04004988}
4989
4990/**
James Smartecfd03c2010-02-12 14:41:27 -05004991 * lpfc_fcf_redisc_wait_start_timer - Start fcf rediscover wait timer
4992 * @phba: Pointer to hba for which this call is being executed.
4993 *
4994 * This routine starts the timer waiting for the FCF rediscovery to complete.
4995 **/
4996void
4997lpfc_fcf_redisc_wait_start_timer(struct lpfc_hba *phba)
4998{
4999 unsigned long fcf_redisc_wait_tmo =
5000 (jiffies + msecs_to_jiffies(LPFC_FCF_REDISCOVER_WAIT_TMO));
5001 /* Start fcf rediscovery wait period timer */
5002 mod_timer(&phba->fcf.redisc_wait, fcf_redisc_wait_tmo);
5003 spin_lock_irq(&phba->hbalock);
5004 /* Allow action to new fcf asynchronous event */
5005 phba->fcf.fcf_flag &= ~(FCF_AVAILABLE | FCF_SCAN_DONE);
5006 /* Mark the FCF rediscovery pending state */
5007 phba->fcf.fcf_flag |= FCF_REDISC_PEND;
5008 spin_unlock_irq(&phba->hbalock);
5009}
5010
5011/**
5012 * lpfc_sli4_fcf_redisc_wait_tmo - FCF table rediscover wait timeout
Lee Jonesfe614ac2020-07-23 13:24:22 +01005013 * @t: Timer context used to obtain the pointer to lpfc hba data structure.
James Smartecfd03c2010-02-12 14:41:27 -05005014 *
5015 * This routine is invoked when waiting for FCF table rediscover has been
5016 * timed out. If new FCF record(s) has (have) been discovered during the
5017 * wait period, a new FCF event shall be added to the FCOE async event
5018 * list, and then worker thread shall be waked up for processing from the
5019 * worker thread context.
5020 **/
Rashika Kheriae399b222014-09-03 12:55:28 -04005021static void
Kees Cookf22eb4d2017-09-06 20:24:26 -07005022lpfc_sli4_fcf_redisc_wait_tmo(struct timer_list *t)
James Smartecfd03c2010-02-12 14:41:27 -05005023{
Kees Cookf22eb4d2017-09-06 20:24:26 -07005024 struct lpfc_hba *phba = from_timer(phba, t, fcf.redisc_wait);
James Smartecfd03c2010-02-12 14:41:27 -05005025
5026 /* Don't send FCF rediscovery event if timer cancelled */
5027 spin_lock_irq(&phba->hbalock);
5028 if (!(phba->fcf.fcf_flag & FCF_REDISC_PEND)) {
5029 spin_unlock_irq(&phba->hbalock);
5030 return;
5031 }
5032 /* Clear FCF rediscovery timer pending flag */
5033 phba->fcf.fcf_flag &= ~FCF_REDISC_PEND;
5034 /* FCF rediscovery event to worker thread */
5035 phba->fcf.fcf_flag |= FCF_REDISC_EVT;
5036 spin_unlock_irq(&phba->hbalock);
James Smart0c9ab6f2010-02-26 14:15:57 -05005037 lpfc_printf_log(phba, KERN_INFO, LOG_FIP,
James Smarta93ff372010-10-22 11:06:08 -04005038 "2776 FCF rediscover quiescent timer expired\n");
James Smartecfd03c2010-02-12 14:41:27 -05005039 /* wake up worker thread */
5040 lpfc_worker_wake_up(phba);
5041}
5042
5043/**
Gaurav Srivastava20397172021-06-08 10:05:54 +05305044 * lpfc_vmid_poll - VMID timeout detection
James Smart50baa152021-07-07 11:43:35 -07005045 * @t: Timer context used to obtain the pointer to lpfc hba data structure.
Gaurav Srivastava20397172021-06-08 10:05:54 +05305046 *
5047 * This routine is invoked when there is no I/O on by a VM for the specified
5048 * amount of time. When this situation is detected, the VMID has to be
5049 * deregistered from the switch and all the local resources freed. The VMID
5050 * will be reassigned to the VM once the I/O begins.
5051 **/
5052static void
5053lpfc_vmid_poll(struct timer_list *t)
5054{
5055 struct lpfc_hba *phba = from_timer(phba, t, inactive_vmid_poll);
5056 u32 wake_up = 0;
5057
5058 /* check if there is a need to issue QFPA */
5059 if (phba->pport->vmid_priority_tagging) {
5060 wake_up = 1;
5061 phba->pport->work_port_events |= WORKER_CHECK_VMID_ISSUE_QFPA;
5062 }
5063
5064 /* Is the vmid inactivity timer enabled */
5065 if (phba->pport->vmid_inactivity_timeout ||
5066 phba->pport->load_flag & FC_DEREGISTER_ALL_APP_ID) {
5067 wake_up = 1;
5068 phba->pport->work_port_events |= WORKER_CHECK_INACTIVE_VMID;
5069 }
5070
5071 if (wake_up)
5072 lpfc_worker_wake_up(phba);
5073
5074 /* restart the timer for the next iteration */
5075 mod_timer(&phba->inactive_vmid_poll, jiffies + msecs_to_jiffies(1000 *
5076 LPFC_VMID_TIMER));
5077}
5078
5079/**
James Smartda0436e2009-05-22 14:51:39 -04005080 * lpfc_sli4_parse_latt_fault - Parse sli4 link-attention link fault code
5081 * @phba: pointer to lpfc hba data structure.
5082 * @acqe_link: pointer to the async link completion queue entry.
5083 *
James Smart23288b72018-05-04 20:37:53 -07005084 * This routine is to parse the SLI4 link-attention link fault code.
James Smartda0436e2009-05-22 14:51:39 -04005085 **/
James Smart23288b72018-05-04 20:37:53 -07005086static void
James Smartda0436e2009-05-22 14:51:39 -04005087lpfc_sli4_parse_latt_fault(struct lpfc_hba *phba,
5088 struct lpfc_acqe_link *acqe_link)
5089{
James Smartda0436e2009-05-22 14:51:39 -04005090 switch (bf_get(lpfc_acqe_link_fault, acqe_link)) {
5091 case LPFC_ASYNC_LINK_FAULT_NONE:
5092 case LPFC_ASYNC_LINK_FAULT_LOCAL:
5093 case LPFC_ASYNC_LINK_FAULT_REMOTE:
James Smart23288b72018-05-04 20:37:53 -07005094 case LPFC_ASYNC_LINK_FAULT_LR_LRR:
James Smartda0436e2009-05-22 14:51:39 -04005095 break;
5096 default:
Dick Kennedy372c1872020-06-30 14:50:00 -07005097 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart23288b72018-05-04 20:37:53 -07005098 "0398 Unknown link fault code: x%x\n",
James Smartda0436e2009-05-22 14:51:39 -04005099 bf_get(lpfc_acqe_link_fault, acqe_link));
James Smartda0436e2009-05-22 14:51:39 -04005100 break;
5101 }
James Smartda0436e2009-05-22 14:51:39 -04005102}
5103
5104/**
5105 * lpfc_sli4_parse_latt_type - Parse sli4 link attention type
5106 * @phba: pointer to lpfc hba data structure.
5107 * @acqe_link: pointer to the async link completion queue entry.
5108 *
5109 * This routine is to parse the SLI4 link attention type and translate it
5110 * into the base driver's link attention type coding.
5111 *
5112 * Return: Link attention type in terms of base driver's coding.
5113 **/
5114static uint8_t
5115lpfc_sli4_parse_latt_type(struct lpfc_hba *phba,
5116 struct lpfc_acqe_link *acqe_link)
5117{
5118 uint8_t att_type;
5119
5120 switch (bf_get(lpfc_acqe_link_status, acqe_link)) {
5121 case LPFC_ASYNC_LINK_STATUS_DOWN:
5122 case LPFC_ASYNC_LINK_STATUS_LOGICAL_DOWN:
James Smart76a95d72010-11-20 23:11:48 -05005123 att_type = LPFC_ATT_LINK_DOWN;
James Smartda0436e2009-05-22 14:51:39 -04005124 break;
5125 case LPFC_ASYNC_LINK_STATUS_UP:
5126 /* Ignore physical link up events - wait for logical link up */
James Smart76a95d72010-11-20 23:11:48 -05005127 att_type = LPFC_ATT_RESERVED;
James Smartda0436e2009-05-22 14:51:39 -04005128 break;
5129 case LPFC_ASYNC_LINK_STATUS_LOGICAL_UP:
James Smart76a95d72010-11-20 23:11:48 -05005130 att_type = LPFC_ATT_LINK_UP;
James Smartda0436e2009-05-22 14:51:39 -04005131 break;
5132 default:
Dick Kennedy372c1872020-06-30 14:50:00 -07005133 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartda0436e2009-05-22 14:51:39 -04005134 "0399 Invalid link attention type: x%x\n",
5135 bf_get(lpfc_acqe_link_status, acqe_link));
James Smart76a95d72010-11-20 23:11:48 -05005136 att_type = LPFC_ATT_RESERVED;
James Smartda0436e2009-05-22 14:51:39 -04005137 break;
5138 }
5139 return att_type;
5140}
5141
5142/**
James Smart8b68cd52012-09-29 11:32:37 -04005143 * lpfc_sli_port_speed_get - Get sli3 link speed code to link speed
5144 * @phba: pointer to lpfc hba data structure.
5145 *
5146 * This routine is to get an SLI3 FC port's link speed in Mbps.
5147 *
5148 * Return: link speed in terms of Mbps.
5149 **/
5150uint32_t
5151lpfc_sli_port_speed_get(struct lpfc_hba *phba)
5152{
5153 uint32_t link_speed;
5154
5155 if (!lpfc_is_link_up(phba))
5156 return 0;
5157
James Smarta085e872015-12-16 18:12:02 -05005158 if (phba->sli_rev <= LPFC_SLI_REV3) {
5159 switch (phba->fc_linkspeed) {
5160 case LPFC_LINK_SPEED_1GHZ:
5161 link_speed = 1000;
5162 break;
5163 case LPFC_LINK_SPEED_2GHZ:
5164 link_speed = 2000;
5165 break;
5166 case LPFC_LINK_SPEED_4GHZ:
5167 link_speed = 4000;
5168 break;
5169 case LPFC_LINK_SPEED_8GHZ:
5170 link_speed = 8000;
5171 break;
5172 case LPFC_LINK_SPEED_10GHZ:
5173 link_speed = 10000;
5174 break;
5175 case LPFC_LINK_SPEED_16GHZ:
5176 link_speed = 16000;
5177 break;
5178 default:
5179 link_speed = 0;
5180 }
5181 } else {
5182 if (phba->sli4_hba.link_state.logical_speed)
5183 link_speed =
5184 phba->sli4_hba.link_state.logical_speed;
5185 else
5186 link_speed = phba->sli4_hba.link_state.speed;
James Smart8b68cd52012-09-29 11:32:37 -04005187 }
5188 return link_speed;
5189}
5190
5191/**
5192 * lpfc_sli4_port_speed_parse - Parse async evt link speed code to link speed
5193 * @phba: pointer to lpfc hba data structure.
5194 * @evt_code: asynchronous event code.
5195 * @speed_code: asynchronous event link speed code.
5196 *
5197 * This routine is to parse the giving SLI4 async event link speed code into
5198 * value of Mbps for the link speed.
5199 *
5200 * Return: link speed in terms of Mbps.
5201 **/
5202static uint32_t
5203lpfc_sli4_port_speed_parse(struct lpfc_hba *phba, uint32_t evt_code,
5204 uint8_t speed_code)
5205{
5206 uint32_t port_speed;
5207
5208 switch (evt_code) {
5209 case LPFC_TRAILER_CODE_LINK:
5210 switch (speed_code) {
James Smart26d830e2015-04-07 15:07:17 -04005211 case LPFC_ASYNC_LINK_SPEED_ZERO:
James Smart8b68cd52012-09-29 11:32:37 -04005212 port_speed = 0;
5213 break;
James Smart26d830e2015-04-07 15:07:17 -04005214 case LPFC_ASYNC_LINK_SPEED_10MBPS:
James Smart8b68cd52012-09-29 11:32:37 -04005215 port_speed = 10;
5216 break;
James Smart26d830e2015-04-07 15:07:17 -04005217 case LPFC_ASYNC_LINK_SPEED_100MBPS:
James Smart8b68cd52012-09-29 11:32:37 -04005218 port_speed = 100;
5219 break;
James Smart26d830e2015-04-07 15:07:17 -04005220 case LPFC_ASYNC_LINK_SPEED_1GBPS:
James Smart8b68cd52012-09-29 11:32:37 -04005221 port_speed = 1000;
5222 break;
James Smart26d830e2015-04-07 15:07:17 -04005223 case LPFC_ASYNC_LINK_SPEED_10GBPS:
James Smart8b68cd52012-09-29 11:32:37 -04005224 port_speed = 10000;
5225 break;
James Smart26d830e2015-04-07 15:07:17 -04005226 case LPFC_ASYNC_LINK_SPEED_20GBPS:
5227 port_speed = 20000;
5228 break;
5229 case LPFC_ASYNC_LINK_SPEED_25GBPS:
5230 port_speed = 25000;
5231 break;
5232 case LPFC_ASYNC_LINK_SPEED_40GBPS:
5233 port_speed = 40000;
5234 break;
Dick Kennedya1e4d3d2020-08-03 14:02:22 -07005235 case LPFC_ASYNC_LINK_SPEED_100GBPS:
5236 port_speed = 100000;
5237 break;
James Smart8b68cd52012-09-29 11:32:37 -04005238 default:
5239 port_speed = 0;
5240 }
5241 break;
5242 case LPFC_TRAILER_CODE_FC:
5243 switch (speed_code) {
James Smart26d830e2015-04-07 15:07:17 -04005244 case LPFC_FC_LA_SPEED_UNKNOWN:
James Smart8b68cd52012-09-29 11:32:37 -04005245 port_speed = 0;
5246 break;
James Smart26d830e2015-04-07 15:07:17 -04005247 case LPFC_FC_LA_SPEED_1G:
James Smart8b68cd52012-09-29 11:32:37 -04005248 port_speed = 1000;
5249 break;
James Smart26d830e2015-04-07 15:07:17 -04005250 case LPFC_FC_LA_SPEED_2G:
James Smart8b68cd52012-09-29 11:32:37 -04005251 port_speed = 2000;
5252 break;
James Smart26d830e2015-04-07 15:07:17 -04005253 case LPFC_FC_LA_SPEED_4G:
James Smart8b68cd52012-09-29 11:32:37 -04005254 port_speed = 4000;
5255 break;
James Smart26d830e2015-04-07 15:07:17 -04005256 case LPFC_FC_LA_SPEED_8G:
James Smart8b68cd52012-09-29 11:32:37 -04005257 port_speed = 8000;
5258 break;
James Smart26d830e2015-04-07 15:07:17 -04005259 case LPFC_FC_LA_SPEED_10G:
James Smart8b68cd52012-09-29 11:32:37 -04005260 port_speed = 10000;
5261 break;
James Smart26d830e2015-04-07 15:07:17 -04005262 case LPFC_FC_LA_SPEED_16G:
James Smart8b68cd52012-09-29 11:32:37 -04005263 port_speed = 16000;
5264 break;
James Smartd38dd522015-08-31 16:48:17 -04005265 case LPFC_FC_LA_SPEED_32G:
5266 port_speed = 32000;
5267 break;
James Smartfbd8a6b2018-02-22 08:18:45 -08005268 case LPFC_FC_LA_SPEED_64G:
5269 port_speed = 64000;
5270 break;
James Smart1dc5ec22018-10-23 13:41:11 -07005271 case LPFC_FC_LA_SPEED_128G:
5272 port_speed = 128000;
5273 break;
James Smartbfc47782021-07-22 15:17:19 -07005274 case LPFC_FC_LA_SPEED_256G:
5275 port_speed = 256000;
5276 break;
James Smart8b68cd52012-09-29 11:32:37 -04005277 default:
5278 port_speed = 0;
5279 }
5280 break;
5281 default:
5282 port_speed = 0;
5283 }
5284 return port_speed;
5285}
5286
5287/**
James Smart70f3c072010-12-15 17:57:33 -05005288 * lpfc_sli4_async_link_evt - Process the asynchronous FCoE link event
James Smartda0436e2009-05-22 14:51:39 -04005289 * @phba: pointer to lpfc hba data structure.
5290 * @acqe_link: pointer to the async link completion queue entry.
5291 *
James Smart70f3c072010-12-15 17:57:33 -05005292 * This routine is to handle the SLI4 asynchronous FCoE link event.
James Smartda0436e2009-05-22 14:51:39 -04005293 **/
5294static void
5295lpfc_sli4_async_link_evt(struct lpfc_hba *phba,
5296 struct lpfc_acqe_link *acqe_link)
5297{
5298 struct lpfc_dmabuf *mp;
5299 LPFC_MBOXQ_t *pmb;
5300 MAILBOX_t *mb;
James Smart76a95d72010-11-20 23:11:48 -05005301 struct lpfc_mbx_read_top *la;
James Smartda0436e2009-05-22 14:51:39 -04005302 uint8_t att_type;
James Smart76a95d72010-11-20 23:11:48 -05005303 int rc;
James Smartda0436e2009-05-22 14:51:39 -04005304
5305 att_type = lpfc_sli4_parse_latt_type(phba, acqe_link);
James Smart76a95d72010-11-20 23:11:48 -05005306 if (att_type != LPFC_ATT_LINK_DOWN && att_type != LPFC_ATT_LINK_UP)
James Smartda0436e2009-05-22 14:51:39 -04005307 return;
James Smart32b97932009-07-19 10:01:21 -04005308 phba->fcoe_eventtag = acqe_link->event_tag;
James Smartda0436e2009-05-22 14:51:39 -04005309 pmb = (LPFC_MBOXQ_t *)mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL);
5310 if (!pmb) {
Dick Kennedy372c1872020-06-30 14:50:00 -07005311 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartda0436e2009-05-22 14:51:39 -04005312 "0395 The mboxq allocation failed\n");
5313 return;
5314 }
5315 mp = kmalloc(sizeof(struct lpfc_dmabuf), GFP_KERNEL);
5316 if (!mp) {
Dick Kennedy372c1872020-06-30 14:50:00 -07005317 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartda0436e2009-05-22 14:51:39 -04005318 "0396 The lpfc_dmabuf allocation failed\n");
5319 goto out_free_pmb;
5320 }
5321 mp->virt = lpfc_mbuf_alloc(phba, 0, &mp->phys);
5322 if (!mp->virt) {
Dick Kennedy372c1872020-06-30 14:50:00 -07005323 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartda0436e2009-05-22 14:51:39 -04005324 "0397 The mbuf allocation failed\n");
5325 goto out_free_dmabuf;
5326 }
5327
5328 /* Cleanup any outstanding ELS commands */
5329 lpfc_els_flush_all_cmd(phba);
5330
5331 /* Block ELS IOCBs until we have done process link event */
James Smart895427b2017-02-12 13:52:30 -08005332 phba->sli4_hba.els_wq->pring->flag |= LPFC_STOP_IOCB_EVENT;
James Smartda0436e2009-05-22 14:51:39 -04005333
5334 /* Update link event statistics */
5335 phba->sli.slistat.link_event++;
5336
James Smart76a95d72010-11-20 23:11:48 -05005337 /* Create lpfc_handle_latt mailbox command from link ACQE */
5338 lpfc_read_topology(phba, pmb, mp);
5339 pmb->mbox_cmpl = lpfc_mbx_cmpl_read_topology;
James Smartda0436e2009-05-22 14:51:39 -04005340 pmb->vport = phba->pport;
5341
James Smartda0436e2009-05-22 14:51:39 -04005342 /* Keep the link status for extra SLI4 state machine reference */
5343 phba->sli4_hba.link_state.speed =
James Smart8b68cd52012-09-29 11:32:37 -04005344 lpfc_sli4_port_speed_parse(phba, LPFC_TRAILER_CODE_LINK,
5345 bf_get(lpfc_acqe_link_speed, acqe_link));
James Smartda0436e2009-05-22 14:51:39 -04005346 phba->sli4_hba.link_state.duplex =
5347 bf_get(lpfc_acqe_link_duplex, acqe_link);
5348 phba->sli4_hba.link_state.status =
5349 bf_get(lpfc_acqe_link_status, acqe_link);
James Smart70f3c072010-12-15 17:57:33 -05005350 phba->sli4_hba.link_state.type =
5351 bf_get(lpfc_acqe_link_type, acqe_link);
5352 phba->sli4_hba.link_state.number =
5353 bf_get(lpfc_acqe_link_number, acqe_link);
James Smartda0436e2009-05-22 14:51:39 -04005354 phba->sli4_hba.link_state.fault =
5355 bf_get(lpfc_acqe_link_fault, acqe_link);
James Smart65467b62010-01-26 23:08:29 -05005356 phba->sli4_hba.link_state.logical_speed =
James Smart8b68cd52012-09-29 11:32:37 -04005357 bf_get(lpfc_acqe_logical_link_speed, acqe_link) * 10;
5358
James Smart70f3c072010-12-15 17:57:33 -05005359 lpfc_printf_log(phba, KERN_INFO, LOG_SLI,
James Smartc31098c2011-04-16 11:03:33 -04005360 "2900 Async FC/FCoE Link event - Speed:%dGBit "
5361 "duplex:x%x LA Type:x%x Port Type:%d Port Number:%d "
5362 "Logical speed:%dMbps Fault:%d\n",
James Smart70f3c072010-12-15 17:57:33 -05005363 phba->sli4_hba.link_state.speed,
5364 phba->sli4_hba.link_state.topology,
5365 phba->sli4_hba.link_state.status,
5366 phba->sli4_hba.link_state.type,
5367 phba->sli4_hba.link_state.number,
James Smart8b68cd52012-09-29 11:32:37 -04005368 phba->sli4_hba.link_state.logical_speed,
James Smart70f3c072010-12-15 17:57:33 -05005369 phba->sli4_hba.link_state.fault);
James Smart76a95d72010-11-20 23:11:48 -05005370 /*
5371 * For FC Mode: issue the READ_TOPOLOGY mailbox command to fetch
5372 * topology info. Note: Optional for non FC-AL ports.
5373 */
5374 if (!(phba->hba_flag & HBA_FCOE_MODE)) {
5375 rc = lpfc_sli_issue_mbox(phba, pmb, MBX_NOWAIT);
5376 if (rc == MBX_NOT_FINISHED)
5377 goto out_free_dmabuf;
5378 return;
5379 }
5380 /*
5381 * For FCoE Mode: fill in all the topology information we need and call
5382 * the READ_TOPOLOGY completion routine to continue without actually
5383 * sending the READ_TOPOLOGY mailbox command to the port.
5384 */
James Smart23288b72018-05-04 20:37:53 -07005385 /* Initialize completion status */
James Smart76a95d72010-11-20 23:11:48 -05005386 mb = &pmb->u.mb;
James Smart23288b72018-05-04 20:37:53 -07005387 mb->mbxStatus = MBX_SUCCESS;
5388
5389 /* Parse port fault information field */
5390 lpfc_sli4_parse_latt_fault(phba, acqe_link);
James Smart76a95d72010-11-20 23:11:48 -05005391
5392 /* Parse and translate link attention fields */
5393 la = (struct lpfc_mbx_read_top *) &pmb->u.mb.un.varReadTop;
5394 la->eventTag = acqe_link->event_tag;
5395 bf_set(lpfc_mbx_read_top_att_type, la, att_type);
5396 bf_set(lpfc_mbx_read_top_link_spd, la,
James Smarta085e872015-12-16 18:12:02 -05005397 (bf_get(lpfc_acqe_link_speed, acqe_link)));
James Smart76a95d72010-11-20 23:11:48 -05005398
5399 /* Fake the the following irrelvant fields */
5400 bf_set(lpfc_mbx_read_top_topology, la, LPFC_TOPOLOGY_PT_PT);
5401 bf_set(lpfc_mbx_read_top_alpa_granted, la, 0);
5402 bf_set(lpfc_mbx_read_top_il, la, 0);
5403 bf_set(lpfc_mbx_read_top_pb, la, 0);
5404 bf_set(lpfc_mbx_read_top_fa, la, 0);
5405 bf_set(lpfc_mbx_read_top_mm, la, 0);
James Smartda0436e2009-05-22 14:51:39 -04005406
5407 /* Invoke the lpfc_handle_latt mailbox command callback function */
James Smart76a95d72010-11-20 23:11:48 -05005408 lpfc_mbx_cmpl_read_topology(phba, pmb);
James Smartda0436e2009-05-22 14:51:39 -04005409
5410 return;
5411
5412out_free_dmabuf:
5413 kfree(mp);
5414out_free_pmb:
5415 mempool_free(pmb, phba->mbox_mem_pool);
5416}
5417
5418/**
James Smart1dc5ec22018-10-23 13:41:11 -07005419 * lpfc_async_link_speed_to_read_top - Parse async evt link speed code to read
5420 * topology.
5421 * @phba: pointer to lpfc hba data structure.
James Smart1dc5ec22018-10-23 13:41:11 -07005422 * @speed_code: asynchronous event link speed code.
5423 *
5424 * This routine is to parse the giving SLI4 async event link speed code into
5425 * value of Read topology link speed.
5426 *
5427 * Return: link speed in terms of Read topology.
5428 **/
5429static uint8_t
5430lpfc_async_link_speed_to_read_top(struct lpfc_hba *phba, uint8_t speed_code)
5431{
5432 uint8_t port_speed;
5433
5434 switch (speed_code) {
5435 case LPFC_FC_LA_SPEED_1G:
5436 port_speed = LPFC_LINK_SPEED_1GHZ;
5437 break;
5438 case LPFC_FC_LA_SPEED_2G:
5439 port_speed = LPFC_LINK_SPEED_2GHZ;
5440 break;
5441 case LPFC_FC_LA_SPEED_4G:
5442 port_speed = LPFC_LINK_SPEED_4GHZ;
5443 break;
5444 case LPFC_FC_LA_SPEED_8G:
5445 port_speed = LPFC_LINK_SPEED_8GHZ;
5446 break;
5447 case LPFC_FC_LA_SPEED_16G:
5448 port_speed = LPFC_LINK_SPEED_16GHZ;
5449 break;
5450 case LPFC_FC_LA_SPEED_32G:
5451 port_speed = LPFC_LINK_SPEED_32GHZ;
5452 break;
5453 case LPFC_FC_LA_SPEED_64G:
5454 port_speed = LPFC_LINK_SPEED_64GHZ;
5455 break;
5456 case LPFC_FC_LA_SPEED_128G:
5457 port_speed = LPFC_LINK_SPEED_128GHZ;
5458 break;
5459 case LPFC_FC_LA_SPEED_256G:
5460 port_speed = LPFC_LINK_SPEED_256GHZ;
5461 break;
5462 default:
5463 port_speed = 0;
5464 break;
5465 }
5466
5467 return port_speed;
5468}
5469
James Smart74a7baa2021-08-16 09:28:58 -07005470void
5471lpfc_cgn_dump_rxmonitor(struct lpfc_hba *phba)
5472{
5473 struct rxtable_entry *entry;
5474 int cnt = 0, head, tail, last, start;
5475
5476 head = atomic_read(&phba->rxtable_idx_head);
5477 tail = atomic_read(&phba->rxtable_idx_tail);
5478 if (!phba->rxtable || head == tail) {
5479 lpfc_printf_log(phba, KERN_ERR, LOG_CGN_MGMT,
5480 "4411 Rxtable is empty\n");
5481 return;
5482 }
5483 last = tail;
5484 start = head;
5485
5486 /* Display the last LPFC_MAX_RXMONITOR_DUMP entries from the rxtable */
5487 while (start != last) {
5488 if (start)
5489 start--;
5490 else
5491 start = LPFC_MAX_RXMONITOR_ENTRY - 1;
5492 entry = &phba->rxtable[start];
5493 lpfc_printf_log(phba, KERN_INFO, LOG_CGN_MGMT,
5494 "4410 %02d: MBPI %lld Xmit %lld Cmpl %lld "
5495 "Lat %lld ASz %lld Info %02d BWUtil %d "
5496 "Int %d slot %d\n",
5497 cnt, entry->max_bytes_per_interval,
5498 entry->total_bytes, entry->rcv_bytes,
5499 entry->avg_io_latency, entry->avg_io_size,
5500 entry->cmf_info, entry->timer_utilization,
5501 entry->timer_interval, start);
5502 cnt++;
5503 if (cnt >= LPFC_MAX_RXMONITOR_DUMP)
5504 return;
5505 }
5506}
5507
James Smart02243832021-08-16 09:28:54 -07005508/**
James Smart7481811c2021-08-16 09:28:56 -07005509 * lpfc_cgn_update_stat - Save data into congestion stats buffer
5510 * @phba: pointer to lpfc hba data structure.
5511 * @dtag: FPIN descriptor received
5512 *
5513 * Increment the FPIN received counter/time when it happens.
5514 */
5515void
5516lpfc_cgn_update_stat(struct lpfc_hba *phba, uint32_t dtag)
5517{
5518 struct lpfc_cgn_info *cp;
5519 struct tm broken;
5520 struct timespec64 cur_time;
5521 u32 cnt;
5522 u16 value;
5523
5524 /* Make sure we have a congestion info buffer */
5525 if (!phba->cgn_i)
5526 return;
5527 cp = (struct lpfc_cgn_info *)phba->cgn_i->virt;
5528 ktime_get_real_ts64(&cur_time);
5529 time64_to_tm(cur_time.tv_sec, 0, &broken);
5530
5531 /* Update congestion statistics */
5532 switch (dtag) {
5533 case ELS_DTAG_LNK_INTEGRITY:
5534 cnt = le32_to_cpu(cp->link_integ_notification);
5535 cnt++;
5536 cp->link_integ_notification = cpu_to_le32(cnt);
5537
5538 cp->cgn_stat_lnk_month = broken.tm_mon + 1;
5539 cp->cgn_stat_lnk_day = broken.tm_mday;
5540 cp->cgn_stat_lnk_year = broken.tm_year - 100;
5541 cp->cgn_stat_lnk_hour = broken.tm_hour;
5542 cp->cgn_stat_lnk_min = broken.tm_min;
5543 cp->cgn_stat_lnk_sec = broken.tm_sec;
5544 break;
5545 case ELS_DTAG_DELIVERY:
5546 cnt = le32_to_cpu(cp->delivery_notification);
5547 cnt++;
5548 cp->delivery_notification = cpu_to_le32(cnt);
5549
5550 cp->cgn_stat_del_month = broken.tm_mon + 1;
5551 cp->cgn_stat_del_day = broken.tm_mday;
5552 cp->cgn_stat_del_year = broken.tm_year - 100;
5553 cp->cgn_stat_del_hour = broken.tm_hour;
5554 cp->cgn_stat_del_min = broken.tm_min;
5555 cp->cgn_stat_del_sec = broken.tm_sec;
5556 break;
5557 case ELS_DTAG_PEER_CONGEST:
5558 cnt = le32_to_cpu(cp->cgn_peer_notification);
5559 cnt++;
5560 cp->cgn_peer_notification = cpu_to_le32(cnt);
5561
5562 cp->cgn_stat_peer_month = broken.tm_mon + 1;
5563 cp->cgn_stat_peer_day = broken.tm_mday;
5564 cp->cgn_stat_peer_year = broken.tm_year - 100;
5565 cp->cgn_stat_peer_hour = broken.tm_hour;
5566 cp->cgn_stat_peer_min = broken.tm_min;
5567 cp->cgn_stat_peer_sec = broken.tm_sec;
5568 break;
5569 case ELS_DTAG_CONGESTION:
5570 cnt = le32_to_cpu(cp->cgn_notification);
5571 cnt++;
5572 cp->cgn_notification = cpu_to_le32(cnt);
5573
5574 cp->cgn_stat_cgn_month = broken.tm_mon + 1;
5575 cp->cgn_stat_cgn_day = broken.tm_mday;
5576 cp->cgn_stat_cgn_year = broken.tm_year - 100;
5577 cp->cgn_stat_cgn_hour = broken.tm_hour;
5578 cp->cgn_stat_cgn_min = broken.tm_min;
5579 cp->cgn_stat_cgn_sec = broken.tm_sec;
5580 }
5581 if (phba->cgn_fpin_frequency &&
5582 phba->cgn_fpin_frequency != LPFC_FPIN_INIT_FREQ) {
5583 value = LPFC_CGN_TIMER_TO_MIN / phba->cgn_fpin_frequency;
James Smart59936432021-08-30 16:12:43 -07005584 cp->cgn_stat_npm = value;
James Smart7481811c2021-08-16 09:28:56 -07005585 }
5586 value = lpfc_cgn_calc_crc32(cp, LPFC_CGN_INFO_SZ,
5587 LPFC_CGN_CRC32_SEED);
5588 cp->cgn_info_crc = cpu_to_le32(value);
5589}
5590
5591/**
5592 * lpfc_cgn_save_evt_cnt - Save data into registered congestion buffer
5593 * @phba: pointer to lpfc hba data structure.
5594 *
5595 * Save the congestion event data every minute.
5596 * On the hour collapse all the minute data into hour data. Every day
5597 * collapse all the hour data into daily data. Separate driver
5598 * and fabrc congestion event counters that will be saved out
5599 * to the registered congestion buffer every minute.
5600 */
5601static void
5602lpfc_cgn_save_evt_cnt(struct lpfc_hba *phba)
5603{
5604 struct lpfc_cgn_info *cp;
5605 struct tm broken;
5606 struct timespec64 cur_time;
5607 uint32_t i, index;
5608 uint16_t value, mvalue;
5609 uint64_t bps;
5610 uint32_t mbps;
5611 uint32_t dvalue, wvalue, lvalue, avalue;
5612 uint64_t latsum;
James Smart59936432021-08-30 16:12:43 -07005613 __le16 *ptr;
5614 __le32 *lptr;
5615 __le16 *mptr;
James Smart7481811c2021-08-16 09:28:56 -07005616
5617 /* Make sure we have a congestion info buffer */
5618 if (!phba->cgn_i)
5619 return;
5620 cp = (struct lpfc_cgn_info *)phba->cgn_i->virt;
5621
5622 if (time_before(jiffies, phba->cgn_evt_timestamp))
5623 return;
5624 phba->cgn_evt_timestamp = jiffies +
5625 msecs_to_jiffies(LPFC_CGN_TIMER_TO_MIN);
5626 phba->cgn_evt_minute++;
5627
5628 /* We should get to this point in the routine on 1 minute intervals */
5629
5630 ktime_get_real_ts64(&cur_time);
5631 time64_to_tm(cur_time.tv_sec, 0, &broken);
5632
5633 if (phba->cgn_fpin_frequency &&
5634 phba->cgn_fpin_frequency != LPFC_FPIN_INIT_FREQ) {
5635 value = LPFC_CGN_TIMER_TO_MIN / phba->cgn_fpin_frequency;
James Smart59936432021-08-30 16:12:43 -07005636 cp->cgn_stat_npm = value;
James Smart7481811c2021-08-16 09:28:56 -07005637 }
5638
5639 /* Read and clear the latency counters for this minute */
5640 lvalue = atomic_read(&phba->cgn_latency_evt_cnt);
5641 latsum = atomic64_read(&phba->cgn_latency_evt);
5642 atomic_set(&phba->cgn_latency_evt_cnt, 0);
5643 atomic64_set(&phba->cgn_latency_evt, 0);
5644
5645 /* We need to store MB/sec bandwidth in the congestion information.
5646 * block_cnt is count of 512 byte blocks for the entire minute,
5647 * bps will get bytes per sec before finally converting to MB/sec.
5648 */
5649 bps = div_u64(phba->rx_block_cnt, LPFC_SEC_MIN) * 512;
5650 phba->rx_block_cnt = 0;
5651 mvalue = bps / (1024 * 1024); /* convert to MB/sec */
5652
5653 /* Every minute */
5654 /* cgn parameters */
5655 cp->cgn_info_mode = phba->cgn_p.cgn_param_mode;
5656 cp->cgn_info_level0 = phba->cgn_p.cgn_param_level0;
5657 cp->cgn_info_level1 = phba->cgn_p.cgn_param_level1;
5658 cp->cgn_info_level2 = phba->cgn_p.cgn_param_level2;
5659
5660 /* Fill in default LUN qdepth */
5661 value = (uint16_t)(phba->pport->cfg_lun_queue_depth);
5662 cp->cgn_lunq = cpu_to_le16(value);
5663
5664 /* Record congestion buffer info - every minute
5665 * cgn_driver_evt_cnt (Driver events)
5666 * cgn_fabric_warn_cnt (Congestion Warnings)
5667 * cgn_latency_evt_cnt / cgn_latency_evt (IO Latency)
5668 * cgn_fabric_alarm_cnt (Congestion Alarms)
5669 */
5670 index = ++cp->cgn_index_minute;
5671 if (cp->cgn_index_minute == LPFC_MIN_HOUR) {
5672 cp->cgn_index_minute = 0;
5673 index = 0;
5674 }
5675
5676 /* Get the number of driver events in this sample and reset counter */
5677 dvalue = atomic_read(&phba->cgn_driver_evt_cnt);
5678 atomic_set(&phba->cgn_driver_evt_cnt, 0);
5679
5680 /* Get the number of warning events - FPIN and Signal for this minute */
5681 wvalue = 0;
5682 if ((phba->cgn_reg_fpin & LPFC_CGN_FPIN_WARN) ||
5683 phba->cgn_reg_signal == EDC_CG_SIG_WARN_ONLY ||
5684 phba->cgn_reg_signal == EDC_CG_SIG_WARN_ALARM)
5685 wvalue = atomic_read(&phba->cgn_fabric_warn_cnt);
5686 atomic_set(&phba->cgn_fabric_warn_cnt, 0);
5687
5688 /* Get the number of alarm events - FPIN and Signal for this minute */
5689 avalue = 0;
5690 if ((phba->cgn_reg_fpin & LPFC_CGN_FPIN_ALARM) ||
5691 phba->cgn_reg_signal == EDC_CG_SIG_WARN_ALARM)
5692 avalue = atomic_read(&phba->cgn_fabric_alarm_cnt);
5693 atomic_set(&phba->cgn_fabric_alarm_cnt, 0);
5694
5695 /* Collect the driver, warning, alarm and latency counts for this
5696 * minute into the driver congestion buffer.
5697 */
5698 ptr = &cp->cgn_drvr_min[index];
5699 value = (uint16_t)dvalue;
5700 *ptr = cpu_to_le16(value);
5701
5702 ptr = &cp->cgn_warn_min[index];
5703 value = (uint16_t)wvalue;
5704 *ptr = cpu_to_le16(value);
5705
5706 ptr = &cp->cgn_alarm_min[index];
5707 value = (uint16_t)avalue;
5708 *ptr = cpu_to_le16(value);
5709
5710 lptr = &cp->cgn_latency_min[index];
5711 if (lvalue) {
5712 lvalue = (uint32_t)div_u64(latsum, lvalue);
5713 *lptr = cpu_to_le32(lvalue);
5714 } else {
5715 *lptr = 0;
5716 }
5717
5718 /* Collect the bandwidth value into the driver's congesion buffer. */
5719 mptr = &cp->cgn_bw_min[index];
5720 *mptr = cpu_to_le16(mvalue);
5721
5722 lpfc_printf_log(phba, KERN_INFO, LOG_CGN_MGMT,
5723 "2418 Congestion Info - minute (%d): %d %d %d %d %d\n",
5724 index, dvalue, wvalue, *lptr, mvalue, avalue);
5725
5726 /* Every hour */
5727 if ((phba->cgn_evt_minute % LPFC_MIN_HOUR) == 0) {
5728 /* Record congestion buffer info - every hour
5729 * Collapse all minutes into an hour
5730 */
5731 index = ++cp->cgn_index_hour;
5732 if (cp->cgn_index_hour == LPFC_HOUR_DAY) {
5733 cp->cgn_index_hour = 0;
5734 index = 0;
5735 }
5736
5737 dvalue = 0;
5738 wvalue = 0;
5739 lvalue = 0;
5740 avalue = 0;
5741 mvalue = 0;
5742 mbps = 0;
5743 for (i = 0; i < LPFC_MIN_HOUR; i++) {
5744 dvalue += le16_to_cpu(cp->cgn_drvr_min[i]);
5745 wvalue += le16_to_cpu(cp->cgn_warn_min[i]);
5746 lvalue += le32_to_cpu(cp->cgn_latency_min[i]);
5747 mbps += le16_to_cpu(cp->cgn_bw_min[i]);
5748 avalue += le16_to_cpu(cp->cgn_alarm_min[i]);
5749 }
5750 if (lvalue) /* Avg of latency averages */
5751 lvalue /= LPFC_MIN_HOUR;
5752 if (mbps) /* Avg of Bandwidth averages */
5753 mvalue = mbps / LPFC_MIN_HOUR;
5754
5755 lptr = &cp->cgn_drvr_hr[index];
5756 *lptr = cpu_to_le32(dvalue);
5757 lptr = &cp->cgn_warn_hr[index];
5758 *lptr = cpu_to_le32(wvalue);
5759 lptr = &cp->cgn_latency_hr[index];
5760 *lptr = cpu_to_le32(lvalue);
5761 mptr = &cp->cgn_bw_hr[index];
5762 *mptr = cpu_to_le16(mvalue);
5763 lptr = &cp->cgn_alarm_hr[index];
5764 *lptr = cpu_to_le32(avalue);
5765
5766 lpfc_printf_log(phba, KERN_INFO, LOG_CGN_MGMT,
5767 "2419 Congestion Info - hour "
5768 "(%d): %d %d %d %d %d\n",
5769 index, dvalue, wvalue, lvalue, mvalue, avalue);
5770 }
5771
5772 /* Every day */
5773 if ((phba->cgn_evt_minute % LPFC_MIN_DAY) == 0) {
5774 /* Record congestion buffer info - every hour
5775 * Collapse all hours into a day. Rotate days
5776 * after LPFC_MAX_CGN_DAYS.
5777 */
5778 index = ++cp->cgn_index_day;
5779 if (cp->cgn_index_day == LPFC_MAX_CGN_DAYS) {
5780 cp->cgn_index_day = 0;
5781 index = 0;
5782 }
5783
5784 /* Anytime we overwrite daily index 0, after we wrap,
5785 * we will be overwriting the oldest day, so we must
5786 * update the congestion data start time for that day.
5787 * That start time should have previously been saved after
5788 * we wrote the last days worth of data.
5789 */
5790 if ((phba->hba_flag & HBA_CGN_DAY_WRAP) && index == 0) {
5791 time64_to_tm(phba->cgn_daily_ts.tv_sec, 0, &broken);
5792
5793 cp->cgn_info_month = broken.tm_mon + 1;
5794 cp->cgn_info_day = broken.tm_mday;
5795 cp->cgn_info_year = broken.tm_year - 100;
5796 cp->cgn_info_hour = broken.tm_hour;
5797 cp->cgn_info_minute = broken.tm_min;
5798 cp->cgn_info_second = broken.tm_sec;
5799
5800 lpfc_printf_log
5801 (phba, KERN_INFO, LOG_CGN_MGMT,
5802 "2646 CGNInfo idx0 Start Time: "
5803 "%d/%d/%d %d:%d:%d\n",
5804 cp->cgn_info_day, cp->cgn_info_month,
5805 cp->cgn_info_year, cp->cgn_info_hour,
5806 cp->cgn_info_minute, cp->cgn_info_second);
5807 }
5808
5809 dvalue = 0;
5810 wvalue = 0;
5811 lvalue = 0;
5812 mvalue = 0;
5813 mbps = 0;
5814 avalue = 0;
5815 for (i = 0; i < LPFC_HOUR_DAY; i++) {
5816 dvalue += le32_to_cpu(cp->cgn_drvr_hr[i]);
5817 wvalue += le32_to_cpu(cp->cgn_warn_hr[i]);
5818 lvalue += le32_to_cpu(cp->cgn_latency_hr[i]);
James Smart59936432021-08-30 16:12:43 -07005819 mbps += le16_to_cpu(cp->cgn_bw_hr[i]);
James Smart7481811c2021-08-16 09:28:56 -07005820 avalue += le32_to_cpu(cp->cgn_alarm_hr[i]);
5821 }
5822 if (lvalue) /* Avg of latency averages */
5823 lvalue /= LPFC_HOUR_DAY;
5824 if (mbps) /* Avg of Bandwidth averages */
5825 mvalue = mbps / LPFC_HOUR_DAY;
5826
5827 lptr = &cp->cgn_drvr_day[index];
5828 *lptr = cpu_to_le32(dvalue);
5829 lptr = &cp->cgn_warn_day[index];
5830 *lptr = cpu_to_le32(wvalue);
5831 lptr = &cp->cgn_latency_day[index];
5832 *lptr = cpu_to_le32(lvalue);
5833 mptr = &cp->cgn_bw_day[index];
5834 *mptr = cpu_to_le16(mvalue);
5835 lptr = &cp->cgn_alarm_day[index];
5836 *lptr = cpu_to_le32(avalue);
5837
5838 lpfc_printf_log(phba, KERN_INFO, LOG_CGN_MGMT,
5839 "2420 Congestion Info - daily (%d): "
5840 "%d %d %d %d %d\n",
5841 index, dvalue, wvalue, lvalue, mvalue, avalue);
5842
5843 /* We just wrote LPFC_MAX_CGN_DAYS of data,
5844 * so we are wrapped on any data after this.
5845 * Save this as the start time for the next day.
5846 */
5847 if (index == (LPFC_MAX_CGN_DAYS - 1)) {
5848 phba->hba_flag |= HBA_CGN_DAY_WRAP;
5849 ktime_get_real_ts64(&phba->cgn_daily_ts);
5850 }
5851 }
5852
5853 /* Use the frequency found in the last rcv'ed FPIN */
5854 value = phba->cgn_fpin_frequency;
5855 if (phba->cgn_reg_fpin & LPFC_CGN_FPIN_WARN)
5856 cp->cgn_warn_freq = cpu_to_le16(value);
5857 if (phba->cgn_reg_fpin & LPFC_CGN_FPIN_ALARM)
5858 cp->cgn_alarm_freq = cpu_to_le16(value);
5859
5860 /* Frequency (in ms) Signal Warning/Signal Congestion Notifications
5861 * are received by the HBA
5862 */
5863 value = phba->cgn_sig_freq;
5864
5865 if (phba->cgn_reg_signal == EDC_CG_SIG_WARN_ONLY ||
5866 phba->cgn_reg_signal == EDC_CG_SIG_WARN_ALARM)
5867 cp->cgn_warn_freq = cpu_to_le16(value);
5868 if (phba->cgn_reg_signal == EDC_CG_SIG_WARN_ALARM)
5869 cp->cgn_alarm_freq = cpu_to_le16(value);
5870
5871 lvalue = lpfc_cgn_calc_crc32(cp, LPFC_CGN_INFO_SZ,
5872 LPFC_CGN_CRC32_SEED);
5873 cp->cgn_info_crc = cpu_to_le32(lvalue);
5874}
5875
5876/**
James Smart02243832021-08-16 09:28:54 -07005877 * lpfc_calc_cmf_latency - latency from start of rxate timer interval
5878 * @phba: The Hba for which this call is being executed.
5879 *
5880 * The routine calculates the latency from the beginning of the CMF timer
5881 * interval to the current point in time. It is called from IO completion
5882 * when we exceed our Bandwidth limitation for the time interval.
5883 */
5884uint32_t
5885lpfc_calc_cmf_latency(struct lpfc_hba *phba)
5886{
5887 struct timespec64 cmpl_time;
5888 uint32_t msec = 0;
5889
5890 ktime_get_real_ts64(&cmpl_time);
5891
5892 /* This routine works on a ms granularity so sec and usec are
5893 * converted accordingly.
5894 */
5895 if (cmpl_time.tv_sec == phba->cmf_latency.tv_sec) {
5896 msec = (cmpl_time.tv_nsec - phba->cmf_latency.tv_nsec) /
5897 NSEC_PER_MSEC;
5898 } else {
5899 if (cmpl_time.tv_nsec >= phba->cmf_latency.tv_nsec) {
5900 msec = (cmpl_time.tv_sec -
5901 phba->cmf_latency.tv_sec) * MSEC_PER_SEC;
5902 msec += ((cmpl_time.tv_nsec -
5903 phba->cmf_latency.tv_nsec) / NSEC_PER_MSEC);
5904 } else {
5905 msec = (cmpl_time.tv_sec - phba->cmf_latency.tv_sec -
5906 1) * MSEC_PER_SEC;
5907 msec += (((NSEC_PER_SEC - phba->cmf_latency.tv_nsec) +
5908 cmpl_time.tv_nsec) / NSEC_PER_MSEC);
5909 }
5910 }
5911 return msec;
5912}
5913
5914/**
5915 * lpfc_cmf_timer - This is the timer function for one congestion
5916 * rate interval.
5917 * @timer: Pointer to the high resolution timer that expired
5918 */
5919static enum hrtimer_restart
5920lpfc_cmf_timer(struct hrtimer *timer)
5921{
5922 struct lpfc_hba *phba = container_of(timer, struct lpfc_hba,
5923 cmf_timer);
James Smart17b27ac2021-08-16 09:28:55 -07005924 struct rxtable_entry *entry;
James Smart02243832021-08-16 09:28:54 -07005925 uint32_t io_cnt;
James Smart17b27ac2021-08-16 09:28:55 -07005926 uint32_t head, tail;
5927 uint32_t busy, max_read;
James Smartd5ac69b2021-09-10 16:31:55 -07005928 uint64_t total, rcv, lat, mbpi, extra;
James Smart02243832021-08-16 09:28:54 -07005929 int timer_interval = LPFC_CMF_INTERVAL;
James Smart17b27ac2021-08-16 09:28:55 -07005930 uint32_t ms;
James Smart02243832021-08-16 09:28:54 -07005931 struct lpfc_cgn_stat *cgs;
5932 int cpu;
5933
5934 /* Only restart the timer if congestion mgmt is on */
5935 if (phba->cmf_active_mode == LPFC_CFG_OFF ||
5936 !phba->cmf_latency.tv_sec) {
5937 lpfc_printf_log(phba, KERN_INFO, LOG_CGN_MGMT,
5938 "6224 CMF timer exit: %d %lld\n",
5939 phba->cmf_active_mode,
5940 (uint64_t)phba->cmf_latency.tv_sec);
5941 return HRTIMER_NORESTART;
5942 }
5943
5944 /* If pport is not ready yet, just exit and wait for
5945 * the next timer cycle to hit.
5946 */
5947 if (!phba->pport)
5948 goto skip;
5949
5950 /* Do not block SCSI IO while in the timer routine since
5951 * total_bytes will be cleared
5952 */
5953 atomic_set(&phba->cmf_stop_io, 1);
5954
James Smart17b27ac2021-08-16 09:28:55 -07005955 /* First we need to calculate the actual ms between
5956 * the last timer interrupt and this one. We ask for
5957 * LPFC_CMF_INTERVAL, however the actual time may
5958 * vary depending on system overhead.
5959 */
5960 ms = lpfc_calc_cmf_latency(phba);
5961
5962
James Smart02243832021-08-16 09:28:54 -07005963 /* Immediately after we calculate the time since the last
5964 * timer interrupt, set the start time for the next
5965 * interrupt
5966 */
5967 ktime_get_real_ts64(&phba->cmf_latency);
5968
5969 phba->cmf_link_byte_count =
5970 div_u64(phba->cmf_max_line_rate * LPFC_CMF_INTERVAL, 1000);
5971
5972 /* Collect all the stats from the prior timer interval */
5973 total = 0;
5974 io_cnt = 0;
5975 lat = 0;
5976 rcv = 0;
5977 for_each_present_cpu(cpu) {
5978 cgs = per_cpu_ptr(phba->cmf_stat, cpu);
5979 total += atomic64_xchg(&cgs->total_bytes, 0);
5980 io_cnt += atomic_xchg(&cgs->rx_io_cnt, 0);
5981 lat += atomic64_xchg(&cgs->rx_latency, 0);
5982 rcv += atomic64_xchg(&cgs->rcv_bytes, 0);
5983 }
5984
5985 /* Before we issue another CMF_SYNC_WQE, retrieve the BW
5986 * returned from the last CMF_SYNC_WQE issued, from
5987 * cmf_last_sync_bw. This will be the target BW for
5988 * this next timer interval.
5989 */
5990 if (phba->cmf_active_mode == LPFC_CFG_MANAGED &&
5991 phba->link_state != LPFC_LINK_DOWN &&
5992 phba->hba_flag & HBA_SETUP) {
5993 mbpi = phba->cmf_last_sync_bw;
5994 phba->cmf_last_sync_bw = 0;
James Smartd5ac69b2021-09-10 16:31:55 -07005995 extra = 0;
5996
5997 /* Calculate any extra bytes needed to account for the
5998 * timer accuracy. If we are less than LPFC_CMF_INTERVAL
5999 * add an extra 3% slop factor, equal to LPFC_CMF_INTERVAL
6000 * add an extra 2%. The goal is to equalize total with a
6001 * time > LPFC_CMF_INTERVAL or <= LPFC_CMF_INTERVAL + 1
6002 */
6003 if (ms == LPFC_CMF_INTERVAL)
6004 extra = div_u64(total, 50);
6005 else if (ms < LPFC_CMF_INTERVAL)
6006 extra = div_u64(total, 33);
6007 lpfc_issue_cmf_sync_wqe(phba, LPFC_CMF_INTERVAL, total + extra);
James Smart02243832021-08-16 09:28:54 -07006008 } else {
6009 /* For Monitor mode or link down we want mbpi
6010 * to be the full link speed
6011 */
6012 mbpi = phba->cmf_link_byte_count;
6013 }
6014 phba->cmf_timer_cnt++;
6015
6016 if (io_cnt) {
6017 /* Update congestion info buffer latency in us */
6018 atomic_add(io_cnt, &phba->cgn_latency_evt_cnt);
6019 atomic64_add(lat, &phba->cgn_latency_evt);
6020 }
James Smart17b27ac2021-08-16 09:28:55 -07006021 busy = atomic_xchg(&phba->cmf_busy, 0);
6022 max_read = atomic_xchg(&phba->rx_max_read_cnt, 0);
James Smart02243832021-08-16 09:28:54 -07006023
6024 /* Calculate MBPI for the next timer interval */
6025 if (mbpi) {
6026 if (mbpi > phba->cmf_link_byte_count ||
6027 phba->cmf_active_mode == LPFC_CFG_MONITOR)
6028 mbpi = phba->cmf_link_byte_count;
6029
6030 /* Change max_bytes_per_interval to what the prior
6031 * CMF_SYNC_WQE cmpl indicated.
6032 */
6033 if (mbpi != phba->cmf_max_bytes_per_interval)
6034 phba->cmf_max_bytes_per_interval = mbpi;
6035 }
6036
James Smart17b27ac2021-08-16 09:28:55 -07006037 /* Save rxmonitor information for debug */
6038 if (phba->rxtable) {
6039 head = atomic_xchg(&phba->rxtable_idx_head,
6040 LPFC_RXMONITOR_TABLE_IN_USE);
6041 entry = &phba->rxtable[head];
6042 entry->total_bytes = total;
6043 entry->rcv_bytes = rcv;
6044 entry->cmf_busy = busy;
6045 entry->cmf_info = phba->cmf_active_info;
6046 if (io_cnt) {
6047 entry->avg_io_latency = div_u64(lat, io_cnt);
6048 entry->avg_io_size = div_u64(rcv, io_cnt);
6049 } else {
6050 entry->avg_io_latency = 0;
6051 entry->avg_io_size = 0;
6052 }
6053 entry->max_read_cnt = max_read;
6054 entry->io_cnt = io_cnt;
6055 entry->max_bytes_per_interval = mbpi;
6056 if (phba->cmf_active_mode == LPFC_CFG_MANAGED)
6057 entry->timer_utilization = phba->cmf_last_ts;
6058 else
6059 entry->timer_utilization = ms;
6060 entry->timer_interval = ms;
6061 phba->cmf_last_ts = 0;
6062
6063 /* Increment rxtable index */
6064 head = (head + 1) % LPFC_MAX_RXMONITOR_ENTRY;
6065 tail = atomic_read(&phba->rxtable_idx_tail);
6066 if (head == tail) {
6067 tail = (tail + 1) % LPFC_MAX_RXMONITOR_ENTRY;
6068 atomic_set(&phba->rxtable_idx_tail, tail);
6069 }
6070 atomic_set(&phba->rxtable_idx_head, head);
6071 }
6072
James Smart02243832021-08-16 09:28:54 -07006073 if (phba->cmf_active_mode == LPFC_CFG_MONITOR) {
6074 /* If Monitor mode, check if we are oversubscribed
6075 * against the full line rate.
6076 */
6077 if (mbpi && total > mbpi)
6078 atomic_inc(&phba->cgn_driver_evt_cnt);
6079 }
6080 phba->rx_block_cnt += div_u64(rcv, 512); /* save 512 byte block cnt */
6081
James Smart7481811c2021-08-16 09:28:56 -07006082 /* Each minute save Fabric and Driver congestion information */
6083 lpfc_cgn_save_evt_cnt(phba);
6084
6085 /* Since we need to call lpfc_cgn_save_evt_cnt every minute, on the
6086 * minute, adjust our next timer interval, if needed, to ensure a
6087 * 1 minute granularity when we get the next timer interrupt.
6088 */
6089 if (time_after(jiffies + msecs_to_jiffies(LPFC_CMF_INTERVAL),
6090 phba->cgn_evt_timestamp)) {
6091 timer_interval = jiffies_to_msecs(phba->cgn_evt_timestamp -
6092 jiffies);
6093 if (timer_interval <= 0)
6094 timer_interval = LPFC_CMF_INTERVAL;
6095
6096 /* If we adjust timer_interval, max_bytes_per_interval
6097 * needs to be adjusted as well.
6098 */
6099 phba->cmf_link_byte_count = div_u64(phba->cmf_max_line_rate *
6100 timer_interval, 1000);
6101 if (phba->cmf_active_mode == LPFC_CFG_MONITOR)
6102 phba->cmf_max_bytes_per_interval =
6103 phba->cmf_link_byte_count;
6104 }
6105
James Smart02243832021-08-16 09:28:54 -07006106 /* Since total_bytes has already been zero'ed, its okay to unblock
6107 * after max_bytes_per_interval is setup.
6108 */
6109 if (atomic_xchg(&phba->cmf_bw_wait, 0))
6110 queue_work(phba->wq, &phba->unblock_request_work);
6111
6112 /* SCSI IO is now unblocked */
6113 atomic_set(&phba->cmf_stop_io, 0);
6114
6115skip:
6116 hrtimer_forward_now(timer,
6117 ktime_set(0, timer_interval * NSEC_PER_MSEC));
6118 return HRTIMER_RESTART;
6119}
6120
James Smart1dc5ec22018-10-23 13:41:11 -07006121#define trunk_link_status(__idx)\
6122 bf_get(lpfc_acqe_fc_la_trunk_config_port##__idx, acqe_fc) ?\
6123 ((phba->trunk_link.link##__idx.state == LPFC_LINK_UP) ?\
6124 "Link up" : "Link down") : "NA"
6125/* Did port __idx reported an error */
6126#define trunk_port_fault(__idx)\
6127 bf_get(lpfc_acqe_fc_la_trunk_config_port##__idx, acqe_fc) ?\
6128 (port_fault & (1 << __idx) ? "YES" : "NO") : "NA"
6129
6130static void
6131lpfc_update_trunk_link_status(struct lpfc_hba *phba,
6132 struct lpfc_acqe_fc_la *acqe_fc)
6133{
6134 uint8_t port_fault = bf_get(lpfc_acqe_fc_la_trunk_linkmask, acqe_fc);
6135 uint8_t err = bf_get(lpfc_acqe_fc_la_trunk_fault, acqe_fc);
6136
6137 phba->sli4_hba.link_state.speed =
6138 lpfc_sli4_port_speed_parse(phba, LPFC_TRAILER_CODE_FC,
6139 bf_get(lpfc_acqe_fc_la_speed, acqe_fc));
6140
6141 phba->sli4_hba.link_state.logical_speed =
James Smartb8e6f132019-05-21 17:49:04 -07006142 bf_get(lpfc_acqe_fc_la_llink_spd, acqe_fc) * 10;
James Smart1dc5ec22018-10-23 13:41:11 -07006143 /* We got FC link speed, convert to fc_linkspeed (READ_TOPOLOGY) */
6144 phba->fc_linkspeed =
6145 lpfc_async_link_speed_to_read_top(
6146 phba,
6147 bf_get(lpfc_acqe_fc_la_speed, acqe_fc));
6148
6149 if (bf_get(lpfc_acqe_fc_la_trunk_config_port0, acqe_fc)) {
6150 phba->trunk_link.link0.state =
6151 bf_get(lpfc_acqe_fc_la_trunk_link_status_port0, acqe_fc)
6152 ? LPFC_LINK_UP : LPFC_LINK_DOWN;
James Smart529b3dd2018-12-13 15:17:54 -08006153 phba->trunk_link.link0.fault = port_fault & 0x1 ? err : 0;
James Smart1dc5ec22018-10-23 13:41:11 -07006154 }
6155 if (bf_get(lpfc_acqe_fc_la_trunk_config_port1, acqe_fc)) {
6156 phba->trunk_link.link1.state =
6157 bf_get(lpfc_acqe_fc_la_trunk_link_status_port1, acqe_fc)
6158 ? LPFC_LINK_UP : LPFC_LINK_DOWN;
James Smart529b3dd2018-12-13 15:17:54 -08006159 phba->trunk_link.link1.fault = port_fault & 0x2 ? err : 0;
James Smart1dc5ec22018-10-23 13:41:11 -07006160 }
6161 if (bf_get(lpfc_acqe_fc_la_trunk_config_port2, acqe_fc)) {
6162 phba->trunk_link.link2.state =
6163 bf_get(lpfc_acqe_fc_la_trunk_link_status_port2, acqe_fc)
6164 ? LPFC_LINK_UP : LPFC_LINK_DOWN;
James Smart529b3dd2018-12-13 15:17:54 -08006165 phba->trunk_link.link2.fault = port_fault & 0x4 ? err : 0;
James Smart1dc5ec22018-10-23 13:41:11 -07006166 }
6167 if (bf_get(lpfc_acqe_fc_la_trunk_config_port3, acqe_fc)) {
6168 phba->trunk_link.link3.state =
6169 bf_get(lpfc_acqe_fc_la_trunk_link_status_port3, acqe_fc)
6170 ? LPFC_LINK_UP : LPFC_LINK_DOWN;
James Smart529b3dd2018-12-13 15:17:54 -08006171 phba->trunk_link.link3.fault = port_fault & 0x8 ? err : 0;
James Smart1dc5ec22018-10-23 13:41:11 -07006172 }
6173
Dick Kennedy372c1872020-06-30 14:50:00 -07006174 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart1dc5ec22018-10-23 13:41:11 -07006175 "2910 Async FC Trunking Event - Speed:%d\n"
6176 "\tLogical speed:%d "
6177 "port0: %s port1: %s port2: %s port3: %s\n",
6178 phba->sli4_hba.link_state.speed,
6179 phba->sli4_hba.link_state.logical_speed,
6180 trunk_link_status(0), trunk_link_status(1),
6181 trunk_link_status(2), trunk_link_status(3));
6182
James Smart02243832021-08-16 09:28:54 -07006183 if (phba->cmf_active_mode != LPFC_CFG_OFF)
6184 lpfc_cmf_signal_init(phba);
6185
James Smart1dc5ec22018-10-23 13:41:11 -07006186 if (port_fault)
Dick Kennedy372c1872020-06-30 14:50:00 -07006187 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart1dc5ec22018-10-23 13:41:11 -07006188 "3202 trunk error:0x%x (%s) seen on port0:%s "
6189 /*
6190 * SLI-4: We have only 0xA error codes
6191 * defined as of now. print an appropriate
6192 * message in case driver needs to be updated.
6193 */
6194 "port1:%s port2:%s port3:%s\n", err, err > 0xA ?
6195 "UNDEFINED. update driver." : trunk_errmsg[err],
6196 trunk_port_fault(0), trunk_port_fault(1),
6197 trunk_port_fault(2), trunk_port_fault(3));
6198}
6199
6200
6201/**
James Smart70f3c072010-12-15 17:57:33 -05006202 * lpfc_sli4_async_fc_evt - Process the asynchronous FC link event
6203 * @phba: pointer to lpfc hba data structure.
6204 * @acqe_fc: pointer to the async fc completion queue entry.
6205 *
6206 * This routine is to handle the SLI4 asynchronous FC event. It will simply log
6207 * that the event was received and then issue a read_topology mailbox command so
6208 * that the rest of the driver will treat it the same as SLI3.
6209 **/
6210static void
6211lpfc_sli4_async_fc_evt(struct lpfc_hba *phba, struct lpfc_acqe_fc_la *acqe_fc)
6212{
6213 struct lpfc_dmabuf *mp;
6214 LPFC_MBOXQ_t *pmb;
James Smart7bdedb32016-07-06 12:36:00 -07006215 MAILBOX_t *mb;
6216 struct lpfc_mbx_read_top *la;
James Smart70f3c072010-12-15 17:57:33 -05006217 int rc;
6218
6219 if (bf_get(lpfc_trailer_type, acqe_fc) !=
6220 LPFC_FC_LA_EVENT_TYPE_FC_LINK) {
Dick Kennedy372c1872020-06-30 14:50:00 -07006221 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart70f3c072010-12-15 17:57:33 -05006222 "2895 Non FC link Event detected.(%d)\n",
6223 bf_get(lpfc_trailer_type, acqe_fc));
6224 return;
6225 }
James Smart1dc5ec22018-10-23 13:41:11 -07006226
6227 if (bf_get(lpfc_acqe_fc_la_att_type, acqe_fc) ==
6228 LPFC_FC_LA_TYPE_TRUNKING_EVENT) {
6229 lpfc_update_trunk_link_status(phba, acqe_fc);
6230 return;
6231 }
6232
James Smart70f3c072010-12-15 17:57:33 -05006233 /* Keep the link status for extra SLI4 state machine reference */
6234 phba->sli4_hba.link_state.speed =
James Smart8b68cd52012-09-29 11:32:37 -04006235 lpfc_sli4_port_speed_parse(phba, LPFC_TRAILER_CODE_FC,
6236 bf_get(lpfc_acqe_fc_la_speed, acqe_fc));
James Smart70f3c072010-12-15 17:57:33 -05006237 phba->sli4_hba.link_state.duplex = LPFC_ASYNC_LINK_DUPLEX_FULL;
6238 phba->sli4_hba.link_state.topology =
6239 bf_get(lpfc_acqe_fc_la_topology, acqe_fc);
6240 phba->sli4_hba.link_state.status =
6241 bf_get(lpfc_acqe_fc_la_att_type, acqe_fc);
6242 phba->sli4_hba.link_state.type =
6243 bf_get(lpfc_acqe_fc_la_port_type, acqe_fc);
6244 phba->sli4_hba.link_state.number =
6245 bf_get(lpfc_acqe_fc_la_port_number, acqe_fc);
6246 phba->sli4_hba.link_state.fault =
6247 bf_get(lpfc_acqe_link_fault, acqe_fc);
James Smartb8e6f132019-05-21 17:49:04 -07006248
6249 if (bf_get(lpfc_acqe_fc_la_att_type, acqe_fc) ==
6250 LPFC_FC_LA_TYPE_LINK_DOWN)
6251 phba->sli4_hba.link_state.logical_speed = 0;
6252 else if (!phba->sli4_hba.conf_trunk)
6253 phba->sli4_hba.link_state.logical_speed =
James Smart8b68cd52012-09-29 11:32:37 -04006254 bf_get(lpfc_acqe_fc_la_llink_spd, acqe_fc) * 10;
James Smartb8e6f132019-05-21 17:49:04 -07006255
James Smart70f3c072010-12-15 17:57:33 -05006256 lpfc_printf_log(phba, KERN_INFO, LOG_SLI,
6257 "2896 Async FC event - Speed:%dGBaud Topology:x%x "
6258 "LA Type:x%x Port Type:%d Port Number:%d Logical speed:"
6259 "%dMbps Fault:%d\n",
6260 phba->sli4_hba.link_state.speed,
6261 phba->sli4_hba.link_state.topology,
6262 phba->sli4_hba.link_state.status,
6263 phba->sli4_hba.link_state.type,
6264 phba->sli4_hba.link_state.number,
James Smart8b68cd52012-09-29 11:32:37 -04006265 phba->sli4_hba.link_state.logical_speed,
James Smart70f3c072010-12-15 17:57:33 -05006266 phba->sli4_hba.link_state.fault);
6267 pmb = (LPFC_MBOXQ_t *)mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL);
6268 if (!pmb) {
Dick Kennedy372c1872020-06-30 14:50:00 -07006269 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart70f3c072010-12-15 17:57:33 -05006270 "2897 The mboxq allocation failed\n");
6271 return;
6272 }
6273 mp = kmalloc(sizeof(struct lpfc_dmabuf), GFP_KERNEL);
6274 if (!mp) {
Dick Kennedy372c1872020-06-30 14:50:00 -07006275 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart70f3c072010-12-15 17:57:33 -05006276 "2898 The lpfc_dmabuf allocation failed\n");
6277 goto out_free_pmb;
6278 }
6279 mp->virt = lpfc_mbuf_alloc(phba, 0, &mp->phys);
6280 if (!mp->virt) {
Dick Kennedy372c1872020-06-30 14:50:00 -07006281 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart70f3c072010-12-15 17:57:33 -05006282 "2899 The mbuf allocation failed\n");
6283 goto out_free_dmabuf;
6284 }
6285
6286 /* Cleanup any outstanding ELS commands */
6287 lpfc_els_flush_all_cmd(phba);
6288
6289 /* Block ELS IOCBs until we have done process link event */
James Smart895427b2017-02-12 13:52:30 -08006290 phba->sli4_hba.els_wq->pring->flag |= LPFC_STOP_IOCB_EVENT;
James Smart70f3c072010-12-15 17:57:33 -05006291
6292 /* Update link event statistics */
6293 phba->sli.slistat.link_event++;
6294
6295 /* Create lpfc_handle_latt mailbox command from link ACQE */
6296 lpfc_read_topology(phba, pmb, mp);
6297 pmb->mbox_cmpl = lpfc_mbx_cmpl_read_topology;
6298 pmb->vport = phba->pport;
6299
James Smart7bdedb32016-07-06 12:36:00 -07006300 if (phba->sli4_hba.link_state.status != LPFC_FC_LA_TYPE_LINK_UP) {
James Smartae9e28f2017-05-15 15:20:51 -07006301 phba->link_flag &= ~(LS_MDS_LINK_DOWN | LS_MDS_LOOPBACK);
6302
6303 switch (phba->sli4_hba.link_state.status) {
6304 case LPFC_FC_LA_TYPE_MDS_LINK_DOWN:
6305 phba->link_flag |= LS_MDS_LINK_DOWN;
6306 break;
6307 case LPFC_FC_LA_TYPE_MDS_LOOPBACK:
6308 phba->link_flag |= LS_MDS_LOOPBACK;
6309 break;
6310 default:
6311 break;
6312 }
6313
James Smart23288b72018-05-04 20:37:53 -07006314 /* Initialize completion status */
James Smart7bdedb32016-07-06 12:36:00 -07006315 mb = &pmb->u.mb;
James Smart23288b72018-05-04 20:37:53 -07006316 mb->mbxStatus = MBX_SUCCESS;
6317
6318 /* Parse port fault information field */
6319 lpfc_sli4_parse_latt_fault(phba, (void *)acqe_fc);
James Smart7bdedb32016-07-06 12:36:00 -07006320
6321 /* Parse and translate link attention fields */
6322 la = (struct lpfc_mbx_read_top *)&pmb->u.mb.un.varReadTop;
6323 la->eventTag = acqe_fc->event_tag;
James Smart7bdedb32016-07-06 12:36:00 -07006324
James Smartaeb3c812017-04-21 16:05:02 -07006325 if (phba->sli4_hba.link_state.status ==
6326 LPFC_FC_LA_TYPE_UNEXP_WWPN) {
6327 bf_set(lpfc_mbx_read_top_att_type, la,
6328 LPFC_FC_LA_TYPE_UNEXP_WWPN);
6329 } else {
6330 bf_set(lpfc_mbx_read_top_att_type, la,
6331 LPFC_FC_LA_TYPE_LINK_DOWN);
6332 }
James Smart7bdedb32016-07-06 12:36:00 -07006333 /* Invoke the mailbox command callback function */
6334 lpfc_mbx_cmpl_read_topology(phba, pmb);
6335
6336 return;
6337 }
6338
James Smart70f3c072010-12-15 17:57:33 -05006339 rc = lpfc_sli_issue_mbox(phba, pmb, MBX_NOWAIT);
6340 if (rc == MBX_NOT_FINISHED)
6341 goto out_free_dmabuf;
6342 return;
6343
6344out_free_dmabuf:
6345 kfree(mp);
6346out_free_pmb:
6347 mempool_free(pmb, phba->mbox_mem_pool);
6348}
6349
6350/**
6351 * lpfc_sli4_async_sli_evt - Process the asynchronous SLI link event
6352 * @phba: pointer to lpfc hba data structure.
Lee Jonesfe614ac2020-07-23 13:24:22 +01006353 * @acqe_sli: pointer to the async SLI completion queue entry.
James Smart70f3c072010-12-15 17:57:33 -05006354 *
6355 * This routine is to handle the SLI4 asynchronous SLI events.
6356 **/
6357static void
6358lpfc_sli4_async_sli_evt(struct lpfc_hba *phba, struct lpfc_acqe_sli *acqe_sli)
6359{
James Smart4b8bae02012-06-12 13:55:07 -04006360 char port_name;
James Smart8c1312e2012-10-31 14:45:09 -04006361 char message[128];
James Smart4b8bae02012-06-12 13:55:07 -04006362 uint8_t status;
James Smart946727d2015-04-07 15:07:09 -04006363 uint8_t evt_type;
James Smart448193b2015-12-16 18:12:05 -05006364 uint8_t operational = 0;
James Smart946727d2015-04-07 15:07:09 -04006365 struct temp_event temp_event_data;
James Smart4b8bae02012-06-12 13:55:07 -04006366 struct lpfc_acqe_misconfigured_event *misconfigured;
James Smart9064aeb2021-08-16 09:28:50 -07006367 struct lpfc_acqe_cgn_signal *cgn_signal;
James Smart946727d2015-04-07 15:07:09 -04006368 struct Scsi_Host *shost;
James Smartcd713482018-10-23 13:41:01 -07006369 struct lpfc_vport **vports;
James Smart9064aeb2021-08-16 09:28:50 -07006370 int rc, i, cnt;
James Smart4b8bae02012-06-12 13:55:07 -04006371
James Smart946727d2015-04-07 15:07:09 -04006372 evt_type = bf_get(lpfc_trailer_type, acqe_sli);
6373
James Smart448193b2015-12-16 18:12:05 -05006374 lpfc_printf_log(phba, KERN_INFO, LOG_SLI,
James Smartd11ed162019-09-21 20:59:03 -07006375 "2901 Async SLI event - Type:%d, Event Data: x%08x "
6376 "x%08x x%08x x%08x\n", evt_type,
James Smart448193b2015-12-16 18:12:05 -05006377 acqe_sli->event_data1, acqe_sli->event_data2,
James Smartd11ed162019-09-21 20:59:03 -07006378 acqe_sli->reserved, acqe_sli->trailer);
James Smart4b8bae02012-06-12 13:55:07 -04006379
6380 port_name = phba->Port[0];
6381 if (port_name == 0x00)
6382 port_name = '?'; /* get port name is empty */
6383
James Smart946727d2015-04-07 15:07:09 -04006384 switch (evt_type) {
6385 case LPFC_SLI_EVENT_TYPE_OVER_TEMP:
6386 temp_event_data.event_type = FC_REG_TEMPERATURE_EVENT;
6387 temp_event_data.event_code = LPFC_THRESHOLD_TEMP;
6388 temp_event_data.data = (uint32_t)acqe_sli->event_data1;
6389
6390 lpfc_printf_log(phba, KERN_WARNING, LOG_SLI,
6391 "3190 Over Temperature:%d Celsius- Port Name %c\n",
6392 acqe_sli->event_data1, port_name);
6393
James Smart310429e2016-07-06 12:35:54 -07006394 phba->sfp_warning |= LPFC_TRANSGRESSION_HIGH_TEMPERATURE;
James Smart946727d2015-04-07 15:07:09 -04006395 shost = lpfc_shost_from_vport(phba->pport);
6396 fc_host_post_vendor_event(shost, fc_get_event_number(),
6397 sizeof(temp_event_data),
6398 (char *)&temp_event_data,
6399 SCSI_NL_VID_TYPE_PCI
6400 | PCI_VENDOR_ID_EMULEX);
6401 break;
6402 case LPFC_SLI_EVENT_TYPE_NORM_TEMP:
6403 temp_event_data.event_type = FC_REG_TEMPERATURE_EVENT;
6404 temp_event_data.event_code = LPFC_NORMAL_TEMP;
6405 temp_event_data.data = (uint32_t)acqe_sli->event_data1;
6406
6407 lpfc_printf_log(phba, KERN_INFO, LOG_SLI,
6408 "3191 Normal Temperature:%d Celsius - Port Name %c\n",
6409 acqe_sli->event_data1, port_name);
6410
6411 shost = lpfc_shost_from_vport(phba->pport);
6412 fc_host_post_vendor_event(shost, fc_get_event_number(),
6413 sizeof(temp_event_data),
6414 (char *)&temp_event_data,
6415 SCSI_NL_VID_TYPE_PCI
6416 | PCI_VENDOR_ID_EMULEX);
6417 break;
6418 case LPFC_SLI_EVENT_TYPE_MISCONFIGURED:
6419 misconfigured = (struct lpfc_acqe_misconfigured_event *)
James Smart4b8bae02012-06-12 13:55:07 -04006420 &acqe_sli->event_data1;
6421
James Smart946727d2015-04-07 15:07:09 -04006422 /* fetch the status for this port */
6423 switch (phba->sli4_hba.lnk_info.lnk_no) {
6424 case LPFC_LINK_NUMBER_0:
James Smart448193b2015-12-16 18:12:05 -05006425 status = bf_get(lpfc_sli_misconfigured_port0_state,
6426 &misconfigured->theEvent);
6427 operational = bf_get(lpfc_sli_misconfigured_port0_op,
James Smart4b8bae02012-06-12 13:55:07 -04006428 &misconfigured->theEvent);
James Smart946727d2015-04-07 15:07:09 -04006429 break;
6430 case LPFC_LINK_NUMBER_1:
James Smart448193b2015-12-16 18:12:05 -05006431 status = bf_get(lpfc_sli_misconfigured_port1_state,
6432 &misconfigured->theEvent);
6433 operational = bf_get(lpfc_sli_misconfigured_port1_op,
James Smart4b8bae02012-06-12 13:55:07 -04006434 &misconfigured->theEvent);
James Smart946727d2015-04-07 15:07:09 -04006435 break;
6436 case LPFC_LINK_NUMBER_2:
James Smart448193b2015-12-16 18:12:05 -05006437 status = bf_get(lpfc_sli_misconfigured_port2_state,
6438 &misconfigured->theEvent);
6439 operational = bf_get(lpfc_sli_misconfigured_port2_op,
James Smart4b8bae02012-06-12 13:55:07 -04006440 &misconfigured->theEvent);
James Smart946727d2015-04-07 15:07:09 -04006441 break;
6442 case LPFC_LINK_NUMBER_3:
James Smart448193b2015-12-16 18:12:05 -05006443 status = bf_get(lpfc_sli_misconfigured_port3_state,
6444 &misconfigured->theEvent);
6445 operational = bf_get(lpfc_sli_misconfigured_port3_op,
James Smart4b8bae02012-06-12 13:55:07 -04006446 &misconfigured->theEvent);
James Smart946727d2015-04-07 15:07:09 -04006447 break;
6448 default:
Dick Kennedy372c1872020-06-30 14:50:00 -07006449 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart448193b2015-12-16 18:12:05 -05006450 "3296 "
6451 "LPFC_SLI_EVENT_TYPE_MISCONFIGURED "
6452 "event: Invalid link %d",
6453 phba->sli4_hba.lnk_info.lnk_no);
6454 return;
James Smart946727d2015-04-07 15:07:09 -04006455 }
James Smart4b8bae02012-06-12 13:55:07 -04006456
James Smart448193b2015-12-16 18:12:05 -05006457 /* Skip if optic state unchanged */
6458 if (phba->sli4_hba.lnk_info.optic_state == status)
6459 return;
6460
James Smart946727d2015-04-07 15:07:09 -04006461 switch (status) {
6462 case LPFC_SLI_EVENT_STATUS_VALID:
James Smart448193b2015-12-16 18:12:05 -05006463 sprintf(message, "Physical Link is functional");
6464 break;
James Smart946727d2015-04-07 15:07:09 -04006465 case LPFC_SLI_EVENT_STATUS_NOT_PRESENT:
6466 sprintf(message, "Optics faulted/incorrectly "
6467 "installed/not installed - Reseat optics, "
6468 "if issue not resolved, replace.");
6469 break;
6470 case LPFC_SLI_EVENT_STATUS_WRONG_TYPE:
6471 sprintf(message,
6472 "Optics of two types installed - Remove one "
6473 "optic or install matching pair of optics.");
6474 break;
6475 case LPFC_SLI_EVENT_STATUS_UNSUPPORTED:
6476 sprintf(message, "Incompatible optics - Replace with "
James Smart292098b2012-09-29 11:31:41 -04006477 "compatible optics for card to function.");
James Smart946727d2015-04-07 15:07:09 -04006478 break;
James Smart448193b2015-12-16 18:12:05 -05006479 case LPFC_SLI_EVENT_STATUS_UNQUALIFIED:
6480 sprintf(message, "Unqualified optics - Replace with "
6481 "Avago optics for Warranty and Technical "
6482 "Support - Link is%s operational",
James Smart2ea259e2017-02-12 13:52:27 -08006483 (operational) ? " not" : "");
James Smart448193b2015-12-16 18:12:05 -05006484 break;
6485 case LPFC_SLI_EVENT_STATUS_UNCERTIFIED:
6486 sprintf(message, "Uncertified optics - Replace with "
6487 "Avago-certified optics to enable link "
6488 "operation - Link is%s operational",
James Smart2ea259e2017-02-12 13:52:27 -08006489 (operational) ? " not" : "");
James Smart448193b2015-12-16 18:12:05 -05006490 break;
James Smart946727d2015-04-07 15:07:09 -04006491 default:
6492 /* firmware is reporting a status we don't know about */
6493 sprintf(message, "Unknown event status x%02x", status);
6494 break;
6495 }
James Smartcd713482018-10-23 13:41:01 -07006496
6497 /* Issue READ_CONFIG mbox command to refresh supported speeds */
6498 rc = lpfc_sli4_read_config(phba);
James Smart3952e912018-10-23 13:41:02 -07006499 if (rc) {
James Smartcd713482018-10-23 13:41:01 -07006500 phba->lmt = 0;
Dick Kennedy372c1872020-06-30 14:50:00 -07006501 lpfc_printf_log(phba, KERN_ERR,
6502 LOG_TRACE_EVENT,
James Smartcd713482018-10-23 13:41:01 -07006503 "3194 Unable to retrieve supported "
James Smart3952e912018-10-23 13:41:02 -07006504 "speeds, rc = 0x%x\n", rc);
James Smartcd713482018-10-23 13:41:01 -07006505 }
James Smart7a1dda92021-10-20 14:14:12 -07006506 rc = lpfc_sli4_refresh_params(phba);
6507 if (rc) {
6508 lpfc_printf_log(phba, KERN_ERR, LOG_SLI,
6509 "3174 Unable to update pls support, "
6510 "rc x%x\n", rc);
6511 }
James Smartcd713482018-10-23 13:41:01 -07006512 vports = lpfc_create_vport_work_array(phba);
6513 if (vports != NULL) {
6514 for (i = 0; i <= phba->max_vports && vports[i] != NULL;
6515 i++) {
6516 shost = lpfc_shost_from_vport(vports[i]);
6517 lpfc_host_supported_speeds_set(shost);
6518 }
6519 }
6520 lpfc_destroy_vport_work_array(phba, vports);
6521
James Smart448193b2015-12-16 18:12:05 -05006522 phba->sli4_hba.lnk_info.optic_state = status;
James Smart946727d2015-04-07 15:07:09 -04006523 lpfc_printf_log(phba, KERN_ERR, LOG_SLI,
James Smart448193b2015-12-16 18:12:05 -05006524 "3176 Port Name %c %s\n", port_name, message);
James Smart946727d2015-04-07 15:07:09 -04006525 break;
6526 case LPFC_SLI_EVENT_TYPE_REMOTE_DPORT:
6527 lpfc_printf_log(phba, KERN_INFO, LOG_SLI,
6528 "3192 Remote DPort Test Initiated - "
6529 "Event Data1:x%08x Event Data2: x%08x\n",
6530 acqe_sli->event_data1, acqe_sli->event_data2);
James Smart4b8bae02012-06-12 13:55:07 -04006531 break;
James Smart02243832021-08-16 09:28:54 -07006532 case LPFC_SLI_EVENT_TYPE_PORT_PARAMS_CHG:
6533 /* Call FW to obtain active parms */
6534 lpfc_sli4_cgn_parm_chg_evt(phba);
6535 break;
James Smarte7d85952019-10-18 14:18:29 -07006536 case LPFC_SLI_EVENT_TYPE_MISCONF_FAWWN:
6537 /* Misconfigured WWN. Reports that the SLI Port is configured
6538 * to use FA-WWN, but the attached device doesn’t support it.
6539 * No driver action is required.
6540 * Event Data1 - N.A, Event Data2 - N.A
6541 */
6542 lpfc_log_msg(phba, KERN_WARNING, LOG_SLI,
6543 "2699 Misconfigured FA-WWN - Attached device does "
6544 "not support FA-WWN\n");
6545 break;
James Smartd11ed162019-09-21 20:59:03 -07006546 case LPFC_SLI_EVENT_TYPE_EEPROM_FAILURE:
6547 /* EEPROM failure. No driver action is required */
6548 lpfc_printf_log(phba, KERN_WARNING, LOG_SLI,
6549 "2518 EEPROM failure - "
6550 "Event Data1: x%08x Event Data2: x%08x\n",
6551 acqe_sli->event_data1, acqe_sli->event_data2);
6552 break;
James Smart9064aeb2021-08-16 09:28:50 -07006553 case LPFC_SLI_EVENT_TYPE_CGN_SIGNAL:
6554 if (phba->cmf_active_mode == LPFC_CFG_OFF)
6555 break;
6556 cgn_signal = (struct lpfc_acqe_cgn_signal *)
6557 &acqe_sli->event_data1;
6558 phba->cgn_acqe_cnt++;
6559
6560 cnt = bf_get(lpfc_warn_acqe, cgn_signal);
6561 atomic64_add(cnt, &phba->cgn_acqe_stat.warn);
6562 atomic64_add(cgn_signal->alarm_cnt, &phba->cgn_acqe_stat.alarm);
6563
6564 /* no threshold for CMF, even 1 signal will trigger an event */
6565
6566 /* Alarm overrides warning, so check that first */
6567 if (cgn_signal->alarm_cnt) {
6568 if (phba->cgn_reg_signal == EDC_CG_SIG_WARN_ALARM) {
6569 /* Keep track of alarm cnt for cgn_info */
6570 atomic_add(cgn_signal->alarm_cnt,
6571 &phba->cgn_fabric_alarm_cnt);
6572 /* Keep track of alarm cnt for CMF_SYNC_WQE */
6573 atomic_add(cgn_signal->alarm_cnt,
6574 &phba->cgn_sync_alarm_cnt);
6575 }
6576 } else if (cnt) {
6577 /* signal action needs to be taken */
6578 if (phba->cgn_reg_signal == EDC_CG_SIG_WARN_ONLY ||
6579 phba->cgn_reg_signal == EDC_CG_SIG_WARN_ALARM) {
6580 /* Keep track of warning cnt for cgn_info */
6581 atomic_add(cnt, &phba->cgn_fabric_warn_cnt);
6582 /* Keep track of warning cnt for CMF_SYNC_WQE */
6583 atomic_add(cnt, &phba->cgn_sync_warn_cnt);
6584 }
6585 }
6586 break;
James Smart4b8bae02012-06-12 13:55:07 -04006587 default:
James Smart946727d2015-04-07 15:07:09 -04006588 lpfc_printf_log(phba, KERN_INFO, LOG_SLI,
James Smartd11ed162019-09-21 20:59:03 -07006589 "3193 Unrecognized SLI event, type: 0x%x",
James Smart946727d2015-04-07 15:07:09 -04006590 evt_type);
James Smart4b8bae02012-06-12 13:55:07 -04006591 break;
6592 }
James Smart70f3c072010-12-15 17:57:33 -05006593}
6594
6595/**
James Smartfc2b9892010-02-26 14:15:29 -05006596 * lpfc_sli4_perform_vport_cvl - Perform clear virtual link on a vport
6597 * @vport: pointer to vport data structure.
6598 *
6599 * This routine is to perform Clear Virtual Link (CVL) on a vport in
6600 * response to a CVL event.
6601 *
6602 * Return the pointer to the ndlp with the vport if successful, otherwise
6603 * return NULL.
6604 **/
6605static struct lpfc_nodelist *
6606lpfc_sli4_perform_vport_cvl(struct lpfc_vport *vport)
6607{
6608 struct lpfc_nodelist *ndlp;
6609 struct Scsi_Host *shost;
6610 struct lpfc_hba *phba;
6611
6612 if (!vport)
6613 return NULL;
James Smartfc2b9892010-02-26 14:15:29 -05006614 phba = vport->phba;
6615 if (!phba)
6616 return NULL;
James Smart78730cf2010-04-06 15:06:30 -04006617 ndlp = lpfc_findnode_did(vport, Fabric_DID);
6618 if (!ndlp) {
6619 /* Cannot find existing Fabric ndlp, so allocate a new one */
James Smart9d3d3402017-04-21 16:05:00 -07006620 ndlp = lpfc_nlp_init(vport, Fabric_DID);
James Smart78730cf2010-04-06 15:06:30 -04006621 if (!ndlp)
Colin Ian King5860d9f2021-09-25 23:41:13 +01006622 return NULL;
James Smart78730cf2010-04-06 15:06:30 -04006623 /* Set the node type */
6624 ndlp->nlp_type |= NLP_FABRIC;
6625 /* Put ndlp onto node list */
6626 lpfc_enqueue_node(vport, ndlp);
James Smart78730cf2010-04-06 15:06:30 -04006627 }
James Smart63e801c2010-11-20 23:14:19 -05006628 if ((phba->pport->port_state < LPFC_FLOGI) &&
6629 (phba->pport->port_state != LPFC_VPORT_FAILED))
James Smartfc2b9892010-02-26 14:15:29 -05006630 return NULL;
6631 /* If virtual link is not yet instantiated ignore CVL */
James Smart63e801c2010-11-20 23:14:19 -05006632 if ((vport != phba->pport) && (vport->port_state < LPFC_FDISC)
6633 && (vport->port_state != LPFC_VPORT_FAILED))
James Smartfc2b9892010-02-26 14:15:29 -05006634 return NULL;
6635 shost = lpfc_shost_from_vport(vport);
6636 if (!shost)
6637 return NULL;
6638 lpfc_linkdown_port(vport);
6639 lpfc_cleanup_pending_mbox(vport);
6640 spin_lock_irq(shost->host_lock);
6641 vport->fc_flag |= FC_VPORT_CVL_RCVD;
6642 spin_unlock_irq(shost->host_lock);
6643
6644 return ndlp;
6645}
6646
6647/**
6648 * lpfc_sli4_perform_all_vport_cvl - Perform clear virtual link on all vports
Lee Jonesfe614ac2020-07-23 13:24:22 +01006649 * @phba: pointer to lpfc hba data structure.
James Smartfc2b9892010-02-26 14:15:29 -05006650 *
6651 * This routine is to perform Clear Virtual Link (CVL) on all vports in
6652 * response to a FCF dead event.
6653 **/
6654static void
6655lpfc_sli4_perform_all_vport_cvl(struct lpfc_hba *phba)
6656{
6657 struct lpfc_vport **vports;
6658 int i;
6659
6660 vports = lpfc_create_vport_work_array(phba);
6661 if (vports)
6662 for (i = 0; i <= phba->max_vports && vports[i] != NULL; i++)
6663 lpfc_sli4_perform_vport_cvl(vports[i]);
6664 lpfc_destroy_vport_work_array(phba, vports);
6665}
6666
6667/**
James Smart76a95d72010-11-20 23:11:48 -05006668 * lpfc_sli4_async_fip_evt - Process the asynchronous FCoE FIP event
James Smartda0436e2009-05-22 14:51:39 -04006669 * @phba: pointer to lpfc hba data structure.
Lee Jonesfe614ac2020-07-23 13:24:22 +01006670 * @acqe_fip: pointer to the async fcoe completion queue entry.
James Smartda0436e2009-05-22 14:51:39 -04006671 *
6672 * This routine is to handle the SLI4 asynchronous fcoe event.
6673 **/
6674static void
James Smart76a95d72010-11-20 23:11:48 -05006675lpfc_sli4_async_fip_evt(struct lpfc_hba *phba,
James Smart70f3c072010-12-15 17:57:33 -05006676 struct lpfc_acqe_fip *acqe_fip)
James Smartda0436e2009-05-22 14:51:39 -04006677{
James Smart70f3c072010-12-15 17:57:33 -05006678 uint8_t event_type = bf_get(lpfc_trailer_type, acqe_fip);
James Smartda0436e2009-05-22 14:51:39 -04006679 int rc;
James Smart6669f9b2009-10-02 15:16:45 -04006680 struct lpfc_vport *vport;
6681 struct lpfc_nodelist *ndlp;
James Smart695a8142010-01-26 23:08:03 -05006682 int active_vlink_present;
6683 struct lpfc_vport **vports;
6684 int i;
James Smartda0436e2009-05-22 14:51:39 -04006685
James Smart70f3c072010-12-15 17:57:33 -05006686 phba->fc_eventTag = acqe_fip->event_tag;
6687 phba->fcoe_eventtag = acqe_fip->event_tag;
James Smartda0436e2009-05-22 14:51:39 -04006688 switch (event_type) {
James Smart70f3c072010-12-15 17:57:33 -05006689 case LPFC_FIP_EVENT_TYPE_NEW_FCF:
6690 case LPFC_FIP_EVENT_TYPE_FCF_PARAM_MOD:
6691 if (event_type == LPFC_FIP_EVENT_TYPE_NEW_FCF)
Dick Kennedy372c1872020-06-30 14:50:00 -07006692 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smarta93ff372010-10-22 11:06:08 -04006693 "2546 New FCF event, evt_tag:x%x, "
6694 "index:x%x\n",
James Smart70f3c072010-12-15 17:57:33 -05006695 acqe_fip->event_tag,
6696 acqe_fip->index);
James Smart999d8132010-03-15 11:24:56 -04006697 else
6698 lpfc_printf_log(phba, KERN_WARNING, LOG_FIP |
6699 LOG_DISCOVERY,
James Smarta93ff372010-10-22 11:06:08 -04006700 "2788 FCF param modified event, "
6701 "evt_tag:x%x, index:x%x\n",
James Smart70f3c072010-12-15 17:57:33 -05006702 acqe_fip->event_tag,
6703 acqe_fip->index);
James Smart38b92ef2010-08-04 16:11:39 -04006704 if (phba->fcf.fcf_flag & FCF_DISCOVERY) {
James Smart0c9ab6f2010-02-26 14:15:57 -05006705 /*
6706 * During period of FCF discovery, read the FCF
6707 * table record indexed by the event to update
James Smarta93ff372010-10-22 11:06:08 -04006708 * FCF roundrobin failover eligible FCF bmask.
James Smart0c9ab6f2010-02-26 14:15:57 -05006709 */
6710 lpfc_printf_log(phba, KERN_INFO, LOG_FIP |
6711 LOG_DISCOVERY,
James Smarta93ff372010-10-22 11:06:08 -04006712 "2779 Read FCF (x%x) for updating "
6713 "roundrobin FCF failover bmask\n",
James Smart70f3c072010-12-15 17:57:33 -05006714 acqe_fip->index);
6715 rc = lpfc_sli4_read_fcf_rec(phba, acqe_fip->index);
James Smart0c9ab6f2010-02-26 14:15:57 -05006716 }
James Smart38b92ef2010-08-04 16:11:39 -04006717
6718 /* If the FCF discovery is in progress, do nothing. */
James Smart3804dc82010-07-14 15:31:37 -04006719 spin_lock_irq(&phba->hbalock);
James Smarta93ff372010-10-22 11:06:08 -04006720 if (phba->hba_flag & FCF_TS_INPROG) {
James Smart38b92ef2010-08-04 16:11:39 -04006721 spin_unlock_irq(&phba->hbalock);
6722 break;
6723 }
6724 /* If fast FCF failover rescan event is pending, do nothing */
James Smart036cad12018-10-23 13:41:06 -07006725 if (phba->fcf.fcf_flag & (FCF_REDISC_EVT | FCF_REDISC_PEND)) {
James Smart38b92ef2010-08-04 16:11:39 -04006726 spin_unlock_irq(&phba->hbalock);
6727 break;
6728 }
6729
James Smartc2b97122013-05-31 17:05:36 -04006730 /* If the FCF has been in discovered state, do nothing. */
6731 if (phba->fcf.fcf_flag & FCF_SCAN_DONE) {
James Smart3804dc82010-07-14 15:31:37 -04006732 spin_unlock_irq(&phba->hbalock);
6733 break;
6734 }
6735 spin_unlock_irq(&phba->hbalock);
James Smart38b92ef2010-08-04 16:11:39 -04006736
James Smart0c9ab6f2010-02-26 14:15:57 -05006737 /* Otherwise, scan the entire FCF table and re-discover SAN */
6738 lpfc_printf_log(phba, KERN_INFO, LOG_FIP | LOG_DISCOVERY,
James Smarta93ff372010-10-22 11:06:08 -04006739 "2770 Start FCF table scan per async FCF "
6740 "event, evt_tag:x%x, index:x%x\n",
James Smart70f3c072010-12-15 17:57:33 -05006741 acqe_fip->event_tag, acqe_fip->index);
James Smart0c9ab6f2010-02-26 14:15:57 -05006742 rc = lpfc_sli4_fcf_scan_read_fcf_rec(phba,
6743 LPFC_FCOE_FCF_GET_FIRST);
James Smartda0436e2009-05-22 14:51:39 -04006744 if (rc)
Dick Kennedy372c1872020-06-30 14:50:00 -07006745 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart0c9ab6f2010-02-26 14:15:57 -05006746 "2547 Issue FCF scan read FCF mailbox "
James Smarta93ff372010-10-22 11:06:08 -04006747 "command failed (x%x)\n", rc);
James Smartda0436e2009-05-22 14:51:39 -04006748 break;
6749
James Smart70f3c072010-12-15 17:57:33 -05006750 case LPFC_FIP_EVENT_TYPE_FCF_TABLE_FULL:
Dick Kennedy372c1872020-06-30 14:50:00 -07006751 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
6752 "2548 FCF Table full count 0x%x tag 0x%x\n",
6753 bf_get(lpfc_acqe_fip_fcf_count, acqe_fip),
6754 acqe_fip->event_tag);
James Smartda0436e2009-05-22 14:51:39 -04006755 break;
6756
James Smart70f3c072010-12-15 17:57:33 -05006757 case LPFC_FIP_EVENT_TYPE_FCF_DEAD:
James Smart80c17842012-03-01 22:35:45 -05006758 phba->fcoe_cvl_eventtag = acqe_fip->event_tag;
Dick Kennedy372c1872020-06-30 14:50:00 -07006759 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
6760 "2549 FCF (x%x) disconnected from network, "
6761 "tag:x%x\n", acqe_fip->index,
6762 acqe_fip->event_tag);
James Smart38b92ef2010-08-04 16:11:39 -04006763 /*
6764 * If we are in the middle of FCF failover process, clear
6765 * the corresponding FCF bit in the roundrobin bitmap.
James Smartda0436e2009-05-22 14:51:39 -04006766 */
James Smartfc2b9892010-02-26 14:15:29 -05006767 spin_lock_irq(&phba->hbalock);
James Smarta1cadfe2016-07-06 12:36:02 -07006768 if ((phba->fcf.fcf_flag & FCF_DISCOVERY) &&
6769 (phba->fcf.current_rec.fcf_indx != acqe_fip->index)) {
James Smartfc2b9892010-02-26 14:15:29 -05006770 spin_unlock_irq(&phba->hbalock);
James Smart0c9ab6f2010-02-26 14:15:57 -05006771 /* Update FLOGI FCF failover eligible FCF bmask */
James Smart70f3c072010-12-15 17:57:33 -05006772 lpfc_sli4_fcf_rr_index_clear(phba, acqe_fip->index);
James Smartfc2b9892010-02-26 14:15:29 -05006773 break;
6774 }
James Smart38b92ef2010-08-04 16:11:39 -04006775 spin_unlock_irq(&phba->hbalock);
6776
6777 /* If the event is not for currently used fcf do nothing */
James Smart70f3c072010-12-15 17:57:33 -05006778 if (phba->fcf.current_rec.fcf_indx != acqe_fip->index)
James Smart38b92ef2010-08-04 16:11:39 -04006779 break;
6780
6781 /*
6782 * Otherwise, request the port to rediscover the entire FCF
6783 * table for a fast recovery from case that the current FCF
6784 * is no longer valid as we are not in the middle of FCF
6785 * failover process already.
6786 */
James Smartc2b97122013-05-31 17:05:36 -04006787 spin_lock_irq(&phba->hbalock);
6788 /* Mark the fast failover process in progress */
6789 phba->fcf.fcf_flag |= FCF_DEAD_DISC;
6790 spin_unlock_irq(&phba->hbalock);
6791
6792 lpfc_printf_log(phba, KERN_INFO, LOG_FIP | LOG_DISCOVERY,
6793 "2771 Start FCF fast failover process due to "
6794 "FCF DEAD event: evt_tag:x%x, fcf_index:x%x "
6795 "\n", acqe_fip->event_tag, acqe_fip->index);
6796 rc = lpfc_sli4_redisc_fcf_table(phba);
6797 if (rc) {
6798 lpfc_printf_log(phba, KERN_ERR, LOG_FIP |
Dick Kennedy372c1872020-06-30 14:50:00 -07006799 LOG_TRACE_EVENT,
Colin Ian King7afc0ce2018-05-03 10:26:12 +01006800 "2772 Issue FCF rediscover mailbox "
James Smartc2b97122013-05-31 17:05:36 -04006801 "command failed, fail through to FCF "
6802 "dead event\n");
6803 spin_lock_irq(&phba->hbalock);
6804 phba->fcf.fcf_flag &= ~FCF_DEAD_DISC;
6805 spin_unlock_irq(&phba->hbalock);
6806 /*
6807 * Last resort will fail over by treating this
6808 * as a link down to FCF registration.
6809 */
6810 lpfc_sli4_fcf_dead_failthrough(phba);
6811 } else {
6812 /* Reset FCF roundrobin bmask for new discovery */
6813 lpfc_sli4_clear_fcf_rr_bmask(phba);
6814 /*
6815 * Handling fast FCF failover to a DEAD FCF event is
6816 * considered equalivant to receiving CVL to all vports.
6817 */
6818 lpfc_sli4_perform_all_vport_cvl(phba);
6819 }
James Smartda0436e2009-05-22 14:51:39 -04006820 break;
James Smart70f3c072010-12-15 17:57:33 -05006821 case LPFC_FIP_EVENT_TYPE_CVL:
James Smart80c17842012-03-01 22:35:45 -05006822 phba->fcoe_cvl_eventtag = acqe_fip->event_tag;
Dick Kennedy372c1872020-06-30 14:50:00 -07006823 lpfc_printf_log(phba, KERN_ERR,
6824 LOG_TRACE_EVENT,
James Smart6669f9b2009-10-02 15:16:45 -04006825 "2718 Clear Virtual Link Received for VPI 0x%x"
James Smart70f3c072010-12-15 17:57:33 -05006826 " tag 0x%x\n", acqe_fip->index, acqe_fip->event_tag);
James Smart6d368e52011-05-24 11:44:12 -04006827
James Smart6669f9b2009-10-02 15:16:45 -04006828 vport = lpfc_find_vport_by_vpid(phba,
James Smart5248a742011-07-22 18:37:06 -04006829 acqe_fip->index);
James Smartfc2b9892010-02-26 14:15:29 -05006830 ndlp = lpfc_sli4_perform_vport_cvl(vport);
James Smart6669f9b2009-10-02 15:16:45 -04006831 if (!ndlp)
6832 break;
James Smart695a8142010-01-26 23:08:03 -05006833 active_vlink_present = 0;
6834
6835 vports = lpfc_create_vport_work_array(phba);
6836 if (vports) {
6837 for (i = 0; i <= phba->max_vports && vports[i] != NULL;
6838 i++) {
6839 if ((!(vports[i]->fc_flag &
6840 FC_VPORT_CVL_RCVD)) &&
6841 (vports[i]->port_state > LPFC_FDISC)) {
6842 active_vlink_present = 1;
6843 break;
6844 }
6845 }
6846 lpfc_destroy_vport_work_array(phba, vports);
6847 }
6848
James Smartcc823552015-05-21 13:55:26 -04006849 /*
6850 * Don't re-instantiate if vport is marked for deletion.
6851 * If we are here first then vport_delete is going to wait
6852 * for discovery to complete.
6853 */
6854 if (!(vport->load_flag & FC_UNLOADING) &&
6855 active_vlink_present) {
James Smart695a8142010-01-26 23:08:03 -05006856 /*
6857 * If there are other active VLinks present,
6858 * re-instantiate the Vlink using FDISC.
6859 */
James Smart256ec0d2013-04-17 20:14:58 -04006860 mod_timer(&ndlp->nlp_delayfunc,
6861 jiffies + msecs_to_jiffies(1000));
James Smartc6adba12020-11-15 11:26:34 -08006862 spin_lock_irq(&ndlp->lock);
James Smart6669f9b2009-10-02 15:16:45 -04006863 ndlp->nlp_flag |= NLP_DELAY_TMO;
James Smartc6adba12020-11-15 11:26:34 -08006864 spin_unlock_irq(&ndlp->lock);
James Smart695a8142010-01-26 23:08:03 -05006865 ndlp->nlp_last_elscmd = ELS_CMD_FDISC;
6866 vport->port_state = LPFC_FDISC;
6867 } else {
James Smartecfd03c2010-02-12 14:41:27 -05006868 /*
6869 * Otherwise, we request port to rediscover
6870 * the entire FCF table for a fast recovery
6871 * from possible case that the current FCF
James Smart0c9ab6f2010-02-26 14:15:57 -05006872 * is no longer valid if we are not already
6873 * in the FCF failover process.
James Smartecfd03c2010-02-12 14:41:27 -05006874 */
James Smartfc2b9892010-02-26 14:15:29 -05006875 spin_lock_irq(&phba->hbalock);
James Smart0c9ab6f2010-02-26 14:15:57 -05006876 if (phba->fcf.fcf_flag & FCF_DISCOVERY) {
James Smartfc2b9892010-02-26 14:15:29 -05006877 spin_unlock_irq(&phba->hbalock);
6878 break;
6879 }
6880 /* Mark the fast failover process in progress */
James Smart0c9ab6f2010-02-26 14:15:57 -05006881 phba->fcf.fcf_flag |= FCF_ACVL_DISC;
James Smartfc2b9892010-02-26 14:15:29 -05006882 spin_unlock_irq(&phba->hbalock);
James Smart0c9ab6f2010-02-26 14:15:57 -05006883 lpfc_printf_log(phba, KERN_INFO, LOG_FIP |
6884 LOG_DISCOVERY,
James Smarta93ff372010-10-22 11:06:08 -04006885 "2773 Start FCF failover per CVL, "
James Smart70f3c072010-12-15 17:57:33 -05006886 "evt_tag:x%x\n", acqe_fip->event_tag);
James Smartecfd03c2010-02-12 14:41:27 -05006887 rc = lpfc_sli4_redisc_fcf_table(phba);
James Smartfc2b9892010-02-26 14:15:29 -05006888 if (rc) {
James Smart0c9ab6f2010-02-26 14:15:57 -05006889 lpfc_printf_log(phba, KERN_ERR, LOG_FIP |
Dick Kennedy372c1872020-06-30 14:50:00 -07006890 LOG_TRACE_EVENT,
James Smart0c9ab6f2010-02-26 14:15:57 -05006891 "2774 Issue FCF rediscover "
Colin Ian King7afc0ce2018-05-03 10:26:12 +01006892 "mailbox command failed, "
James Smart0c9ab6f2010-02-26 14:15:57 -05006893 "through to CVL event\n");
James Smartfc2b9892010-02-26 14:15:29 -05006894 spin_lock_irq(&phba->hbalock);
James Smart0c9ab6f2010-02-26 14:15:57 -05006895 phba->fcf.fcf_flag &= ~FCF_ACVL_DISC;
James Smartfc2b9892010-02-26 14:15:29 -05006896 spin_unlock_irq(&phba->hbalock);
James Smartecfd03c2010-02-12 14:41:27 -05006897 /*
6898 * Last resort will be re-try on the
6899 * the current registered FCF entry.
6900 */
6901 lpfc_retry_pport_discovery(phba);
James Smart38b92ef2010-08-04 16:11:39 -04006902 } else
6903 /*
6904 * Reset FCF roundrobin bmask for new
6905 * discovery.
6906 */
James Smart7d791df2011-07-22 18:37:52 -04006907 lpfc_sli4_clear_fcf_rr_bmask(phba);
James Smart6669f9b2009-10-02 15:16:45 -04006908 }
6909 break;
James Smartda0436e2009-05-22 14:51:39 -04006910 default:
Dick Kennedy372c1872020-06-30 14:50:00 -07006911 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
6912 "0288 Unknown FCoE event type 0x%x event tag "
6913 "0x%x\n", event_type, acqe_fip->event_tag);
James Smartda0436e2009-05-22 14:51:39 -04006914 break;
6915 }
6916}
6917
6918/**
6919 * lpfc_sli4_async_dcbx_evt - Process the asynchronous dcbx event
6920 * @phba: pointer to lpfc hba data structure.
Lee Jonesfe614ac2020-07-23 13:24:22 +01006921 * @acqe_dcbx: pointer to the async dcbx completion queue entry.
James Smartda0436e2009-05-22 14:51:39 -04006922 *
6923 * This routine is to handle the SLI4 asynchronous dcbx event.
6924 **/
6925static void
6926lpfc_sli4_async_dcbx_evt(struct lpfc_hba *phba,
6927 struct lpfc_acqe_dcbx *acqe_dcbx)
6928{
James Smart4d9ab992009-10-02 15:16:39 -04006929 phba->fc_eventTag = acqe_dcbx->event_tag;
Dick Kennedy372c1872020-06-30 14:50:00 -07006930 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartda0436e2009-05-22 14:51:39 -04006931 "0290 The SLI4 DCBX asynchronous event is not "
6932 "handled yet\n");
6933}
6934
6935/**
James Smartb19a0612010-04-06 14:48:51 -04006936 * lpfc_sli4_async_grp5_evt - Process the asynchronous group5 event
6937 * @phba: pointer to lpfc hba data structure.
Lee Jonesfe614ac2020-07-23 13:24:22 +01006938 * @acqe_grp5: pointer to the async grp5 completion queue entry.
James Smartb19a0612010-04-06 14:48:51 -04006939 *
6940 * This routine is to handle the SLI4 asynchronous grp5 event. A grp5 event
6941 * is an asynchronous notified of a logical link speed change. The Port
6942 * reports the logical link speed in units of 10Mbps.
6943 **/
6944static void
6945lpfc_sli4_async_grp5_evt(struct lpfc_hba *phba,
6946 struct lpfc_acqe_grp5 *acqe_grp5)
6947{
6948 uint16_t prev_ll_spd;
6949
6950 phba->fc_eventTag = acqe_grp5->event_tag;
6951 phba->fcoe_eventtag = acqe_grp5->event_tag;
6952 prev_ll_spd = phba->sli4_hba.link_state.logical_speed;
6953 phba->sli4_hba.link_state.logical_speed =
James Smart8b68cd52012-09-29 11:32:37 -04006954 (bf_get(lpfc_acqe_grp5_llink_spd, acqe_grp5)) * 10;
James Smartb19a0612010-04-06 14:48:51 -04006955 lpfc_printf_log(phba, KERN_INFO, LOG_SLI,
6956 "2789 GRP5 Async Event: Updating logical link speed "
James Smart8b68cd52012-09-29 11:32:37 -04006957 "from %dMbps to %dMbps\n", prev_ll_spd,
6958 phba->sli4_hba.link_state.logical_speed);
James Smartb19a0612010-04-06 14:48:51 -04006959}
6960
6961/**
James Smart02243832021-08-16 09:28:54 -07006962 * lpfc_sli4_async_cmstat_evt - Process the asynchronous cmstat event
6963 * @phba: pointer to lpfc hba data structure.
6964 *
6965 * This routine is to handle the SLI4 asynchronous cmstat event. A cmstat event
6966 * is an asynchronous notification of a request to reset CM stats.
6967 **/
6968static void
6969lpfc_sli4_async_cmstat_evt(struct lpfc_hba *phba)
6970{
6971 if (!phba->cgn_i)
6972 return;
6973 lpfc_init_congestion_stat(phba);
6974}
6975
6976/**
James Smart72df8a452021-08-16 09:28:52 -07006977 * lpfc_cgn_params_val - Validate FW congestion parameters.
6978 * @phba: pointer to lpfc hba data structure.
6979 * @p_cfg_param: pointer to FW provided congestion parameters.
6980 *
6981 * This routine validates the congestion parameters passed
6982 * by the FW to the driver via an ACQE event.
6983 **/
6984static void
6985lpfc_cgn_params_val(struct lpfc_hba *phba, struct lpfc_cgn_param *p_cfg_param)
6986{
6987 spin_lock_irq(&phba->hbalock);
6988
6989 if (!lpfc_rangecheck(p_cfg_param->cgn_param_mode, LPFC_CFG_OFF,
6990 LPFC_CFG_MONITOR)) {
6991 lpfc_printf_log(phba, KERN_ERR, LOG_CGN_MGMT,
6992 "6225 CMF mode param out of range: %d\n",
6993 p_cfg_param->cgn_param_mode);
6994 p_cfg_param->cgn_param_mode = LPFC_CFG_OFF;
6995 }
6996
6997 spin_unlock_irq(&phba->hbalock);
6998}
6999
7000/**
7001 * lpfc_cgn_params_parse - Process a FW cong parm change event
7002 * @phba: pointer to lpfc hba data structure.
7003 * @p_cgn_param: pointer to a data buffer with the FW cong params.
7004 * @len: the size of pdata in bytes.
7005 *
7006 * This routine validates the congestion management buffer signature
7007 * from the FW, validates the contents and makes corrections for
7008 * valid, in-range values. If the signature magic is correct and
7009 * after parameter validation, the contents are copied to the driver's
7010 * @phba structure. If the magic is incorrect, an error message is
7011 * logged.
7012 **/
7013static void
7014lpfc_cgn_params_parse(struct lpfc_hba *phba,
7015 struct lpfc_cgn_param *p_cgn_param, uint32_t len)
7016{
James Smart7481811c2021-08-16 09:28:56 -07007017 struct lpfc_cgn_info *cp;
7018 uint32_t crc, oldmode;
James Smart72df8a452021-08-16 09:28:52 -07007019
7020 /* Make sure the FW has encoded the correct magic number to
7021 * validate the congestion parameter in FW memory.
7022 */
7023 if (p_cgn_param->cgn_param_magic == LPFC_CFG_PARAM_MAGIC_NUM) {
7024 lpfc_printf_log(phba, KERN_INFO, LOG_CGN_MGMT | LOG_INIT,
7025 "4668 FW cgn parm buffer data: "
7026 "magic 0x%x version %d mode %d "
7027 "level0 %d level1 %d "
7028 "level2 %d byte13 %d "
7029 "byte14 %d byte15 %d "
7030 "byte11 %d byte12 %d activeMode %d\n",
7031 p_cgn_param->cgn_param_magic,
7032 p_cgn_param->cgn_param_version,
7033 p_cgn_param->cgn_param_mode,
7034 p_cgn_param->cgn_param_level0,
7035 p_cgn_param->cgn_param_level1,
7036 p_cgn_param->cgn_param_level2,
7037 p_cgn_param->byte13,
7038 p_cgn_param->byte14,
7039 p_cgn_param->byte15,
7040 p_cgn_param->byte11,
7041 p_cgn_param->byte12,
7042 phba->cmf_active_mode);
7043
7044 oldmode = phba->cmf_active_mode;
7045
7046 /* Any parameters out of range are corrected to defaults
7047 * by this routine. No need to fail.
7048 */
7049 lpfc_cgn_params_val(phba, p_cgn_param);
7050
7051 /* Parameters are verified, move them into driver storage */
7052 spin_lock_irq(&phba->hbalock);
7053 memcpy(&phba->cgn_p, p_cgn_param,
7054 sizeof(struct lpfc_cgn_param));
7055
James Smart7481811c2021-08-16 09:28:56 -07007056 /* Update parameters in congestion info buffer now */
7057 if (phba->cgn_i) {
7058 cp = (struct lpfc_cgn_info *)phba->cgn_i->virt;
7059 cp->cgn_info_mode = phba->cgn_p.cgn_param_mode;
7060 cp->cgn_info_level0 = phba->cgn_p.cgn_param_level0;
7061 cp->cgn_info_level1 = phba->cgn_p.cgn_param_level1;
7062 cp->cgn_info_level2 = phba->cgn_p.cgn_param_level2;
7063 crc = lpfc_cgn_calc_crc32(cp, LPFC_CGN_INFO_SZ,
7064 LPFC_CGN_CRC32_SEED);
7065 cp->cgn_info_crc = cpu_to_le32(crc);
7066 }
James Smart72df8a452021-08-16 09:28:52 -07007067 spin_unlock_irq(&phba->hbalock);
7068
7069 phba->cmf_active_mode = phba->cgn_p.cgn_param_mode;
7070
7071 switch (oldmode) {
7072 case LPFC_CFG_OFF:
7073 if (phba->cgn_p.cgn_param_mode != LPFC_CFG_OFF) {
7074 /* Turning CMF on */
James Smart02243832021-08-16 09:28:54 -07007075 lpfc_cmf_start(phba);
James Smart72df8a452021-08-16 09:28:52 -07007076
7077 if (phba->link_state >= LPFC_LINK_UP) {
7078 phba->cgn_reg_fpin =
7079 phba->cgn_init_reg_fpin;
7080 phba->cgn_reg_signal =
7081 phba->cgn_init_reg_signal;
7082 lpfc_issue_els_edc(phba->pport, 0);
7083 }
7084 }
7085 break;
7086 case LPFC_CFG_MANAGED:
7087 switch (phba->cgn_p.cgn_param_mode) {
7088 case LPFC_CFG_OFF:
7089 /* Turning CMF off */
James Smart02243832021-08-16 09:28:54 -07007090 lpfc_cmf_stop(phba);
James Smart72df8a452021-08-16 09:28:52 -07007091 if (phba->link_state >= LPFC_LINK_UP)
7092 lpfc_issue_els_edc(phba->pport, 0);
7093 break;
7094 case LPFC_CFG_MONITOR:
7095 lpfc_printf_log(phba, KERN_INFO, LOG_CGN_MGMT,
7096 "4661 Switch from MANAGED to "
7097 "`MONITOR mode\n");
James Smart02243832021-08-16 09:28:54 -07007098 phba->cmf_max_bytes_per_interval =
7099 phba->cmf_link_byte_count;
7100
7101 /* Resume blocked IO - unblock on workqueue */
7102 queue_work(phba->wq,
7103 &phba->unblock_request_work);
James Smart72df8a452021-08-16 09:28:52 -07007104 break;
7105 }
7106 break;
7107 case LPFC_CFG_MONITOR:
7108 switch (phba->cgn_p.cgn_param_mode) {
7109 case LPFC_CFG_OFF:
7110 /* Turning CMF off */
James Smart02243832021-08-16 09:28:54 -07007111 lpfc_cmf_stop(phba);
James Smart72df8a452021-08-16 09:28:52 -07007112 if (phba->link_state >= LPFC_LINK_UP)
7113 lpfc_issue_els_edc(phba->pport, 0);
7114 break;
7115 case LPFC_CFG_MANAGED:
7116 lpfc_printf_log(phba, KERN_INFO, LOG_CGN_MGMT,
7117 "4662 Switch from MONITOR to "
7118 "MANAGED mode\n");
James Smart02243832021-08-16 09:28:54 -07007119 lpfc_cmf_signal_init(phba);
James Smart72df8a452021-08-16 09:28:52 -07007120 break;
7121 }
7122 break;
7123 }
7124 } else {
7125 lpfc_printf_log(phba, KERN_ERR, LOG_CGN_MGMT | LOG_INIT,
7126 "4669 FW cgn parm buf wrong magic 0x%x "
7127 "version %d\n", p_cgn_param->cgn_param_magic,
7128 p_cgn_param->cgn_param_version);
7129 }
7130}
7131
7132/**
7133 * lpfc_sli4_cgn_params_read - Read and Validate FW congestion parameters.
7134 * @phba: pointer to lpfc hba data structure.
7135 *
7136 * This routine issues a read_object mailbox command to
7137 * get the congestion management parameters from the FW
7138 * parses it and updates the driver maintained values.
7139 *
7140 * Returns
7141 * 0 if the object was empty
7142 * -Eval if an error was encountered
7143 * Count if bytes were read from object
7144 **/
7145int
7146lpfc_sli4_cgn_params_read(struct lpfc_hba *phba)
7147{
7148 int ret = 0;
7149 struct lpfc_cgn_param *p_cgn_param = NULL;
7150 u32 *pdata = NULL;
7151 u32 len = 0;
7152
7153 /* Find out if the FW has a new set of congestion parameters. */
7154 len = sizeof(struct lpfc_cgn_param);
7155 pdata = kzalloc(len, GFP_KERNEL);
7156 ret = lpfc_read_object(phba, (char *)LPFC_PORT_CFG_NAME,
7157 pdata, len);
7158
7159 /* 0 means no data. A negative means error. A positive means
7160 * bytes were copied.
7161 */
7162 if (!ret) {
7163 lpfc_printf_log(phba, KERN_ERR, LOG_CGN_MGMT | LOG_INIT,
7164 "4670 CGN RD OBJ returns no data\n");
7165 goto rd_obj_err;
7166 } else if (ret < 0) {
7167 /* Some error. Just exit and return it to the caller.*/
7168 goto rd_obj_err;
7169 }
7170
7171 lpfc_printf_log(phba, KERN_INFO, LOG_CGN_MGMT | LOG_INIT,
7172 "6234 READ CGN PARAMS Successful %d\n", len);
7173
7174 /* Parse data pointer over len and update the phba congestion
7175 * parameters with values passed back. The receive rate values
7176 * may have been altered in FW, but take no action here.
7177 */
7178 p_cgn_param = (struct lpfc_cgn_param *)pdata;
7179 lpfc_cgn_params_parse(phba, p_cgn_param, len);
7180
7181 rd_obj_err:
7182 kfree(pdata);
7183 return ret;
7184}
7185
7186/**
James Smart02243832021-08-16 09:28:54 -07007187 * lpfc_sli4_cgn_parm_chg_evt - Process a FW congestion param change event
7188 * @phba: pointer to lpfc hba data structure.
7189 *
7190 * The FW generated Async ACQE SLI event calls this routine when
7191 * the event type is an SLI Internal Port Event and the Event Code
7192 * indicates a change to the FW maintained congestion parameters.
7193 *
7194 * This routine executes a Read_Object mailbox call to obtain the
7195 * current congestion parameters maintained in FW and corrects
7196 * the driver's active congestion parameters.
7197 *
7198 * The acqe event is not passed because there is no further data
7199 * required.
7200 *
7201 * Returns nonzero error if event processing encountered an error.
7202 * Zero otherwise for success.
7203 **/
7204static int
7205lpfc_sli4_cgn_parm_chg_evt(struct lpfc_hba *phba)
7206{
7207 int ret = 0;
7208
7209 if (!phba->sli4_hba.pc_sli4_params.cmf) {
7210 lpfc_printf_log(phba, KERN_ERR, LOG_CGN_MGMT | LOG_INIT,
7211 "4664 Cgn Evt when E2E off. Drop event\n");
7212 return -EACCES;
7213 }
7214
7215 /* If the event is claiming an empty object, it's ok. A write
7216 * could have cleared it. Only error is a negative return
7217 * status.
7218 */
7219 ret = lpfc_sli4_cgn_params_read(phba);
7220 if (ret < 0) {
7221 lpfc_printf_log(phba, KERN_ERR, LOG_CGN_MGMT | LOG_INIT,
7222 "4667 Error reading Cgn Params (%d)\n",
7223 ret);
7224 } else if (!ret) {
7225 lpfc_printf_log(phba, KERN_ERR, LOG_CGN_MGMT | LOG_INIT,
7226 "4673 CGN Event empty object.\n");
7227 }
7228 return ret;
7229}
7230
7231/**
James Smartda0436e2009-05-22 14:51:39 -04007232 * lpfc_sli4_async_event_proc - Process all the pending asynchronous event
7233 * @phba: pointer to lpfc hba data structure.
7234 *
7235 * This routine is invoked by the worker thread to process all the pending
7236 * SLI4 asynchronous events.
7237 **/
7238void lpfc_sli4_async_event_proc(struct lpfc_hba *phba)
7239{
7240 struct lpfc_cq_event *cq_event;
James Smarte7dab162020-10-20 13:27:12 -07007241 unsigned long iflags;
James Smartda0436e2009-05-22 14:51:39 -04007242
7243 /* First, declare the async event has been handled */
James Smarte7dab162020-10-20 13:27:12 -07007244 spin_lock_irqsave(&phba->hbalock, iflags);
James Smartda0436e2009-05-22 14:51:39 -04007245 phba->hba_flag &= ~ASYNC_EVENT;
James Smarte7dab162020-10-20 13:27:12 -07007246 spin_unlock_irqrestore(&phba->hbalock, iflags);
7247
James Smartda0436e2009-05-22 14:51:39 -04007248 /* Now, handle all the async events */
James Smarte7dab162020-10-20 13:27:12 -07007249 spin_lock_irqsave(&phba->sli4_hba.asynce_list_lock, iflags);
James Smartda0436e2009-05-22 14:51:39 -04007250 while (!list_empty(&phba->sli4_hba.sp_asynce_work_queue)) {
James Smartda0436e2009-05-22 14:51:39 -04007251 list_remove_head(&phba->sli4_hba.sp_asynce_work_queue,
7252 cq_event, struct lpfc_cq_event, list);
James Smarte7dab162020-10-20 13:27:12 -07007253 spin_unlock_irqrestore(&phba->sli4_hba.asynce_list_lock,
7254 iflags);
7255
James Smartda0436e2009-05-22 14:51:39 -04007256 /* Process the asynchronous event */
7257 switch (bf_get(lpfc_trailer_code, &cq_event->cqe.mcqe_cmpl)) {
7258 case LPFC_TRAILER_CODE_LINK:
7259 lpfc_sli4_async_link_evt(phba,
7260 &cq_event->cqe.acqe_link);
7261 break;
7262 case LPFC_TRAILER_CODE_FCOE:
James Smart70f3c072010-12-15 17:57:33 -05007263 lpfc_sli4_async_fip_evt(phba, &cq_event->cqe.acqe_fip);
James Smartda0436e2009-05-22 14:51:39 -04007264 break;
7265 case LPFC_TRAILER_CODE_DCBX:
7266 lpfc_sli4_async_dcbx_evt(phba,
7267 &cq_event->cqe.acqe_dcbx);
7268 break;
James Smartb19a0612010-04-06 14:48:51 -04007269 case LPFC_TRAILER_CODE_GRP5:
7270 lpfc_sli4_async_grp5_evt(phba,
7271 &cq_event->cqe.acqe_grp5);
7272 break;
James Smart70f3c072010-12-15 17:57:33 -05007273 case LPFC_TRAILER_CODE_FC:
7274 lpfc_sli4_async_fc_evt(phba, &cq_event->cqe.acqe_fc);
7275 break;
7276 case LPFC_TRAILER_CODE_SLI:
7277 lpfc_sli4_async_sli_evt(phba, &cq_event->cqe.acqe_sli);
7278 break;
James Smart02243832021-08-16 09:28:54 -07007279 case LPFC_TRAILER_CODE_CMSTAT:
7280 lpfc_sli4_async_cmstat_evt(phba);
7281 break;
James Smartda0436e2009-05-22 14:51:39 -04007282 default:
Dick Kennedy372c1872020-06-30 14:50:00 -07007283 lpfc_printf_log(phba, KERN_ERR,
7284 LOG_TRACE_EVENT,
Colin Ian King291c2542019-12-18 08:43:01 +00007285 "1804 Invalid asynchronous event code: "
James Smartda0436e2009-05-22 14:51:39 -04007286 "x%x\n", bf_get(lpfc_trailer_code,
7287 &cq_event->cqe.mcqe_cmpl));
7288 break;
7289 }
James Smarte7dab162020-10-20 13:27:12 -07007290
James Smartda0436e2009-05-22 14:51:39 -04007291 /* Free the completion event processed to the free pool */
7292 lpfc_sli4_cq_event_release(phba, cq_event);
James Smarte7dab162020-10-20 13:27:12 -07007293 spin_lock_irqsave(&phba->sli4_hba.asynce_list_lock, iflags);
James Smartda0436e2009-05-22 14:51:39 -04007294 }
James Smarte7dab162020-10-20 13:27:12 -07007295 spin_unlock_irqrestore(&phba->sli4_hba.asynce_list_lock, iflags);
James Smartda0436e2009-05-22 14:51:39 -04007296}
7297
7298/**
James Smartecfd03c2010-02-12 14:41:27 -05007299 * lpfc_sli4_fcf_redisc_event_proc - Process fcf table rediscovery event
7300 * @phba: pointer to lpfc hba data structure.
7301 *
7302 * This routine is invoked by the worker thread to process FCF table
7303 * rediscovery pending completion event.
7304 **/
7305void lpfc_sli4_fcf_redisc_event_proc(struct lpfc_hba *phba)
7306{
7307 int rc;
7308
7309 spin_lock_irq(&phba->hbalock);
7310 /* Clear FCF rediscovery timeout event */
7311 phba->fcf.fcf_flag &= ~FCF_REDISC_EVT;
7312 /* Clear driver fast failover FCF record flag */
7313 phba->fcf.failover_rec.flag = 0;
7314 /* Set state for FCF fast failover */
7315 phba->fcf.fcf_flag |= FCF_REDISC_FOV;
7316 spin_unlock_irq(&phba->hbalock);
7317
7318 /* Scan FCF table from the first entry to re-discover SAN */
James Smart0c9ab6f2010-02-26 14:15:57 -05007319 lpfc_printf_log(phba, KERN_INFO, LOG_FIP | LOG_DISCOVERY,
James Smarta93ff372010-10-22 11:06:08 -04007320 "2777 Start post-quiescent FCF table scan\n");
James Smart0c9ab6f2010-02-26 14:15:57 -05007321 rc = lpfc_sli4_fcf_scan_read_fcf_rec(phba, LPFC_FCOE_FCF_GET_FIRST);
James Smartecfd03c2010-02-12 14:41:27 -05007322 if (rc)
Dick Kennedy372c1872020-06-30 14:50:00 -07007323 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart0c9ab6f2010-02-26 14:15:57 -05007324 "2747 Issue FCF scan read FCF mailbox "
7325 "command failed 0x%x\n", rc);
James Smartecfd03c2010-02-12 14:41:27 -05007326}
7327
7328/**
James Smartda0436e2009-05-22 14:51:39 -04007329 * lpfc_api_table_setup - Set up per hba pci-device group func api jump table
7330 * @phba: pointer to lpfc hba data structure.
7331 * @dev_grp: The HBA PCI-Device group number.
7332 *
7333 * This routine is invoked to set up the per HBA PCI-Device group function
7334 * API jump table entries.
7335 *
7336 * Return: 0 if success, otherwise -ENODEV
7337 **/
7338int
7339lpfc_api_table_setup(struct lpfc_hba *phba, uint8_t dev_grp)
James Smart5b75da22008-12-04 22:39:35 -05007340{
7341 int rc;
7342
James Smartda0436e2009-05-22 14:51:39 -04007343 /* Set up lpfc PCI-device group */
7344 phba->pci_dev_grp = dev_grp;
James Smart5b75da22008-12-04 22:39:35 -05007345
James Smartda0436e2009-05-22 14:51:39 -04007346 /* The LPFC_PCI_DEV_OC uses SLI4 */
7347 if (dev_grp == LPFC_PCI_DEV_OC)
7348 phba->sli_rev = LPFC_SLI_REV4;
James Smart5b75da22008-12-04 22:39:35 -05007349
James Smartda0436e2009-05-22 14:51:39 -04007350 /* Set up device INIT API function jump table */
7351 rc = lpfc_init_api_table_setup(phba, dev_grp);
7352 if (rc)
7353 return -ENODEV;
7354 /* Set up SCSI API function jump table */
7355 rc = lpfc_scsi_api_table_setup(phba, dev_grp);
7356 if (rc)
7357 return -ENODEV;
7358 /* Set up SLI API function jump table */
7359 rc = lpfc_sli_api_table_setup(phba, dev_grp);
7360 if (rc)
7361 return -ENODEV;
7362 /* Set up MBOX API function jump table */
7363 rc = lpfc_mbox_api_table_setup(phba, dev_grp);
7364 if (rc)
7365 return -ENODEV;
James Smart5b75da22008-12-04 22:39:35 -05007366
James Smartda0436e2009-05-22 14:51:39 -04007367 return 0;
James Smart5b75da22008-12-04 22:39:35 -05007368}
7369
7370/**
James Smart3621a712009-04-06 18:47:14 -04007371 * lpfc_log_intr_mode - Log the active interrupt mode
James Smart5b75da22008-12-04 22:39:35 -05007372 * @phba: pointer to lpfc hba data structure.
7373 * @intr_mode: active interrupt mode adopted.
7374 *
7375 * This routine it invoked to log the currently used active interrupt mode
7376 * to the device.
James Smart3772a992009-05-22 14:50:54 -04007377 **/
7378static void lpfc_log_intr_mode(struct lpfc_hba *phba, uint32_t intr_mode)
James Smart5b75da22008-12-04 22:39:35 -05007379{
7380 switch (intr_mode) {
7381 case 0:
7382 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
7383 "0470 Enable INTx interrupt mode.\n");
7384 break;
7385 case 1:
7386 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
7387 "0481 Enabled MSI interrupt mode.\n");
7388 break;
7389 case 2:
7390 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
7391 "0480 Enabled MSI-X interrupt mode.\n");
7392 break;
7393 default:
Dick Kennedy372c1872020-06-30 14:50:00 -07007394 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart5b75da22008-12-04 22:39:35 -05007395 "0482 Illegal interrupt mode.\n");
7396 break;
7397 }
7398 return;
7399}
7400
James Smart5b75da22008-12-04 22:39:35 -05007401/**
James Smart3772a992009-05-22 14:50:54 -04007402 * lpfc_enable_pci_dev - Enable a generic PCI device.
James Smart5b75da22008-12-04 22:39:35 -05007403 * @phba: pointer to lpfc hba data structure.
7404 *
James Smart3772a992009-05-22 14:50:54 -04007405 * This routine is invoked to enable the PCI device that is common to all
7406 * PCI devices.
James Smart5b75da22008-12-04 22:39:35 -05007407 *
7408 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -02007409 * 0 - successful
James Smart3772a992009-05-22 14:50:54 -04007410 * other values - error
James Smart5b75da22008-12-04 22:39:35 -05007411 **/
James Smart3772a992009-05-22 14:50:54 -04007412static int
7413lpfc_enable_pci_dev(struct lpfc_hba *phba)
James Smart5b75da22008-12-04 22:39:35 -05007414{
James Smart3772a992009-05-22 14:50:54 -04007415 struct pci_dev *pdev;
James Smart5b75da22008-12-04 22:39:35 -05007416
James Smart3772a992009-05-22 14:50:54 -04007417 /* Obtain PCI device reference */
7418 if (!phba->pcidev)
7419 goto out_error;
7420 else
7421 pdev = phba->pcidev;
James Smart3772a992009-05-22 14:50:54 -04007422 /* Enable PCI device */
7423 if (pci_enable_device_mem(pdev))
7424 goto out_error;
7425 /* Request PCI resource for the device */
Johannes Thumshirne0c04832016-06-07 09:44:03 +02007426 if (pci_request_mem_regions(pdev, LPFC_DRIVER_NAME))
James Smart3772a992009-05-22 14:50:54 -04007427 goto out_disable_device;
7428 /* Set up device as PCI master and save state for EEH */
7429 pci_set_master(pdev);
7430 pci_try_set_mwi(pdev);
7431 pci_save_state(pdev);
James Smart5b75da22008-12-04 22:39:35 -05007432
James Smart05580562011-05-24 11:40:48 -04007433 /* PCIe EEH recovery on powerpc platforms needs fundamental reset */
Jon Mason453193e2013-09-11 15:38:13 -07007434 if (pci_is_pcie(pdev))
James Smart05580562011-05-24 11:40:48 -04007435 pdev->needs_freset = 1;
7436
James Smart3772a992009-05-22 14:50:54 -04007437 return 0;
James Smart5b75da22008-12-04 22:39:35 -05007438
James Smart3772a992009-05-22 14:50:54 -04007439out_disable_device:
7440 pci_disable_device(pdev);
7441out_error:
James Smarta5160742021-10-20 14:14:10 -07007442 lpfc_printf_log(phba, KERN_ERR, LOG_INIT,
Johannes Thumshirne0c04832016-06-07 09:44:03 +02007443 "1401 Failed to enable pci device\n");
James Smart3772a992009-05-22 14:50:54 -04007444 return -ENODEV;
James Smart5b75da22008-12-04 22:39:35 -05007445}
7446
7447/**
James Smart3772a992009-05-22 14:50:54 -04007448 * lpfc_disable_pci_dev - Disable a generic PCI device.
James Smart5b75da22008-12-04 22:39:35 -05007449 * @phba: pointer to lpfc hba data structure.
7450 *
James Smart3772a992009-05-22 14:50:54 -04007451 * This routine is invoked to disable the PCI device that is common to all
7452 * PCI devices.
James Smart5b75da22008-12-04 22:39:35 -05007453 **/
7454static void
James Smart3772a992009-05-22 14:50:54 -04007455lpfc_disable_pci_dev(struct lpfc_hba *phba)
James Smart5b75da22008-12-04 22:39:35 -05007456{
James Smart3772a992009-05-22 14:50:54 -04007457 struct pci_dev *pdev;
James Smart5b75da22008-12-04 22:39:35 -05007458
James Smart3772a992009-05-22 14:50:54 -04007459 /* Obtain PCI device reference */
7460 if (!phba->pcidev)
7461 return;
7462 else
7463 pdev = phba->pcidev;
James Smart3772a992009-05-22 14:50:54 -04007464 /* Release PCI resource and disable PCI device */
Johannes Thumshirne0c04832016-06-07 09:44:03 +02007465 pci_release_mem_regions(pdev);
James Smart3772a992009-05-22 14:50:54 -04007466 pci_disable_device(pdev);
James Smart5b75da22008-12-04 22:39:35 -05007467
7468 return;
7469}
7470
7471/**
James Smart3772a992009-05-22 14:50:54 -04007472 * lpfc_reset_hba - Reset a hba
7473 * @phba: pointer to lpfc hba data structure.
James Smarte59058c2008-08-24 21:49:00 -04007474 *
James Smart3772a992009-05-22 14:50:54 -04007475 * This routine is invoked to reset a hba device. It brings the HBA
7476 * offline, performs a board restart, and then brings the board back
7477 * online. The lpfc_offline calls lpfc_sli_hba_down which will clean up
7478 * on outstanding mailbox commands.
James Smarte59058c2008-08-24 21:49:00 -04007479 **/
James Smart3772a992009-05-22 14:50:54 -04007480void
7481lpfc_reset_hba(struct lpfc_hba *phba)
dea31012005-04-17 16:05:31 -05007482{
James Smart3772a992009-05-22 14:50:54 -04007483 /* If resets are disabled then set error state and return. */
7484 if (!phba->cfg_enable_hba_reset) {
7485 phba->link_state = LPFC_HBA_ERROR;
7486 return;
7487 }
James Smart9ec58ec2021-01-04 10:02:35 -08007488
7489 /* If not LPFC_SLI_ACTIVE, force all IO to be flushed */
7490 if (phba->sli.sli_flag & LPFC_SLI_ACTIVE) {
James Smartee620212014-04-04 13:51:53 -04007491 lpfc_offline_prep(phba, LPFC_MBX_WAIT);
James Smart9ec58ec2021-01-04 10:02:35 -08007492 } else {
James Smartee620212014-04-04 13:51:53 -04007493 lpfc_offline_prep(phba, LPFC_MBX_NO_WAIT);
James Smart9ec58ec2021-01-04 10:02:35 -08007494 lpfc_sli_flush_io_rings(phba);
7495 }
James Smart3772a992009-05-22 14:50:54 -04007496 lpfc_offline(phba);
7497 lpfc_sli_brdrestart(phba);
7498 lpfc_online(phba);
7499 lpfc_unblock_mgmt_io(phba);
7500}
dea31012005-04-17 16:05:31 -05007501
James Smart3772a992009-05-22 14:50:54 -04007502/**
James Smart0a96e972011-07-22 18:37:28 -04007503 * lpfc_sli_sriov_nr_virtfn_get - Get the number of sr-iov virtual functions
7504 * @phba: pointer to lpfc hba data structure.
7505 *
7506 * This function enables the PCI SR-IOV virtual functions to a physical
7507 * function. It invokes the PCI SR-IOV api with the @nr_vfn provided to
7508 * enable the number of virtual functions to the physical function. As
7509 * not all devices support SR-IOV, the return code from the pci_enable_sriov()
7510 * API call does not considered as an error condition for most of the device.
7511 **/
7512uint16_t
7513lpfc_sli_sriov_nr_virtfn_get(struct lpfc_hba *phba)
7514{
7515 struct pci_dev *pdev = phba->pcidev;
7516 uint16_t nr_virtfn;
7517 int pos;
7518
James Smart0a96e972011-07-22 18:37:28 -04007519 pos = pci_find_ext_capability(pdev, PCI_EXT_CAP_ID_SRIOV);
7520 if (pos == 0)
7521 return 0;
7522
7523 pci_read_config_word(pdev, pos + PCI_SRIOV_TOTAL_VF, &nr_virtfn);
7524 return nr_virtfn;
7525}
7526
7527/**
James Smart912e3ac2011-05-24 11:42:11 -04007528 * lpfc_sli_probe_sriov_nr_virtfn - Enable a number of sr-iov virtual functions
7529 * @phba: pointer to lpfc hba data structure.
7530 * @nr_vfn: number of virtual functions to be enabled.
7531 *
7532 * This function enables the PCI SR-IOV virtual functions to a physical
7533 * function. It invokes the PCI SR-IOV api with the @nr_vfn provided to
7534 * enable the number of virtual functions to the physical function. As
7535 * not all devices support SR-IOV, the return code from the pci_enable_sriov()
7536 * API call does not considered as an error condition for most of the device.
7537 **/
7538int
7539lpfc_sli_probe_sriov_nr_virtfn(struct lpfc_hba *phba, int nr_vfn)
7540{
7541 struct pci_dev *pdev = phba->pcidev;
James Smart0a96e972011-07-22 18:37:28 -04007542 uint16_t max_nr_vfn;
James Smart912e3ac2011-05-24 11:42:11 -04007543 int rc;
7544
James Smart0a96e972011-07-22 18:37:28 -04007545 max_nr_vfn = lpfc_sli_sriov_nr_virtfn_get(phba);
7546 if (nr_vfn > max_nr_vfn) {
Dick Kennedy372c1872020-06-30 14:50:00 -07007547 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart0a96e972011-07-22 18:37:28 -04007548 "3057 Requested vfs (%d) greater than "
7549 "supported vfs (%d)", nr_vfn, max_nr_vfn);
7550 return -EINVAL;
7551 }
7552
James Smart912e3ac2011-05-24 11:42:11 -04007553 rc = pci_enable_sriov(pdev, nr_vfn);
7554 if (rc) {
7555 lpfc_printf_log(phba, KERN_WARNING, LOG_INIT,
7556 "2806 Failed to enable sriov on this device "
7557 "with vfn number nr_vf:%d, rc:%d\n",
7558 nr_vfn, rc);
7559 } else
7560 lpfc_printf_log(phba, KERN_WARNING, LOG_INIT,
7561 "2807 Successful enable sriov on this device "
7562 "with vfn number nr_vf:%d\n", nr_vfn);
7563 return rc;
7564}
7565
James Smart02243832021-08-16 09:28:54 -07007566static void
7567lpfc_unblock_requests_work(struct work_struct *work)
7568{
7569 struct lpfc_hba *phba = container_of(work, struct lpfc_hba,
7570 unblock_request_work);
7571
7572 lpfc_unblock_requests(phba);
7573}
7574
James Smart912e3ac2011-05-24 11:42:11 -04007575/**
James Smart895427b2017-02-12 13:52:30 -08007576 * lpfc_setup_driver_resource_phase1 - Phase1 etup driver internal resources.
James Smart3772a992009-05-22 14:50:54 -04007577 * @phba: pointer to lpfc hba data structure.
7578 *
James Smart895427b2017-02-12 13:52:30 -08007579 * This routine is invoked to set up the driver internal resources before the
7580 * device specific resource setup to support the HBA device it attached to.
James Smart3772a992009-05-22 14:50:54 -04007581 *
7582 * Return codes
James Smart895427b2017-02-12 13:52:30 -08007583 * 0 - successful
7584 * other values - error
James Smart3772a992009-05-22 14:50:54 -04007585 **/
7586static int
James Smart895427b2017-02-12 13:52:30 -08007587lpfc_setup_driver_resource_phase1(struct lpfc_hba *phba)
James Smart3772a992009-05-22 14:50:54 -04007588{
James Smart895427b2017-02-12 13:52:30 -08007589 struct lpfc_sli *psli = &phba->sli;
dea31012005-04-17 16:05:31 -05007590
James Smart3772a992009-05-22 14:50:54 -04007591 /*
James Smart895427b2017-02-12 13:52:30 -08007592 * Driver resources common to all SLI revisions
James Smart3772a992009-05-22 14:50:54 -04007593 */
James Smart895427b2017-02-12 13:52:30 -08007594 atomic_set(&phba->fast_event_count, 0);
Dick Kennedy372c1872020-06-30 14:50:00 -07007595 atomic_set(&phba->dbg_log_idx, 0);
7596 atomic_set(&phba->dbg_log_cnt, 0);
7597 atomic_set(&phba->dbg_log_dmping, 0);
James Smart895427b2017-02-12 13:52:30 -08007598 spin_lock_init(&phba->hbalock);
dea31012005-04-17 16:05:31 -05007599
James Smart523128e2018-09-10 10:30:46 -07007600 /* Initialize port_list spinlock */
7601 spin_lock_init(&phba->port_list_lock);
James Smart895427b2017-02-12 13:52:30 -08007602 INIT_LIST_HEAD(&phba->port_list);
James Smart523128e2018-09-10 10:30:46 -07007603
James Smart895427b2017-02-12 13:52:30 -08007604 INIT_LIST_HEAD(&phba->work_list);
7605 init_waitqueue_head(&phba->wait_4_mlo_m_q);
7606
7607 /* Initialize the wait queue head for the kernel thread */
7608 init_waitqueue_head(&phba->work_waitq);
7609
7610 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
James Smartf358dd02017-02-12 13:52:34 -08007611 "1403 Protocols supported %s %s %s\n",
James Smart895427b2017-02-12 13:52:30 -08007612 ((phba->cfg_enable_fc4_type & LPFC_ENABLE_FCP) ?
7613 "SCSI" : " "),
7614 ((phba->cfg_enable_fc4_type & LPFC_ENABLE_NVME) ?
James Smartf358dd02017-02-12 13:52:34 -08007615 "NVME" : " "),
7616 (phba->nvmet_support ? "NVMET" : " "));
James Smart895427b2017-02-12 13:52:30 -08007617
James Smart0794d602019-01-28 11:14:19 -08007618 /* Initialize the IO buffer list used by driver for SLI3 SCSI */
7619 spin_lock_init(&phba->scsi_buf_list_get_lock);
7620 INIT_LIST_HEAD(&phba->lpfc_scsi_buf_list_get);
7621 spin_lock_init(&phba->scsi_buf_list_put_lock);
7622 INIT_LIST_HEAD(&phba->lpfc_scsi_buf_list_put);
James Smart895427b2017-02-12 13:52:30 -08007623
7624 /* Initialize the fabric iocb list */
7625 INIT_LIST_HEAD(&phba->fabric_iocb_list);
7626
7627 /* Initialize list to save ELS buffers */
7628 INIT_LIST_HEAD(&phba->elsbuf);
7629
7630 /* Initialize FCF connection rec list */
7631 INIT_LIST_HEAD(&phba->fcf_conn_rec_list);
7632
7633 /* Initialize OAS configuration list */
7634 spin_lock_init(&phba->devicelock);
7635 INIT_LIST_HEAD(&phba->luns);
7636
James Smart3772a992009-05-22 14:50:54 -04007637 /* MBOX heartbeat timer */
Kees Cookf22eb4d2017-09-06 20:24:26 -07007638 timer_setup(&psli->mbox_tmo, lpfc_mbox_timeout, 0);
James Smart3772a992009-05-22 14:50:54 -04007639 /* Fabric block timer */
Kees Cookf22eb4d2017-09-06 20:24:26 -07007640 timer_setup(&phba->fabric_block_timer, lpfc_fabric_block_timeout, 0);
James Smart3772a992009-05-22 14:50:54 -04007641 /* EA polling mode timer */
Kees Cookf22eb4d2017-09-06 20:24:26 -07007642 timer_setup(&phba->eratt_poll, lpfc_poll_eratt, 0);
James Smart895427b2017-02-12 13:52:30 -08007643 /* Heartbeat timer */
Kees Cookf22eb4d2017-09-06 20:24:26 -07007644 timer_setup(&phba->hb_tmofunc, lpfc_hb_timeout, 0);
James Smart895427b2017-02-12 13:52:30 -08007645
James Smart32517fc2019-01-28 11:14:33 -08007646 INIT_DELAYED_WORK(&phba->eq_delay_work, lpfc_hb_eq_delay_work);
7647
Dick Kennedy317aeb82020-06-30 14:49:59 -07007648 INIT_DELAYED_WORK(&phba->idle_stat_delay_work,
7649 lpfc_idle_stat_delay_work);
James Smart02243832021-08-16 09:28:54 -07007650 INIT_WORK(&phba->unblock_request_work, lpfc_unblock_requests_work);
James Smart895427b2017-02-12 13:52:30 -08007651 return 0;
7652}
7653
7654/**
7655 * lpfc_sli_driver_resource_setup - Setup driver internal resources for SLI3 dev
7656 * @phba: pointer to lpfc hba data structure.
7657 *
7658 * This routine is invoked to set up the driver internal resources specific to
7659 * support the SLI-3 HBA device it attached to.
7660 *
7661 * Return codes
7662 * 0 - successful
7663 * other values - error
7664 **/
7665static int
7666lpfc_sli_driver_resource_setup(struct lpfc_hba *phba)
7667{
James Smart0794d602019-01-28 11:14:19 -08007668 int rc, entry_sz;
James Smart895427b2017-02-12 13:52:30 -08007669
7670 /*
7671 * Initialize timers used by driver
7672 */
7673
7674 /* FCP polling mode timer */
Kees Cookf22eb4d2017-09-06 20:24:26 -07007675 timer_setup(&phba->fcp_poll_timer, lpfc_poll_timeout, 0);
James Smart3772a992009-05-22 14:50:54 -04007676
7677 /* Host attention work mask setup */
7678 phba->work_ha_mask = (HA_ERATT | HA_MBATT | HA_LATT);
7679 phba->work_ha_mask |= (HA_RXMASK << (LPFC_ELS_RING * 4));
7680
7681 /* Get all the module params for configuring this host */
7682 lpfc_get_cfgparam(phba);
James Smart895427b2017-02-12 13:52:30 -08007683 /* Set up phase-1 common device driver resources */
7684
7685 rc = lpfc_setup_driver_resource_phase1(phba);
7686 if (rc)
7687 return -ENODEV;
7688
James Smart49198b32010-04-06 15:04:33 -04007689 if (phba->pcidev->device == PCI_DEVICE_ID_HORNET) {
7690 phba->menlo_flag |= HBA_MENLO_SUPPORT;
7691 /* check for menlo minimum sg count */
7692 if (phba->cfg_sg_seg_cnt < LPFC_DEFAULT_MENLO_SG_SEG_CNT)
7693 phba->cfg_sg_seg_cnt = LPFC_DEFAULT_MENLO_SG_SEG_CNT;
7694 }
7695
James Smart895427b2017-02-12 13:52:30 -08007696 if (!phba->sli.sli3_ring)
Kees Cook6396bb22018-06-12 14:03:40 -07007697 phba->sli.sli3_ring = kcalloc(LPFC_SLI3_MAX_RING,
7698 sizeof(struct lpfc_sli_ring),
7699 GFP_KERNEL);
James Smart895427b2017-02-12 13:52:30 -08007700 if (!phba->sli.sli3_ring)
James Smart2a76a282012-08-03 12:35:54 -04007701 return -ENOMEM;
7702
James Smart3772a992009-05-22 14:50:54 -04007703 /*
James Smart96f70772013-04-17 20:16:15 -04007704 * Since lpfc_sg_seg_cnt is module parameter, the sg_dma_buf_size
James Smart3772a992009-05-22 14:50:54 -04007705 * used to create the sg_dma_buf_pool must be dynamically calculated.
James Smart3772a992009-05-22 14:50:54 -04007706 */
James Smart3772a992009-05-22 14:50:54 -04007707
James Smart0794d602019-01-28 11:14:19 -08007708 if (phba->sli_rev == LPFC_SLI_REV4)
7709 entry_sz = sizeof(struct sli4_sge);
7710 else
7711 entry_sz = sizeof(struct ulp_bde64);
7712
James Smart96f70772013-04-17 20:16:15 -04007713 /* There are going to be 2 reserved BDEs: 1 FCP cmnd + 1 FCP rsp */
7714 if (phba->cfg_enable_bg) {
7715 /*
7716 * The scsi_buf for a T10-DIF I/O will hold the FCP cmnd,
7717 * the FCP rsp, and a BDE for each. Sice we have no control
7718 * over how many protection data segments the SCSI Layer
7719 * will hand us (ie: there could be one for every block
7720 * in the IO), we just allocate enough BDEs to accomidate
7721 * our max amount and we need to limit lpfc_sg_seg_cnt to
7722 * minimize the risk of running out.
7723 */
7724 phba->cfg_sg_dma_buf_size = sizeof(struct fcp_cmnd) +
7725 sizeof(struct fcp_rsp) +
James Smart0794d602019-01-28 11:14:19 -08007726 (LPFC_MAX_SG_SEG_CNT * entry_sz);
James Smart96f70772013-04-17 20:16:15 -04007727
7728 if (phba->cfg_sg_seg_cnt > LPFC_MAX_SG_SEG_CNT_DIF)
7729 phba->cfg_sg_seg_cnt = LPFC_MAX_SG_SEG_CNT_DIF;
7730
7731 /* Total BDEs in BPL for scsi_sg_list and scsi_sg_prot_list */
7732 phba->cfg_total_seg_cnt = LPFC_MAX_SG_SEG_CNT;
7733 } else {
7734 /*
7735 * The scsi_buf for a regular I/O will hold the FCP cmnd,
7736 * the FCP rsp, a BDE for each, and a BDE for up to
7737 * cfg_sg_seg_cnt data segments.
7738 */
7739 phba->cfg_sg_dma_buf_size = sizeof(struct fcp_cmnd) +
7740 sizeof(struct fcp_rsp) +
James Smart0794d602019-01-28 11:14:19 -08007741 ((phba->cfg_sg_seg_cnt + 2) * entry_sz);
James Smart96f70772013-04-17 20:16:15 -04007742
7743 /* Total BDEs in BPL for scsi_sg_list */
7744 phba->cfg_total_seg_cnt = phba->cfg_sg_seg_cnt + 2;
7745 }
7746
7747 lpfc_printf_log(phba, KERN_INFO, LOG_INIT | LOG_FCP,
James Smartc90b4482020-03-22 11:12:56 -07007748 "9088 INIT sg_tablesize:%d dmabuf_size:%d total_bde:%d\n",
James Smart96f70772013-04-17 20:16:15 -04007749 phba->cfg_sg_seg_cnt, phba->cfg_sg_dma_buf_size,
7750 phba->cfg_total_seg_cnt);
7751
James Smart3772a992009-05-22 14:50:54 -04007752 phba->max_vpi = LPFC_MAX_VPI;
7753 /* This will be set to correct value after config_port mbox */
7754 phba->max_vports = 0;
7755
7756 /*
7757 * Initialize the SLI Layer to run with lpfc HBAs.
7758 */
7759 lpfc_sli_setup(phba);
James Smart895427b2017-02-12 13:52:30 -08007760 lpfc_sli_queue_init(phba);
James Smart3772a992009-05-22 14:50:54 -04007761
7762 /* Allocate device driver memory */
7763 if (lpfc_mem_alloc(phba, BPL_ALIGN_SZ))
7764 return -ENOMEM;
7765
James Smartd79c9e92019-08-14 16:57:09 -07007766 phba->lpfc_sg_dma_buf_pool =
7767 dma_pool_create("lpfc_sg_dma_buf_pool",
7768 &phba->pcidev->dev, phba->cfg_sg_dma_buf_size,
7769 BPL_ALIGN_SZ, 0);
7770
7771 if (!phba->lpfc_sg_dma_buf_pool)
7772 goto fail_free_mem;
7773
7774 phba->lpfc_cmd_rsp_buf_pool =
7775 dma_pool_create("lpfc_cmd_rsp_buf_pool",
7776 &phba->pcidev->dev,
7777 sizeof(struct fcp_cmnd) +
7778 sizeof(struct fcp_rsp),
7779 BPL_ALIGN_SZ, 0);
7780
7781 if (!phba->lpfc_cmd_rsp_buf_pool)
7782 goto fail_free_dma_buf_pool;
7783
James Smart912e3ac2011-05-24 11:42:11 -04007784 /*
7785 * Enable sr-iov virtual functions if supported and configured
7786 * through the module parameter.
7787 */
7788 if (phba->cfg_sriov_nr_virtfn > 0) {
7789 rc = lpfc_sli_probe_sriov_nr_virtfn(phba,
7790 phba->cfg_sriov_nr_virtfn);
7791 if (rc) {
7792 lpfc_printf_log(phba, KERN_WARNING, LOG_INIT,
7793 "2808 Requested number of SR-IOV "
7794 "virtual functions (%d) is not "
7795 "supported\n",
7796 phba->cfg_sriov_nr_virtfn);
7797 phba->cfg_sriov_nr_virtfn = 0;
7798 }
7799 }
7800
James Smart3772a992009-05-22 14:50:54 -04007801 return 0;
James Smartd79c9e92019-08-14 16:57:09 -07007802
7803fail_free_dma_buf_pool:
7804 dma_pool_destroy(phba->lpfc_sg_dma_buf_pool);
7805 phba->lpfc_sg_dma_buf_pool = NULL;
7806fail_free_mem:
7807 lpfc_mem_free(phba);
7808 return -ENOMEM;
James Smart3772a992009-05-22 14:50:54 -04007809}
7810
7811/**
7812 * lpfc_sli_driver_resource_unset - Unset drvr internal resources for SLI3 dev
7813 * @phba: pointer to lpfc hba data structure.
7814 *
7815 * This routine is invoked to unset the driver internal resources set up
7816 * specific for supporting the SLI-3 HBA device it attached to.
7817 **/
7818static void
7819lpfc_sli_driver_resource_unset(struct lpfc_hba *phba)
7820{
7821 /* Free device driver memory allocated */
7822 lpfc_mem_free_all(phba);
7823
7824 return;
7825}
7826
7827/**
James Smartda0436e2009-05-22 14:51:39 -04007828 * lpfc_sli4_driver_resource_setup - Setup drvr internal resources for SLI4 dev
7829 * @phba: pointer to lpfc hba data structure.
7830 *
7831 * This routine is invoked to set up the driver internal resources specific to
7832 * support the SLI-4 HBA device it attached to.
7833 *
7834 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -02007835 * 0 - successful
James Smartda0436e2009-05-22 14:51:39 -04007836 * other values - error
7837 **/
7838static int
7839lpfc_sli4_driver_resource_setup(struct lpfc_hba *phba)
7840{
James Smart28baac72010-02-12 14:42:03 -05007841 LPFC_MBOXQ_t *mboxq;
James Smartf358dd02017-02-12 13:52:34 -08007842 MAILBOX_t *mb;
James Smart895427b2017-02-12 13:52:30 -08007843 int rc, i, max_buf_size;
James Smart09294d42013-04-17 20:16:05 -04007844 int longs;
James Smart81e6a632017-11-20 16:00:43 -08007845 int extra;
James Smartf358dd02017-02-12 13:52:34 -08007846 uint64_t wwn;
James Smartb92dc722018-05-24 21:09:01 -07007847 u32 if_type;
7848 u32 if_fam;
James Smartda0436e2009-05-22 14:51:39 -04007849
James Smart895427b2017-02-12 13:52:30 -08007850 phba->sli4_hba.num_present_cpu = lpfc_present_cpu;
James Smarteede4972019-11-21 09:55:56 -08007851 phba->sli4_hba.num_possible_cpu = cpumask_last(cpu_possible_mask) + 1;
James Smart895427b2017-02-12 13:52:30 -08007852 phba->sli4_hba.curr_disp_cpu = 0;
7853
James Smart716d3bc2013-09-06 12:18:28 -04007854 /* Get all the module params for configuring this host */
7855 lpfc_get_cfgparam(phba);
7856
James Smart895427b2017-02-12 13:52:30 -08007857 /* Set up phase-1 common device driver resources */
7858 rc = lpfc_setup_driver_resource_phase1(phba);
7859 if (rc)
7860 return -ENODEV;
7861
James Smartda0436e2009-05-22 14:51:39 -04007862 /* Before proceed, wait for POST done and device ready */
7863 rc = lpfc_sli4_post_status_check(phba);
7864 if (rc)
7865 return -ENODEV;
7866
James Smart3cee98d2019-08-14 16:56:34 -07007867 /* Allocate all driver workqueues here */
7868
7869 /* The lpfc_wq workqueue for deferred irq use */
7870 phba->wq = alloc_workqueue("lpfc_wq", WQ_MEM_RECLAIM, 0);
7871
James Smartda0436e2009-05-22 14:51:39 -04007872 /*
7873 * Initialize timers used by driver
7874 */
7875
Kees Cookf22eb4d2017-09-06 20:24:26 -07007876 timer_setup(&phba->rrq_tmr, lpfc_rrq_timeout, 0);
James Smartda0436e2009-05-22 14:51:39 -04007877
James Smartecfd03c2010-02-12 14:41:27 -05007878 /* FCF rediscover timer */
Kees Cookf22eb4d2017-09-06 20:24:26 -07007879 timer_setup(&phba->fcf.redisc_wait, lpfc_sli4_fcf_redisc_wait_tmo, 0);
James Smartecfd03c2010-02-12 14:41:27 -05007880
James Smart02243832021-08-16 09:28:54 -07007881 /* CMF congestion timer */
7882 hrtimer_init(&phba->cmf_timer, CLOCK_MONOTONIC, HRTIMER_MODE_REL);
7883 phba->cmf_timer.function = lpfc_cmf_timer;
7884
James Smartda0436e2009-05-22 14:51:39 -04007885 /*
James Smart7ad20aa2011-05-24 11:44:28 -04007886 * Control structure for handling external multi-buffer mailbox
7887 * command pass-through.
7888 */
7889 memset((uint8_t *)&phba->mbox_ext_buf_ctx, 0,
7890 sizeof(struct lpfc_mbox_ext_buf_ctx));
7891 INIT_LIST_HEAD(&phba->mbox_ext_buf_ctx.ext_dmabuf_list);
7892
James Smartda0436e2009-05-22 14:51:39 -04007893 phba->max_vpi = LPFC_MAX_VPI;
James Smart67d12732012-08-03 12:36:13 -04007894
James Smartda0436e2009-05-22 14:51:39 -04007895 /* This will be set to correct value after the read_config mbox */
7896 phba->max_vports = 0;
7897
7898 /* Program the default value of vlan_id and fc_map */
7899 phba->valid_vlan = 0;
7900 phba->fc_map[0] = LPFC_FCOE_FCF_MAP0;
7901 phba->fc_map[1] = LPFC_FCOE_FCF_MAP1;
7902 phba->fc_map[2] = LPFC_FCOE_FCF_MAP2;
7903
7904 /*
James Smart2a76a282012-08-03 12:35:54 -04007905 * For SLI4, instead of using ring 0 (LPFC_FCP_RING) for FCP commands
James Smart895427b2017-02-12 13:52:30 -08007906 * we will associate a new ring, for each EQ/CQ/WQ tuple.
7907 * The WQ create will allocate the ring.
James Smart2a76a282012-08-03 12:35:54 -04007908 */
James Smart085c6472010-11-20 23:11:37 -05007909
James Smartda0436e2009-05-22 14:51:39 -04007910 /* Initialize buffer queue management fields */
James Smart895427b2017-02-12 13:52:30 -08007911 INIT_LIST_HEAD(&phba->hbqs[LPFC_ELS_HBQ].hbq_buffer_list);
James Smartda0436e2009-05-22 14:51:39 -04007912 phba->hbqs[LPFC_ELS_HBQ].hbq_alloc_buffer = lpfc_sli4_rb_alloc;
7913 phba->hbqs[LPFC_ELS_HBQ].hbq_free_buffer = lpfc_sli4_rb_free;
7914
Gaurav Srivastava20397172021-06-08 10:05:54 +05307915 /* for VMID idle timeout if VMID is enabled */
7916 if (lpfc_is_vmid_enabled(phba))
7917 timer_setup(&phba->inactive_vmid_poll, lpfc_vmid_poll, 0);
7918
James Smartda0436e2009-05-22 14:51:39 -04007919 /*
7920 * Initialize the SLI Layer to run with lpfc SLI4 HBAs.
7921 */
James Smartc00f62e2019-08-14 16:57:11 -07007922 /* Initialize the Abort buffer list used by driver */
7923 spin_lock_init(&phba->sli4_hba.abts_io_buf_list_lock);
7924 INIT_LIST_HEAD(&phba->sli4_hba.lpfc_abts_io_buf_list);
James Smart895427b2017-02-12 13:52:30 -08007925
7926 if (phba->cfg_enable_fc4_type & LPFC_ENABLE_NVME) {
7927 /* Initialize the Abort nvme buffer list used by driver */
James Smart5e5b5112019-01-28 11:14:22 -08007928 spin_lock_init(&phba->sli4_hba.abts_nvmet_buf_list_lock);
James Smart86c67372017-04-21 16:05:04 -07007929 INIT_LIST_HEAD(&phba->sli4_hba.lpfc_abts_nvmet_ctx_list);
James Smarta8cf5df2017-05-15 15:20:46 -07007930 INIT_LIST_HEAD(&phba->sli4_hba.lpfc_nvmet_io_wait_list);
James Smart79d8c4c2019-05-21 17:48:56 -07007931 spin_lock_init(&phba->sli4_hba.t_active_list_lock);
7932 INIT_LIST_HEAD(&phba->sli4_hba.t_active_ctx_list);
James Smart895427b2017-02-12 13:52:30 -08007933 }
7934
James Smartda0436e2009-05-22 14:51:39 -04007935 /* This abort list used by worker thread */
James Smart895427b2017-02-12 13:52:30 -08007936 spin_lock_init(&phba->sli4_hba.sgl_list_lock);
James Smarta8cf5df2017-05-15 15:20:46 -07007937 spin_lock_init(&phba->sli4_hba.nvmet_io_wait_lock);
James Smarte7dab162020-10-20 13:27:12 -07007938 spin_lock_init(&phba->sli4_hba.asynce_list_lock);
7939 spin_lock_init(&phba->sli4_hba.els_xri_abrt_list_lock);
James Smartda0436e2009-05-22 14:51:39 -04007940
7941 /*
James Smart6d368e52011-05-24 11:44:12 -04007942 * Initialize driver internal slow-path work queues
James Smartda0436e2009-05-22 14:51:39 -04007943 */
7944
7945 /* Driver internel slow-path CQ Event pool */
7946 INIT_LIST_HEAD(&phba->sli4_hba.sp_cqe_event_pool);
7947 /* Response IOCB work queue list */
James Smart45ed1192009-10-02 15:17:02 -04007948 INIT_LIST_HEAD(&phba->sli4_hba.sp_queue_event);
James Smartda0436e2009-05-22 14:51:39 -04007949 /* Asynchronous event CQ Event work queue list */
7950 INIT_LIST_HEAD(&phba->sli4_hba.sp_asynce_work_queue);
James Smartda0436e2009-05-22 14:51:39 -04007951 /* Slow-path XRI aborted CQ Event work queue list */
7952 INIT_LIST_HEAD(&phba->sli4_hba.sp_els_xri_aborted_work_queue);
7953 /* Receive queue CQ Event work queue list */
7954 INIT_LIST_HEAD(&phba->sli4_hba.sp_unsol_work_queue);
7955
James Smart6d368e52011-05-24 11:44:12 -04007956 /* Initialize extent block lists. */
7957 INIT_LIST_HEAD(&phba->sli4_hba.lpfc_rpi_blk_list);
7958 INIT_LIST_HEAD(&phba->sli4_hba.lpfc_xri_blk_list);
7959 INIT_LIST_HEAD(&phba->sli4_hba.lpfc_vfi_blk_list);
7960 INIT_LIST_HEAD(&phba->lpfc_vpi_blk_list);
7961
James Smartd1f525a2017-04-21 16:04:55 -07007962 /* Initialize mboxq lists. If the early init routines fail
7963 * these lists need to be correctly initialized.
7964 */
7965 INIT_LIST_HEAD(&phba->sli.mboxq);
7966 INIT_LIST_HEAD(&phba->sli.mboxq_cmpl);
7967
James Smart448193b2015-12-16 18:12:05 -05007968 /* initialize optic_state to 0xFF */
7969 phba->sli4_hba.lnk_info.optic_state = 0xff;
7970
James Smartda0436e2009-05-22 14:51:39 -04007971 /* Allocate device driver memory */
7972 rc = lpfc_mem_alloc(phba, SGL_ALIGN_SZ);
7973 if (rc)
7974 return -ENOMEM;
7975
James Smart2fcee4b2010-12-15 17:57:46 -05007976 /* IF Type 2 ports get initialized now. */
James Smart27d6ac02018-02-22 08:18:42 -08007977 if (bf_get(lpfc_sli_intf_if_type, &phba->sli4_hba.sli_intf) >=
James Smart2fcee4b2010-12-15 17:57:46 -05007978 LPFC_SLI_INTF_IF_TYPE_2) {
7979 rc = lpfc_pci_function_reset(phba);
James Smart895427b2017-02-12 13:52:30 -08007980 if (unlikely(rc)) {
7981 rc = -ENODEV;
7982 goto out_free_mem;
7983 }
James Smart946727d2015-04-07 15:07:09 -04007984 phba->temp_sensor_support = 1;
James Smart2fcee4b2010-12-15 17:57:46 -05007985 }
7986
James Smartda0436e2009-05-22 14:51:39 -04007987 /* Create the bootstrap mailbox command */
7988 rc = lpfc_create_bootstrap_mbox(phba);
7989 if (unlikely(rc))
7990 goto out_free_mem;
7991
7992 /* Set up the host's endian order with the device. */
7993 rc = lpfc_setup_endian_order(phba);
7994 if (unlikely(rc))
7995 goto out_free_bsmbx;
7996
7997 /* Set up the hba's configuration parameters. */
7998 rc = lpfc_sli4_read_config(phba);
7999 if (unlikely(rc))
8000 goto out_free_bsmbx;
James Smartcff261f2013-12-17 20:29:47 -05008001 rc = lpfc_mem_alloc_active_rrq_pool_s4(phba);
8002 if (unlikely(rc))
8003 goto out_free_bsmbx;
James Smartda0436e2009-05-22 14:51:39 -04008004
James Smart2fcee4b2010-12-15 17:57:46 -05008005 /* IF Type 0 ports get initialized now. */
8006 if (bf_get(lpfc_sli_intf_if_type, &phba->sli4_hba.sli_intf) ==
8007 LPFC_SLI_INTF_IF_TYPE_0) {
8008 rc = lpfc_pci_function_reset(phba);
8009 if (unlikely(rc))
8010 goto out_free_bsmbx;
8011 }
James Smartda0436e2009-05-22 14:51:39 -04008012
James Smartcb5172e2010-03-15 11:25:07 -04008013 mboxq = (LPFC_MBOXQ_t *) mempool_alloc(phba->mbox_mem_pool,
8014 GFP_KERNEL);
8015 if (!mboxq) {
8016 rc = -ENOMEM;
8017 goto out_free_bsmbx;
8018 }
8019
James Smartf358dd02017-02-12 13:52:34 -08008020 /* Check for NVMET being configured */
James Smart895427b2017-02-12 13:52:30 -08008021 phba->nvmet_support = 0;
James Smartf358dd02017-02-12 13:52:34 -08008022 if (lpfc_enable_nvmet_cnt) {
8023
8024 /* First get WWN of HBA instance */
8025 lpfc_read_nv(phba, mboxq);
8026 rc = lpfc_sli_issue_mbox(phba, mboxq, MBX_POLL);
8027 if (rc != MBX_SUCCESS) {
Dick Kennedy372c1872020-06-30 14:50:00 -07008028 lpfc_printf_log(phba, KERN_ERR,
8029 LOG_TRACE_EVENT,
James Smartf358dd02017-02-12 13:52:34 -08008030 "6016 Mailbox failed , mbxCmd x%x "
8031 "READ_NV, mbxStatus x%x\n",
8032 bf_get(lpfc_mqe_command, &mboxq->u.mqe),
8033 bf_get(lpfc_mqe_status, &mboxq->u.mqe));
James Smartd1f525a2017-04-21 16:04:55 -07008034 mempool_free(mboxq, phba->mbox_mem_pool);
James Smartf358dd02017-02-12 13:52:34 -08008035 rc = -EIO;
8036 goto out_free_bsmbx;
8037 }
8038 mb = &mboxq->u.mb;
8039 memcpy(&wwn, (char *)mb->un.varRDnvp.nodename,
8040 sizeof(uint64_t));
8041 wwn = cpu_to_be64(wwn);
8042 phba->sli4_hba.wwnn.u.name = wwn;
8043 memcpy(&wwn, (char *)mb->un.varRDnvp.portname,
8044 sizeof(uint64_t));
8045 /* wwn is WWPN of HBA instance */
8046 wwn = cpu_to_be64(wwn);
8047 phba->sli4_hba.wwpn.u.name = wwn;
8048
8049 /* Check to see if it matches any module parameter */
8050 for (i = 0; i < lpfc_enable_nvmet_cnt; i++) {
8051 if (wwn == lpfc_enable_nvmet[i]) {
James Smart7d708032017-03-08 14:36:01 -08008052#if (IS_ENABLED(CONFIG_NVME_TARGET_FC))
James Smart3c603be2017-05-15 15:20:44 -07008053 if (lpfc_nvmet_mem_alloc(phba))
8054 break;
8055
8056 phba->nvmet_support = 1; /* a match */
8057
Dick Kennedy372c1872020-06-30 14:50:00 -07008058 lpfc_printf_log(phba, KERN_ERR,
8059 LOG_TRACE_EVENT,
James Smartf358dd02017-02-12 13:52:34 -08008060 "6017 NVME Target %016llx\n",
8061 wwn);
James Smart7d708032017-03-08 14:36:01 -08008062#else
Dick Kennedy372c1872020-06-30 14:50:00 -07008063 lpfc_printf_log(phba, KERN_ERR,
8064 LOG_TRACE_EVENT,
James Smart7d708032017-03-08 14:36:01 -08008065 "6021 Can't enable NVME Target."
8066 " NVME_TARGET_FC infrastructure"
8067 " is not in kernel\n");
8068#endif
James Smartc4908502019-01-28 11:14:28 -08008069 /* Not supported for NVMET */
8070 phba->cfg_xri_rebalancing = 0;
Dick Kennedy3048e3e2020-05-01 14:43:06 -07008071 if (phba->irq_chann_mode == NHT_MODE) {
8072 phba->cfg_irq_chann =
8073 phba->sli4_hba.num_present_cpu;
8074 phba->cfg_hdw_queue =
8075 phba->sli4_hba.num_present_cpu;
8076 phba->irq_chann_mode = NORMAL_MODE;
8077 }
James Smart3c603be2017-05-15 15:20:44 -07008078 break;
James Smartf358dd02017-02-12 13:52:34 -08008079 }
8080 }
8081 }
James Smart895427b2017-02-12 13:52:30 -08008082
8083 lpfc_nvme_mod_param_dep(phba);
8084
James Smartfedd3b72011-02-16 12:39:24 -05008085 /*
8086 * Get sli4 parameters that override parameters from Port capabilities.
James Smart6d368e52011-05-24 11:44:12 -04008087 * If this call fails, it isn't critical unless the SLI4 parameters come
8088 * back in conflict.
James Smartfedd3b72011-02-16 12:39:24 -05008089 */
James Smart6d368e52011-05-24 11:44:12 -04008090 rc = lpfc_get_sli4_parameters(phba, mboxq);
8091 if (rc) {
James Smartb92dc722018-05-24 21:09:01 -07008092 if_type = bf_get(lpfc_sli_intf_if_type,
8093 &phba->sli4_hba.sli_intf);
8094 if_fam = bf_get(lpfc_sli_intf_sli_family,
8095 &phba->sli4_hba.sli_intf);
James Smart6d368e52011-05-24 11:44:12 -04008096 if (phba->sli4_hba.extents_in_use &&
8097 phba->sli4_hba.rpi_hdrs_in_use) {
Dick Kennedy372c1872020-06-30 14:50:00 -07008098 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
8099 "2999 Unsupported SLI4 Parameters "
8100 "Extents and RPI headers enabled.\n");
James Smartb92dc722018-05-24 21:09:01 -07008101 if (if_type == LPFC_SLI_INTF_IF_TYPE_0 &&
8102 if_fam == LPFC_SLI_INTF_FAMILY_BE2) {
8103 mempool_free(mboxq, phba->mbox_mem_pool);
8104 rc = -EIO;
8105 goto out_free_bsmbx;
8106 }
James Smart6d368e52011-05-24 11:44:12 -04008107 }
James Smartb92dc722018-05-24 21:09:01 -07008108 if (!(if_type == LPFC_SLI_INTF_IF_TYPE_0 &&
8109 if_fam == LPFC_SLI_INTF_FAMILY_BE2)) {
8110 mempool_free(mboxq, phba->mbox_mem_pool);
8111 rc = -EIO;
8112 goto out_free_bsmbx;
8113 }
James Smart6d368e52011-05-24 11:44:12 -04008114 }
James Smart895427b2017-02-12 13:52:30 -08008115
James Smartd79c9e92019-08-14 16:57:09 -07008116 /*
8117 * 1 for cmd, 1 for rsp, NVME adds an extra one
8118 * for boundary conditions in its max_sgl_segment template.
8119 */
8120 extra = 2;
8121 if (phba->cfg_enable_fc4_type & LPFC_ENABLE_NVME)
8122 extra++;
8123
8124 /*
8125 * It doesn't matter what family our adapter is in, we are
8126 * limited to 2 Pages, 512 SGEs, for our SGL.
8127 * There are going to be 2 reserved SGEs: 1 FCP cmnd + 1 FCP rsp
8128 */
8129 max_buf_size = (2 * SLI4_PAGE_SIZE);
8130
8131 /*
8132 * Since lpfc_sg_seg_cnt is module param, the sg_dma_buf_size
8133 * used to create the sg_dma_buf_pool must be calculated.
8134 */
8135 if (phba->sli3_options & LPFC_SLI3_BG_ENABLED) {
8136 /* Both cfg_enable_bg and cfg_external_dif code paths */
8137
8138 /*
8139 * The scsi_buf for a T10-DIF I/O holds the FCP cmnd,
8140 * the FCP rsp, and a SGE. Sice we have no control
8141 * over how many protection segments the SCSI Layer
8142 * will hand us (ie: there could be one for every block
8143 * in the IO), just allocate enough SGEs to accomidate
8144 * our max amount and we need to limit lpfc_sg_seg_cnt
8145 * to minimize the risk of running out.
8146 */
8147 phba->cfg_sg_dma_buf_size = sizeof(struct fcp_cmnd) +
8148 sizeof(struct fcp_rsp) + max_buf_size;
8149
8150 /* Total SGEs for scsi_sg_list and scsi_sg_prot_list */
8151 phba->cfg_total_seg_cnt = LPFC_MAX_SGL_SEG_CNT;
8152
8153 /*
8154 * If supporting DIF, reduce the seg count for scsi to
8155 * allow room for the DIF sges.
8156 */
8157 if (phba->cfg_enable_bg &&
8158 phba->cfg_sg_seg_cnt > LPFC_MAX_BG_SLI4_SEG_CNT_DIF)
8159 phba->cfg_scsi_seg_cnt = LPFC_MAX_BG_SLI4_SEG_CNT_DIF;
8160 else
8161 phba->cfg_scsi_seg_cnt = phba->cfg_sg_seg_cnt;
8162
8163 } else {
8164 /*
8165 * The scsi_buf for a regular I/O holds the FCP cmnd,
8166 * the FCP rsp, a SGE for each, and a SGE for up to
8167 * cfg_sg_seg_cnt data segments.
8168 */
8169 phba->cfg_sg_dma_buf_size = sizeof(struct fcp_cmnd) +
8170 sizeof(struct fcp_rsp) +
8171 ((phba->cfg_sg_seg_cnt + extra) *
8172 sizeof(struct sli4_sge));
8173
8174 /* Total SGEs for scsi_sg_list */
8175 phba->cfg_total_seg_cnt = phba->cfg_sg_seg_cnt + extra;
8176 phba->cfg_scsi_seg_cnt = phba->cfg_sg_seg_cnt;
8177
8178 /*
8179 * NOTE: if (phba->cfg_sg_seg_cnt + extra) <= 256 we only
8180 * need to post 1 page for the SGL.
8181 */
8182 }
8183
8184 if (phba->cfg_xpsgl && !phba->nvmet_support)
8185 phba->cfg_sg_dma_buf_size = LPFC_DEFAULT_XPSGL_SIZE;
8186 else if (phba->cfg_sg_dma_buf_size <= LPFC_MIN_SG_SLI4_BUF_SZ)
8187 phba->cfg_sg_dma_buf_size = LPFC_MIN_SG_SLI4_BUF_SZ;
8188 else
8189 phba->cfg_sg_dma_buf_size =
8190 SLI4_PAGE_ALIGN(phba->cfg_sg_dma_buf_size);
8191
8192 phba->border_sge_num = phba->cfg_sg_dma_buf_size /
8193 sizeof(struct sli4_sge);
8194
8195 /* Limit to LPFC_MAX_NVME_SEG_CNT for NVME. */
8196 if (phba->cfg_enable_fc4_type & LPFC_ENABLE_NVME) {
8197 if (phba->cfg_sg_seg_cnt > LPFC_MAX_NVME_SEG_CNT) {
8198 lpfc_printf_log(phba, KERN_INFO, LOG_NVME | LOG_INIT,
8199 "6300 Reducing NVME sg segment "
8200 "cnt to %d\n",
8201 LPFC_MAX_NVME_SEG_CNT);
8202 phba->cfg_nvme_seg_cnt = LPFC_MAX_NVME_SEG_CNT;
8203 } else
8204 phba->cfg_nvme_seg_cnt = phba->cfg_sg_seg_cnt;
8205 }
8206
James Smartd79c9e92019-08-14 16:57:09 -07008207 lpfc_printf_log(phba, KERN_INFO, LOG_INIT | LOG_FCP,
8208 "9087 sg_seg_cnt:%d dmabuf_size:%d "
8209 "total:%d scsi:%d nvme:%d\n",
8210 phba->cfg_sg_seg_cnt, phba->cfg_sg_dma_buf_size,
8211 phba->cfg_total_seg_cnt, phba->cfg_scsi_seg_cnt,
8212 phba->cfg_nvme_seg_cnt);
8213
8214 if (phba->cfg_sg_dma_buf_size < SLI4_PAGE_SIZE)
8215 i = phba->cfg_sg_dma_buf_size;
8216 else
8217 i = SLI4_PAGE_SIZE;
8218
8219 phba->lpfc_sg_dma_buf_pool =
8220 dma_pool_create("lpfc_sg_dma_buf_pool",
8221 &phba->pcidev->dev,
8222 phba->cfg_sg_dma_buf_size,
8223 i, 0);
8224 if (!phba->lpfc_sg_dma_buf_pool)
8225 goto out_free_bsmbx;
8226
8227 phba->lpfc_cmd_rsp_buf_pool =
8228 dma_pool_create("lpfc_cmd_rsp_buf_pool",
8229 &phba->pcidev->dev,
8230 sizeof(struct fcp_cmnd) +
8231 sizeof(struct fcp_rsp),
8232 i, 0);
8233 if (!phba->lpfc_cmd_rsp_buf_pool)
8234 goto out_free_sg_dma_buf;
8235
James Smartcb5172e2010-03-15 11:25:07 -04008236 mempool_free(mboxq, phba->mbox_mem_pool);
James Smart1ba981f2014-02-20 09:56:45 -05008237
8238 /* Verify OAS is supported */
8239 lpfc_sli4_oas_verify(phba);
James Smart1ba981f2014-02-20 09:56:45 -05008240
James Smartd2cc9bc2018-09-10 10:30:50 -07008241 /* Verify RAS support on adapter */
8242 lpfc_sli4_ras_init(phba);
8243
James Smart5350d872011-10-10 21:33:49 -04008244 /* Verify all the SLI4 queues */
8245 rc = lpfc_sli4_queue_verify(phba);
James Smartda0436e2009-05-22 14:51:39 -04008246 if (rc)
James Smartd79c9e92019-08-14 16:57:09 -07008247 goto out_free_cmd_rsp_buf;
James Smartda0436e2009-05-22 14:51:39 -04008248
8249 /* Create driver internal CQE event pool */
8250 rc = lpfc_sli4_cq_event_pool_create(phba);
8251 if (rc)
James Smartd79c9e92019-08-14 16:57:09 -07008252 goto out_free_cmd_rsp_buf;
James Smartda0436e2009-05-22 14:51:39 -04008253
James Smart8a9d2e82012-05-09 21:16:12 -04008254 /* Initialize sgl lists per host */
8255 lpfc_init_sgl_list(phba);
8256
8257 /* Allocate and initialize active sgl array */
James Smartda0436e2009-05-22 14:51:39 -04008258 rc = lpfc_init_active_sgl_array(phba);
8259 if (rc) {
Dick Kennedy372c1872020-06-30 14:50:00 -07008260 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartda0436e2009-05-22 14:51:39 -04008261 "1430 Failed to initialize sgl list.\n");
James Smart8a9d2e82012-05-09 21:16:12 -04008262 goto out_destroy_cq_event_pool;
James Smartda0436e2009-05-22 14:51:39 -04008263 }
James Smartda0436e2009-05-22 14:51:39 -04008264 rc = lpfc_sli4_init_rpi_hdrs(phba);
8265 if (rc) {
Dick Kennedy372c1872020-06-30 14:50:00 -07008266 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartda0436e2009-05-22 14:51:39 -04008267 "1432 Failed to initialize rpi headers.\n");
8268 goto out_free_active_sgl;
8269 }
8270
James Smarta93ff372010-10-22 11:06:08 -04008271 /* Allocate eligible FCF bmask memory for FCF roundrobin failover */
James Smart0c9ab6f2010-02-26 14:15:57 -05008272 longs = (LPFC_SLI4_FCF_TBL_INDX_MAX + BITS_PER_LONG - 1)/BITS_PER_LONG;
Kees Cook6396bb22018-06-12 14:03:40 -07008273 phba->fcf.fcf_rr_bmask = kcalloc(longs, sizeof(unsigned long),
James Smart0c9ab6f2010-02-26 14:15:57 -05008274 GFP_KERNEL);
8275 if (!phba->fcf.fcf_rr_bmask) {
Dick Kennedy372c1872020-06-30 14:50:00 -07008276 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart0c9ab6f2010-02-26 14:15:57 -05008277 "2759 Failed allocate memory for FCF round "
8278 "robin failover bmask\n");
James Smart05580562011-05-24 11:40:48 -04008279 rc = -ENOMEM;
James Smart0c9ab6f2010-02-26 14:15:57 -05008280 goto out_remove_rpi_hdrs;
8281 }
8282
James Smart6a828b02019-01-28 11:14:31 -08008283 phba->sli4_hba.hba_eq_hdl = kcalloc(phba->cfg_irq_chann,
James Smartcdb42be2019-01-28 11:14:21 -08008284 sizeof(struct lpfc_hba_eq_hdl),
8285 GFP_KERNEL);
James Smart895427b2017-02-12 13:52:30 -08008286 if (!phba->sli4_hba.hba_eq_hdl) {
Dick Kennedy372c1872020-06-30 14:50:00 -07008287 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart67d12732012-08-03 12:36:13 -04008288 "2572 Failed allocate memory for "
8289 "fast-path per-EQ handle array\n");
8290 rc = -ENOMEM;
8291 goto out_free_fcf_rr_bmask;
James Smartda0436e2009-05-22 14:51:39 -04008292 }
8293
James Smart222e9232019-01-28 11:14:35 -08008294 phba->sli4_hba.cpu_map = kcalloc(phba->sli4_hba.num_possible_cpu,
James Smart895427b2017-02-12 13:52:30 -08008295 sizeof(struct lpfc_vector_map_info),
8296 GFP_KERNEL);
James Smart7bb03bb2013-04-17 20:19:16 -04008297 if (!phba->sli4_hba.cpu_map) {
Dick Kennedy372c1872020-06-30 14:50:00 -07008298 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart7bb03bb2013-04-17 20:19:16 -04008299 "3327 Failed allocate memory for msi-x "
8300 "interrupt vector mapping\n");
8301 rc = -ENOMEM;
James Smart895427b2017-02-12 13:52:30 -08008302 goto out_free_hba_eq_hdl;
James Smart7bb03bb2013-04-17 20:19:16 -04008303 }
James Smartb246de12013-05-31 17:03:07 -04008304
James Smart32517fc2019-01-28 11:14:33 -08008305 phba->sli4_hba.eq_info = alloc_percpu(struct lpfc_eq_intr_info);
8306 if (!phba->sli4_hba.eq_info) {
Dick Kennedy372c1872020-06-30 14:50:00 -07008307 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart32517fc2019-01-28 11:14:33 -08008308 "3321 Failed allocation for per_cpu stats\n");
8309 rc = -ENOMEM;
8310 goto out_free_hba_cpu_map;
8311 }
James Smart840eda92020-03-22 11:13:00 -07008312
Dick Kennedy317aeb82020-06-30 14:49:59 -07008313 phba->sli4_hba.idle_stat = kcalloc(phba->sli4_hba.num_possible_cpu,
8314 sizeof(*phba->sli4_hba.idle_stat),
8315 GFP_KERNEL);
8316 if (!phba->sli4_hba.idle_stat) {
Dick Kennedy372c1872020-06-30 14:50:00 -07008317 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
Dick Kennedy317aeb82020-06-30 14:49:59 -07008318 "3390 Failed allocation for idle_stat\n");
8319 rc = -ENOMEM;
8320 goto out_free_hba_eq_info;
8321 }
8322
James Smart840eda92020-03-22 11:13:00 -07008323#ifdef CONFIG_SCSI_LPFC_DEBUG_FS
8324 phba->sli4_hba.c_stat = alloc_percpu(struct lpfc_hdwq_stat);
8325 if (!phba->sli4_hba.c_stat) {
Dick Kennedy372c1872020-06-30 14:50:00 -07008326 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart840eda92020-03-22 11:13:00 -07008327 "3332 Failed allocating per cpu hdwq stats\n");
8328 rc = -ENOMEM;
Dick Kennedy317aeb82020-06-30 14:49:59 -07008329 goto out_free_hba_idle_stat;
James Smart840eda92020-03-22 11:13:00 -07008330 }
8331#endif
8332
James Smart02243832021-08-16 09:28:54 -07008333 phba->cmf_stat = alloc_percpu(struct lpfc_cgn_stat);
8334 if (!phba->cmf_stat) {
8335 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
8336 "3331 Failed allocating per cpu cgn stats\n");
8337 rc = -ENOMEM;
8338 goto out_free_hba_hdwq_info;
8339 }
8340
James Smart912e3ac2011-05-24 11:42:11 -04008341 /*
8342 * Enable sr-iov virtual functions if supported and configured
8343 * through the module parameter.
8344 */
8345 if (phba->cfg_sriov_nr_virtfn > 0) {
8346 rc = lpfc_sli_probe_sriov_nr_virtfn(phba,
8347 phba->cfg_sriov_nr_virtfn);
8348 if (rc) {
8349 lpfc_printf_log(phba, KERN_WARNING, LOG_INIT,
8350 "3020 Requested number of SR-IOV "
8351 "virtual functions (%d) is not "
8352 "supported\n",
8353 phba->cfg_sriov_nr_virtfn);
8354 phba->cfg_sriov_nr_virtfn = 0;
8355 }
8356 }
8357
James Smart5248a742011-07-22 18:37:06 -04008358 return 0;
James Smartda0436e2009-05-22 14:51:39 -04008359
James Smart02243832021-08-16 09:28:54 -07008360out_free_hba_hdwq_info:
James Smart840eda92020-03-22 11:13:00 -07008361#ifdef CONFIG_SCSI_LPFC_DEBUG_FS
James Smart37e38402021-09-07 22:09:27 -07008362 free_percpu(phba->sli4_hba.c_stat);
Dick Kennedy317aeb82020-06-30 14:49:59 -07008363out_free_hba_idle_stat:
Dick Kennedy317aeb82020-06-30 14:49:59 -07008364#endif
James Smart37e38402021-09-07 22:09:27 -07008365 kfree(phba->sli4_hba.idle_stat);
James Smart840eda92020-03-22 11:13:00 -07008366out_free_hba_eq_info:
8367 free_percpu(phba->sli4_hba.eq_info);
James Smart32517fc2019-01-28 11:14:33 -08008368out_free_hba_cpu_map:
8369 kfree(phba->sli4_hba.cpu_map);
James Smart895427b2017-02-12 13:52:30 -08008370out_free_hba_eq_hdl:
8371 kfree(phba->sli4_hba.hba_eq_hdl);
James Smart0c9ab6f2010-02-26 14:15:57 -05008372out_free_fcf_rr_bmask:
8373 kfree(phba->fcf.fcf_rr_bmask);
James Smartda0436e2009-05-22 14:51:39 -04008374out_remove_rpi_hdrs:
8375 lpfc_sli4_remove_rpi_hdrs(phba);
8376out_free_active_sgl:
8377 lpfc_free_active_sgl(phba);
James Smartda0436e2009-05-22 14:51:39 -04008378out_destroy_cq_event_pool:
8379 lpfc_sli4_cq_event_pool_destroy(phba);
James Smartd79c9e92019-08-14 16:57:09 -07008380out_free_cmd_rsp_buf:
8381 dma_pool_destroy(phba->lpfc_cmd_rsp_buf_pool);
8382 phba->lpfc_cmd_rsp_buf_pool = NULL;
8383out_free_sg_dma_buf:
8384 dma_pool_destroy(phba->lpfc_sg_dma_buf_pool);
8385 phba->lpfc_sg_dma_buf_pool = NULL;
James Smartda0436e2009-05-22 14:51:39 -04008386out_free_bsmbx:
8387 lpfc_destroy_bootstrap_mbox(phba);
8388out_free_mem:
8389 lpfc_mem_free(phba);
8390 return rc;
8391}
8392
8393/**
8394 * lpfc_sli4_driver_resource_unset - Unset drvr internal resources for SLI4 dev
8395 * @phba: pointer to lpfc hba data structure.
8396 *
8397 * This routine is invoked to unset the driver internal resources set up
8398 * specific for supporting the SLI-4 HBA device it attached to.
8399 **/
8400static void
8401lpfc_sli4_driver_resource_unset(struct lpfc_hba *phba)
8402{
8403 struct lpfc_fcf_conn_entry *conn_entry, *next_conn_entry;
8404
James Smart32517fc2019-01-28 11:14:33 -08008405 free_percpu(phba->sli4_hba.eq_info);
James Smart840eda92020-03-22 11:13:00 -07008406#ifdef CONFIG_SCSI_LPFC_DEBUG_FS
8407 free_percpu(phba->sli4_hba.c_stat);
8408#endif
James Smart02243832021-08-16 09:28:54 -07008409 free_percpu(phba->cmf_stat);
Dick Kennedy317aeb82020-06-30 14:49:59 -07008410 kfree(phba->sli4_hba.idle_stat);
James Smart32517fc2019-01-28 11:14:33 -08008411
James Smart7bb03bb2013-04-17 20:19:16 -04008412 /* Free memory allocated for msi-x interrupt vector to CPU mapping */
8413 kfree(phba->sli4_hba.cpu_map);
James Smart222e9232019-01-28 11:14:35 -08008414 phba->sli4_hba.num_possible_cpu = 0;
James Smart7bb03bb2013-04-17 20:19:16 -04008415 phba->sli4_hba.num_present_cpu = 0;
James Smart76fd07a2014-02-20 09:57:18 -05008416 phba->sli4_hba.curr_disp_cpu = 0;
Dick Kennedy3048e3e2020-05-01 14:43:06 -07008417 cpumask_clear(&phba->sli4_hba.irq_aff_mask);
James Smart7bb03bb2013-04-17 20:19:16 -04008418
James Smartda0436e2009-05-22 14:51:39 -04008419 /* Free memory allocated for fast-path work queue handles */
James Smart895427b2017-02-12 13:52:30 -08008420 kfree(phba->sli4_hba.hba_eq_hdl);
James Smartda0436e2009-05-22 14:51:39 -04008421
8422 /* Free the allocated rpi headers. */
8423 lpfc_sli4_remove_rpi_hdrs(phba);
James Smartd11e31d2009-06-10 17:23:06 -04008424 lpfc_sli4_remove_rpis(phba);
James Smartda0436e2009-05-22 14:51:39 -04008425
James Smart0c9ab6f2010-02-26 14:15:57 -05008426 /* Free eligible FCF index bmask */
8427 kfree(phba->fcf.fcf_rr_bmask);
8428
James Smartda0436e2009-05-22 14:51:39 -04008429 /* Free the ELS sgl list */
8430 lpfc_free_active_sgl(phba);
James Smart8a9d2e82012-05-09 21:16:12 -04008431 lpfc_free_els_sgl_list(phba);
James Smartf358dd02017-02-12 13:52:34 -08008432 lpfc_free_nvmet_sgl_list(phba);
James Smartda0436e2009-05-22 14:51:39 -04008433
James Smartda0436e2009-05-22 14:51:39 -04008434 /* Free the completion queue EQ event pool */
8435 lpfc_sli4_cq_event_release_all(phba);
8436 lpfc_sli4_cq_event_pool_destroy(phba);
8437
James Smart6d368e52011-05-24 11:44:12 -04008438 /* Release resource identifiers. */
8439 lpfc_sli4_dealloc_resource_identifiers(phba);
8440
James Smartda0436e2009-05-22 14:51:39 -04008441 /* Free the bsmbx region. */
8442 lpfc_destroy_bootstrap_mbox(phba);
8443
8444 /* Free the SLI Layer memory with SLI4 HBAs */
8445 lpfc_mem_free_all(phba);
8446
8447 /* Free the current connect table */
8448 list_for_each_entry_safe(conn_entry, next_conn_entry,
James Smart4d9ab992009-10-02 15:16:39 -04008449 &phba->fcf_conn_rec_list, list) {
8450 list_del_init(&conn_entry->list);
James Smartda0436e2009-05-22 14:51:39 -04008451 kfree(conn_entry);
James Smart4d9ab992009-10-02 15:16:39 -04008452 }
James Smartda0436e2009-05-22 14:51:39 -04008453
8454 return;
8455}
8456
8457/**
Lucas De Marchi25985ed2011-03-30 22:57:33 -03008458 * lpfc_init_api_table_setup - Set up init api function jump table
James Smart3772a992009-05-22 14:50:54 -04008459 * @phba: The hba struct for which this call is being executed.
8460 * @dev_grp: The HBA PCI-Device group number.
8461 *
8462 * This routine sets up the device INIT interface API function jump table
8463 * in @phba struct.
8464 *
8465 * Returns: 0 - success, -ENODEV - failure.
8466 **/
8467int
8468lpfc_init_api_table_setup(struct lpfc_hba *phba, uint8_t dev_grp)
8469{
James Smart84d1b002010-02-12 14:42:33 -05008470 phba->lpfc_hba_init_link = lpfc_hba_init_link;
8471 phba->lpfc_hba_down_link = lpfc_hba_down_link;
James Smart7f860592011-03-11 16:05:52 -05008472 phba->lpfc_selective_reset = lpfc_selective_reset;
James Smart3772a992009-05-22 14:50:54 -04008473 switch (dev_grp) {
8474 case LPFC_PCI_DEV_LP:
8475 phba->lpfc_hba_down_post = lpfc_hba_down_post_s3;
8476 phba->lpfc_handle_eratt = lpfc_handle_eratt_s3;
8477 phba->lpfc_stop_port = lpfc_stop_port_s3;
8478 break;
James Smartda0436e2009-05-22 14:51:39 -04008479 case LPFC_PCI_DEV_OC:
8480 phba->lpfc_hba_down_post = lpfc_hba_down_post_s4;
8481 phba->lpfc_handle_eratt = lpfc_handle_eratt_s4;
8482 phba->lpfc_stop_port = lpfc_stop_port_s4;
8483 break;
James Smart3772a992009-05-22 14:50:54 -04008484 default:
James Smarta5160742021-10-20 14:14:10 -07008485 lpfc_printf_log(phba, KERN_ERR, LOG_INIT,
James Smart3772a992009-05-22 14:50:54 -04008486 "1431 Invalid HBA PCI-device group: 0x%x\n",
8487 dev_grp);
8488 return -ENODEV;
James Smart3772a992009-05-22 14:50:54 -04008489 }
8490 return 0;
8491}
8492
8493/**
James Smart3772a992009-05-22 14:50:54 -04008494 * lpfc_setup_driver_resource_phase2 - Phase2 setup driver internal resources.
8495 * @phba: pointer to lpfc hba data structure.
8496 *
8497 * This routine is invoked to set up the driver internal resources after the
8498 * device specific resource setup to support the HBA device it attached to.
8499 *
8500 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -02008501 * 0 - successful
James Smart3772a992009-05-22 14:50:54 -04008502 * other values - error
8503 **/
8504static int
8505lpfc_setup_driver_resource_phase2(struct lpfc_hba *phba)
8506{
8507 int error;
8508
8509 /* Startup the kernel thread for this host adapter. */
8510 phba->worker_thread = kthread_run(lpfc_do_work, phba,
8511 "lpfc_worker_%d", phba->brd_no);
8512 if (IS_ERR(phba->worker_thread)) {
8513 error = PTR_ERR(phba->worker_thread);
8514 return error;
Jamie Wellnitz901a9202006-02-28 19:25:19 -05008515 }
8516
James Smart3772a992009-05-22 14:50:54 -04008517 return 0;
8518}
8519
8520/**
8521 * lpfc_unset_driver_resource_phase2 - Phase2 unset driver internal resources.
8522 * @phba: pointer to lpfc hba data structure.
8523 *
8524 * This routine is invoked to unset the driver internal resources set up after
8525 * the device specific resource setup for supporting the HBA device it
8526 * attached to.
8527 **/
8528static void
8529lpfc_unset_driver_resource_phase2(struct lpfc_hba *phba)
8530{
Dick Kennedyf485c182017-09-29 17:34:34 -07008531 if (phba->wq) {
8532 flush_workqueue(phba->wq);
8533 destroy_workqueue(phba->wq);
8534 phba->wq = NULL;
8535 }
8536
James Smart3772a992009-05-22 14:50:54 -04008537 /* Stop kernel worker thread */
James Smart0cdb84e2018-04-09 14:24:26 -07008538 if (phba->worker_thread)
8539 kthread_stop(phba->worker_thread);
James Smart3772a992009-05-22 14:50:54 -04008540}
8541
8542/**
8543 * lpfc_free_iocb_list - Free iocb list.
8544 * @phba: pointer to lpfc hba data structure.
8545 *
8546 * This routine is invoked to free the driver's IOCB list and memory.
8547 **/
James Smart6c621a22017-05-15 15:20:45 -07008548void
James Smart3772a992009-05-22 14:50:54 -04008549lpfc_free_iocb_list(struct lpfc_hba *phba)
8550{
8551 struct lpfc_iocbq *iocbq_entry = NULL, *iocbq_next = NULL;
8552
8553 spin_lock_irq(&phba->hbalock);
8554 list_for_each_entry_safe(iocbq_entry, iocbq_next,
8555 &phba->lpfc_iocb_list, list) {
8556 list_del(&iocbq_entry->list);
8557 kfree(iocbq_entry);
8558 phba->total_iocbq_bufs--;
Jamie Wellnitz901a9202006-02-28 19:25:19 -05008559 }
James Smart3772a992009-05-22 14:50:54 -04008560 spin_unlock_irq(&phba->hbalock);
dea31012005-04-17 16:05:31 -05008561
James Smart3772a992009-05-22 14:50:54 -04008562 return;
8563}
dea31012005-04-17 16:05:31 -05008564
James Smart3772a992009-05-22 14:50:54 -04008565/**
8566 * lpfc_init_iocb_list - Allocate and initialize iocb list.
8567 * @phba: pointer to lpfc hba data structure.
Lee Jonesfe614ac2020-07-23 13:24:22 +01008568 * @iocb_count: number of requested iocbs
James Smart3772a992009-05-22 14:50:54 -04008569 *
8570 * This routine is invoked to allocate and initizlize the driver's IOCB
8571 * list and set up the IOCB tag array accordingly.
8572 *
8573 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -02008574 * 0 - successful
James Smart3772a992009-05-22 14:50:54 -04008575 * other values - error
8576 **/
James Smart6c621a22017-05-15 15:20:45 -07008577int
James Smart3772a992009-05-22 14:50:54 -04008578lpfc_init_iocb_list(struct lpfc_hba *phba, int iocb_count)
8579{
8580 struct lpfc_iocbq *iocbq_entry = NULL;
8581 uint16_t iotag;
8582 int i;
dea31012005-04-17 16:05:31 -05008583
8584 /* Initialize and populate the iocb list per host. */
8585 INIT_LIST_HEAD(&phba->lpfc_iocb_list);
James Smart3772a992009-05-22 14:50:54 -04008586 for (i = 0; i < iocb_count; i++) {
Yoann Padioleaudd00cc42007-07-19 01:49:03 -07008587 iocbq_entry = kzalloc(sizeof(struct lpfc_iocbq), GFP_KERNEL);
dea31012005-04-17 16:05:31 -05008588 if (iocbq_entry == NULL) {
8589 printk(KERN_ERR "%s: only allocated %d iocbs of "
8590 "expected %d count. Unloading driver.\n",
James Smarta5f73372019-09-21 20:58:50 -07008591 __func__, i, iocb_count);
dea31012005-04-17 16:05:31 -05008592 goto out_free_iocbq;
8593 }
8594
James Bottomley604a3e32005-10-29 10:28:33 -05008595 iotag = lpfc_sli_next_iotag(phba, iocbq_entry);
8596 if (iotag == 0) {
James Smart3772a992009-05-22 14:50:54 -04008597 kfree(iocbq_entry);
James Bottomley604a3e32005-10-29 10:28:33 -05008598 printk(KERN_ERR "%s: failed to allocate IOTAG. "
James Smart3772a992009-05-22 14:50:54 -04008599 "Unloading driver.\n", __func__);
James Bottomley604a3e32005-10-29 10:28:33 -05008600 goto out_free_iocbq;
8601 }
James Smart6d368e52011-05-24 11:44:12 -04008602 iocbq_entry->sli4_lxritag = NO_XRI;
James Smart3772a992009-05-22 14:50:54 -04008603 iocbq_entry->sli4_xritag = NO_XRI;
James Smart2e0fef82007-06-17 19:56:36 -05008604
8605 spin_lock_irq(&phba->hbalock);
dea31012005-04-17 16:05:31 -05008606 list_add(&iocbq_entry->list, &phba->lpfc_iocb_list);
8607 phba->total_iocbq_bufs++;
James Smart2e0fef82007-06-17 19:56:36 -05008608 spin_unlock_irq(&phba->hbalock);
dea31012005-04-17 16:05:31 -05008609 }
8610
James Smart3772a992009-05-22 14:50:54 -04008611 return 0;
8612
8613out_free_iocbq:
8614 lpfc_free_iocb_list(phba);
8615
8616 return -ENOMEM;
8617}
8618
8619/**
James Smart8a9d2e82012-05-09 21:16:12 -04008620 * lpfc_free_sgl_list - Free a given sgl list.
James Smartda0436e2009-05-22 14:51:39 -04008621 * @phba: pointer to lpfc hba data structure.
James Smart8a9d2e82012-05-09 21:16:12 -04008622 * @sglq_list: pointer to the head of sgl list.
James Smartda0436e2009-05-22 14:51:39 -04008623 *
James Smart8a9d2e82012-05-09 21:16:12 -04008624 * This routine is invoked to free a give sgl list and memory.
James Smartda0436e2009-05-22 14:51:39 -04008625 **/
James Smart8a9d2e82012-05-09 21:16:12 -04008626void
8627lpfc_free_sgl_list(struct lpfc_hba *phba, struct list_head *sglq_list)
James Smartda0436e2009-05-22 14:51:39 -04008628{
8629 struct lpfc_sglq *sglq_entry = NULL, *sglq_next = NULL;
James Smart8a9d2e82012-05-09 21:16:12 -04008630
8631 list_for_each_entry_safe(sglq_entry, sglq_next, sglq_list, list) {
8632 list_del(&sglq_entry->list);
8633 lpfc_mbuf_free(phba, sglq_entry->virt, sglq_entry->phys);
8634 kfree(sglq_entry);
8635 }
8636}
8637
8638/**
8639 * lpfc_free_els_sgl_list - Free els sgl list.
8640 * @phba: pointer to lpfc hba data structure.
8641 *
8642 * This routine is invoked to free the driver's els sgl list and memory.
8643 **/
8644static void
8645lpfc_free_els_sgl_list(struct lpfc_hba *phba)
8646{
James Smartda0436e2009-05-22 14:51:39 -04008647 LIST_HEAD(sglq_list);
James Smartda0436e2009-05-22 14:51:39 -04008648
James Smart8a9d2e82012-05-09 21:16:12 -04008649 /* Retrieve all els sgls from driver list */
James Smarta7892412021-04-11 18:31:15 -07008650 spin_lock_irq(&phba->sli4_hba.sgl_list_lock);
James Smart895427b2017-02-12 13:52:30 -08008651 list_splice_init(&phba->sli4_hba.lpfc_els_sgl_list, &sglq_list);
James Smarta7892412021-04-11 18:31:15 -07008652 spin_unlock_irq(&phba->sli4_hba.sgl_list_lock);
James Smartda0436e2009-05-22 14:51:39 -04008653
James Smart8a9d2e82012-05-09 21:16:12 -04008654 /* Now free the sgl list */
8655 lpfc_free_sgl_list(phba, &sglq_list);
James Smartda0436e2009-05-22 14:51:39 -04008656}
8657
8658/**
James Smartf358dd02017-02-12 13:52:34 -08008659 * lpfc_free_nvmet_sgl_list - Free nvmet sgl list.
8660 * @phba: pointer to lpfc hba data structure.
8661 *
8662 * This routine is invoked to free the driver's nvmet sgl list and memory.
8663 **/
8664static void
8665lpfc_free_nvmet_sgl_list(struct lpfc_hba *phba)
8666{
8667 struct lpfc_sglq *sglq_entry = NULL, *sglq_next = NULL;
8668 LIST_HEAD(sglq_list);
8669
8670 /* Retrieve all nvmet sgls from driver list */
8671 spin_lock_irq(&phba->hbalock);
8672 spin_lock(&phba->sli4_hba.sgl_list_lock);
8673 list_splice_init(&phba->sli4_hba.lpfc_nvmet_sgl_list, &sglq_list);
8674 spin_unlock(&phba->sli4_hba.sgl_list_lock);
8675 spin_unlock_irq(&phba->hbalock);
8676
8677 /* Now free the sgl list */
8678 list_for_each_entry_safe(sglq_entry, sglq_next, &sglq_list, list) {
8679 list_del(&sglq_entry->list);
8680 lpfc_nvmet_buf_free(phba, sglq_entry->virt, sglq_entry->phys);
8681 kfree(sglq_entry);
8682 }
Dick Kennedy4b40d022017-08-23 16:55:38 -07008683
8684 /* Update the nvmet_xri_cnt to reflect no current sgls.
8685 * The next initialization cycle sets the count and allocates
8686 * the sgls over again.
8687 */
8688 phba->sli4_hba.nvmet_xri_cnt = 0;
James Smartf358dd02017-02-12 13:52:34 -08008689}
8690
8691/**
James Smartda0436e2009-05-22 14:51:39 -04008692 * lpfc_init_active_sgl_array - Allocate the buf to track active ELS XRIs.
8693 * @phba: pointer to lpfc hba data structure.
8694 *
8695 * This routine is invoked to allocate the driver's active sgl memory.
8696 * This array will hold the sglq_entry's for active IOs.
8697 **/
8698static int
8699lpfc_init_active_sgl_array(struct lpfc_hba *phba)
8700{
8701 int size;
8702 size = sizeof(struct lpfc_sglq *);
8703 size *= phba->sli4_hba.max_cfg_param.max_xri;
8704
8705 phba->sli4_hba.lpfc_sglq_active_list =
8706 kzalloc(size, GFP_KERNEL);
8707 if (!phba->sli4_hba.lpfc_sglq_active_list)
8708 return -ENOMEM;
8709 return 0;
8710}
8711
8712/**
8713 * lpfc_free_active_sgl - Free the buf that tracks active ELS XRIs.
8714 * @phba: pointer to lpfc hba data structure.
8715 *
8716 * This routine is invoked to walk through the array of active sglq entries
8717 * and free all of the resources.
8718 * This is just a place holder for now.
8719 **/
8720static void
8721lpfc_free_active_sgl(struct lpfc_hba *phba)
8722{
8723 kfree(phba->sli4_hba.lpfc_sglq_active_list);
8724}
8725
8726/**
8727 * lpfc_init_sgl_list - Allocate and initialize sgl list.
8728 * @phba: pointer to lpfc hba data structure.
8729 *
8730 * This routine is invoked to allocate and initizlize the driver's sgl
8731 * list and set up the sgl xritag tag array accordingly.
8732 *
James Smartda0436e2009-05-22 14:51:39 -04008733 **/
James Smart8a9d2e82012-05-09 21:16:12 -04008734static void
James Smartda0436e2009-05-22 14:51:39 -04008735lpfc_init_sgl_list(struct lpfc_hba *phba)
8736{
James Smartda0436e2009-05-22 14:51:39 -04008737 /* Initialize and populate the sglq list per host/VF. */
James Smart895427b2017-02-12 13:52:30 -08008738 INIT_LIST_HEAD(&phba->sli4_hba.lpfc_els_sgl_list);
James Smartda0436e2009-05-22 14:51:39 -04008739 INIT_LIST_HEAD(&phba->sli4_hba.lpfc_abts_els_sgl_list);
James Smartf358dd02017-02-12 13:52:34 -08008740 INIT_LIST_HEAD(&phba->sli4_hba.lpfc_nvmet_sgl_list);
James Smart86c67372017-04-21 16:05:04 -07008741 INIT_LIST_HEAD(&phba->sli4_hba.lpfc_abts_nvmet_ctx_list);
James Smartda0436e2009-05-22 14:51:39 -04008742
James Smart8a9d2e82012-05-09 21:16:12 -04008743 /* els xri-sgl book keeping */
8744 phba->sli4_hba.els_xri_cnt = 0;
James Smartda0436e2009-05-22 14:51:39 -04008745
James Smart895427b2017-02-12 13:52:30 -08008746 /* nvme xri-buffer book keeping */
James Smart5e5b5112019-01-28 11:14:22 -08008747 phba->sli4_hba.io_xri_cnt = 0;
James Smartda0436e2009-05-22 14:51:39 -04008748}
8749
8750/**
8751 * lpfc_sli4_init_rpi_hdrs - Post the rpi header memory region to the port
8752 * @phba: pointer to lpfc hba data structure.
8753 *
8754 * This routine is invoked to post rpi header templates to the
James Smart88a2cfb2011-07-22 18:36:33 -04008755 * port for those SLI4 ports that do not support extents. This routine
James Smartda0436e2009-05-22 14:51:39 -04008756 * posts a PAGE_SIZE memory region to the port to hold up to
James Smart88a2cfb2011-07-22 18:36:33 -04008757 * PAGE_SIZE modulo 64 rpi context headers. This is an initialization routine
8758 * and should be called only when interrupts are disabled.
James Smartda0436e2009-05-22 14:51:39 -04008759 *
8760 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -02008761 * 0 - successful
James Smart88a2cfb2011-07-22 18:36:33 -04008762 * -ERROR - otherwise.
James Smartda0436e2009-05-22 14:51:39 -04008763 **/
8764int
8765lpfc_sli4_init_rpi_hdrs(struct lpfc_hba *phba)
8766{
8767 int rc = 0;
James Smartda0436e2009-05-22 14:51:39 -04008768 struct lpfc_rpi_hdr *rpi_hdr;
8769
8770 INIT_LIST_HEAD(&phba->sli4_hba.lpfc_rpi_hdr_list);
James Smartff78d8f2011-12-13 13:21:35 -05008771 if (!phba->sli4_hba.rpi_hdrs_in_use)
James Smart6d368e52011-05-24 11:44:12 -04008772 return rc;
James Smart6d368e52011-05-24 11:44:12 -04008773 if (phba->sli4_hba.extents_in_use)
8774 return -EIO;
James Smartda0436e2009-05-22 14:51:39 -04008775
8776 rpi_hdr = lpfc_sli4_create_rpi_hdr(phba);
8777 if (!rpi_hdr) {
Dick Kennedy372c1872020-06-30 14:50:00 -07008778 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartda0436e2009-05-22 14:51:39 -04008779 "0391 Error during rpi post operation\n");
8780 lpfc_sli4_remove_rpis(phba);
8781 rc = -ENODEV;
8782 }
8783
8784 return rc;
8785}
8786
8787/**
8788 * lpfc_sli4_create_rpi_hdr - Allocate an rpi header memory region
8789 * @phba: pointer to lpfc hba data structure.
8790 *
8791 * This routine is invoked to allocate a single 4KB memory region to
8792 * support rpis and stores them in the phba. This single region
8793 * provides support for up to 64 rpis. The region is used globally
8794 * by the device.
8795 *
8796 * Returns:
8797 * A valid rpi hdr on success.
8798 * A NULL pointer on any failure.
8799 **/
8800struct lpfc_rpi_hdr *
8801lpfc_sli4_create_rpi_hdr(struct lpfc_hba *phba)
8802{
8803 uint16_t rpi_limit, curr_rpi_range;
8804 struct lpfc_dmabuf *dmabuf;
8805 struct lpfc_rpi_hdr *rpi_hdr;
8806
James Smart6d368e52011-05-24 11:44:12 -04008807 /*
8808 * If the SLI4 port supports extents, posting the rpi header isn't
8809 * required. Set the expected maximum count and let the actual value
8810 * get set when extents are fully allocated.
8811 */
8812 if (!phba->sli4_hba.rpi_hdrs_in_use)
8813 return NULL;
8814 if (phba->sli4_hba.extents_in_use)
8815 return NULL;
8816
8817 /* The limit on the logical index is just the max_rpi count. */
James Smart845d9e82017-05-15 15:20:38 -07008818 rpi_limit = phba->sli4_hba.max_cfg_param.max_rpi;
James Smartda0436e2009-05-22 14:51:39 -04008819
8820 spin_lock_irq(&phba->hbalock);
James Smart6d368e52011-05-24 11:44:12 -04008821 /*
8822 * Establish the starting RPI in this header block. The starting
8823 * rpi is normalized to a zero base because the physical rpi is
8824 * port based.
8825 */
James Smart97f2ecf2012-03-01 22:35:23 -05008826 curr_rpi_range = phba->sli4_hba.next_rpi;
James Smartda0436e2009-05-22 14:51:39 -04008827 spin_unlock_irq(&phba->hbalock);
8828
James Smart845d9e82017-05-15 15:20:38 -07008829 /* Reached full RPI range */
8830 if (curr_rpi_range == rpi_limit)
James Smart6d368e52011-05-24 11:44:12 -04008831 return NULL;
James Smart845d9e82017-05-15 15:20:38 -07008832
James Smartda0436e2009-05-22 14:51:39 -04008833 /*
8834 * First allocate the protocol header region for the port. The
8835 * port expects a 4KB DMA-mapped memory region that is 4K aligned.
8836 */
8837 dmabuf = kzalloc(sizeof(struct lpfc_dmabuf), GFP_KERNEL);
8838 if (!dmabuf)
8839 return NULL;
8840
Luis Chamberlain750afb02019-01-04 09:23:09 +01008841 dmabuf->virt = dma_alloc_coherent(&phba->pcidev->dev,
8842 LPFC_HDR_TEMPLATE_SIZE,
8843 &dmabuf->phys, GFP_KERNEL);
James Smartda0436e2009-05-22 14:51:39 -04008844 if (!dmabuf->virt) {
8845 rpi_hdr = NULL;
8846 goto err_free_dmabuf;
8847 }
8848
James Smartda0436e2009-05-22 14:51:39 -04008849 if (!IS_ALIGNED(dmabuf->phys, LPFC_HDR_TEMPLATE_SIZE)) {
8850 rpi_hdr = NULL;
8851 goto err_free_coherent;
8852 }
8853
8854 /* Save the rpi header data for cleanup later. */
8855 rpi_hdr = kzalloc(sizeof(struct lpfc_rpi_hdr), GFP_KERNEL);
8856 if (!rpi_hdr)
8857 goto err_free_coherent;
8858
8859 rpi_hdr->dmabuf = dmabuf;
8860 rpi_hdr->len = LPFC_HDR_TEMPLATE_SIZE;
8861 rpi_hdr->page_count = 1;
8862 spin_lock_irq(&phba->hbalock);
James Smart6d368e52011-05-24 11:44:12 -04008863
8864 /* The rpi_hdr stores the logical index only. */
8865 rpi_hdr->start_rpi = curr_rpi_range;
James Smart845d9e82017-05-15 15:20:38 -07008866 rpi_hdr->next_rpi = phba->sli4_hba.next_rpi + LPFC_RPI_HDR_COUNT;
James Smartda0436e2009-05-22 14:51:39 -04008867 list_add_tail(&rpi_hdr->list, &phba->sli4_hba.lpfc_rpi_hdr_list);
8868
James Smartda0436e2009-05-22 14:51:39 -04008869 spin_unlock_irq(&phba->hbalock);
8870 return rpi_hdr;
8871
8872 err_free_coherent:
8873 dma_free_coherent(&phba->pcidev->dev, LPFC_HDR_TEMPLATE_SIZE,
8874 dmabuf->virt, dmabuf->phys);
8875 err_free_dmabuf:
8876 kfree(dmabuf);
8877 return NULL;
8878}
8879
8880/**
8881 * lpfc_sli4_remove_rpi_hdrs - Remove all rpi header memory regions
8882 * @phba: pointer to lpfc hba data structure.
8883 *
8884 * This routine is invoked to remove all memory resources allocated
James Smart6d368e52011-05-24 11:44:12 -04008885 * to support rpis for SLI4 ports not supporting extents. This routine
8886 * presumes the caller has released all rpis consumed by fabric or port
8887 * logins and is prepared to have the header pages removed.
James Smartda0436e2009-05-22 14:51:39 -04008888 **/
8889void
8890lpfc_sli4_remove_rpi_hdrs(struct lpfc_hba *phba)
8891{
8892 struct lpfc_rpi_hdr *rpi_hdr, *next_rpi_hdr;
8893
James Smart6d368e52011-05-24 11:44:12 -04008894 if (!phba->sli4_hba.rpi_hdrs_in_use)
8895 goto exit;
8896
James Smartda0436e2009-05-22 14:51:39 -04008897 list_for_each_entry_safe(rpi_hdr, next_rpi_hdr,
8898 &phba->sli4_hba.lpfc_rpi_hdr_list, list) {
8899 list_del(&rpi_hdr->list);
8900 dma_free_coherent(&phba->pcidev->dev, rpi_hdr->len,
8901 rpi_hdr->dmabuf->virt, rpi_hdr->dmabuf->phys);
8902 kfree(rpi_hdr->dmabuf);
8903 kfree(rpi_hdr);
8904 }
James Smart6d368e52011-05-24 11:44:12 -04008905 exit:
8906 /* There are no rpis available to the port now. */
8907 phba->sli4_hba.next_rpi = 0;
James Smartda0436e2009-05-22 14:51:39 -04008908}
8909
8910/**
James Smart3772a992009-05-22 14:50:54 -04008911 * lpfc_hba_alloc - Allocate driver hba data structure for a device.
8912 * @pdev: pointer to pci device data structure.
8913 *
8914 * This routine is invoked to allocate the driver hba data structure for an
8915 * HBA device. If the allocation is successful, the phba reference to the
8916 * PCI device data structure is set.
8917 *
8918 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -02008919 * pointer to @phba - successful
James Smart3772a992009-05-22 14:50:54 -04008920 * NULL - error
8921 **/
8922static struct lpfc_hba *
8923lpfc_hba_alloc(struct pci_dev *pdev)
8924{
8925 struct lpfc_hba *phba;
8926
8927 /* Allocate memory for HBA structure */
8928 phba = kzalloc(sizeof(struct lpfc_hba), GFP_KERNEL);
8929 if (!phba) {
Jiri Slabye34ccdf2009-07-13 23:25:54 +02008930 dev_err(&pdev->dev, "failed to allocate hba struct\n");
James Smart3772a992009-05-22 14:50:54 -04008931 return NULL;
8932 }
8933
8934 /* Set reference to PCI device in HBA structure */
8935 phba->pcidev = pdev;
8936
8937 /* Assign an unused board number */
8938 phba->brd_no = lpfc_get_instance();
8939 if (phba->brd_no < 0) {
8940 kfree(phba);
8941 return NULL;
8942 }
James Smart65791f12016-07-06 12:35:56 -07008943 phba->eratt_poll_interval = LPFC_ERATT_POLL_INTERVAL;
James Smart3772a992009-05-22 14:50:54 -04008944
James Smart4fede782010-01-26 23:08:55 -05008945 spin_lock_init(&phba->ct_ev_lock);
James Smartf1c3b0f2009-07-19 10:01:32 -04008946 INIT_LIST_HEAD(&phba->ct_ev_waiters);
8947
James Smart3772a992009-05-22 14:50:54 -04008948 return phba;
8949}
8950
8951/**
8952 * lpfc_hba_free - Free driver hba data structure with a device.
8953 * @phba: pointer to lpfc hba data structure.
8954 *
8955 * This routine is invoked to free the driver hba data structure with an
8956 * HBA device.
8957 **/
8958static void
8959lpfc_hba_free(struct lpfc_hba *phba)
8960{
James Smart5e5b5112019-01-28 11:14:22 -08008961 if (phba->sli_rev == LPFC_SLI_REV4)
8962 kfree(phba->sli4_hba.hdwq);
8963
James Smart3772a992009-05-22 14:50:54 -04008964 /* Release the driver assigned board number */
8965 idr_remove(&lpfc_hba_index, phba->brd_no);
8966
James Smart895427b2017-02-12 13:52:30 -08008967 /* Free memory allocated with sli3 rings */
8968 kfree(phba->sli.sli3_ring);
8969 phba->sli.sli3_ring = NULL;
James Smart2a76a282012-08-03 12:35:54 -04008970
James Smart3772a992009-05-22 14:50:54 -04008971 kfree(phba);
8972 return;
8973}
8974
8975/**
8976 * lpfc_create_shost - Create hba physical port with associated scsi host.
8977 * @phba: pointer to lpfc hba data structure.
8978 *
8979 * This routine is invoked to create HBA physical port and associate a SCSI
8980 * host with it.
8981 *
8982 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -02008983 * 0 - successful
James Smart3772a992009-05-22 14:50:54 -04008984 * other values - error
8985 **/
8986static int
8987lpfc_create_shost(struct lpfc_hba *phba)
8988{
8989 struct lpfc_vport *vport;
8990 struct Scsi_Host *shost;
8991
8992 /* Initialize HBA FC structure */
dea31012005-04-17 16:05:31 -05008993 phba->fc_edtov = FF_DEF_EDTOV;
8994 phba->fc_ratov = FF_DEF_RATOV;
8995 phba->fc_altov = FF_DEF_ALTOV;
8996 phba->fc_arbtov = FF_DEF_ARBTOV;
8997
James Smartd7c47992010-06-08 18:31:54 -04008998 atomic_set(&phba->sdev_cnt, 0);
James Smart3de2a652007-08-02 11:09:59 -04008999 vport = lpfc_create_port(phba, phba->brd_no, &phba->pcidev->dev);
James Smart2e0fef82007-06-17 19:56:36 -05009000 if (!vport)
James Smart3772a992009-05-22 14:50:54 -04009001 return -ENODEV;
James Smart2e0fef82007-06-17 19:56:36 -05009002
9003 shost = lpfc_shost_from_vport(vport);
James Smart2e0fef82007-06-17 19:56:36 -05009004 phba->pport = vport;
James Smart2ea259e2017-02-12 13:52:27 -08009005
James Smartf358dd02017-02-12 13:52:34 -08009006 if (phba->nvmet_support) {
9007 /* Only 1 vport (pport) will support NVME target */
James Smartea85a202019-10-18 14:18:25 -07009008 phba->targetport = NULL;
9009 phba->cfg_enable_fc4_type = LPFC_ENABLE_NVME;
9010 lpfc_printf_log(phba, KERN_INFO, LOG_INIT | LOG_NVME_DISC,
9011 "6076 NVME Target Found\n");
James Smartf358dd02017-02-12 13:52:34 -08009012 }
9013
James Smart858c9f62007-06-17 19:56:39 -05009014 lpfc_debugfs_initialize(vport);
James Smart3772a992009-05-22 14:50:54 -04009015 /* Put reference to SCSI host to driver's device private data */
9016 pci_set_drvdata(phba->pcidev, shost);
James Smart2e0fef82007-06-17 19:56:36 -05009017
James Smart4258e982015-12-16 18:11:58 -05009018 /*
9019 * At this point we are fully registered with PSA. In addition,
9020 * any initial discovery should be completed.
9021 */
9022 vport->load_flag |= FC_ALLOW_FDMI;
James Smart8663cbb2016-03-31 14:12:33 -07009023 if (phba->cfg_enable_SmartSAN ||
9024 (phba->cfg_fdmi_on == LPFC_FDMI_SUPPORT)) {
James Smart4258e982015-12-16 18:11:58 -05009025
9026 /* Setup appropriate attribute masks */
9027 vport->fdmi_hba_mask = LPFC_FDMI2_HBA_ATTR;
James Smart8663cbb2016-03-31 14:12:33 -07009028 if (phba->cfg_enable_SmartSAN)
James Smart4258e982015-12-16 18:11:58 -05009029 vport->fdmi_port_mask = LPFC_FDMI2_SMART_ATTR;
9030 else
9031 vport->fdmi_port_mask = LPFC_FDMI2_PORT_ATTR;
9032 }
James Smart3772a992009-05-22 14:50:54 -04009033 return 0;
9034}
dea31012005-04-17 16:05:31 -05009035
James Smart3772a992009-05-22 14:50:54 -04009036/**
9037 * lpfc_destroy_shost - Destroy hba physical port with associated scsi host.
9038 * @phba: pointer to lpfc hba data structure.
9039 *
9040 * This routine is invoked to destroy HBA physical port and the associated
9041 * SCSI host.
9042 **/
9043static void
9044lpfc_destroy_shost(struct lpfc_hba *phba)
9045{
9046 struct lpfc_vport *vport = phba->pport;
James Smart93996272008-08-24 21:50:30 -04009047
James Smart3772a992009-05-22 14:50:54 -04009048 /* Destroy physical port that associated with the SCSI host */
9049 destroy_port(vport);
9050
9051 return;
9052}
9053
9054/**
9055 * lpfc_setup_bg - Setup Block guard structures and debug areas.
9056 * @phba: pointer to lpfc hba data structure.
9057 * @shost: the shost to be used to detect Block guard settings.
9058 *
9059 * This routine sets up the local Block guard protocol settings for @shost.
9060 * This routine also allocates memory for debugging bg buffers.
9061 **/
9062static void
9063lpfc_setup_bg(struct lpfc_hba *phba, struct Scsi_Host *shost)
9064{
James Smartbbeb79b2012-06-12 13:54:27 -04009065 uint32_t old_mask;
9066 uint32_t old_guard;
9067
James Smartb3b98b72016-10-13 15:06:06 -07009068 if (phba->cfg_prot_mask && phba->cfg_prot_guard) {
James Smart3772a992009-05-22 14:50:54 -04009069 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
9070 "1478 Registering BlockGuard with the "
9071 "SCSI layer\n");
James Smartbbeb79b2012-06-12 13:54:27 -04009072
James Smartb3b98b72016-10-13 15:06:06 -07009073 old_mask = phba->cfg_prot_mask;
9074 old_guard = phba->cfg_prot_guard;
James Smartbbeb79b2012-06-12 13:54:27 -04009075
9076 /* Only allow supported values */
James Smartb3b98b72016-10-13 15:06:06 -07009077 phba->cfg_prot_mask &= (SHOST_DIF_TYPE1_PROTECTION |
James Smartbbeb79b2012-06-12 13:54:27 -04009078 SHOST_DIX_TYPE0_PROTECTION |
9079 SHOST_DIX_TYPE1_PROTECTION);
James Smartb3b98b72016-10-13 15:06:06 -07009080 phba->cfg_prot_guard &= (SHOST_DIX_GUARD_IP |
9081 SHOST_DIX_GUARD_CRC);
James Smartbbeb79b2012-06-12 13:54:27 -04009082
9083 /* DIF Type 1 protection for profiles AST1/C1 is end to end */
James Smartb3b98b72016-10-13 15:06:06 -07009084 if (phba->cfg_prot_mask == SHOST_DIX_TYPE1_PROTECTION)
9085 phba->cfg_prot_mask |= SHOST_DIF_TYPE1_PROTECTION;
James Smartbbeb79b2012-06-12 13:54:27 -04009086
James Smartb3b98b72016-10-13 15:06:06 -07009087 if (phba->cfg_prot_mask && phba->cfg_prot_guard) {
9088 if ((old_mask != phba->cfg_prot_mask) ||
9089 (old_guard != phba->cfg_prot_guard))
Dick Kennedy372c1872020-06-30 14:50:00 -07009090 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartbbeb79b2012-06-12 13:54:27 -04009091 "1475 Registering BlockGuard with the "
9092 "SCSI layer: mask %d guard %d\n",
James Smartb3b98b72016-10-13 15:06:06 -07009093 phba->cfg_prot_mask,
9094 phba->cfg_prot_guard);
James Smartbbeb79b2012-06-12 13:54:27 -04009095
James Smartb3b98b72016-10-13 15:06:06 -07009096 scsi_host_set_prot(shost, phba->cfg_prot_mask);
9097 scsi_host_set_guard(shost, phba->cfg_prot_guard);
James Smartbbeb79b2012-06-12 13:54:27 -04009098 } else
Dick Kennedy372c1872020-06-30 14:50:00 -07009099 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartbbeb79b2012-06-12 13:54:27 -04009100 "1479 Not Registering BlockGuard with the SCSI "
9101 "layer, Bad protection parameters: %d %d\n",
9102 old_mask, old_guard);
James Smart98c9ea52007-10-27 13:37:33 -04009103 }
James Smart3772a992009-05-22 14:50:54 -04009104}
9105
9106/**
9107 * lpfc_post_init_setup - Perform necessary device post initialization setup.
9108 * @phba: pointer to lpfc hba data structure.
9109 *
9110 * This routine is invoked to perform all the necessary post initialization
9111 * setup for the device.
9112 **/
9113static void
9114lpfc_post_init_setup(struct lpfc_hba *phba)
9115{
9116 struct Scsi_Host *shost;
9117 struct lpfc_adapter_event_header adapter_event;
9118
9119 /* Get the default values for Model Name and Description */
9120 lpfc_get_hba_model_desc(phba, phba->ModelName, phba->ModelDesc);
9121
9122 /*
9123 * hba setup may have changed the hba_queue_depth so we need to
9124 * adjust the value of can_queue.
9125 */
9126 shost = pci_get_drvdata(phba->pcidev);
9127 shost->can_queue = phba->cfg_hba_queue_depth - 10;
James Smart858c9f62007-06-17 19:56:39 -05009128
9129 lpfc_host_attrib_init(shost);
9130
James Smart2e0fef82007-06-17 19:56:36 -05009131 if (phba->cfg_poll & DISABLE_FCP_RING_INT) {
9132 spin_lock_irq(shost->host_lock);
9133 lpfc_poll_start_timer(phba);
9134 spin_unlock_irq(shost->host_lock);
9135 }
James Smart8f6d98d2006-08-01 07:34:00 -04009136
James Smart93996272008-08-24 21:50:30 -04009137 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
9138 "0428 Perform SCSI scan\n");
James Smartea2151b2008-09-07 11:52:10 -04009139 /* Send board arrival event to upper layer */
9140 adapter_event.event_type = FC_REG_ADAPTER_EVENT;
9141 adapter_event.subcategory = LPFC_EVENT_ARRIVAL;
9142 fc_host_post_vendor_event(shost, fc_get_event_number(),
James Smart3772a992009-05-22 14:50:54 -04009143 sizeof(adapter_event),
9144 (char *) &adapter_event,
9145 LPFC_NL_VENDOR_ID);
9146 return;
9147}
9148
9149/**
9150 * lpfc_sli_pci_mem_setup - Setup SLI3 HBA PCI memory space.
9151 * @phba: pointer to lpfc hba data structure.
9152 *
9153 * This routine is invoked to set up the PCI device memory space for device
9154 * with SLI-3 interface spec.
9155 *
9156 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -02009157 * 0 - successful
James Smart3772a992009-05-22 14:50:54 -04009158 * other values - error
9159 **/
9160static int
9161lpfc_sli_pci_mem_setup(struct lpfc_hba *phba)
9162{
Christoph Hellwigf30e1bf2018-10-18 15:10:21 +02009163 struct pci_dev *pdev = phba->pcidev;
James Smart3772a992009-05-22 14:50:54 -04009164 unsigned long bar0map_len, bar2map_len;
9165 int i, hbq_count;
9166 void *ptr;
Hannes Reinecke56de8352019-02-18 08:34:19 +01009167 int error;
James Smart3772a992009-05-22 14:50:54 -04009168
Christoph Hellwigf30e1bf2018-10-18 15:10:21 +02009169 if (!pdev)
Hannes Reinecke56de8352019-02-18 08:34:19 +01009170 return -ENODEV;
James Smart3772a992009-05-22 14:50:54 -04009171
9172 /* Set the device DMA mask size */
Hannes Reinecke56de8352019-02-18 08:34:19 +01009173 error = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(64));
9174 if (error)
9175 error = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(32));
9176 if (error)
Christoph Hellwigf30e1bf2018-10-18 15:10:21 +02009177 return error;
Hannes Reinecke56de8352019-02-18 08:34:19 +01009178 error = -ENODEV;
James Smart3772a992009-05-22 14:50:54 -04009179
9180 /* Get the bus address of Bar0 and Bar2 and the number of bytes
9181 * required by each mapping.
9182 */
9183 phba->pci_bar0_map = pci_resource_start(pdev, 0);
9184 bar0map_len = pci_resource_len(pdev, 0);
9185
9186 phba->pci_bar2_map = pci_resource_start(pdev, 2);
9187 bar2map_len = pci_resource_len(pdev, 2);
9188
9189 /* Map HBA SLIM to a kernel virtual address. */
9190 phba->slim_memmap_p = ioremap(phba->pci_bar0_map, bar0map_len);
9191 if (!phba->slim_memmap_p) {
9192 dev_printk(KERN_ERR, &pdev->dev,
9193 "ioremap failed for SLIM memory.\n");
9194 goto out;
9195 }
9196
9197 /* Map HBA Control Registers to a kernel virtual address. */
9198 phba->ctrl_regs_memmap_p = ioremap(phba->pci_bar2_map, bar2map_len);
9199 if (!phba->ctrl_regs_memmap_p) {
9200 dev_printk(KERN_ERR, &pdev->dev,
9201 "ioremap failed for HBA control registers.\n");
9202 goto out_iounmap_slim;
9203 }
9204
9205 /* Allocate memory for SLI-2 structures */
Luis Chamberlain750afb02019-01-04 09:23:09 +01009206 phba->slim2p.virt = dma_alloc_coherent(&pdev->dev, SLI2_SLIM_SIZE,
9207 &phba->slim2p.phys, GFP_KERNEL);
James Smart3772a992009-05-22 14:50:54 -04009208 if (!phba->slim2p.virt)
9209 goto out_iounmap;
9210
James Smart3772a992009-05-22 14:50:54 -04009211 phba->mbox = phba->slim2p.virt + offsetof(struct lpfc_sli2_slim, mbx);
James Smart7a470272010-03-15 11:25:20 -04009212 phba->mbox_ext = (phba->slim2p.virt +
9213 offsetof(struct lpfc_sli2_slim, mbx_ext_words));
James Smart3772a992009-05-22 14:50:54 -04009214 phba->pcb = (phba->slim2p.virt + offsetof(struct lpfc_sli2_slim, pcb));
9215 phba->IOCBs = (phba->slim2p.virt +
9216 offsetof(struct lpfc_sli2_slim, IOCBs));
9217
9218 phba->hbqslimp.virt = dma_alloc_coherent(&pdev->dev,
9219 lpfc_sli_hbq_size(),
9220 &phba->hbqslimp.phys,
9221 GFP_KERNEL);
9222 if (!phba->hbqslimp.virt)
9223 goto out_free_slim;
9224
9225 hbq_count = lpfc_sli_hbq_count();
9226 ptr = phba->hbqslimp.virt;
9227 for (i = 0; i < hbq_count; ++i) {
9228 phba->hbqs[i].hbq_virt = ptr;
9229 INIT_LIST_HEAD(&phba->hbqs[i].hbq_buffer_list);
9230 ptr += (lpfc_hbq_defs[i]->entry_count *
9231 sizeof(struct lpfc_hbq_entry));
9232 }
9233 phba->hbqs[LPFC_ELS_HBQ].hbq_alloc_buffer = lpfc_els_hbq_alloc;
9234 phba->hbqs[LPFC_ELS_HBQ].hbq_free_buffer = lpfc_els_hbq_free;
9235
9236 memset(phba->hbqslimp.virt, 0, lpfc_sli_hbq_size());
9237
James Smart3772a992009-05-22 14:50:54 -04009238 phba->MBslimaddr = phba->slim_memmap_p;
9239 phba->HAregaddr = phba->ctrl_regs_memmap_p + HA_REG_OFFSET;
9240 phba->CAregaddr = phba->ctrl_regs_memmap_p + CA_REG_OFFSET;
9241 phba->HSregaddr = phba->ctrl_regs_memmap_p + HS_REG_OFFSET;
9242 phba->HCregaddr = phba->ctrl_regs_memmap_p + HC_REG_OFFSET;
James Smartea2151b2008-09-07 11:52:10 -04009243
dea31012005-04-17 16:05:31 -05009244 return 0;
9245
dea31012005-04-17 16:05:31 -05009246out_free_slim:
James Smart34b02dc2008-08-24 21:49:55 -04009247 dma_free_coherent(&pdev->dev, SLI2_SLIM_SIZE,
9248 phba->slim2p.virt, phba->slim2p.phys);
dea31012005-04-17 16:05:31 -05009249out_iounmap:
9250 iounmap(phba->ctrl_regs_memmap_p);
Jamie Wellnitz901a9202006-02-28 19:25:19 -05009251out_iounmap_slim:
dea31012005-04-17 16:05:31 -05009252 iounmap(phba->slim_memmap_p);
dea31012005-04-17 16:05:31 -05009253out:
9254 return error;
9255}
9256
James Smarte59058c2008-08-24 21:49:00 -04009257/**
James Smart3772a992009-05-22 14:50:54 -04009258 * lpfc_sli_pci_mem_unset - Unset SLI3 HBA PCI memory space.
9259 * @phba: pointer to lpfc hba data structure.
9260 *
9261 * This routine is invoked to unset the PCI device memory space for device
9262 * with SLI-3 interface spec.
9263 **/
9264static void
9265lpfc_sli_pci_mem_unset(struct lpfc_hba *phba)
9266{
9267 struct pci_dev *pdev;
9268
9269 /* Obtain PCI device reference */
9270 if (!phba->pcidev)
9271 return;
9272 else
9273 pdev = phba->pcidev;
9274
9275 /* Free coherent DMA memory allocated */
9276 dma_free_coherent(&pdev->dev, lpfc_sli_hbq_size(),
9277 phba->hbqslimp.virt, phba->hbqslimp.phys);
9278 dma_free_coherent(&pdev->dev, SLI2_SLIM_SIZE,
9279 phba->slim2p.virt, phba->slim2p.phys);
9280
9281 /* I/O memory unmap */
9282 iounmap(phba->ctrl_regs_memmap_p);
9283 iounmap(phba->slim_memmap_p);
9284
9285 return;
9286}
9287
9288/**
James Smartda0436e2009-05-22 14:51:39 -04009289 * lpfc_sli4_post_status_check - Wait for SLI4 POST done and check status
9290 * @phba: pointer to lpfc hba data structure.
9291 *
9292 * This routine is invoked to wait for SLI4 device Power On Self Test (POST)
9293 * done and check status.
9294 *
9295 * Return 0 if successful, otherwise -ENODEV.
9296 **/
9297int
9298lpfc_sli4_post_status_check(struct lpfc_hba *phba)
9299{
James Smart2fcee4b2010-12-15 17:57:46 -05009300 struct lpfc_register portsmphr_reg, uerrlo_reg, uerrhi_reg;
9301 struct lpfc_register reg_data;
9302 int i, port_error = 0;
9303 uint32_t if_type;
James Smartda0436e2009-05-22 14:51:39 -04009304
James Smart9940b972011-03-11 16:06:12 -05009305 memset(&portsmphr_reg, 0, sizeof(portsmphr_reg));
9306 memset(&reg_data, 0, sizeof(reg_data));
James Smart2fcee4b2010-12-15 17:57:46 -05009307 if (!phba->sli4_hba.PSMPHRregaddr)
James Smartda0436e2009-05-22 14:51:39 -04009308 return -ENODEV;
9309
James Smartda0436e2009-05-22 14:51:39 -04009310 /* Wait up to 30 seconds for the SLI Port POST done and ready */
9311 for (i = 0; i < 3000; i++) {
James Smart9940b972011-03-11 16:06:12 -05009312 if (lpfc_readl(phba->sli4_hba.PSMPHRregaddr,
9313 &portsmphr_reg.word0) ||
9314 (bf_get(lpfc_port_smphr_perr, &portsmphr_reg))) {
James Smart2fcee4b2010-12-15 17:57:46 -05009315 /* Port has a fatal POST error, break out */
James Smartda0436e2009-05-22 14:51:39 -04009316 port_error = -ENODEV;
9317 break;
9318 }
James Smart2fcee4b2010-12-15 17:57:46 -05009319 if (LPFC_POST_STAGE_PORT_READY ==
9320 bf_get(lpfc_port_smphr_port_status, &portsmphr_reg))
James Smartda0436e2009-05-22 14:51:39 -04009321 break;
James Smartda0436e2009-05-22 14:51:39 -04009322 msleep(10);
9323 }
9324
James Smart2fcee4b2010-12-15 17:57:46 -05009325 /*
9326 * If there was a port error during POST, then don't proceed with
9327 * other register reads as the data may not be valid. Just exit.
9328 */
9329 if (port_error) {
Dick Kennedy372c1872020-06-30 14:50:00 -07009330 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart2fcee4b2010-12-15 17:57:46 -05009331 "1408 Port Failed POST - portsmphr=0x%x, "
9332 "perr=x%x, sfi=x%x, nip=x%x, ipc=x%x, scr1=x%x, "
9333 "scr2=x%x, hscratch=x%x, pstatus=x%x\n",
9334 portsmphr_reg.word0,
9335 bf_get(lpfc_port_smphr_perr, &portsmphr_reg),
9336 bf_get(lpfc_port_smphr_sfi, &portsmphr_reg),
9337 bf_get(lpfc_port_smphr_nip, &portsmphr_reg),
9338 bf_get(lpfc_port_smphr_ipc, &portsmphr_reg),
9339 bf_get(lpfc_port_smphr_scr1, &portsmphr_reg),
9340 bf_get(lpfc_port_smphr_scr2, &portsmphr_reg),
9341 bf_get(lpfc_port_smphr_host_scratch, &portsmphr_reg),
9342 bf_get(lpfc_port_smphr_port_status, &portsmphr_reg));
9343 } else {
James Smart28baac72010-02-12 14:42:03 -05009344 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
James Smart2fcee4b2010-12-15 17:57:46 -05009345 "2534 Device Info: SLIFamily=0x%x, "
9346 "SLIRev=0x%x, IFType=0x%x, SLIHint_1=0x%x, "
9347 "SLIHint_2=0x%x, FT=0x%x\n",
James Smart28baac72010-02-12 14:42:03 -05009348 bf_get(lpfc_sli_intf_sli_family,
9349 &phba->sli4_hba.sli_intf),
9350 bf_get(lpfc_sli_intf_slirev,
9351 &phba->sli4_hba.sli_intf),
James Smart085c6472010-11-20 23:11:37 -05009352 bf_get(lpfc_sli_intf_if_type,
James Smart28baac72010-02-12 14:42:03 -05009353 &phba->sli4_hba.sli_intf),
James Smart085c6472010-11-20 23:11:37 -05009354 bf_get(lpfc_sli_intf_sli_hint1,
9355 &phba->sli4_hba.sli_intf),
9356 bf_get(lpfc_sli_intf_sli_hint2,
9357 &phba->sli4_hba.sli_intf),
9358 bf_get(lpfc_sli_intf_func_type,
James Smart28baac72010-02-12 14:42:03 -05009359 &phba->sli4_hba.sli_intf));
James Smart2fcee4b2010-12-15 17:57:46 -05009360 /*
9361 * Check for other Port errors during the initialization
9362 * process. Fail the load if the port did not come up
9363 * correctly.
9364 */
9365 if_type = bf_get(lpfc_sli_intf_if_type,
9366 &phba->sli4_hba.sli_intf);
9367 switch (if_type) {
9368 case LPFC_SLI_INTF_IF_TYPE_0:
9369 phba->sli4_hba.ue_mask_lo =
9370 readl(phba->sli4_hba.u.if_type0.UEMASKLOregaddr);
9371 phba->sli4_hba.ue_mask_hi =
9372 readl(phba->sli4_hba.u.if_type0.UEMASKHIregaddr);
9373 uerrlo_reg.word0 =
9374 readl(phba->sli4_hba.u.if_type0.UERRLOregaddr);
9375 uerrhi_reg.word0 =
9376 readl(phba->sli4_hba.u.if_type0.UERRHIregaddr);
9377 if ((~phba->sli4_hba.ue_mask_lo & uerrlo_reg.word0) ||
9378 (~phba->sli4_hba.ue_mask_hi & uerrhi_reg.word0)) {
Dick Kennedy372c1872020-06-30 14:50:00 -07009379 lpfc_printf_log(phba, KERN_ERR,
9380 LOG_TRACE_EVENT,
James Smart2fcee4b2010-12-15 17:57:46 -05009381 "1422 Unrecoverable Error "
9382 "Detected during POST "
9383 "uerr_lo_reg=0x%x, "
9384 "uerr_hi_reg=0x%x, "
9385 "ue_mask_lo_reg=0x%x, "
9386 "ue_mask_hi_reg=0x%x\n",
9387 uerrlo_reg.word0,
9388 uerrhi_reg.word0,
9389 phba->sli4_hba.ue_mask_lo,
9390 phba->sli4_hba.ue_mask_hi);
9391 port_error = -ENODEV;
9392 }
9393 break;
9394 case LPFC_SLI_INTF_IF_TYPE_2:
James Smart27d6ac02018-02-22 08:18:42 -08009395 case LPFC_SLI_INTF_IF_TYPE_6:
James Smart2fcee4b2010-12-15 17:57:46 -05009396 /* Final checks. The port status should be clean. */
James Smart9940b972011-03-11 16:06:12 -05009397 if (lpfc_readl(phba->sli4_hba.u.if_type2.STATUSregaddr,
9398 &reg_data.word0) ||
James Smart05580562011-05-24 11:40:48 -04009399 (bf_get(lpfc_sliport_status_err, &reg_data) &&
9400 !bf_get(lpfc_sliport_status_rn, &reg_data))) {
James Smart2fcee4b2010-12-15 17:57:46 -05009401 phba->work_status[0] =
9402 readl(phba->sli4_hba.u.if_type2.
9403 ERR1regaddr);
9404 phba->work_status[1] =
9405 readl(phba->sli4_hba.u.if_type2.
9406 ERR2regaddr);
Dick Kennedy372c1872020-06-30 14:50:00 -07009407 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart8fcb8ac2012-03-01 22:35:58 -05009408 "2888 Unrecoverable port error "
9409 "following POST: port status reg "
9410 "0x%x, port_smphr reg 0x%x, "
James Smart2fcee4b2010-12-15 17:57:46 -05009411 "error 1=0x%x, error 2=0x%x\n",
9412 reg_data.word0,
9413 portsmphr_reg.word0,
9414 phba->work_status[0],
9415 phba->work_status[1]);
9416 port_error = -ENODEV;
James Smarta5b141a2021-09-27 11:35:18 -07009417 break;
James Smart2fcee4b2010-12-15 17:57:46 -05009418 }
James Smarta5b141a2021-09-27 11:35:18 -07009419
9420 if (lpfc_pldv_detect &&
9421 bf_get(lpfc_sli_intf_sli_family,
9422 &phba->sli4_hba.sli_intf) ==
9423 LPFC_SLI_INTF_FAMILY_G6)
9424 pci_write_config_byte(phba->pcidev,
9425 LPFC_SLI_INTF, CFG_PLD);
James Smart2fcee4b2010-12-15 17:57:46 -05009426 break;
9427 case LPFC_SLI_INTF_IF_TYPE_1:
9428 default:
9429 break;
9430 }
James Smart28baac72010-02-12 14:42:03 -05009431 }
James Smartda0436e2009-05-22 14:51:39 -04009432 return port_error;
9433}
9434
9435/**
9436 * lpfc_sli4_bar0_register_memmap - Set up SLI4 BAR0 register memory map.
9437 * @phba: pointer to lpfc hba data structure.
James Smart2fcee4b2010-12-15 17:57:46 -05009438 * @if_type: The SLI4 interface type getting configured.
James Smartda0436e2009-05-22 14:51:39 -04009439 *
9440 * This routine is invoked to set up SLI4 BAR0 PCI config space register
9441 * memory map.
9442 **/
9443static void
James Smart2fcee4b2010-12-15 17:57:46 -05009444lpfc_sli4_bar0_register_memmap(struct lpfc_hba *phba, uint32_t if_type)
James Smartda0436e2009-05-22 14:51:39 -04009445{
James Smart2fcee4b2010-12-15 17:57:46 -05009446 switch (if_type) {
9447 case LPFC_SLI_INTF_IF_TYPE_0:
9448 phba->sli4_hba.u.if_type0.UERRLOregaddr =
9449 phba->sli4_hba.conf_regs_memmap_p + LPFC_UERR_STATUS_LO;
9450 phba->sli4_hba.u.if_type0.UERRHIregaddr =
9451 phba->sli4_hba.conf_regs_memmap_p + LPFC_UERR_STATUS_HI;
9452 phba->sli4_hba.u.if_type0.UEMASKLOregaddr =
9453 phba->sli4_hba.conf_regs_memmap_p + LPFC_UE_MASK_LO;
9454 phba->sli4_hba.u.if_type0.UEMASKHIregaddr =
9455 phba->sli4_hba.conf_regs_memmap_p + LPFC_UE_MASK_HI;
9456 phba->sli4_hba.SLIINTFregaddr =
9457 phba->sli4_hba.conf_regs_memmap_p + LPFC_SLI_INTF;
9458 break;
9459 case LPFC_SLI_INTF_IF_TYPE_2:
James Smart0cf07f842017-06-01 21:07:10 -07009460 phba->sli4_hba.u.if_type2.EQDregaddr =
9461 phba->sli4_hba.conf_regs_memmap_p +
9462 LPFC_CTL_PORT_EQ_DELAY_OFFSET;
James Smart2fcee4b2010-12-15 17:57:46 -05009463 phba->sli4_hba.u.if_type2.ERR1regaddr =
James Smart88a2cfb2011-07-22 18:36:33 -04009464 phba->sli4_hba.conf_regs_memmap_p +
9465 LPFC_CTL_PORT_ER1_OFFSET;
James Smart2fcee4b2010-12-15 17:57:46 -05009466 phba->sli4_hba.u.if_type2.ERR2regaddr =
James Smart88a2cfb2011-07-22 18:36:33 -04009467 phba->sli4_hba.conf_regs_memmap_p +
9468 LPFC_CTL_PORT_ER2_OFFSET;
James Smart2fcee4b2010-12-15 17:57:46 -05009469 phba->sli4_hba.u.if_type2.CTRLregaddr =
James Smart88a2cfb2011-07-22 18:36:33 -04009470 phba->sli4_hba.conf_regs_memmap_p +
9471 LPFC_CTL_PORT_CTL_OFFSET;
James Smart2fcee4b2010-12-15 17:57:46 -05009472 phba->sli4_hba.u.if_type2.STATUSregaddr =
James Smart88a2cfb2011-07-22 18:36:33 -04009473 phba->sli4_hba.conf_regs_memmap_p +
9474 LPFC_CTL_PORT_STA_OFFSET;
James Smart2fcee4b2010-12-15 17:57:46 -05009475 phba->sli4_hba.SLIINTFregaddr =
9476 phba->sli4_hba.conf_regs_memmap_p + LPFC_SLI_INTF;
9477 phba->sli4_hba.PSMPHRregaddr =
James Smart88a2cfb2011-07-22 18:36:33 -04009478 phba->sli4_hba.conf_regs_memmap_p +
9479 LPFC_CTL_PORT_SEM_OFFSET;
James Smart2fcee4b2010-12-15 17:57:46 -05009480 phba->sli4_hba.RQDBregaddr =
James Smart962bc512013-01-03 15:44:00 -05009481 phba->sli4_hba.conf_regs_memmap_p +
9482 LPFC_ULP0_RQ_DOORBELL;
James Smart2fcee4b2010-12-15 17:57:46 -05009483 phba->sli4_hba.WQDBregaddr =
James Smart962bc512013-01-03 15:44:00 -05009484 phba->sli4_hba.conf_regs_memmap_p +
9485 LPFC_ULP0_WQ_DOORBELL;
James Smart9dd35422018-02-22 08:18:41 -08009486 phba->sli4_hba.CQDBregaddr =
James Smart2fcee4b2010-12-15 17:57:46 -05009487 phba->sli4_hba.conf_regs_memmap_p + LPFC_EQCQ_DOORBELL;
James Smart9dd35422018-02-22 08:18:41 -08009488 phba->sli4_hba.EQDBregaddr = phba->sli4_hba.CQDBregaddr;
James Smart2fcee4b2010-12-15 17:57:46 -05009489 phba->sli4_hba.MQDBregaddr =
9490 phba->sli4_hba.conf_regs_memmap_p + LPFC_MQ_DOORBELL;
9491 phba->sli4_hba.BMBXregaddr =
9492 phba->sli4_hba.conf_regs_memmap_p + LPFC_BMBX;
9493 break;
James Smart27d6ac02018-02-22 08:18:42 -08009494 case LPFC_SLI_INTF_IF_TYPE_6:
9495 phba->sli4_hba.u.if_type2.EQDregaddr =
9496 phba->sli4_hba.conf_regs_memmap_p +
9497 LPFC_CTL_PORT_EQ_DELAY_OFFSET;
9498 phba->sli4_hba.u.if_type2.ERR1regaddr =
9499 phba->sli4_hba.conf_regs_memmap_p +
9500 LPFC_CTL_PORT_ER1_OFFSET;
9501 phba->sli4_hba.u.if_type2.ERR2regaddr =
9502 phba->sli4_hba.conf_regs_memmap_p +
9503 LPFC_CTL_PORT_ER2_OFFSET;
9504 phba->sli4_hba.u.if_type2.CTRLregaddr =
9505 phba->sli4_hba.conf_regs_memmap_p +
9506 LPFC_CTL_PORT_CTL_OFFSET;
9507 phba->sli4_hba.u.if_type2.STATUSregaddr =
9508 phba->sli4_hba.conf_regs_memmap_p +
9509 LPFC_CTL_PORT_STA_OFFSET;
9510 phba->sli4_hba.PSMPHRregaddr =
9511 phba->sli4_hba.conf_regs_memmap_p +
9512 LPFC_CTL_PORT_SEM_OFFSET;
9513 phba->sli4_hba.BMBXregaddr =
9514 phba->sli4_hba.conf_regs_memmap_p + LPFC_BMBX;
9515 break;
James Smart2fcee4b2010-12-15 17:57:46 -05009516 case LPFC_SLI_INTF_IF_TYPE_1:
9517 default:
9518 dev_printk(KERN_ERR, &phba->pcidev->dev,
9519 "FATAL - unsupported SLI4 interface type - %d\n",
9520 if_type);
9521 break;
9522 }
James Smartda0436e2009-05-22 14:51:39 -04009523}
9524
9525/**
9526 * lpfc_sli4_bar1_register_memmap - Set up SLI4 BAR1 register memory map.
9527 * @phba: pointer to lpfc hba data structure.
Lee Jonesfe614ac2020-07-23 13:24:22 +01009528 * @if_type: sli if type to operate on.
James Smartda0436e2009-05-22 14:51:39 -04009529 *
James Smart27d6ac02018-02-22 08:18:42 -08009530 * This routine is invoked to set up SLI4 BAR1 register memory map.
James Smartda0436e2009-05-22 14:51:39 -04009531 **/
9532static void
James Smart27d6ac02018-02-22 08:18:42 -08009533lpfc_sli4_bar1_register_memmap(struct lpfc_hba *phba, uint32_t if_type)
James Smartda0436e2009-05-22 14:51:39 -04009534{
James Smart27d6ac02018-02-22 08:18:42 -08009535 switch (if_type) {
9536 case LPFC_SLI_INTF_IF_TYPE_0:
9537 phba->sli4_hba.PSMPHRregaddr =
9538 phba->sli4_hba.ctrl_regs_memmap_p +
9539 LPFC_SLIPORT_IF0_SMPHR;
9540 phba->sli4_hba.ISRregaddr = phba->sli4_hba.ctrl_regs_memmap_p +
9541 LPFC_HST_ISR0;
9542 phba->sli4_hba.IMRregaddr = phba->sli4_hba.ctrl_regs_memmap_p +
9543 LPFC_HST_IMR0;
9544 phba->sli4_hba.ISCRregaddr = phba->sli4_hba.ctrl_regs_memmap_p +
9545 LPFC_HST_ISCR0;
9546 break;
9547 case LPFC_SLI_INTF_IF_TYPE_6:
9548 phba->sli4_hba.RQDBregaddr = phba->sli4_hba.drbl_regs_memmap_p +
9549 LPFC_IF6_RQ_DOORBELL;
9550 phba->sli4_hba.WQDBregaddr = phba->sli4_hba.drbl_regs_memmap_p +
9551 LPFC_IF6_WQ_DOORBELL;
9552 phba->sli4_hba.CQDBregaddr = phba->sli4_hba.drbl_regs_memmap_p +
9553 LPFC_IF6_CQ_DOORBELL;
9554 phba->sli4_hba.EQDBregaddr = phba->sli4_hba.drbl_regs_memmap_p +
9555 LPFC_IF6_EQ_DOORBELL;
9556 phba->sli4_hba.MQDBregaddr = phba->sli4_hba.drbl_regs_memmap_p +
9557 LPFC_IF6_MQ_DOORBELL;
9558 break;
9559 case LPFC_SLI_INTF_IF_TYPE_2:
9560 case LPFC_SLI_INTF_IF_TYPE_1:
9561 default:
9562 dev_err(&phba->pcidev->dev,
9563 "FATAL - unsupported SLI4 interface type - %d\n",
9564 if_type);
9565 break;
9566 }
James Smartda0436e2009-05-22 14:51:39 -04009567}
9568
9569/**
9570 * lpfc_sli4_bar2_register_memmap - Set up SLI4 BAR2 register memory map.
9571 * @phba: pointer to lpfc hba data structure.
9572 * @vf: virtual function number
9573 *
9574 * This routine is invoked to set up SLI4 BAR2 doorbell register memory map
9575 * based on the given viftual function number, @vf.
9576 *
9577 * Return 0 if successful, otherwise -ENODEV.
9578 **/
9579static int
9580lpfc_sli4_bar2_register_memmap(struct lpfc_hba *phba, uint32_t vf)
9581{
9582 if (vf > LPFC_VIR_FUNC_MAX)
9583 return -ENODEV;
9584
9585 phba->sli4_hba.RQDBregaddr = (phba->sli4_hba.drbl_regs_memmap_p +
James Smart962bc512013-01-03 15:44:00 -05009586 vf * LPFC_VFR_PAGE_SIZE +
9587 LPFC_ULP0_RQ_DOORBELL);
James Smartda0436e2009-05-22 14:51:39 -04009588 phba->sli4_hba.WQDBregaddr = (phba->sli4_hba.drbl_regs_memmap_p +
James Smart962bc512013-01-03 15:44:00 -05009589 vf * LPFC_VFR_PAGE_SIZE +
9590 LPFC_ULP0_WQ_DOORBELL);
James Smart9dd35422018-02-22 08:18:41 -08009591 phba->sli4_hba.CQDBregaddr = (phba->sli4_hba.drbl_regs_memmap_p +
9592 vf * LPFC_VFR_PAGE_SIZE +
9593 LPFC_EQCQ_DOORBELL);
9594 phba->sli4_hba.EQDBregaddr = phba->sli4_hba.CQDBregaddr;
James Smartda0436e2009-05-22 14:51:39 -04009595 phba->sli4_hba.MQDBregaddr = (phba->sli4_hba.drbl_regs_memmap_p +
9596 vf * LPFC_VFR_PAGE_SIZE + LPFC_MQ_DOORBELL);
9597 phba->sli4_hba.BMBXregaddr = (phba->sli4_hba.drbl_regs_memmap_p +
9598 vf * LPFC_VFR_PAGE_SIZE + LPFC_BMBX);
9599 return 0;
9600}
9601
9602/**
9603 * lpfc_create_bootstrap_mbox - Create the bootstrap mailbox
9604 * @phba: pointer to lpfc hba data structure.
9605 *
9606 * This routine is invoked to create the bootstrap mailbox
9607 * region consistent with the SLI-4 interface spec. This
9608 * routine allocates all memory necessary to communicate
9609 * mailbox commands to the port and sets up all alignment
9610 * needs. No locks are expected to be held when calling
9611 * this routine.
9612 *
9613 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -02009614 * 0 - successful
James Smartd439d282010-09-29 11:18:45 -04009615 * -ENOMEM - could not allocated memory.
James Smartda0436e2009-05-22 14:51:39 -04009616 **/
9617static int
9618lpfc_create_bootstrap_mbox(struct lpfc_hba *phba)
9619{
9620 uint32_t bmbx_size;
9621 struct lpfc_dmabuf *dmabuf;
9622 struct dma_address *dma_address;
9623 uint32_t pa_addr;
9624 uint64_t phys_addr;
9625
9626 dmabuf = kzalloc(sizeof(struct lpfc_dmabuf), GFP_KERNEL);
9627 if (!dmabuf)
9628 return -ENOMEM;
9629
9630 /*
9631 * The bootstrap mailbox region is comprised of 2 parts
9632 * plus an alignment restriction of 16 bytes.
9633 */
9634 bmbx_size = sizeof(struct lpfc_bmbx_create) + (LPFC_ALIGN_16_BYTE - 1);
Luis Chamberlain750afb02019-01-04 09:23:09 +01009635 dmabuf->virt = dma_alloc_coherent(&phba->pcidev->dev, bmbx_size,
9636 &dmabuf->phys, GFP_KERNEL);
James Smartda0436e2009-05-22 14:51:39 -04009637 if (!dmabuf->virt) {
9638 kfree(dmabuf);
9639 return -ENOMEM;
9640 }
James Smartda0436e2009-05-22 14:51:39 -04009641
9642 /*
9643 * Initialize the bootstrap mailbox pointers now so that the register
9644 * operations are simple later. The mailbox dma address is required
9645 * to be 16-byte aligned. Also align the virtual memory as each
9646 * maibox is copied into the bmbx mailbox region before issuing the
9647 * command to the port.
9648 */
9649 phba->sli4_hba.bmbx.dmabuf = dmabuf;
9650 phba->sli4_hba.bmbx.bmbx_size = bmbx_size;
9651
9652 phba->sli4_hba.bmbx.avirt = PTR_ALIGN(dmabuf->virt,
9653 LPFC_ALIGN_16_BYTE);
9654 phba->sli4_hba.bmbx.aphys = ALIGN(dmabuf->phys,
9655 LPFC_ALIGN_16_BYTE);
9656
9657 /*
9658 * Set the high and low physical addresses now. The SLI4 alignment
9659 * requirement is 16 bytes and the mailbox is posted to the port
9660 * as two 30-bit addresses. The other data is a bit marking whether
9661 * the 30-bit address is the high or low address.
9662 * Upcast bmbx aphys to 64bits so shift instruction compiles
9663 * clean on 32 bit machines.
9664 */
9665 dma_address = &phba->sli4_hba.bmbx.dma_address;
9666 phys_addr = (uint64_t)phba->sli4_hba.bmbx.aphys;
9667 pa_addr = (uint32_t) ((phys_addr >> 34) & 0x3fffffff);
9668 dma_address->addr_hi = (uint32_t) ((pa_addr << 2) |
9669 LPFC_BMBX_BIT1_ADDR_HI);
9670
9671 pa_addr = (uint32_t) ((phba->sli4_hba.bmbx.aphys >> 4) & 0x3fffffff);
9672 dma_address->addr_lo = (uint32_t) ((pa_addr << 2) |
9673 LPFC_BMBX_BIT1_ADDR_LO);
9674 return 0;
9675}
9676
9677/**
9678 * lpfc_destroy_bootstrap_mbox - Destroy all bootstrap mailbox resources
9679 * @phba: pointer to lpfc hba data structure.
9680 *
9681 * This routine is invoked to teardown the bootstrap mailbox
9682 * region and release all host resources. This routine requires
9683 * the caller to ensure all mailbox commands recovered, no
9684 * additional mailbox comands are sent, and interrupts are disabled
9685 * before calling this routine.
9686 *
9687 **/
9688static void
9689lpfc_destroy_bootstrap_mbox(struct lpfc_hba *phba)
9690{
9691 dma_free_coherent(&phba->pcidev->dev,
9692 phba->sli4_hba.bmbx.bmbx_size,
9693 phba->sli4_hba.bmbx.dmabuf->virt,
9694 phba->sli4_hba.bmbx.dmabuf->phys);
9695
9696 kfree(phba->sli4_hba.bmbx.dmabuf);
9697 memset(&phba->sli4_hba.bmbx, 0, sizeof(struct lpfc_bmbx));
9698}
9699
James Smart83c6cb12019-10-18 14:18:30 -07009700static const char * const lpfc_topo_to_str[] = {
9701 "Loop then P2P",
9702 "Loopback",
9703 "P2P Only",
9704 "Unsupported",
9705 "Loop Only",
9706 "Unsupported",
9707 "P2P then Loop",
9708};
9709
Lee Jonesfe614ac2020-07-23 13:24:22 +01009710#define LINK_FLAGS_DEF 0x0
9711#define LINK_FLAGS_P2P 0x1
9712#define LINK_FLAGS_LOOP 0x2
James Smart83c6cb12019-10-18 14:18:30 -07009713/**
9714 * lpfc_map_topology - Map the topology read from READ_CONFIG
9715 * @phba: pointer to lpfc hba data structure.
Lee Jonesfe614ac2020-07-23 13:24:22 +01009716 * @rd_config: pointer to read config data
James Smart83c6cb12019-10-18 14:18:30 -07009717 *
9718 * This routine is invoked to map the topology values as read
9719 * from the read config mailbox command. If the persistent
9720 * topology feature is supported, the firmware will provide the
9721 * saved topology information to be used in INIT_LINK
James Smart83c6cb12019-10-18 14:18:30 -07009722 **/
James Smart83c6cb12019-10-18 14:18:30 -07009723static void
9724lpfc_map_topology(struct lpfc_hba *phba, struct lpfc_mbx_read_config *rd_config)
9725{
9726 u8 ptv, tf, pt;
9727
9728 ptv = bf_get(lpfc_mbx_rd_conf_ptv, rd_config);
9729 tf = bf_get(lpfc_mbx_rd_conf_tf, rd_config);
9730 pt = bf_get(lpfc_mbx_rd_conf_pt, rd_config);
9731
9732 lpfc_printf_log(phba, KERN_INFO, LOG_SLI,
9733 "2027 Read Config Data : ptv:0x%x, tf:0x%x pt:0x%x",
9734 ptv, tf, pt);
9735 if (!ptv) {
9736 lpfc_printf_log(phba, KERN_WARNING, LOG_SLI,
9737 "2019 FW does not support persistent topology "
9738 "Using driver parameter defined value [%s]",
9739 lpfc_topo_to_str[phba->cfg_topology]);
9740 return;
9741 }
9742 /* FW supports persistent topology - override module parameter value */
9743 phba->hba_flag |= HBA_PERSISTENT_TOPO;
James Smartf6c5e6c2021-07-22 15:17:18 -07009744
9745 /* if ASIC_GEN_NUM >= 0xC) */
9746 if ((bf_get(lpfc_sli_intf_if_type, &phba->sli4_hba.sli_intf) ==
9747 LPFC_SLI_INTF_IF_TYPE_6) ||
9748 (bf_get(lpfc_sli_intf_sli_family, &phba->sli4_hba.sli_intf) ==
9749 LPFC_SLI_INTF_FAMILY_G6)) {
James Smart83c6cb12019-10-18 14:18:30 -07009750 if (!tf) {
9751 phba->cfg_topology = ((pt == LINK_FLAGS_LOOP)
9752 ? FLAGS_TOPOLOGY_MODE_LOOP
9753 : FLAGS_TOPOLOGY_MODE_PT_PT);
9754 } else {
9755 phba->hba_flag &= ~HBA_PERSISTENT_TOPO;
9756 }
James Smartf6c5e6c2021-07-22 15:17:18 -07009757 } else { /* G5 */
James Smart83c6cb12019-10-18 14:18:30 -07009758 if (tf) {
9759 /* If topology failover set - pt is '0' or '1' */
9760 phba->cfg_topology = (pt ? FLAGS_TOPOLOGY_MODE_PT_LOOP :
9761 FLAGS_TOPOLOGY_MODE_LOOP_PT);
9762 } else {
9763 phba->cfg_topology = ((pt == LINK_FLAGS_P2P)
9764 ? FLAGS_TOPOLOGY_MODE_PT_PT
9765 : FLAGS_TOPOLOGY_MODE_LOOP);
9766 }
James Smart83c6cb12019-10-18 14:18:30 -07009767 }
9768 if (phba->hba_flag & HBA_PERSISTENT_TOPO) {
9769 lpfc_printf_log(phba, KERN_INFO, LOG_SLI,
9770 "2020 Using persistent topology value [%s]",
9771 lpfc_topo_to_str[phba->cfg_topology]);
9772 } else {
9773 lpfc_printf_log(phba, KERN_WARNING, LOG_SLI,
9774 "2021 Invalid topology values from FW "
9775 "Using driver parameter defined value [%s]",
9776 lpfc_topo_to_str[phba->cfg_topology]);
9777 }
9778}
9779
James Smartda0436e2009-05-22 14:51:39 -04009780/**
9781 * lpfc_sli4_read_config - Get the config parameters.
9782 * @phba: pointer to lpfc hba data structure.
9783 *
9784 * This routine is invoked to read the configuration parameters from the HBA.
9785 * The configuration parameters are used to set the base and maximum values
9786 * for RPI's XRI's VPI's VFI's and FCFIs. These values also affect the resource
9787 * allocation for the port.
9788 *
9789 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -02009790 * 0 - successful
Lucas De Marchi25985ed2011-03-30 22:57:33 -03009791 * -ENOMEM - No available memory
James Smartd439d282010-09-29 11:18:45 -04009792 * -EIO - The mailbox failed to complete successfully.
James Smartda0436e2009-05-22 14:51:39 -04009793 **/
James Smartff78d8f2011-12-13 13:21:35 -05009794int
James Smartda0436e2009-05-22 14:51:39 -04009795lpfc_sli4_read_config(struct lpfc_hba *phba)
9796{
9797 LPFC_MBOXQ_t *pmb;
9798 struct lpfc_mbx_read_config *rd_config;
James Smart912e3ac2011-05-24 11:42:11 -04009799 union lpfc_sli4_cfg_shdr *shdr;
9800 uint32_t shdr_status, shdr_add_status;
9801 struct lpfc_mbx_get_func_cfg *get_func_cfg;
9802 struct lpfc_rsrc_desc_fcfcoe *desc;
James Smart8aa134a2012-08-14 14:25:29 -04009803 char *pdesc_0;
James Smartc6918162016-10-13 15:06:16 -07009804 uint16_t forced_link_speed;
James Smart6a828b02019-01-28 11:14:31 -08009805 uint32_t if_type, qmin;
James Smart8aa134a2012-08-14 14:25:29 -04009806 int length, i, rc = 0, rc2;
James Smartda0436e2009-05-22 14:51:39 -04009807
9808 pmb = (LPFC_MBOXQ_t *) mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL);
9809 if (!pmb) {
Dick Kennedy372c1872020-06-30 14:50:00 -07009810 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartda0436e2009-05-22 14:51:39 -04009811 "2011 Unable to allocate memory for issuing "
9812 "SLI_CONFIG_SPECIAL mailbox command\n");
9813 return -ENOMEM;
9814 }
9815
9816 lpfc_read_config(phba, pmb);
9817
9818 rc = lpfc_sli_issue_mbox(phba, pmb, MBX_POLL);
9819 if (rc != MBX_SUCCESS) {
Dick Kennedy372c1872020-06-30 14:50:00 -07009820 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
9821 "2012 Mailbox failed , mbxCmd x%x "
9822 "READ_CONFIG, mbxStatus x%x\n",
9823 bf_get(lpfc_mqe_command, &pmb->u.mqe),
9824 bf_get(lpfc_mqe_status, &pmb->u.mqe));
James Smartda0436e2009-05-22 14:51:39 -04009825 rc = -EIO;
9826 } else {
9827 rd_config = &pmb->u.mqe.un.rd_config;
James Smartff78d8f2011-12-13 13:21:35 -05009828 if (bf_get(lpfc_mbx_rd_conf_lnk_ldv, rd_config)) {
9829 phba->sli4_hba.lnk_info.lnk_dv = LPFC_LNK_DAT_VAL;
9830 phba->sli4_hba.lnk_info.lnk_tp =
9831 bf_get(lpfc_mbx_rd_conf_lnk_type, rd_config);
9832 phba->sli4_hba.lnk_info.lnk_no =
9833 bf_get(lpfc_mbx_rd_conf_lnk_numb, rd_config);
9834 lpfc_printf_log(phba, KERN_INFO, LOG_SLI,
9835 "3081 lnk_type:%d, lnk_numb:%d\n",
9836 phba->sli4_hba.lnk_info.lnk_tp,
9837 phba->sli4_hba.lnk_info.lnk_no);
9838 } else
9839 lpfc_printf_log(phba, KERN_WARNING, LOG_SLI,
9840 "3082 Mailbox (x%x) returned ldv:x0\n",
9841 bf_get(lpfc_mqe_command, &pmb->u.mqe));
James Smart44fd7fe2017-08-23 16:55:47 -07009842 if (bf_get(lpfc_mbx_rd_conf_bbscn_def, rd_config)) {
9843 phba->bbcredit_support = 1;
9844 phba->sli4_hba.bbscn_params.word0 = rd_config->word8;
9845 }
9846
James Smart1dc5ec22018-10-23 13:41:11 -07009847 phba->sli4_hba.conf_trunk =
9848 bf_get(lpfc_mbx_rd_conf_trunk, rd_config);
James Smart6d368e52011-05-24 11:44:12 -04009849 phba->sli4_hba.extents_in_use =
9850 bf_get(lpfc_mbx_rd_conf_extnts_inuse, rd_config);
James Smartda0436e2009-05-22 14:51:39 -04009851 phba->sli4_hba.max_cfg_param.max_xri =
9852 bf_get(lpfc_mbx_rd_conf_xri_count, rd_config);
James Smart31f06d22019-08-14 16:56:31 -07009853 /* Reduce resource usage in kdump environment */
9854 if (is_kdump_kernel() &&
9855 phba->sli4_hba.max_cfg_param.max_xri > 512)
9856 phba->sli4_hba.max_cfg_param.max_xri = 512;
James Smartda0436e2009-05-22 14:51:39 -04009857 phba->sli4_hba.max_cfg_param.xri_base =
9858 bf_get(lpfc_mbx_rd_conf_xri_base, rd_config);
9859 phba->sli4_hba.max_cfg_param.max_vpi =
9860 bf_get(lpfc_mbx_rd_conf_vpi_count, rd_config);
James Smart8b47ae62018-11-29 16:09:33 -08009861 /* Limit the max we support */
9862 if (phba->sli4_hba.max_cfg_param.max_vpi > LPFC_MAX_VPORTS)
9863 phba->sli4_hba.max_cfg_param.max_vpi = LPFC_MAX_VPORTS;
James Smartda0436e2009-05-22 14:51:39 -04009864 phba->sli4_hba.max_cfg_param.vpi_base =
9865 bf_get(lpfc_mbx_rd_conf_vpi_base, rd_config);
9866 phba->sli4_hba.max_cfg_param.max_rpi =
9867 bf_get(lpfc_mbx_rd_conf_rpi_count, rd_config);
9868 phba->sli4_hba.max_cfg_param.rpi_base =
9869 bf_get(lpfc_mbx_rd_conf_rpi_base, rd_config);
9870 phba->sli4_hba.max_cfg_param.max_vfi =
9871 bf_get(lpfc_mbx_rd_conf_vfi_count, rd_config);
9872 phba->sli4_hba.max_cfg_param.vfi_base =
9873 bf_get(lpfc_mbx_rd_conf_vfi_base, rd_config);
9874 phba->sli4_hba.max_cfg_param.max_fcfi =
9875 bf_get(lpfc_mbx_rd_conf_fcfi_count, rd_config);
James Smartda0436e2009-05-22 14:51:39 -04009876 phba->sli4_hba.max_cfg_param.max_eq =
9877 bf_get(lpfc_mbx_rd_conf_eq_count, rd_config);
9878 phba->sli4_hba.max_cfg_param.max_rq =
9879 bf_get(lpfc_mbx_rd_conf_rq_count, rd_config);
9880 phba->sli4_hba.max_cfg_param.max_wq =
9881 bf_get(lpfc_mbx_rd_conf_wq_count, rd_config);
9882 phba->sli4_hba.max_cfg_param.max_cq =
9883 bf_get(lpfc_mbx_rd_conf_cq_count, rd_config);
9884 phba->lmt = bf_get(lpfc_mbx_rd_conf_lmt, rd_config);
9885 phba->sli4_hba.next_xri = phba->sli4_hba.max_cfg_param.xri_base;
9886 phba->vpi_base = phba->sli4_hba.max_cfg_param.vpi_base;
9887 phba->vfi_base = phba->sli4_hba.max_cfg_param.vfi_base;
James Smart5ffc2662009-11-18 15:39:44 -05009888 phba->max_vpi = (phba->sli4_hba.max_cfg_param.max_vpi > 0) ?
9889 (phba->sli4_hba.max_cfg_param.max_vpi - 1) : 0;
James Smartda0436e2009-05-22 14:51:39 -04009890 phba->max_vports = phba->max_vpi;
James Smart9064aeb2021-08-16 09:28:50 -07009891
9892 /* Next decide on FPIN or Signal E2E CGN support
9893 * For congestion alarms and warnings valid combination are:
9894 * 1. FPIN alarms / FPIN warnings
9895 * 2. Signal alarms / Signal warnings
9896 * 3. FPIN alarms / Signal warnings
9897 * 4. Signal alarms / FPIN warnings
9898 *
9899 * Initialize the adapter frequency to 100 mSecs
9900 */
9901 phba->cgn_reg_fpin = LPFC_CGN_FPIN_BOTH;
9902 phba->cgn_reg_signal = EDC_CG_SIG_NOTSUPPORTED;
9903 phba->cgn_sig_freq = lpfc_fabric_cgn_frequency;
9904
9905 if (lpfc_use_cgn_signal) {
9906 if (bf_get(lpfc_mbx_rd_conf_wcs, rd_config)) {
9907 phba->cgn_reg_signal = EDC_CG_SIG_WARN_ONLY;
9908 phba->cgn_reg_fpin &= ~LPFC_CGN_FPIN_WARN;
9909 }
9910 if (bf_get(lpfc_mbx_rd_conf_acs, rd_config)) {
9911 /* MUST support both alarm and warning
9912 * because EDC does not support alarm alone.
9913 */
9914 if (phba->cgn_reg_signal !=
9915 EDC_CG_SIG_WARN_ONLY) {
9916 /* Must support both or none */
9917 phba->cgn_reg_fpin = LPFC_CGN_FPIN_BOTH;
9918 phba->cgn_reg_signal =
9919 EDC_CG_SIG_NOTSUPPORTED;
9920 } else {
9921 phba->cgn_reg_signal =
9922 EDC_CG_SIG_WARN_ALARM;
9923 phba->cgn_reg_fpin =
9924 LPFC_CGN_FPIN_NONE;
9925 }
9926 }
9927 }
9928
9929 /* Set the congestion initial signal and fpin values. */
9930 phba->cgn_init_reg_fpin = phba->cgn_reg_fpin;
9931 phba->cgn_init_reg_signal = phba->cgn_reg_signal;
9932
9933 lpfc_printf_log(phba, KERN_INFO, LOG_CGN_MGMT,
9934 "6446 READ_CONFIG reg_sig x%x reg_fpin:x%x\n",
9935 phba->cgn_reg_signal, phba->cgn_reg_fpin);
9936
James Smart83c6cb12019-10-18 14:18:30 -07009937 lpfc_map_topology(phba, rd_config);
James Smartda0436e2009-05-22 14:51:39 -04009938 lpfc_printf_log(phba, KERN_INFO, LOG_SLI,
James Smart6d368e52011-05-24 11:44:12 -04009939 "2003 cfg params Extents? %d "
9940 "XRI(B:%d M:%d), "
James Smartda0436e2009-05-22 14:51:39 -04009941 "VPI(B:%d M:%d) "
9942 "VFI(B:%d M:%d) "
9943 "RPI(B:%d M:%d) "
Dick Kennedya1e4d3d2020-08-03 14:02:22 -07009944 "FCFI:%d EQ:%d CQ:%d WQ:%d RQ:%d lmt:x%x\n",
James Smart6d368e52011-05-24 11:44:12 -04009945 phba->sli4_hba.extents_in_use,
James Smartda0436e2009-05-22 14:51:39 -04009946 phba->sli4_hba.max_cfg_param.xri_base,
9947 phba->sli4_hba.max_cfg_param.max_xri,
9948 phba->sli4_hba.max_cfg_param.vpi_base,
9949 phba->sli4_hba.max_cfg_param.max_vpi,
9950 phba->sli4_hba.max_cfg_param.vfi_base,
9951 phba->sli4_hba.max_cfg_param.max_vfi,
9952 phba->sli4_hba.max_cfg_param.rpi_base,
9953 phba->sli4_hba.max_cfg_param.max_rpi,
James Smart2ea259e2017-02-12 13:52:27 -08009954 phba->sli4_hba.max_cfg_param.max_fcfi,
9955 phba->sli4_hba.max_cfg_param.max_eq,
9956 phba->sli4_hba.max_cfg_param.max_cq,
9957 phba->sli4_hba.max_cfg_param.max_wq,
Dick Kennedya1e4d3d2020-08-03 14:02:22 -07009958 phba->sli4_hba.max_cfg_param.max_rq,
9959 phba->lmt);
James Smart2ea259e2017-02-12 13:52:27 -08009960
James Smartd38f33b2018-05-04 20:37:54 -07009961 /*
James Smart6a828b02019-01-28 11:14:31 -08009962 * Calculate queue resources based on how
9963 * many WQ/CQ/EQs are available.
James Smartd38f33b2018-05-04 20:37:54 -07009964 */
James Smart6a828b02019-01-28 11:14:31 -08009965 qmin = phba->sli4_hba.max_cfg_param.max_wq;
9966 if (phba->sli4_hba.max_cfg_param.max_cq < qmin)
9967 qmin = phba->sli4_hba.max_cfg_param.max_cq;
9968 if (phba->sli4_hba.max_cfg_param.max_eq < qmin)
9969 qmin = phba->sli4_hba.max_cfg_param.max_eq;
9970 /*
9971 * Whats left after this can go toward NVME / FCP.
9972 * The minus 4 accounts for ELS, NVME LS, MBOX
9973 * plus one extra. When configured for
9974 * NVMET, FCP io channel WQs are not created.
9975 */
9976 qmin -= 4;
James Smartd38f33b2018-05-04 20:37:54 -07009977
James Smart6a828b02019-01-28 11:14:31 -08009978 /* Check to see if there is enough for NVME */
9979 if ((phba->cfg_irq_chann > qmin) ||
9980 (phba->cfg_hdw_queue > qmin)) {
Dick Kennedy372c1872020-06-30 14:50:00 -07009981 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
Dick Kennedy9e3e3652020-08-03 14:02:23 -07009982 "2005 Reducing Queues - "
9983 "FW resource limitation: "
James Smart6a828b02019-01-28 11:14:31 -08009984 "WQ %d CQ %d EQ %d: min %d: "
9985 "IRQ %d HDWQ %d\n",
James Smartd38f33b2018-05-04 20:37:54 -07009986 phba->sli4_hba.max_cfg_param.max_wq,
9987 phba->sli4_hba.max_cfg_param.max_cq,
James Smart6a828b02019-01-28 11:14:31 -08009988 phba->sli4_hba.max_cfg_param.max_eq,
9989 qmin, phba->cfg_irq_chann,
James Smartcdb42be2019-01-28 11:14:21 -08009990 phba->cfg_hdw_queue);
James Smartd38f33b2018-05-04 20:37:54 -07009991
James Smart6a828b02019-01-28 11:14:31 -08009992 if (phba->cfg_irq_chann > qmin)
9993 phba->cfg_irq_chann = qmin;
9994 if (phba->cfg_hdw_queue > qmin)
9995 phba->cfg_hdw_queue = qmin;
James Smartd38f33b2018-05-04 20:37:54 -07009996 }
James Smartda0436e2009-05-22 14:51:39 -04009997 }
James Smart912e3ac2011-05-24 11:42:11 -04009998
9999 if (rc)
10000 goto read_cfg_out;
James Smartda0436e2009-05-22 14:51:39 -040010001
James Smartc6918162016-10-13 15:06:16 -070010002 /* Update link speed if forced link speed is supported */
10003 if_type = bf_get(lpfc_sli_intf_if_type, &phba->sli4_hba.sli_intf);
James Smart27d6ac02018-02-22 08:18:42 -080010004 if (if_type >= LPFC_SLI_INTF_IF_TYPE_2) {
James Smartc6918162016-10-13 15:06:16 -070010005 forced_link_speed =
10006 bf_get(lpfc_mbx_rd_conf_link_speed, rd_config);
10007 if (forced_link_speed) {
10008 phba->hba_flag |= HBA_FORCED_LINK_SPEED;
10009
10010 switch (forced_link_speed) {
10011 case LINK_SPEED_1G:
10012 phba->cfg_link_speed =
10013 LPFC_USER_LINK_SPEED_1G;
10014 break;
10015 case LINK_SPEED_2G:
10016 phba->cfg_link_speed =
10017 LPFC_USER_LINK_SPEED_2G;
10018 break;
10019 case LINK_SPEED_4G:
10020 phba->cfg_link_speed =
10021 LPFC_USER_LINK_SPEED_4G;
10022 break;
10023 case LINK_SPEED_8G:
10024 phba->cfg_link_speed =
10025 LPFC_USER_LINK_SPEED_8G;
10026 break;
10027 case LINK_SPEED_10G:
10028 phba->cfg_link_speed =
10029 LPFC_USER_LINK_SPEED_10G;
10030 break;
10031 case LINK_SPEED_16G:
10032 phba->cfg_link_speed =
10033 LPFC_USER_LINK_SPEED_16G;
10034 break;
10035 case LINK_SPEED_32G:
10036 phba->cfg_link_speed =
10037 LPFC_USER_LINK_SPEED_32G;
10038 break;
James Smartfbd8a6b2018-02-22 08:18:45 -080010039 case LINK_SPEED_64G:
10040 phba->cfg_link_speed =
10041 LPFC_USER_LINK_SPEED_64G;
10042 break;
James Smartc6918162016-10-13 15:06:16 -070010043 case 0xffff:
10044 phba->cfg_link_speed =
10045 LPFC_USER_LINK_SPEED_AUTO;
10046 break;
10047 default:
Dick Kennedy372c1872020-06-30 14:50:00 -070010048 lpfc_printf_log(phba, KERN_ERR,
10049 LOG_TRACE_EVENT,
James Smartc6918162016-10-13 15:06:16 -070010050 "0047 Unrecognized link "
10051 "speed : %d\n",
10052 forced_link_speed);
10053 phba->cfg_link_speed =
10054 LPFC_USER_LINK_SPEED_AUTO;
10055 }
10056 }
10057 }
10058
James Smartda0436e2009-05-22 14:51:39 -040010059 /* Reset the DFT_HBA_Q_DEPTH to the max xri */
James Smart572709e2013-07-15 18:32:43 -040010060 length = phba->sli4_hba.max_cfg_param.max_xri -
10061 lpfc_sli4_get_els_iocb_cnt(phba);
10062 if (phba->cfg_hba_queue_depth > length) {
10063 lpfc_printf_log(phba, KERN_WARNING, LOG_INIT,
10064 "3361 HBA queue depth changed from %d to %d\n",
10065 phba->cfg_hba_queue_depth, length);
10066 phba->cfg_hba_queue_depth = length;
10067 }
James Smart912e3ac2011-05-24 11:42:11 -040010068
James Smart27d6ac02018-02-22 08:18:42 -080010069 if (bf_get(lpfc_sli_intf_if_type, &phba->sli4_hba.sli_intf) <
James Smart912e3ac2011-05-24 11:42:11 -040010070 LPFC_SLI_INTF_IF_TYPE_2)
10071 goto read_cfg_out;
10072
10073 /* get the pf# and vf# for SLI4 if_type 2 port */
10074 length = (sizeof(struct lpfc_mbx_get_func_cfg) -
10075 sizeof(struct lpfc_sli4_cfg_mhdr));
10076 lpfc_sli4_config(phba, pmb, LPFC_MBOX_SUBSYSTEM_COMMON,
10077 LPFC_MBOX_OPCODE_GET_FUNCTION_CONFIG,
10078 length, LPFC_SLI4_MBX_EMBED);
10079
James Smart8aa134a2012-08-14 14:25:29 -040010080 rc2 = lpfc_sli_issue_mbox(phba, pmb, MBX_POLL);
James Smart912e3ac2011-05-24 11:42:11 -040010081 shdr = (union lpfc_sli4_cfg_shdr *)
10082 &pmb->u.mqe.un.sli4_config.header.cfg_shdr;
10083 shdr_status = bf_get(lpfc_mbox_hdr_status, &shdr->response);
10084 shdr_add_status = bf_get(lpfc_mbox_hdr_add_status, &shdr->response);
James Smart8aa134a2012-08-14 14:25:29 -040010085 if (rc2 || shdr_status || shdr_add_status) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010086 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart912e3ac2011-05-24 11:42:11 -040010087 "3026 Mailbox failed , mbxCmd x%x "
10088 "GET_FUNCTION_CONFIG, mbxStatus x%x\n",
10089 bf_get(lpfc_mqe_command, &pmb->u.mqe),
10090 bf_get(lpfc_mqe_status, &pmb->u.mqe));
James Smart912e3ac2011-05-24 11:42:11 -040010091 goto read_cfg_out;
10092 }
10093
10094 /* search for fc_fcoe resrouce descriptor */
10095 get_func_cfg = &pmb->u.mqe.un.get_func_cfg;
James Smart912e3ac2011-05-24 11:42:11 -040010096
James Smart8aa134a2012-08-14 14:25:29 -040010097 pdesc_0 = (char *)&get_func_cfg->func_cfg.desc[0];
10098 desc = (struct lpfc_rsrc_desc_fcfcoe *)pdesc_0;
10099 length = bf_get(lpfc_rsrc_desc_fcfcoe_length, desc);
10100 if (length == LPFC_RSRC_DESC_TYPE_FCFCOE_V0_RSVD)
10101 length = LPFC_RSRC_DESC_TYPE_FCFCOE_V0_LENGTH;
10102 else if (length != LPFC_RSRC_DESC_TYPE_FCFCOE_V1_LENGTH)
10103 goto read_cfg_out;
10104
James Smart912e3ac2011-05-24 11:42:11 -040010105 for (i = 0; i < LPFC_RSRC_DESC_MAX_NUM; i++) {
James Smart8aa134a2012-08-14 14:25:29 -040010106 desc = (struct lpfc_rsrc_desc_fcfcoe *)(pdesc_0 + length * i);
James Smart912e3ac2011-05-24 11:42:11 -040010107 if (LPFC_RSRC_DESC_TYPE_FCFCOE ==
James Smart8aa134a2012-08-14 14:25:29 -040010108 bf_get(lpfc_rsrc_desc_fcfcoe_type, desc)) {
James Smart912e3ac2011-05-24 11:42:11 -040010109 phba->sli4_hba.iov.pf_number =
10110 bf_get(lpfc_rsrc_desc_fcfcoe_pfnum, desc);
10111 phba->sli4_hba.iov.vf_number =
10112 bf_get(lpfc_rsrc_desc_fcfcoe_vfnum, desc);
10113 break;
10114 }
10115 }
10116
10117 if (i < LPFC_RSRC_DESC_MAX_NUM)
10118 lpfc_printf_log(phba, KERN_INFO, LOG_SLI,
10119 "3027 GET_FUNCTION_CONFIG: pf_number:%d, "
10120 "vf_number:%d\n", phba->sli4_hba.iov.pf_number,
10121 phba->sli4_hba.iov.vf_number);
James Smart8aa134a2012-08-14 14:25:29 -040010122 else
Dick Kennedy372c1872020-06-30 14:50:00 -070010123 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart912e3ac2011-05-24 11:42:11 -040010124 "3028 GET_FUNCTION_CONFIG: failed to find "
Colin Ian Kingc4dba182018-10-16 18:28:53 +010010125 "Resource Descriptor:x%x\n",
James Smart912e3ac2011-05-24 11:42:11 -040010126 LPFC_RSRC_DESC_TYPE_FCFCOE);
James Smart912e3ac2011-05-24 11:42:11 -040010127
10128read_cfg_out:
10129 mempool_free(pmb, phba->mbox_mem_pool);
James Smartda0436e2009-05-22 14:51:39 -040010130 return rc;
10131}
10132
10133/**
James Smart2fcee4b2010-12-15 17:57:46 -050010134 * lpfc_setup_endian_order - Write endian order to an SLI4 if_type 0 port.
James Smartda0436e2009-05-22 14:51:39 -040010135 * @phba: pointer to lpfc hba data structure.
10136 *
James Smart2fcee4b2010-12-15 17:57:46 -050010137 * This routine is invoked to setup the port-side endian order when
10138 * the port if_type is 0. This routine has no function for other
10139 * if_types.
James Smartda0436e2009-05-22 14:51:39 -040010140 *
10141 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -020010142 * 0 - successful
Lucas De Marchi25985ed2011-03-30 22:57:33 -030010143 * -ENOMEM - No available memory
James Smartd439d282010-09-29 11:18:45 -040010144 * -EIO - The mailbox failed to complete successfully.
James Smartda0436e2009-05-22 14:51:39 -040010145 **/
10146static int
10147lpfc_setup_endian_order(struct lpfc_hba *phba)
10148{
10149 LPFC_MBOXQ_t *mboxq;
James Smart2fcee4b2010-12-15 17:57:46 -050010150 uint32_t if_type, rc = 0;
James Smartda0436e2009-05-22 14:51:39 -040010151 uint32_t endian_mb_data[2] = {HOST_ENDIAN_LOW_WORD0,
10152 HOST_ENDIAN_HIGH_WORD1};
10153
James Smart2fcee4b2010-12-15 17:57:46 -050010154 if_type = bf_get(lpfc_sli_intf_if_type, &phba->sli4_hba.sli_intf);
10155 switch (if_type) {
10156 case LPFC_SLI_INTF_IF_TYPE_0:
10157 mboxq = (LPFC_MBOXQ_t *) mempool_alloc(phba->mbox_mem_pool,
10158 GFP_KERNEL);
10159 if (!mboxq) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010160 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart2fcee4b2010-12-15 17:57:46 -050010161 "0492 Unable to allocate memory for "
10162 "issuing SLI_CONFIG_SPECIAL mailbox "
10163 "command\n");
10164 return -ENOMEM;
10165 }
James Smartda0436e2009-05-22 14:51:39 -040010166
James Smart2fcee4b2010-12-15 17:57:46 -050010167 /*
10168 * The SLI4_CONFIG_SPECIAL mailbox command requires the first
10169 * two words to contain special data values and no other data.
10170 */
10171 memset(mboxq, 0, sizeof(LPFC_MBOXQ_t));
10172 memcpy(&mboxq->u.mqe, &endian_mb_data, sizeof(endian_mb_data));
10173 rc = lpfc_sli_issue_mbox(phba, mboxq, MBX_POLL);
10174 if (rc != MBX_SUCCESS) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010175 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart2fcee4b2010-12-15 17:57:46 -050010176 "0493 SLI_CONFIG_SPECIAL mailbox "
10177 "failed with status x%x\n",
10178 rc);
10179 rc = -EIO;
10180 }
10181 mempool_free(mboxq, phba->mbox_mem_pool);
10182 break;
James Smart27d6ac02018-02-22 08:18:42 -080010183 case LPFC_SLI_INTF_IF_TYPE_6:
James Smart2fcee4b2010-12-15 17:57:46 -050010184 case LPFC_SLI_INTF_IF_TYPE_2:
10185 case LPFC_SLI_INTF_IF_TYPE_1:
10186 default:
10187 break;
James Smartda0436e2009-05-22 14:51:39 -040010188 }
James Smartda0436e2009-05-22 14:51:39 -040010189 return rc;
10190}
10191
10192/**
James Smart895427b2017-02-12 13:52:30 -080010193 * lpfc_sli4_queue_verify - Verify and update EQ counts
James Smartda0436e2009-05-22 14:51:39 -040010194 * @phba: pointer to lpfc hba data structure.
10195 *
James Smart895427b2017-02-12 13:52:30 -080010196 * This routine is invoked to check the user settable queue counts for EQs.
10197 * After this routine is called the counts will be set to valid values that
James Smart5350d872011-10-10 21:33:49 -040010198 * adhere to the constraints of the system's interrupt vectors and the port's
10199 * queue resources.
James Smartda0436e2009-05-22 14:51:39 -040010200 *
10201 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -020010202 * 0 - successful
Lucas De Marchi25985ed2011-03-30 22:57:33 -030010203 * -ENOMEM - No available memory
James Smartda0436e2009-05-22 14:51:39 -040010204 **/
10205static int
James Smart5350d872011-10-10 21:33:49 -040010206lpfc_sli4_queue_verify(struct lpfc_hba *phba)
James Smartda0436e2009-05-22 14:51:39 -040010207{
James Smartda0436e2009-05-22 14:51:39 -040010208 /*
James Smart67d12732012-08-03 12:36:13 -040010209 * Sanity check for configured queue parameters against the run-time
James Smartda0436e2009-05-22 14:51:39 -040010210 * device parameters
10211 */
10212
James Smartbcb24f62017-11-20 16:00:36 -080010213 if (phba->nvmet_support) {
James Smart97a9ed32019-10-18 14:18:17 -070010214 if (phba->cfg_hdw_queue < phba->cfg_nvmet_mrq)
10215 phba->cfg_nvmet_mrq = phba->cfg_hdw_queue;
James Smart982ab122019-03-12 16:30:10 -070010216 if (phba->cfg_nvmet_mrq > LPFC_NVMET_MRQ_MAX)
10217 phba->cfg_nvmet_mrq = LPFC_NVMET_MRQ_MAX;
James Smartbcb24f62017-11-20 16:00:36 -080010218 }
James Smart895427b2017-02-12 13:52:30 -080010219
10220 lpfc_printf_log(phba, KERN_ERR, LOG_INIT,
James Smart6a828b02019-01-28 11:14:31 -080010221 "2574 IO channels: hdwQ %d IRQ %d MRQ: %d\n",
10222 phba->cfg_hdw_queue, phba->cfg_irq_chann,
10223 phba->cfg_nvmet_mrq);
James Smartda0436e2009-05-22 14:51:39 -040010224
James Smartda0436e2009-05-22 14:51:39 -040010225 /* Get EQ depth from module parameter, fake the default for now */
10226 phba->sli4_hba.eq_esize = LPFC_EQE_SIZE_4B;
10227 phba->sli4_hba.eq_ecount = LPFC_EQE_DEF_COUNT;
10228
James Smart5350d872011-10-10 21:33:49 -040010229 /* Get CQ depth from module parameter, fake the default for now */
10230 phba->sli4_hba.cq_esize = LPFC_CQE_SIZE;
10231 phba->sli4_hba.cq_ecount = LPFC_CQE_DEF_COUNT;
James Smart5350d872011-10-10 21:33:49 -040010232 return 0;
James Smart895427b2017-02-12 13:52:30 -080010233}
10234
10235static int
James Smartc00f62e2019-08-14 16:57:11 -070010236lpfc_alloc_io_wq_cq(struct lpfc_hba *phba, int idx)
James Smart895427b2017-02-12 13:52:30 -080010237{
10238 struct lpfc_queue *qdesc;
James Smartc00f62e2019-08-14 16:57:11 -070010239 u32 wqesize;
James Smartc1a21eb2019-03-12 16:30:29 -070010240 int cpu;
James Smart895427b2017-02-12 13:52:30 -080010241
James Smartc00f62e2019-08-14 16:57:11 -070010242 cpu = lpfc_find_cpu_handle(phba, idx, LPFC_FIND_BY_HDWQ);
10243 /* Create Fast Path IO CQs */
James Smartc176ffa2018-01-30 15:58:46 -080010244 if (phba->enab_exp_wqcq_pages)
James Smarta51e41b2017-12-08 17:18:06 -080010245 /* Increase the CQ size when WQEs contain an embedded cdb */
10246 qdesc = lpfc_sli4_queue_alloc(phba, LPFC_EXPANDED_PAGE_SIZE,
10247 phba->sli4_hba.cq_esize,
James Smartc1a21eb2019-03-12 16:30:29 -070010248 LPFC_CQE_EXP_COUNT, cpu);
James Smarta51e41b2017-12-08 17:18:06 -080010249
10250 else
10251 qdesc = lpfc_sli4_queue_alloc(phba, LPFC_DEFAULT_PAGE_SIZE,
10252 phba->sli4_hba.cq_esize,
James Smartc1a21eb2019-03-12 16:30:29 -070010253 phba->sli4_hba.cq_ecount, cpu);
James Smart895427b2017-02-12 13:52:30 -080010254 if (!qdesc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010255 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
10256 "0499 Failed allocate fast-path IO CQ (%d)\n",
10257 idx);
James Smart895427b2017-02-12 13:52:30 -080010258 return 1;
10259 }
James Smart7365f6f2018-02-22 08:18:46 -080010260 qdesc->qe_valid = 1;
James Smartc00f62e2019-08-14 16:57:11 -070010261 qdesc->hdwq = idx;
James Smartc1a21eb2019-03-12 16:30:29 -070010262 qdesc->chann = cpu;
James Smartc00f62e2019-08-14 16:57:11 -070010263 phba->sli4_hba.hdwq[idx].io_cq = qdesc;
James Smart895427b2017-02-12 13:52:30 -080010264
James Smartc00f62e2019-08-14 16:57:11 -070010265 /* Create Fast Path IO WQs */
James Smartc176ffa2018-01-30 15:58:46 -080010266 if (phba->enab_exp_wqcq_pages) {
James Smarta51e41b2017-12-08 17:18:06 -080010267 /* Increase the WQ size when WQEs contain an embedded cdb */
James Smartc176ffa2018-01-30 15:58:46 -080010268 wqesize = (phba->fcp_embed_io) ?
10269 LPFC_WQE128_SIZE : phba->sli4_hba.wq_esize;
James Smarta51e41b2017-12-08 17:18:06 -080010270 qdesc = lpfc_sli4_queue_alloc(phba, LPFC_EXPANDED_PAGE_SIZE,
James Smartc176ffa2018-01-30 15:58:46 -080010271 wqesize,
James Smartc1a21eb2019-03-12 16:30:29 -070010272 LPFC_WQE_EXP_COUNT, cpu);
James Smartc176ffa2018-01-30 15:58:46 -080010273 } else
James Smarta51e41b2017-12-08 17:18:06 -080010274 qdesc = lpfc_sli4_queue_alloc(phba, LPFC_DEFAULT_PAGE_SIZE,
10275 phba->sli4_hba.wq_esize,
James Smartc1a21eb2019-03-12 16:30:29 -070010276 phba->sli4_hba.wq_ecount, cpu);
James Smartc176ffa2018-01-30 15:58:46 -080010277
James Smart895427b2017-02-12 13:52:30 -080010278 if (!qdesc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010279 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartc00f62e2019-08-14 16:57:11 -070010280 "0503 Failed allocate fast-path IO WQ (%d)\n",
10281 idx);
James Smart895427b2017-02-12 13:52:30 -080010282 return 1;
10283 }
James Smartc00f62e2019-08-14 16:57:11 -070010284 qdesc->hdwq = idx;
10285 qdesc->chann = cpu;
10286 phba->sli4_hba.hdwq[idx].io_wq = qdesc;
James Smart895427b2017-02-12 13:52:30 -080010287 list_add_tail(&qdesc->wq_list, &phba->sli4_hba.lpfc_wq_list);
10288 return 0;
James Smart5350d872011-10-10 21:33:49 -040010289}
10290
10291/**
10292 * lpfc_sli4_queue_create - Create all the SLI4 queues
10293 * @phba: pointer to lpfc hba data structure.
10294 *
10295 * This routine is invoked to allocate all the SLI4 queues for the FCoE HBA
10296 * operation. For each SLI4 queue type, the parameters such as queue entry
10297 * count (queue depth) shall be taken from the module parameter. For now,
10298 * we just use some constant number as place holder.
10299 *
10300 * Return codes
Anatol Pomozov4907cb72012-09-01 10:31:09 -070010301 * 0 - successful
James Smart5350d872011-10-10 21:33:49 -040010302 * -ENOMEM - No availble memory
10303 * -EIO - The mailbox failed to complete successfully.
10304 **/
10305int
10306lpfc_sli4_queue_create(struct lpfc_hba *phba)
10307{
10308 struct lpfc_queue *qdesc;
James Smart657add42019-05-21 17:49:06 -070010309 int idx, cpu, eqcpu;
James Smart5e5b5112019-01-28 11:14:22 -080010310 struct lpfc_sli4_hdw_queue *qp;
James Smart657add42019-05-21 17:49:06 -070010311 struct lpfc_vector_map_info *cpup;
10312 struct lpfc_vector_map_info *eqcpup;
James Smart32517fc2019-01-28 11:14:33 -080010313 struct lpfc_eq_intr_info *eqi;
James Smart5350d872011-10-10 21:33:49 -040010314
10315 /*
James Smart67d12732012-08-03 12:36:13 -040010316 * Create HBA Record arrays.
James Smart895427b2017-02-12 13:52:30 -080010317 * Both NVME and FCP will share that same vectors / EQs
James Smart5350d872011-10-10 21:33:49 -040010318 */
James Smart67d12732012-08-03 12:36:13 -040010319 phba->sli4_hba.mq_esize = LPFC_MQE_SIZE;
10320 phba->sli4_hba.mq_ecount = LPFC_MQE_DEF_COUNT;
10321 phba->sli4_hba.wq_esize = LPFC_WQE_SIZE;
10322 phba->sli4_hba.wq_ecount = LPFC_WQE_DEF_COUNT;
10323 phba->sli4_hba.rq_esize = LPFC_RQE_SIZE;
10324 phba->sli4_hba.rq_ecount = LPFC_RQE_DEF_COUNT;
James Smart895427b2017-02-12 13:52:30 -080010325 phba->sli4_hba.eq_esize = LPFC_EQE_SIZE_4B;
10326 phba->sli4_hba.eq_ecount = LPFC_EQE_DEF_COUNT;
10327 phba->sli4_hba.cq_esize = LPFC_CQE_SIZE;
10328 phba->sli4_hba.cq_ecount = LPFC_CQE_DEF_COUNT;
James Smart67d12732012-08-03 12:36:13 -040010329
James Smartcdb42be2019-01-28 11:14:21 -080010330 if (!phba->sli4_hba.hdwq) {
James Smart5e5b5112019-01-28 11:14:22 -080010331 phba->sli4_hba.hdwq = kcalloc(
10332 phba->cfg_hdw_queue, sizeof(struct lpfc_sli4_hdw_queue),
10333 GFP_KERNEL);
10334 if (!phba->sli4_hba.hdwq) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010335 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart5e5b5112019-01-28 11:14:22 -080010336 "6427 Failed allocate memory for "
10337 "fast-path Hardware Queue array\n");
James Smart895427b2017-02-12 13:52:30 -080010338 goto out_error;
10339 }
James Smart5e5b5112019-01-28 11:14:22 -080010340 /* Prepare hardware queues to take IO buffers */
10341 for (idx = 0; idx < phba->cfg_hdw_queue; idx++) {
10342 qp = &phba->sli4_hba.hdwq[idx];
10343 spin_lock_init(&qp->io_buf_list_get_lock);
10344 spin_lock_init(&qp->io_buf_list_put_lock);
10345 INIT_LIST_HEAD(&qp->lpfc_io_buf_list_get);
10346 INIT_LIST_HEAD(&qp->lpfc_io_buf_list_put);
10347 qp->get_io_bufs = 0;
10348 qp->put_io_bufs = 0;
10349 qp->total_io_bufs = 0;
James Smartc00f62e2019-08-14 16:57:11 -070010350 spin_lock_init(&qp->abts_io_buf_list_lock);
10351 INIT_LIST_HEAD(&qp->lpfc_abts_io_buf_list);
James Smart5e5b5112019-01-28 11:14:22 -080010352 qp->abts_scsi_io_bufs = 0;
James Smart5e5b5112019-01-28 11:14:22 -080010353 qp->abts_nvme_io_bufs = 0;
James Smartd79c9e92019-08-14 16:57:09 -070010354 INIT_LIST_HEAD(&qp->sgl_list);
10355 INIT_LIST_HEAD(&qp->cmd_rsp_buf_list);
10356 spin_lock_init(&qp->hdwq_lock);
James Smart895427b2017-02-12 13:52:30 -080010357 }
James Smart67d12732012-08-03 12:36:13 -040010358 }
10359
James Smartcdb42be2019-01-28 11:14:21 -080010360 if (phba->cfg_enable_fc4_type & LPFC_ENABLE_NVME) {
James Smart2d7dbc42017-02-12 13:52:35 -080010361 if (phba->nvmet_support) {
10362 phba->sli4_hba.nvmet_cqset = kcalloc(
10363 phba->cfg_nvmet_mrq,
10364 sizeof(struct lpfc_queue *),
10365 GFP_KERNEL);
10366 if (!phba->sli4_hba.nvmet_cqset) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010367 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart2d7dbc42017-02-12 13:52:35 -080010368 "3121 Fail allocate memory for "
10369 "fast-path CQ set array\n");
10370 goto out_error;
10371 }
10372 phba->sli4_hba.nvmet_mrq_hdr = kcalloc(
10373 phba->cfg_nvmet_mrq,
10374 sizeof(struct lpfc_queue *),
10375 GFP_KERNEL);
10376 if (!phba->sli4_hba.nvmet_mrq_hdr) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010377 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart2d7dbc42017-02-12 13:52:35 -080010378 "3122 Fail allocate memory for "
10379 "fast-path RQ set hdr array\n");
10380 goto out_error;
10381 }
10382 phba->sli4_hba.nvmet_mrq_data = kcalloc(
10383 phba->cfg_nvmet_mrq,
10384 sizeof(struct lpfc_queue *),
10385 GFP_KERNEL);
10386 if (!phba->sli4_hba.nvmet_mrq_data) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010387 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart2d7dbc42017-02-12 13:52:35 -080010388 "3124 Fail allocate memory for "
10389 "fast-path RQ set data array\n");
10390 goto out_error;
10391 }
10392 }
James Smart67d12732012-08-03 12:36:13 -040010393 }
James Smartda0436e2009-05-22 14:51:39 -040010394
James Smart895427b2017-02-12 13:52:30 -080010395 INIT_LIST_HEAD(&phba->sli4_hba.lpfc_wq_list);
James Smart67d12732012-08-03 12:36:13 -040010396
James Smart895427b2017-02-12 13:52:30 -080010397 /* Create HBA Event Queues (EQs) */
James Smart657add42019-05-21 17:49:06 -070010398 for_each_present_cpu(cpu) {
10399 /* We only want to create 1 EQ per vector, even though
10400 * multiple CPUs might be using that vector. so only
10401 * selects the CPUs that are LPFC_CPU_FIRST_IRQ.
James Smart6a828b02019-01-28 11:14:31 -080010402 */
James Smart657add42019-05-21 17:49:06 -070010403 cpup = &phba->sli4_hba.cpu_map[cpu];
10404 if (!(cpup->flag & LPFC_CPU_FIRST_IRQ))
James Smart6a828b02019-01-28 11:14:31 -080010405 continue;
James Smart657add42019-05-21 17:49:06 -070010406
10407 /* Get a ptr to the Hardware Queue associated with this CPU */
10408 qp = &phba->sli4_hba.hdwq[cpup->hdwq];
10409
10410 /* Allocate an EQ */
James Smart81b96ed2017-11-20 16:00:29 -080010411 qdesc = lpfc_sli4_queue_alloc(phba, LPFC_DEFAULT_PAGE_SIZE,
10412 phba->sli4_hba.eq_esize,
James Smartc1a21eb2019-03-12 16:30:29 -070010413 phba->sli4_hba.eq_ecount, cpu);
James Smartda0436e2009-05-22 14:51:39 -040010414 if (!qdesc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010415 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart657add42019-05-21 17:49:06 -070010416 "0497 Failed allocate EQ (%d)\n",
10417 cpup->hdwq);
James Smart67d12732012-08-03 12:36:13 -040010418 goto out_error;
James Smartda0436e2009-05-22 14:51:39 -040010419 }
James Smart7365f6f2018-02-22 08:18:46 -080010420 qdesc->qe_valid = 1;
James Smart657add42019-05-21 17:49:06 -070010421 qdesc->hdwq = cpup->hdwq;
James Smart3ad348d2019-08-14 16:56:43 -070010422 qdesc->chann = cpu; /* First CPU this EQ is affinitized to */
James Smart32517fc2019-01-28 11:14:33 -080010423 qdesc->last_cpu = qdesc->chann;
James Smart657add42019-05-21 17:49:06 -070010424
10425 /* Save the allocated EQ in the Hardware Queue */
10426 qp->hba_eq = qdesc;
10427
James Smart32517fc2019-01-28 11:14:33 -080010428 eqi = per_cpu_ptr(phba->sli4_hba.eq_info, qdesc->last_cpu);
10429 list_add(&qdesc->cpu_list, &eqi->list);
James Smartda0436e2009-05-22 14:51:39 -040010430 }
10431
James Smart657add42019-05-21 17:49:06 -070010432 /* Now we need to populate the other Hardware Queues, that share
10433 * an IRQ vector, with the associated EQ ptr.
10434 */
10435 for_each_present_cpu(cpu) {
10436 cpup = &phba->sli4_hba.cpu_map[cpu];
10437
10438 /* Check for EQ already allocated in previous loop */
10439 if (cpup->flag & LPFC_CPU_FIRST_IRQ)
10440 continue;
10441
10442 /* Check for multiple CPUs per hdwq */
10443 qp = &phba->sli4_hba.hdwq[cpup->hdwq];
10444 if (qp->hba_eq)
10445 continue;
10446
10447 /* We need to share an EQ for this hdwq */
10448 eqcpu = lpfc_find_cpu_handle(phba, cpup->eq, LPFC_FIND_BY_EQ);
10449 eqcpup = &phba->sli4_hba.cpu_map[eqcpu];
10450 qp->hba_eq = phba->sli4_hba.hdwq[eqcpup->hdwq].hba_eq;
10451 }
James Smart895427b2017-02-12 13:52:30 -080010452
James Smartc00f62e2019-08-14 16:57:11 -070010453 /* Allocate IO Path SLI4 CQ/WQs */
James Smart6a828b02019-01-28 11:14:31 -080010454 for (idx = 0; idx < phba->cfg_hdw_queue; idx++) {
James Smartc00f62e2019-08-14 16:57:11 -070010455 if (lpfc_alloc_io_wq_cq(phba, idx))
James Smart895427b2017-02-12 13:52:30 -080010456 goto out_error;
James Smart6a828b02019-01-28 11:14:31 -080010457 }
James Smart895427b2017-02-12 13:52:30 -080010458
James Smartc00f62e2019-08-14 16:57:11 -070010459 if (phba->nvmet_support) {
10460 for (idx = 0; idx < phba->cfg_nvmet_mrq; idx++) {
10461 cpu = lpfc_find_cpu_handle(phba, idx,
10462 LPFC_FIND_BY_HDWQ);
10463 qdesc = lpfc_sli4_queue_alloc(phba,
James Smart81b96ed2017-11-20 16:00:29 -080010464 LPFC_DEFAULT_PAGE_SIZE,
10465 phba->sli4_hba.cq_esize,
James Smartc1a21eb2019-03-12 16:30:29 -070010466 phba->sli4_hba.cq_ecount,
10467 cpu);
James Smartc00f62e2019-08-14 16:57:11 -070010468 if (!qdesc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010469 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartcdb42be2019-01-28 11:14:21 -080010470 "3142 Failed allocate NVME "
10471 "CQ Set (%d)\n", idx);
James Smartc00f62e2019-08-14 16:57:11 -070010472 goto out_error;
James Smart2d7dbc42017-02-12 13:52:35 -080010473 }
James Smartc00f62e2019-08-14 16:57:11 -070010474 qdesc->qe_valid = 1;
10475 qdesc->hdwq = idx;
10476 qdesc->chann = cpu;
10477 phba->sli4_hba.nvmet_cqset[idx] = qdesc;
James Smart2d7dbc42017-02-12 13:52:35 -080010478 }
10479 }
10480
James Smartda0436e2009-05-22 14:51:39 -040010481 /*
James Smart67d12732012-08-03 12:36:13 -040010482 * Create Slow Path Completion Queues (CQs)
James Smartda0436e2009-05-22 14:51:39 -040010483 */
10484
James Smartc1a21eb2019-03-12 16:30:29 -070010485 cpu = lpfc_find_cpu_handle(phba, 0, LPFC_FIND_BY_EQ);
James Smartda0436e2009-05-22 14:51:39 -040010486 /* Create slow-path Mailbox Command Complete Queue */
James Smart81b96ed2017-11-20 16:00:29 -080010487 qdesc = lpfc_sli4_queue_alloc(phba, LPFC_DEFAULT_PAGE_SIZE,
10488 phba->sli4_hba.cq_esize,
James Smartc1a21eb2019-03-12 16:30:29 -070010489 phba->sli4_hba.cq_ecount, cpu);
James Smartda0436e2009-05-22 14:51:39 -040010490 if (!qdesc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010491 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartda0436e2009-05-22 14:51:39 -040010492 "0500 Failed allocate slow-path mailbox CQ\n");
James Smart67d12732012-08-03 12:36:13 -040010493 goto out_error;
James Smartda0436e2009-05-22 14:51:39 -040010494 }
James Smart7365f6f2018-02-22 08:18:46 -080010495 qdesc->qe_valid = 1;
James Smartda0436e2009-05-22 14:51:39 -040010496 phba->sli4_hba.mbx_cq = qdesc;
10497
10498 /* Create slow-path ELS Complete Queue */
James Smart81b96ed2017-11-20 16:00:29 -080010499 qdesc = lpfc_sli4_queue_alloc(phba, LPFC_DEFAULT_PAGE_SIZE,
10500 phba->sli4_hba.cq_esize,
James Smartc1a21eb2019-03-12 16:30:29 -070010501 phba->sli4_hba.cq_ecount, cpu);
James Smartda0436e2009-05-22 14:51:39 -040010502 if (!qdesc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010503 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartda0436e2009-05-22 14:51:39 -040010504 "0501 Failed allocate slow-path ELS CQ\n");
James Smart67d12732012-08-03 12:36:13 -040010505 goto out_error;
James Smartda0436e2009-05-22 14:51:39 -040010506 }
James Smart7365f6f2018-02-22 08:18:46 -080010507 qdesc->qe_valid = 1;
James Smartc00f62e2019-08-14 16:57:11 -070010508 qdesc->chann = cpu;
James Smartda0436e2009-05-22 14:51:39 -040010509 phba->sli4_hba.els_cq = qdesc;
10510
James Smartda0436e2009-05-22 14:51:39 -040010511
James Smart5350d872011-10-10 21:33:49 -040010512 /*
James Smart67d12732012-08-03 12:36:13 -040010513 * Create Slow Path Work Queues (WQs)
James Smart5350d872011-10-10 21:33:49 -040010514 */
James Smartda0436e2009-05-22 14:51:39 -040010515
10516 /* Create Mailbox Command Queue */
James Smartda0436e2009-05-22 14:51:39 -040010517
James Smart81b96ed2017-11-20 16:00:29 -080010518 qdesc = lpfc_sli4_queue_alloc(phba, LPFC_DEFAULT_PAGE_SIZE,
10519 phba->sli4_hba.mq_esize,
James Smartc1a21eb2019-03-12 16:30:29 -070010520 phba->sli4_hba.mq_ecount, cpu);
James Smartda0436e2009-05-22 14:51:39 -040010521 if (!qdesc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010522 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartda0436e2009-05-22 14:51:39 -040010523 "0505 Failed allocate slow-path MQ\n");
James Smart67d12732012-08-03 12:36:13 -040010524 goto out_error;
James Smartda0436e2009-05-22 14:51:39 -040010525 }
James Smartc00f62e2019-08-14 16:57:11 -070010526 qdesc->chann = cpu;
James Smartda0436e2009-05-22 14:51:39 -040010527 phba->sli4_hba.mbx_wq = qdesc;
10528
10529 /*
James Smart67d12732012-08-03 12:36:13 -040010530 * Create ELS Work Queues
James Smartda0436e2009-05-22 14:51:39 -040010531 */
James Smartda0436e2009-05-22 14:51:39 -040010532
10533 /* Create slow-path ELS Work Queue */
James Smart81b96ed2017-11-20 16:00:29 -080010534 qdesc = lpfc_sli4_queue_alloc(phba, LPFC_DEFAULT_PAGE_SIZE,
10535 phba->sli4_hba.wq_esize,
James Smartc1a21eb2019-03-12 16:30:29 -070010536 phba->sli4_hba.wq_ecount, cpu);
James Smartda0436e2009-05-22 14:51:39 -040010537 if (!qdesc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010538 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartda0436e2009-05-22 14:51:39 -040010539 "0504 Failed allocate slow-path ELS WQ\n");
James Smart67d12732012-08-03 12:36:13 -040010540 goto out_error;
James Smartda0436e2009-05-22 14:51:39 -040010541 }
James Smartc00f62e2019-08-14 16:57:11 -070010542 qdesc->chann = cpu;
James Smartda0436e2009-05-22 14:51:39 -040010543 phba->sli4_hba.els_wq = qdesc;
James Smart895427b2017-02-12 13:52:30 -080010544 list_add_tail(&qdesc->wq_list, &phba->sli4_hba.lpfc_wq_list);
10545
10546 if (phba->cfg_enable_fc4_type & LPFC_ENABLE_NVME) {
10547 /* Create NVME LS Complete Queue */
James Smart81b96ed2017-11-20 16:00:29 -080010548 qdesc = lpfc_sli4_queue_alloc(phba, LPFC_DEFAULT_PAGE_SIZE,
10549 phba->sli4_hba.cq_esize,
James Smartc1a21eb2019-03-12 16:30:29 -070010550 phba->sli4_hba.cq_ecount, cpu);
James Smart895427b2017-02-12 13:52:30 -080010551 if (!qdesc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010552 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart895427b2017-02-12 13:52:30 -080010553 "6079 Failed allocate NVME LS CQ\n");
10554 goto out_error;
10555 }
James Smartc00f62e2019-08-14 16:57:11 -070010556 qdesc->chann = cpu;
James Smart7365f6f2018-02-22 08:18:46 -080010557 qdesc->qe_valid = 1;
James Smart895427b2017-02-12 13:52:30 -080010558 phba->sli4_hba.nvmels_cq = qdesc;
10559
10560 /* Create NVME LS Work Queue */
James Smart81b96ed2017-11-20 16:00:29 -080010561 qdesc = lpfc_sli4_queue_alloc(phba, LPFC_DEFAULT_PAGE_SIZE,
10562 phba->sli4_hba.wq_esize,
James Smartc1a21eb2019-03-12 16:30:29 -070010563 phba->sli4_hba.wq_ecount, cpu);
James Smart895427b2017-02-12 13:52:30 -080010564 if (!qdesc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010565 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart895427b2017-02-12 13:52:30 -080010566 "6080 Failed allocate NVME LS WQ\n");
10567 goto out_error;
10568 }
James Smartc00f62e2019-08-14 16:57:11 -070010569 qdesc->chann = cpu;
James Smart895427b2017-02-12 13:52:30 -080010570 phba->sli4_hba.nvmels_wq = qdesc;
10571 list_add_tail(&qdesc->wq_list, &phba->sli4_hba.lpfc_wq_list);
10572 }
James Smartda0436e2009-05-22 14:51:39 -040010573
James Smartda0436e2009-05-22 14:51:39 -040010574 /*
10575 * Create Receive Queue (RQ)
10576 */
James Smartda0436e2009-05-22 14:51:39 -040010577
10578 /* Create Receive Queue for header */
James Smart81b96ed2017-11-20 16:00:29 -080010579 qdesc = lpfc_sli4_queue_alloc(phba, LPFC_DEFAULT_PAGE_SIZE,
10580 phba->sli4_hba.rq_esize,
James Smartc1a21eb2019-03-12 16:30:29 -070010581 phba->sli4_hba.rq_ecount, cpu);
James Smartda0436e2009-05-22 14:51:39 -040010582 if (!qdesc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010583 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartda0436e2009-05-22 14:51:39 -040010584 "0506 Failed allocate receive HRQ\n");
James Smart67d12732012-08-03 12:36:13 -040010585 goto out_error;
James Smartda0436e2009-05-22 14:51:39 -040010586 }
10587 phba->sli4_hba.hdr_rq = qdesc;
10588
10589 /* Create Receive Queue for data */
James Smart81b96ed2017-11-20 16:00:29 -080010590 qdesc = lpfc_sli4_queue_alloc(phba, LPFC_DEFAULT_PAGE_SIZE,
10591 phba->sli4_hba.rq_esize,
James Smartc1a21eb2019-03-12 16:30:29 -070010592 phba->sli4_hba.rq_ecount, cpu);
James Smartda0436e2009-05-22 14:51:39 -040010593 if (!qdesc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010594 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartda0436e2009-05-22 14:51:39 -040010595 "0507 Failed allocate receive DRQ\n");
James Smart67d12732012-08-03 12:36:13 -040010596 goto out_error;
James Smartda0436e2009-05-22 14:51:39 -040010597 }
10598 phba->sli4_hba.dat_rq = qdesc;
10599
James Smartcdb42be2019-01-28 11:14:21 -080010600 if ((phba->cfg_enable_fc4_type & LPFC_ENABLE_NVME) &&
10601 phba->nvmet_support) {
James Smart2d7dbc42017-02-12 13:52:35 -080010602 for (idx = 0; idx < phba->cfg_nvmet_mrq; idx++) {
James Smartc1a21eb2019-03-12 16:30:29 -070010603 cpu = lpfc_find_cpu_handle(phba, idx,
10604 LPFC_FIND_BY_HDWQ);
James Smart2d7dbc42017-02-12 13:52:35 -080010605 /* Create NVMET Receive Queue for header */
10606 qdesc = lpfc_sli4_queue_alloc(phba,
James Smart81b96ed2017-11-20 16:00:29 -080010607 LPFC_DEFAULT_PAGE_SIZE,
James Smart2d7dbc42017-02-12 13:52:35 -080010608 phba->sli4_hba.rq_esize,
James Smartc1a21eb2019-03-12 16:30:29 -070010609 LPFC_NVMET_RQE_DEF_COUNT,
10610 cpu);
James Smart2d7dbc42017-02-12 13:52:35 -080010611 if (!qdesc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010612 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart2d7dbc42017-02-12 13:52:35 -080010613 "3146 Failed allocate "
10614 "receive HRQ\n");
10615 goto out_error;
10616 }
James Smart5e5b5112019-01-28 11:14:22 -080010617 qdesc->hdwq = idx;
James Smart2d7dbc42017-02-12 13:52:35 -080010618 phba->sli4_hba.nvmet_mrq_hdr[idx] = qdesc;
10619
10620 /* Only needed for header of RQ pair */
James Smartc1a21eb2019-03-12 16:30:29 -070010621 qdesc->rqbp = kzalloc_node(sizeof(*qdesc->rqbp),
10622 GFP_KERNEL,
10623 cpu_to_node(cpu));
James Smart2d7dbc42017-02-12 13:52:35 -080010624 if (qdesc->rqbp == NULL) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010625 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart2d7dbc42017-02-12 13:52:35 -080010626 "6131 Failed allocate "
10627 "Header RQBP\n");
10628 goto out_error;
10629 }
10630
Dick Kennedy4b40d022017-08-23 16:55:38 -070010631 /* Put list in known state in case driver load fails. */
10632 INIT_LIST_HEAD(&qdesc->rqbp->rqb_buffer_list);
10633
James Smart2d7dbc42017-02-12 13:52:35 -080010634 /* Create NVMET Receive Queue for data */
10635 qdesc = lpfc_sli4_queue_alloc(phba,
James Smart81b96ed2017-11-20 16:00:29 -080010636 LPFC_DEFAULT_PAGE_SIZE,
James Smart2d7dbc42017-02-12 13:52:35 -080010637 phba->sli4_hba.rq_esize,
James Smartc1a21eb2019-03-12 16:30:29 -070010638 LPFC_NVMET_RQE_DEF_COUNT,
10639 cpu);
James Smart2d7dbc42017-02-12 13:52:35 -080010640 if (!qdesc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010641 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart2d7dbc42017-02-12 13:52:35 -080010642 "3156 Failed allocate "
10643 "receive DRQ\n");
10644 goto out_error;
10645 }
James Smart5e5b5112019-01-28 11:14:22 -080010646 qdesc->hdwq = idx;
James Smart2d7dbc42017-02-12 13:52:35 -080010647 phba->sli4_hba.nvmet_mrq_data[idx] = qdesc;
10648 }
10649 }
10650
James Smart4c47efc2019-01-28 11:14:25 -080010651 /* Clear NVME stats */
10652 if (phba->cfg_enable_fc4_type & LPFC_ENABLE_NVME) {
10653 for (idx = 0; idx < phba->cfg_hdw_queue; idx++) {
10654 memset(&phba->sli4_hba.hdwq[idx].nvme_cstat, 0,
10655 sizeof(phba->sli4_hba.hdwq[idx].nvme_cstat));
10656 }
10657 }
James Smart4c47efc2019-01-28 11:14:25 -080010658
10659 /* Clear SCSI stats */
10660 if (phba->cfg_enable_fc4_type & LPFC_ENABLE_FCP) {
10661 for (idx = 0; idx < phba->cfg_hdw_queue; idx++) {
10662 memset(&phba->sli4_hba.hdwq[idx].scsi_cstat, 0,
10663 sizeof(phba->sli4_hba.hdwq[idx].scsi_cstat));
10664 }
10665 }
10666
James Smartda0436e2009-05-22 14:51:39 -040010667 return 0;
10668
James Smartda0436e2009-05-22 14:51:39 -040010669out_error:
James Smart67d12732012-08-03 12:36:13 -040010670 lpfc_sli4_queue_destroy(phba);
James Smartda0436e2009-05-22 14:51:39 -040010671 return -ENOMEM;
10672}
10673
James Smart895427b2017-02-12 13:52:30 -080010674static inline void
10675__lpfc_sli4_release_queue(struct lpfc_queue **qp)
10676{
10677 if (*qp != NULL) {
10678 lpfc_sli4_queue_free(*qp);
10679 *qp = NULL;
10680 }
10681}
10682
10683static inline void
10684lpfc_sli4_release_queues(struct lpfc_queue ***qs, int max)
10685{
10686 int idx;
10687
10688 if (*qs == NULL)
10689 return;
10690
10691 for (idx = 0; idx < max; idx++)
10692 __lpfc_sli4_release_queue(&(*qs)[idx]);
10693
10694 kfree(*qs);
10695 *qs = NULL;
10696}
10697
10698static inline void
James Smart6a828b02019-01-28 11:14:31 -080010699lpfc_sli4_release_hdwq(struct lpfc_hba *phba)
James Smart895427b2017-02-12 13:52:30 -080010700{
James Smart6a828b02019-01-28 11:14:31 -080010701 struct lpfc_sli4_hdw_queue *hdwq;
James Smart657add42019-05-21 17:49:06 -070010702 struct lpfc_queue *eq;
James Smartcdb42be2019-01-28 11:14:21 -080010703 uint32_t idx;
10704
James Smart6a828b02019-01-28 11:14:31 -080010705 hdwq = phba->sli4_hba.hdwq;
James Smart6a828b02019-01-28 11:14:31 -080010706
James Smart657add42019-05-21 17:49:06 -070010707 /* Loop thru all Hardware Queues */
10708 for (idx = 0; idx < phba->cfg_hdw_queue; idx++) {
10709 /* Free the CQ/WQ corresponding to the Hardware Queue */
James Smartc00f62e2019-08-14 16:57:11 -070010710 lpfc_sli4_queue_free(hdwq[idx].io_cq);
10711 lpfc_sli4_queue_free(hdwq[idx].io_wq);
James Smart821bc882020-01-27 16:23:05 -080010712 hdwq[idx].hba_eq = NULL;
James Smartc00f62e2019-08-14 16:57:11 -070010713 hdwq[idx].io_cq = NULL;
10714 hdwq[idx].io_wq = NULL;
James Smartd79c9e92019-08-14 16:57:09 -070010715 if (phba->cfg_xpsgl && !phba->nvmet_support)
10716 lpfc_free_sgl_per_hdwq(phba, &hdwq[idx]);
10717 lpfc_free_cmd_rsp_buf_per_hdwq(phba, &hdwq[idx]);
James Smart895427b2017-02-12 13:52:30 -080010718 }
James Smart657add42019-05-21 17:49:06 -070010719 /* Loop thru all IRQ vectors */
10720 for (idx = 0; idx < phba->cfg_irq_chann; idx++) {
10721 /* Free the EQ corresponding to the IRQ vector */
10722 eq = phba->sli4_hba.hba_eq_hdl[idx].eq;
10723 lpfc_sli4_queue_free(eq);
10724 phba->sli4_hba.hba_eq_hdl[idx].eq = NULL;
10725 }
James Smart895427b2017-02-12 13:52:30 -080010726}
10727
James Smartda0436e2009-05-22 14:51:39 -040010728/**
10729 * lpfc_sli4_queue_destroy - Destroy all the SLI4 queues
10730 * @phba: pointer to lpfc hba data structure.
10731 *
10732 * This routine is invoked to release all the SLI4 queues with the FCoE HBA
10733 * operation.
10734 *
10735 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -020010736 * 0 - successful
Lucas De Marchi25985ed2011-03-30 22:57:33 -030010737 * -ENOMEM - No available memory
James Smartd439d282010-09-29 11:18:45 -040010738 * -EIO - The mailbox failed to complete successfully.
James Smartda0436e2009-05-22 14:51:39 -040010739 **/
James Smart5350d872011-10-10 21:33:49 -040010740void
James Smartda0436e2009-05-22 14:51:39 -040010741lpfc_sli4_queue_destroy(struct lpfc_hba *phba)
10742{
James Smart4645f7b2019-03-12 16:30:14 -070010743 /*
10744 * Set FREE_INIT before beginning to free the queues.
10745 * Wait until the users of queues to acknowledge to
10746 * release queues by clearing FREE_WAIT.
10747 */
10748 spin_lock_irq(&phba->hbalock);
10749 phba->sli.sli_flag |= LPFC_QUEUE_FREE_INIT;
10750 while (phba->sli.sli_flag & LPFC_QUEUE_FREE_WAIT) {
10751 spin_unlock_irq(&phba->hbalock);
10752 msleep(20);
10753 spin_lock_irq(&phba->hbalock);
10754 }
10755 spin_unlock_irq(&phba->hbalock);
10756
James Smart93a4d6f2019-11-04 16:57:05 -080010757 lpfc_sli4_cleanup_poll_list(phba);
10758
James Smart895427b2017-02-12 13:52:30 -080010759 /* Release HBA eqs */
James Smartcdb42be2019-01-28 11:14:21 -080010760 if (phba->sli4_hba.hdwq)
James Smart6a828b02019-01-28 11:14:31 -080010761 lpfc_sli4_release_hdwq(phba);
James Smartda0436e2009-05-22 14:51:39 -040010762
James Smartbcb24f62017-11-20 16:00:36 -080010763 if (phba->nvmet_support) {
10764 lpfc_sli4_release_queues(&phba->sli4_hba.nvmet_cqset,
10765 phba->cfg_nvmet_mrq);
James Smart2d7dbc42017-02-12 13:52:35 -080010766
James Smartbcb24f62017-11-20 16:00:36 -080010767 lpfc_sli4_release_queues(&phba->sli4_hba.nvmet_mrq_hdr,
10768 phba->cfg_nvmet_mrq);
10769 lpfc_sli4_release_queues(&phba->sli4_hba.nvmet_mrq_data,
10770 phba->cfg_nvmet_mrq);
10771 }
James Smart2d7dbc42017-02-12 13:52:35 -080010772
James Smartda0436e2009-05-22 14:51:39 -040010773 /* Release mailbox command work queue */
James Smart895427b2017-02-12 13:52:30 -080010774 __lpfc_sli4_release_queue(&phba->sli4_hba.mbx_wq);
James Smartda0436e2009-05-22 14:51:39 -040010775
10776 /* Release ELS work queue */
James Smart895427b2017-02-12 13:52:30 -080010777 __lpfc_sli4_release_queue(&phba->sli4_hba.els_wq);
10778
10779 /* Release ELS work queue */
10780 __lpfc_sli4_release_queue(&phba->sli4_hba.nvmels_wq);
James Smartda0436e2009-05-22 14:51:39 -040010781
10782 /* Release unsolicited receive queue */
James Smart895427b2017-02-12 13:52:30 -080010783 __lpfc_sli4_release_queue(&phba->sli4_hba.hdr_rq);
10784 __lpfc_sli4_release_queue(&phba->sli4_hba.dat_rq);
James Smartda0436e2009-05-22 14:51:39 -040010785
James Smartda0436e2009-05-22 14:51:39 -040010786 /* Release ELS complete queue */
James Smart895427b2017-02-12 13:52:30 -080010787 __lpfc_sli4_release_queue(&phba->sli4_hba.els_cq);
10788
10789 /* Release NVME LS complete queue */
10790 __lpfc_sli4_release_queue(&phba->sli4_hba.nvmels_cq);
James Smartda0436e2009-05-22 14:51:39 -040010791
10792 /* Release mailbox command complete queue */
James Smart895427b2017-02-12 13:52:30 -080010793 __lpfc_sli4_release_queue(&phba->sli4_hba.mbx_cq);
10794
10795 /* Everything on this list has been freed */
10796 INIT_LIST_HEAD(&phba->sli4_hba.lpfc_wq_list);
James Smart4645f7b2019-03-12 16:30:14 -070010797
10798 /* Done with freeing the queues */
10799 spin_lock_irq(&phba->hbalock);
10800 phba->sli.sli_flag &= ~LPFC_QUEUE_FREE_INIT;
10801 spin_unlock_irq(&phba->hbalock);
James Smart895427b2017-02-12 13:52:30 -080010802}
10803
10804int
James Smart895427b2017-02-12 13:52:30 -080010805lpfc_free_rq_buffer(struct lpfc_hba *phba, struct lpfc_queue *rq)
10806{
10807 struct lpfc_rqb *rqbp;
10808 struct lpfc_dmabuf *h_buf;
10809 struct rqb_dmabuf *rqb_buffer;
10810
10811 rqbp = rq->rqbp;
10812 while (!list_empty(&rqbp->rqb_buffer_list)) {
10813 list_remove_head(&rqbp->rqb_buffer_list, h_buf,
10814 struct lpfc_dmabuf, list);
10815
10816 rqb_buffer = container_of(h_buf, struct rqb_dmabuf, hbuf);
10817 (rqbp->rqb_free_buffer)(phba, rqb_buffer);
10818 rqbp->buffer_count--;
10819 }
10820 return 1;
10821}
10822
10823static int
10824lpfc_create_wq_cq(struct lpfc_hba *phba, struct lpfc_queue *eq,
10825 struct lpfc_queue *cq, struct lpfc_queue *wq, uint16_t *cq_map,
10826 int qidx, uint32_t qtype)
10827{
10828 struct lpfc_sli_ring *pring;
10829 int rc;
10830
10831 if (!eq || !cq || !wq) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010832 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart895427b2017-02-12 13:52:30 -080010833 "6085 Fast-path %s (%d) not allocated\n",
10834 ((eq) ? ((cq) ? "WQ" : "CQ") : "EQ"), qidx);
10835 return -ENOMEM;
James Smart67d12732012-08-03 12:36:13 -040010836 }
James Smartda0436e2009-05-22 14:51:39 -040010837
James Smart895427b2017-02-12 13:52:30 -080010838 /* create the Cq first */
10839 rc = lpfc_cq_create(phba, cq, eq,
10840 (qtype == LPFC_MBOX) ? LPFC_MCQ : LPFC_WCQ, qtype);
10841 if (rc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010842 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
10843 "6086 Failed setup of CQ (%d), rc = 0x%x\n",
10844 qidx, (uint32_t)rc);
James Smart895427b2017-02-12 13:52:30 -080010845 return rc;
10846 }
10847
10848 if (qtype != LPFC_MBOX) {
James Smartcdb42be2019-01-28 11:14:21 -080010849 /* Setup cq_map for fast lookup */
James Smart895427b2017-02-12 13:52:30 -080010850 if (cq_map)
10851 *cq_map = cq->queue_id;
10852
10853 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
10854 "6087 CQ setup: cq[%d]-id=%d, parent eq[%d]-id=%d\n",
10855 qidx, cq->queue_id, qidx, eq->queue_id);
10856
10857 /* create the wq */
10858 rc = lpfc_wq_create(phba, wq, cq, qtype);
10859 if (rc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010860 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartc835c082019-03-12 16:30:30 -070010861 "4618 Fail setup fastpath WQ (%d), rc = 0x%x\n",
James Smart895427b2017-02-12 13:52:30 -080010862 qidx, (uint32_t)rc);
10863 /* no need to tear down cq - caller will do so */
10864 return rc;
10865 }
10866
10867 /* Bind this CQ/WQ to the NVME ring */
10868 pring = wq->pring;
10869 pring->sli.sli4.wqp = (void *)wq;
10870 cq->pring = pring;
10871
10872 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
10873 "2593 WQ setup: wq[%d]-id=%d assoc=%d, cq[%d]-id=%d\n",
10874 qidx, wq->queue_id, wq->assoc_qid, qidx, cq->queue_id);
10875 } else {
10876 rc = lpfc_mq_create(phba, wq, cq, LPFC_MBOX);
10877 if (rc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010878 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
10879 "0539 Failed setup of slow-path MQ: "
10880 "rc = 0x%x\n", rc);
James Smart895427b2017-02-12 13:52:30 -080010881 /* no need to tear down cq - caller will do so */
10882 return rc;
10883 }
10884
10885 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
10886 "2589 MBX MQ setup: wq-id=%d, parent cq-id=%d\n",
10887 phba->sli4_hba.mbx_wq->queue_id,
10888 phba->sli4_hba.mbx_cq->queue_id);
10889 }
10890
10891 return 0;
James Smartda0436e2009-05-22 14:51:39 -040010892}
10893
10894/**
James Smart6a828b02019-01-28 11:14:31 -080010895 * lpfc_setup_cq_lookup - Setup the CQ lookup table
10896 * @phba: pointer to lpfc hba data structure.
10897 *
10898 * This routine will populate the cq_lookup table by all
10899 * available CQ queue_id's.
10900 **/
Bart Van Assche3999df72019-03-28 11:06:16 -070010901static void
James Smart6a828b02019-01-28 11:14:31 -080010902lpfc_setup_cq_lookup(struct lpfc_hba *phba)
10903{
10904 struct lpfc_queue *eq, *childq;
James Smart6a828b02019-01-28 11:14:31 -080010905 int qidx;
10906
James Smart6a828b02019-01-28 11:14:31 -080010907 memset(phba->sli4_hba.cq_lookup, 0,
10908 (sizeof(struct lpfc_queue *) * (phba->sli4_hba.cq_max + 1)));
James Smart657add42019-05-21 17:49:06 -070010909 /* Loop thru all IRQ vectors */
James Smart6a828b02019-01-28 11:14:31 -080010910 for (qidx = 0; qidx < phba->cfg_irq_chann; qidx++) {
James Smart657add42019-05-21 17:49:06 -070010911 /* Get the EQ corresponding to the IRQ vector */
10912 eq = phba->sli4_hba.hba_eq_hdl[qidx].eq;
James Smart6a828b02019-01-28 11:14:31 -080010913 if (!eq)
10914 continue;
James Smart657add42019-05-21 17:49:06 -070010915 /* Loop through all CQs associated with that EQ */
James Smart6a828b02019-01-28 11:14:31 -080010916 list_for_each_entry(childq, &eq->child_list, list) {
10917 if (childq->queue_id > phba->sli4_hba.cq_max)
10918 continue;
James Smartc00f62e2019-08-14 16:57:11 -070010919 if (childq->subtype == LPFC_IO)
James Smart6a828b02019-01-28 11:14:31 -080010920 phba->sli4_hba.cq_lookup[childq->queue_id] =
10921 childq;
10922 }
10923 }
10924}
10925
10926/**
James Smartda0436e2009-05-22 14:51:39 -040010927 * lpfc_sli4_queue_setup - Set up all the SLI4 queues
10928 * @phba: pointer to lpfc hba data structure.
10929 *
10930 * This routine is invoked to set up all the SLI4 queues for the FCoE HBA
10931 * operation.
10932 *
10933 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -020010934 * 0 - successful
Lucas De Marchi25985ed2011-03-30 22:57:33 -030010935 * -ENOMEM - No available memory
James Smartd439d282010-09-29 11:18:45 -040010936 * -EIO - The mailbox failed to complete successfully.
James Smartda0436e2009-05-22 14:51:39 -040010937 **/
10938int
10939lpfc_sli4_queue_setup(struct lpfc_hba *phba)
10940{
James Smart962bc512013-01-03 15:44:00 -050010941 uint32_t shdr_status, shdr_add_status;
10942 union lpfc_sli4_cfg_shdr *shdr;
James Smart657add42019-05-21 17:49:06 -070010943 struct lpfc_vector_map_info *cpup;
James Smartcdb42be2019-01-28 11:14:21 -080010944 struct lpfc_sli4_hdw_queue *qp;
James Smart962bc512013-01-03 15:44:00 -050010945 LPFC_MBOXQ_t *mboxq;
James Smart657add42019-05-21 17:49:06 -070010946 int qidx, cpu;
James Smartcb733e32019-01-28 11:14:32 -080010947 uint32_t length, usdelay;
James Smart895427b2017-02-12 13:52:30 -080010948 int rc = -ENOMEM;
James Smart962bc512013-01-03 15:44:00 -050010949
10950 /* Check for dual-ULP support */
10951 mboxq = (LPFC_MBOXQ_t *)mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL);
10952 if (!mboxq) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010953 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart962bc512013-01-03 15:44:00 -050010954 "3249 Unable to allocate memory for "
10955 "QUERY_FW_CFG mailbox command\n");
10956 return -ENOMEM;
10957 }
10958 length = (sizeof(struct lpfc_mbx_query_fw_config) -
10959 sizeof(struct lpfc_sli4_cfg_mhdr));
10960 lpfc_sli4_config(phba, mboxq, LPFC_MBOX_SUBSYSTEM_COMMON,
10961 LPFC_MBOX_OPCODE_QUERY_FW_CFG,
10962 length, LPFC_SLI4_MBX_EMBED);
10963
10964 rc = lpfc_sli_issue_mbox(phba, mboxq, MBX_POLL);
10965
10966 shdr = (union lpfc_sli4_cfg_shdr *)
10967 &mboxq->u.mqe.un.sli4_config.header.cfg_shdr;
10968 shdr_status = bf_get(lpfc_mbox_hdr_status, &shdr->response);
10969 shdr_add_status = bf_get(lpfc_mbox_hdr_add_status, &shdr->response);
10970 if (shdr_status || shdr_add_status || rc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070010971 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart962bc512013-01-03 15:44:00 -050010972 "3250 QUERY_FW_CFG mailbox failed with status "
10973 "x%x add_status x%x, mbx status x%x\n",
10974 shdr_status, shdr_add_status, rc);
James Smart304ee432021-04-11 18:31:17 -070010975 mempool_free(mboxq, phba->mbox_mem_pool);
James Smart962bc512013-01-03 15:44:00 -050010976 rc = -ENXIO;
10977 goto out_error;
10978 }
10979
10980 phba->sli4_hba.fw_func_mode =
10981 mboxq->u.mqe.un.query_fw_cfg.rsp.function_mode;
10982 phba->sli4_hba.ulp0_mode = mboxq->u.mqe.un.query_fw_cfg.rsp.ulp0_mode;
10983 phba->sli4_hba.ulp1_mode = mboxq->u.mqe.un.query_fw_cfg.rsp.ulp1_mode;
James Smart8b017a32015-05-21 13:55:18 -040010984 phba->sli4_hba.physical_port =
10985 mboxq->u.mqe.un.query_fw_cfg.rsp.physical_port;
James Smart962bc512013-01-03 15:44:00 -050010986 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
10987 "3251 QUERY_FW_CFG: func_mode:x%x, ulp0_mode:x%x, "
10988 "ulp1_mode:x%x\n", phba->sli4_hba.fw_func_mode,
10989 phba->sli4_hba.ulp0_mode, phba->sli4_hba.ulp1_mode);
10990
James Smart304ee432021-04-11 18:31:17 -070010991 mempool_free(mboxq, phba->mbox_mem_pool);
James Smartda0436e2009-05-22 14:51:39 -040010992
10993 /*
James Smart67d12732012-08-03 12:36:13 -040010994 * Set up HBA Event Queues (EQs)
James Smartda0436e2009-05-22 14:51:39 -040010995 */
James Smartcdb42be2019-01-28 11:14:21 -080010996 qp = phba->sli4_hba.hdwq;
James Smartda0436e2009-05-22 14:51:39 -040010997
James Smart67d12732012-08-03 12:36:13 -040010998 /* Set up HBA event queue */
James Smartcdb42be2019-01-28 11:14:21 -080010999 if (!qp) {
Dick Kennedy372c1872020-06-30 14:50:00 -070011000 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart2e90f4b2011-12-13 13:22:37 -050011001 "3147 Fast-path EQs not allocated\n");
James Smart1b511972011-12-13 13:23:09 -050011002 rc = -ENOMEM;
James Smart67d12732012-08-03 12:36:13 -040011003 goto out_error;
James Smart2e90f4b2011-12-13 13:22:37 -050011004 }
James Smart657add42019-05-21 17:49:06 -070011005
11006 /* Loop thru all IRQ vectors */
James Smart6a828b02019-01-28 11:14:31 -080011007 for (qidx = 0; qidx < phba->cfg_irq_chann; qidx++) {
James Smart657add42019-05-21 17:49:06 -070011008 /* Create HBA Event Queues (EQs) in order */
11009 for_each_present_cpu(cpu) {
11010 cpup = &phba->sli4_hba.cpu_map[cpu];
11011
11012 /* Look for the CPU thats using that vector with
11013 * LPFC_CPU_FIRST_IRQ set.
11014 */
11015 if (!(cpup->flag & LPFC_CPU_FIRST_IRQ))
11016 continue;
11017 if (qidx != cpup->eq)
11018 continue;
11019
11020 /* Create an EQ for that vector */
11021 rc = lpfc_eq_create(phba, qp[cpup->hdwq].hba_eq,
11022 phba->cfg_fcp_imax);
11023 if (rc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070011024 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart657add42019-05-21 17:49:06 -070011025 "0523 Failed setup of fast-path"
11026 " EQ (%d), rc = 0x%x\n",
11027 cpup->eq, (uint32_t)rc);
11028 goto out_destroy;
11029 }
11030
11031 /* Save the EQ for that vector in the hba_eq_hdl */
11032 phba->sli4_hba.hba_eq_hdl[cpup->eq].eq =
11033 qp[cpup->hdwq].hba_eq;
11034
11035 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
11036 "2584 HBA EQ setup: queue[%d]-id=%d\n",
11037 cpup->eq,
11038 qp[cpup->hdwq].hba_eq->queue_id);
James Smartda0436e2009-05-22 14:51:39 -040011039 }
James Smartda0436e2009-05-22 14:51:39 -040011040 }
11041
James Smart657add42019-05-21 17:49:06 -070011042 /* Loop thru all Hardware Queues */
James Smartcdb42be2019-01-28 11:14:21 -080011043 for (qidx = 0; qidx < phba->cfg_hdw_queue; qidx++) {
James Smart657add42019-05-21 17:49:06 -070011044 cpu = lpfc_find_cpu_handle(phba, qidx, LPFC_FIND_BY_HDWQ);
11045 cpup = &phba->sli4_hba.cpu_map[cpu];
11046
11047 /* Create the CQ/WQ corresponding to the Hardware Queue */
James Smartcdb42be2019-01-28 11:14:21 -080011048 rc = lpfc_create_wq_cq(phba,
James Smart657add42019-05-21 17:49:06 -070011049 phba->sli4_hba.hdwq[cpup->hdwq].hba_eq,
James Smartc00f62e2019-08-14 16:57:11 -070011050 qp[qidx].io_cq,
11051 qp[qidx].io_wq,
11052 &phba->sli4_hba.hdwq[qidx].io_cq_map,
11053 qidx,
11054 LPFC_IO);
James Smartcdb42be2019-01-28 11:14:21 -080011055 if (rc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070011056 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart895427b2017-02-12 13:52:30 -080011057 "0535 Failed to setup fastpath "
James Smartc00f62e2019-08-14 16:57:11 -070011058 "IO WQ/CQ (%d), rc = 0x%x\n",
James Smart895427b2017-02-12 13:52:30 -080011059 qidx, (uint32_t)rc);
James Smartcdb42be2019-01-28 11:14:21 -080011060 goto out_destroy;
James Smart895427b2017-02-12 13:52:30 -080011061 }
James Smart67d12732012-08-03 12:36:13 -040011062 }
James Smartda0436e2009-05-22 14:51:39 -040011063
11064 /*
James Smart895427b2017-02-12 13:52:30 -080011065 * Set up Slow Path Complete Queues (CQs)
James Smartda0436e2009-05-22 14:51:39 -040011066 */
11067
James Smart895427b2017-02-12 13:52:30 -080011068 /* Set up slow-path MBOX CQ/MQ */
11069
11070 if (!phba->sli4_hba.mbx_cq || !phba->sli4_hba.mbx_wq) {
Dick Kennedy372c1872020-06-30 14:50:00 -070011071 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart895427b2017-02-12 13:52:30 -080011072 "0528 %s not allocated\n",
11073 phba->sli4_hba.mbx_cq ?
James Smartd1f525a2017-04-21 16:04:55 -070011074 "Mailbox WQ" : "Mailbox CQ");
James Smart1b511972011-12-13 13:23:09 -050011075 rc = -ENOMEM;
James Smart895427b2017-02-12 13:52:30 -080011076 goto out_destroy;
James Smartda0436e2009-05-22 14:51:39 -040011077 }
James Smart895427b2017-02-12 13:52:30 -080011078
James Smartcdb42be2019-01-28 11:14:21 -080011079 rc = lpfc_create_wq_cq(phba, qp[0].hba_eq,
James Smartd1f525a2017-04-21 16:04:55 -070011080 phba->sli4_hba.mbx_cq,
11081 phba->sli4_hba.mbx_wq,
11082 NULL, 0, LPFC_MBOX);
James Smartda0436e2009-05-22 14:51:39 -040011083 if (rc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070011084 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart895427b2017-02-12 13:52:30 -080011085 "0529 Failed setup of mailbox WQ/CQ: rc = 0x%x\n",
11086 (uint32_t)rc);
11087 goto out_destroy;
James Smartda0436e2009-05-22 14:51:39 -040011088 }
James Smart2d7dbc42017-02-12 13:52:35 -080011089 if (phba->nvmet_support) {
11090 if (!phba->sli4_hba.nvmet_cqset) {
Dick Kennedy372c1872020-06-30 14:50:00 -070011091 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart2d7dbc42017-02-12 13:52:35 -080011092 "3165 Fast-path NVME CQ Set "
11093 "array not allocated\n");
11094 rc = -ENOMEM;
11095 goto out_destroy;
11096 }
11097 if (phba->cfg_nvmet_mrq > 1) {
11098 rc = lpfc_cq_create_set(phba,
11099 phba->sli4_hba.nvmet_cqset,
James Smartcdb42be2019-01-28 11:14:21 -080011100 qp,
James Smart2d7dbc42017-02-12 13:52:35 -080011101 LPFC_WCQ, LPFC_NVMET);
11102 if (rc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070011103 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart2d7dbc42017-02-12 13:52:35 -080011104 "3164 Failed setup of NVME CQ "
11105 "Set, rc = 0x%x\n",
11106 (uint32_t)rc);
11107 goto out_destroy;
11108 }
11109 } else {
11110 /* Set up NVMET Receive Complete Queue */
11111 rc = lpfc_cq_create(phba, phba->sli4_hba.nvmet_cqset[0],
James Smartcdb42be2019-01-28 11:14:21 -080011112 qp[0].hba_eq,
James Smart2d7dbc42017-02-12 13:52:35 -080011113 LPFC_WCQ, LPFC_NVMET);
11114 if (rc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070011115 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart2d7dbc42017-02-12 13:52:35 -080011116 "6089 Failed setup NVMET CQ: "
11117 "rc = 0x%x\n", (uint32_t)rc);
11118 goto out_destroy;
11119 }
James Smart81b96ed2017-11-20 16:00:29 -080011120 phba->sli4_hba.nvmet_cqset[0]->chann = 0;
11121
James Smart2d7dbc42017-02-12 13:52:35 -080011122 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
11123 "6090 NVMET CQ setup: cq-id=%d, "
11124 "parent eq-id=%d\n",
11125 phba->sli4_hba.nvmet_cqset[0]->queue_id,
James Smartcdb42be2019-01-28 11:14:21 -080011126 qp[0].hba_eq->queue_id);
James Smart2d7dbc42017-02-12 13:52:35 -080011127 }
11128 }
James Smartda0436e2009-05-22 14:51:39 -040011129
James Smart895427b2017-02-12 13:52:30 -080011130 /* Set up slow-path ELS WQ/CQ */
11131 if (!phba->sli4_hba.els_cq || !phba->sli4_hba.els_wq) {
Dick Kennedy372c1872020-06-30 14:50:00 -070011132 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart895427b2017-02-12 13:52:30 -080011133 "0530 ELS %s not allocated\n",
11134 phba->sli4_hba.els_cq ? "WQ" : "CQ");
James Smart1b511972011-12-13 13:23:09 -050011135 rc = -ENOMEM;
James Smart895427b2017-02-12 13:52:30 -080011136 goto out_destroy;
James Smartda0436e2009-05-22 14:51:39 -040011137 }
James Smartcdb42be2019-01-28 11:14:21 -080011138 rc = lpfc_create_wq_cq(phba, qp[0].hba_eq,
11139 phba->sli4_hba.els_cq,
11140 phba->sli4_hba.els_wq,
11141 NULL, 0, LPFC_ELS);
James Smartda0436e2009-05-22 14:51:39 -040011142 if (rc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070011143 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartcdb42be2019-01-28 11:14:21 -080011144 "0525 Failed setup of ELS WQ/CQ: rc = 0x%x\n",
11145 (uint32_t)rc);
James Smart895427b2017-02-12 13:52:30 -080011146 goto out_destroy;
James Smartda0436e2009-05-22 14:51:39 -040011147 }
11148 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
11149 "2590 ELS WQ setup: wq-id=%d, parent cq-id=%d\n",
11150 phba->sli4_hba.els_wq->queue_id,
11151 phba->sli4_hba.els_cq->queue_id);
11152
James Smartcdb42be2019-01-28 11:14:21 -080011153 if (phba->cfg_enable_fc4_type & LPFC_ENABLE_NVME) {
James Smart895427b2017-02-12 13:52:30 -080011154 /* Set up NVME LS Complete Queue */
11155 if (!phba->sli4_hba.nvmels_cq || !phba->sli4_hba.nvmels_wq) {
Dick Kennedy372c1872020-06-30 14:50:00 -070011156 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart895427b2017-02-12 13:52:30 -080011157 "6091 LS %s not allocated\n",
11158 phba->sli4_hba.nvmels_cq ? "WQ" : "CQ");
11159 rc = -ENOMEM;
11160 goto out_destroy;
11161 }
James Smartcdb42be2019-01-28 11:14:21 -080011162 rc = lpfc_create_wq_cq(phba, qp[0].hba_eq,
11163 phba->sli4_hba.nvmels_cq,
11164 phba->sli4_hba.nvmels_wq,
11165 NULL, 0, LPFC_NVME_LS);
James Smart895427b2017-02-12 13:52:30 -080011166 if (rc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070011167 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartcdb42be2019-01-28 11:14:21 -080011168 "0526 Failed setup of NVVME LS WQ/CQ: "
11169 "rc = 0x%x\n", (uint32_t)rc);
James Smart895427b2017-02-12 13:52:30 -080011170 goto out_destroy;
11171 }
11172
11173 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
11174 "6096 ELS WQ setup: wq-id=%d, "
11175 "parent cq-id=%d\n",
11176 phba->sli4_hba.nvmels_wq->queue_id,
11177 phba->sli4_hba.nvmels_cq->queue_id);
11178 }
11179
James Smart2d7dbc42017-02-12 13:52:35 -080011180 /*
11181 * Create NVMET Receive Queue (RQ)
11182 */
11183 if (phba->nvmet_support) {
11184 if ((!phba->sli4_hba.nvmet_cqset) ||
11185 (!phba->sli4_hba.nvmet_mrq_hdr) ||
11186 (!phba->sli4_hba.nvmet_mrq_data)) {
Dick Kennedy372c1872020-06-30 14:50:00 -070011187 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart2d7dbc42017-02-12 13:52:35 -080011188 "6130 MRQ CQ Queues not "
11189 "allocated\n");
11190 rc = -ENOMEM;
11191 goto out_destroy;
11192 }
11193 if (phba->cfg_nvmet_mrq > 1) {
11194 rc = lpfc_mrq_create(phba,
11195 phba->sli4_hba.nvmet_mrq_hdr,
11196 phba->sli4_hba.nvmet_mrq_data,
11197 phba->sli4_hba.nvmet_cqset,
11198 LPFC_NVMET);
11199 if (rc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070011200 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart2d7dbc42017-02-12 13:52:35 -080011201 "6098 Failed setup of NVMET "
11202 "MRQ: rc = 0x%x\n",
11203 (uint32_t)rc);
11204 goto out_destroy;
11205 }
11206
11207 } else {
11208 rc = lpfc_rq_create(phba,
11209 phba->sli4_hba.nvmet_mrq_hdr[0],
11210 phba->sli4_hba.nvmet_mrq_data[0],
11211 phba->sli4_hba.nvmet_cqset[0],
11212 LPFC_NVMET);
11213 if (rc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070011214 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart2d7dbc42017-02-12 13:52:35 -080011215 "6057 Failed setup of NVMET "
11216 "Receive Queue: rc = 0x%x\n",
11217 (uint32_t)rc);
11218 goto out_destroy;
11219 }
11220
11221 lpfc_printf_log(
11222 phba, KERN_INFO, LOG_INIT,
11223 "6099 NVMET RQ setup: hdr-rq-id=%d, "
11224 "dat-rq-id=%d parent cq-id=%d\n",
11225 phba->sli4_hba.nvmet_mrq_hdr[0]->queue_id,
11226 phba->sli4_hba.nvmet_mrq_data[0]->queue_id,
11227 phba->sli4_hba.nvmet_cqset[0]->queue_id);
11228
11229 }
11230 }
11231
James Smartda0436e2009-05-22 14:51:39 -040011232 if (!phba->sli4_hba.hdr_rq || !phba->sli4_hba.dat_rq) {
Dick Kennedy372c1872020-06-30 14:50:00 -070011233 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartda0436e2009-05-22 14:51:39 -040011234 "0540 Receive Queue not allocated\n");
James Smart1b511972011-12-13 13:23:09 -050011235 rc = -ENOMEM;
James Smart895427b2017-02-12 13:52:30 -080011236 goto out_destroy;
James Smartda0436e2009-05-22 14:51:39 -040011237 }
James Smart73d91e52011-10-10 21:32:10 -040011238
James Smartda0436e2009-05-22 14:51:39 -040011239 rc = lpfc_rq_create(phba, phba->sli4_hba.hdr_rq, phba->sli4_hba.dat_rq,
James Smart4d9ab992009-10-02 15:16:39 -040011240 phba->sli4_hba.els_cq, LPFC_USOL);
James Smartda0436e2009-05-22 14:51:39 -040011241 if (rc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070011242 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartda0436e2009-05-22 14:51:39 -040011243 "0541 Failed setup of Receive Queue: "
James Smarta2fc4aef2014-09-03 12:57:55 -040011244 "rc = 0x%x\n", (uint32_t)rc);
James Smart895427b2017-02-12 13:52:30 -080011245 goto out_destroy;
James Smartda0436e2009-05-22 14:51:39 -040011246 }
James Smart73d91e52011-10-10 21:32:10 -040011247
James Smartda0436e2009-05-22 14:51:39 -040011248 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
11249 "2592 USL RQ setup: hdr-rq-id=%d, dat-rq-id=%d "
11250 "parent cq-id=%d\n",
11251 phba->sli4_hba.hdr_rq->queue_id,
11252 phba->sli4_hba.dat_rq->queue_id,
James Smart4d9ab992009-10-02 15:16:39 -040011253 phba->sli4_hba.els_cq->queue_id);
James Smart1ba981f2014-02-20 09:56:45 -050011254
James Smartcb733e32019-01-28 11:14:32 -080011255 if (phba->cfg_fcp_imax)
11256 usdelay = LPFC_SEC_TO_USEC / phba->cfg_fcp_imax;
11257 else
11258 usdelay = 0;
11259
James Smart6a828b02019-01-28 11:14:31 -080011260 for (qidx = 0; qidx < phba->cfg_irq_chann;
James Smartcdb42be2019-01-28 11:14:21 -080011261 qidx += LPFC_MAX_EQ_DELAY_EQID_CNT)
James Smart0cf07f842017-06-01 21:07:10 -070011262 lpfc_modify_hba_eq_delay(phba, qidx, LPFC_MAX_EQ_DELAY_EQID_CNT,
James Smartcb733e32019-01-28 11:14:32 -080011263 usdelay);
James Smart43140ca2017-03-04 09:30:34 -080011264
James Smart6a828b02019-01-28 11:14:31 -080011265 if (phba->sli4_hba.cq_max) {
11266 kfree(phba->sli4_hba.cq_lookup);
11267 phba->sli4_hba.cq_lookup = kcalloc((phba->sli4_hba.cq_max + 1),
11268 sizeof(struct lpfc_queue *), GFP_KERNEL);
11269 if (!phba->sli4_hba.cq_lookup) {
Dick Kennedy372c1872020-06-30 14:50:00 -070011270 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart6a828b02019-01-28 11:14:31 -080011271 "0549 Failed setup of CQ Lookup table: "
11272 "size 0x%x\n", phba->sli4_hba.cq_max);
Dan Carpenterfad28e32019-02-11 21:43:00 +030011273 rc = -ENOMEM;
James Smart1ba981f2014-02-20 09:56:45 -050011274 goto out_destroy;
James Smart895427b2017-02-12 13:52:30 -080011275 }
James Smart6a828b02019-01-28 11:14:31 -080011276 lpfc_setup_cq_lookup(phba);
James Smart1ba981f2014-02-20 09:56:45 -050011277 }
James Smartda0436e2009-05-22 14:51:39 -040011278 return 0;
11279
James Smart895427b2017-02-12 13:52:30 -080011280out_destroy:
11281 lpfc_sli4_queue_unset(phba);
James Smartda0436e2009-05-22 14:51:39 -040011282out_error:
11283 return rc;
11284}
11285
11286/**
11287 * lpfc_sli4_queue_unset - Unset all the SLI4 queues
11288 * @phba: pointer to lpfc hba data structure.
11289 *
11290 * This routine is invoked to unset all the SLI4 queues with the FCoE HBA
11291 * operation.
11292 *
11293 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -020011294 * 0 - successful
Lucas De Marchi25985ed2011-03-30 22:57:33 -030011295 * -ENOMEM - No available memory
James Smartd439d282010-09-29 11:18:45 -040011296 * -EIO - The mailbox failed to complete successfully.
James Smartda0436e2009-05-22 14:51:39 -040011297 **/
11298void
11299lpfc_sli4_queue_unset(struct lpfc_hba *phba)
11300{
James Smartcdb42be2019-01-28 11:14:21 -080011301 struct lpfc_sli4_hdw_queue *qp;
James Smart657add42019-05-21 17:49:06 -070011302 struct lpfc_queue *eq;
James Smart895427b2017-02-12 13:52:30 -080011303 int qidx;
James Smartda0436e2009-05-22 14:51:39 -040011304
11305 /* Unset mailbox command work queue */
James Smart895427b2017-02-12 13:52:30 -080011306 if (phba->sli4_hba.mbx_wq)
11307 lpfc_mq_destroy(phba, phba->sli4_hba.mbx_wq);
11308
11309 /* Unset NVME LS work queue */
11310 if (phba->sli4_hba.nvmels_wq)
11311 lpfc_wq_destroy(phba, phba->sli4_hba.nvmels_wq);
11312
James Smartda0436e2009-05-22 14:51:39 -040011313 /* Unset ELS work queue */
Colin Ian King019c0d62017-05-06 23:13:55 +010011314 if (phba->sli4_hba.els_wq)
James Smart895427b2017-02-12 13:52:30 -080011315 lpfc_wq_destroy(phba, phba->sli4_hba.els_wq);
11316
James Smartda0436e2009-05-22 14:51:39 -040011317 /* Unset unsolicited receive queue */
James Smart895427b2017-02-12 13:52:30 -080011318 if (phba->sli4_hba.hdr_rq)
11319 lpfc_rq_destroy(phba, phba->sli4_hba.hdr_rq,
11320 phba->sli4_hba.dat_rq);
11321
James Smartda0436e2009-05-22 14:51:39 -040011322 /* Unset mailbox command complete queue */
James Smart895427b2017-02-12 13:52:30 -080011323 if (phba->sli4_hba.mbx_cq)
11324 lpfc_cq_destroy(phba, phba->sli4_hba.mbx_cq);
11325
James Smartda0436e2009-05-22 14:51:39 -040011326 /* Unset ELS complete queue */
James Smart895427b2017-02-12 13:52:30 -080011327 if (phba->sli4_hba.els_cq)
11328 lpfc_cq_destroy(phba, phba->sli4_hba.els_cq);
11329
11330 /* Unset NVME LS complete queue */
11331 if (phba->sli4_hba.nvmels_cq)
11332 lpfc_cq_destroy(phba, phba->sli4_hba.nvmels_cq);
11333
James Smartbcb24f62017-11-20 16:00:36 -080011334 if (phba->nvmet_support) {
11335 /* Unset NVMET MRQ queue */
11336 if (phba->sli4_hba.nvmet_mrq_hdr) {
11337 for (qidx = 0; qidx < phba->cfg_nvmet_mrq; qidx++)
11338 lpfc_rq_destroy(
11339 phba,
James Smart2d7dbc42017-02-12 13:52:35 -080011340 phba->sli4_hba.nvmet_mrq_hdr[qidx],
11341 phba->sli4_hba.nvmet_mrq_data[qidx]);
James Smartbcb24f62017-11-20 16:00:36 -080011342 }
James Smart2d7dbc42017-02-12 13:52:35 -080011343
James Smartbcb24f62017-11-20 16:00:36 -080011344 /* Unset NVMET CQ Set complete queue */
11345 if (phba->sli4_hba.nvmet_cqset) {
11346 for (qidx = 0; qidx < phba->cfg_nvmet_mrq; qidx++)
11347 lpfc_cq_destroy(
11348 phba, phba->sli4_hba.nvmet_cqset[qidx]);
11349 }
James Smart2d7dbc42017-02-12 13:52:35 -080011350 }
11351
James Smartcdb42be2019-01-28 11:14:21 -080011352 /* Unset fast-path SLI4 queues */
11353 if (phba->sli4_hba.hdwq) {
James Smart657add42019-05-21 17:49:06 -070011354 /* Loop thru all Hardware Queues */
James Smartcdb42be2019-01-28 11:14:21 -080011355 for (qidx = 0; qidx < phba->cfg_hdw_queue; qidx++) {
James Smart657add42019-05-21 17:49:06 -070011356 /* Destroy the CQ/WQ corresponding to Hardware Queue */
James Smartcdb42be2019-01-28 11:14:21 -080011357 qp = &phba->sli4_hba.hdwq[qidx];
James Smartc00f62e2019-08-14 16:57:11 -070011358 lpfc_wq_destroy(phba, qp->io_wq);
11359 lpfc_cq_destroy(phba, qp->io_cq);
James Smart657add42019-05-21 17:49:06 -070011360 }
11361 /* Loop thru all IRQ vectors */
11362 for (qidx = 0; qidx < phba->cfg_irq_chann; qidx++) {
11363 /* Destroy the EQ corresponding to the IRQ vector */
11364 eq = phba->sli4_hba.hba_eq_hdl[qidx].eq;
11365 lpfc_eq_destroy(phba, eq);
James Smartcdb42be2019-01-28 11:14:21 -080011366 }
11367 }
James Smart895427b2017-02-12 13:52:30 -080011368
James Smart6a828b02019-01-28 11:14:31 -080011369 kfree(phba->sli4_hba.cq_lookup);
11370 phba->sli4_hba.cq_lookup = NULL;
11371 phba->sli4_hba.cq_max = 0;
James Smartda0436e2009-05-22 14:51:39 -040011372}
11373
11374/**
11375 * lpfc_sli4_cq_event_pool_create - Create completion-queue event free pool
11376 * @phba: pointer to lpfc hba data structure.
11377 *
11378 * This routine is invoked to allocate and set up a pool of completion queue
11379 * events. The body of the completion queue event is a completion queue entry
11380 * CQE. For now, this pool is used for the interrupt service routine to queue
11381 * the following HBA completion queue events for the worker thread to process:
11382 * - Mailbox asynchronous events
11383 * - Receive queue completion unsolicited events
11384 * Later, this can be used for all the slow-path events.
11385 *
11386 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -020011387 * 0 - successful
Lucas De Marchi25985ed2011-03-30 22:57:33 -030011388 * -ENOMEM - No available memory
James Smartda0436e2009-05-22 14:51:39 -040011389 **/
11390static int
11391lpfc_sli4_cq_event_pool_create(struct lpfc_hba *phba)
11392{
11393 struct lpfc_cq_event *cq_event;
11394 int i;
11395
11396 for (i = 0; i < (4 * phba->sli4_hba.cq_ecount); i++) {
11397 cq_event = kmalloc(sizeof(struct lpfc_cq_event), GFP_KERNEL);
11398 if (!cq_event)
11399 goto out_pool_create_fail;
11400 list_add_tail(&cq_event->list,
11401 &phba->sli4_hba.sp_cqe_event_pool);
11402 }
11403 return 0;
11404
11405out_pool_create_fail:
11406 lpfc_sli4_cq_event_pool_destroy(phba);
11407 return -ENOMEM;
11408}
11409
11410/**
11411 * lpfc_sli4_cq_event_pool_destroy - Free completion-queue event free pool
11412 * @phba: pointer to lpfc hba data structure.
11413 *
11414 * This routine is invoked to free the pool of completion queue events at
11415 * driver unload time. Note that, it is the responsibility of the driver
11416 * cleanup routine to free all the outstanding completion-queue events
11417 * allocated from this pool back into the pool before invoking this routine
11418 * to destroy the pool.
11419 **/
11420static void
11421lpfc_sli4_cq_event_pool_destroy(struct lpfc_hba *phba)
11422{
11423 struct lpfc_cq_event *cq_event, *next_cq_event;
11424
11425 list_for_each_entry_safe(cq_event, next_cq_event,
11426 &phba->sli4_hba.sp_cqe_event_pool, list) {
11427 list_del(&cq_event->list);
11428 kfree(cq_event);
11429 }
11430}
11431
11432/**
11433 * __lpfc_sli4_cq_event_alloc - Allocate a completion-queue event from free pool
11434 * @phba: pointer to lpfc hba data structure.
11435 *
11436 * This routine is the lock free version of the API invoked to allocate a
11437 * completion-queue event from the free pool.
11438 *
11439 * Return: Pointer to the newly allocated completion-queue event if successful
11440 * NULL otherwise.
11441 **/
11442struct lpfc_cq_event *
11443__lpfc_sli4_cq_event_alloc(struct lpfc_hba *phba)
11444{
11445 struct lpfc_cq_event *cq_event = NULL;
11446
11447 list_remove_head(&phba->sli4_hba.sp_cqe_event_pool, cq_event,
11448 struct lpfc_cq_event, list);
11449 return cq_event;
11450}
11451
11452/**
11453 * lpfc_sli4_cq_event_alloc - Allocate a completion-queue event from free pool
11454 * @phba: pointer to lpfc hba data structure.
11455 *
11456 * This routine is the lock version of the API invoked to allocate a
11457 * completion-queue event from the free pool.
11458 *
11459 * Return: Pointer to the newly allocated completion-queue event if successful
11460 * NULL otherwise.
11461 **/
11462struct lpfc_cq_event *
11463lpfc_sli4_cq_event_alloc(struct lpfc_hba *phba)
11464{
11465 struct lpfc_cq_event *cq_event;
11466 unsigned long iflags;
11467
11468 spin_lock_irqsave(&phba->hbalock, iflags);
11469 cq_event = __lpfc_sli4_cq_event_alloc(phba);
11470 spin_unlock_irqrestore(&phba->hbalock, iflags);
11471 return cq_event;
11472}
11473
11474/**
11475 * __lpfc_sli4_cq_event_release - Release a completion-queue event to free pool
11476 * @phba: pointer to lpfc hba data structure.
11477 * @cq_event: pointer to the completion queue event to be freed.
11478 *
11479 * This routine is the lock free version of the API invoked to release a
11480 * completion-queue event back into the free pool.
11481 **/
11482void
11483__lpfc_sli4_cq_event_release(struct lpfc_hba *phba,
11484 struct lpfc_cq_event *cq_event)
11485{
11486 list_add_tail(&cq_event->list, &phba->sli4_hba.sp_cqe_event_pool);
11487}
11488
11489/**
11490 * lpfc_sli4_cq_event_release - Release a completion-queue event to free pool
11491 * @phba: pointer to lpfc hba data structure.
11492 * @cq_event: pointer to the completion queue event to be freed.
11493 *
11494 * This routine is the lock version of the API invoked to release a
11495 * completion-queue event back into the free pool.
11496 **/
11497void
11498lpfc_sli4_cq_event_release(struct lpfc_hba *phba,
11499 struct lpfc_cq_event *cq_event)
11500{
11501 unsigned long iflags;
11502 spin_lock_irqsave(&phba->hbalock, iflags);
11503 __lpfc_sli4_cq_event_release(phba, cq_event);
11504 spin_unlock_irqrestore(&phba->hbalock, iflags);
11505}
11506
11507/**
11508 * lpfc_sli4_cq_event_release_all - Release all cq events to the free pool
11509 * @phba: pointer to lpfc hba data structure.
11510 *
11511 * This routine is to free all the pending completion-queue events to the
11512 * back into the free pool for device reset.
11513 **/
11514static void
11515lpfc_sli4_cq_event_release_all(struct lpfc_hba *phba)
11516{
James Smarte7dab162020-10-20 13:27:12 -070011517 LIST_HEAD(cq_event_list);
11518 struct lpfc_cq_event *cq_event;
James Smartda0436e2009-05-22 14:51:39 -040011519 unsigned long iflags;
11520
11521 /* Retrieve all the pending WCQEs from pending WCQE lists */
James Smartda0436e2009-05-22 14:51:39 -040011522
James Smarte7dab162020-10-20 13:27:12 -070011523 /* Pending ELS XRI abort events */
11524 spin_lock_irqsave(&phba->sli4_hba.els_xri_abrt_list_lock, iflags);
11525 list_splice_init(&phba->sli4_hba.sp_els_xri_aborted_work_queue,
11526 &cq_event_list);
11527 spin_unlock_irqrestore(&phba->sli4_hba.els_xri_abrt_list_lock, iflags);
11528
11529 /* Pending asynnc events */
11530 spin_lock_irqsave(&phba->sli4_hba.asynce_list_lock, iflags);
11531 list_splice_init(&phba->sli4_hba.sp_asynce_work_queue,
11532 &cq_event_list);
11533 spin_unlock_irqrestore(&phba->sli4_hba.asynce_list_lock, iflags);
11534
11535 while (!list_empty(&cq_event_list)) {
11536 list_remove_head(&cq_event_list, cq_event,
11537 struct lpfc_cq_event, list);
11538 lpfc_sli4_cq_event_release(phba, cq_event);
James Smartda0436e2009-05-22 14:51:39 -040011539 }
11540}
11541
11542/**
11543 * lpfc_pci_function_reset - Reset pci function.
11544 * @phba: pointer to lpfc hba data structure.
11545 *
11546 * This routine is invoked to request a PCI function reset. It will destroys
11547 * all resources assigned to the PCI function which originates this request.
11548 *
11549 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -020011550 * 0 - successful
Lucas De Marchi25985ed2011-03-30 22:57:33 -030011551 * -ENOMEM - No available memory
James Smartd439d282010-09-29 11:18:45 -040011552 * -EIO - The mailbox failed to complete successfully.
James Smartda0436e2009-05-22 14:51:39 -040011553 **/
11554int
11555lpfc_pci_function_reset(struct lpfc_hba *phba)
11556{
11557 LPFC_MBOXQ_t *mboxq;
James Smart2fcee4b2010-12-15 17:57:46 -050011558 uint32_t rc = 0, if_type;
James Smartda0436e2009-05-22 14:51:39 -040011559 uint32_t shdr_status, shdr_add_status;
James Smart2f6fa2c2014-09-03 12:57:08 -040011560 uint32_t rdy_chk;
11561 uint32_t port_reset = 0;
James Smartda0436e2009-05-22 14:51:39 -040011562 union lpfc_sli4_cfg_shdr *shdr;
James Smart2fcee4b2010-12-15 17:57:46 -050011563 struct lpfc_register reg_data;
James Smart2b81f942012-03-01 22:37:18 -050011564 uint16_t devid;
James Smartda0436e2009-05-22 14:51:39 -040011565
James Smart2fcee4b2010-12-15 17:57:46 -050011566 if_type = bf_get(lpfc_sli_intf_if_type, &phba->sli4_hba.sli_intf);
11567 switch (if_type) {
11568 case LPFC_SLI_INTF_IF_TYPE_0:
11569 mboxq = (LPFC_MBOXQ_t *) mempool_alloc(phba->mbox_mem_pool,
11570 GFP_KERNEL);
11571 if (!mboxq) {
Dick Kennedy372c1872020-06-30 14:50:00 -070011572 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart2fcee4b2010-12-15 17:57:46 -050011573 "0494 Unable to allocate memory for "
11574 "issuing SLI_FUNCTION_RESET mailbox "
11575 "command\n");
11576 return -ENOMEM;
11577 }
11578
11579 /* Setup PCI function reset mailbox-ioctl command */
11580 lpfc_sli4_config(phba, mboxq, LPFC_MBOX_SUBSYSTEM_COMMON,
11581 LPFC_MBOX_OPCODE_FUNCTION_RESET, 0,
11582 LPFC_SLI4_MBX_EMBED);
11583 rc = lpfc_sli_issue_mbox(phba, mboxq, MBX_POLL);
11584 shdr = (union lpfc_sli4_cfg_shdr *)
11585 &mboxq->u.mqe.un.sli4_config.header.cfg_shdr;
11586 shdr_status = bf_get(lpfc_mbox_hdr_status, &shdr->response);
11587 shdr_add_status = bf_get(lpfc_mbox_hdr_add_status,
11588 &shdr->response);
James Smart304ee432021-04-11 18:31:17 -070011589 mempool_free(mboxq, phba->mbox_mem_pool);
James Smart2fcee4b2010-12-15 17:57:46 -050011590 if (shdr_status || shdr_add_status || rc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070011591 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart2fcee4b2010-12-15 17:57:46 -050011592 "0495 SLI_FUNCTION_RESET mailbox "
11593 "failed with status x%x add_status x%x,"
11594 " mbx status x%x\n",
11595 shdr_status, shdr_add_status, rc);
11596 rc = -ENXIO;
11597 }
11598 break;
11599 case LPFC_SLI_INTF_IF_TYPE_2:
James Smart27d6ac02018-02-22 08:18:42 -080011600 case LPFC_SLI_INTF_IF_TYPE_6:
James Smart2f6fa2c2014-09-03 12:57:08 -040011601wait:
11602 /*
11603 * Poll the Port Status Register and wait for RDY for
11604 * up to 30 seconds. If the port doesn't respond, treat
11605 * it as an error.
11606 */
James Smart77d093f2015-04-07 15:07:08 -040011607 for (rdy_chk = 0; rdy_chk < 1500; rdy_chk++) {
James Smart2f6fa2c2014-09-03 12:57:08 -040011608 if (lpfc_readl(phba->sli4_hba.u.if_type2.
11609 STATUSregaddr, &reg_data.word0)) {
11610 rc = -ENODEV;
11611 goto out;
11612 }
11613 if (bf_get(lpfc_sliport_status_rdy, &reg_data))
11614 break;
11615 msleep(20);
11616 }
11617
11618 if (!bf_get(lpfc_sliport_status_rdy, &reg_data)) {
11619 phba->work_status[0] = readl(
11620 phba->sli4_hba.u.if_type2.ERR1regaddr);
11621 phba->work_status[1] = readl(
11622 phba->sli4_hba.u.if_type2.ERR2regaddr);
Dick Kennedy372c1872020-06-30 14:50:00 -070011623 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart2f6fa2c2014-09-03 12:57:08 -040011624 "2890 Port not ready, port status reg "
11625 "0x%x error 1=0x%x, error 2=0x%x\n",
11626 reg_data.word0,
11627 phba->work_status[0],
11628 phba->work_status[1]);
11629 rc = -ENODEV;
11630 goto out;
11631 }
11632
James Smarta5b141a2021-09-27 11:35:18 -070011633 if (bf_get(lpfc_sliport_status_pldv, &reg_data))
11634 lpfc_pldv_detect = true;
11635
James Smart2f6fa2c2014-09-03 12:57:08 -040011636 if (!port_reset) {
11637 /*
11638 * Reset the port now
11639 */
James Smart2fcee4b2010-12-15 17:57:46 -050011640 reg_data.word0 = 0;
11641 bf_set(lpfc_sliport_ctrl_end, &reg_data,
11642 LPFC_SLIPORT_LITTLE_ENDIAN);
11643 bf_set(lpfc_sliport_ctrl_ip, &reg_data,
11644 LPFC_SLIPORT_INIT_PORT);
11645 writel(reg_data.word0, phba->sli4_hba.u.if_type2.
11646 CTRLregaddr);
James Smart8fcb8ac2012-03-01 22:35:58 -050011647 /* flush */
James Smart2b81f942012-03-01 22:37:18 -050011648 pci_read_config_word(phba->pcidev,
11649 PCI_DEVICE_ID, &devid);
James Smart2fcee4b2010-12-15 17:57:46 -050011650
James Smart2f6fa2c2014-09-03 12:57:08 -040011651 port_reset = 1;
11652 msleep(20);
11653 goto wait;
11654 } else if (bf_get(lpfc_sliport_status_rn, &reg_data)) {
11655 rc = -ENODEV;
11656 goto out;
James Smart2fcee4b2010-12-15 17:57:46 -050011657 }
11658 break;
James Smart2f6fa2c2014-09-03 12:57:08 -040011659
James Smart2fcee4b2010-12-15 17:57:46 -050011660 case LPFC_SLI_INTF_IF_TYPE_1:
11661 default:
11662 break;
James Smartda0436e2009-05-22 14:51:39 -040011663 }
11664
James Smart73d91e52011-10-10 21:32:10 -040011665out:
James Smart2fcee4b2010-12-15 17:57:46 -050011666 /* Catch the not-ready port failure after a port reset. */
James Smart2f6fa2c2014-09-03 12:57:08 -040011667 if (rc) {
Dick Kennedy372c1872020-06-30 14:50:00 -070011668 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart229adb02013-04-17 20:16:51 -040011669 "3317 HBA not functional: IP Reset Failed "
James Smart2f6fa2c2014-09-03 12:57:08 -040011670 "try: echo fw_reset > board_mode\n");
James Smart2fcee4b2010-12-15 17:57:46 -050011671 rc = -ENODEV;
James Smart229adb02013-04-17 20:16:51 -040011672 }
James Smart2fcee4b2010-12-15 17:57:46 -050011673
James Smartda0436e2009-05-22 14:51:39 -040011674 return rc;
11675}
11676
11677/**
James Smartda0436e2009-05-22 14:51:39 -040011678 * lpfc_sli4_pci_mem_setup - Setup SLI4 HBA PCI memory space.
11679 * @phba: pointer to lpfc hba data structure.
11680 *
11681 * This routine is invoked to set up the PCI device memory space for device
11682 * with SLI-4 interface spec.
11683 *
11684 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -020011685 * 0 - successful
James Smartda0436e2009-05-22 14:51:39 -040011686 * other values - error
11687 **/
11688static int
11689lpfc_sli4_pci_mem_setup(struct lpfc_hba *phba)
11690{
Christoph Hellwigf30e1bf2018-10-18 15:10:21 +020011691 struct pci_dev *pdev = phba->pcidev;
James Smartda0436e2009-05-22 14:51:39 -040011692 unsigned long bar0map_len, bar1map_len, bar2map_len;
Dan Carpenter3a487ff2019-03-07 08:33:44 +030011693 int error;
James Smart2fcee4b2010-12-15 17:57:46 -050011694 uint32_t if_type;
James Smartda0436e2009-05-22 14:51:39 -040011695
Christoph Hellwigf30e1bf2018-10-18 15:10:21 +020011696 if (!pdev)
Hannes Reinecke56de8352019-02-18 08:34:19 +010011697 return -ENODEV;
James Smartda0436e2009-05-22 14:51:39 -040011698
11699 /* Set the device DMA mask size */
Hannes Reinecke56de8352019-02-18 08:34:19 +010011700 error = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(64));
11701 if (error)
11702 error = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(32));
11703 if (error)
Christoph Hellwigf30e1bf2018-10-18 15:10:21 +020011704 return error;
James Smartda0436e2009-05-22 14:51:39 -040011705
James Smart2fcee4b2010-12-15 17:57:46 -050011706 /*
11707 * The BARs and register set definitions and offset locations are
11708 * dependent on the if_type.
11709 */
11710 if (pci_read_config_dword(pdev, LPFC_SLI_INTF,
11711 &phba->sli4_hba.sli_intf.word0)) {
Dan Carpenter3a487ff2019-03-07 08:33:44 +030011712 return -ENODEV;
James Smart2fcee4b2010-12-15 17:57:46 -050011713 }
11714
11715 /* There is no SLI3 failback for SLI4 devices. */
11716 if (bf_get(lpfc_sli_intf_valid, &phba->sli4_hba.sli_intf) !=
11717 LPFC_SLI_INTF_VALID) {
James Smarta5160742021-10-20 14:14:10 -070011718 lpfc_printf_log(phba, KERN_ERR, LOG_INIT,
James Smart2fcee4b2010-12-15 17:57:46 -050011719 "2894 SLI_INTF reg contents invalid "
11720 "sli_intf reg 0x%x\n",
11721 phba->sli4_hba.sli_intf.word0);
Dan Carpenter3a487ff2019-03-07 08:33:44 +030011722 return -ENODEV;
James Smart2fcee4b2010-12-15 17:57:46 -050011723 }
11724
11725 if_type = bf_get(lpfc_sli_intf_if_type, &phba->sli4_hba.sli_intf);
11726 /*
11727 * Get the bus address of SLI4 device Bar regions and the
11728 * number of bytes required by each mapping. The mapping of the
11729 * particular PCI BARs regions is dependent on the type of
11730 * SLI4 device.
James Smartda0436e2009-05-22 14:51:39 -040011731 */
James Smartf5ca6f22013-09-06 12:21:09 -040011732 if (pci_resource_start(pdev, PCI_64BIT_BAR0)) {
11733 phba->pci_bar0_map = pci_resource_start(pdev, PCI_64BIT_BAR0);
11734 bar0map_len = pci_resource_len(pdev, PCI_64BIT_BAR0);
James Smart2fcee4b2010-12-15 17:57:46 -050011735
11736 /*
11737 * Map SLI4 PCI Config Space Register base to a kernel virtual
11738 * addr
11739 */
11740 phba->sli4_hba.conf_regs_memmap_p =
11741 ioremap(phba->pci_bar0_map, bar0map_len);
11742 if (!phba->sli4_hba.conf_regs_memmap_p) {
11743 dev_printk(KERN_ERR, &pdev->dev,
11744 "ioremap failed for SLI4 PCI config "
11745 "registers.\n");
Dan Carpenter3a487ff2019-03-07 08:33:44 +030011746 return -ENODEV;
James Smart2fcee4b2010-12-15 17:57:46 -050011747 }
James Smartf5ca6f22013-09-06 12:21:09 -040011748 phba->pci_bar0_memmap_p = phba->sli4_hba.conf_regs_memmap_p;
James Smart2fcee4b2010-12-15 17:57:46 -050011749 /* Set up BAR0 PCI config space register memory map */
11750 lpfc_sli4_bar0_register_memmap(phba, if_type);
James Smart1dfb5a42010-02-12 14:40:50 -050011751 } else {
11752 phba->pci_bar0_map = pci_resource_start(pdev, 1);
11753 bar0map_len = pci_resource_len(pdev, 1);
James Smart27d6ac02018-02-22 08:18:42 -080011754 if (if_type >= LPFC_SLI_INTF_IF_TYPE_2) {
James Smart2fcee4b2010-12-15 17:57:46 -050011755 dev_printk(KERN_ERR, &pdev->dev,
11756 "FATAL - No BAR0 mapping for SLI4, if_type 2\n");
Dan Carpenter3a487ff2019-03-07 08:33:44 +030011757 return -ENODEV;
James Smart2fcee4b2010-12-15 17:57:46 -050011758 }
11759 phba->sli4_hba.conf_regs_memmap_p =
James Smartda0436e2009-05-22 14:51:39 -040011760 ioremap(phba->pci_bar0_map, bar0map_len);
James Smart2fcee4b2010-12-15 17:57:46 -050011761 if (!phba->sli4_hba.conf_regs_memmap_p) {
11762 dev_printk(KERN_ERR, &pdev->dev,
11763 "ioremap failed for SLI4 PCI config "
11764 "registers.\n");
Dan Carpenter3a487ff2019-03-07 08:33:44 +030011765 return -ENODEV;
James Smart2fcee4b2010-12-15 17:57:46 -050011766 }
11767 lpfc_sli4_bar0_register_memmap(phba, if_type);
James Smartda0436e2009-05-22 14:51:39 -040011768 }
11769
James Smarte4b97942017-11-20 16:00:31 -080011770 if (if_type == LPFC_SLI_INTF_IF_TYPE_0) {
11771 if (pci_resource_start(pdev, PCI_64BIT_BAR2)) {
11772 /*
11773 * Map SLI4 if type 0 HBA Control Register base to a
11774 * kernel virtual address and setup the registers.
11775 */
11776 phba->pci_bar1_map = pci_resource_start(pdev,
11777 PCI_64BIT_BAR2);
11778 bar1map_len = pci_resource_len(pdev, PCI_64BIT_BAR2);
11779 phba->sli4_hba.ctrl_regs_memmap_p =
11780 ioremap(phba->pci_bar1_map,
11781 bar1map_len);
11782 if (!phba->sli4_hba.ctrl_regs_memmap_p) {
11783 dev_err(&pdev->dev,
11784 "ioremap failed for SLI4 HBA "
11785 "control registers.\n");
11786 error = -ENOMEM;
11787 goto out_iounmap_conf;
11788 }
11789 phba->pci_bar2_memmap_p =
11790 phba->sli4_hba.ctrl_regs_memmap_p;
James Smart27d6ac02018-02-22 08:18:42 -080011791 lpfc_sli4_bar1_register_memmap(phba, if_type);
James Smarte4b97942017-11-20 16:00:31 -080011792 } else {
11793 error = -ENOMEM;
James Smart2fcee4b2010-12-15 17:57:46 -050011794 goto out_iounmap_conf;
11795 }
James Smartda0436e2009-05-22 14:51:39 -040011796 }
11797
James Smart27d6ac02018-02-22 08:18:42 -080011798 if ((if_type == LPFC_SLI_INTF_IF_TYPE_6) &&
11799 (pci_resource_start(pdev, PCI_64BIT_BAR2))) {
11800 /*
11801 * Map SLI4 if type 6 HBA Doorbell Register base to a kernel
11802 * virtual address and setup the registers.
11803 */
11804 phba->pci_bar1_map = pci_resource_start(pdev, PCI_64BIT_BAR2);
11805 bar1map_len = pci_resource_len(pdev, PCI_64BIT_BAR2);
11806 phba->sli4_hba.drbl_regs_memmap_p =
11807 ioremap(phba->pci_bar1_map, bar1map_len);
11808 if (!phba->sli4_hba.drbl_regs_memmap_p) {
11809 dev_err(&pdev->dev,
11810 "ioremap failed for SLI4 HBA doorbell registers.\n");
Dan Carpenter3a487ff2019-03-07 08:33:44 +030011811 error = -ENOMEM;
James Smart27d6ac02018-02-22 08:18:42 -080011812 goto out_iounmap_conf;
11813 }
11814 phba->pci_bar2_memmap_p = phba->sli4_hba.drbl_regs_memmap_p;
11815 lpfc_sli4_bar1_register_memmap(phba, if_type);
11816 }
11817
James Smarte4b97942017-11-20 16:00:31 -080011818 if (if_type == LPFC_SLI_INTF_IF_TYPE_0) {
11819 if (pci_resource_start(pdev, PCI_64BIT_BAR4)) {
11820 /*
11821 * Map SLI4 if type 0 HBA Doorbell Register base to
11822 * a kernel virtual address and setup the registers.
11823 */
11824 phba->pci_bar2_map = pci_resource_start(pdev,
11825 PCI_64BIT_BAR4);
11826 bar2map_len = pci_resource_len(pdev, PCI_64BIT_BAR4);
11827 phba->sli4_hba.drbl_regs_memmap_p =
11828 ioremap(phba->pci_bar2_map,
11829 bar2map_len);
11830 if (!phba->sli4_hba.drbl_regs_memmap_p) {
11831 dev_err(&pdev->dev,
11832 "ioremap failed for SLI4 HBA"
11833 " doorbell registers.\n");
11834 error = -ENOMEM;
11835 goto out_iounmap_ctrl;
11836 }
11837 phba->pci_bar4_memmap_p =
11838 phba->sli4_hba.drbl_regs_memmap_p;
11839 error = lpfc_sli4_bar2_register_memmap(phba, LPFC_VF0);
11840 if (error)
11841 goto out_iounmap_all;
11842 } else {
11843 error = -ENOMEM;
James Smart2fcee4b2010-12-15 17:57:46 -050011844 goto out_iounmap_all;
James Smarte4b97942017-11-20 16:00:31 -080011845 }
James Smartda0436e2009-05-22 14:51:39 -040011846 }
11847
James Smart1351e692018-02-22 08:18:43 -080011848 if (if_type == LPFC_SLI_INTF_IF_TYPE_6 &&
11849 pci_resource_start(pdev, PCI_64BIT_BAR4)) {
11850 /*
11851 * Map SLI4 if type 6 HBA DPP Register base to a kernel
11852 * virtual address and setup the registers.
11853 */
11854 phba->pci_bar2_map = pci_resource_start(pdev, PCI_64BIT_BAR4);
11855 bar2map_len = pci_resource_len(pdev, PCI_64BIT_BAR4);
11856 phba->sli4_hba.dpp_regs_memmap_p =
11857 ioremap(phba->pci_bar2_map, bar2map_len);
11858 if (!phba->sli4_hba.dpp_regs_memmap_p) {
11859 dev_err(&pdev->dev,
11860 "ioremap failed for SLI4 HBA dpp registers.\n");
Dan Carpenter3a487ff2019-03-07 08:33:44 +030011861 error = -ENOMEM;
James Smart1351e692018-02-22 08:18:43 -080011862 goto out_iounmap_ctrl;
11863 }
11864 phba->pci_bar4_memmap_p = phba->sli4_hba.dpp_regs_memmap_p;
11865 }
11866
James Smartb71413d2018-02-22 08:18:40 -080011867 /* Set up the EQ/CQ register handeling functions now */
James Smart27d6ac02018-02-22 08:18:42 -080011868 switch (if_type) {
11869 case LPFC_SLI_INTF_IF_TYPE_0:
11870 case LPFC_SLI_INTF_IF_TYPE_2:
James Smartb71413d2018-02-22 08:18:40 -080011871 phba->sli4_hba.sli4_eq_clr_intr = lpfc_sli4_eq_clr_intr;
James Smart32517fc2019-01-28 11:14:33 -080011872 phba->sli4_hba.sli4_write_eq_db = lpfc_sli4_write_eq_db;
11873 phba->sli4_hba.sli4_write_cq_db = lpfc_sli4_write_cq_db;
James Smart27d6ac02018-02-22 08:18:42 -080011874 break;
11875 case LPFC_SLI_INTF_IF_TYPE_6:
11876 phba->sli4_hba.sli4_eq_clr_intr = lpfc_sli4_if6_eq_clr_intr;
James Smart32517fc2019-01-28 11:14:33 -080011877 phba->sli4_hba.sli4_write_eq_db = lpfc_sli4_if6_write_eq_db;
11878 phba->sli4_hba.sli4_write_cq_db = lpfc_sli4_if6_write_cq_db;
James Smart27d6ac02018-02-22 08:18:42 -080011879 break;
11880 default:
11881 break;
James Smartb71413d2018-02-22 08:18:40 -080011882 }
11883
James Smartda0436e2009-05-22 14:51:39 -040011884 return 0;
11885
11886out_iounmap_all:
11887 iounmap(phba->sli4_hba.drbl_regs_memmap_p);
11888out_iounmap_ctrl:
11889 iounmap(phba->sli4_hba.ctrl_regs_memmap_p);
11890out_iounmap_conf:
11891 iounmap(phba->sli4_hba.conf_regs_memmap_p);
Dan Carpenter3a487ff2019-03-07 08:33:44 +030011892
James Smartda0436e2009-05-22 14:51:39 -040011893 return error;
11894}
11895
11896/**
11897 * lpfc_sli4_pci_mem_unset - Unset SLI4 HBA PCI memory space.
11898 * @phba: pointer to lpfc hba data structure.
11899 *
11900 * This routine is invoked to unset the PCI device memory space for device
11901 * with SLI-4 interface spec.
11902 **/
11903static void
11904lpfc_sli4_pci_mem_unset(struct lpfc_hba *phba)
11905{
James Smart2e90f4b2011-12-13 13:22:37 -050011906 uint32_t if_type;
11907 if_type = bf_get(lpfc_sli_intf_if_type, &phba->sli4_hba.sli_intf);
James Smartda0436e2009-05-22 14:51:39 -040011908
James Smart2e90f4b2011-12-13 13:22:37 -050011909 switch (if_type) {
11910 case LPFC_SLI_INTF_IF_TYPE_0:
11911 iounmap(phba->sli4_hba.drbl_regs_memmap_p);
11912 iounmap(phba->sli4_hba.ctrl_regs_memmap_p);
11913 iounmap(phba->sli4_hba.conf_regs_memmap_p);
11914 break;
11915 case LPFC_SLI_INTF_IF_TYPE_2:
11916 iounmap(phba->sli4_hba.conf_regs_memmap_p);
11917 break;
James Smart27d6ac02018-02-22 08:18:42 -080011918 case LPFC_SLI_INTF_IF_TYPE_6:
11919 iounmap(phba->sli4_hba.drbl_regs_memmap_p);
11920 iounmap(phba->sli4_hba.conf_regs_memmap_p);
James Smart0b439192019-12-18 15:58:05 -080011921 if (phba->sli4_hba.dpp_regs_memmap_p)
11922 iounmap(phba->sli4_hba.dpp_regs_memmap_p);
James Smart27d6ac02018-02-22 08:18:42 -080011923 break;
James Smart2e90f4b2011-12-13 13:22:37 -050011924 case LPFC_SLI_INTF_IF_TYPE_1:
11925 default:
11926 dev_printk(KERN_ERR, &phba->pcidev->dev,
11927 "FATAL - unsupported SLI4 interface type - %d\n",
11928 if_type);
11929 break;
11930 }
James Smartda0436e2009-05-22 14:51:39 -040011931}
11932
11933/**
James Smart3772a992009-05-22 14:50:54 -040011934 * lpfc_sli_enable_msix - Enable MSI-X interrupt mode on SLI-3 device
11935 * @phba: pointer to lpfc hba data structure.
11936 *
11937 * This routine is invoked to enable the MSI-X interrupt vectors to device
Christoph Hellwig45ffac12017-02-12 13:52:26 -080011938 * with SLI-3 interface specs.
James Smart3772a992009-05-22 14:50:54 -040011939 *
11940 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -020011941 * 0 - successful
James Smart3772a992009-05-22 14:50:54 -040011942 * other values - error
11943 **/
11944static int
11945lpfc_sli_enable_msix(struct lpfc_hba *phba)
11946{
Christoph Hellwig45ffac12017-02-12 13:52:26 -080011947 int rc;
James Smart3772a992009-05-22 14:50:54 -040011948 LPFC_MBOXQ_t *pmb;
11949
11950 /* Set up MSI-X multi-message vectors */
Christoph Hellwig45ffac12017-02-12 13:52:26 -080011951 rc = pci_alloc_irq_vectors(phba->pcidev,
11952 LPFC_MSIX_VECTORS, LPFC_MSIX_VECTORS, PCI_IRQ_MSIX);
11953 if (rc < 0) {
James Smart3772a992009-05-22 14:50:54 -040011954 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
11955 "0420 PCI enable MSI-X failed (%d)\n", rc);
Alexander Gordeev029165a2014-07-16 20:05:15 +020011956 goto vec_fail_out;
James Smart3772a992009-05-22 14:50:54 -040011957 }
Christoph Hellwig45ffac12017-02-12 13:52:26 -080011958
James Smart3772a992009-05-22 14:50:54 -040011959 /*
11960 * Assign MSI-X vectors to interrupt handlers
11961 */
11962
11963 /* vector-0 is associated to slow-path handler */
Christoph Hellwig45ffac12017-02-12 13:52:26 -080011964 rc = request_irq(pci_irq_vector(phba->pcidev, 0),
James Smarted243d32015-05-21 13:55:25 -040011965 &lpfc_sli_sp_intr_handler, 0,
James Smart3772a992009-05-22 14:50:54 -040011966 LPFC_SP_DRIVER_HANDLER_NAME, phba);
11967 if (rc) {
11968 lpfc_printf_log(phba, KERN_WARNING, LOG_INIT,
11969 "0421 MSI-X slow-path request_irq failed "
11970 "(%d)\n", rc);
11971 goto msi_fail_out;
11972 }
11973
11974 /* vector-1 is associated to fast-path handler */
Christoph Hellwig45ffac12017-02-12 13:52:26 -080011975 rc = request_irq(pci_irq_vector(phba->pcidev, 1),
James Smarted243d32015-05-21 13:55:25 -040011976 &lpfc_sli_fp_intr_handler, 0,
James Smart3772a992009-05-22 14:50:54 -040011977 LPFC_FP_DRIVER_HANDLER_NAME, phba);
11978
11979 if (rc) {
11980 lpfc_printf_log(phba, KERN_WARNING, LOG_INIT,
11981 "0429 MSI-X fast-path request_irq failed "
11982 "(%d)\n", rc);
11983 goto irq_fail_out;
11984 }
11985
11986 /*
11987 * Configure HBA MSI-X attention conditions to messages
11988 */
11989 pmb = (LPFC_MBOXQ_t *) mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL);
11990
11991 if (!pmb) {
11992 rc = -ENOMEM;
Dick Kennedy372c1872020-06-30 14:50:00 -070011993 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart3772a992009-05-22 14:50:54 -040011994 "0474 Unable to allocate memory for issuing "
11995 "MBOX_CONFIG_MSI command\n");
11996 goto mem_fail_out;
11997 }
11998 rc = lpfc_config_msi(phba, pmb);
11999 if (rc)
12000 goto mbx_fail_out;
12001 rc = lpfc_sli_issue_mbox(phba, pmb, MBX_POLL);
12002 if (rc != MBX_SUCCESS) {
12003 lpfc_printf_log(phba, KERN_WARNING, LOG_MBOX,
12004 "0351 Config MSI mailbox command failed, "
12005 "mbxCmd x%x, mbxStatus x%x\n",
12006 pmb->u.mb.mbxCommand, pmb->u.mb.mbxStatus);
12007 goto mbx_fail_out;
12008 }
12009
12010 /* Free memory allocated for mailbox command */
12011 mempool_free(pmb, phba->mbox_mem_pool);
12012 return rc;
12013
12014mbx_fail_out:
12015 /* Free memory allocated for mailbox command */
12016 mempool_free(pmb, phba->mbox_mem_pool);
12017
12018mem_fail_out:
12019 /* free the irq already requested */
Christoph Hellwig45ffac12017-02-12 13:52:26 -080012020 free_irq(pci_irq_vector(phba->pcidev, 1), phba);
James Smart3772a992009-05-22 14:50:54 -040012021
12022irq_fail_out:
12023 /* free the irq already requested */
Christoph Hellwig45ffac12017-02-12 13:52:26 -080012024 free_irq(pci_irq_vector(phba->pcidev, 0), phba);
James Smart3772a992009-05-22 14:50:54 -040012025
12026msi_fail_out:
12027 /* Unconfigure MSI-X capability structure */
Christoph Hellwig45ffac12017-02-12 13:52:26 -080012028 pci_free_irq_vectors(phba->pcidev);
Alexander Gordeev029165a2014-07-16 20:05:15 +020012029
12030vec_fail_out:
James Smart3772a992009-05-22 14:50:54 -040012031 return rc;
12032}
12033
12034/**
James Smart3772a992009-05-22 14:50:54 -040012035 * lpfc_sli_enable_msi - Enable MSI interrupt mode on SLI-3 device.
12036 * @phba: pointer to lpfc hba data structure.
12037 *
12038 * This routine is invoked to enable the MSI interrupt mode to device with
12039 * SLI-3 interface spec. The kernel function pci_enable_msi() is called to
12040 * enable the MSI vector. The device driver is responsible for calling the
12041 * request_irq() to register MSI vector with a interrupt the handler, which
12042 * is done in this function.
12043 *
12044 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -020012045 * 0 - successful
James Smart3772a992009-05-22 14:50:54 -040012046 * other values - error
12047 */
12048static int
12049lpfc_sli_enable_msi(struct lpfc_hba *phba)
12050{
12051 int rc;
12052
12053 rc = pci_enable_msi(phba->pcidev);
12054 if (!rc)
12055 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
12056 "0462 PCI enable MSI mode success.\n");
12057 else {
12058 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
12059 "0471 PCI enable MSI mode failed (%d)\n", rc);
12060 return rc;
12061 }
12062
12063 rc = request_irq(phba->pcidev->irq, lpfc_sli_intr_handler,
James Smarted243d32015-05-21 13:55:25 -040012064 0, LPFC_DRIVER_NAME, phba);
James Smart3772a992009-05-22 14:50:54 -040012065 if (rc) {
12066 pci_disable_msi(phba->pcidev);
12067 lpfc_printf_log(phba, KERN_WARNING, LOG_INIT,
12068 "0478 MSI request_irq failed (%d)\n", rc);
12069 }
12070 return rc;
12071}
12072
12073/**
James Smart3772a992009-05-22 14:50:54 -040012074 * lpfc_sli_enable_intr - Enable device interrupt to SLI-3 device.
12075 * @phba: pointer to lpfc hba data structure.
Lee Jonesfe614ac2020-07-23 13:24:22 +010012076 * @cfg_mode: Interrupt configuration mode (INTx, MSI or MSI-X).
James Smart3772a992009-05-22 14:50:54 -040012077 *
12078 * This routine is invoked to enable device interrupt and associate driver's
12079 * interrupt handler(s) to interrupt vector(s) to device with SLI-3 interface
12080 * spec. Depends on the interrupt mode configured to the driver, the driver
12081 * will try to fallback from the configured interrupt mode to an interrupt
12082 * mode which is supported by the platform, kernel, and device in the order
12083 * of:
12084 * MSI-X -> MSI -> IRQ.
12085 *
12086 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -020012087 * 0 - successful
James Smart3772a992009-05-22 14:50:54 -040012088 * other values - error
12089 **/
12090static uint32_t
12091lpfc_sli_enable_intr(struct lpfc_hba *phba, uint32_t cfg_mode)
12092{
12093 uint32_t intr_mode = LPFC_INTR_ERROR;
12094 int retval;
12095
James Smartd2f25472021-01-04 10:02:27 -080012096 /* Need to issue conf_port mbox cmd before conf_msi mbox cmd */
12097 retval = lpfc_sli_config_port(phba, LPFC_SLI_REV3);
12098 if (retval)
12099 return intr_mode;
12100 phba->hba_flag &= ~HBA_NEEDS_CFG_PORT;
12101
James Smart3772a992009-05-22 14:50:54 -040012102 if (cfg_mode == 2) {
James Smartd2f25472021-01-04 10:02:27 -080012103 /* Now, try to enable MSI-X interrupt mode */
12104 retval = lpfc_sli_enable_msix(phba);
James Smart3772a992009-05-22 14:50:54 -040012105 if (!retval) {
James Smartd2f25472021-01-04 10:02:27 -080012106 /* Indicate initialization to MSI-X mode */
12107 phba->intr_type = MSIX;
12108 intr_mode = 2;
James Smart3772a992009-05-22 14:50:54 -040012109 }
12110 }
12111
12112 /* Fallback to MSI if MSI-X initialization failed */
12113 if (cfg_mode >= 1 && phba->intr_type == NONE) {
12114 retval = lpfc_sli_enable_msi(phba);
12115 if (!retval) {
12116 /* Indicate initialization to MSI mode */
12117 phba->intr_type = MSI;
12118 intr_mode = 1;
12119 }
12120 }
12121
12122 /* Fallback to INTx if both MSI-X/MSI initalization failed */
12123 if (phba->intr_type == NONE) {
12124 retval = request_irq(phba->pcidev->irq, lpfc_sli_intr_handler,
12125 IRQF_SHARED, LPFC_DRIVER_NAME, phba);
12126 if (!retval) {
12127 /* Indicate initialization to INTx mode */
12128 phba->intr_type = INTx;
12129 intr_mode = 0;
12130 }
12131 }
12132 return intr_mode;
12133}
12134
12135/**
12136 * lpfc_sli_disable_intr - Disable device interrupt to SLI-3 device.
12137 * @phba: pointer to lpfc hba data structure.
12138 *
12139 * This routine is invoked to disable device interrupt and disassociate the
12140 * driver's interrupt handler(s) from interrupt vector(s) to device with
12141 * SLI-3 interface spec. Depending on the interrupt mode, the driver will
12142 * release the interrupt vector(s) for the message signaled interrupt.
12143 **/
12144static void
12145lpfc_sli_disable_intr(struct lpfc_hba *phba)
12146{
Christoph Hellwig45ffac12017-02-12 13:52:26 -080012147 int nr_irqs, i;
12148
James Smart3772a992009-05-22 14:50:54 -040012149 if (phba->intr_type == MSIX)
Christoph Hellwig45ffac12017-02-12 13:52:26 -080012150 nr_irqs = LPFC_MSIX_VECTORS;
12151 else
12152 nr_irqs = 1;
12153
12154 for (i = 0; i < nr_irqs; i++)
12155 free_irq(pci_irq_vector(phba->pcidev, i), phba);
12156 pci_free_irq_vectors(phba->pcidev);
James Smart3772a992009-05-22 14:50:54 -040012157
12158 /* Reset interrupt management states */
12159 phba->intr_type = NONE;
12160 phba->sli.slistat.sli_intr = 0;
James Smart3772a992009-05-22 14:50:54 -040012161}
12162
12163/**
James Smart657add42019-05-21 17:49:06 -070012164 * lpfc_find_cpu_handle - Find the CPU that corresponds to the specified Queue
James Smart6a828b02019-01-28 11:14:31 -080012165 * @phba: pointer to lpfc hba data structure.
12166 * @id: EQ vector index or Hardware Queue index
12167 * @match: LPFC_FIND_BY_EQ = match by EQ
12168 * LPFC_FIND_BY_HDWQ = match by Hardware Queue
James Smart657add42019-05-21 17:49:06 -070012169 * Return the CPU that matches the selection criteria
James Smart6a828b02019-01-28 11:14:31 -080012170 */
12171static uint16_t
12172lpfc_find_cpu_handle(struct lpfc_hba *phba, uint16_t id, int match)
12173{
12174 struct lpfc_vector_map_info *cpup;
12175 int cpu;
12176
James Smart657add42019-05-21 17:49:06 -070012177 /* Loop through all CPUs */
James Smart222e9232019-01-28 11:14:35 -080012178 for_each_present_cpu(cpu) {
12179 cpup = &phba->sli4_hba.cpu_map[cpu];
James Smart657add42019-05-21 17:49:06 -070012180
12181 /* If we are matching by EQ, there may be multiple CPUs using
12182 * using the same vector, so select the one with
12183 * LPFC_CPU_FIRST_IRQ set.
12184 */
James Smart6a828b02019-01-28 11:14:31 -080012185 if ((match == LPFC_FIND_BY_EQ) &&
James Smart657add42019-05-21 17:49:06 -070012186 (cpup->flag & LPFC_CPU_FIRST_IRQ) &&
James Smart6a828b02019-01-28 11:14:31 -080012187 (cpup->eq == id))
12188 return cpu;
James Smart657add42019-05-21 17:49:06 -070012189
12190 /* If matching by HDWQ, select the first CPU that matches */
James Smart6a828b02019-01-28 11:14:31 -080012191 if ((match == LPFC_FIND_BY_HDWQ) && (cpup->hdwq == id))
12192 return cpu;
James Smart6a828b02019-01-28 11:14:31 -080012193 }
12194 return 0;
12195}
12196
James Smart6a828b02019-01-28 11:14:31 -080012197#ifdef CONFIG_X86
12198/**
12199 * lpfc_find_hyper - Determine if the CPU map entry is hyper-threaded
12200 * @phba: pointer to lpfc hba data structure.
12201 * @cpu: CPU map index
12202 * @phys_id: CPU package physical id
12203 * @core_id: CPU core id
12204 */
12205static int
12206lpfc_find_hyper(struct lpfc_hba *phba, int cpu,
12207 uint16_t phys_id, uint16_t core_id)
12208{
12209 struct lpfc_vector_map_info *cpup;
12210 int idx;
12211
James Smart222e9232019-01-28 11:14:35 -080012212 for_each_present_cpu(idx) {
12213 cpup = &phba->sli4_hba.cpu_map[idx];
James Smart6a828b02019-01-28 11:14:31 -080012214 /* Does the cpup match the one we are looking for */
12215 if ((cpup->phys_id == phys_id) &&
12216 (cpup->core_id == core_id) &&
James Smart222e9232019-01-28 11:14:35 -080012217 (cpu != idx))
James Smart6a828b02019-01-28 11:14:31 -080012218 return 1;
James Smart6a828b02019-01-28 11:14:31 -080012219 }
12220 return 0;
12221}
12222#endif
12223
James Smartdcaa2132019-11-04 16:57:06 -080012224/*
12225 * lpfc_assign_eq_map_info - Assigns eq for vector_map structure
12226 * @phba: pointer to lpfc hba data structure.
12227 * @eqidx: index for eq and irq vector
12228 * @flag: flags to set for vector_map structure
12229 * @cpu: cpu used to index vector_map structure
12230 *
12231 * The routine assigns eq info into vector_map structure
12232 */
12233static inline void
12234lpfc_assign_eq_map_info(struct lpfc_hba *phba, uint16_t eqidx, uint16_t flag,
12235 unsigned int cpu)
12236{
12237 struct lpfc_vector_map_info *cpup = &phba->sli4_hba.cpu_map[cpu];
12238 struct lpfc_hba_eq_hdl *eqhdl = lpfc_get_eq_hdl(eqidx);
12239
12240 cpup->eq = eqidx;
12241 cpup->flag |= flag;
12242
12243 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
12244 "3336 Set Affinity: CPU %d irq %d eq %d flag x%x\n",
12245 cpu, eqhdl->irq, cpup->eq, cpup->flag);
12246}
12247
12248/**
12249 * lpfc_cpu_map_array_init - Initialize cpu_map structure
12250 * @phba: pointer to lpfc hba data structure.
12251 *
12252 * The routine initializes the cpu_map array structure
12253 */
12254static void
12255lpfc_cpu_map_array_init(struct lpfc_hba *phba)
12256{
12257 struct lpfc_vector_map_info *cpup;
12258 struct lpfc_eq_intr_info *eqi;
12259 int cpu;
12260
12261 for_each_possible_cpu(cpu) {
12262 cpup = &phba->sli4_hba.cpu_map[cpu];
12263 cpup->phys_id = LPFC_VECTOR_MAP_EMPTY;
12264 cpup->core_id = LPFC_VECTOR_MAP_EMPTY;
12265 cpup->hdwq = LPFC_VECTOR_MAP_EMPTY;
12266 cpup->eq = LPFC_VECTOR_MAP_EMPTY;
12267 cpup->flag = 0;
12268 eqi = per_cpu_ptr(phba->sli4_hba.eq_info, cpu);
12269 INIT_LIST_HEAD(&eqi->list);
12270 eqi->icnt = 0;
12271 }
12272}
12273
12274/**
12275 * lpfc_hba_eq_hdl_array_init - Initialize hba_eq_hdl structure
12276 * @phba: pointer to lpfc hba data structure.
12277 *
12278 * The routine initializes the hba_eq_hdl array structure
12279 */
12280static void
12281lpfc_hba_eq_hdl_array_init(struct lpfc_hba *phba)
12282{
12283 struct lpfc_hba_eq_hdl *eqhdl;
12284 int i;
12285
12286 for (i = 0; i < phba->cfg_irq_chann; i++) {
12287 eqhdl = lpfc_get_eq_hdl(i);
12288 eqhdl->irq = LPFC_VECTOR_MAP_EMPTY;
12289 eqhdl->phba = phba;
12290 }
12291}
12292
James Smart6a828b02019-01-28 11:14:31 -080012293/**
James Smart895427b2017-02-12 13:52:30 -080012294 * lpfc_cpu_affinity_check - Check vector CPU affinity mappings
James Smart7bb03bb2013-04-17 20:19:16 -040012295 * @phba: pointer to lpfc hba data structure.
James Smart895427b2017-02-12 13:52:30 -080012296 * @vectors: number of msix vectors allocated.
James Smart7bb03bb2013-04-17 20:19:16 -040012297 *
James Smart895427b2017-02-12 13:52:30 -080012298 * The routine will figure out the CPU affinity assignment for every
James Smart6a828b02019-01-28 11:14:31 -080012299 * MSI-X vector allocated for the HBA.
James Smart895427b2017-02-12 13:52:30 -080012300 * In addition, the CPU to IO channel mapping will be calculated
12301 * and the phba->sli4_hba.cpu_map array will reflect this.
James Smart7bb03bb2013-04-17 20:19:16 -040012302 */
James Smart895427b2017-02-12 13:52:30 -080012303static void
12304lpfc_cpu_affinity_check(struct lpfc_hba *phba, int vectors)
James Smart7bb03bb2013-04-17 20:19:16 -040012305{
James Smart3ad348d2019-08-14 16:56:43 -070012306 int i, cpu, idx, next_idx, new_cpu, start_cpu, first_cpu;
James Smart6a828b02019-01-28 11:14:31 -080012307 int max_phys_id, min_phys_id;
12308 int max_core_id, min_core_id;
James Smart7bb03bb2013-04-17 20:19:16 -040012309 struct lpfc_vector_map_info *cpup;
James Smartd9954a22019-05-21 17:49:05 -070012310 struct lpfc_vector_map_info *new_cpup;
James Smart7bb03bb2013-04-17 20:19:16 -040012311#ifdef CONFIG_X86
12312 struct cpuinfo_x86 *cpuinfo;
12313#endif
James Smart840eda92020-03-22 11:13:00 -070012314#ifdef CONFIG_SCSI_LPFC_DEBUG_FS
12315 struct lpfc_hdwq_stat *c_stat;
12316#endif
James Smart7bb03bb2013-04-17 20:19:16 -040012317
James Smart6a828b02019-01-28 11:14:31 -080012318 max_phys_id = 0;
James Smartd9954a22019-05-21 17:49:05 -070012319 min_phys_id = LPFC_VECTOR_MAP_EMPTY;
James Smart6a828b02019-01-28 11:14:31 -080012320 max_core_id = 0;
James Smartd9954a22019-05-21 17:49:05 -070012321 min_core_id = LPFC_VECTOR_MAP_EMPTY;
James Smart7bb03bb2013-04-17 20:19:16 -040012322
12323 /* Update CPU map with physical id and core id of each CPU */
James Smart222e9232019-01-28 11:14:35 -080012324 for_each_present_cpu(cpu) {
12325 cpup = &phba->sli4_hba.cpu_map[cpu];
James Smart7bb03bb2013-04-17 20:19:16 -040012326#ifdef CONFIG_X86
12327 cpuinfo = &cpu_data(cpu);
12328 cpup->phys_id = cpuinfo->phys_proc_id;
12329 cpup->core_id = cpuinfo->cpu_core_id;
James Smartd9954a22019-05-21 17:49:05 -070012330 if (lpfc_find_hyper(phba, cpu, cpup->phys_id, cpup->core_id))
12331 cpup->flag |= LPFC_CPU_MAP_HYPER;
James Smart7bb03bb2013-04-17 20:19:16 -040012332#else
12333 /* No distinction between CPUs for other platforms */
12334 cpup->phys_id = 0;
James Smart6a828b02019-01-28 11:14:31 -080012335 cpup->core_id = cpu;
James Smart7bb03bb2013-04-17 20:19:16 -040012336#endif
James Smart7bb03bb2013-04-17 20:19:16 -040012337
James Smartb3295c22019-01-28 11:14:30 -080012338 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
James Smart3ad348d2019-08-14 16:56:43 -070012339 "3328 CPU %d physid %d coreid %d flag x%x\n",
12340 cpu, cpup->phys_id, cpup->core_id, cpup->flag);
James Smart6a828b02019-01-28 11:14:31 -080012341
12342 if (cpup->phys_id > max_phys_id)
12343 max_phys_id = cpup->phys_id;
12344 if (cpup->phys_id < min_phys_id)
12345 min_phys_id = cpup->phys_id;
12346
12347 if (cpup->core_id > max_core_id)
12348 max_core_id = cpup->core_id;
12349 if (cpup->core_id < min_core_id)
12350 min_core_id = cpup->core_id;
James Smart7bb03bb2013-04-17 20:19:16 -040012351 }
James Smartb3295c22019-01-28 11:14:30 -080012352
James Smartd9954a22019-05-21 17:49:05 -070012353 /* After looking at each irq vector assigned to this pcidev, its
12354 * possible to see that not ALL CPUs have been accounted for.
James Smart657add42019-05-21 17:49:06 -070012355 * Next we will set any unassigned (unaffinitized) cpu map
12356 * entries to a IRQ on the same phys_id.
James Smartd9954a22019-05-21 17:49:05 -070012357 */
12358 first_cpu = cpumask_first(cpu_present_mask);
12359 start_cpu = first_cpu;
12360
12361 for_each_present_cpu(cpu) {
12362 cpup = &phba->sli4_hba.cpu_map[cpu];
12363
12364 /* Is this CPU entry unassigned */
12365 if (cpup->eq == LPFC_VECTOR_MAP_EMPTY) {
12366 /* Mark CPU as IRQ not assigned by the kernel */
12367 cpup->flag |= LPFC_CPU_MAP_UNASSIGN;
12368
James Smart657add42019-05-21 17:49:06 -070012369 /* If so, find a new_cpup thats on the the SAME
James Smartd9954a22019-05-21 17:49:05 -070012370 * phys_id as cpup. start_cpu will start where we
12371 * left off so all unassigned entries don't get assgined
12372 * the IRQ of the first entry.
12373 */
12374 new_cpu = start_cpu;
12375 for (i = 0; i < phba->sli4_hba.num_present_cpu; i++) {
12376 new_cpup = &phba->sli4_hba.cpu_map[new_cpu];
12377 if (!(new_cpup->flag & LPFC_CPU_MAP_UNASSIGN) &&
James Smartdcaa2132019-11-04 16:57:06 -080012378 (new_cpup->eq != LPFC_VECTOR_MAP_EMPTY) &&
James Smartd9954a22019-05-21 17:49:05 -070012379 (new_cpup->phys_id == cpup->phys_id))
12380 goto found_same;
12381 new_cpu = cpumask_next(
12382 new_cpu, cpu_present_mask);
12383 if (new_cpu == nr_cpumask_bits)
12384 new_cpu = first_cpu;
12385 }
12386 /* At this point, we leave the CPU as unassigned */
12387 continue;
12388found_same:
12389 /* We found a matching phys_id, so copy the IRQ info */
12390 cpup->eq = new_cpup->eq;
James Smartd9954a22019-05-21 17:49:05 -070012391
12392 /* Bump start_cpu to the next slot to minmize the
12393 * chance of having multiple unassigned CPU entries
12394 * selecting the same IRQ.
12395 */
12396 start_cpu = cpumask_next(new_cpu, cpu_present_mask);
12397 if (start_cpu == nr_cpumask_bits)
12398 start_cpu = first_cpu;
12399
James Smart657add42019-05-21 17:49:06 -070012400 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
James Smartd9954a22019-05-21 17:49:05 -070012401 "3337 Set Affinity: CPU %d "
James Smartdcaa2132019-11-04 16:57:06 -080012402 "eq %d from peer cpu %d same "
James Smartd9954a22019-05-21 17:49:05 -070012403 "phys_id (%d)\n",
James Smartdcaa2132019-11-04 16:57:06 -080012404 cpu, cpup->eq, new_cpu,
12405 cpup->phys_id);
James Smartd9954a22019-05-21 17:49:05 -070012406 }
12407 }
12408
12409 /* Set any unassigned cpu map entries to a IRQ on any phys_id */
12410 start_cpu = first_cpu;
12411
12412 for_each_present_cpu(cpu) {
12413 cpup = &phba->sli4_hba.cpu_map[cpu];
12414
12415 /* Is this entry unassigned */
12416 if (cpup->eq == LPFC_VECTOR_MAP_EMPTY) {
12417 /* Mark it as IRQ not assigned by the kernel */
12418 cpup->flag |= LPFC_CPU_MAP_UNASSIGN;
12419
James Smart657add42019-05-21 17:49:06 -070012420 /* If so, find a new_cpup thats on ANY phys_id
James Smartd9954a22019-05-21 17:49:05 -070012421 * as the cpup. start_cpu will start where we
12422 * left off so all unassigned entries don't get
12423 * assigned the IRQ of the first entry.
12424 */
12425 new_cpu = start_cpu;
12426 for (i = 0; i < phba->sli4_hba.num_present_cpu; i++) {
12427 new_cpup = &phba->sli4_hba.cpu_map[new_cpu];
12428 if (!(new_cpup->flag & LPFC_CPU_MAP_UNASSIGN) &&
James Smartdcaa2132019-11-04 16:57:06 -080012429 (new_cpup->eq != LPFC_VECTOR_MAP_EMPTY))
James Smartd9954a22019-05-21 17:49:05 -070012430 goto found_any;
12431 new_cpu = cpumask_next(
12432 new_cpu, cpu_present_mask);
12433 if (new_cpu == nr_cpumask_bits)
12434 new_cpu = first_cpu;
12435 }
12436 /* We should never leave an entry unassigned */
12437 lpfc_printf_log(phba, KERN_ERR, LOG_INIT,
12438 "3339 Set Affinity: CPU %d "
James Smartdcaa2132019-11-04 16:57:06 -080012439 "eq %d UNASSIGNED\n",
12440 cpup->hdwq, cpup->eq);
James Smartd9954a22019-05-21 17:49:05 -070012441 continue;
12442found_any:
12443 /* We found an available entry, copy the IRQ info */
12444 cpup->eq = new_cpup->eq;
James Smartd9954a22019-05-21 17:49:05 -070012445
12446 /* Bump start_cpu to the next slot to minmize the
12447 * chance of having multiple unassigned CPU entries
12448 * selecting the same IRQ.
12449 */
12450 start_cpu = cpumask_next(new_cpu, cpu_present_mask);
12451 if (start_cpu == nr_cpumask_bits)
12452 start_cpu = first_cpu;
12453
James Smart657add42019-05-21 17:49:06 -070012454 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
James Smartd9954a22019-05-21 17:49:05 -070012455 "3338 Set Affinity: CPU %d "
James Smartdcaa2132019-11-04 16:57:06 -080012456 "eq %d from peer cpu %d (%d/%d)\n",
12457 cpu, cpup->eq, new_cpu,
James Smartd9954a22019-05-21 17:49:05 -070012458 new_cpup->phys_id, new_cpup->core_id);
12459 }
12460 }
James Smart657add42019-05-21 17:49:06 -070012461
James Smart3ad348d2019-08-14 16:56:43 -070012462 /* Assign hdwq indices that are unique across all cpus in the map
12463 * that are also FIRST_CPUs.
12464 */
12465 idx = 0;
12466 for_each_present_cpu(cpu) {
12467 cpup = &phba->sli4_hba.cpu_map[cpu];
12468
12469 /* Only FIRST IRQs get a hdwq index assignment. */
12470 if (!(cpup->flag & LPFC_CPU_FIRST_IRQ))
12471 continue;
12472
12473 /* 1 to 1, the first LPFC_CPU_FIRST_IRQ cpus to a unique hdwq */
12474 cpup->hdwq = idx;
12475 idx++;
Anton Blanchardbc2736e2020-07-13 18:39:08 +100012476 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
James Smart3ad348d2019-08-14 16:56:43 -070012477 "3333 Set Affinity: CPU %d (phys %d core %d): "
James Smartdcaa2132019-11-04 16:57:06 -080012478 "hdwq %d eq %d flg x%x\n",
James Smart3ad348d2019-08-14 16:56:43 -070012479 cpu, cpup->phys_id, cpup->core_id,
James Smartdcaa2132019-11-04 16:57:06 -080012480 cpup->hdwq, cpup->eq, cpup->flag);
James Smart3ad348d2019-08-14 16:56:43 -070012481 }
James Smartbc227dd2019-11-11 15:03:59 -080012482 /* Associate a hdwq with each cpu_map entry
James Smart657add42019-05-21 17:49:06 -070012483 * This will be 1 to 1 - hdwq to cpu, unless there are less
12484 * hardware queues then CPUs. For that case we will just round-robin
12485 * the available hardware queues as they get assigned to CPUs.
James Smart3ad348d2019-08-14 16:56:43 -070012486 * The next_idx is the idx from the FIRST_CPU loop above to account
12487 * for irq_chann < hdwq. The idx is used for round-robin assignments
12488 * and needs to start at 0.
James Smart657add42019-05-21 17:49:06 -070012489 */
James Smart3ad348d2019-08-14 16:56:43 -070012490 next_idx = idx;
James Smart657add42019-05-21 17:49:06 -070012491 start_cpu = 0;
James Smart3ad348d2019-08-14 16:56:43 -070012492 idx = 0;
James Smart657add42019-05-21 17:49:06 -070012493 for_each_present_cpu(cpu) {
12494 cpup = &phba->sli4_hba.cpu_map[cpu];
James Smart657add42019-05-21 17:49:06 -070012495
James Smart3ad348d2019-08-14 16:56:43 -070012496 /* FIRST cpus are already mapped. */
12497 if (cpup->flag & LPFC_CPU_FIRST_IRQ)
12498 continue;
James Smart657add42019-05-21 17:49:06 -070012499
James Smart3ad348d2019-08-14 16:56:43 -070012500 /* If the cfg_irq_chann < cfg_hdw_queue, set the hdwq
12501 * of the unassigned cpus to the next idx so that all
12502 * hdw queues are fully utilized.
12503 */
12504 if (next_idx < phba->cfg_hdw_queue) {
12505 cpup->hdwq = next_idx;
12506 next_idx++;
12507 continue;
James Smart657add42019-05-21 17:49:06 -070012508 }
James Smart3ad348d2019-08-14 16:56:43 -070012509
12510 /* Not a First CPU and all hdw_queues are used. Reuse a
12511 * Hardware Queue for another CPU, so be smart about it
12512 * and pick one that has its IRQ/EQ mapped to the same phys_id
12513 * (CPU package) and core_id.
12514 */
12515 new_cpu = start_cpu;
12516 for (i = 0; i < phba->sli4_hba.num_present_cpu; i++) {
12517 new_cpup = &phba->sli4_hba.cpu_map[new_cpu];
12518 if (new_cpup->hdwq != LPFC_VECTOR_MAP_EMPTY &&
12519 new_cpup->phys_id == cpup->phys_id &&
12520 new_cpup->core_id == cpup->core_id) {
12521 goto found_hdwq;
12522 }
12523 new_cpu = cpumask_next(new_cpu, cpu_present_mask);
12524 if (new_cpu == nr_cpumask_bits)
12525 new_cpu = first_cpu;
12526 }
12527
12528 /* If we can't match both phys_id and core_id,
12529 * settle for just a phys_id match.
12530 */
12531 new_cpu = start_cpu;
12532 for (i = 0; i < phba->sli4_hba.num_present_cpu; i++) {
12533 new_cpup = &phba->sli4_hba.cpu_map[new_cpu];
12534 if (new_cpup->hdwq != LPFC_VECTOR_MAP_EMPTY &&
12535 new_cpup->phys_id == cpup->phys_id)
12536 goto found_hdwq;
12537
12538 new_cpu = cpumask_next(new_cpu, cpu_present_mask);
12539 if (new_cpu == nr_cpumask_bits)
12540 new_cpu = first_cpu;
12541 }
12542
12543 /* Otherwise just round robin on cfg_hdw_queue */
12544 cpup->hdwq = idx % phba->cfg_hdw_queue;
12545 idx++;
12546 goto logit;
12547 found_hdwq:
12548 /* We found an available entry, copy the IRQ info */
12549 start_cpu = cpumask_next(new_cpu, cpu_present_mask);
12550 if (start_cpu == nr_cpumask_bits)
12551 start_cpu = first_cpu;
12552 cpup->hdwq = new_cpup->hdwq;
12553 logit:
Anton Blanchardbc2736e2020-07-13 18:39:08 +100012554 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
James Smart657add42019-05-21 17:49:06 -070012555 "3335 Set Affinity: CPU %d (phys %d core %d): "
James Smartdcaa2132019-11-04 16:57:06 -080012556 "hdwq %d eq %d flg x%x\n",
James Smart657add42019-05-21 17:49:06 -070012557 cpu, cpup->phys_id, cpup->core_id,
James Smartdcaa2132019-11-04 16:57:06 -080012558 cpup->hdwq, cpup->eq, cpup->flag);
James Smart657add42019-05-21 17:49:06 -070012559 }
12560
James Smartbc227dd2019-11-11 15:03:59 -080012561 /*
12562 * Initialize the cpu_map slots for not-present cpus in case
12563 * a cpu is hot-added. Perform a simple hdwq round robin assignment.
12564 */
12565 idx = 0;
12566 for_each_possible_cpu(cpu) {
12567 cpup = &phba->sli4_hba.cpu_map[cpu];
James Smart840eda92020-03-22 11:13:00 -070012568#ifdef CONFIG_SCSI_LPFC_DEBUG_FS
12569 c_stat = per_cpu_ptr(phba->sli4_hba.c_stat, cpu);
12570 c_stat->hdwq_no = cpup->hdwq;
12571#endif
James Smartbc227dd2019-11-11 15:03:59 -080012572 if (cpup->hdwq != LPFC_VECTOR_MAP_EMPTY)
12573 continue;
12574
12575 cpup->hdwq = idx++ % phba->cfg_hdw_queue;
James Smart840eda92020-03-22 11:13:00 -070012576#ifdef CONFIG_SCSI_LPFC_DEBUG_FS
12577 c_stat->hdwq_no = cpup->hdwq;
12578#endif
James Smartbc227dd2019-11-11 15:03:59 -080012579 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
12580 "3340 Set Affinity: not present "
12581 "CPU %d hdwq %d\n",
12582 cpu, cpup->hdwq);
James Smart657add42019-05-21 17:49:06 -070012583 }
12584
12585 /* The cpu_map array will be used later during initialization
12586 * when EQ / CQ / WQs are allocated and configured.
12587 */
James Smartb3295c22019-01-28 11:14:30 -080012588 return;
James Smart7bb03bb2013-04-17 20:19:16 -040012589}
James Smart7bb03bb2013-04-17 20:19:16 -040012590
12591/**
James Smart93a4d6f2019-11-04 16:57:05 -080012592 * lpfc_cpuhp_get_eq
12593 *
12594 * @phba: pointer to lpfc hba data structure.
12595 * @cpu: cpu going offline
Lee Jonesfe614ac2020-07-23 13:24:22 +010012596 * @eqlist: eq list to append to
James Smart93a4d6f2019-11-04 16:57:05 -080012597 */
James Smarta99c8072020-01-27 16:23:06 -080012598static int
James Smart93a4d6f2019-11-04 16:57:05 -080012599lpfc_cpuhp_get_eq(struct lpfc_hba *phba, unsigned int cpu,
12600 struct list_head *eqlist)
12601{
James Smart93a4d6f2019-11-04 16:57:05 -080012602 const struct cpumask *maskp;
12603 struct lpfc_queue *eq;
James Smarta99c8072020-01-27 16:23:06 -080012604 struct cpumask *tmp;
James Smart93a4d6f2019-11-04 16:57:05 -080012605 u16 idx;
12606
James Smarta99c8072020-01-27 16:23:06 -080012607 tmp = kzalloc(cpumask_size(), GFP_KERNEL);
12608 if (!tmp)
12609 return -ENOMEM;
12610
James Smart93a4d6f2019-11-04 16:57:05 -080012611 for (idx = 0; idx < phba->cfg_irq_chann; idx++) {
12612 maskp = pci_irq_get_affinity(phba->pcidev, idx);
12613 if (!maskp)
12614 continue;
12615 /*
12616 * if irq is not affinitized to the cpu going
12617 * then we don't need to poll the eq attached
12618 * to it.
12619 */
James Smarta99c8072020-01-27 16:23:06 -080012620 if (!cpumask_and(tmp, maskp, cpumask_of(cpu)))
James Smart93a4d6f2019-11-04 16:57:05 -080012621 continue;
12622 /* get the cpus that are online and are affini-
12623 * tized to this irq vector. If the count is
12624 * more than 1 then cpuhp is not going to shut-
12625 * down this vector. Since this cpu has not
12626 * gone offline yet, we need >1.
12627 */
James Smarta99c8072020-01-27 16:23:06 -080012628 cpumask_and(tmp, maskp, cpu_online_mask);
12629 if (cpumask_weight(tmp) > 1)
James Smart93a4d6f2019-11-04 16:57:05 -080012630 continue;
12631
12632 /* Now that we have an irq to shutdown, get the eq
12633 * mapped to this irq. Note: multiple hdwq's in
12634 * the software can share an eq, but eventually
12635 * only eq will be mapped to this vector
12636 */
James Smartdcaa2132019-11-04 16:57:06 -080012637 eq = phba->sli4_hba.hba_eq_hdl[idx].eq;
12638 list_add(&eq->_poll_list, eqlist);
James Smart93a4d6f2019-11-04 16:57:05 -080012639 }
James Smarta99c8072020-01-27 16:23:06 -080012640 kfree(tmp);
12641 return 0;
James Smart93a4d6f2019-11-04 16:57:05 -080012642}
12643
12644static void __lpfc_cpuhp_remove(struct lpfc_hba *phba)
12645{
12646 if (phba->sli_rev != LPFC_SLI_REV4)
12647 return;
12648
12649 cpuhp_state_remove_instance_nocalls(lpfc_cpuhp_state,
12650 &phba->cpuhp);
12651 /*
12652 * unregistering the instance doesn't stop the polling
12653 * timer. Wait for the poll timer to retire.
12654 */
12655 synchronize_rcu();
12656 del_timer_sync(&phba->cpuhp_poll_timer);
12657}
12658
12659static void lpfc_cpuhp_remove(struct lpfc_hba *phba)
12660{
12661 if (phba->pport->fc_flag & FC_OFFLINE_MODE)
12662 return;
12663
12664 __lpfc_cpuhp_remove(phba);
12665}
12666
12667static void lpfc_cpuhp_add(struct lpfc_hba *phba)
12668{
12669 if (phba->sli_rev != LPFC_SLI_REV4)
12670 return;
12671
12672 rcu_read_lock();
12673
James Smartf861f592020-03-22 11:12:54 -070012674 if (!list_empty(&phba->poll_list))
James Smart93a4d6f2019-11-04 16:57:05 -080012675 mod_timer(&phba->cpuhp_poll_timer,
12676 jiffies + msecs_to_jiffies(LPFC_POLL_HB));
James Smart93a4d6f2019-11-04 16:57:05 -080012677
12678 rcu_read_unlock();
12679
12680 cpuhp_state_add_instance_nocalls(lpfc_cpuhp_state,
12681 &phba->cpuhp);
12682}
12683
12684static int __lpfc_cpuhp_checks(struct lpfc_hba *phba, int *retval)
12685{
12686 if (phba->pport->load_flag & FC_UNLOADING) {
12687 *retval = -EAGAIN;
12688 return true;
12689 }
12690
12691 if (phba->sli_rev != LPFC_SLI_REV4) {
12692 *retval = 0;
12693 return true;
12694 }
12695
12696 /* proceed with the hotplug */
12697 return false;
12698}
12699
James Smartdcaa2132019-11-04 16:57:06 -080012700/**
12701 * lpfc_irq_set_aff - set IRQ affinity
12702 * @eqhdl: EQ handle
12703 * @cpu: cpu to set affinity
12704 *
12705 **/
12706static inline void
12707lpfc_irq_set_aff(struct lpfc_hba_eq_hdl *eqhdl, unsigned int cpu)
12708{
12709 cpumask_clear(&eqhdl->aff_mask);
12710 cpumask_set_cpu(cpu, &eqhdl->aff_mask);
12711 irq_set_status_flags(eqhdl->irq, IRQ_NO_BALANCING);
12712 irq_set_affinity_hint(eqhdl->irq, &eqhdl->aff_mask);
12713}
12714
12715/**
12716 * lpfc_irq_clear_aff - clear IRQ affinity
12717 * @eqhdl: EQ handle
12718 *
12719 **/
12720static inline void
12721lpfc_irq_clear_aff(struct lpfc_hba_eq_hdl *eqhdl)
12722{
12723 cpumask_clear(&eqhdl->aff_mask);
12724 irq_clear_status_flags(eqhdl->irq, IRQ_NO_BALANCING);
James Smartdcaa2132019-11-04 16:57:06 -080012725}
12726
12727/**
12728 * lpfc_irq_rebalance - rebalances IRQ affinity according to cpuhp event
12729 * @phba: pointer to HBA context object.
12730 * @cpu: cpu going offline/online
12731 * @offline: true, cpu is going offline. false, cpu is coming online.
12732 *
12733 * If cpu is going offline, we'll try our best effort to find the next
Dick Kennedy3048e3e2020-05-01 14:43:06 -070012734 * online cpu on the phba's original_mask and migrate all offlining IRQ
12735 * affinities.
James Smartdcaa2132019-11-04 16:57:06 -080012736 *
Dick Kennedy3048e3e2020-05-01 14:43:06 -070012737 * If cpu is coming online, reaffinitize the IRQ back to the onlining cpu.
James Smartdcaa2132019-11-04 16:57:06 -080012738 *
Dick Kennedy3048e3e2020-05-01 14:43:06 -070012739 * Note: Call only if NUMA or NHT mode is enabled, otherwise rely on
James Smartdcaa2132019-11-04 16:57:06 -080012740 * PCI_IRQ_AFFINITY to auto-manage IRQ affinity.
12741 *
12742 **/
12743static void
12744lpfc_irq_rebalance(struct lpfc_hba *phba, unsigned int cpu, bool offline)
12745{
12746 struct lpfc_vector_map_info *cpup;
12747 struct cpumask *aff_mask;
12748 unsigned int cpu_select, cpu_next, idx;
Dick Kennedy3048e3e2020-05-01 14:43:06 -070012749 const struct cpumask *orig_mask;
James Smartdcaa2132019-11-04 16:57:06 -080012750
Dick Kennedy3048e3e2020-05-01 14:43:06 -070012751 if (phba->irq_chann_mode == NORMAL_MODE)
James Smartdcaa2132019-11-04 16:57:06 -080012752 return;
12753
Dick Kennedy3048e3e2020-05-01 14:43:06 -070012754 orig_mask = &phba->sli4_hba.irq_aff_mask;
James Smartdcaa2132019-11-04 16:57:06 -080012755
Dick Kennedy3048e3e2020-05-01 14:43:06 -070012756 if (!cpumask_test_cpu(cpu, orig_mask))
James Smartdcaa2132019-11-04 16:57:06 -080012757 return;
12758
12759 cpup = &phba->sli4_hba.cpu_map[cpu];
12760
12761 if (!(cpup->flag & LPFC_CPU_FIRST_IRQ))
12762 return;
12763
12764 if (offline) {
Dick Kennedy3048e3e2020-05-01 14:43:06 -070012765 /* Find next online CPU on original mask */
12766 cpu_next = cpumask_next_wrap(cpu, orig_mask, cpu, true);
12767 cpu_select = lpfc_next_online_cpu(orig_mask, cpu_next);
James Smartdcaa2132019-11-04 16:57:06 -080012768
12769 /* Found a valid CPU */
12770 if ((cpu_select < nr_cpu_ids) && (cpu_select != cpu)) {
12771 /* Go through each eqhdl and ensure offlining
12772 * cpu aff_mask is migrated
12773 */
12774 for (idx = 0; idx < phba->cfg_irq_chann; idx++) {
12775 aff_mask = lpfc_get_aff_mask(idx);
12776
12777 /* Migrate affinity */
12778 if (cpumask_test_cpu(cpu, aff_mask))
12779 lpfc_irq_set_aff(lpfc_get_eq_hdl(idx),
12780 cpu_select);
12781 }
12782 } else {
12783 /* Rely on irqbalance if no online CPUs left on NUMA */
12784 for (idx = 0; idx < phba->cfg_irq_chann; idx++)
12785 lpfc_irq_clear_aff(lpfc_get_eq_hdl(idx));
12786 }
12787 } else {
12788 /* Migrate affinity back to this CPU */
12789 lpfc_irq_set_aff(lpfc_get_eq_hdl(cpup->eq), cpu);
12790 }
12791}
12792
James Smart93a4d6f2019-11-04 16:57:05 -080012793static int lpfc_cpu_offline(unsigned int cpu, struct hlist_node *node)
12794{
12795 struct lpfc_hba *phba = hlist_entry_safe(node, struct lpfc_hba, cpuhp);
12796 struct lpfc_queue *eq, *next;
12797 LIST_HEAD(eqlist);
12798 int retval;
12799
12800 if (!phba) {
12801 WARN_ONCE(!phba, "cpu: %u. phba:NULL", raw_smp_processor_id());
12802 return 0;
12803 }
12804
12805 if (__lpfc_cpuhp_checks(phba, &retval))
12806 return retval;
12807
James Smartdcaa2132019-11-04 16:57:06 -080012808 lpfc_irq_rebalance(phba, cpu, true);
12809
James Smarta99c8072020-01-27 16:23:06 -080012810 retval = lpfc_cpuhp_get_eq(phba, cpu, &eqlist);
12811 if (retval)
12812 return retval;
James Smart93a4d6f2019-11-04 16:57:05 -080012813
12814 /* start polling on these eq's */
12815 list_for_each_entry_safe(eq, next, &eqlist, _poll_list) {
12816 list_del_init(&eq->_poll_list);
12817 lpfc_sli4_start_polling(eq);
12818 }
12819
12820 return 0;
12821}
12822
12823static int lpfc_cpu_online(unsigned int cpu, struct hlist_node *node)
12824{
12825 struct lpfc_hba *phba = hlist_entry_safe(node, struct lpfc_hba, cpuhp);
12826 struct lpfc_queue *eq, *next;
12827 unsigned int n;
12828 int retval;
12829
12830 if (!phba) {
12831 WARN_ONCE(!phba, "cpu: %u. phba:NULL", raw_smp_processor_id());
12832 return 0;
12833 }
12834
12835 if (__lpfc_cpuhp_checks(phba, &retval))
12836 return retval;
12837
James Smartdcaa2132019-11-04 16:57:06 -080012838 lpfc_irq_rebalance(phba, cpu, false);
12839
James Smart93a4d6f2019-11-04 16:57:05 -080012840 list_for_each_entry_safe(eq, next, &phba->poll_list, _poll_list) {
12841 n = lpfc_find_cpu_handle(phba, eq->hdwq, LPFC_FIND_BY_HDWQ);
12842 if (n == cpu)
12843 lpfc_sli4_stop_polling(eq);
12844 }
12845
12846 return 0;
12847}
12848
12849/**
James Smartda0436e2009-05-22 14:51:39 -040012850 * lpfc_sli4_enable_msix - Enable MSI-X interrupt mode to SLI-4 device
12851 * @phba: pointer to lpfc hba data structure.
12852 *
12853 * This routine is invoked to enable the MSI-X interrupt vectors to device
James Smartdcaa2132019-11-04 16:57:06 -080012854 * with SLI-4 interface spec. It also allocates MSI-X vectors and maps them
12855 * to cpus on the system.
12856 *
12857 * When cfg_irq_numa is enabled, the adapter will only allocate vectors for
12858 * the number of cpus on the same numa node as this adapter. The vectors are
12859 * allocated without requesting OS affinity mapping. A vector will be
12860 * allocated and assigned to each online and offline cpu. If the cpu is
12861 * online, then affinity will be set to that cpu. If the cpu is offline, then
12862 * affinity will be set to the nearest peer cpu within the numa node that is
12863 * online. If there are no online cpus within the numa node, affinity is not
12864 * assigned and the OS may do as it pleases. Note: cpu vector affinity mapping
12865 * is consistent with the way cpu online/offline is handled when cfg_irq_numa is
12866 * configured.
12867 *
12868 * If numa mode is not enabled and there is more than 1 vector allocated, then
12869 * the driver relies on the managed irq interface where the OS assigns vector to
12870 * cpu affinity. The driver will then use that affinity mapping to setup its
12871 * cpu mapping table.
James Smartda0436e2009-05-22 14:51:39 -040012872 *
12873 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -020012874 * 0 - successful
James Smartda0436e2009-05-22 14:51:39 -040012875 * other values - error
12876 **/
12877static int
12878lpfc_sli4_enable_msix(struct lpfc_hba *phba)
12879{
James Smart75baf692010-06-08 18:31:21 -040012880 int vectors, rc, index;
James Smartb83d0052017-06-01 21:07:05 -070012881 char *name;
Dick Kennedy3048e3e2020-05-01 14:43:06 -070012882 const struct cpumask *aff_mask = NULL;
James Smartdcaa2132019-11-04 16:57:06 -080012883 unsigned int cpu = 0, cpu_cnt = 0, cpu_select = nr_cpu_ids;
Dick Kennedy17105d92020-07-06 13:42:30 -070012884 struct lpfc_vector_map_info *cpup;
James Smartdcaa2132019-11-04 16:57:06 -080012885 struct lpfc_hba_eq_hdl *eqhdl;
12886 const struct cpumask *maskp;
James Smartdcaa2132019-11-04 16:57:06 -080012887 unsigned int flags = PCI_IRQ_MSIX;
James Smartda0436e2009-05-22 14:51:39 -040012888
12889 /* Set up MSI-X multi-message vectors */
James Smart6a828b02019-01-28 11:14:31 -080012890 vectors = phba->cfg_irq_chann;
Christoph Hellwig45ffac12017-02-12 13:52:26 -080012891
Dick Kennedy3048e3e2020-05-01 14:43:06 -070012892 if (phba->irq_chann_mode != NORMAL_MODE)
12893 aff_mask = &phba->sli4_hba.irq_aff_mask;
12894
12895 if (aff_mask) {
12896 cpu_cnt = cpumask_weight(aff_mask);
James Smartdcaa2132019-11-04 16:57:06 -080012897 vectors = min(phba->cfg_irq_chann, cpu_cnt);
12898
Dick Kennedy3048e3e2020-05-01 14:43:06 -070012899 /* cpu: iterates over aff_mask including offline or online
12900 * cpu_select: iterates over online aff_mask to set affinity
James Smartdcaa2132019-11-04 16:57:06 -080012901 */
Dick Kennedy3048e3e2020-05-01 14:43:06 -070012902 cpu = cpumask_first(aff_mask);
12903 cpu_select = lpfc_next_online_cpu(aff_mask, cpu);
James Smartdcaa2132019-11-04 16:57:06 -080012904 } else {
12905 flags |= PCI_IRQ_AFFINITY;
12906 }
12907
12908 rc = pci_alloc_irq_vectors(phba->pcidev, 1, vectors, flags);
Alexander Gordeev4f871e12014-09-03 12:56:29 -040012909 if (rc < 0) {
James Smartda0436e2009-05-22 14:51:39 -040012910 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
12911 "0484 PCI enable MSI-X failed (%d)\n", rc);
Alexander Gordeev029165a2014-07-16 20:05:15 +020012912 goto vec_fail_out;
James Smartda0436e2009-05-22 14:51:39 -040012913 }
Alexander Gordeev4f871e12014-09-03 12:56:29 -040012914 vectors = rc;
James Smart75baf692010-06-08 18:31:21 -040012915
James Smart7bb03bb2013-04-17 20:19:16 -040012916 /* Assign MSI-X vectors to interrupt handlers */
James Smart67d12732012-08-03 12:36:13 -040012917 for (index = 0; index < vectors; index++) {
James Smartdcaa2132019-11-04 16:57:06 -080012918 eqhdl = lpfc_get_eq_hdl(index);
12919 name = eqhdl->handler_name;
James Smartb83d0052017-06-01 21:07:05 -070012920 memset(name, 0, LPFC_SLI4_HANDLER_NAME_SZ);
12921 snprintf(name, LPFC_SLI4_HANDLER_NAME_SZ,
James Smart4305f182012-08-03 12:36:33 -040012922 LPFC_DRIVER_HANDLER_NAME"%d", index);
James Smartda0436e2009-05-22 14:51:39 -040012923
James Smartdcaa2132019-11-04 16:57:06 -080012924 eqhdl->idx = index;
James Smart7370d102019-01-28 11:14:20 -080012925 rc = request_irq(pci_irq_vector(phba->pcidev, index),
12926 &lpfc_sli4_hba_intr_handler, 0,
James Smartdcaa2132019-11-04 16:57:06 -080012927 name, eqhdl);
James Smartda0436e2009-05-22 14:51:39 -040012928 if (rc) {
12929 lpfc_printf_log(phba, KERN_WARNING, LOG_INIT,
12930 "0486 MSI-X fast-path (%d) "
12931 "request_irq failed (%d)\n", index, rc);
12932 goto cfg_fail_out;
12933 }
James Smartdcaa2132019-11-04 16:57:06 -080012934
12935 eqhdl->irq = pci_irq_vector(phba->pcidev, index);
12936
Dick Kennedy3048e3e2020-05-01 14:43:06 -070012937 if (aff_mask) {
James Smartdcaa2132019-11-04 16:57:06 -080012938 /* If found a neighboring online cpu, set affinity */
12939 if (cpu_select < nr_cpu_ids)
12940 lpfc_irq_set_aff(eqhdl, cpu_select);
12941
12942 /* Assign EQ to cpu_map */
12943 lpfc_assign_eq_map_info(phba, index,
12944 LPFC_CPU_FIRST_IRQ,
12945 cpu);
12946
Dick Kennedy3048e3e2020-05-01 14:43:06 -070012947 /* Iterate to next offline or online cpu in aff_mask */
12948 cpu = cpumask_next(cpu, aff_mask);
James Smartdcaa2132019-11-04 16:57:06 -080012949
Dick Kennedy3048e3e2020-05-01 14:43:06 -070012950 /* Find next online cpu in aff_mask to set affinity */
12951 cpu_select = lpfc_next_online_cpu(aff_mask, cpu);
James Smartdcaa2132019-11-04 16:57:06 -080012952 } else if (vectors == 1) {
12953 cpu = cpumask_first(cpu_present_mask);
12954 lpfc_assign_eq_map_info(phba, index, LPFC_CPU_FIRST_IRQ,
12955 cpu);
12956 } else {
12957 maskp = pci_irq_get_affinity(phba->pcidev, index);
12958
James Smartdcaa2132019-11-04 16:57:06 -080012959 /* Loop through all CPUs associated with vector index */
12960 for_each_cpu_and(cpu, maskp, cpu_present_mask) {
Dick Kennedy17105d92020-07-06 13:42:30 -070012961 cpup = &phba->sli4_hba.cpu_map[cpu];
12962
James Smartdcaa2132019-11-04 16:57:06 -080012963 /* If this is the first CPU thats assigned to
12964 * this vector, set LPFC_CPU_FIRST_IRQ.
Dick Kennedy17105d92020-07-06 13:42:30 -070012965 *
12966 * With certain platforms its possible that irq
12967 * vectors are affinitized to all the cpu's.
12968 * This can result in each cpu_map.eq to be set
12969 * to the last vector, resulting in overwrite
12970 * of all the previous cpu_map.eq. Ensure that
12971 * each vector receives a place in cpu_map.
12972 * Later call to lpfc_cpu_affinity_check will
12973 * ensure we are nicely balanced out.
James Smartdcaa2132019-11-04 16:57:06 -080012974 */
Dick Kennedy17105d92020-07-06 13:42:30 -070012975 if (cpup->eq != LPFC_VECTOR_MAP_EMPTY)
12976 continue;
James Smartdcaa2132019-11-04 16:57:06 -080012977 lpfc_assign_eq_map_info(phba, index,
Dick Kennedy17105d92020-07-06 13:42:30 -070012978 LPFC_CPU_FIRST_IRQ,
James Smartdcaa2132019-11-04 16:57:06 -080012979 cpu);
Dick Kennedy17105d92020-07-06 13:42:30 -070012980 break;
James Smartdcaa2132019-11-04 16:57:06 -080012981 }
12982 }
James Smartda0436e2009-05-22 14:51:39 -040012983 }
12984
James Smart6a828b02019-01-28 11:14:31 -080012985 if (vectors != phba->cfg_irq_chann) {
Dick Kennedy372c1872020-06-30 14:50:00 -070012986 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart82c3e9b2012-09-29 11:29:50 -040012987 "3238 Reducing IO channels to match number of "
12988 "MSI-X vectors, requested %d got %d\n",
James Smart6a828b02019-01-28 11:14:31 -080012989 phba->cfg_irq_chann, vectors);
12990 if (phba->cfg_irq_chann > vectors)
12991 phba->cfg_irq_chann = vectors;
James Smart82c3e9b2012-09-29 11:29:50 -040012992 }
James Smart7bb03bb2013-04-17 20:19:16 -040012993
James Smartda0436e2009-05-22 14:51:39 -040012994 return rc;
12995
12996cfg_fail_out:
12997 /* free the irq already requested */
James Smartdcaa2132019-11-04 16:57:06 -080012998 for (--index; index >= 0; index--) {
12999 eqhdl = lpfc_get_eq_hdl(index);
13000 lpfc_irq_clear_aff(eqhdl);
13001 irq_set_affinity_hint(eqhdl->irq, NULL);
13002 free_irq(eqhdl->irq, eqhdl);
13003 }
James Smartda0436e2009-05-22 14:51:39 -040013004
James Smartda0436e2009-05-22 14:51:39 -040013005 /* Unconfigure MSI-X capability structure */
Christoph Hellwig45ffac12017-02-12 13:52:26 -080013006 pci_free_irq_vectors(phba->pcidev);
Alexander Gordeev029165a2014-07-16 20:05:15 +020013007
13008vec_fail_out:
James Smartda0436e2009-05-22 14:51:39 -040013009 return rc;
13010}
13011
13012/**
James Smartda0436e2009-05-22 14:51:39 -040013013 * lpfc_sli4_enable_msi - Enable MSI interrupt mode to SLI-4 device
13014 * @phba: pointer to lpfc hba data structure.
13015 *
13016 * This routine is invoked to enable the MSI interrupt mode to device with
James Smart07b1b912019-08-14 16:56:58 -070013017 * SLI-4 interface spec. The kernel function pci_alloc_irq_vectors() is
13018 * called to enable the MSI vector. The device driver is responsible for
13019 * calling the request_irq() to register MSI vector with a interrupt the
13020 * handler, which is done in this function.
James Smartda0436e2009-05-22 14:51:39 -040013021 *
13022 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -020013023 * 0 - successful
James Smartda0436e2009-05-22 14:51:39 -040013024 * other values - error
13025 **/
13026static int
13027lpfc_sli4_enable_msi(struct lpfc_hba *phba)
13028{
13029 int rc, index;
James Smartdcaa2132019-11-04 16:57:06 -080013030 unsigned int cpu;
13031 struct lpfc_hba_eq_hdl *eqhdl;
James Smartda0436e2009-05-22 14:51:39 -040013032
James Smart07b1b912019-08-14 16:56:58 -070013033 rc = pci_alloc_irq_vectors(phba->pcidev, 1, 1,
13034 PCI_IRQ_MSI | PCI_IRQ_AFFINITY);
13035 if (rc > 0)
James Smartda0436e2009-05-22 14:51:39 -040013036 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
13037 "0487 PCI enable MSI mode success.\n");
13038 else {
13039 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
13040 "0488 PCI enable MSI mode failed (%d)\n", rc);
James Smart07b1b912019-08-14 16:56:58 -070013041 return rc ? rc : -1;
James Smartda0436e2009-05-22 14:51:39 -040013042 }
13043
13044 rc = request_irq(phba->pcidev->irq, lpfc_sli4_intr_handler,
James Smarted243d32015-05-21 13:55:25 -040013045 0, LPFC_DRIVER_NAME, phba);
James Smartda0436e2009-05-22 14:51:39 -040013046 if (rc) {
James Smart07b1b912019-08-14 16:56:58 -070013047 pci_free_irq_vectors(phba->pcidev);
James Smartda0436e2009-05-22 14:51:39 -040013048 lpfc_printf_log(phba, KERN_WARNING, LOG_INIT,
13049 "0490 MSI request_irq failed (%d)\n", rc);
James Smart75baf692010-06-08 18:31:21 -040013050 return rc;
James Smartda0436e2009-05-22 14:51:39 -040013051 }
13052
James Smartdcaa2132019-11-04 16:57:06 -080013053 eqhdl = lpfc_get_eq_hdl(0);
13054 eqhdl->irq = pci_irq_vector(phba->pcidev, 0);
13055
13056 cpu = cpumask_first(cpu_present_mask);
13057 lpfc_assign_eq_map_info(phba, 0, LPFC_CPU_FIRST_IRQ, cpu);
13058
James Smart6a828b02019-01-28 11:14:31 -080013059 for (index = 0; index < phba->cfg_irq_chann; index++) {
James Smartdcaa2132019-11-04 16:57:06 -080013060 eqhdl = lpfc_get_eq_hdl(index);
13061 eqhdl->idx = index;
James Smartda0436e2009-05-22 14:51:39 -040013062 }
13063
James Smart75baf692010-06-08 18:31:21 -040013064 return 0;
James Smartda0436e2009-05-22 14:51:39 -040013065}
13066
13067/**
James Smartda0436e2009-05-22 14:51:39 -040013068 * lpfc_sli4_enable_intr - Enable device interrupt to SLI-4 device
13069 * @phba: pointer to lpfc hba data structure.
Lee Jonesfe614ac2020-07-23 13:24:22 +010013070 * @cfg_mode: Interrupt configuration mode (INTx, MSI or MSI-X).
James Smartda0436e2009-05-22 14:51:39 -040013071 *
13072 * This routine is invoked to enable device interrupt and associate driver's
13073 * interrupt handler(s) to interrupt vector(s) to device with SLI-4
13074 * interface spec. Depends on the interrupt mode configured to the driver,
13075 * the driver will try to fallback from the configured interrupt mode to an
13076 * interrupt mode which is supported by the platform, kernel, and device in
13077 * the order of:
13078 * MSI-X -> MSI -> IRQ.
13079 *
13080 * Return codes
André Goddard Rosaaf901ca2009-11-14 13:09:05 -020013081 * 0 - successful
James Smartda0436e2009-05-22 14:51:39 -040013082 * other values - error
13083 **/
13084static uint32_t
13085lpfc_sli4_enable_intr(struct lpfc_hba *phba, uint32_t cfg_mode)
13086{
13087 uint32_t intr_mode = LPFC_INTR_ERROR;
James Smart895427b2017-02-12 13:52:30 -080013088 int retval, idx;
James Smartda0436e2009-05-22 14:51:39 -040013089
13090 if (cfg_mode == 2) {
13091 /* Preparation before conf_msi mbox cmd */
13092 retval = 0;
13093 if (!retval) {
13094 /* Now, try to enable MSI-X interrupt mode */
13095 retval = lpfc_sli4_enable_msix(phba);
13096 if (!retval) {
13097 /* Indicate initialization to MSI-X mode */
13098 phba->intr_type = MSIX;
13099 intr_mode = 2;
13100 }
13101 }
13102 }
13103
13104 /* Fallback to MSI if MSI-X initialization failed */
13105 if (cfg_mode >= 1 && phba->intr_type == NONE) {
13106 retval = lpfc_sli4_enable_msi(phba);
13107 if (!retval) {
13108 /* Indicate initialization to MSI mode */
13109 phba->intr_type = MSI;
13110 intr_mode = 1;
13111 }
13112 }
13113
13114 /* Fallback to INTx if both MSI-X/MSI initalization failed */
13115 if (phba->intr_type == NONE) {
13116 retval = request_irq(phba->pcidev->irq, lpfc_sli4_intr_handler,
13117 IRQF_SHARED, LPFC_DRIVER_NAME, phba);
13118 if (!retval) {
James Smart895427b2017-02-12 13:52:30 -080013119 struct lpfc_hba_eq_hdl *eqhdl;
James Smartdcaa2132019-11-04 16:57:06 -080013120 unsigned int cpu;
James Smart895427b2017-02-12 13:52:30 -080013121
James Smartda0436e2009-05-22 14:51:39 -040013122 /* Indicate initialization to INTx mode */
13123 phba->intr_type = INTx;
13124 intr_mode = 0;
James Smart895427b2017-02-12 13:52:30 -080013125
James Smartdcaa2132019-11-04 16:57:06 -080013126 eqhdl = lpfc_get_eq_hdl(0);
13127 eqhdl->irq = pci_irq_vector(phba->pcidev, 0);
13128
13129 cpu = cpumask_first(cpu_present_mask);
13130 lpfc_assign_eq_map_info(phba, 0, LPFC_CPU_FIRST_IRQ,
13131 cpu);
James Smart6a828b02019-01-28 11:14:31 -080013132 for (idx = 0; idx < phba->cfg_irq_chann; idx++) {
James Smartdcaa2132019-11-04 16:57:06 -080013133 eqhdl = lpfc_get_eq_hdl(idx);
James Smart895427b2017-02-12 13:52:30 -080013134 eqhdl->idx = idx;
James Smart1ba981f2014-02-20 09:56:45 -050013135 }
James Smartda0436e2009-05-22 14:51:39 -040013136 }
13137 }
13138 return intr_mode;
13139}
13140
13141/**
13142 * lpfc_sli4_disable_intr - Disable device interrupt to SLI-4 device
13143 * @phba: pointer to lpfc hba data structure.
13144 *
13145 * This routine is invoked to disable device interrupt and disassociate
13146 * the driver's interrupt handler(s) from interrupt vector(s) to device
13147 * with SLI-4 interface spec. Depending on the interrupt mode, the driver
13148 * will release the interrupt vector(s) for the message signaled interrupt.
13149 **/
13150static void
13151lpfc_sli4_disable_intr(struct lpfc_hba *phba)
13152{
13153 /* Disable the currently initialized interrupt mode */
Christoph Hellwig45ffac12017-02-12 13:52:26 -080013154 if (phba->intr_type == MSIX) {
13155 int index;
James Smartdcaa2132019-11-04 16:57:06 -080013156 struct lpfc_hba_eq_hdl *eqhdl;
Christoph Hellwig45ffac12017-02-12 13:52:26 -080013157
13158 /* Free up MSI-X multi-message vectors */
James Smart6a828b02019-01-28 11:14:31 -080013159 for (index = 0; index < phba->cfg_irq_chann; index++) {
James Smartdcaa2132019-11-04 16:57:06 -080013160 eqhdl = lpfc_get_eq_hdl(index);
13161 lpfc_irq_clear_aff(eqhdl);
13162 irq_set_affinity_hint(eqhdl->irq, NULL);
13163 free_irq(eqhdl->irq, eqhdl);
James Smartb3295c22019-01-28 11:14:30 -080013164 }
Christoph Hellwig45ffac12017-02-12 13:52:26 -080013165 } else {
James Smartda0436e2009-05-22 14:51:39 -040013166 free_irq(phba->pcidev->irq, phba);
Christoph Hellwig45ffac12017-02-12 13:52:26 -080013167 }
13168
13169 pci_free_irq_vectors(phba->pcidev);
James Smartda0436e2009-05-22 14:51:39 -040013170
13171 /* Reset interrupt management states */
13172 phba->intr_type = NONE;
13173 phba->sli.slistat.sli_intr = 0;
James Smartda0436e2009-05-22 14:51:39 -040013174}
13175
13176/**
James Smart3772a992009-05-22 14:50:54 -040013177 * lpfc_unset_hba - Unset SLI3 hba device initialization
13178 * @phba: pointer to lpfc hba data structure.
13179 *
13180 * This routine is invoked to unset the HBA device initialization steps to
13181 * a device with SLI-3 interface spec.
13182 **/
13183static void
13184lpfc_unset_hba(struct lpfc_hba *phba)
13185{
13186 struct lpfc_vport *vport = phba->pport;
13187 struct Scsi_Host *shost = lpfc_shost_from_vport(vport);
13188
13189 spin_lock_irq(shost->host_lock);
13190 vport->load_flag |= FC_UNLOADING;
13191 spin_unlock_irq(shost->host_lock);
13192
James Smart72859902012-01-18 16:25:38 -050013193 kfree(phba->vpi_bmask);
13194 kfree(phba->vpi_ids);
13195
James Smart3772a992009-05-22 14:50:54 -040013196 lpfc_stop_hba_timers(phba);
13197
13198 phba->pport->work_port_events = 0;
13199
13200 lpfc_sli_hba_down(phba);
13201
13202 lpfc_sli_brdrestart(phba);
13203
13204 lpfc_sli_disable_intr(phba);
13205
13206 return;
13207}
13208
13209/**
James Smart5af5eee2010-10-22 11:06:38 -040013210 * lpfc_sli4_xri_exchange_busy_wait - Wait for device XRI exchange busy
13211 * @phba: Pointer to HBA context object.
13212 *
13213 * This function is called in the SLI4 code path to wait for completion
13214 * of device's XRIs exchange busy. It will check the XRI exchange busy
13215 * on outstanding FCP and ELS I/Os every 10ms for up to 10 seconds; after
13216 * that, it will check the XRI exchange busy on outstanding FCP and ELS
13217 * I/Os every 30 seconds, log error message, and wait forever. Only when
13218 * all XRI exchange busy complete, the driver unload shall proceed with
13219 * invoking the function reset ioctl mailbox command to the CNA and the
13220 * the rest of the driver unload resource release.
13221 **/
13222static void
13223lpfc_sli4_xri_exchange_busy_wait(struct lpfc_hba *phba)
13224{
James Smart5e5b5112019-01-28 11:14:22 -080013225 struct lpfc_sli4_hdw_queue *qp;
James Smartc00f62e2019-08-14 16:57:11 -070013226 int idx, ccnt;
James Smart5af5eee2010-10-22 11:06:38 -040013227 int wait_time = 0;
James Smart5e5b5112019-01-28 11:14:22 -080013228 int io_xri_cmpl = 1;
James Smart86c67372017-04-21 16:05:04 -070013229 int nvmet_xri_cmpl = 1;
James Smart5af5eee2010-10-22 11:06:38 -040013230 int els_xri_cmpl = list_empty(&phba->sli4_hba.lpfc_abts_els_sgl_list);
13231
James Smartc3725bd2017-11-20 16:00:42 -080013232 /* Driver just aborted IOs during the hba_unset process. Pause
13233 * here to give the HBA time to complete the IO and get entries
13234 * into the abts lists.
13235 */
13236 msleep(LPFC_XRI_EXCH_BUSY_WAIT_T1 * 5);
13237
13238 /* Wait for NVME pending IO to flush back to transport. */
13239 if (phba->cfg_enable_fc4_type & LPFC_ENABLE_NVME)
13240 lpfc_nvme_wait_for_io_drain(phba);
13241
James Smart5e5b5112019-01-28 11:14:22 -080013242 ccnt = 0;
James Smart5e5b5112019-01-28 11:14:22 -080013243 for (idx = 0; idx < phba->cfg_hdw_queue; idx++) {
13244 qp = &phba->sli4_hba.hdwq[idx];
James Smartc00f62e2019-08-14 16:57:11 -070013245 io_xri_cmpl = list_empty(&qp->lpfc_abts_io_buf_list);
13246 if (!io_xri_cmpl) /* if list is NOT empty */
13247 ccnt++;
James Smart5e5b5112019-01-28 11:14:22 -080013248 }
13249 if (ccnt)
13250 io_xri_cmpl = 0;
James Smart5e5b5112019-01-28 11:14:22 -080013251
James Smart86c67372017-04-21 16:05:04 -070013252 if (phba->cfg_enable_fc4_type & LPFC_ENABLE_NVME) {
James Smart86c67372017-04-21 16:05:04 -070013253 nvmet_xri_cmpl =
13254 list_empty(&phba->sli4_hba.lpfc_abts_nvmet_ctx_list);
13255 }
James Smart895427b2017-02-12 13:52:30 -080013256
James Smartc00f62e2019-08-14 16:57:11 -070013257 while (!els_xri_cmpl || !io_xri_cmpl || !nvmet_xri_cmpl) {
James Smart5af5eee2010-10-22 11:06:38 -040013258 if (wait_time > LPFC_XRI_EXCH_BUSY_WAIT_TMO) {
James Smart68c9b552018-06-26 08:24:25 -070013259 if (!nvmet_xri_cmpl)
Dick Kennedy372c1872020-06-30 14:50:00 -070013260 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart68c9b552018-06-26 08:24:25 -070013261 "6424 NVMET XRI exchange busy "
13262 "wait time: %d seconds.\n",
13263 wait_time/1000);
James Smart5e5b5112019-01-28 11:14:22 -080013264 if (!io_xri_cmpl)
Dick Kennedy372c1872020-06-30 14:50:00 -070013265 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartc00f62e2019-08-14 16:57:11 -070013266 "6100 IO XRI exchange busy "
James Smart5af5eee2010-10-22 11:06:38 -040013267 "wait time: %d seconds.\n",
13268 wait_time/1000);
13269 if (!els_xri_cmpl)
Dick Kennedy372c1872020-06-30 14:50:00 -070013270 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart5af5eee2010-10-22 11:06:38 -040013271 "2878 ELS XRI exchange busy "
13272 "wait time: %d seconds.\n",
13273 wait_time/1000);
13274 msleep(LPFC_XRI_EXCH_BUSY_WAIT_T2);
13275 wait_time += LPFC_XRI_EXCH_BUSY_WAIT_T2;
13276 } else {
13277 msleep(LPFC_XRI_EXCH_BUSY_WAIT_T1);
13278 wait_time += LPFC_XRI_EXCH_BUSY_WAIT_T1;
13279 }
James Smart5e5b5112019-01-28 11:14:22 -080013280
13281 ccnt = 0;
James Smart5e5b5112019-01-28 11:14:22 -080013282 for (idx = 0; idx < phba->cfg_hdw_queue; idx++) {
13283 qp = &phba->sli4_hba.hdwq[idx];
James Smartc00f62e2019-08-14 16:57:11 -070013284 io_xri_cmpl = list_empty(
13285 &qp->lpfc_abts_io_buf_list);
13286 if (!io_xri_cmpl) /* if list is NOT empty */
13287 ccnt++;
James Smart5e5b5112019-01-28 11:14:22 -080013288 }
13289 if (ccnt)
13290 io_xri_cmpl = 0;
James Smart5e5b5112019-01-28 11:14:22 -080013291
James Smart86c67372017-04-21 16:05:04 -070013292 if (phba->cfg_enable_fc4_type & LPFC_ENABLE_NVME) {
James Smart86c67372017-04-21 16:05:04 -070013293 nvmet_xri_cmpl = list_empty(
13294 &phba->sli4_hba.lpfc_abts_nvmet_ctx_list);
13295 }
James Smart5af5eee2010-10-22 11:06:38 -040013296 els_xri_cmpl =
13297 list_empty(&phba->sli4_hba.lpfc_abts_els_sgl_list);
James Smartf358dd02017-02-12 13:52:34 -080013298
James Smart5af5eee2010-10-22 11:06:38 -040013299 }
13300}
13301
13302/**
James Smartda0436e2009-05-22 14:51:39 -040013303 * lpfc_sli4_hba_unset - Unset the fcoe hba
13304 * @phba: Pointer to HBA context object.
13305 *
13306 * This function is called in the SLI4 code path to reset the HBA's FCoE
13307 * function. The caller is not required to hold any lock. This routine
13308 * issues PCI function reset mailbox command to reset the FCoE function.
13309 * At the end of the function, it calls lpfc_hba_down_post function to
13310 * free any pending commands.
13311 **/
13312static void
13313lpfc_sli4_hba_unset(struct lpfc_hba *phba)
13314{
13315 int wait_cnt = 0;
13316 LPFC_MBOXQ_t *mboxq;
James Smart912e3ac2011-05-24 11:42:11 -040013317 struct pci_dev *pdev = phba->pcidev;
James Smartda0436e2009-05-22 14:51:39 -040013318
13319 lpfc_stop_hba_timers(phba);
James Smart02243832021-08-16 09:28:54 -070013320 hrtimer_cancel(&phba->cmf_timer);
13321
James Smartcdb42be2019-01-28 11:14:21 -080013322 if (phba->pport)
13323 phba->sli4_hba.intr_enable = 0;
James Smartda0436e2009-05-22 14:51:39 -040013324
13325 /*
13326 * Gracefully wait out the potential current outstanding asynchronous
13327 * mailbox command.
13328 */
13329
13330 /* First, block any pending async mailbox command from posted */
13331 spin_lock_irq(&phba->hbalock);
13332 phba->sli.sli_flag |= LPFC_SLI_ASYNC_MBX_BLK;
13333 spin_unlock_irq(&phba->hbalock);
13334 /* Now, trying to wait it out if we can */
13335 while (phba->sli.sli_flag & LPFC_SLI_MBOX_ACTIVE) {
13336 msleep(10);
13337 if (++wait_cnt > LPFC_ACTIVE_MBOX_WAIT_CNT)
13338 break;
13339 }
13340 /* Forcefully release the outstanding mailbox command if timed out */
13341 if (phba->sli.sli_flag & LPFC_SLI_MBOX_ACTIVE) {
13342 spin_lock_irq(&phba->hbalock);
13343 mboxq = phba->sli.mbox_active;
13344 mboxq->u.mb.mbxStatus = MBX_NOT_FINISHED;
13345 __lpfc_mbox_cmpl_put(phba, mboxq);
13346 phba->sli.sli_flag &= ~LPFC_SLI_MBOX_ACTIVE;
13347 phba->sli.mbox_active = NULL;
13348 spin_unlock_irq(&phba->hbalock);
13349 }
13350
James Smart5af5eee2010-10-22 11:06:38 -040013351 /* Abort all iocbs associated with the hba */
13352 lpfc_sli_hba_iocb_abort(phba);
13353
13354 /* Wait for completion of device XRI exchange busy */
13355 lpfc_sli4_xri_exchange_busy_wait(phba);
13356
James Smart93a4d6f2019-11-04 16:57:05 -080013357 /* per-phba callback de-registration for hotplug event */
SeongJae Park46da5472020-06-23 10:41:22 +020013358 if (phba->pport)
13359 lpfc_cpuhp_remove(phba);
James Smart93a4d6f2019-11-04 16:57:05 -080013360
James Smartda0436e2009-05-22 14:51:39 -040013361 /* Disable PCI subsystem interrupt */
13362 lpfc_sli4_disable_intr(phba);
13363
James Smart912e3ac2011-05-24 11:42:11 -040013364 /* Disable SR-IOV if enabled */
13365 if (phba->cfg_sriov_nr_virtfn)
13366 pci_disable_sriov(pdev);
13367
James Smartda0436e2009-05-22 14:51:39 -040013368 /* Stop kthread signal shall trigger work_done one more time */
13369 kthread_stop(phba->worker_thread);
13370
James Smartd2cc9bc2018-09-10 10:30:50 -070013371 /* Disable FW logging to host memory */
James Smart1165a5c2018-11-29 16:09:39 -080013372 lpfc_ras_stop_fwlog(phba);
James Smartd2cc9bc2018-09-10 10:30:50 -070013373
James Smartd1f525a2017-04-21 16:04:55 -070013374 /* Unset the queues shared with the hardware then release all
13375 * allocated resources.
13376 */
13377 lpfc_sli4_queue_unset(phba);
13378 lpfc_sli4_queue_destroy(phba);
13379
James Smart3677a3a2010-09-29 11:19:14 -040013380 /* Reset SLI4 HBA FCoE function */
13381 lpfc_pci_function_reset(phba);
13382
James Smart1165a5c2018-11-29 16:09:39 -080013383 /* Free RAS DMA memory */
13384 if (phba->ras_fwlog.ras_enabled)
13385 lpfc_sli4_ras_dma_free(phba);
13386
James Smartda0436e2009-05-22 14:51:39 -040013387 /* Stop the SLI4 device port */
James Smart1ffdd2c2019-03-04 15:27:51 -080013388 if (phba->pport)
13389 phba->pport->work_port_events = 0;
James Smartda0436e2009-05-22 14:51:39 -040013390}
13391
James Smart7481811c2021-08-16 09:28:56 -070013392static uint32_t
13393lpfc_cgn_crc32(uint32_t crc, u8 byte)
13394{
13395 uint32_t msb = 0;
13396 uint32_t bit;
13397
13398 for (bit = 0; bit < 8; bit++) {
13399 msb = (crc >> 31) & 1;
13400 crc <<= 1;
13401
13402 if (msb ^ (byte & 1)) {
13403 crc ^= LPFC_CGN_CRC32_MAGIC_NUMBER;
13404 crc |= 1;
13405 }
13406 byte >>= 1;
13407 }
13408 return crc;
13409}
13410
13411static uint32_t
13412lpfc_cgn_reverse_bits(uint32_t wd)
13413{
13414 uint32_t result = 0;
13415 uint32_t i;
13416
13417 for (i = 0; i < 32; i++) {
13418 result <<= 1;
13419 result |= (1 & (wd >> i));
13420 }
13421 return result;
13422}
13423
13424/*
13425 * The routine corresponds with the algorithm the HBA firmware
13426 * uses to validate the data integrity.
13427 */
13428uint32_t
13429lpfc_cgn_calc_crc32(void *ptr, uint32_t byteLen, uint32_t crc)
13430{
13431 uint32_t i;
13432 uint32_t result;
13433 uint8_t *data = (uint8_t *)ptr;
13434
13435 for (i = 0; i < byteLen; ++i)
13436 crc = lpfc_cgn_crc32(crc, data[i]);
13437
13438 result = ~lpfc_cgn_reverse_bits(crc);
13439 return result;
13440}
13441
James Smart8c42a652021-08-16 09:28:51 -070013442void
13443lpfc_init_congestion_buf(struct lpfc_hba *phba)
13444{
James Smart7481811c2021-08-16 09:28:56 -070013445 struct lpfc_cgn_info *cp;
13446 struct timespec64 cmpl_time;
13447 struct tm broken;
13448 uint16_t size;
13449 uint32_t crc;
13450
James Smart8c42a652021-08-16 09:28:51 -070013451 lpfc_printf_log(phba, KERN_INFO, LOG_CGN_MGMT,
13452 "6235 INIT Congestion Buffer %p\n", phba->cgn_i);
13453
13454 if (!phba->cgn_i)
13455 return;
James Smart7481811c2021-08-16 09:28:56 -070013456 cp = (struct lpfc_cgn_info *)phba->cgn_i->virt;
James Smart8c42a652021-08-16 09:28:51 -070013457
13458 atomic_set(&phba->cgn_fabric_warn_cnt, 0);
13459 atomic_set(&phba->cgn_fabric_alarm_cnt, 0);
13460 atomic_set(&phba->cgn_sync_alarm_cnt, 0);
13461 atomic_set(&phba->cgn_sync_warn_cnt, 0);
13462
James Smart8c42a652021-08-16 09:28:51 -070013463 atomic_set(&phba->cgn_driver_evt_cnt, 0);
13464 atomic_set(&phba->cgn_latency_evt_cnt, 0);
13465 atomic64_set(&phba->cgn_latency_evt, 0);
13466 phba->cgn_evt_minute = 0;
James Smart7481811c2021-08-16 09:28:56 -070013467 phba->hba_flag &= ~HBA_CGN_DAY_WRAP;
13468
13469 memset(cp, 0xff, LPFC_CGN_DATA_SIZE);
13470 cp->cgn_info_size = cpu_to_le16(LPFC_CGN_INFO_SZ);
13471 cp->cgn_info_version = LPFC_CGN_INFO_V3;
13472
13473 /* cgn parameters */
13474 cp->cgn_info_mode = phba->cgn_p.cgn_param_mode;
13475 cp->cgn_info_level0 = phba->cgn_p.cgn_param_level0;
13476 cp->cgn_info_level1 = phba->cgn_p.cgn_param_level1;
13477 cp->cgn_info_level2 = phba->cgn_p.cgn_param_level2;
13478
13479 ktime_get_real_ts64(&cmpl_time);
13480 time64_to_tm(cmpl_time.tv_sec, 0, &broken);
13481
13482 cp->cgn_info_month = broken.tm_mon + 1;
13483 cp->cgn_info_day = broken.tm_mday;
13484 cp->cgn_info_year = broken.tm_year - 100; /* relative to 2000 */
13485 cp->cgn_info_hour = broken.tm_hour;
13486 cp->cgn_info_minute = broken.tm_min;
13487 cp->cgn_info_second = broken.tm_sec;
13488
13489 lpfc_printf_log(phba, KERN_INFO, LOG_CGN_MGMT | LOG_INIT,
13490 "2643 CGNInfo Init: Start Time "
13491 "%d/%d/%d %d:%d:%d\n",
13492 cp->cgn_info_day, cp->cgn_info_month,
13493 cp->cgn_info_year, cp->cgn_info_hour,
13494 cp->cgn_info_minute, cp->cgn_info_second);
13495
13496 /* Fill in default LUN qdepth */
13497 if (phba->pport) {
13498 size = (uint16_t)(phba->pport->cfg_lun_queue_depth);
13499 cp->cgn_lunq = cpu_to_le16(size);
13500 }
13501
13502 /* last used Index initialized to 0xff already */
13503
James Smart59936432021-08-30 16:12:43 -070013504 cp->cgn_warn_freq = cpu_to_le16(LPFC_FPIN_INIT_FREQ);
13505 cp->cgn_alarm_freq = cpu_to_le16(LPFC_FPIN_INIT_FREQ);
James Smart7481811c2021-08-16 09:28:56 -070013506 crc = lpfc_cgn_calc_crc32(cp, LPFC_CGN_INFO_SZ, LPFC_CGN_CRC32_SEED);
13507 cp->cgn_info_crc = cpu_to_le32(crc);
James Smart8c42a652021-08-16 09:28:51 -070013508
13509 phba->cgn_evt_timestamp = jiffies +
13510 msecs_to_jiffies(LPFC_CGN_TIMER_TO_MIN);
13511}
13512
13513void
13514lpfc_init_congestion_stat(struct lpfc_hba *phba)
13515{
James Smart7481811c2021-08-16 09:28:56 -070013516 struct lpfc_cgn_info *cp;
13517 struct timespec64 cmpl_time;
13518 struct tm broken;
13519 uint32_t crc;
13520
James Smart8c42a652021-08-16 09:28:51 -070013521 lpfc_printf_log(phba, KERN_INFO, LOG_CGN_MGMT,
13522 "6236 INIT Congestion Stat %p\n", phba->cgn_i);
13523
13524 if (!phba->cgn_i)
13525 return;
James Smart7481811c2021-08-16 09:28:56 -070013526
13527 cp = (struct lpfc_cgn_info *)phba->cgn_i->virt;
13528 memset(&cp->cgn_stat_npm, 0, LPFC_CGN_STAT_SIZE);
13529
13530 ktime_get_real_ts64(&cmpl_time);
13531 time64_to_tm(cmpl_time.tv_sec, 0, &broken);
13532
13533 cp->cgn_stat_month = broken.tm_mon + 1;
13534 cp->cgn_stat_day = broken.tm_mday;
13535 cp->cgn_stat_year = broken.tm_year - 100; /* relative to 2000 */
13536 cp->cgn_stat_hour = broken.tm_hour;
13537 cp->cgn_stat_minute = broken.tm_min;
13538
13539 lpfc_printf_log(phba, KERN_INFO, LOG_CGN_MGMT | LOG_INIT,
13540 "2647 CGNstat Init: Start Time "
13541 "%d/%d/%d %d:%d\n",
13542 cp->cgn_stat_day, cp->cgn_stat_month,
13543 cp->cgn_stat_year, cp->cgn_stat_hour,
13544 cp->cgn_stat_minute);
13545
13546 crc = lpfc_cgn_calc_crc32(cp, LPFC_CGN_INFO_SZ, LPFC_CGN_CRC32_SEED);
13547 cp->cgn_info_crc = cpu_to_le32(crc);
James Smart8c42a652021-08-16 09:28:51 -070013548}
13549
13550/**
13551 * __lpfc_reg_congestion_buf - register congestion info buffer with HBA
13552 * @phba: Pointer to hba context object.
13553 * @reg: flag to determine register or unregister.
13554 */
13555static int
13556__lpfc_reg_congestion_buf(struct lpfc_hba *phba, int reg)
13557{
13558 struct lpfc_mbx_reg_congestion_buf *reg_congestion_buf;
13559 union lpfc_sli4_cfg_shdr *shdr;
13560 uint32_t shdr_status, shdr_add_status;
13561 LPFC_MBOXQ_t *mboxq;
13562 int length, rc;
13563
13564 if (!phba->cgn_i)
13565 return -ENXIO;
13566
13567 mboxq = mempool_alloc(phba->mbox_mem_pool, GFP_KERNEL);
13568 if (!mboxq) {
13569 lpfc_printf_log(phba, KERN_ERR, LOG_MBOX,
13570 "2641 REG_CONGESTION_BUF mbox allocation fail: "
13571 "HBA state x%x reg %d\n",
13572 phba->pport->port_state, reg);
13573 return -ENOMEM;
13574 }
13575
13576 length = (sizeof(struct lpfc_mbx_reg_congestion_buf) -
13577 sizeof(struct lpfc_sli4_cfg_mhdr));
13578 lpfc_sli4_config(phba, mboxq, LPFC_MBOX_SUBSYSTEM_COMMON,
13579 LPFC_MBOX_OPCODE_REG_CONGESTION_BUF, length,
13580 LPFC_SLI4_MBX_EMBED);
13581 reg_congestion_buf = &mboxq->u.mqe.un.reg_congestion_buf;
13582 bf_set(lpfc_mbx_reg_cgn_buf_type, reg_congestion_buf, 1);
13583 if (reg > 0)
13584 bf_set(lpfc_mbx_reg_cgn_buf_cnt, reg_congestion_buf, 1);
13585 else
13586 bf_set(lpfc_mbx_reg_cgn_buf_cnt, reg_congestion_buf, 0);
13587 reg_congestion_buf->length = sizeof(struct lpfc_cgn_info);
13588 reg_congestion_buf->addr_lo =
13589 putPaddrLow(phba->cgn_i->phys);
13590 reg_congestion_buf->addr_hi =
13591 putPaddrHigh(phba->cgn_i->phys);
13592
13593 rc = lpfc_sli_issue_mbox(phba, mboxq, MBX_POLL);
13594 shdr = (union lpfc_sli4_cfg_shdr *)
13595 &mboxq->u.mqe.un.sli4_config.header.cfg_shdr;
13596 shdr_status = bf_get(lpfc_mbox_hdr_status, &shdr->response);
13597 shdr_add_status = bf_get(lpfc_mbox_hdr_add_status,
13598 &shdr->response);
13599 mempool_free(mboxq, phba->mbox_mem_pool);
13600 if (shdr_status || shdr_add_status || rc) {
13601 lpfc_printf_log(phba, KERN_ERR, LOG_INIT,
13602 "2642 REG_CONGESTION_BUF mailbox "
13603 "failed with status x%x add_status x%x,"
13604 " mbx status x%x reg %d\n",
13605 shdr_status, shdr_add_status, rc, reg);
13606 return -ENXIO;
13607 }
13608 return 0;
13609}
13610
James Smart02243832021-08-16 09:28:54 -070013611int
James Smart8c42a652021-08-16 09:28:51 -070013612lpfc_unreg_congestion_buf(struct lpfc_hba *phba)
13613{
James Smart02243832021-08-16 09:28:54 -070013614 lpfc_cmf_stop(phba);
James Smart8c42a652021-08-16 09:28:51 -070013615 return __lpfc_reg_congestion_buf(phba, 0);
13616}
13617
13618int
13619lpfc_reg_congestion_buf(struct lpfc_hba *phba)
13620{
13621 return __lpfc_reg_congestion_buf(phba, 1);
13622}
13623
James Smartda0436e2009-05-22 14:51:39 -040013624/**
James Smartfedd3b72011-02-16 12:39:24 -050013625 * lpfc_get_sli4_parameters - Get the SLI4 Config PARAMETERS.
13626 * @phba: Pointer to HBA context object.
13627 * @mboxq: Pointer to the mailboxq memory for the mailbox command response.
13628 *
13629 * This function is called in the SLI4 code path to read the port's
13630 * sli4 capabilities.
13631 *
13632 * This function may be be called from any context that can block-wait
13633 * for the completion. The expectation is that this routine is called
13634 * typically from probe_one or from the online routine.
13635 **/
13636int
13637lpfc_get_sli4_parameters(struct lpfc_hba *phba, LPFC_MBOXQ_t *mboxq)
13638{
13639 int rc;
13640 struct lpfc_mqe *mqe = &mboxq->u.mqe;
13641 struct lpfc_pc_sli4_params *sli4_params;
James Smarta183a152011-10-10 21:32:43 -040013642 uint32_t mbox_tmo;
James Smartfedd3b72011-02-16 12:39:24 -050013643 int length;
James Smartbf316c72018-04-09 14:24:28 -070013644 bool exp_wqcq_pages = true;
James Smartfedd3b72011-02-16 12:39:24 -050013645 struct lpfc_sli4_parameters *mbx_sli4_parameters;
13646
James Smart6d368e52011-05-24 11:44:12 -040013647 /*
13648 * By default, the driver assumes the SLI4 port requires RPI
13649 * header postings. The SLI4_PARAM response will correct this
13650 * assumption.
13651 */
13652 phba->sli4_hba.rpi_hdrs_in_use = 1;
13653
James Smartfedd3b72011-02-16 12:39:24 -050013654 /* Read the port's SLI4 Config Parameters */
13655 length = (sizeof(struct lpfc_mbx_get_sli4_parameters) -
13656 sizeof(struct lpfc_sli4_cfg_mhdr));
13657 lpfc_sli4_config(phba, mboxq, LPFC_MBOX_SUBSYSTEM_COMMON,
13658 LPFC_MBOX_OPCODE_GET_SLI4_PARAMETERS,
13659 length, LPFC_SLI4_MBX_EMBED);
13660 if (!phba->sli4_hba.intr_enable)
13661 rc = lpfc_sli_issue_mbox(phba, mboxq, MBX_POLL);
James Smarta183a152011-10-10 21:32:43 -040013662 else {
13663 mbox_tmo = lpfc_mbox_tmo_val(phba, mboxq);
13664 rc = lpfc_sli_issue_mbox_wait(phba, mboxq, mbox_tmo);
13665 }
James Smartfedd3b72011-02-16 12:39:24 -050013666 if (unlikely(rc))
13667 return rc;
13668 sli4_params = &phba->sli4_hba.pc_sli4_params;
13669 mbx_sli4_parameters = &mqe->un.get_sli4_parameters.sli4_parameters;
13670 sli4_params->if_type = bf_get(cfg_if_type, mbx_sli4_parameters);
13671 sli4_params->sli_rev = bf_get(cfg_sli_rev, mbx_sli4_parameters);
13672 sli4_params->sli_family = bf_get(cfg_sli_family, mbx_sli4_parameters);
13673 sli4_params->featurelevel_1 = bf_get(cfg_sli_hint_1,
13674 mbx_sli4_parameters);
13675 sli4_params->featurelevel_2 = bf_get(cfg_sli_hint_2,
13676 mbx_sli4_parameters);
13677 if (bf_get(cfg_phwq, mbx_sli4_parameters))
13678 phba->sli3_options |= LPFC_SLI4_PHWQ_ENABLED;
13679 else
13680 phba->sli3_options &= ~LPFC_SLI4_PHWQ_ENABLED;
13681 sli4_params->sge_supp_len = mbx_sli4_parameters->sge_supp_len;
James Smartb62232b2021-04-11 18:31:22 -070013682 sli4_params->loopbk_scope = bf_get(cfg_loopbk_scope,
13683 mbx_sli4_parameters);
James Smart1ba981f2014-02-20 09:56:45 -050013684 sli4_params->oas_supported = bf_get(cfg_oas, mbx_sli4_parameters);
James Smartfedd3b72011-02-16 12:39:24 -050013685 sli4_params->cqv = bf_get(cfg_cqv, mbx_sli4_parameters);
13686 sli4_params->mqv = bf_get(cfg_mqv, mbx_sli4_parameters);
13687 sli4_params->wqv = bf_get(cfg_wqv, mbx_sli4_parameters);
13688 sli4_params->rqv = bf_get(cfg_rqv, mbx_sli4_parameters);
James Smart7365f6f2018-02-22 08:18:46 -080013689 sli4_params->eqav = bf_get(cfg_eqav, mbx_sli4_parameters);
13690 sli4_params->cqav = bf_get(cfg_cqav, mbx_sli4_parameters);
James Smart0c651872013-07-15 18:33:23 -040013691 sli4_params->wqsize = bf_get(cfg_wqsize, mbx_sli4_parameters);
James Smart66e9e6b2018-06-26 08:24:27 -070013692 sli4_params->bv1s = bf_get(cfg_bv1s, mbx_sli4_parameters);
James Smart83c6cb12019-10-18 14:18:30 -070013693 sli4_params->pls = bf_get(cfg_pvl, mbx_sli4_parameters);
James Smartfedd3b72011-02-16 12:39:24 -050013694 sli4_params->sgl_pages_max = bf_get(cfg_sgl_page_cnt,
13695 mbx_sli4_parameters);
James Smart895427b2017-02-12 13:52:30 -080013696 sli4_params->wqpcnt = bf_get(cfg_wqpcnt, mbx_sli4_parameters);
James Smartfedd3b72011-02-16 12:39:24 -050013697 sli4_params->sgl_pp_align = bf_get(cfg_sgl_pp_align,
13698 mbx_sli4_parameters);
James Smart6d368e52011-05-24 11:44:12 -040013699 phba->sli4_hba.extents_in_use = bf_get(cfg_ext, mbx_sli4_parameters);
13700 phba->sli4_hba.rpi_hdrs_in_use = bf_get(cfg_hdrr, mbx_sli4_parameters);
James Smart895427b2017-02-12 13:52:30 -080013701
James Smartd79c9e92019-08-14 16:57:09 -070013702 /* Check for Extended Pre-Registered SGL support */
13703 phba->cfg_xpsgl = bf_get(cfg_xpsgl, mbx_sli4_parameters);
13704
James Smartc15e0702019-05-21 17:49:02 -070013705 /* Check for firmware nvme support */
13706 rc = (bf_get(cfg_nvme, mbx_sli4_parameters) &&
13707 bf_get(cfg_xib, mbx_sli4_parameters));
James Smart895427b2017-02-12 13:52:30 -080013708
James Smartc15e0702019-05-21 17:49:02 -070013709 if (rc) {
13710 /* Save this to indicate the Firmware supports NVME */
13711 sli4_params->nvme = 1;
13712
13713 /* Firmware NVME support, check driver FC4 NVME support */
13714 if (phba->cfg_enable_fc4_type == LPFC_ENABLE_FCP) {
13715 lpfc_printf_log(phba, KERN_INFO, LOG_INIT | LOG_NVME,
13716 "6133 Disabling NVME support: "
13717 "FC4 type not supported: x%x\n",
13718 phba->cfg_enable_fc4_type);
13719 goto fcponly;
13720 }
13721 } else {
13722 /* No firmware NVME support, check driver FC4 NVME support */
13723 sli4_params->nvme = 0;
13724 if (phba->cfg_enable_fc4_type & LPFC_ENABLE_NVME) {
13725 lpfc_printf_log(phba, KERN_ERR, LOG_INIT | LOG_NVME,
13726 "6101 Disabling NVME support: Not "
13727 "supported by firmware (%d %d) x%x\n",
13728 bf_get(cfg_nvme, mbx_sli4_parameters),
13729 bf_get(cfg_xib, mbx_sli4_parameters),
13730 phba->cfg_enable_fc4_type);
13731fcponly:
James Smartc15e0702019-05-21 17:49:02 -070013732 phba->nvmet_support = 0;
13733 phba->cfg_nvmet_mrq = 0;
James Smart6a224b42019-08-14 16:56:57 -070013734 phba->cfg_nvme_seg_cnt = 0;
James Smartc15e0702019-05-21 17:49:02 -070013735
13736 /* If no FC4 type support, move to just SCSI support */
13737 if (!(phba->cfg_enable_fc4_type & LPFC_ENABLE_FCP))
13738 return -ENODEV;
13739 phba->cfg_enable_fc4_type = LPFC_ENABLE_FCP;
13740 }
James Smart895427b2017-02-12 13:52:30 -080013741 }
James Smart05580562011-05-24 11:40:48 -040013742
James Smartc26c2652019-08-14 16:56:49 -070013743 /* If the NVME FC4 type is enabled, scale the sg_seg_cnt to
James Smarta5f73372019-09-21 20:58:50 -070013744 * accommodate 512K and 1M IOs in a single nvme buf.
James Smartc26c2652019-08-14 16:56:49 -070013745 */
James Smarta5f73372019-09-21 20:58:50 -070013746 if (phba->cfg_enable_fc4_type & LPFC_ENABLE_NVME)
James Smartc26c2652019-08-14 16:56:49 -070013747 phba->cfg_sg_seg_cnt = LPFC_MAX_NVME_SEG_CNT;
James Smartc26c2652019-08-14 16:56:49 -070013748
James Smart137ddf02021-07-07 11:43:44 -070013749 /* Enable embedded Payload BDE if support is indicated */
13750 if (bf_get(cfg_pbde, mbx_sli4_parameters))
13751 phba->cfg_enable_pbde = 1;
13752 else
James Smart414abe02018-06-26 08:24:26 -070013753 phba->cfg_enable_pbde = 0;
James Smart0bc2b7c2018-02-22 08:18:48 -080013754
James Smart20aefac2018-01-30 15:58:58 -080013755 /*
13756 * To support Suppress Response feature we must satisfy 3 conditions.
13757 * lpfc_suppress_rsp module parameter must be set (default).
13758 * In SLI4-Parameters Descriptor:
13759 * Extended Inline Buffers (XIB) must be supported.
13760 * Suppress Response IU Not Supported (SRIUNS) must NOT be supported
13761 * (double negative).
13762 */
13763 if (phba->cfg_suppress_rsp && bf_get(cfg_xib, mbx_sli4_parameters) &&
13764 !(bf_get(cfg_nosr, mbx_sli4_parameters)))
James Smartf358dd02017-02-12 13:52:34 -080013765 phba->sli.sli_flag |= LPFC_SLI_SUPPRESS_RSP;
James Smart20aefac2018-01-30 15:58:58 -080013766 else
13767 phba->cfg_suppress_rsp = 0;
James Smartf358dd02017-02-12 13:52:34 -080013768
James Smart0cf07f842017-06-01 21:07:10 -070013769 if (bf_get(cfg_eqdr, mbx_sli4_parameters))
13770 phba->sli.sli_flag |= LPFC_SLI_USE_EQDR;
13771
James Smart05580562011-05-24 11:40:48 -040013772 /* Make sure that sge_supp_len can be handled by the driver */
13773 if (sli4_params->sge_supp_len > LPFC_MAX_SGE_SIZE)
13774 sli4_params->sge_supp_len = LPFC_MAX_SGE_SIZE;
13775
James Smartb5c53952016-03-31 14:12:30 -070013776 /*
James Smartc176ffa2018-01-30 15:58:46 -080013777 * Check whether the adapter supports an embedded copy of the
13778 * FCP CMD IU within the WQE for FCP_Ixxx commands. In order
13779 * to use this option, 128-byte WQEs must be used.
James Smartb5c53952016-03-31 14:12:30 -070013780 */
13781 if (bf_get(cfg_ext_embed_cb, mbx_sli4_parameters))
13782 phba->fcp_embed_io = 1;
13783 else
13784 phba->fcp_embed_io = 0;
James Smart7bdedb32016-07-06 12:36:00 -070013785
James Smart0bc2b7c2018-02-22 08:18:48 -080013786 lpfc_printf_log(phba, KERN_INFO, LOG_INIT | LOG_NVME,
James Smart414abe02018-06-26 08:24:26 -070013787 "6422 XIB %d PBDE %d: FCP %d NVME %d %d %d\n",
James Smart0bc2b7c2018-02-22 08:18:48 -080013788 bf_get(cfg_xib, mbx_sli4_parameters),
James Smart414abe02018-06-26 08:24:26 -070013789 phba->cfg_enable_pbde,
James Smartae463b62021-07-07 11:43:32 -070013790 phba->fcp_embed_io, sli4_params->nvme,
James Smart4e565cf2018-02-22 08:18:50 -080013791 phba->cfg_nvme_embed_cmd, phba->cfg_suppress_rsp);
James Smart0bc2b7c2018-02-22 08:18:48 -080013792
James Smartbf316c72018-04-09 14:24:28 -070013793 if ((bf_get(lpfc_sli_intf_if_type, &phba->sli4_hba.sli_intf) ==
13794 LPFC_SLI_INTF_IF_TYPE_2) &&
13795 (bf_get(lpfc_sli_intf_sli_family, &phba->sli4_hba.sli_intf) ==
James Smartc2217682018-05-24 21:09:00 -070013796 LPFC_SLI_INTF_FAMILY_LNCR_A0))
James Smartbf316c72018-04-09 14:24:28 -070013797 exp_wqcq_pages = false;
13798
James Smartc176ffa2018-01-30 15:58:46 -080013799 if ((bf_get(cfg_cqpsize, mbx_sli4_parameters) & LPFC_CQ_16K_PAGE_SZ) &&
13800 (bf_get(cfg_wqpsize, mbx_sli4_parameters) & LPFC_WQ_16K_PAGE_SZ) &&
James Smartbf316c72018-04-09 14:24:28 -070013801 exp_wqcq_pages &&
James Smartc176ffa2018-01-30 15:58:46 -080013802 (sli4_params->wqsize & LPFC_WQ_SZ128_SUPPORT))
13803 phba->enab_exp_wqcq_pages = 1;
13804 else
13805 phba->enab_exp_wqcq_pages = 0;
James Smart7bdedb32016-07-06 12:36:00 -070013806 /*
13807 * Check if the SLI port supports MDS Diagnostics
13808 */
13809 if (bf_get(cfg_mds_diags, mbx_sli4_parameters))
13810 phba->mds_diags_support = 1;
13811 else
13812 phba->mds_diags_support = 0;
James Smartd2cc9bc2018-09-10 10:30:50 -070013813
James Smart0d8af092019-08-14 16:57:10 -070013814 /*
13815 * Check if the SLI port supports NSLER
13816 */
13817 if (bf_get(cfg_nsler, mbx_sli4_parameters))
13818 phba->nsler = 1;
13819 else
13820 phba->nsler = 0;
13821
James Smartfedd3b72011-02-16 12:39:24 -050013822 return 0;
13823}
13824
13825/**
James Smart3772a992009-05-22 14:50:54 -040013826 * lpfc_pci_probe_one_s3 - PCI probe func to reg SLI-3 device to PCI subsystem.
13827 * @pdev: pointer to PCI device
13828 * @pid: pointer to PCI device identifier
13829 *
13830 * This routine is to be called to attach a device with SLI-3 interface spec
13831 * to the PCI subsystem. When an Emulex HBA with SLI-3 interface spec is
13832 * presented on PCI bus, the kernel PCI subsystem looks at PCI device-specific
13833 * information of the device and driver to see if the driver state that it can
13834 * support this kind of device. If the match is successful, the driver core
13835 * invokes this routine. If this routine determines it can claim the HBA, it
13836 * does all the initialization that it needs to do to handle the HBA properly.
13837 *
13838 * Return code
13839 * 0 - driver can claim the device
13840 * negative value - driver can not claim the device
13841 **/
Greg Kroah-Hartman6f039792012-12-21 13:08:55 -080013842static int
James Smart3772a992009-05-22 14:50:54 -040013843lpfc_pci_probe_one_s3(struct pci_dev *pdev, const struct pci_device_id *pid)
13844{
13845 struct lpfc_hba *phba;
13846 struct lpfc_vport *vport = NULL;
James Smart6669f9b2009-10-02 15:16:45 -040013847 struct Scsi_Host *shost = NULL;
James Smart3772a992009-05-22 14:50:54 -040013848 int error;
13849 uint32_t cfg_mode, intr_mode;
13850
13851 /* Allocate memory for HBA structure */
13852 phba = lpfc_hba_alloc(pdev);
13853 if (!phba)
13854 return -ENOMEM;
13855
13856 /* Perform generic PCI device enabling operation */
13857 error = lpfc_enable_pci_dev(phba);
James Smart079b5c92011-08-21 21:48:49 -040013858 if (error)
James Smart3772a992009-05-22 14:50:54 -040013859 goto out_free_phba;
James Smart3772a992009-05-22 14:50:54 -040013860
13861 /* Set up SLI API function jump table for PCI-device group-0 HBAs */
13862 error = lpfc_api_table_setup(phba, LPFC_PCI_DEV_LP);
13863 if (error)
13864 goto out_disable_pci_dev;
13865
13866 /* Set up SLI-3 specific device PCI memory space */
13867 error = lpfc_sli_pci_mem_setup(phba);
13868 if (error) {
13869 lpfc_printf_log(phba, KERN_ERR, LOG_INIT,
13870 "1402 Failed to set up pci memory space.\n");
13871 goto out_disable_pci_dev;
13872 }
13873
James Smart3772a992009-05-22 14:50:54 -040013874 /* Set up SLI-3 specific device driver resources */
13875 error = lpfc_sli_driver_resource_setup(phba);
13876 if (error) {
13877 lpfc_printf_log(phba, KERN_ERR, LOG_INIT,
13878 "1404 Failed to set up driver resource.\n");
13879 goto out_unset_pci_mem_s3;
13880 }
13881
13882 /* Initialize and populate the iocb list per host */
James Smartd1f525a2017-04-21 16:04:55 -070013883
James Smart3772a992009-05-22 14:50:54 -040013884 error = lpfc_init_iocb_list(phba, LPFC_IOCB_LIST_CNT);
13885 if (error) {
13886 lpfc_printf_log(phba, KERN_ERR, LOG_INIT,
13887 "1405 Failed to initialize iocb list.\n");
13888 goto out_unset_driver_resource_s3;
13889 }
13890
13891 /* Set up common device driver resources */
13892 error = lpfc_setup_driver_resource_phase2(phba);
13893 if (error) {
13894 lpfc_printf_log(phba, KERN_ERR, LOG_INIT,
13895 "1406 Failed to set up driver resource.\n");
13896 goto out_free_iocb_list;
13897 }
13898
James Smart079b5c92011-08-21 21:48:49 -040013899 /* Get the default values for Model Name and Description */
13900 lpfc_get_hba_model_desc(phba, phba->ModelName, phba->ModelDesc);
13901
James Smart3772a992009-05-22 14:50:54 -040013902 /* Create SCSI host to the physical port */
13903 error = lpfc_create_shost(phba);
13904 if (error) {
13905 lpfc_printf_log(phba, KERN_ERR, LOG_INIT,
13906 "1407 Failed to create scsi host.\n");
13907 goto out_unset_driver_resource;
13908 }
13909
13910 /* Configure sysfs attributes */
13911 vport = phba->pport;
13912 error = lpfc_alloc_sysfs_attr(vport);
13913 if (error) {
13914 lpfc_printf_log(phba, KERN_ERR, LOG_INIT,
13915 "1476 Failed to allocate sysfs attr\n");
13916 goto out_destroy_shost;
13917 }
13918
James Smart6669f9b2009-10-02 15:16:45 -040013919 shost = lpfc_shost_from_vport(vport); /* save shost for error cleanup */
James Smart3772a992009-05-22 14:50:54 -040013920 /* Now, trying to enable interrupt and bring up the device */
13921 cfg_mode = phba->cfg_use_msi;
13922 while (true) {
13923 /* Put device to a known state before enabling interrupt */
13924 lpfc_stop_port(phba);
13925 /* Configure and enable interrupt */
13926 intr_mode = lpfc_sli_enable_intr(phba, cfg_mode);
13927 if (intr_mode == LPFC_INTR_ERROR) {
Dick Kennedy372c1872020-06-30 14:50:00 -070013928 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart3772a992009-05-22 14:50:54 -040013929 "0431 Failed to enable interrupt.\n");
13930 error = -ENODEV;
13931 goto out_free_sysfs_attr;
13932 }
13933 /* SLI-3 HBA setup */
13934 if (lpfc_sli_hba_setup(phba)) {
Dick Kennedy372c1872020-06-30 14:50:00 -070013935 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart3772a992009-05-22 14:50:54 -040013936 "1477 Failed to set up hba\n");
13937 error = -ENODEV;
13938 goto out_remove_device;
13939 }
13940
13941 /* Wait 50ms for the interrupts of previous mailbox commands */
13942 msleep(50);
13943 /* Check active interrupts on message signaled interrupts */
13944 if (intr_mode == 0 ||
13945 phba->sli.slistat.sli_intr > LPFC_MSIX_VECTORS) {
13946 /* Log the current active interrupt mode */
13947 phba->intr_mode = intr_mode;
13948 lpfc_log_intr_mode(phba, intr_mode);
13949 break;
13950 } else {
13951 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
13952 "0447 Configure interrupt mode (%d) "
13953 "failed active interrupt test.\n",
13954 intr_mode);
13955 /* Disable the current interrupt mode */
13956 lpfc_sli_disable_intr(phba);
13957 /* Try next level of interrupt mode */
13958 cfg_mode = --intr_mode;
13959 }
13960 }
13961
13962 /* Perform post initialization setup */
13963 lpfc_post_init_setup(phba);
13964
13965 /* Check if there are static vports to be created. */
13966 lpfc_create_static_vport(phba);
13967
13968 return 0;
13969
13970out_remove_device:
13971 lpfc_unset_hba(phba);
13972out_free_sysfs_attr:
13973 lpfc_free_sysfs_attr(vport);
13974out_destroy_shost:
13975 lpfc_destroy_shost(phba);
13976out_unset_driver_resource:
13977 lpfc_unset_driver_resource_phase2(phba);
13978out_free_iocb_list:
13979 lpfc_free_iocb_list(phba);
13980out_unset_driver_resource_s3:
13981 lpfc_sli_driver_resource_unset(phba);
13982out_unset_pci_mem_s3:
13983 lpfc_sli_pci_mem_unset(phba);
13984out_disable_pci_dev:
13985 lpfc_disable_pci_dev(phba);
James Smart6669f9b2009-10-02 15:16:45 -040013986 if (shost)
13987 scsi_host_put(shost);
James Smart3772a992009-05-22 14:50:54 -040013988out_free_phba:
13989 lpfc_hba_free(phba);
13990 return error;
13991}
13992
13993/**
13994 * lpfc_pci_remove_one_s3 - PCI func to unreg SLI-3 device from PCI subsystem.
James Smarte59058c2008-08-24 21:49:00 -040013995 * @pdev: pointer to PCI device
13996 *
James Smart3772a992009-05-22 14:50:54 -040013997 * This routine is to be called to disattach a device with SLI-3 interface
13998 * spec from PCI subsystem. When an Emulex HBA with SLI-3 interface spec is
13999 * removed from PCI bus, it performs all the necessary cleanup for the HBA
14000 * device to be removed from the PCI subsystem properly.
James Smarte59058c2008-08-24 21:49:00 -040014001 **/
Greg Kroah-Hartman6f039792012-12-21 13:08:55 -080014002static void
James Smart3772a992009-05-22 14:50:54 -040014003lpfc_pci_remove_one_s3(struct pci_dev *pdev)
dea31012005-04-17 16:05:31 -050014004{
James Smart2e0fef82007-06-17 19:56:36 -050014005 struct Scsi_Host *shost = pci_get_drvdata(pdev);
14006 struct lpfc_vport *vport = (struct lpfc_vport *) shost->hostdata;
James Smarteada2722008-12-04 22:39:13 -050014007 struct lpfc_vport **vports;
James Smart2e0fef82007-06-17 19:56:36 -050014008 struct lpfc_hba *phba = vport->phba;
James Smarteada2722008-12-04 22:39:13 -050014009 int i;
Tomohiro Kusumi8a4df1202008-01-11 01:53:00 -050014010
James Smart549e55c2007-08-02 11:09:51 -040014011 spin_lock_irq(&phba->hbalock);
James Smart51ef4c22007-08-02 11:10:31 -040014012 vport->load_flag |= FC_UNLOADING;
James Smart549e55c2007-08-02 11:09:51 -040014013 spin_unlock_irq(&phba->hbalock);
dea31012005-04-17 16:05:31 -050014014
James Smart858c9f62007-06-17 19:56:39 -050014015 lpfc_free_sysfs_attr(vport);
14016
James Smarteada2722008-12-04 22:39:13 -050014017 /* Release all the vports against this physical port */
14018 vports = lpfc_create_vport_work_array(phba);
14019 if (vports != NULL)
James Smart587a37f2012-05-09 21:16:03 -040014020 for (i = 0; i <= phba->max_vports && vports[i] != NULL; i++) {
14021 if (vports[i]->port_type == LPFC_PHYSICAL_PORT)
14022 continue;
James Smarteada2722008-12-04 22:39:13 -050014023 fc_vport_terminate(vports[i]->fc_vport);
James Smart587a37f2012-05-09 21:16:03 -040014024 }
James Smarteada2722008-12-04 22:39:13 -050014025 lpfc_destroy_vport_work_array(phba, vports);
14026
James Smart95f0ef82020-11-15 11:26:32 -080014027 /* Remove FC host with the physical port */
James Smart858c9f62007-06-17 19:56:39 -050014028 fc_remove_host(shost);
James Smarte9b11082020-11-15 11:26:33 -080014029 scsi_remove_host(shost);
James Smartd613b6a2017-02-12 13:52:37 -080014030
James Smart95f0ef82020-11-15 11:26:32 -080014031 /* Clean up all nodes, mailboxes and IOs. */
James Smart87af33f2007-10-27 13:37:43 -040014032 lpfc_cleanup(vport);
14033
James Smart2e0fef82007-06-17 19:56:36 -050014034 /*
14035 * Bring down the SLI Layer. This step disable all interrupts,
14036 * clears the rings, discards all mailbox commands, and resets
14037 * the HBA.
14038 */
James Smarta257bf92009-04-06 18:48:10 -040014039
Justin P. Mattock48e34d02010-12-30 15:07:58 -080014040 /* HBA interrupt will be disabled after this call */
James Smart2e0fef82007-06-17 19:56:36 -050014041 lpfc_sli_hba_down(phba);
James Smarta257bf92009-04-06 18:48:10 -040014042 /* Stop kthread signal shall trigger work_done one more time */
14043 kthread_stop(phba->worker_thread);
14044 /* Final cleanup of txcmplq and reset the HBA */
James Smart2e0fef82007-06-17 19:56:36 -050014045 lpfc_sli_brdrestart(phba);
14046
James Smart72859902012-01-18 16:25:38 -050014047 kfree(phba->vpi_bmask);
14048 kfree(phba->vpi_ids);
14049
James Smart3772a992009-05-22 14:50:54 -040014050 lpfc_stop_hba_timers(phba);
James Smart523128e2018-09-10 10:30:46 -070014051 spin_lock_irq(&phba->port_list_lock);
James Smart858c9f62007-06-17 19:56:39 -050014052 list_del_init(&vport->listentry);
James Smart523128e2018-09-10 10:30:46 -070014053 spin_unlock_irq(&phba->port_list_lock);
James Smart858c9f62007-06-17 19:56:39 -050014054
James Smart858c9f62007-06-17 19:56:39 -050014055 lpfc_debugfs_terminate(vport);
James Smart2e0fef82007-06-17 19:56:36 -050014056
James Smart912e3ac2011-05-24 11:42:11 -040014057 /* Disable SR-IOV if enabled */
14058 if (phba->cfg_sriov_nr_virtfn)
14059 pci_disable_sriov(pdev);
14060
James Smart5b75da22008-12-04 22:39:35 -050014061 /* Disable interrupt */
James Smart3772a992009-05-22 14:50:54 -040014062 lpfc_sli_disable_intr(phba);
dea31012005-04-17 16:05:31 -050014063
James Smart858c9f62007-06-17 19:56:39 -050014064 scsi_host_put(shost);
James Smart2e0fef82007-06-17 19:56:36 -050014065
14066 /*
14067 * Call scsi_free before mem_free since scsi bufs are released to their
14068 * corresponding pools here.
14069 */
14070 lpfc_scsi_free(phba);
James Smart0794d602019-01-28 11:14:19 -080014071 lpfc_free_iocb_list(phba);
14072
James Smart3772a992009-05-22 14:50:54 -040014073 lpfc_mem_free_all(phba);
James Smart2e0fef82007-06-17 19:56:36 -050014074
James Smart34b02dc2008-08-24 21:49:55 -040014075 dma_free_coherent(&pdev->dev, lpfc_sli_hbq_size(),
14076 phba->hbqslimp.virt, phba->hbqslimp.phys);
James Smarted957682007-06-17 19:56:37 -050014077
James Smart2e0fef82007-06-17 19:56:36 -050014078 /* Free resources associated with SLI2 interface */
14079 dma_free_coherent(&pdev->dev, SLI2_SLIM_SIZE,
James Smart34b02dc2008-08-24 21:49:55 -040014080 phba->slim2p.virt, phba->slim2p.phys);
James Smart2e0fef82007-06-17 19:56:36 -050014081
14082 /* unmap adapter SLIM and Control Registers */
14083 iounmap(phba->ctrl_regs_memmap_p);
14084 iounmap(phba->slim_memmap_p);
14085
James Smart3772a992009-05-22 14:50:54 -040014086 lpfc_hba_free(phba);
James Smart2e0fef82007-06-17 19:56:36 -050014087
Johannes Thumshirne0c04832016-06-07 09:44:03 +020014088 pci_release_mem_regions(pdev);
James Smart2e0fef82007-06-17 19:56:36 -050014089 pci_disable_device(pdev);
dea31012005-04-17 16:05:31 -050014090}
14091
Linas Vepstas8d63f372007-02-14 14:28:36 -060014092/**
James Smart3772a992009-05-22 14:50:54 -040014093 * lpfc_pci_suspend_one_s3 - PCI func to suspend SLI-3 device for power mgmnt
Vaibhav Guptaef6fa162020-11-02 22:17:18 +053014094 * @dev_d: pointer to device
James Smart3a55b532008-12-04 22:38:54 -050014095 *
James Smart3772a992009-05-22 14:50:54 -040014096 * This routine is to be called from the kernel's PCI subsystem to support
14097 * system Power Management (PM) to device with SLI-3 interface spec. When
14098 * PM invokes this method, it quiesces the device by stopping the driver's
14099 * worker thread for the device, turning off device's interrupt and DMA,
14100 * and bring the device offline. Note that as the driver implements the
14101 * minimum PM requirements to a power-aware driver's PM support for the
14102 * suspend/resume -- all the possible PM messages (SUSPEND, HIBERNATE, FREEZE)
14103 * to the suspend() method call will be treated as SUSPEND and the driver will
14104 * fully reinitialize its device during resume() method call, the driver will
14105 * set device to PCI_D3hot state in PCI config space instead of setting it
14106 * according to the @msg provided by the PM.
James Smart3a55b532008-12-04 22:38:54 -050014107 *
14108 * Return code
James Smart3772a992009-05-22 14:50:54 -040014109 * 0 - driver suspended the device
14110 * Error otherwise
James Smart3a55b532008-12-04 22:38:54 -050014111 **/
Vaibhav Guptaef6fa162020-11-02 22:17:18 +053014112static int __maybe_unused
14113lpfc_pci_suspend_one_s3(struct device *dev_d)
James Smart3a55b532008-12-04 22:38:54 -050014114{
Vaibhav Guptaef6fa162020-11-02 22:17:18 +053014115 struct Scsi_Host *shost = dev_get_drvdata(dev_d);
James Smart3a55b532008-12-04 22:38:54 -050014116 struct lpfc_hba *phba = ((struct lpfc_vport *)shost->hostdata)->phba;
14117
14118 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
14119 "0473 PCI device Power Management suspend.\n");
14120
14121 /* Bring down the device */
James Smart618a5232012-06-12 13:54:36 -040014122 lpfc_offline_prep(phba, LPFC_MBX_WAIT);
James Smart3a55b532008-12-04 22:38:54 -050014123 lpfc_offline(phba);
14124 kthread_stop(phba->worker_thread);
14125
14126 /* Disable interrupt from device */
James Smart3772a992009-05-22 14:50:54 -040014127 lpfc_sli_disable_intr(phba);
James Smart3a55b532008-12-04 22:38:54 -050014128
James Smart3a55b532008-12-04 22:38:54 -050014129 return 0;
14130}
14131
14132/**
James Smart3772a992009-05-22 14:50:54 -040014133 * lpfc_pci_resume_one_s3 - PCI func to resume SLI-3 device for power mgmnt
Vaibhav Guptaef6fa162020-11-02 22:17:18 +053014134 * @dev_d: pointer to device
James Smart3a55b532008-12-04 22:38:54 -050014135 *
James Smart3772a992009-05-22 14:50:54 -040014136 * This routine is to be called from the kernel's PCI subsystem to support
14137 * system Power Management (PM) to device with SLI-3 interface spec. When PM
14138 * invokes this method, it restores the device's PCI config space state and
14139 * fully reinitializes the device and brings it online. Note that as the
14140 * driver implements the minimum PM requirements to a power-aware driver's
14141 * PM for suspend/resume -- all the possible PM messages (SUSPEND, HIBERNATE,
14142 * FREEZE) to the suspend() method call will be treated as SUSPEND and the
14143 * driver will fully reinitialize its device during resume() method call,
14144 * the device will be set to PCI_D0 directly in PCI config space before
14145 * restoring the state.
James Smart3a55b532008-12-04 22:38:54 -050014146 *
14147 * Return code
James Smart3772a992009-05-22 14:50:54 -040014148 * 0 - driver suspended the device
14149 * Error otherwise
James Smart3a55b532008-12-04 22:38:54 -050014150 **/
Vaibhav Guptaef6fa162020-11-02 22:17:18 +053014151static int __maybe_unused
14152lpfc_pci_resume_one_s3(struct device *dev_d)
James Smart3a55b532008-12-04 22:38:54 -050014153{
Vaibhav Guptaef6fa162020-11-02 22:17:18 +053014154 struct Scsi_Host *shost = dev_get_drvdata(dev_d);
James Smart3a55b532008-12-04 22:38:54 -050014155 struct lpfc_hba *phba = ((struct lpfc_vport *)shost->hostdata)->phba;
James Smart5b75da22008-12-04 22:39:35 -050014156 uint32_t intr_mode;
James Smart3a55b532008-12-04 22:38:54 -050014157 int error;
14158
14159 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
14160 "0452 PCI device Power Management resume.\n");
14161
James Smart3a55b532008-12-04 22:38:54 -050014162 /* Startup the kernel thread for this host adapter. */
14163 phba->worker_thread = kthread_run(lpfc_do_work, phba,
14164 "lpfc_worker_%d", phba->brd_no);
14165 if (IS_ERR(phba->worker_thread)) {
14166 error = PTR_ERR(phba->worker_thread);
14167 lpfc_printf_log(phba, KERN_ERR, LOG_INIT,
14168 "0434 PM resume failed to start worker "
14169 "thread: error=x%x.\n", error);
14170 return error;
14171 }
14172
James Smart25ac2c972021-09-10 16:31:54 -070014173 /* Init cpu_map array */
14174 lpfc_cpu_map_array_init(phba);
14175 /* Init hba_eq_hdl array */
14176 lpfc_hba_eq_hdl_array_init(phba);
James Smart5b75da22008-12-04 22:39:35 -050014177 /* Configure and enable interrupt */
James Smart3772a992009-05-22 14:50:54 -040014178 intr_mode = lpfc_sli_enable_intr(phba, phba->intr_mode);
James Smart5b75da22008-12-04 22:39:35 -050014179 if (intr_mode == LPFC_INTR_ERROR) {
Dick Kennedy372c1872020-06-30 14:50:00 -070014180 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart5b75da22008-12-04 22:39:35 -050014181 "0430 PM resume Failed to enable interrupt\n");
14182 return -EIO;
14183 } else
14184 phba->intr_mode = intr_mode;
James Smart3a55b532008-12-04 22:38:54 -050014185
14186 /* Restart HBA and bring it online */
14187 lpfc_sli_brdrestart(phba);
14188 lpfc_online(phba);
14189
James Smart5b75da22008-12-04 22:39:35 -050014190 /* Log the current active interrupt mode */
14191 lpfc_log_intr_mode(phba, phba->intr_mode);
14192
James Smart3a55b532008-12-04 22:38:54 -050014193 return 0;
14194}
14195
14196/**
James Smart891478a2009-11-18 15:40:23 -050014197 * lpfc_sli_prep_dev_for_recover - Prepare SLI3 device for pci slot recover
14198 * @phba: pointer to lpfc hba data structure.
14199 *
14200 * This routine is called to prepare the SLI3 device for PCI slot recover. It
James Smarte2af0d22010-03-15 11:25:32 -040014201 * aborts all the outstanding SCSI I/Os to the pci device.
James Smart891478a2009-11-18 15:40:23 -050014202 **/
14203static void
14204lpfc_sli_prep_dev_for_recover(struct lpfc_hba *phba)
14205{
Dick Kennedy372c1872020-06-30 14:50:00 -070014206 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart891478a2009-11-18 15:40:23 -050014207 "2723 PCI channel I/O abort preparing for recovery\n");
James Smarte2af0d22010-03-15 11:25:32 -040014208
14209 /*
14210 * There may be errored I/Os through HBA, abort all I/Os on txcmplq
14211 * and let the SCSI mid-layer to retry them to recover.
14212 */
James Smartdb55fba2014-04-04 13:52:02 -040014213 lpfc_sli_abort_fcp_rings(phba);
James Smart891478a2009-11-18 15:40:23 -050014214}
14215
14216/**
James Smart0d878412009-10-02 15:16:56 -040014217 * lpfc_sli_prep_dev_for_reset - Prepare SLI3 device for pci slot reset
14218 * @phba: pointer to lpfc hba data structure.
14219 *
14220 * This routine is called to prepare the SLI3 device for PCI slot reset. It
14221 * disables the device interrupt and pci device, and aborts the internal FCP
14222 * pending I/Os.
14223 **/
14224static void
14225lpfc_sli_prep_dev_for_reset(struct lpfc_hba *phba)
14226{
Dick Kennedy372c1872020-06-30 14:50:00 -070014227 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart891478a2009-11-18 15:40:23 -050014228 "2710 PCI channel disable preparing for reset\n");
James Smarte2af0d22010-03-15 11:25:32 -040014229
James Smart75baf692010-06-08 18:31:21 -040014230 /* Block any management I/Os to the device */
James Smart618a5232012-06-12 13:54:36 -040014231 lpfc_block_mgmt_io(phba, LPFC_MBX_WAIT);
James Smart75baf692010-06-08 18:31:21 -040014232
James Smarte2af0d22010-03-15 11:25:32 -040014233 /* Block all SCSI devices' I/Os on the host */
14234 lpfc_scsi_dev_block(phba);
14235
James Smartea714f32013-04-17 20:18:47 -040014236 /* Flush all driver's outstanding SCSI I/Os as we are to reset */
James Smartc00f62e2019-08-14 16:57:11 -070014237 lpfc_sli_flush_io_rings(phba);
James Smartea714f32013-04-17 20:18:47 -040014238
James Smarte2af0d22010-03-15 11:25:32 -040014239 /* stop all timers */
14240 lpfc_stop_hba_timers(phba);
14241
James Smart0d878412009-10-02 15:16:56 -040014242 /* Disable interrupt and pci device */
14243 lpfc_sli_disable_intr(phba);
14244 pci_disable_device(phba->pcidev);
James Smart0d878412009-10-02 15:16:56 -040014245}
14246
14247/**
14248 * lpfc_sli_prep_dev_for_perm_failure - Prepare SLI3 dev for pci slot disable
14249 * @phba: pointer to lpfc hba data structure.
14250 *
14251 * This routine is called to prepare the SLI3 device for PCI slot permanently
14252 * disabling. It blocks the SCSI transport layer traffic and flushes the FCP
14253 * pending I/Os.
14254 **/
14255static void
James Smart75baf692010-06-08 18:31:21 -040014256lpfc_sli_prep_dev_for_perm_failure(struct lpfc_hba *phba)
James Smart0d878412009-10-02 15:16:56 -040014257{
Dick Kennedy372c1872020-06-30 14:50:00 -070014258 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart891478a2009-11-18 15:40:23 -050014259 "2711 PCI channel permanent disable for failure\n");
James Smarte2af0d22010-03-15 11:25:32 -040014260 /* Block all SCSI devices' I/Os on the host */
14261 lpfc_scsi_dev_block(phba);
14262
14263 /* stop all timers */
14264 lpfc_stop_hba_timers(phba);
14265
James Smart0d878412009-10-02 15:16:56 -040014266 /* Clean up all driver's outstanding SCSI I/Os */
James Smartc00f62e2019-08-14 16:57:11 -070014267 lpfc_sli_flush_io_rings(phba);
James Smart0d878412009-10-02 15:16:56 -040014268}
14269
14270/**
James Smart3772a992009-05-22 14:50:54 -040014271 * lpfc_io_error_detected_s3 - Method for handling SLI-3 device PCI I/O error
James Smarte59058c2008-08-24 21:49:00 -040014272 * @pdev: pointer to PCI device.
14273 * @state: the current PCI connection state.
Linas Vepstas8d63f372007-02-14 14:28:36 -060014274 *
James Smart3772a992009-05-22 14:50:54 -040014275 * This routine is called from the PCI subsystem for I/O error handling to
14276 * device with SLI-3 interface spec. This function is called by the PCI
14277 * subsystem after a PCI bus error affecting this device has been detected.
14278 * When this function is invoked, it will need to stop all the I/Os and
14279 * interrupt(s) to the device. Once that is done, it will return
14280 * PCI_ERS_RESULT_NEED_RESET for the PCI subsystem to perform proper recovery
14281 * as desired.
James Smarte59058c2008-08-24 21:49:00 -040014282 *
14283 * Return codes
James Smart0d878412009-10-02 15:16:56 -040014284 * PCI_ERS_RESULT_CAN_RECOVER - can be recovered with reset_link
James Smart3772a992009-05-22 14:50:54 -040014285 * PCI_ERS_RESULT_NEED_RESET - need to reset before recovery
14286 * PCI_ERS_RESULT_DISCONNECT - device could not be recovered
James Smarte59058c2008-08-24 21:49:00 -040014287 **/
James Smart3772a992009-05-22 14:50:54 -040014288static pci_ers_result_t
14289lpfc_io_error_detected_s3(struct pci_dev *pdev, pci_channel_state_t state)
Linas Vepstas8d63f372007-02-14 14:28:36 -060014290{
James Smart51ef4c22007-08-02 11:10:31 -040014291 struct Scsi_Host *shost = pci_get_drvdata(pdev);
14292 struct lpfc_hba *phba = ((struct lpfc_vport *)shost->hostdata)->phba;
Linas Vepstas8d63f372007-02-14 14:28:36 -060014293
James Smart0d878412009-10-02 15:16:56 -040014294 switch (state) {
14295 case pci_channel_io_normal:
James Smart891478a2009-11-18 15:40:23 -050014296 /* Non-fatal error, prepare for recovery */
14297 lpfc_sli_prep_dev_for_recover(phba);
James Smart0d878412009-10-02 15:16:56 -040014298 return PCI_ERS_RESULT_CAN_RECOVER;
14299 case pci_channel_io_frozen:
14300 /* Fatal error, prepare for slot reset */
14301 lpfc_sli_prep_dev_for_reset(phba);
14302 return PCI_ERS_RESULT_NEED_RESET;
14303 case pci_channel_io_perm_failure:
14304 /* Permanent failure, prepare for device down */
James Smart75baf692010-06-08 18:31:21 -040014305 lpfc_sli_prep_dev_for_perm_failure(phba);
Linas Vepstas8d63f372007-02-14 14:28:36 -060014306 return PCI_ERS_RESULT_DISCONNECT;
James Smart0d878412009-10-02 15:16:56 -040014307 default:
14308 /* Unknown state, prepare and request slot reset */
Dick Kennedy372c1872020-06-30 14:50:00 -070014309 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart0d878412009-10-02 15:16:56 -040014310 "0472 Unknown PCI error state: x%x\n", state);
14311 lpfc_sli_prep_dev_for_reset(phba);
14312 return PCI_ERS_RESULT_NEED_RESET;
James Smarta8e497d2008-08-24 21:50:11 -040014313 }
Linas Vepstas8d63f372007-02-14 14:28:36 -060014314}
14315
14316/**
James Smart3772a992009-05-22 14:50:54 -040014317 * lpfc_io_slot_reset_s3 - Method for restarting PCI SLI-3 device from scratch.
James Smarte59058c2008-08-24 21:49:00 -040014318 * @pdev: pointer to PCI device.
Linas Vepstas8d63f372007-02-14 14:28:36 -060014319 *
James Smart3772a992009-05-22 14:50:54 -040014320 * This routine is called from the PCI subsystem for error handling to
14321 * device with SLI-3 interface spec. This is called after PCI bus has been
14322 * reset to restart the PCI card from scratch, as if from a cold-boot.
14323 * During the PCI subsystem error recovery, after driver returns
14324 * PCI_ERS_RESULT_NEED_RESET, the PCI subsystem will perform proper error
14325 * recovery and then call this routine before calling the .resume method
14326 * to recover the device. This function will initialize the HBA device,
14327 * enable the interrupt, but it will just put the HBA to offline state
14328 * without passing any I/O traffic.
James Smarte59058c2008-08-24 21:49:00 -040014329 *
14330 * Return codes
James Smart3772a992009-05-22 14:50:54 -040014331 * PCI_ERS_RESULT_RECOVERED - the device has been recovered
14332 * PCI_ERS_RESULT_DISCONNECT - device could not be recovered
Linas Vepstas8d63f372007-02-14 14:28:36 -060014333 */
James Smart3772a992009-05-22 14:50:54 -040014334static pci_ers_result_t
14335lpfc_io_slot_reset_s3(struct pci_dev *pdev)
Linas Vepstas8d63f372007-02-14 14:28:36 -060014336{
James Smart51ef4c22007-08-02 11:10:31 -040014337 struct Scsi_Host *shost = pci_get_drvdata(pdev);
14338 struct lpfc_hba *phba = ((struct lpfc_vport *)shost->hostdata)->phba;
Linas Vepstas8d63f372007-02-14 14:28:36 -060014339 struct lpfc_sli *psli = &phba->sli;
James Smart5b75da22008-12-04 22:39:35 -050014340 uint32_t intr_mode;
Linas Vepstas8d63f372007-02-14 14:28:36 -060014341
14342 dev_printk(KERN_INFO, &pdev->dev, "recovering from a slot reset.\n");
Benjamin Herrenschmidt09483912007-12-20 15:28:09 +110014343 if (pci_enable_device_mem(pdev)) {
Linas Vepstas8d63f372007-02-14 14:28:36 -060014344 printk(KERN_ERR "lpfc: Cannot re-enable "
14345 "PCI device after reset.\n");
14346 return PCI_ERS_RESULT_DISCONNECT;
14347 }
14348
James Smart97207482008-12-04 22:39:19 -050014349 pci_restore_state(pdev);
James Smart1dfb5a42010-02-12 14:40:50 -050014350
14351 /*
14352 * As the new kernel behavior of pci_restore_state() API call clears
14353 * device saved_state flag, need to save the restored state again.
14354 */
14355 pci_save_state(pdev);
14356
James Smart97207482008-12-04 22:39:19 -050014357 if (pdev->is_busmaster)
14358 pci_set_master(pdev);
Linas Vepstas8d63f372007-02-14 14:28:36 -060014359
James Smart92d7f7b2007-06-17 19:56:38 -050014360 spin_lock_irq(&phba->hbalock);
James Smart3772a992009-05-22 14:50:54 -040014361 psli->sli_flag &= ~LPFC_SLI_ACTIVE;
James Smart92d7f7b2007-06-17 19:56:38 -050014362 spin_unlock_irq(&phba->hbalock);
Linas Vepstas8d63f372007-02-14 14:28:36 -060014363
James Smart5b75da22008-12-04 22:39:35 -050014364 /* Configure and enable interrupt */
James Smart3772a992009-05-22 14:50:54 -040014365 intr_mode = lpfc_sli_enable_intr(phba, phba->intr_mode);
James Smart5b75da22008-12-04 22:39:35 -050014366 if (intr_mode == LPFC_INTR_ERROR) {
Dick Kennedy372c1872020-06-30 14:50:00 -070014367 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart5b75da22008-12-04 22:39:35 -050014368 "0427 Cannot re-enable interrupt after "
14369 "slot reset.\n");
14370 return PCI_ERS_RESULT_DISCONNECT;
14371 } else
14372 phba->intr_mode = intr_mode;
Linas Vepstas8d63f372007-02-14 14:28:36 -060014373
James Smart75baf692010-06-08 18:31:21 -040014374 /* Take device offline, it will perform cleanup */
James Smart618a5232012-06-12 13:54:36 -040014375 lpfc_offline_prep(phba, LPFC_MBX_WAIT);
Linas Vepstas8d63f372007-02-14 14:28:36 -060014376 lpfc_offline(phba);
14377 lpfc_sli_brdrestart(phba);
14378
James Smart5b75da22008-12-04 22:39:35 -050014379 /* Log the current active interrupt mode */
14380 lpfc_log_intr_mode(phba, phba->intr_mode);
14381
Linas Vepstas8d63f372007-02-14 14:28:36 -060014382 return PCI_ERS_RESULT_RECOVERED;
14383}
14384
14385/**
James Smart3772a992009-05-22 14:50:54 -040014386 * lpfc_io_resume_s3 - Method for resuming PCI I/O operation on SLI-3 device.
James Smarte59058c2008-08-24 21:49:00 -040014387 * @pdev: pointer to PCI device
Linas Vepstas8d63f372007-02-14 14:28:36 -060014388 *
James Smart3772a992009-05-22 14:50:54 -040014389 * This routine is called from the PCI subsystem for error handling to device
14390 * with SLI-3 interface spec. It is called when kernel error recovery tells
14391 * the lpfc driver that it is ok to resume normal PCI operation after PCI bus
14392 * error recovery. After this call, traffic can start to flow from this device
14393 * again.
Linas Vepstas8d63f372007-02-14 14:28:36 -060014394 */
James Smart3772a992009-05-22 14:50:54 -040014395static void
14396lpfc_io_resume_s3(struct pci_dev *pdev)
Linas Vepstas8d63f372007-02-14 14:28:36 -060014397{
James Smart51ef4c22007-08-02 11:10:31 -040014398 struct Scsi_Host *shost = pci_get_drvdata(pdev);
14399 struct lpfc_hba *phba = ((struct lpfc_vport *)shost->hostdata)->phba;
Linas Vepstas8d63f372007-02-14 14:28:36 -060014400
James Smarte2af0d22010-03-15 11:25:32 -040014401 /* Bring device online, it will be no-op for non-fatal error resume */
James Smart58da1ff2008-04-07 10:15:56 -040014402 lpfc_online(phba);
Linas Vepstas8d63f372007-02-14 14:28:36 -060014403}
14404
James Smart3772a992009-05-22 14:50:54 -040014405/**
James Smartda0436e2009-05-22 14:51:39 -040014406 * lpfc_sli4_get_els_iocb_cnt - Calculate the # of ELS IOCBs to reserve
14407 * @phba: pointer to lpfc hba data structure.
14408 *
14409 * returns the number of ELS/CT IOCBs to reserve
14410 **/
14411int
14412lpfc_sli4_get_els_iocb_cnt(struct lpfc_hba *phba)
14413{
14414 int max_xri = phba->sli4_hba.max_cfg_param.max_xri;
14415
James Smartf1126682009-06-10 17:22:44 -040014416 if (phba->sli_rev == LPFC_SLI_REV4) {
14417 if (max_xri <= 100)
James Smart6a9c52c2009-10-02 15:16:51 -040014418 return 10;
James Smartf1126682009-06-10 17:22:44 -040014419 else if (max_xri <= 256)
James Smart6a9c52c2009-10-02 15:16:51 -040014420 return 25;
James Smartf1126682009-06-10 17:22:44 -040014421 else if (max_xri <= 512)
James Smart6a9c52c2009-10-02 15:16:51 -040014422 return 50;
James Smartf1126682009-06-10 17:22:44 -040014423 else if (max_xri <= 1024)
James Smart6a9c52c2009-10-02 15:16:51 -040014424 return 100;
James Smart8a9d2e82012-05-09 21:16:12 -040014425 else if (max_xri <= 1536)
James Smart6a9c52c2009-10-02 15:16:51 -040014426 return 150;
James Smart8a9d2e82012-05-09 21:16:12 -040014427 else if (max_xri <= 2048)
14428 return 200;
14429 else
14430 return 250;
James Smartf1126682009-06-10 17:22:44 -040014431 } else
14432 return 0;
James Smartda0436e2009-05-22 14:51:39 -040014433}
14434
14435/**
James Smart895427b2017-02-12 13:52:30 -080014436 * lpfc_sli4_get_iocb_cnt - Calculate the # of total IOCBs to reserve
14437 * @phba: pointer to lpfc hba data structure.
14438 *
James Smartf358dd02017-02-12 13:52:34 -080014439 * returns the number of ELS/CT + NVMET IOCBs to reserve
James Smart895427b2017-02-12 13:52:30 -080014440 **/
14441int
14442lpfc_sli4_get_iocb_cnt(struct lpfc_hba *phba)
14443{
14444 int max_xri = lpfc_sli4_get_els_iocb_cnt(phba);
14445
James Smartf358dd02017-02-12 13:52:34 -080014446 if (phba->nvmet_support)
14447 max_xri += LPFC_NVMET_BUF_POST;
James Smart895427b2017-02-12 13:52:30 -080014448 return max_xri;
14449}
14450
14451
James Smart0a5ce732019-10-18 14:18:18 -070014452static int
James Smart1feb8202018-02-22 08:18:47 -080014453lpfc_log_write_firmware_error(struct lpfc_hba *phba, uint32_t offset,
14454 uint32_t magic_number, uint32_t ftype, uint32_t fid, uint32_t fsize,
14455 const struct firmware *fw)
14456{
James Smart0a5ce732019-10-18 14:18:18 -070014457 int rc;
James Smartf6c5e6c2021-07-22 15:17:18 -070014458 u8 sli_family;
James Smart1feb8202018-02-22 08:18:47 -080014459
James Smartf6c5e6c2021-07-22 15:17:18 -070014460 sli_family = bf_get(lpfc_sli_intf_sli_family, &phba->sli4_hba.sli_intf);
James Smart0a5ce732019-10-18 14:18:18 -070014461 /* Three cases: (1) FW was not supported on the detected adapter.
14462 * (2) FW update has been locked out administratively.
14463 * (3) Some other error during FW update.
14464 * In each case, an unmaskable message is written to the console
14465 * for admin diagnosis.
14466 */
14467 if (offset == ADD_STATUS_FW_NOT_SUPPORTED ||
James Smartf6c5e6c2021-07-22 15:17:18 -070014468 (sli_family == LPFC_SLI_INTF_FAMILY_G6 &&
James Smart5792a0e2019-10-25 11:43:42 -070014469 magic_number != MAGIC_NUMBER_G6) ||
James Smartf6c5e6c2021-07-22 15:17:18 -070014470 (sli_family == LPFC_SLI_INTF_FAMILY_G7 &&
14471 magic_number != MAGIC_NUMBER_G7) ||
14472 (sli_family == LPFC_SLI_INTF_FAMILY_G7P &&
14473 magic_number != MAGIC_NUMBER_G7P)) {
Dick Kennedy372c1872020-06-30 14:50:00 -070014474 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart0a5ce732019-10-18 14:18:18 -070014475 "3030 This firmware version is not supported on"
14476 " this HBA model. Device:%x Magic:%x Type:%x "
14477 "ID:%x Size %d %zd\n",
14478 phba->pcidev->device, magic_number, ftype, fid,
14479 fsize, fw->size);
14480 rc = -EINVAL;
14481 } else if (offset == ADD_STATUS_FW_DOWNLOAD_HW_DISABLED) {
Dick Kennedy372c1872020-06-30 14:50:00 -070014482 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart0a5ce732019-10-18 14:18:18 -070014483 "3021 Firmware downloads have been prohibited "
14484 "by a system configuration setting on "
14485 "Device:%x Magic:%x Type:%x ID:%x Size %d "
14486 "%zd\n",
14487 phba->pcidev->device, magic_number, ftype, fid,
14488 fsize, fw->size);
14489 rc = -EACCES;
14490 } else {
Dick Kennedy372c1872020-06-30 14:50:00 -070014491 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart0a5ce732019-10-18 14:18:18 -070014492 "3022 FW Download failed. Add Status x%x "
14493 "Device:%x Magic:%x Type:%x ID:%x Size %d "
14494 "%zd\n",
14495 offset, phba->pcidev->device, magic_number,
14496 ftype, fid, fsize, fw->size);
14497 rc = -EIO;
14498 }
14499 return rc;
James Smart1feb8202018-02-22 08:18:47 -080014500}
James Smart1feb8202018-02-22 08:18:47 -080014501
James Smart895427b2017-02-12 13:52:30 -080014502/**
James Smart52d52442011-05-24 11:42:45 -040014503 * lpfc_write_firmware - attempt to write a firmware image to the port
James Smart52d52442011-05-24 11:42:45 -040014504 * @fw: pointer to firmware image returned from request_firmware.
James Smart0a5ce732019-10-18 14:18:18 -070014505 * @context: pointer to firmware image returned from request_firmware.
James Smart52d52442011-05-24 11:42:45 -040014506 *
James Smart52d52442011-05-24 11:42:45 -040014507 **/
James Smartce396282012-09-29 11:30:56 -040014508static void
14509lpfc_write_firmware(const struct firmware *fw, void *context)
James Smart52d52442011-05-24 11:42:45 -040014510{
James Smartce396282012-09-29 11:30:56 -040014511 struct lpfc_hba *phba = (struct lpfc_hba *)context;
James Smart6b5151f2012-01-18 16:24:06 -050014512 char fwrev[FW_REV_STR_SIZE];
James Smartce396282012-09-29 11:30:56 -040014513 struct lpfc_grp_hdr *image;
James Smart52d52442011-05-24 11:42:45 -040014514 struct list_head dma_buffer_list;
14515 int i, rc = 0;
14516 struct lpfc_dmabuf *dmabuf, *next;
14517 uint32_t offset = 0, temp_offset = 0;
James Smart6b6ef5d2016-10-13 15:06:17 -070014518 uint32_t magic_number, ftype, fid, fsize;
James Smart52d52442011-05-24 11:42:45 -040014519
James Smartc71ab862012-10-31 14:44:33 -040014520 /* It can be null in no-wait mode, sanity check */
James Smartce396282012-09-29 11:30:56 -040014521 if (!fw) {
14522 rc = -ENXIO;
14523 goto out;
14524 }
14525 image = (struct lpfc_grp_hdr *)fw->data;
14526
James Smart6b6ef5d2016-10-13 15:06:17 -070014527 magic_number = be32_to_cpu(image->magic_number);
14528 ftype = bf_get_be32(lpfc_grp_hdr_file_type, image);
James Smart1feb8202018-02-22 08:18:47 -080014529 fid = bf_get_be32(lpfc_grp_hdr_id, image);
James Smart6b6ef5d2016-10-13 15:06:17 -070014530 fsize = be32_to_cpu(image->size);
14531
James Smart52d52442011-05-24 11:42:45 -040014532 INIT_LIST_HEAD(&dma_buffer_list);
James Smart52d52442011-05-24 11:42:45 -040014533 lpfc_decode_firmware_rev(phba, fwrev, 1);
James Smart88a2cfb2011-07-22 18:36:33 -040014534 if (strncmp(fwrev, image->revision, strnlen(image->revision, 16))) {
Dick Kennedy372c1872020-06-30 14:50:00 -070014535 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartce396282012-09-29 11:30:56 -040014536 "3023 Updating Firmware, Current Version:%s "
James Smart52d52442011-05-24 11:42:45 -040014537 "New Version:%s\n",
James Smart88a2cfb2011-07-22 18:36:33 -040014538 fwrev, image->revision);
James Smart52d52442011-05-24 11:42:45 -040014539 for (i = 0; i < LPFC_MBX_WR_CONFIG_MAX_BDE; i++) {
14540 dmabuf = kzalloc(sizeof(struct lpfc_dmabuf),
14541 GFP_KERNEL);
14542 if (!dmabuf) {
14543 rc = -ENOMEM;
James Smartce396282012-09-29 11:30:56 -040014544 goto release_out;
James Smart52d52442011-05-24 11:42:45 -040014545 }
14546 dmabuf->virt = dma_alloc_coherent(&phba->pcidev->dev,
14547 SLI4_PAGE_SIZE,
14548 &dmabuf->phys,
14549 GFP_KERNEL);
14550 if (!dmabuf->virt) {
14551 kfree(dmabuf);
14552 rc = -ENOMEM;
James Smartce396282012-09-29 11:30:56 -040014553 goto release_out;
James Smart52d52442011-05-24 11:42:45 -040014554 }
14555 list_add_tail(&dmabuf->list, &dma_buffer_list);
14556 }
14557 while (offset < fw->size) {
14558 temp_offset = offset;
14559 list_for_each_entry(dmabuf, &dma_buffer_list, list) {
James Smart079b5c92011-08-21 21:48:49 -040014560 if (temp_offset + SLI4_PAGE_SIZE > fw->size) {
James Smart52d52442011-05-24 11:42:45 -040014561 memcpy(dmabuf->virt,
14562 fw->data + temp_offset,
James Smart079b5c92011-08-21 21:48:49 -040014563 fw->size - temp_offset);
14564 temp_offset = fw->size;
James Smart52d52442011-05-24 11:42:45 -040014565 break;
14566 }
James Smart52d52442011-05-24 11:42:45 -040014567 memcpy(dmabuf->virt, fw->data + temp_offset,
14568 SLI4_PAGE_SIZE);
James Smart88a2cfb2011-07-22 18:36:33 -040014569 temp_offset += SLI4_PAGE_SIZE;
James Smart52d52442011-05-24 11:42:45 -040014570 }
14571 rc = lpfc_wr_object(phba, &dma_buffer_list,
14572 (fw->size - offset), &offset);
James Smart1feb8202018-02-22 08:18:47 -080014573 if (rc) {
James Smart0a5ce732019-10-18 14:18:18 -070014574 rc = lpfc_log_write_firmware_error(phba, offset,
14575 magic_number,
14576 ftype,
14577 fid,
14578 fsize,
14579 fw);
James Smartce396282012-09-29 11:30:56 -040014580 goto release_out;
James Smart1feb8202018-02-22 08:18:47 -080014581 }
James Smart52d52442011-05-24 11:42:45 -040014582 }
14583 rc = offset;
James Smart1feb8202018-02-22 08:18:47 -080014584 } else
Dick Kennedy372c1872020-06-30 14:50:00 -070014585 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart1feb8202018-02-22 08:18:47 -080014586 "3029 Skipped Firmware update, Current "
14587 "Version:%s New Version:%s\n",
14588 fwrev, image->revision);
James Smartce396282012-09-29 11:30:56 -040014589
14590release_out:
James Smart52d52442011-05-24 11:42:45 -040014591 list_for_each_entry_safe(dmabuf, next, &dma_buffer_list, list) {
14592 list_del(&dmabuf->list);
14593 dma_free_coherent(&phba->pcidev->dev, SLI4_PAGE_SIZE,
14594 dmabuf->virt, dmabuf->phys);
14595 kfree(dmabuf);
14596 }
James Smartce396282012-09-29 11:30:56 -040014597 release_firmware(fw);
14598out:
James Smart0a5ce732019-10-18 14:18:18 -070014599 if (rc < 0)
Dick Kennedy372c1872020-06-30 14:50:00 -070014600 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart0a5ce732019-10-18 14:18:18 -070014601 "3062 Firmware update error, status %d.\n", rc);
14602 else
Dick Kennedy372c1872020-06-30 14:50:00 -070014603 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart0a5ce732019-10-18 14:18:18 -070014604 "3024 Firmware update success: size %d.\n", rc);
James Smart52d52442011-05-24 11:42:45 -040014605}
14606
14607/**
James Smartc71ab862012-10-31 14:44:33 -040014608 * lpfc_sli4_request_firmware_update - Request linux generic firmware upgrade
14609 * @phba: pointer to lpfc hba data structure.
Lee Jonesfe614ac2020-07-23 13:24:22 +010014610 * @fw_upgrade: which firmware to update.
James Smartc71ab862012-10-31 14:44:33 -040014611 *
14612 * This routine is called to perform Linux generic firmware upgrade on device
14613 * that supports such feature.
14614 **/
14615int
14616lpfc_sli4_request_firmware_update(struct lpfc_hba *phba, uint8_t fw_upgrade)
14617{
14618 uint8_t file_name[ELX_MODEL_NAME_SIZE];
14619 int ret;
14620 const struct firmware *fw;
14621
14622 /* Only supported on SLI4 interface type 2 for now */
James Smart27d6ac02018-02-22 08:18:42 -080014623 if (bf_get(lpfc_sli_intf_if_type, &phba->sli4_hba.sli_intf) <
James Smartc71ab862012-10-31 14:44:33 -040014624 LPFC_SLI_INTF_IF_TYPE_2)
14625 return -EPERM;
14626
14627 snprintf(file_name, ELX_MODEL_NAME_SIZE, "%s.grp", phba->ModelName);
14628
14629 if (fw_upgrade == INT_FW_UPGRADE) {
Shawn Guo0733d832021-04-25 10:00:24 +080014630 ret = request_firmware_nowait(THIS_MODULE, FW_ACTION_UEVENT,
James Smartc71ab862012-10-31 14:44:33 -040014631 file_name, &phba->pcidev->dev,
14632 GFP_KERNEL, (void *)phba,
14633 lpfc_write_firmware);
14634 } else if (fw_upgrade == RUN_FW_UPGRADE) {
14635 ret = request_firmware(&fw, file_name, &phba->pcidev->dev);
14636 if (!ret)
14637 lpfc_write_firmware(fw, (void *)phba);
14638 } else {
14639 ret = -EINVAL;
14640 }
14641
14642 return ret;
14643}
14644
14645/**
James Smartda0436e2009-05-22 14:51:39 -040014646 * lpfc_pci_probe_one_s4 - PCI probe func to reg SLI-4 device to PCI subsys
14647 * @pdev: pointer to PCI device
14648 * @pid: pointer to PCI device identifier
14649 *
14650 * This routine is called from the kernel's PCI subsystem to device with
14651 * SLI-4 interface spec. When an Emulex HBA with SLI-4 interface spec is
14652 * presented on PCI bus, the kernel PCI subsystem looks at PCI device-specific
14653 * information of the device and driver to see if the driver state that it
14654 * can support this kind of device. If the match is successful, the driver
14655 * core invokes this routine. If this routine determines it can claim the HBA,
14656 * it does all the initialization that it needs to do to handle the HBA
14657 * properly.
14658 *
14659 * Return code
14660 * 0 - driver can claim the device
14661 * negative value - driver can not claim the device
14662 **/
Greg Kroah-Hartman6f039792012-12-21 13:08:55 -080014663static int
James Smartda0436e2009-05-22 14:51:39 -040014664lpfc_pci_probe_one_s4(struct pci_dev *pdev, const struct pci_device_id *pid)
14665{
14666 struct lpfc_hba *phba;
14667 struct lpfc_vport *vport = NULL;
James Smart6669f9b2009-10-02 15:16:45 -040014668 struct Scsi_Host *shost = NULL;
James Smart6c621a22017-05-15 15:20:45 -070014669 int error;
James Smartda0436e2009-05-22 14:51:39 -040014670 uint32_t cfg_mode, intr_mode;
James Smartda0436e2009-05-22 14:51:39 -040014671
14672 /* Allocate memory for HBA structure */
14673 phba = lpfc_hba_alloc(pdev);
14674 if (!phba)
14675 return -ENOMEM;
14676
Ewan D. Milne9977d882021-08-09 11:09:47 -040014677 INIT_LIST_HEAD(&phba->poll_list);
14678
James Smartda0436e2009-05-22 14:51:39 -040014679 /* Perform generic PCI device enabling operation */
14680 error = lpfc_enable_pci_dev(phba);
James Smart079b5c92011-08-21 21:48:49 -040014681 if (error)
James Smartda0436e2009-05-22 14:51:39 -040014682 goto out_free_phba;
James Smartda0436e2009-05-22 14:51:39 -040014683
14684 /* Set up SLI API function jump table for PCI-device group-1 HBAs */
14685 error = lpfc_api_table_setup(phba, LPFC_PCI_DEV_OC);
14686 if (error)
14687 goto out_disable_pci_dev;
14688
14689 /* Set up SLI-4 specific device PCI memory space */
14690 error = lpfc_sli4_pci_mem_setup(phba);
14691 if (error) {
14692 lpfc_printf_log(phba, KERN_ERR, LOG_INIT,
14693 "1410 Failed to set up pci memory space.\n");
14694 goto out_disable_pci_dev;
14695 }
14696
James Smartda0436e2009-05-22 14:51:39 -040014697 /* Set up SLI-4 Specific device driver resources */
14698 error = lpfc_sli4_driver_resource_setup(phba);
14699 if (error) {
14700 lpfc_printf_log(phba, KERN_ERR, LOG_INIT,
14701 "1412 Failed to set up driver resource.\n");
14702 goto out_unset_pci_mem_s4;
14703 }
14704
James Smart19ca7602010-11-20 23:11:55 -050014705 INIT_LIST_HEAD(&phba->active_rrq_list);
James Smart7d791df2011-07-22 18:37:52 -040014706 INIT_LIST_HEAD(&phba->fcf.fcf_pri_list);
James Smart19ca7602010-11-20 23:11:55 -050014707
James Smartda0436e2009-05-22 14:51:39 -040014708 /* Set up common device driver resources */
14709 error = lpfc_setup_driver_resource_phase2(phba);
14710 if (error) {
14711 lpfc_printf_log(phba, KERN_ERR, LOG_INIT,
14712 "1414 Failed to set up driver resource.\n");
James Smart6c621a22017-05-15 15:20:45 -070014713 goto out_unset_driver_resource_s4;
James Smartda0436e2009-05-22 14:51:39 -040014714 }
14715
James Smart079b5c92011-08-21 21:48:49 -040014716 /* Get the default values for Model Name and Description */
14717 lpfc_get_hba_model_desc(phba, phba->ModelName, phba->ModelDesc);
14718
James Smartda0436e2009-05-22 14:51:39 -040014719 /* Now, trying to enable interrupt and bring up the device */
14720 cfg_mode = phba->cfg_use_msi;
James Smartda0436e2009-05-22 14:51:39 -040014721
James Smart7b15db32013-01-03 15:43:29 -050014722 /* Put device to a known state before enabling interrupt */
James Smartcdb42be2019-01-28 11:14:21 -080014723 phba->pport = NULL;
James Smart7b15db32013-01-03 15:43:29 -050014724 lpfc_stop_port(phba);
James Smart895427b2017-02-12 13:52:30 -080014725
James Smartdcaa2132019-11-04 16:57:06 -080014726 /* Init cpu_map array */
14727 lpfc_cpu_map_array_init(phba);
14728
14729 /* Init hba_eq_hdl array */
14730 lpfc_hba_eq_hdl_array_init(phba);
14731
James Smart7b15db32013-01-03 15:43:29 -050014732 /* Configure and enable interrupt */
14733 intr_mode = lpfc_sli4_enable_intr(phba, cfg_mode);
14734 if (intr_mode == LPFC_INTR_ERROR) {
Dick Kennedy372c1872020-06-30 14:50:00 -070014735 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart7b15db32013-01-03 15:43:29 -050014736 "0426 Failed to enable interrupt.\n");
14737 error = -ENODEV;
James Smartcdb42be2019-01-28 11:14:21 -080014738 goto out_unset_driver_resource;
James Smartda0436e2009-05-22 14:51:39 -040014739 }
James Smart7b15db32013-01-03 15:43:29 -050014740 /* Default to single EQ for non-MSI-X */
James Smart895427b2017-02-12 13:52:30 -080014741 if (phba->intr_type != MSIX) {
James Smart6a828b02019-01-28 11:14:31 -080014742 phba->cfg_irq_chann = 1;
James Smart2d7dbc42017-02-12 13:52:35 -080014743 if (phba->cfg_enable_fc4_type & LPFC_ENABLE_NVME) {
James Smart2d7dbc42017-02-12 13:52:35 -080014744 if (phba->nvmet_support)
14745 phba->cfg_nvmet_mrq = 1;
14746 }
James Smartcdb42be2019-01-28 11:14:21 -080014747 }
James Smart6a828b02019-01-28 11:14:31 -080014748 lpfc_cpu_affinity_check(phba, phba->cfg_irq_chann);
James Smartcdb42be2019-01-28 11:14:21 -080014749
14750 /* Create SCSI host to the physical port */
14751 error = lpfc_create_shost(phba);
14752 if (error) {
14753 lpfc_printf_log(phba, KERN_ERR, LOG_INIT,
14754 "1415 Failed to create scsi host.\n");
14755 goto out_disable_intr;
14756 }
14757 vport = phba->pport;
14758 shost = lpfc_shost_from_vport(vport); /* save shost for error cleanup */
14759
14760 /* Configure sysfs attributes */
14761 error = lpfc_alloc_sysfs_attr(vport);
14762 if (error) {
14763 lpfc_printf_log(phba, KERN_ERR, LOG_INIT,
14764 "1416 Failed to allocate sysfs attr\n");
14765 goto out_destroy_shost;
James Smart895427b2017-02-12 13:52:30 -080014766 }
14767
James Smart7b15db32013-01-03 15:43:29 -050014768 /* Set up SLI-4 HBA */
14769 if (lpfc_sli4_hba_setup(phba)) {
Dick Kennedy372c1872020-06-30 14:50:00 -070014770 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart7b15db32013-01-03 15:43:29 -050014771 "1421 Failed to set up hba\n");
14772 error = -ENODEV;
James Smartcdb42be2019-01-28 11:14:21 -080014773 goto out_free_sysfs_attr;
James Smart7b15db32013-01-03 15:43:29 -050014774 }
14775
14776 /* Log the current active interrupt mode */
14777 phba->intr_mode = intr_mode;
14778 lpfc_log_intr_mode(phba, intr_mode);
James Smartda0436e2009-05-22 14:51:39 -040014779
14780 /* Perform post initialization setup */
14781 lpfc_post_init_setup(phba);
14782
James Smart01649562017-02-12 13:52:32 -080014783 /* NVME support in FW earlier in the driver load corrects the
14784 * FC4 type making a check for nvme_support unnecessary.
14785 */
James Smart0794d602019-01-28 11:14:19 -080014786 if (phba->nvmet_support == 0) {
14787 if (phba->cfg_enable_fc4_type & LPFC_ENABLE_NVME) {
14788 /* Create NVME binding with nvme_fc_transport. This
14789 * ensures the vport is initialized. If the localport
14790 * create fails, it should not unload the driver to
14791 * support field issues.
14792 */
14793 error = lpfc_nvme_create_localport(vport);
14794 if (error) {
Dick Kennedy372c1872020-06-30 14:50:00 -070014795 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart0794d602019-01-28 11:14:19 -080014796 "6004 NVME registration "
14797 "failed, error x%x\n",
14798 error);
14799 }
James Smart01649562017-02-12 13:52:32 -080014800 }
14801 }
James Smart895427b2017-02-12 13:52:30 -080014802
James Smartc71ab862012-10-31 14:44:33 -040014803 /* check for firmware upgrade or downgrade */
14804 if (phba->cfg_request_firmware_upgrade)
Sebastian Herbsztdb6f1c22015-08-31 16:48:14 -040014805 lpfc_sli4_request_firmware_update(phba, INT_FW_UPGRADE);
James Smart52d52442011-05-24 11:42:45 -040014806
James Smart1c6834a2009-07-19 10:01:26 -040014807 /* Check if there are static vports to be created. */
14808 lpfc_create_static_vport(phba);
James Smartd2cc9bc2018-09-10 10:30:50 -070014809
14810 /* Enable RAS FW log support */
14811 lpfc_sli4_ras_setup(phba);
14812
James Smartf861f592020-03-22 11:12:54 -070014813 timer_setup(&phba->cpuhp_poll_timer, lpfc_sli4_poll_hbtimer, 0);
James Smart93a4d6f2019-11-04 16:57:05 -080014814 cpuhp_state_add_instance_nocalls(lpfc_cpuhp_state, &phba->cpuhp);
14815
James Smartda0436e2009-05-22 14:51:39 -040014816 return 0;
14817
James Smartda0436e2009-05-22 14:51:39 -040014818out_free_sysfs_attr:
14819 lpfc_free_sysfs_attr(vport);
14820out_destroy_shost:
14821 lpfc_destroy_shost(phba);
James Smartcdb42be2019-01-28 11:14:21 -080014822out_disable_intr:
14823 lpfc_sli4_disable_intr(phba);
James Smartda0436e2009-05-22 14:51:39 -040014824out_unset_driver_resource:
14825 lpfc_unset_driver_resource_phase2(phba);
James Smartda0436e2009-05-22 14:51:39 -040014826out_unset_driver_resource_s4:
14827 lpfc_sli4_driver_resource_unset(phba);
14828out_unset_pci_mem_s4:
14829 lpfc_sli4_pci_mem_unset(phba);
14830out_disable_pci_dev:
14831 lpfc_disable_pci_dev(phba);
James Smart6669f9b2009-10-02 15:16:45 -040014832 if (shost)
14833 scsi_host_put(shost);
James Smartda0436e2009-05-22 14:51:39 -040014834out_free_phba:
14835 lpfc_hba_free(phba);
14836 return error;
14837}
14838
14839/**
14840 * lpfc_pci_remove_one_s4 - PCI func to unreg SLI-4 device from PCI subsystem
14841 * @pdev: pointer to PCI device
14842 *
14843 * This routine is called from the kernel's PCI subsystem to device with
14844 * SLI-4 interface spec. When an Emulex HBA with SLI-4 interface spec is
14845 * removed from PCI bus, it performs all the necessary cleanup for the HBA
14846 * device to be removed from the PCI subsystem properly.
14847 **/
Greg Kroah-Hartman6f039792012-12-21 13:08:55 -080014848static void
James Smartda0436e2009-05-22 14:51:39 -040014849lpfc_pci_remove_one_s4(struct pci_dev *pdev)
14850{
14851 struct Scsi_Host *shost = pci_get_drvdata(pdev);
14852 struct lpfc_vport *vport = (struct lpfc_vport *) shost->hostdata;
14853 struct lpfc_vport **vports;
14854 struct lpfc_hba *phba = vport->phba;
14855 int i;
14856
14857 /* Mark the device unloading flag */
14858 spin_lock_irq(&phba->hbalock);
14859 vport->load_flag |= FC_UNLOADING;
14860 spin_unlock_irq(&phba->hbalock);
James Smart02243832021-08-16 09:28:54 -070014861 if (phba->cgn_i)
14862 lpfc_unreg_congestion_buf(phba);
James Smartda0436e2009-05-22 14:51:39 -040014863
James Smartda0436e2009-05-22 14:51:39 -040014864 lpfc_free_sysfs_attr(vport);
14865
14866 /* Release all the vports against this physical port */
14867 vports = lpfc_create_vport_work_array(phba);
14868 if (vports != NULL)
James Smart587a37f2012-05-09 21:16:03 -040014869 for (i = 0; i <= phba->max_vports && vports[i] != NULL; i++) {
14870 if (vports[i]->port_type == LPFC_PHYSICAL_PORT)
14871 continue;
James Smartda0436e2009-05-22 14:51:39 -040014872 fc_vport_terminate(vports[i]->fc_vport);
James Smart587a37f2012-05-09 21:16:03 -040014873 }
James Smartda0436e2009-05-22 14:51:39 -040014874 lpfc_destroy_vport_work_array(phba, vports);
14875
James Smart95f0ef82020-11-15 11:26:32 -080014876 /* Remove FC host with the physical port */
James Smartda0436e2009-05-22 14:51:39 -040014877 fc_remove_host(shost);
James Smarte9b11082020-11-15 11:26:33 -080014878 scsi_remove_host(shost);
James Smartda0436e2009-05-22 14:51:39 -040014879
James Smartda0436e2009-05-22 14:51:39 -040014880 /* Perform ndlp cleanup on the physical port. The nvme and nvmet
14881 * localports are destroyed after to cleanup all transport memory.
14882 */
14883 lpfc_cleanup(vport);
14884 lpfc_nvmet_destroy_targetport(phba);
14885 lpfc_nvme_destroy_localport(vport);
14886
James Smartc4908502019-01-28 11:14:28 -080014887 /* De-allocate multi-XRI pools */
14888 if (phba->cfg_xri_rebalancing)
14889 lpfc_destroy_multixri_pools(phba);
14890
James Smart281d6192018-01-30 15:58:47 -080014891 /*
14892 * Bring down the SLI Layer. This step disables all interrupts,
14893 * clears the rings, discards all mailbox commands, and resets
14894 * the HBA FCoE function.
14895 */
14896 lpfc_debugfs_terminate(vport);
James Smartda0436e2009-05-22 14:51:39 -040014897
Dick Kennedy19017622017-09-29 17:34:27 -070014898 lpfc_stop_hba_timers(phba);
James Smart523128e2018-09-10 10:30:46 -070014899 spin_lock_irq(&phba->port_list_lock);
James Smartda0436e2009-05-22 14:51:39 -040014900 list_del_init(&vport->listentry);
James Smart523128e2018-09-10 10:30:46 -070014901 spin_unlock_irq(&phba->port_list_lock);
James Smartda0436e2009-05-22 14:51:39 -040014902
James Smart3677a3a2010-09-29 11:19:14 -040014903 /* Perform scsi free before driver resource_unset since scsi
James Smartda0436e2009-05-22 14:51:39 -040014904 * buffers are released to their corresponding pools here.
14905 */
James Smart5e5b5112019-01-28 11:14:22 -080014906 lpfc_io_free(phba);
James Smart01649562017-02-12 13:52:32 -080014907 lpfc_free_iocb_list(phba);
James Smart5e5b5112019-01-28 11:14:22 -080014908 lpfc_sli4_hba_unset(phba);
James Smart67d12732012-08-03 12:36:13 -040014909
James Smart0cdb84e2018-04-09 14:24:26 -070014910 lpfc_unset_driver_resource_phase2(phba);
James Smartda0436e2009-05-22 14:51:39 -040014911 lpfc_sli4_driver_resource_unset(phba);
14912
14913 /* Unmap adapter Control and Doorbell registers */
14914 lpfc_sli4_pci_mem_unset(phba);
14915
14916 /* Release PCI resources and disable device's PCI function */
14917 scsi_host_put(shost);
14918 lpfc_disable_pci_dev(phba);
14919
14920 /* Finally, free the driver's device data structure */
14921 lpfc_hba_free(phba);
14922
14923 return;
14924}
14925
14926/**
14927 * lpfc_pci_suspend_one_s4 - PCI func to suspend SLI-4 device for power mgmnt
Vaibhav Guptaef6fa162020-11-02 22:17:18 +053014928 * @dev_d: pointer to device
James Smartda0436e2009-05-22 14:51:39 -040014929 *
14930 * This routine is called from the kernel's PCI subsystem to support system
14931 * Power Management (PM) to device with SLI-4 interface spec. When PM invokes
14932 * this method, it quiesces the device by stopping the driver's worker
14933 * thread for the device, turning off device's interrupt and DMA, and bring
14934 * the device offline. Note that as the driver implements the minimum PM
14935 * requirements to a power-aware driver's PM support for suspend/resume -- all
14936 * the possible PM messages (SUSPEND, HIBERNATE, FREEZE) to the suspend()
14937 * method call will be treated as SUSPEND and the driver will fully
14938 * reinitialize its device during resume() method call, the driver will set
14939 * device to PCI_D3hot state in PCI config space instead of setting it
14940 * according to the @msg provided by the PM.
14941 *
14942 * Return code
14943 * 0 - driver suspended the device
14944 * Error otherwise
14945 **/
Vaibhav Guptaef6fa162020-11-02 22:17:18 +053014946static int __maybe_unused
14947lpfc_pci_suspend_one_s4(struct device *dev_d)
James Smartda0436e2009-05-22 14:51:39 -040014948{
Vaibhav Guptaef6fa162020-11-02 22:17:18 +053014949 struct Scsi_Host *shost = dev_get_drvdata(dev_d);
James Smartda0436e2009-05-22 14:51:39 -040014950 struct lpfc_hba *phba = ((struct lpfc_vport *)shost->hostdata)->phba;
14951
14952 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
James Smart75baf692010-06-08 18:31:21 -040014953 "2843 PCI device Power Management suspend.\n");
James Smartda0436e2009-05-22 14:51:39 -040014954
14955 /* Bring down the device */
James Smart618a5232012-06-12 13:54:36 -040014956 lpfc_offline_prep(phba, LPFC_MBX_WAIT);
James Smartda0436e2009-05-22 14:51:39 -040014957 lpfc_offline(phba);
14958 kthread_stop(phba->worker_thread);
14959
14960 /* Disable interrupt from device */
14961 lpfc_sli4_disable_intr(phba);
James Smart5350d872011-10-10 21:33:49 -040014962 lpfc_sli4_queue_destroy(phba);
James Smartda0436e2009-05-22 14:51:39 -040014963
James Smartda0436e2009-05-22 14:51:39 -040014964 return 0;
14965}
14966
14967/**
14968 * lpfc_pci_resume_one_s4 - PCI func to resume SLI-4 device for power mgmnt
Vaibhav Guptaef6fa162020-11-02 22:17:18 +053014969 * @dev_d: pointer to device
James Smartda0436e2009-05-22 14:51:39 -040014970 *
14971 * This routine is called from the kernel's PCI subsystem to support system
14972 * Power Management (PM) to device with SLI-4 interface spac. When PM invokes
14973 * this method, it restores the device's PCI config space state and fully
14974 * reinitializes the device and brings it online. Note that as the driver
14975 * implements the minimum PM requirements to a power-aware driver's PM for
14976 * suspend/resume -- all the possible PM messages (SUSPEND, HIBERNATE, FREEZE)
14977 * to the suspend() method call will be treated as SUSPEND and the driver
14978 * will fully reinitialize its device during resume() method call, the device
14979 * will be set to PCI_D0 directly in PCI config space before restoring the
14980 * state.
14981 *
14982 * Return code
14983 * 0 - driver suspended the device
14984 * Error otherwise
14985 **/
Vaibhav Guptaef6fa162020-11-02 22:17:18 +053014986static int __maybe_unused
14987lpfc_pci_resume_one_s4(struct device *dev_d)
James Smartda0436e2009-05-22 14:51:39 -040014988{
Vaibhav Guptaef6fa162020-11-02 22:17:18 +053014989 struct Scsi_Host *shost = dev_get_drvdata(dev_d);
James Smartda0436e2009-05-22 14:51:39 -040014990 struct lpfc_hba *phba = ((struct lpfc_vport *)shost->hostdata)->phba;
14991 uint32_t intr_mode;
14992 int error;
14993
14994 lpfc_printf_log(phba, KERN_INFO, LOG_INIT,
14995 "0292 PCI device Power Management resume.\n");
14996
James Smartda0436e2009-05-22 14:51:39 -040014997 /* Startup the kernel thread for this host adapter. */
14998 phba->worker_thread = kthread_run(lpfc_do_work, phba,
14999 "lpfc_worker_%d", phba->brd_no);
15000 if (IS_ERR(phba->worker_thread)) {
15001 error = PTR_ERR(phba->worker_thread);
15002 lpfc_printf_log(phba, KERN_ERR, LOG_INIT,
15003 "0293 PM resume failed to start worker "
15004 "thread: error=x%x.\n", error);
15005 return error;
15006 }
15007
15008 /* Configure and enable interrupt */
15009 intr_mode = lpfc_sli4_enable_intr(phba, phba->intr_mode);
15010 if (intr_mode == LPFC_INTR_ERROR) {
Dick Kennedy372c1872020-06-30 14:50:00 -070015011 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smartda0436e2009-05-22 14:51:39 -040015012 "0294 PM resume Failed to enable interrupt\n");
15013 return -EIO;
15014 } else
15015 phba->intr_mode = intr_mode;
15016
15017 /* Restart HBA and bring it online */
15018 lpfc_sli_brdrestart(phba);
15019 lpfc_online(phba);
15020
15021 /* Log the current active interrupt mode */
15022 lpfc_log_intr_mode(phba, phba->intr_mode);
15023
15024 return 0;
15025}
15026
15027/**
James Smart75baf692010-06-08 18:31:21 -040015028 * lpfc_sli4_prep_dev_for_recover - Prepare SLI4 device for pci slot recover
15029 * @phba: pointer to lpfc hba data structure.
15030 *
15031 * This routine is called to prepare the SLI4 device for PCI slot recover. It
15032 * aborts all the outstanding SCSI I/Os to the pci device.
15033 **/
15034static void
15035lpfc_sli4_prep_dev_for_recover(struct lpfc_hba *phba)
15036{
Dick Kennedy372c1872020-06-30 14:50:00 -070015037 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart75baf692010-06-08 18:31:21 -040015038 "2828 PCI channel I/O abort preparing for recovery\n");
15039 /*
15040 * There may be errored I/Os through HBA, abort all I/Os on txcmplq
15041 * and let the SCSI mid-layer to retry them to recover.
15042 */
James Smartdb55fba2014-04-04 13:52:02 -040015043 lpfc_sli_abort_fcp_rings(phba);
James Smart75baf692010-06-08 18:31:21 -040015044}
15045
15046/**
15047 * lpfc_sli4_prep_dev_for_reset - Prepare SLI4 device for pci slot reset
15048 * @phba: pointer to lpfc hba data structure.
15049 *
15050 * This routine is called to prepare the SLI4 device for PCI slot reset. It
15051 * disables the device interrupt and pci device, and aborts the internal FCP
15052 * pending I/Os.
15053 **/
15054static void
15055lpfc_sli4_prep_dev_for_reset(struct lpfc_hba *phba)
15056{
Dick Kennedy372c1872020-06-30 14:50:00 -070015057 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart75baf692010-06-08 18:31:21 -040015058 "2826 PCI channel disable preparing for reset\n");
15059
15060 /* Block any management I/Os to the device */
James Smart618a5232012-06-12 13:54:36 -040015061 lpfc_block_mgmt_io(phba, LPFC_MBX_NO_WAIT);
James Smart75baf692010-06-08 18:31:21 -040015062
15063 /* Block all SCSI devices' I/Os on the host */
15064 lpfc_scsi_dev_block(phba);
15065
James Smartc00f62e2019-08-14 16:57:11 -070015066 /* Flush all driver's outstanding I/Os as we are to reset */
15067 lpfc_sli_flush_io_rings(phba);
James Smartc3725bd2017-11-20 16:00:42 -080015068
James Smart75baf692010-06-08 18:31:21 -040015069 /* stop all timers */
15070 lpfc_stop_hba_timers(phba);
15071
15072 /* Disable interrupt and pci device */
15073 lpfc_sli4_disable_intr(phba);
James Smart5350d872011-10-10 21:33:49 -040015074 lpfc_sli4_queue_destroy(phba);
James Smart75baf692010-06-08 18:31:21 -040015075 pci_disable_device(phba->pcidev);
James Smart75baf692010-06-08 18:31:21 -040015076}
15077
15078/**
15079 * lpfc_sli4_prep_dev_for_perm_failure - Prepare SLI4 dev for pci slot disable
15080 * @phba: pointer to lpfc hba data structure.
15081 *
15082 * This routine is called to prepare the SLI4 device for PCI slot permanently
15083 * disabling. It blocks the SCSI transport layer traffic and flushes the FCP
15084 * pending I/Os.
15085 **/
15086static void
15087lpfc_sli4_prep_dev_for_perm_failure(struct lpfc_hba *phba)
15088{
Dick Kennedy372c1872020-06-30 14:50:00 -070015089 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart75baf692010-06-08 18:31:21 -040015090 "2827 PCI channel permanent disable for failure\n");
15091
15092 /* Block all SCSI devices' I/Os on the host */
15093 lpfc_scsi_dev_block(phba);
15094
15095 /* stop all timers */
15096 lpfc_stop_hba_timers(phba);
15097
James Smartc00f62e2019-08-14 16:57:11 -070015098 /* Clean up all driver's outstanding I/Os */
15099 lpfc_sli_flush_io_rings(phba);
James Smart75baf692010-06-08 18:31:21 -040015100}
15101
15102/**
James Smartda0436e2009-05-22 14:51:39 -040015103 * lpfc_io_error_detected_s4 - Method for handling PCI I/O error to SLI-4 device
15104 * @pdev: pointer to PCI device.
15105 * @state: the current PCI connection state.
15106 *
15107 * This routine is called from the PCI subsystem for error handling to device
15108 * with SLI-4 interface spec. This function is called by the PCI subsystem
15109 * after a PCI bus error affecting this device has been detected. When this
15110 * function is invoked, it will need to stop all the I/Os and interrupt(s)
15111 * to the device. Once that is done, it will return PCI_ERS_RESULT_NEED_RESET
15112 * for the PCI subsystem to perform proper recovery as desired.
15113 *
15114 * Return codes
15115 * PCI_ERS_RESULT_NEED_RESET - need to reset before recovery
15116 * PCI_ERS_RESULT_DISCONNECT - device could not be recovered
15117 **/
15118static pci_ers_result_t
15119lpfc_io_error_detected_s4(struct pci_dev *pdev, pci_channel_state_t state)
15120{
James Smart75baf692010-06-08 18:31:21 -040015121 struct Scsi_Host *shost = pci_get_drvdata(pdev);
15122 struct lpfc_hba *phba = ((struct lpfc_vport *)shost->hostdata)->phba;
15123
15124 switch (state) {
15125 case pci_channel_io_normal:
15126 /* Non-fatal error, prepare for recovery */
15127 lpfc_sli4_prep_dev_for_recover(phba);
15128 return PCI_ERS_RESULT_CAN_RECOVER;
15129 case pci_channel_io_frozen:
James Smart25ac2c972021-09-10 16:31:54 -070015130 phba->hba_flag |= HBA_PCI_ERR;
James Smart75baf692010-06-08 18:31:21 -040015131 /* Fatal error, prepare for slot reset */
15132 lpfc_sli4_prep_dev_for_reset(phba);
15133 return PCI_ERS_RESULT_NEED_RESET;
15134 case pci_channel_io_perm_failure:
James Smart25ac2c972021-09-10 16:31:54 -070015135 phba->hba_flag |= HBA_PCI_ERR;
James Smart75baf692010-06-08 18:31:21 -040015136 /* Permanent failure, prepare for device down */
15137 lpfc_sli4_prep_dev_for_perm_failure(phba);
15138 return PCI_ERS_RESULT_DISCONNECT;
15139 default:
James Smart25ac2c972021-09-10 16:31:54 -070015140 phba->hba_flag |= HBA_PCI_ERR;
James Smart75baf692010-06-08 18:31:21 -040015141 /* Unknown state, prepare and request slot reset */
Dick Kennedy372c1872020-06-30 14:50:00 -070015142 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart75baf692010-06-08 18:31:21 -040015143 "2825 Unknown PCI error state: x%x\n", state);
15144 lpfc_sli4_prep_dev_for_reset(phba);
15145 return PCI_ERS_RESULT_NEED_RESET;
15146 }
James Smartda0436e2009-05-22 14:51:39 -040015147}
15148
15149/**
15150 * lpfc_io_slot_reset_s4 - Method for restart PCI SLI-4 device from scratch
15151 * @pdev: pointer to PCI device.
15152 *
15153 * This routine is called from the PCI subsystem for error handling to device
15154 * with SLI-4 interface spec. It is called after PCI bus has been reset to
15155 * restart the PCI card from scratch, as if from a cold-boot. During the
15156 * PCI subsystem error recovery, after the driver returns
15157 * PCI_ERS_RESULT_NEED_RESET, the PCI subsystem will perform proper error
15158 * recovery and then call this routine before calling the .resume method to
15159 * recover the device. This function will initialize the HBA device, enable
15160 * the interrupt, but it will just put the HBA to offline state without
15161 * passing any I/O traffic.
15162 *
15163 * Return codes
15164 * PCI_ERS_RESULT_RECOVERED - the device has been recovered
15165 * PCI_ERS_RESULT_DISCONNECT - device could not be recovered
15166 */
15167static pci_ers_result_t
15168lpfc_io_slot_reset_s4(struct pci_dev *pdev)
15169{
James Smart75baf692010-06-08 18:31:21 -040015170 struct Scsi_Host *shost = pci_get_drvdata(pdev);
15171 struct lpfc_hba *phba = ((struct lpfc_vport *)shost->hostdata)->phba;
15172 struct lpfc_sli *psli = &phba->sli;
15173 uint32_t intr_mode;
15174
15175 dev_printk(KERN_INFO, &pdev->dev, "recovering from a slot reset.\n");
15176 if (pci_enable_device_mem(pdev)) {
15177 printk(KERN_ERR "lpfc: Cannot re-enable "
15178 "PCI device after reset.\n");
15179 return PCI_ERS_RESULT_DISCONNECT;
15180 }
15181
15182 pci_restore_state(pdev);
James Smart0a96e972011-07-22 18:37:28 -040015183
James Smart25ac2c972021-09-10 16:31:54 -070015184 phba->hba_flag &= ~HBA_PCI_ERR;
James Smart0a96e972011-07-22 18:37:28 -040015185 /*
15186 * As the new kernel behavior of pci_restore_state() API call clears
15187 * device saved_state flag, need to save the restored state again.
15188 */
15189 pci_save_state(pdev);
15190
James Smart75baf692010-06-08 18:31:21 -040015191 if (pdev->is_busmaster)
15192 pci_set_master(pdev);
15193
15194 spin_lock_irq(&phba->hbalock);
15195 psli->sli_flag &= ~LPFC_SLI_ACTIVE;
15196 spin_unlock_irq(&phba->hbalock);
15197
15198 /* Configure and enable interrupt */
15199 intr_mode = lpfc_sli4_enable_intr(phba, phba->intr_mode);
15200 if (intr_mode == LPFC_INTR_ERROR) {
Dick Kennedy372c1872020-06-30 14:50:00 -070015201 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart75baf692010-06-08 18:31:21 -040015202 "2824 Cannot re-enable interrupt after "
15203 "slot reset.\n");
15204 return PCI_ERS_RESULT_DISCONNECT;
15205 } else
15206 phba->intr_mode = intr_mode;
James Smart25ac2c972021-09-10 16:31:54 -070015207 lpfc_cpu_affinity_check(phba, phba->cfg_irq_chann);
James Smart75baf692010-06-08 18:31:21 -040015208
15209 /* Log the current active interrupt mode */
15210 lpfc_log_intr_mode(phba, phba->intr_mode);
15211
James Smartda0436e2009-05-22 14:51:39 -040015212 return PCI_ERS_RESULT_RECOVERED;
15213}
15214
15215/**
15216 * lpfc_io_resume_s4 - Method for resuming PCI I/O operation to SLI-4 device
15217 * @pdev: pointer to PCI device
15218 *
15219 * This routine is called from the PCI subsystem for error handling to device
15220 * with SLI-4 interface spec. It is called when kernel error recovery tells
15221 * the lpfc driver that it is ok to resume normal PCI operation after PCI bus
15222 * error recovery. After this call, traffic can start to flow from this device
15223 * again.
15224 **/
15225static void
15226lpfc_io_resume_s4(struct pci_dev *pdev)
15227{
James Smart75baf692010-06-08 18:31:21 -040015228 struct Scsi_Host *shost = pci_get_drvdata(pdev);
15229 struct lpfc_hba *phba = ((struct lpfc_vport *)shost->hostdata)->phba;
15230
15231 /*
15232 * In case of slot reset, as function reset is performed through
15233 * mailbox command which needs DMA to be enabled, this operation
15234 * has to be moved to the io resume phase. Taking device offline
15235 * will perform the necessary cleanup.
15236 */
15237 if (!(phba->sli.sli_flag & LPFC_SLI_ACTIVE)) {
15238 /* Perform device reset */
James Smart618a5232012-06-12 13:54:36 -040015239 lpfc_offline_prep(phba, LPFC_MBX_WAIT);
James Smart75baf692010-06-08 18:31:21 -040015240 lpfc_offline(phba);
15241 lpfc_sli_brdrestart(phba);
15242 /* Bring the device back online */
15243 lpfc_online(phba);
15244 }
James Smartda0436e2009-05-22 14:51:39 -040015245}
15246
15247/**
James Smart3772a992009-05-22 14:50:54 -040015248 * lpfc_pci_probe_one - lpfc PCI probe func to reg dev to PCI subsystem
15249 * @pdev: pointer to PCI device
15250 * @pid: pointer to PCI device identifier
15251 *
15252 * This routine is to be registered to the kernel's PCI subsystem. When an
15253 * Emulex HBA device is presented on PCI bus, the kernel PCI subsystem looks
15254 * at PCI device-specific information of the device and driver to see if the
15255 * driver state that it can support this kind of device. If the match is
15256 * successful, the driver core invokes this routine. This routine dispatches
15257 * the action to the proper SLI-3 or SLI-4 device probing routine, which will
15258 * do all the initialization that it needs to do to handle the HBA device
15259 * properly.
15260 *
15261 * Return code
15262 * 0 - driver can claim the device
15263 * negative value - driver can not claim the device
15264 **/
Greg Kroah-Hartman6f039792012-12-21 13:08:55 -080015265static int
James Smart3772a992009-05-22 14:50:54 -040015266lpfc_pci_probe_one(struct pci_dev *pdev, const struct pci_device_id *pid)
15267{
15268 int rc;
James Smart8fa38512009-07-19 10:01:03 -040015269 struct lpfc_sli_intf intf;
James Smart3772a992009-05-22 14:50:54 -040015270
James Smart28baac72010-02-12 14:42:03 -050015271 if (pci_read_config_dword(pdev, LPFC_SLI_INTF, &intf.word0))
James Smart3772a992009-05-22 14:50:54 -040015272 return -ENODEV;
15273
James Smart8fa38512009-07-19 10:01:03 -040015274 if ((bf_get(lpfc_sli_intf_valid, &intf) == LPFC_SLI_INTF_VALID) &&
James Smart28baac72010-02-12 14:42:03 -050015275 (bf_get(lpfc_sli_intf_slirev, &intf) == LPFC_SLI_INTF_REV_SLI4))
James Smartda0436e2009-05-22 14:51:39 -040015276 rc = lpfc_pci_probe_one_s4(pdev, pid);
James Smart8fa38512009-07-19 10:01:03 -040015277 else
James Smart3772a992009-05-22 14:50:54 -040015278 rc = lpfc_pci_probe_one_s3(pdev, pid);
James Smart8fa38512009-07-19 10:01:03 -040015279
James Smart3772a992009-05-22 14:50:54 -040015280 return rc;
15281}
15282
15283/**
15284 * lpfc_pci_remove_one - lpfc PCI func to unreg dev from PCI subsystem
15285 * @pdev: pointer to PCI device
15286 *
15287 * This routine is to be registered to the kernel's PCI subsystem. When an
15288 * Emulex HBA is removed from PCI bus, the driver core invokes this routine.
15289 * This routine dispatches the action to the proper SLI-3 or SLI-4 device
15290 * remove routine, which will perform all the necessary cleanup for the
15291 * device to be removed from the PCI subsystem properly.
15292 **/
Greg Kroah-Hartman6f039792012-12-21 13:08:55 -080015293static void
James Smart3772a992009-05-22 14:50:54 -040015294lpfc_pci_remove_one(struct pci_dev *pdev)
15295{
15296 struct Scsi_Host *shost = pci_get_drvdata(pdev);
15297 struct lpfc_hba *phba = ((struct lpfc_vport *)shost->hostdata)->phba;
15298
15299 switch (phba->pci_dev_grp) {
15300 case LPFC_PCI_DEV_LP:
15301 lpfc_pci_remove_one_s3(pdev);
15302 break;
James Smartda0436e2009-05-22 14:51:39 -040015303 case LPFC_PCI_DEV_OC:
15304 lpfc_pci_remove_one_s4(pdev);
15305 break;
James Smart3772a992009-05-22 14:50:54 -040015306 default:
Dick Kennedy372c1872020-06-30 14:50:00 -070015307 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart3772a992009-05-22 14:50:54 -040015308 "1424 Invalid PCI device group: 0x%x\n",
15309 phba->pci_dev_grp);
15310 break;
15311 }
15312 return;
15313}
15314
15315/**
15316 * lpfc_pci_suspend_one - lpfc PCI func to suspend dev for power management
Vaibhav Guptaef6fa162020-11-02 22:17:18 +053015317 * @dev: pointer to device
James Smart3772a992009-05-22 14:50:54 -040015318 *
15319 * This routine is to be registered to the kernel's PCI subsystem to support
15320 * system Power Management (PM). When PM invokes this method, it dispatches
15321 * the action to the proper SLI-3 or SLI-4 device suspend routine, which will
15322 * suspend the device.
15323 *
15324 * Return code
15325 * 0 - driver suspended the device
15326 * Error otherwise
15327 **/
Vaibhav Guptaef6fa162020-11-02 22:17:18 +053015328static int __maybe_unused
15329lpfc_pci_suspend_one(struct device *dev)
James Smart3772a992009-05-22 14:50:54 -040015330{
Vaibhav Guptaef6fa162020-11-02 22:17:18 +053015331 struct Scsi_Host *shost = dev_get_drvdata(dev);
James Smart3772a992009-05-22 14:50:54 -040015332 struct lpfc_hba *phba = ((struct lpfc_vport *)shost->hostdata)->phba;
15333 int rc = -ENODEV;
15334
15335 switch (phba->pci_dev_grp) {
15336 case LPFC_PCI_DEV_LP:
Vaibhav Guptaef6fa162020-11-02 22:17:18 +053015337 rc = lpfc_pci_suspend_one_s3(dev);
James Smart3772a992009-05-22 14:50:54 -040015338 break;
James Smartda0436e2009-05-22 14:51:39 -040015339 case LPFC_PCI_DEV_OC:
Vaibhav Guptaef6fa162020-11-02 22:17:18 +053015340 rc = lpfc_pci_suspend_one_s4(dev);
James Smartda0436e2009-05-22 14:51:39 -040015341 break;
James Smart3772a992009-05-22 14:50:54 -040015342 default:
Dick Kennedy372c1872020-06-30 14:50:00 -070015343 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart3772a992009-05-22 14:50:54 -040015344 "1425 Invalid PCI device group: 0x%x\n",
15345 phba->pci_dev_grp);
15346 break;
15347 }
15348 return rc;
15349}
15350
15351/**
15352 * lpfc_pci_resume_one - lpfc PCI func to resume dev for power management
Vaibhav Guptaef6fa162020-11-02 22:17:18 +053015353 * @dev: pointer to device
James Smart3772a992009-05-22 14:50:54 -040015354 *
15355 * This routine is to be registered to the kernel's PCI subsystem to support
15356 * system Power Management (PM). When PM invokes this method, it dispatches
15357 * the action to the proper SLI-3 or SLI-4 device resume routine, which will
15358 * resume the device.
15359 *
15360 * Return code
15361 * 0 - driver suspended the device
15362 * Error otherwise
15363 **/
Vaibhav Guptaef6fa162020-11-02 22:17:18 +053015364static int __maybe_unused
15365lpfc_pci_resume_one(struct device *dev)
James Smart3772a992009-05-22 14:50:54 -040015366{
Vaibhav Guptaef6fa162020-11-02 22:17:18 +053015367 struct Scsi_Host *shost = dev_get_drvdata(dev);
James Smart3772a992009-05-22 14:50:54 -040015368 struct lpfc_hba *phba = ((struct lpfc_vport *)shost->hostdata)->phba;
15369 int rc = -ENODEV;
15370
15371 switch (phba->pci_dev_grp) {
15372 case LPFC_PCI_DEV_LP:
Vaibhav Guptaef6fa162020-11-02 22:17:18 +053015373 rc = lpfc_pci_resume_one_s3(dev);
James Smart3772a992009-05-22 14:50:54 -040015374 break;
James Smartda0436e2009-05-22 14:51:39 -040015375 case LPFC_PCI_DEV_OC:
Vaibhav Guptaef6fa162020-11-02 22:17:18 +053015376 rc = lpfc_pci_resume_one_s4(dev);
James Smartda0436e2009-05-22 14:51:39 -040015377 break;
James Smart3772a992009-05-22 14:50:54 -040015378 default:
Dick Kennedy372c1872020-06-30 14:50:00 -070015379 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart3772a992009-05-22 14:50:54 -040015380 "1426 Invalid PCI device group: 0x%x\n",
15381 phba->pci_dev_grp);
15382 break;
15383 }
15384 return rc;
15385}
15386
15387/**
15388 * lpfc_io_error_detected - lpfc method for handling PCI I/O error
15389 * @pdev: pointer to PCI device.
15390 * @state: the current PCI connection state.
15391 *
15392 * This routine is registered to the PCI subsystem for error handling. This
15393 * function is called by the PCI subsystem after a PCI bus error affecting
15394 * this device has been detected. When this routine is invoked, it dispatches
15395 * the action to the proper SLI-3 or SLI-4 device error detected handling
15396 * routine, which will perform the proper error detected operation.
15397 *
15398 * Return codes
15399 * PCI_ERS_RESULT_NEED_RESET - need to reset before recovery
15400 * PCI_ERS_RESULT_DISCONNECT - device could not be recovered
15401 **/
15402static pci_ers_result_t
15403lpfc_io_error_detected(struct pci_dev *pdev, pci_channel_state_t state)
15404{
15405 struct Scsi_Host *shost = pci_get_drvdata(pdev);
15406 struct lpfc_hba *phba = ((struct lpfc_vport *)shost->hostdata)->phba;
15407 pci_ers_result_t rc = PCI_ERS_RESULT_DISCONNECT;
15408
James Smart25ac2c972021-09-10 16:31:54 -070015409 if (phba->link_state == LPFC_HBA_ERROR &&
15410 phba->hba_flag & HBA_IOQ_FLUSH)
15411 return PCI_ERS_RESULT_NEED_RESET;
15412
James Smart3772a992009-05-22 14:50:54 -040015413 switch (phba->pci_dev_grp) {
15414 case LPFC_PCI_DEV_LP:
15415 rc = lpfc_io_error_detected_s3(pdev, state);
15416 break;
James Smartda0436e2009-05-22 14:51:39 -040015417 case LPFC_PCI_DEV_OC:
15418 rc = lpfc_io_error_detected_s4(pdev, state);
15419 break;
James Smart3772a992009-05-22 14:50:54 -040015420 default:
Dick Kennedy372c1872020-06-30 14:50:00 -070015421 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart3772a992009-05-22 14:50:54 -040015422 "1427 Invalid PCI device group: 0x%x\n",
15423 phba->pci_dev_grp);
15424 break;
15425 }
15426 return rc;
15427}
15428
15429/**
15430 * lpfc_io_slot_reset - lpfc method for restart PCI dev from scratch
15431 * @pdev: pointer to PCI device.
15432 *
15433 * This routine is registered to the PCI subsystem for error handling. This
15434 * function is called after PCI bus has been reset to restart the PCI card
15435 * from scratch, as if from a cold-boot. When this routine is invoked, it
15436 * dispatches the action to the proper SLI-3 or SLI-4 device reset handling
15437 * routine, which will perform the proper device reset.
15438 *
15439 * Return codes
15440 * PCI_ERS_RESULT_RECOVERED - the device has been recovered
15441 * PCI_ERS_RESULT_DISCONNECT - device could not be recovered
15442 **/
15443static pci_ers_result_t
15444lpfc_io_slot_reset(struct pci_dev *pdev)
15445{
15446 struct Scsi_Host *shost = pci_get_drvdata(pdev);
15447 struct lpfc_hba *phba = ((struct lpfc_vport *)shost->hostdata)->phba;
15448 pci_ers_result_t rc = PCI_ERS_RESULT_DISCONNECT;
15449
15450 switch (phba->pci_dev_grp) {
15451 case LPFC_PCI_DEV_LP:
15452 rc = lpfc_io_slot_reset_s3(pdev);
15453 break;
James Smartda0436e2009-05-22 14:51:39 -040015454 case LPFC_PCI_DEV_OC:
15455 rc = lpfc_io_slot_reset_s4(pdev);
15456 break;
James Smart3772a992009-05-22 14:50:54 -040015457 default:
Dick Kennedy372c1872020-06-30 14:50:00 -070015458 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart3772a992009-05-22 14:50:54 -040015459 "1428 Invalid PCI device group: 0x%x\n",
15460 phba->pci_dev_grp);
15461 break;
15462 }
15463 return rc;
15464}
15465
15466/**
15467 * lpfc_io_resume - lpfc method for resuming PCI I/O operation
15468 * @pdev: pointer to PCI device
15469 *
15470 * This routine is registered to the PCI subsystem for error handling. It
15471 * is called when kernel error recovery tells the lpfc driver that it is
15472 * OK to resume normal PCI operation after PCI bus error recovery. When
15473 * this routine is invoked, it dispatches the action to the proper SLI-3
15474 * or SLI-4 device io_resume routine, which will resume the device operation.
15475 **/
15476static void
15477lpfc_io_resume(struct pci_dev *pdev)
15478{
15479 struct Scsi_Host *shost = pci_get_drvdata(pdev);
15480 struct lpfc_hba *phba = ((struct lpfc_vport *)shost->hostdata)->phba;
15481
15482 switch (phba->pci_dev_grp) {
15483 case LPFC_PCI_DEV_LP:
15484 lpfc_io_resume_s3(pdev);
15485 break;
James Smartda0436e2009-05-22 14:51:39 -040015486 case LPFC_PCI_DEV_OC:
15487 lpfc_io_resume_s4(pdev);
15488 break;
James Smart3772a992009-05-22 14:50:54 -040015489 default:
Dick Kennedy372c1872020-06-30 14:50:00 -070015490 lpfc_printf_log(phba, KERN_ERR, LOG_TRACE_EVENT,
James Smart3772a992009-05-22 14:50:54 -040015491 "1429 Invalid PCI device group: 0x%x\n",
15492 phba->pci_dev_grp);
15493 break;
15494 }
15495 return;
15496}
15497
James Smart1ba981f2014-02-20 09:56:45 -050015498/**
15499 * lpfc_sli4_oas_verify - Verify OAS is supported by this adapter
15500 * @phba: pointer to lpfc hba data structure.
15501 *
15502 * This routine checks to see if OAS is supported for this adapter. If
15503 * supported, the configure Flash Optimized Fabric flag is set. Otherwise,
15504 * the enable oas flag is cleared and the pool created for OAS device data
15505 * is destroyed.
15506 *
15507 **/
YueHaibingc7092972019-04-17 23:05:54 +080015508static void
James Smart1ba981f2014-02-20 09:56:45 -050015509lpfc_sli4_oas_verify(struct lpfc_hba *phba)
15510{
15511
15512 if (!phba->cfg_EnableXLane)
15513 return;
15514
15515 if (phba->sli4_hba.pc_sli4_params.oas_supported) {
15516 phba->cfg_fof = 1;
15517 } else {
James Smartf38fa0b2014-04-04 13:52:21 -040015518 phba->cfg_fof = 0;
Saurav Girepunjec3e5aac2019-10-27 01:17:17 +053015519 mempool_destroy(phba->device_data_mem_pool);
James Smart1ba981f2014-02-20 09:56:45 -050015520 phba->device_data_mem_pool = NULL;
15521 }
15522
15523 return;
15524}
15525
15526/**
James Smartd2cc9bc2018-09-10 10:30:50 -070015527 * lpfc_sli4_ras_init - Verify RAS-FW log is supported by this adapter
15528 * @phba: pointer to lpfc hba data structure.
15529 *
15530 * This routine checks to see if RAS is supported by the adapter. Check the
15531 * function through which RAS support enablement is to be done.
15532 **/
15533void
15534lpfc_sli4_ras_init(struct lpfc_hba *phba)
15535{
James Smartf6c5e6c2021-07-22 15:17:18 -070015536 /* if ASIC_GEN_NUM >= 0xC) */
15537 if ((bf_get(lpfc_sli_intf_if_type, &phba->sli4_hba.sli_intf) ==
15538 LPFC_SLI_INTF_IF_TYPE_6) ||
15539 (bf_get(lpfc_sli_intf_sli_family, &phba->sli4_hba.sli_intf) ==
15540 LPFC_SLI_INTF_FAMILY_G6)) {
James Smartd2cc9bc2018-09-10 10:30:50 -070015541 phba->ras_fwlog.ras_hwsupport = true;
James Smartcb349902018-11-29 16:09:27 -080015542 if (phba->cfg_ras_fwlog_func == PCI_FUNC(phba->pcidev->devfn) &&
15543 phba->cfg_ras_fwlog_buffsize)
James Smartd2cc9bc2018-09-10 10:30:50 -070015544 phba->ras_fwlog.ras_enabled = true;
15545 else
15546 phba->ras_fwlog.ras_enabled = false;
James Smartf6c5e6c2021-07-22 15:17:18 -070015547 } else {
James Smartd2cc9bc2018-09-10 10:30:50 -070015548 phba->ras_fwlog.ras_hwsupport = false;
15549 }
15550}
15551
James Smart1ba981f2014-02-20 09:56:45 -050015552
dea31012005-04-17 16:05:31 -050015553MODULE_DEVICE_TABLE(pci, lpfc_id_table);
15554
Stephen Hemmingera55b2d22012-09-07 09:33:16 -070015555static const struct pci_error_handlers lpfc_err_handler = {
Linas Vepstas8d63f372007-02-14 14:28:36 -060015556 .error_detected = lpfc_io_error_detected,
15557 .slot_reset = lpfc_io_slot_reset,
15558 .resume = lpfc_io_resume,
15559};
15560
Vaibhav Guptaef6fa162020-11-02 22:17:18 +053015561static SIMPLE_DEV_PM_OPS(lpfc_pci_pm_ops_one,
15562 lpfc_pci_suspend_one,
15563 lpfc_pci_resume_one);
15564
dea31012005-04-17 16:05:31 -050015565static struct pci_driver lpfc_driver = {
15566 .name = LPFC_DRIVER_NAME,
15567 .id_table = lpfc_id_table,
15568 .probe = lpfc_pci_probe_one,
Greg Kroah-Hartman6f039792012-12-21 13:08:55 -080015569 .remove = lpfc_pci_remove_one,
Anton Blanchard85e8a232017-02-13 08:49:20 +110015570 .shutdown = lpfc_pci_remove_one,
Vaibhav Guptaef6fa162020-11-02 22:17:18 +053015571 .driver.pm = &lpfc_pci_pm_ops_one,
James Smart2e0fef82007-06-17 19:56:36 -050015572 .err_handler = &lpfc_err_handler,
dea31012005-04-17 16:05:31 -050015573};
15574
James Smart3ef6d242012-01-18 16:23:48 -050015575static const struct file_operations lpfc_mgmt_fop = {
Al Viro858feac2013-04-14 22:39:37 -040015576 .owner = THIS_MODULE,
James Smart3ef6d242012-01-18 16:23:48 -050015577};
15578
15579static struct miscdevice lpfc_mgmt_dev = {
15580 .minor = MISC_DYNAMIC_MINOR,
15581 .name = "lpfcmgmt",
15582 .fops = &lpfc_mgmt_fop,
15583};
15584
James Smarte59058c2008-08-24 21:49:00 -040015585/**
James Smart3621a712009-04-06 18:47:14 -040015586 * lpfc_init - lpfc module initialization routine
James Smarte59058c2008-08-24 21:49:00 -040015587 *
15588 * This routine is to be invoked when the lpfc module is loaded into the
15589 * kernel. The special kernel macro module_init() is used to indicate the
15590 * role of this routine to the kernel as lpfc module entry point.
15591 *
15592 * Return codes
15593 * 0 - successful
15594 * -ENOMEM - FC attach transport failed
15595 * all others - failed
15596 */
dea31012005-04-17 16:05:31 -050015597static int __init
15598lpfc_init(void)
15599{
15600 int error = 0;
15601
Anton Blanchardbc2736e2020-07-13 18:39:08 +100015602 pr_info(LPFC_MODULE_DESC "\n");
15603 pr_info(LPFC_COPYRIGHT "\n");
dea31012005-04-17 16:05:31 -050015604
James Smart3ef6d242012-01-18 16:23:48 -050015605 error = misc_register(&lpfc_mgmt_dev);
15606 if (error)
15607 printk(KERN_ERR "Could not register lpfcmgmt device, "
15608 "misc_register returned with status %d", error);
15609
Jing Xiangfeng1eaff532020-07-31 14:56:39 +080015610 error = -ENOMEM;
James Smart458c0832016-07-06 12:36:07 -070015611 lpfc_transport_functions.vport_create = lpfc_vport_create;
15612 lpfc_transport_functions.vport_delete = lpfc_vport_delete;
dea31012005-04-17 16:05:31 -050015613 lpfc_transport_template =
15614 fc_attach_transport(&lpfc_transport_functions);
James Smart7ee5d432007-10-27 13:37:17 -040015615 if (lpfc_transport_template == NULL)
Jing Xiangfeng1eaff532020-07-31 14:56:39 +080015616 goto unregister;
James Smart458c0832016-07-06 12:36:07 -070015617 lpfc_vport_transport_template =
15618 fc_attach_transport(&lpfc_vport_transport_functions);
15619 if (lpfc_vport_transport_template == NULL) {
15620 fc_release_transport(lpfc_transport_template);
Jing Xiangfeng1eaff532020-07-31 14:56:39 +080015621 goto unregister;
James Smart7ee5d432007-10-27 13:37:17 -040015622 }
James Smart840a4702020-11-15 11:26:40 -080015623 lpfc_wqe_cmd_template();
James Smartbd3061b2018-03-05 12:04:05 -080015624 lpfc_nvmet_cmd_template();
James Smart7bb03bb2013-04-17 20:19:16 -040015625
15626 /* Initialize in case vector mapping is needed */
James Smart2ea259e2017-02-12 13:52:27 -080015627 lpfc_present_cpu = num_present_cpus();
James Smart7bb03bb2013-04-17 20:19:16 -040015628
James Smarta5b141a2021-09-27 11:35:18 -070015629 lpfc_pldv_detect = false;
15630
James Smart93a4d6f2019-11-04 16:57:05 -080015631 error = cpuhp_setup_state_multi(CPUHP_AP_ONLINE_DYN,
15632 "lpfc/sli4:online",
15633 lpfc_cpu_online, lpfc_cpu_offline);
15634 if (error < 0)
15635 goto cpuhp_failure;
15636 lpfc_cpuhp_state = error;
15637
dea31012005-04-17 16:05:31 -050015638 error = pci_register_driver(&lpfc_driver);
James Smart93a4d6f2019-11-04 16:57:05 -080015639 if (error)
15640 goto unwind;
15641
15642 return error;
15643
15644unwind:
15645 cpuhp_remove_multi_state(lpfc_cpuhp_state);
15646cpuhp_failure:
15647 fc_release_transport(lpfc_transport_template);
15648 fc_release_transport(lpfc_vport_transport_template);
Jing Xiangfeng1eaff532020-07-31 14:56:39 +080015649unregister:
15650 misc_deregister(&lpfc_mgmt_dev);
dea31012005-04-17 16:05:31 -050015651
15652 return error;
15653}
15654
Dick Kennedy372c1872020-06-30 14:50:00 -070015655void lpfc_dmp_dbg(struct lpfc_hba *phba)
15656{
15657 unsigned int start_idx;
15658 unsigned int dbg_cnt;
15659 unsigned int temp_idx;
15660 int i;
15661 int j = 0;
James Smarte8613082021-07-07 11:43:33 -070015662 unsigned long rem_nsec, iflags;
15663 bool log_verbose = false;
15664 struct lpfc_vport *port_iterator;
Dick Kennedy372c1872020-06-30 14:50:00 -070015665
James Smart0b3ad322021-01-04 10:02:39 -080015666 /* Don't dump messages if we explicitly set log_verbose for the
15667 * physical port or any vport.
15668 */
Dick Kennedy372c1872020-06-30 14:50:00 -070015669 if (phba->cfg_log_verbose)
15670 return;
15671
James Smarte8613082021-07-07 11:43:33 -070015672 spin_lock_irqsave(&phba->port_list_lock, iflags);
15673 list_for_each_entry(port_iterator, &phba->port_list, listentry) {
15674 if (port_iterator->load_flag & FC_UNLOADING)
15675 continue;
15676 if (scsi_host_get(lpfc_shost_from_vport(port_iterator))) {
15677 if (port_iterator->cfg_log_verbose)
15678 log_verbose = true;
15679
15680 scsi_host_put(lpfc_shost_from_vport(port_iterator));
15681
15682 if (log_verbose) {
15683 spin_unlock_irqrestore(&phba->port_list_lock,
15684 iflags);
James Smart0b3ad322021-01-04 10:02:39 -080015685 return;
15686 }
15687 }
15688 }
James Smarte8613082021-07-07 11:43:33 -070015689 spin_unlock_irqrestore(&phba->port_list_lock, iflags);
James Smart0b3ad322021-01-04 10:02:39 -080015690
Dick Kennedy372c1872020-06-30 14:50:00 -070015691 if (atomic_cmpxchg(&phba->dbg_log_dmping, 0, 1) != 0)
15692 return;
15693
15694 start_idx = (unsigned int)atomic_read(&phba->dbg_log_idx) % DBG_LOG_SZ;
15695 dbg_cnt = (unsigned int)atomic_read(&phba->dbg_log_cnt);
James Smart0b3ad322021-01-04 10:02:39 -080015696 if (!dbg_cnt)
15697 goto out;
Dick Kennedy372c1872020-06-30 14:50:00 -070015698 temp_idx = start_idx;
15699 if (dbg_cnt >= DBG_LOG_SZ) {
15700 dbg_cnt = DBG_LOG_SZ;
15701 temp_idx -= 1;
15702 } else {
15703 if ((start_idx + dbg_cnt) > (DBG_LOG_SZ - 1)) {
15704 temp_idx = (start_idx + dbg_cnt) % DBG_LOG_SZ;
15705 } else {
Dick Kennedy77dd7d72020-07-06 13:42:46 -070015706 if (start_idx < dbg_cnt)
Dick Kennedy372c1872020-06-30 14:50:00 -070015707 start_idx = DBG_LOG_SZ - (dbg_cnt - start_idx);
Dick Kennedy77dd7d72020-07-06 13:42:46 -070015708 else
Dick Kennedy372c1872020-06-30 14:50:00 -070015709 start_idx -= dbg_cnt;
Dick Kennedy372c1872020-06-30 14:50:00 -070015710 }
15711 }
15712 dev_info(&phba->pcidev->dev, "start %d end %d cnt %d\n",
15713 start_idx, temp_idx, dbg_cnt);
15714
15715 for (i = 0; i < dbg_cnt; i++) {
15716 if ((start_idx + i) < DBG_LOG_SZ)
Dick Kennedy77dd7d72020-07-06 13:42:46 -070015717 temp_idx = (start_idx + i) % DBG_LOG_SZ;
Dick Kennedy372c1872020-06-30 14:50:00 -070015718 else
15719 temp_idx = j++;
15720 rem_nsec = do_div(phba->dbg_log[temp_idx].t_ns, NSEC_PER_SEC);
15721 dev_info(&phba->pcidev->dev, "%d: [%5lu.%06lu] %s",
15722 temp_idx,
15723 (unsigned long)phba->dbg_log[temp_idx].t_ns,
15724 rem_nsec / 1000,
15725 phba->dbg_log[temp_idx].log);
15726 }
James Smart0b3ad322021-01-04 10:02:39 -080015727out:
Dick Kennedy372c1872020-06-30 14:50:00 -070015728 atomic_set(&phba->dbg_log_cnt, 0);
15729 atomic_set(&phba->dbg_log_dmping, 0);
15730}
15731
Lee Jones7fa03c72020-07-23 13:24:21 +010015732__printf(2, 3)
Dick Kennedy372c1872020-06-30 14:50:00 -070015733void lpfc_dbg_print(struct lpfc_hba *phba, const char *fmt, ...)
15734{
15735 unsigned int idx;
15736 va_list args;
15737 int dbg_dmping = atomic_read(&phba->dbg_log_dmping);
15738 struct va_format vaf;
15739
15740
15741 va_start(args, fmt);
15742 if (unlikely(dbg_dmping)) {
15743 vaf.fmt = fmt;
15744 vaf.va = &args;
15745 dev_info(&phba->pcidev->dev, "%pV", &vaf);
15746 va_end(args);
15747 return;
15748 }
15749 idx = (unsigned int)atomic_fetch_add(1, &phba->dbg_log_idx) %
15750 DBG_LOG_SZ;
15751
15752 atomic_inc(&phba->dbg_log_cnt);
15753
15754 vscnprintf(phba->dbg_log[idx].log,
15755 sizeof(phba->dbg_log[idx].log), fmt, args);
15756 va_end(args);
15757
15758 phba->dbg_log[idx].t_ns = local_clock();
15759}
15760
James Smarte59058c2008-08-24 21:49:00 -040015761/**
James Smart3621a712009-04-06 18:47:14 -040015762 * lpfc_exit - lpfc module removal routine
James Smarte59058c2008-08-24 21:49:00 -040015763 *
15764 * This routine is invoked when the lpfc module is removed from the kernel.
15765 * The special kernel macro module_exit() is used to indicate the role of
15766 * this routine to the kernel as lpfc module exit point.
15767 */
dea31012005-04-17 16:05:31 -050015768static void __exit
15769lpfc_exit(void)
15770{
James Smart3ef6d242012-01-18 16:23:48 -050015771 misc_deregister(&lpfc_mgmt_dev);
dea31012005-04-17 16:05:31 -050015772 pci_unregister_driver(&lpfc_driver);
James Smart93a4d6f2019-11-04 16:57:05 -080015773 cpuhp_remove_multi_state(lpfc_cpuhp_state);
dea31012005-04-17 16:05:31 -050015774 fc_release_transport(lpfc_transport_template);
James Smart458c0832016-07-06 12:36:07 -070015775 fc_release_transport(lpfc_vport_transport_template);
Johannes Thumshirn79739672015-08-31 16:48:11 -040015776 idr_destroy(&lpfc_hba_index);
dea31012005-04-17 16:05:31 -050015777}
15778
15779module_init(lpfc_init);
15780module_exit(lpfc_exit);
15781MODULE_LICENSE("GPL");
15782MODULE_DESCRIPTION(LPFC_MODULE_DESC);
James Smartd080abe2017-02-12 13:52:39 -080015783MODULE_AUTHOR("Broadcom");
dea31012005-04-17 16:05:31 -050015784MODULE_VERSION("0:" LPFC_DRIVER_VERSION);