xref: /linux/drivers/scsi/mpt3sas/mpt3sas_scsih.c (revision d642ef71)
1 /*
2  * Scsi Host Layer for MPT (Message Passing Technology) based controllers
3  *
4  * This code is based on drivers/scsi/mpt3sas/mpt3sas_scsih.c
5  * Copyright (C) 2012-2014  LSI Corporation
6  * Copyright (C) 2013-2014 Avago Technologies
7  *  (mailto: MPT-FusionLinux.pdl@avagotech.com)
8  *
9  * This program is free software; you can redistribute it and/or
10  * modify it under the terms of the GNU General Public License
11  * as published by the Free Software Foundation; either version 2
12  * of the License, or (at your option) any later version.
13  *
14  * This program is distributed in the hope that it will be useful,
15  * but WITHOUT ANY WARRANTY; without even the implied warranty of
16  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
17  * GNU General Public License for more details.
18  *
19  * NO WARRANTY
20  * THE PROGRAM IS PROVIDED ON AN "AS IS" BASIS, WITHOUT WARRANTIES OR
21  * CONDITIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED INCLUDING, WITHOUT
22  * LIMITATION, ANY WARRANTIES OR CONDITIONS OF TITLE, NON-INFRINGEMENT,
23  * MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Each Recipient is
24  * solely responsible for determining the appropriateness of using and
25  * distributing the Program and assumes all risks associated with its
26  * exercise of rights under this Agreement, including but not limited to
27  * the risks and costs of program errors, damage to or loss of data,
28  * programs or equipment, and unavailability or interruption of operations.
29 
30  * DISCLAIMER OF LIABILITY
31  * NEITHER RECIPIENT NOR ANY CONTRIBUTORS SHALL HAVE ANY LIABILITY FOR ANY
32  * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
33  * DAMAGES (INCLUDING WITHOUT LIMITATION LOST PROFITS), HOWEVER CAUSED AND
34  * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR
35  * TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE
36  * USE OR DISTRIBUTION OF THE PROGRAM OR THE EXERCISE OF ANY RIGHTS GRANTED
37  * HEREUNDER, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGES
38 
39  * You should have received a copy of the GNU General Public License
40  * along with this program; if not, write to the Free Software
41  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA  02110-1301,
42  * USA.
43  */
44 
45 #include <linux/module.h>
46 #include <linux/kernel.h>
47 #include <linux/init.h>
48 #include <linux/errno.h>
49 #include <linux/blkdev.h>
50 #include <linux/sched.h>
51 #include <linux/workqueue.h>
52 #include <linux/delay.h>
53 #include <linux/pci.h>
54 #include <linux/interrupt.h>
55 #include <linux/raid_class.h>
56 #include <linux/blk-mq-pci.h>
57 #include <asm/unaligned.h>
58 
59 #include "mpt3sas_base.h"
60 
61 #define RAID_CHANNEL 1
62 
63 #define PCIE_CHANNEL 2
64 
65 /* forward proto's */
66 static void _scsih_expander_node_remove(struct MPT3SAS_ADAPTER *ioc,
67 	struct _sas_node *sas_expander);
68 static void _firmware_event_work(struct work_struct *work);
69 
70 static void _scsih_remove_device(struct MPT3SAS_ADAPTER *ioc,
71 	struct _sas_device *sas_device);
72 static int _scsih_add_device(struct MPT3SAS_ADAPTER *ioc, u16 handle,
73 	u8 retry_count, u8 is_pd);
74 static int _scsih_pcie_add_device(struct MPT3SAS_ADAPTER *ioc, u16 handle);
75 static void _scsih_pcie_device_remove_from_sml(struct MPT3SAS_ADAPTER *ioc,
76 	struct _pcie_device *pcie_device);
77 static void
78 _scsih_pcie_check_device(struct MPT3SAS_ADAPTER *ioc, u16 handle);
79 static u8 _scsih_check_for_pending_tm(struct MPT3SAS_ADAPTER *ioc, u16 smid);
80 static void _scsih_complete_devices_scanning(struct MPT3SAS_ADAPTER *ioc);
81 
82 /* global parameters */
83 LIST_HEAD(mpt3sas_ioc_list);
84 /* global ioc lock for list operations */
85 DEFINE_SPINLOCK(gioc_lock);
86 
87 MODULE_AUTHOR(MPT3SAS_AUTHOR);
88 MODULE_DESCRIPTION(MPT3SAS_DESCRIPTION);
89 MODULE_LICENSE("GPL");
90 MODULE_VERSION(MPT3SAS_DRIVER_VERSION);
91 MODULE_ALIAS("mpt2sas");
92 
93 /* local parameters */
94 static u8 scsi_io_cb_idx = -1;
95 static u8 tm_cb_idx = -1;
96 static u8 ctl_cb_idx = -1;
97 static u8 base_cb_idx = -1;
98 static u8 port_enable_cb_idx = -1;
99 static u8 transport_cb_idx = -1;
100 static u8 scsih_cb_idx = -1;
101 static u8 config_cb_idx = -1;
102 static int mpt2_ids;
103 static int mpt3_ids;
104 
105 static u8 tm_tr_cb_idx = -1 ;
106 static u8 tm_tr_volume_cb_idx = -1 ;
107 static u8 tm_sas_control_cb_idx = -1;
108 
109 /* command line options */
110 static u32 logging_level;
111 MODULE_PARM_DESC(logging_level,
112 	" bits for enabling additional logging info (default=0)");
113 
114 
115 static ushort max_sectors = 0xFFFF;
116 module_param(max_sectors, ushort, 0444);
117 MODULE_PARM_DESC(max_sectors, "max sectors, range 64 to 32767  default=32767");
118 
119 
120 static int missing_delay[2] = {-1, -1};
121 module_param_array(missing_delay, int, NULL, 0444);
122 MODULE_PARM_DESC(missing_delay, " device missing delay , io missing delay");
123 
124 /* scsi-mid layer global parmeter is max_report_luns, which is 511 */
125 #define MPT3SAS_MAX_LUN (16895)
126 static u64 max_lun = MPT3SAS_MAX_LUN;
127 module_param(max_lun, ullong, 0444);
128 MODULE_PARM_DESC(max_lun, " max lun, default=16895 ");
129 
130 static ushort hbas_to_enumerate;
131 module_param(hbas_to_enumerate, ushort, 0444);
132 MODULE_PARM_DESC(hbas_to_enumerate,
133 		" 0 - enumerates both SAS 2.0 & SAS 3.0 generation HBAs\n \
134 		  1 - enumerates only SAS 2.0 generation HBAs\n \
135 		  2 - enumerates only SAS 3.0 generation HBAs (default=0)");
136 
137 /* diag_buffer_enable is bitwise
138  * bit 0 set = TRACE
139  * bit 1 set = SNAPSHOT
140  * bit 2 set = EXTENDED
141  *
142  * Either bit can be set, or both
143  */
144 static int diag_buffer_enable = -1;
145 module_param(diag_buffer_enable, int, 0444);
146 MODULE_PARM_DESC(diag_buffer_enable,
147 	" post diag buffers (TRACE=1/SNAPSHOT=2/EXTENDED=4/default=0)");
148 static int disable_discovery = -1;
149 module_param(disable_discovery, int, 0444);
150 MODULE_PARM_DESC(disable_discovery, " disable discovery ");
151 
152 
153 /* permit overriding the host protection capabilities mask (EEDP/T10 PI) */
154 static int prot_mask = -1;
155 module_param(prot_mask, int, 0444);
156 MODULE_PARM_DESC(prot_mask, " host protection capabilities mask, def=7 ");
157 
158 static bool enable_sdev_max_qd;
159 module_param(enable_sdev_max_qd, bool, 0444);
160 MODULE_PARM_DESC(enable_sdev_max_qd,
161 	"Enable sdev max qd as can_queue, def=disabled(0)");
162 
163 static int multipath_on_hba = -1;
164 module_param(multipath_on_hba, int, 0);
165 MODULE_PARM_DESC(multipath_on_hba,
166 	"Multipath support to add same target device\n\t\t"
167 	"as many times as it is visible to HBA from various paths\n\t\t"
168 	"(by default:\n\t\t"
169 	"\t SAS 2.0 & SAS 3.0 HBA - This will be disabled,\n\t\t"
170 	"\t SAS 3.5 HBA - This will be enabled)");
171 
172 static int host_tagset_enable = 1;
173 module_param(host_tagset_enable, int, 0444);
174 MODULE_PARM_DESC(host_tagset_enable,
175 	"Shared host tagset enable/disable Default: enable(1)");
176 
177 /* raid transport support */
178 static struct raid_template *mpt3sas_raid_template;
179 static struct raid_template *mpt2sas_raid_template;
180 
181 
182 /**
183  * struct sense_info - common structure for obtaining sense keys
184  * @skey: sense key
185  * @asc: additional sense code
186  * @ascq: additional sense code qualifier
187  */
188 struct sense_info {
189 	u8 skey;
190 	u8 asc;
191 	u8 ascq;
192 };
193 
194 #define MPT3SAS_PROCESS_TRIGGER_DIAG (0xFFFB)
195 #define MPT3SAS_TURN_ON_PFA_LED (0xFFFC)
196 #define MPT3SAS_PORT_ENABLE_COMPLETE (0xFFFD)
197 #define MPT3SAS_ABRT_TASK_SET (0xFFFE)
198 #define MPT3SAS_REMOVE_UNRESPONDING_DEVICES (0xFFFF)
199 /**
200  * struct fw_event_work - firmware event struct
201  * @list: link list framework
202  * @work: work object (ioc->fault_reset_work_q)
203  * @ioc: per adapter object
204  * @device_handle: device handle
205  * @VF_ID: virtual function id
206  * @VP_ID: virtual port id
207  * @ignore: flag meaning this event has been marked to ignore
208  * @event: firmware event MPI2_EVENT_XXX defined in mpi2_ioc.h
209  * @refcount: kref for this event
210  * @event_data: reply event data payload follows
211  *
212  * This object stored on ioc->fw_event_list.
213  */
214 struct fw_event_work {
215 	struct list_head	list;
216 	struct work_struct	work;
217 
218 	struct MPT3SAS_ADAPTER *ioc;
219 	u16			device_handle;
220 	u8			VF_ID;
221 	u8			VP_ID;
222 	u8			ignore;
223 	u16			event;
224 	struct kref		refcount;
225 	char			event_data[] __aligned(4);
226 };
227 
228 static void fw_event_work_free(struct kref *r)
229 {
230 	kfree(container_of(r, struct fw_event_work, refcount));
231 }
232 
233 static void fw_event_work_get(struct fw_event_work *fw_work)
234 {
235 	kref_get(&fw_work->refcount);
236 }
237 
238 static void fw_event_work_put(struct fw_event_work *fw_work)
239 {
240 	kref_put(&fw_work->refcount, fw_event_work_free);
241 }
242 
243 static struct fw_event_work *alloc_fw_event_work(int len)
244 {
245 	struct fw_event_work *fw_event;
246 
247 	fw_event = kzalloc(sizeof(*fw_event) + len, GFP_ATOMIC);
248 	if (!fw_event)
249 		return NULL;
250 
251 	kref_init(&fw_event->refcount);
252 	return fw_event;
253 }
254 
255 /**
256  * struct _scsi_io_transfer - scsi io transfer
257  * @handle: sas device handle (assigned by firmware)
258  * @is_raid: flag set for hidden raid components
259  * @dir: DMA_TO_DEVICE, DMA_FROM_DEVICE,
260  * @data_length: data transfer length
261  * @data_dma: dma pointer to data
262  * @sense: sense data
263  * @lun: lun number
264  * @cdb_length: cdb length
265  * @cdb: cdb contents
266  * @timeout: timeout for this command
267  * @VF_ID: virtual function id
268  * @VP_ID: virtual port id
269  * @valid_reply: flag set for reply message
270  * @sense_length: sense length
271  * @ioc_status: ioc status
272  * @scsi_state: scsi state
273  * @scsi_status: scsi staus
274  * @log_info: log information
275  * @transfer_length: data length transfer when there is a reply message
276  *
277  * Used for sending internal scsi commands to devices within this module.
278  * Refer to _scsi_send_scsi_io().
279  */
280 struct _scsi_io_transfer {
281 	u16	handle;
282 	u8	is_raid;
283 	enum dma_data_direction dir;
284 	u32	data_length;
285 	dma_addr_t data_dma;
286 	u8	sense[SCSI_SENSE_BUFFERSIZE];
287 	u32	lun;
288 	u8	cdb_length;
289 	u8	cdb[32];
290 	u8	timeout;
291 	u8	VF_ID;
292 	u8	VP_ID;
293 	u8	valid_reply;
294   /* the following bits are only valid when 'valid_reply = 1' */
295 	u32	sense_length;
296 	u16	ioc_status;
297 	u8	scsi_state;
298 	u8	scsi_status;
299 	u32	log_info;
300 	u32	transfer_length;
301 };
302 
303 /**
304  * _scsih_set_debug_level - global setting of ioc->logging_level.
305  * @val: ?
306  * @kp: ?
307  *
308  * Note: The logging levels are defined in mpt3sas_debug.h.
309  */
310 static int
311 _scsih_set_debug_level(const char *val, const struct kernel_param *kp)
312 {
313 	int ret = param_set_int(val, kp);
314 	struct MPT3SAS_ADAPTER *ioc;
315 
316 	if (ret)
317 		return ret;
318 
319 	pr_info("setting logging_level(0x%08x)\n", logging_level);
320 	spin_lock(&gioc_lock);
321 	list_for_each_entry(ioc, &mpt3sas_ioc_list, list)
322 		ioc->logging_level = logging_level;
323 	spin_unlock(&gioc_lock);
324 	return 0;
325 }
326 module_param_call(logging_level, _scsih_set_debug_level, param_get_int,
327 	&logging_level, 0644);
328 
329 /**
330  * _scsih_srch_boot_sas_address - search based on sas_address
331  * @sas_address: sas address
332  * @boot_device: boot device object from bios page 2
333  *
334  * Return: 1 when there's a match, 0 means no match.
335  */
336 static inline int
337 _scsih_srch_boot_sas_address(u64 sas_address,
338 	Mpi2BootDeviceSasWwid_t *boot_device)
339 {
340 	return (sas_address == le64_to_cpu(boot_device->SASAddress)) ?  1 : 0;
341 }
342 
343 /**
344  * _scsih_srch_boot_device_name - search based on device name
345  * @device_name: device name specified in INDENTIFY fram
346  * @boot_device: boot device object from bios page 2
347  *
348  * Return: 1 when there's a match, 0 means no match.
349  */
350 static inline int
351 _scsih_srch_boot_device_name(u64 device_name,
352 	Mpi2BootDeviceDeviceName_t *boot_device)
353 {
354 	return (device_name == le64_to_cpu(boot_device->DeviceName)) ? 1 : 0;
355 }
356 
357 /**
358  * _scsih_srch_boot_encl_slot - search based on enclosure_logical_id/slot
359  * @enclosure_logical_id: enclosure logical id
360  * @slot_number: slot number
361  * @boot_device: boot device object from bios page 2
362  *
363  * Return: 1 when there's a match, 0 means no match.
364  */
365 static inline int
366 _scsih_srch_boot_encl_slot(u64 enclosure_logical_id, u16 slot_number,
367 	Mpi2BootDeviceEnclosureSlot_t *boot_device)
368 {
369 	return (enclosure_logical_id == le64_to_cpu(boot_device->
370 	    EnclosureLogicalID) && slot_number == le16_to_cpu(boot_device->
371 	    SlotNumber)) ? 1 : 0;
372 }
373 
374 /**
375  * mpt3sas_get_port_by_id - get hba port entry corresponding to provided
376  *			  port number from port list
377  * @ioc: per adapter object
378  * @port_id: port number
379  * @bypass_dirty_port_flag: when set look the matching hba port entry even
380  *			if hba port entry is marked as dirty.
381  *
382  * Search for hba port entry corresponding to provided port number,
383  * if available return port object otherwise return NULL.
384  */
385 struct hba_port *
386 mpt3sas_get_port_by_id(struct MPT3SAS_ADAPTER *ioc,
387 	u8 port_id, u8 bypass_dirty_port_flag)
388 {
389 	struct hba_port *port, *port_next;
390 
391 	/*
392 	 * When multipath_on_hba is disabled then
393 	 * search the hba_port entry using default
394 	 * port id i.e. 255
395 	 */
396 	if (!ioc->multipath_on_hba)
397 		port_id = MULTIPATH_DISABLED_PORT_ID;
398 
399 	list_for_each_entry_safe(port, port_next,
400 	    &ioc->port_table_list, list) {
401 		if (port->port_id != port_id)
402 			continue;
403 		if (bypass_dirty_port_flag)
404 			return port;
405 		if (port->flags & HBA_PORT_FLAG_DIRTY_PORT)
406 			continue;
407 		return port;
408 	}
409 
410 	/*
411 	 * Allocate hba_port object for default port id (i.e. 255)
412 	 * when multipath_on_hba is disabled for the HBA.
413 	 * And add this object to port_table_list.
414 	 */
415 	if (!ioc->multipath_on_hba) {
416 		port = kzalloc(sizeof(struct hba_port), GFP_ATOMIC);
417 		if (!port)
418 			return NULL;
419 
420 		port->port_id = port_id;
421 		ioc_info(ioc,
422 		   "hba_port entry: %p, port: %d is added to hba_port list\n",
423 		   port, port->port_id);
424 		list_add_tail(&port->list,
425 		    &ioc->port_table_list);
426 		return port;
427 	}
428 	return NULL;
429 }
430 
431 /**
432  * mpt3sas_get_vphy_by_phy - get virtual_phy object corresponding to phy number
433  * @ioc: per adapter object
434  * @port: hba_port object
435  * @phy: phy number
436  *
437  * Return virtual_phy object corresponding to phy number.
438  */
439 struct virtual_phy *
440 mpt3sas_get_vphy_by_phy(struct MPT3SAS_ADAPTER *ioc,
441 	struct hba_port *port, u32 phy)
442 {
443 	struct virtual_phy *vphy, *vphy_next;
444 
445 	if (!port->vphys_mask)
446 		return NULL;
447 
448 	list_for_each_entry_safe(vphy, vphy_next, &port->vphys_list, list) {
449 		if (vphy->phy_mask & (1 << phy))
450 			return vphy;
451 	}
452 	return NULL;
453 }
454 
455 /**
456  * _scsih_is_boot_device - search for matching boot device.
457  * @sas_address: sas address
458  * @device_name: device name specified in INDENTIFY fram
459  * @enclosure_logical_id: enclosure logical id
460  * @slot: slot number
461  * @form: specifies boot device form
462  * @boot_device: boot device object from bios page 2
463  *
464  * Return: 1 when there's a match, 0 means no match.
465  */
466 static int
467 _scsih_is_boot_device(u64 sas_address, u64 device_name,
468 	u64 enclosure_logical_id, u16 slot, u8 form,
469 	Mpi2BiosPage2BootDevice_t *boot_device)
470 {
471 	int rc = 0;
472 
473 	switch (form) {
474 	case MPI2_BIOSPAGE2_FORM_SAS_WWID:
475 		if (!sas_address)
476 			break;
477 		rc = _scsih_srch_boot_sas_address(
478 		    sas_address, &boot_device->SasWwid);
479 		break;
480 	case MPI2_BIOSPAGE2_FORM_ENCLOSURE_SLOT:
481 		if (!enclosure_logical_id)
482 			break;
483 		rc = _scsih_srch_boot_encl_slot(
484 		    enclosure_logical_id,
485 		    slot, &boot_device->EnclosureSlot);
486 		break;
487 	case MPI2_BIOSPAGE2_FORM_DEVICE_NAME:
488 		if (!device_name)
489 			break;
490 		rc = _scsih_srch_boot_device_name(
491 		    device_name, &boot_device->DeviceName);
492 		break;
493 	case MPI2_BIOSPAGE2_FORM_NO_DEVICE_SPECIFIED:
494 		break;
495 	}
496 
497 	return rc;
498 }
499 
500 /**
501  * _scsih_get_sas_address - set the sas_address for given device handle
502  * @ioc: ?
503  * @handle: device handle
504  * @sas_address: sas address
505  *
506  * Return: 0 success, non-zero when failure
507  */
508 static int
509 _scsih_get_sas_address(struct MPT3SAS_ADAPTER *ioc, u16 handle,
510 	u64 *sas_address)
511 {
512 	Mpi2SasDevicePage0_t sas_device_pg0;
513 	Mpi2ConfigReply_t mpi_reply;
514 	u32 ioc_status;
515 
516 	*sas_address = 0;
517 
518 	if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0,
519 	    MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) {
520 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
521 			__FILE__, __LINE__, __func__);
522 		return -ENXIO;
523 	}
524 
525 	ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK;
526 	if (ioc_status == MPI2_IOCSTATUS_SUCCESS) {
527 		/* For HBA, vSES doesn't return HBA SAS address. Instead return
528 		 * vSES's sas address.
529 		 */
530 		if ((handle <= ioc->sas_hba.num_phys) &&
531 		   (!(le32_to_cpu(sas_device_pg0.DeviceInfo) &
532 		   MPI2_SAS_DEVICE_INFO_SEP)))
533 			*sas_address = ioc->sas_hba.sas_address;
534 		else
535 			*sas_address = le64_to_cpu(sas_device_pg0.SASAddress);
536 		return 0;
537 	}
538 
539 	/* we hit this because the given parent handle doesn't exist */
540 	if (ioc_status == MPI2_IOCSTATUS_CONFIG_INVALID_PAGE)
541 		return -ENXIO;
542 
543 	/* else error case */
544 	ioc_err(ioc, "handle(0x%04x), ioc_status(0x%04x), failure at %s:%d/%s()!\n",
545 		handle, ioc_status, __FILE__, __LINE__, __func__);
546 	return -EIO;
547 }
548 
549 /**
550  * _scsih_determine_boot_device - determine boot device.
551  * @ioc: per adapter object
552  * @device: sas_device or pcie_device object
553  * @channel: SAS or PCIe channel
554  *
555  * Determines whether this device should be first reported device to
556  * to scsi-ml or sas transport, this purpose is for persistent boot device.
557  * There are primary, alternate, and current entries in bios page 2. The order
558  * priority is primary, alternate, then current.  This routine saves
559  * the corresponding device object.
560  * The saved data to be used later in _scsih_probe_boot_devices().
561  */
562 static void
563 _scsih_determine_boot_device(struct MPT3SAS_ADAPTER *ioc, void *device,
564 	u32 channel)
565 {
566 	struct _sas_device *sas_device;
567 	struct _pcie_device *pcie_device;
568 	struct _raid_device *raid_device;
569 	u64 sas_address;
570 	u64 device_name;
571 	u64 enclosure_logical_id;
572 	u16 slot;
573 
574 	 /* only process this function when driver loads */
575 	if (!ioc->is_driver_loading)
576 		return;
577 
578 	 /* no Bios, return immediately */
579 	if (!ioc->bios_pg3.BiosVersion)
580 		return;
581 
582 	if (channel == RAID_CHANNEL) {
583 		raid_device = device;
584 		sas_address = raid_device->wwid;
585 		device_name = 0;
586 		enclosure_logical_id = 0;
587 		slot = 0;
588 	} else if (channel == PCIE_CHANNEL) {
589 		pcie_device = device;
590 		sas_address = pcie_device->wwid;
591 		device_name = 0;
592 		enclosure_logical_id = 0;
593 		slot = 0;
594 	} else {
595 		sas_device = device;
596 		sas_address = sas_device->sas_address;
597 		device_name = sas_device->device_name;
598 		enclosure_logical_id = sas_device->enclosure_logical_id;
599 		slot = sas_device->slot;
600 	}
601 
602 	if (!ioc->req_boot_device.device) {
603 		if (_scsih_is_boot_device(sas_address, device_name,
604 		    enclosure_logical_id, slot,
605 		    (ioc->bios_pg2.ReqBootDeviceForm &
606 		    MPI2_BIOSPAGE2_FORM_MASK),
607 		    &ioc->bios_pg2.RequestedBootDevice)) {
608 			dinitprintk(ioc,
609 				    ioc_info(ioc, "%s: req_boot_device(0x%016llx)\n",
610 					     __func__, (u64)sas_address));
611 			ioc->req_boot_device.device = device;
612 			ioc->req_boot_device.channel = channel;
613 		}
614 	}
615 
616 	if (!ioc->req_alt_boot_device.device) {
617 		if (_scsih_is_boot_device(sas_address, device_name,
618 		    enclosure_logical_id, slot,
619 		    (ioc->bios_pg2.ReqAltBootDeviceForm &
620 		    MPI2_BIOSPAGE2_FORM_MASK),
621 		    &ioc->bios_pg2.RequestedAltBootDevice)) {
622 			dinitprintk(ioc,
623 				    ioc_info(ioc, "%s: req_alt_boot_device(0x%016llx)\n",
624 					     __func__, (u64)sas_address));
625 			ioc->req_alt_boot_device.device = device;
626 			ioc->req_alt_boot_device.channel = channel;
627 		}
628 	}
629 
630 	if (!ioc->current_boot_device.device) {
631 		if (_scsih_is_boot_device(sas_address, device_name,
632 		    enclosure_logical_id, slot,
633 		    (ioc->bios_pg2.CurrentBootDeviceForm &
634 		    MPI2_BIOSPAGE2_FORM_MASK),
635 		    &ioc->bios_pg2.CurrentBootDevice)) {
636 			dinitprintk(ioc,
637 				    ioc_info(ioc, "%s: current_boot_device(0x%016llx)\n",
638 					     __func__, (u64)sas_address));
639 			ioc->current_boot_device.device = device;
640 			ioc->current_boot_device.channel = channel;
641 		}
642 	}
643 }
644 
645 static struct _sas_device *
646 __mpt3sas_get_sdev_from_target(struct MPT3SAS_ADAPTER *ioc,
647 		struct MPT3SAS_TARGET *tgt_priv)
648 {
649 	struct _sas_device *ret;
650 
651 	assert_spin_locked(&ioc->sas_device_lock);
652 
653 	ret = tgt_priv->sas_dev;
654 	if (ret)
655 		sas_device_get(ret);
656 
657 	return ret;
658 }
659 
660 static struct _sas_device *
661 mpt3sas_get_sdev_from_target(struct MPT3SAS_ADAPTER *ioc,
662 		struct MPT3SAS_TARGET *tgt_priv)
663 {
664 	struct _sas_device *ret;
665 	unsigned long flags;
666 
667 	spin_lock_irqsave(&ioc->sas_device_lock, flags);
668 	ret = __mpt3sas_get_sdev_from_target(ioc, tgt_priv);
669 	spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
670 
671 	return ret;
672 }
673 
674 static struct _pcie_device *
675 __mpt3sas_get_pdev_from_target(struct MPT3SAS_ADAPTER *ioc,
676 	struct MPT3SAS_TARGET *tgt_priv)
677 {
678 	struct _pcie_device *ret;
679 
680 	assert_spin_locked(&ioc->pcie_device_lock);
681 
682 	ret = tgt_priv->pcie_dev;
683 	if (ret)
684 		pcie_device_get(ret);
685 
686 	return ret;
687 }
688 
689 /**
690  * mpt3sas_get_pdev_from_target - pcie device search
691  * @ioc: per adapter object
692  * @tgt_priv: starget private object
693  *
694  * Context: This function will acquire ioc->pcie_device_lock and will release
695  * before returning the pcie_device object.
696  *
697  * This searches for pcie_device from target, then return pcie_device object.
698  */
699 static struct _pcie_device *
700 mpt3sas_get_pdev_from_target(struct MPT3SAS_ADAPTER *ioc,
701 	struct MPT3SAS_TARGET *tgt_priv)
702 {
703 	struct _pcie_device *ret;
704 	unsigned long flags;
705 
706 	spin_lock_irqsave(&ioc->pcie_device_lock, flags);
707 	ret = __mpt3sas_get_pdev_from_target(ioc, tgt_priv);
708 	spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
709 
710 	return ret;
711 }
712 
713 
714 /**
715  * __mpt3sas_get_sdev_by_rphy - sas device search
716  * @ioc: per adapter object
717  * @rphy: sas_rphy pointer
718  *
719  * Context: This function will acquire ioc->sas_device_lock and will release
720  * before returning the sas_device object.
721  *
722  * This searches for sas_device from rphy object
723  * then return sas_device object.
724  */
725 struct _sas_device *
726 __mpt3sas_get_sdev_by_rphy(struct MPT3SAS_ADAPTER *ioc,
727 	struct sas_rphy *rphy)
728 {
729 	struct _sas_device *sas_device;
730 
731 	assert_spin_locked(&ioc->sas_device_lock);
732 
733 	list_for_each_entry(sas_device, &ioc->sas_device_list, list) {
734 		if (sas_device->rphy != rphy)
735 			continue;
736 		sas_device_get(sas_device);
737 		return sas_device;
738 	}
739 
740 	sas_device = NULL;
741 	list_for_each_entry(sas_device, &ioc->sas_device_init_list, list) {
742 		if (sas_device->rphy != rphy)
743 			continue;
744 		sas_device_get(sas_device);
745 		return sas_device;
746 	}
747 
748 	return NULL;
749 }
750 
751 /**
752  * __mpt3sas_get_sdev_by_addr - get _sas_device object corresponding to provided
753  *				sas address from sas_device_list list
754  * @ioc: per adapter object
755  * @sas_address: device sas address
756  * @port: port number
757  *
758  * Search for _sas_device object corresponding to provided sas address,
759  * if available return _sas_device object address otherwise return NULL.
760  */
761 struct _sas_device *
762 __mpt3sas_get_sdev_by_addr(struct MPT3SAS_ADAPTER *ioc,
763 	u64 sas_address, struct hba_port *port)
764 {
765 	struct _sas_device *sas_device;
766 
767 	if (!port)
768 		return NULL;
769 
770 	assert_spin_locked(&ioc->sas_device_lock);
771 
772 	list_for_each_entry(sas_device, &ioc->sas_device_list, list) {
773 		if (sas_device->sas_address != sas_address)
774 			continue;
775 		if (sas_device->port != port)
776 			continue;
777 		sas_device_get(sas_device);
778 		return sas_device;
779 	}
780 
781 	list_for_each_entry(sas_device, &ioc->sas_device_init_list, list) {
782 		if (sas_device->sas_address != sas_address)
783 			continue;
784 		if (sas_device->port != port)
785 			continue;
786 		sas_device_get(sas_device);
787 		return sas_device;
788 	}
789 
790 	return NULL;
791 }
792 
793 /**
794  * mpt3sas_get_sdev_by_addr - sas device search
795  * @ioc: per adapter object
796  * @sas_address: sas address
797  * @port: hba port entry
798  * Context: Calling function should acquire ioc->sas_device_lock
799  *
800  * This searches for sas_device based on sas_address & port number,
801  * then return sas_device object.
802  */
803 struct _sas_device *
804 mpt3sas_get_sdev_by_addr(struct MPT3SAS_ADAPTER *ioc,
805 	u64 sas_address, struct hba_port *port)
806 {
807 	struct _sas_device *sas_device;
808 	unsigned long flags;
809 
810 	spin_lock_irqsave(&ioc->sas_device_lock, flags);
811 	sas_device = __mpt3sas_get_sdev_by_addr(ioc,
812 	    sas_address, port);
813 	spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
814 
815 	return sas_device;
816 }
817 
818 static struct _sas_device *
819 __mpt3sas_get_sdev_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle)
820 {
821 	struct _sas_device *sas_device;
822 
823 	assert_spin_locked(&ioc->sas_device_lock);
824 
825 	list_for_each_entry(sas_device, &ioc->sas_device_list, list)
826 		if (sas_device->handle == handle)
827 			goto found_device;
828 
829 	list_for_each_entry(sas_device, &ioc->sas_device_init_list, list)
830 		if (sas_device->handle == handle)
831 			goto found_device;
832 
833 	return NULL;
834 
835 found_device:
836 	sas_device_get(sas_device);
837 	return sas_device;
838 }
839 
840 /**
841  * mpt3sas_get_sdev_by_handle - sas device search
842  * @ioc: per adapter object
843  * @handle: sas device handle (assigned by firmware)
844  * Context: Calling function should acquire ioc->sas_device_lock
845  *
846  * This searches for sas_device based on sas_address, then return sas_device
847  * object.
848  */
849 struct _sas_device *
850 mpt3sas_get_sdev_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle)
851 {
852 	struct _sas_device *sas_device;
853 	unsigned long flags;
854 
855 	spin_lock_irqsave(&ioc->sas_device_lock, flags);
856 	sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle);
857 	spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
858 
859 	return sas_device;
860 }
861 
862 /**
863  * _scsih_display_enclosure_chassis_info - display device location info
864  * @ioc: per adapter object
865  * @sas_device: per sas device object
866  * @sdev: scsi device struct
867  * @starget: scsi target struct
868  */
869 static void
870 _scsih_display_enclosure_chassis_info(struct MPT3SAS_ADAPTER *ioc,
871 	struct _sas_device *sas_device, struct scsi_device *sdev,
872 	struct scsi_target *starget)
873 {
874 	if (sdev) {
875 		if (sas_device->enclosure_handle != 0)
876 			sdev_printk(KERN_INFO, sdev,
877 			    "enclosure logical id (0x%016llx), slot(%d) \n",
878 			    (unsigned long long)
879 			    sas_device->enclosure_logical_id,
880 			    sas_device->slot);
881 		if (sas_device->connector_name[0] != '\0')
882 			sdev_printk(KERN_INFO, sdev,
883 			    "enclosure level(0x%04x), connector name( %s)\n",
884 			    sas_device->enclosure_level,
885 			    sas_device->connector_name);
886 		if (sas_device->is_chassis_slot_valid)
887 			sdev_printk(KERN_INFO, sdev, "chassis slot(0x%04x)\n",
888 			    sas_device->chassis_slot);
889 	} else if (starget) {
890 		if (sas_device->enclosure_handle != 0)
891 			starget_printk(KERN_INFO, starget,
892 			    "enclosure logical id(0x%016llx), slot(%d) \n",
893 			    (unsigned long long)
894 			    sas_device->enclosure_logical_id,
895 			    sas_device->slot);
896 		if (sas_device->connector_name[0] != '\0')
897 			starget_printk(KERN_INFO, starget,
898 			    "enclosure level(0x%04x), connector name( %s)\n",
899 			    sas_device->enclosure_level,
900 			    sas_device->connector_name);
901 		if (sas_device->is_chassis_slot_valid)
902 			starget_printk(KERN_INFO, starget,
903 			    "chassis slot(0x%04x)\n",
904 			    sas_device->chassis_slot);
905 	} else {
906 		if (sas_device->enclosure_handle != 0)
907 			ioc_info(ioc, "enclosure logical id(0x%016llx), slot(%d)\n",
908 				 (u64)sas_device->enclosure_logical_id,
909 				 sas_device->slot);
910 		if (sas_device->connector_name[0] != '\0')
911 			ioc_info(ioc, "enclosure level(0x%04x), connector name( %s)\n",
912 				 sas_device->enclosure_level,
913 				 sas_device->connector_name);
914 		if (sas_device->is_chassis_slot_valid)
915 			ioc_info(ioc, "chassis slot(0x%04x)\n",
916 				 sas_device->chassis_slot);
917 	}
918 }
919 
920 /**
921  * _scsih_sas_device_remove - remove sas_device from list.
922  * @ioc: per adapter object
923  * @sas_device: the sas_device object
924  * Context: This function will acquire ioc->sas_device_lock.
925  *
926  * If sas_device is on the list, remove it and decrement its reference count.
927  */
928 static void
929 _scsih_sas_device_remove(struct MPT3SAS_ADAPTER *ioc,
930 	struct _sas_device *sas_device)
931 {
932 	unsigned long flags;
933 
934 	if (!sas_device)
935 		return;
936 	ioc_info(ioc, "removing handle(0x%04x), sas_addr(0x%016llx)\n",
937 		 sas_device->handle, (u64)sas_device->sas_address);
938 
939 	_scsih_display_enclosure_chassis_info(ioc, sas_device, NULL, NULL);
940 
941 	/*
942 	 * The lock serializes access to the list, but we still need to verify
943 	 * that nobody removed the entry while we were waiting on the lock.
944 	 */
945 	spin_lock_irqsave(&ioc->sas_device_lock, flags);
946 	if (!list_empty(&sas_device->list)) {
947 		list_del_init(&sas_device->list);
948 		sas_device_put(sas_device);
949 	}
950 	spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
951 }
952 
953 /**
954  * _scsih_device_remove_by_handle - removing device object by handle
955  * @ioc: per adapter object
956  * @handle: device handle
957  */
958 static void
959 _scsih_device_remove_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle)
960 {
961 	struct _sas_device *sas_device;
962 	unsigned long flags;
963 
964 	if (ioc->shost_recovery)
965 		return;
966 
967 	spin_lock_irqsave(&ioc->sas_device_lock, flags);
968 	sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle);
969 	if (sas_device) {
970 		list_del_init(&sas_device->list);
971 		sas_device_put(sas_device);
972 	}
973 	spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
974 	if (sas_device) {
975 		_scsih_remove_device(ioc, sas_device);
976 		sas_device_put(sas_device);
977 	}
978 }
979 
980 /**
981  * mpt3sas_device_remove_by_sas_address - removing device object by
982  *					sas address & port number
983  * @ioc: per adapter object
984  * @sas_address: device sas_address
985  * @port: hba port entry
986  *
987  * Return nothing.
988  */
989 void
990 mpt3sas_device_remove_by_sas_address(struct MPT3SAS_ADAPTER *ioc,
991 	u64 sas_address, struct hba_port *port)
992 {
993 	struct _sas_device *sas_device;
994 	unsigned long flags;
995 
996 	if (ioc->shost_recovery)
997 		return;
998 
999 	spin_lock_irqsave(&ioc->sas_device_lock, flags);
1000 	sas_device = __mpt3sas_get_sdev_by_addr(ioc, sas_address, port);
1001 	if (sas_device) {
1002 		list_del_init(&sas_device->list);
1003 		sas_device_put(sas_device);
1004 	}
1005 	spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
1006 	if (sas_device) {
1007 		_scsih_remove_device(ioc, sas_device);
1008 		sas_device_put(sas_device);
1009 	}
1010 }
1011 
1012 /**
1013  * _scsih_sas_device_add - insert sas_device to the list.
1014  * @ioc: per adapter object
1015  * @sas_device: the sas_device object
1016  * Context: This function will acquire ioc->sas_device_lock.
1017  *
1018  * Adding new object to the ioc->sas_device_list.
1019  */
1020 static void
1021 _scsih_sas_device_add(struct MPT3SAS_ADAPTER *ioc,
1022 	struct _sas_device *sas_device)
1023 {
1024 	unsigned long flags;
1025 
1026 	dewtprintk(ioc,
1027 		   ioc_info(ioc, "%s: handle(0x%04x), sas_addr(0x%016llx)\n",
1028 			    __func__, sas_device->handle,
1029 			    (u64)sas_device->sas_address));
1030 
1031 	dewtprintk(ioc, _scsih_display_enclosure_chassis_info(ioc, sas_device,
1032 	    NULL, NULL));
1033 
1034 	spin_lock_irqsave(&ioc->sas_device_lock, flags);
1035 	sas_device_get(sas_device);
1036 	list_add_tail(&sas_device->list, &ioc->sas_device_list);
1037 	spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
1038 
1039 	if (ioc->hide_drives) {
1040 		clear_bit(sas_device->handle, ioc->pend_os_device_add);
1041 		return;
1042 	}
1043 
1044 	if (!mpt3sas_transport_port_add(ioc, sas_device->handle,
1045 	     sas_device->sas_address_parent, sas_device->port)) {
1046 		_scsih_sas_device_remove(ioc, sas_device);
1047 	} else if (!sas_device->starget) {
1048 		/*
1049 		 * When asyn scanning is enabled, its not possible to remove
1050 		 * devices while scanning is turned on due to an oops in
1051 		 * scsi_sysfs_add_sdev()->add_device()->sysfs_addrm_start()
1052 		 */
1053 		if (!ioc->is_driver_loading) {
1054 			mpt3sas_transport_port_remove(ioc,
1055 			    sas_device->sas_address,
1056 			    sas_device->sas_address_parent,
1057 			    sas_device->port);
1058 			_scsih_sas_device_remove(ioc, sas_device);
1059 		}
1060 	} else
1061 		clear_bit(sas_device->handle, ioc->pend_os_device_add);
1062 }
1063 
1064 /**
1065  * _scsih_sas_device_init_add - insert sas_device to the list.
1066  * @ioc: per adapter object
1067  * @sas_device: the sas_device object
1068  * Context: This function will acquire ioc->sas_device_lock.
1069  *
1070  * Adding new object at driver load time to the ioc->sas_device_init_list.
1071  */
1072 static void
1073 _scsih_sas_device_init_add(struct MPT3SAS_ADAPTER *ioc,
1074 	struct _sas_device *sas_device)
1075 {
1076 	unsigned long flags;
1077 
1078 	dewtprintk(ioc,
1079 		   ioc_info(ioc, "%s: handle(0x%04x), sas_addr(0x%016llx)\n",
1080 			    __func__, sas_device->handle,
1081 			    (u64)sas_device->sas_address));
1082 
1083 	dewtprintk(ioc, _scsih_display_enclosure_chassis_info(ioc, sas_device,
1084 	    NULL, NULL));
1085 
1086 	spin_lock_irqsave(&ioc->sas_device_lock, flags);
1087 	sas_device_get(sas_device);
1088 	list_add_tail(&sas_device->list, &ioc->sas_device_init_list);
1089 	_scsih_determine_boot_device(ioc, sas_device, 0);
1090 	spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
1091 }
1092 
1093 
1094 static struct _pcie_device *
1095 __mpt3sas_get_pdev_by_wwid(struct MPT3SAS_ADAPTER *ioc, u64 wwid)
1096 {
1097 	struct _pcie_device *pcie_device;
1098 
1099 	assert_spin_locked(&ioc->pcie_device_lock);
1100 
1101 	list_for_each_entry(pcie_device, &ioc->pcie_device_list, list)
1102 		if (pcie_device->wwid == wwid)
1103 			goto found_device;
1104 
1105 	list_for_each_entry(pcie_device, &ioc->pcie_device_init_list, list)
1106 		if (pcie_device->wwid == wwid)
1107 			goto found_device;
1108 
1109 	return NULL;
1110 
1111 found_device:
1112 	pcie_device_get(pcie_device);
1113 	return pcie_device;
1114 }
1115 
1116 
1117 /**
1118  * mpt3sas_get_pdev_by_wwid - pcie device search
1119  * @ioc: per adapter object
1120  * @wwid: wwid
1121  *
1122  * Context: This function will acquire ioc->pcie_device_lock and will release
1123  * before returning the pcie_device object.
1124  *
1125  * This searches for pcie_device based on wwid, then return pcie_device object.
1126  */
1127 static struct _pcie_device *
1128 mpt3sas_get_pdev_by_wwid(struct MPT3SAS_ADAPTER *ioc, u64 wwid)
1129 {
1130 	struct _pcie_device *pcie_device;
1131 	unsigned long flags;
1132 
1133 	spin_lock_irqsave(&ioc->pcie_device_lock, flags);
1134 	pcie_device = __mpt3sas_get_pdev_by_wwid(ioc, wwid);
1135 	spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
1136 
1137 	return pcie_device;
1138 }
1139 
1140 
1141 static struct _pcie_device *
1142 __mpt3sas_get_pdev_by_idchannel(struct MPT3SAS_ADAPTER *ioc, int id,
1143 	int channel)
1144 {
1145 	struct _pcie_device *pcie_device;
1146 
1147 	assert_spin_locked(&ioc->pcie_device_lock);
1148 
1149 	list_for_each_entry(pcie_device, &ioc->pcie_device_list, list)
1150 		if (pcie_device->id == id && pcie_device->channel == channel)
1151 			goto found_device;
1152 
1153 	list_for_each_entry(pcie_device, &ioc->pcie_device_init_list, list)
1154 		if (pcie_device->id == id && pcie_device->channel == channel)
1155 			goto found_device;
1156 
1157 	return NULL;
1158 
1159 found_device:
1160 	pcie_device_get(pcie_device);
1161 	return pcie_device;
1162 }
1163 
1164 static struct _pcie_device *
1165 __mpt3sas_get_pdev_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle)
1166 {
1167 	struct _pcie_device *pcie_device;
1168 
1169 	assert_spin_locked(&ioc->pcie_device_lock);
1170 
1171 	list_for_each_entry(pcie_device, &ioc->pcie_device_list, list)
1172 		if (pcie_device->handle == handle)
1173 			goto found_device;
1174 
1175 	list_for_each_entry(pcie_device, &ioc->pcie_device_init_list, list)
1176 		if (pcie_device->handle == handle)
1177 			goto found_device;
1178 
1179 	return NULL;
1180 
1181 found_device:
1182 	pcie_device_get(pcie_device);
1183 	return pcie_device;
1184 }
1185 
1186 
1187 /**
1188  * mpt3sas_get_pdev_by_handle - pcie device search
1189  * @ioc: per adapter object
1190  * @handle: Firmware device handle
1191  *
1192  * Context: This function will acquire ioc->pcie_device_lock and will release
1193  * before returning the pcie_device object.
1194  *
1195  * This searches for pcie_device based on handle, then return pcie_device
1196  * object.
1197  */
1198 struct _pcie_device *
1199 mpt3sas_get_pdev_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle)
1200 {
1201 	struct _pcie_device *pcie_device;
1202 	unsigned long flags;
1203 
1204 	spin_lock_irqsave(&ioc->pcie_device_lock, flags);
1205 	pcie_device = __mpt3sas_get_pdev_by_handle(ioc, handle);
1206 	spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
1207 
1208 	return pcie_device;
1209 }
1210 
1211 /**
1212  * _scsih_set_nvme_max_shutdown_latency - Update max_shutdown_latency.
1213  * @ioc: per adapter object
1214  * Context: This function will acquire ioc->pcie_device_lock
1215  *
1216  * Update ioc->max_shutdown_latency to that NVMe drives RTD3 Entry Latency
1217  * which has reported maximum among all available NVMe drives.
1218  * Minimum max_shutdown_latency will be six seconds.
1219  */
1220 static void
1221 _scsih_set_nvme_max_shutdown_latency(struct MPT3SAS_ADAPTER *ioc)
1222 {
1223 	struct _pcie_device *pcie_device;
1224 	unsigned long flags;
1225 	u16 shutdown_latency = IO_UNIT_CONTROL_SHUTDOWN_TIMEOUT;
1226 
1227 	spin_lock_irqsave(&ioc->pcie_device_lock, flags);
1228 	list_for_each_entry(pcie_device, &ioc->pcie_device_list, list) {
1229 		if (pcie_device->shutdown_latency) {
1230 			if (shutdown_latency < pcie_device->shutdown_latency)
1231 				shutdown_latency =
1232 					pcie_device->shutdown_latency;
1233 		}
1234 	}
1235 	ioc->max_shutdown_latency = shutdown_latency;
1236 	spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
1237 }
1238 
1239 /**
1240  * _scsih_pcie_device_remove - remove pcie_device from list.
1241  * @ioc: per adapter object
1242  * @pcie_device: the pcie_device object
1243  * Context: This function will acquire ioc->pcie_device_lock.
1244  *
1245  * If pcie_device is on the list, remove it and decrement its reference count.
1246  */
1247 static void
1248 _scsih_pcie_device_remove(struct MPT3SAS_ADAPTER *ioc,
1249 	struct _pcie_device *pcie_device)
1250 {
1251 	unsigned long flags;
1252 	int was_on_pcie_device_list = 0;
1253 	u8 update_latency = 0;
1254 
1255 	if (!pcie_device)
1256 		return;
1257 	ioc_info(ioc, "removing handle(0x%04x), wwid(0x%016llx)\n",
1258 		 pcie_device->handle, (u64)pcie_device->wwid);
1259 	if (pcie_device->enclosure_handle != 0)
1260 		ioc_info(ioc, "removing enclosure logical id(0x%016llx), slot(%d)\n",
1261 			 (u64)pcie_device->enclosure_logical_id,
1262 			 pcie_device->slot);
1263 	if (pcie_device->connector_name[0] != '\0')
1264 		ioc_info(ioc, "removing enclosure level(0x%04x), connector name( %s)\n",
1265 			 pcie_device->enclosure_level,
1266 			 pcie_device->connector_name);
1267 
1268 	spin_lock_irqsave(&ioc->pcie_device_lock, flags);
1269 	if (!list_empty(&pcie_device->list)) {
1270 		list_del_init(&pcie_device->list);
1271 		was_on_pcie_device_list = 1;
1272 	}
1273 	if (pcie_device->shutdown_latency == ioc->max_shutdown_latency)
1274 		update_latency = 1;
1275 	spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
1276 	if (was_on_pcie_device_list) {
1277 		kfree(pcie_device->serial_number);
1278 		pcie_device_put(pcie_device);
1279 	}
1280 
1281 	/*
1282 	 * This device's RTD3 Entry Latency matches IOC's
1283 	 * max_shutdown_latency. Recalculate IOC's max_shutdown_latency
1284 	 * from the available drives as current drive is getting removed.
1285 	 */
1286 	if (update_latency)
1287 		_scsih_set_nvme_max_shutdown_latency(ioc);
1288 }
1289 
1290 
1291 /**
1292  * _scsih_pcie_device_remove_by_handle - removing pcie device object by handle
1293  * @ioc: per adapter object
1294  * @handle: device handle
1295  */
1296 static void
1297 _scsih_pcie_device_remove_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle)
1298 {
1299 	struct _pcie_device *pcie_device;
1300 	unsigned long flags;
1301 	int was_on_pcie_device_list = 0;
1302 	u8 update_latency = 0;
1303 
1304 	if (ioc->shost_recovery)
1305 		return;
1306 
1307 	spin_lock_irqsave(&ioc->pcie_device_lock, flags);
1308 	pcie_device = __mpt3sas_get_pdev_by_handle(ioc, handle);
1309 	if (pcie_device) {
1310 		if (!list_empty(&pcie_device->list)) {
1311 			list_del_init(&pcie_device->list);
1312 			was_on_pcie_device_list = 1;
1313 			pcie_device_put(pcie_device);
1314 		}
1315 		if (pcie_device->shutdown_latency == ioc->max_shutdown_latency)
1316 			update_latency = 1;
1317 	}
1318 	spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
1319 	if (was_on_pcie_device_list) {
1320 		_scsih_pcie_device_remove_from_sml(ioc, pcie_device);
1321 		pcie_device_put(pcie_device);
1322 	}
1323 
1324 	/*
1325 	 * This device's RTD3 Entry Latency matches IOC's
1326 	 * max_shutdown_latency. Recalculate IOC's max_shutdown_latency
1327 	 * from the available drives as current drive is getting removed.
1328 	 */
1329 	if (update_latency)
1330 		_scsih_set_nvme_max_shutdown_latency(ioc);
1331 }
1332 
1333 /**
1334  * _scsih_pcie_device_add - add pcie_device object
1335  * @ioc: per adapter object
1336  * @pcie_device: pcie_device object
1337  *
1338  * This is added to the pcie_device_list link list.
1339  */
1340 static void
1341 _scsih_pcie_device_add(struct MPT3SAS_ADAPTER *ioc,
1342 	struct _pcie_device *pcie_device)
1343 {
1344 	unsigned long flags;
1345 
1346 	dewtprintk(ioc,
1347 		   ioc_info(ioc, "%s: handle (0x%04x), wwid(0x%016llx)\n",
1348 			    __func__,
1349 			    pcie_device->handle, (u64)pcie_device->wwid));
1350 	if (pcie_device->enclosure_handle != 0)
1351 		dewtprintk(ioc,
1352 			   ioc_info(ioc, "%s: enclosure logical id(0x%016llx), slot( %d)\n",
1353 				    __func__,
1354 				    (u64)pcie_device->enclosure_logical_id,
1355 				    pcie_device->slot));
1356 	if (pcie_device->connector_name[0] != '\0')
1357 		dewtprintk(ioc,
1358 			   ioc_info(ioc, "%s: enclosure level(0x%04x), connector name( %s)\n",
1359 				    __func__, pcie_device->enclosure_level,
1360 				    pcie_device->connector_name));
1361 
1362 	spin_lock_irqsave(&ioc->pcie_device_lock, flags);
1363 	pcie_device_get(pcie_device);
1364 	list_add_tail(&pcie_device->list, &ioc->pcie_device_list);
1365 	spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
1366 
1367 	if (pcie_device->access_status ==
1368 	    MPI26_PCIEDEV0_ASTATUS_DEVICE_BLOCKED) {
1369 		clear_bit(pcie_device->handle, ioc->pend_os_device_add);
1370 		return;
1371 	}
1372 	if (scsi_add_device(ioc->shost, PCIE_CHANNEL, pcie_device->id, 0)) {
1373 		_scsih_pcie_device_remove(ioc, pcie_device);
1374 	} else if (!pcie_device->starget) {
1375 		if (!ioc->is_driver_loading) {
1376 /*TODO-- Need to find out whether this condition will occur or not*/
1377 			clear_bit(pcie_device->handle, ioc->pend_os_device_add);
1378 		}
1379 	} else
1380 		clear_bit(pcie_device->handle, ioc->pend_os_device_add);
1381 }
1382 
1383 /*
1384  * _scsih_pcie_device_init_add - insert pcie_device to the init list.
1385  * @ioc: per adapter object
1386  * @pcie_device: the pcie_device object
1387  * Context: This function will acquire ioc->pcie_device_lock.
1388  *
1389  * Adding new object at driver load time to the ioc->pcie_device_init_list.
1390  */
1391 static void
1392 _scsih_pcie_device_init_add(struct MPT3SAS_ADAPTER *ioc,
1393 				struct _pcie_device *pcie_device)
1394 {
1395 	unsigned long flags;
1396 
1397 	dewtprintk(ioc,
1398 		   ioc_info(ioc, "%s: handle (0x%04x), wwid(0x%016llx)\n",
1399 			    __func__,
1400 			    pcie_device->handle, (u64)pcie_device->wwid));
1401 	if (pcie_device->enclosure_handle != 0)
1402 		dewtprintk(ioc,
1403 			   ioc_info(ioc, "%s: enclosure logical id(0x%016llx), slot( %d)\n",
1404 				    __func__,
1405 				    (u64)pcie_device->enclosure_logical_id,
1406 				    pcie_device->slot));
1407 	if (pcie_device->connector_name[0] != '\0')
1408 		dewtprintk(ioc,
1409 			   ioc_info(ioc, "%s: enclosure level(0x%04x), connector name( %s)\n",
1410 				    __func__, pcie_device->enclosure_level,
1411 				    pcie_device->connector_name));
1412 
1413 	spin_lock_irqsave(&ioc->pcie_device_lock, flags);
1414 	pcie_device_get(pcie_device);
1415 	list_add_tail(&pcie_device->list, &ioc->pcie_device_init_list);
1416 	if (pcie_device->access_status !=
1417 	    MPI26_PCIEDEV0_ASTATUS_DEVICE_BLOCKED)
1418 		_scsih_determine_boot_device(ioc, pcie_device, PCIE_CHANNEL);
1419 	spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
1420 }
1421 /**
1422  * _scsih_raid_device_find_by_id - raid device search
1423  * @ioc: per adapter object
1424  * @id: sas device target id
1425  * @channel: sas device channel
1426  * Context: Calling function should acquire ioc->raid_device_lock
1427  *
1428  * This searches for raid_device based on target id, then return raid_device
1429  * object.
1430  */
1431 static struct _raid_device *
1432 _scsih_raid_device_find_by_id(struct MPT3SAS_ADAPTER *ioc, int id, int channel)
1433 {
1434 	struct _raid_device *raid_device, *r;
1435 
1436 	r = NULL;
1437 	list_for_each_entry(raid_device, &ioc->raid_device_list, list) {
1438 		if (raid_device->id == id && raid_device->channel == channel) {
1439 			r = raid_device;
1440 			goto out;
1441 		}
1442 	}
1443 
1444  out:
1445 	return r;
1446 }
1447 
1448 /**
1449  * mpt3sas_raid_device_find_by_handle - raid device search
1450  * @ioc: per adapter object
1451  * @handle: sas device handle (assigned by firmware)
1452  * Context: Calling function should acquire ioc->raid_device_lock
1453  *
1454  * This searches for raid_device based on handle, then return raid_device
1455  * object.
1456  */
1457 struct _raid_device *
1458 mpt3sas_raid_device_find_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle)
1459 {
1460 	struct _raid_device *raid_device, *r;
1461 
1462 	r = NULL;
1463 	list_for_each_entry(raid_device, &ioc->raid_device_list, list) {
1464 		if (raid_device->handle != handle)
1465 			continue;
1466 		r = raid_device;
1467 		goto out;
1468 	}
1469 
1470  out:
1471 	return r;
1472 }
1473 
1474 /**
1475  * _scsih_raid_device_find_by_wwid - raid device search
1476  * @ioc: per adapter object
1477  * @wwid: ?
1478  * Context: Calling function should acquire ioc->raid_device_lock
1479  *
1480  * This searches for raid_device based on wwid, then return raid_device
1481  * object.
1482  */
1483 static struct _raid_device *
1484 _scsih_raid_device_find_by_wwid(struct MPT3SAS_ADAPTER *ioc, u64 wwid)
1485 {
1486 	struct _raid_device *raid_device, *r;
1487 
1488 	r = NULL;
1489 	list_for_each_entry(raid_device, &ioc->raid_device_list, list) {
1490 		if (raid_device->wwid != wwid)
1491 			continue;
1492 		r = raid_device;
1493 		goto out;
1494 	}
1495 
1496  out:
1497 	return r;
1498 }
1499 
1500 /**
1501  * _scsih_raid_device_add - add raid_device object
1502  * @ioc: per adapter object
1503  * @raid_device: raid_device object
1504  *
1505  * This is added to the raid_device_list link list.
1506  */
1507 static void
1508 _scsih_raid_device_add(struct MPT3SAS_ADAPTER *ioc,
1509 	struct _raid_device *raid_device)
1510 {
1511 	unsigned long flags;
1512 
1513 	dewtprintk(ioc,
1514 		   ioc_info(ioc, "%s: handle(0x%04x), wwid(0x%016llx)\n",
1515 			    __func__,
1516 			    raid_device->handle, (u64)raid_device->wwid));
1517 
1518 	spin_lock_irqsave(&ioc->raid_device_lock, flags);
1519 	list_add_tail(&raid_device->list, &ioc->raid_device_list);
1520 	spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
1521 }
1522 
1523 /**
1524  * _scsih_raid_device_remove - delete raid_device object
1525  * @ioc: per adapter object
1526  * @raid_device: raid_device object
1527  *
1528  */
1529 static void
1530 _scsih_raid_device_remove(struct MPT3SAS_ADAPTER *ioc,
1531 	struct _raid_device *raid_device)
1532 {
1533 	unsigned long flags;
1534 
1535 	spin_lock_irqsave(&ioc->raid_device_lock, flags);
1536 	list_del(&raid_device->list);
1537 	kfree(raid_device);
1538 	spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
1539 }
1540 
1541 /**
1542  * mpt3sas_scsih_expander_find_by_handle - expander device search
1543  * @ioc: per adapter object
1544  * @handle: expander handle (assigned by firmware)
1545  * Context: Calling function should acquire ioc->sas_device_lock
1546  *
1547  * This searches for expander device based on handle, then returns the
1548  * sas_node object.
1549  */
1550 struct _sas_node *
1551 mpt3sas_scsih_expander_find_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle)
1552 {
1553 	struct _sas_node *sas_expander, *r;
1554 
1555 	r = NULL;
1556 	list_for_each_entry(sas_expander, &ioc->sas_expander_list, list) {
1557 		if (sas_expander->handle != handle)
1558 			continue;
1559 		r = sas_expander;
1560 		goto out;
1561 	}
1562  out:
1563 	return r;
1564 }
1565 
1566 /**
1567  * mpt3sas_scsih_enclosure_find_by_handle - exclosure device search
1568  * @ioc: per adapter object
1569  * @handle: enclosure handle (assigned by firmware)
1570  * Context: Calling function should acquire ioc->sas_device_lock
1571  *
1572  * This searches for enclosure device based on handle, then returns the
1573  * enclosure object.
1574  */
1575 static struct _enclosure_node *
1576 mpt3sas_scsih_enclosure_find_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle)
1577 {
1578 	struct _enclosure_node *enclosure_dev, *r;
1579 
1580 	r = NULL;
1581 	list_for_each_entry(enclosure_dev, &ioc->enclosure_list, list) {
1582 		if (le16_to_cpu(enclosure_dev->pg0.EnclosureHandle) != handle)
1583 			continue;
1584 		r = enclosure_dev;
1585 		goto out;
1586 	}
1587 out:
1588 	return r;
1589 }
1590 /**
1591  * mpt3sas_scsih_expander_find_by_sas_address - expander device search
1592  * @ioc: per adapter object
1593  * @sas_address: sas address
1594  * @port: hba port entry
1595  * Context: Calling function should acquire ioc->sas_node_lock.
1596  *
1597  * This searches for expander device based on sas_address & port number,
1598  * then returns the sas_node object.
1599  */
1600 struct _sas_node *
1601 mpt3sas_scsih_expander_find_by_sas_address(struct MPT3SAS_ADAPTER *ioc,
1602 	u64 sas_address, struct hba_port *port)
1603 {
1604 	struct _sas_node *sas_expander, *r = NULL;
1605 
1606 	if (!port)
1607 		return r;
1608 
1609 	list_for_each_entry(sas_expander, &ioc->sas_expander_list, list) {
1610 		if (sas_expander->sas_address != sas_address)
1611 			continue;
1612 		if (sas_expander->port != port)
1613 			continue;
1614 		r = sas_expander;
1615 		goto out;
1616 	}
1617  out:
1618 	return r;
1619 }
1620 
1621 /**
1622  * _scsih_expander_node_add - insert expander device to the list.
1623  * @ioc: per adapter object
1624  * @sas_expander: the sas_device object
1625  * Context: This function will acquire ioc->sas_node_lock.
1626  *
1627  * Adding new object to the ioc->sas_expander_list.
1628  */
1629 static void
1630 _scsih_expander_node_add(struct MPT3SAS_ADAPTER *ioc,
1631 	struct _sas_node *sas_expander)
1632 {
1633 	unsigned long flags;
1634 
1635 	spin_lock_irqsave(&ioc->sas_node_lock, flags);
1636 	list_add_tail(&sas_expander->list, &ioc->sas_expander_list);
1637 	spin_unlock_irqrestore(&ioc->sas_node_lock, flags);
1638 }
1639 
1640 /**
1641  * _scsih_is_end_device - determines if device is an end device
1642  * @device_info: bitfield providing information about the device.
1643  * Context: none
1644  *
1645  * Return: 1 if end device.
1646  */
1647 static int
1648 _scsih_is_end_device(u32 device_info)
1649 {
1650 	if (device_info & MPI2_SAS_DEVICE_INFO_END_DEVICE &&
1651 		((device_info & MPI2_SAS_DEVICE_INFO_SSP_TARGET) |
1652 		(device_info & MPI2_SAS_DEVICE_INFO_STP_TARGET) |
1653 		(device_info & MPI2_SAS_DEVICE_INFO_SATA_DEVICE)))
1654 		return 1;
1655 	else
1656 		return 0;
1657 }
1658 
1659 /**
1660  * _scsih_is_nvme_pciescsi_device - determines if
1661  *			device is an pcie nvme/scsi device
1662  * @device_info: bitfield providing information about the device.
1663  * Context: none
1664  *
1665  * Returns 1 if device is pcie device type nvme/scsi.
1666  */
1667 static int
1668 _scsih_is_nvme_pciescsi_device(u32 device_info)
1669 {
1670 	if (((device_info & MPI26_PCIE_DEVINFO_MASK_DEVICE_TYPE)
1671 	    == MPI26_PCIE_DEVINFO_NVME) ||
1672 	    ((device_info & MPI26_PCIE_DEVINFO_MASK_DEVICE_TYPE)
1673 	    == MPI26_PCIE_DEVINFO_SCSI))
1674 		return 1;
1675 	else
1676 		return 0;
1677 }
1678 
1679 /**
1680  * _scsih_scsi_lookup_find_by_target - search for matching channel:id
1681  * @ioc: per adapter object
1682  * @id: target id
1683  * @channel: channel
1684  * Context: This function will acquire ioc->scsi_lookup_lock.
1685  *
1686  * This will search for a matching channel:id in the scsi_lookup array,
1687  * returning 1 if found.
1688  */
1689 static u8
1690 _scsih_scsi_lookup_find_by_target(struct MPT3SAS_ADAPTER *ioc, int id,
1691 	int channel)
1692 {
1693 	int smid;
1694 	struct scsi_cmnd *scmd;
1695 
1696 	for (smid = 1;
1697 	     smid <= ioc->shost->can_queue; smid++) {
1698 		scmd = mpt3sas_scsih_scsi_lookup_get(ioc, smid);
1699 		if (!scmd)
1700 			continue;
1701 		if (scmd->device->id == id &&
1702 		    scmd->device->channel == channel)
1703 			return 1;
1704 	}
1705 	return 0;
1706 }
1707 
1708 /**
1709  * _scsih_scsi_lookup_find_by_lun - search for matching channel:id:lun
1710  * @ioc: per adapter object
1711  * @id: target id
1712  * @lun: lun number
1713  * @channel: channel
1714  * Context: This function will acquire ioc->scsi_lookup_lock.
1715  *
1716  * This will search for a matching channel:id:lun in the scsi_lookup array,
1717  * returning 1 if found.
1718  */
1719 static u8
1720 _scsih_scsi_lookup_find_by_lun(struct MPT3SAS_ADAPTER *ioc, int id,
1721 	unsigned int lun, int channel)
1722 {
1723 	int smid;
1724 	struct scsi_cmnd *scmd;
1725 
1726 	for (smid = 1; smid <= ioc->shost->can_queue; smid++) {
1727 
1728 		scmd = mpt3sas_scsih_scsi_lookup_get(ioc, smid);
1729 		if (!scmd)
1730 			continue;
1731 		if (scmd->device->id == id &&
1732 		    scmd->device->channel == channel &&
1733 		    scmd->device->lun == lun)
1734 			return 1;
1735 	}
1736 	return 0;
1737 }
1738 
1739 /**
1740  * mpt3sas_scsih_scsi_lookup_get - returns scmd entry
1741  * @ioc: per adapter object
1742  * @smid: system request message index
1743  *
1744  * Return: the smid stored scmd pointer.
1745  * Then will dereference the stored scmd pointer.
1746  */
1747 struct scsi_cmnd *
1748 mpt3sas_scsih_scsi_lookup_get(struct MPT3SAS_ADAPTER *ioc, u16 smid)
1749 {
1750 	struct scsi_cmnd *scmd = NULL;
1751 	struct scsiio_tracker *st;
1752 	Mpi25SCSIIORequest_t *mpi_request;
1753 	u16 tag = smid - 1;
1754 
1755 	if (smid > 0  &&
1756 	    smid <= ioc->scsiio_depth - INTERNAL_SCSIIO_CMDS_COUNT) {
1757 		u32 unique_tag =
1758 		    ioc->io_queue_num[tag] << BLK_MQ_UNIQUE_TAG_BITS | tag;
1759 
1760 		mpi_request = mpt3sas_base_get_msg_frame(ioc, smid);
1761 
1762 		/*
1763 		 * If SCSI IO request is outstanding at driver level then
1764 		 * DevHandle filed must be non-zero. If DevHandle is zero
1765 		 * then it means that this smid is free at driver level,
1766 		 * so return NULL.
1767 		 */
1768 		if (!mpi_request->DevHandle)
1769 			return scmd;
1770 
1771 		scmd = scsi_host_find_tag(ioc->shost, unique_tag);
1772 		if (scmd) {
1773 			st = scsi_cmd_priv(scmd);
1774 			if (st->cb_idx == 0xFF || st->smid == 0)
1775 				scmd = NULL;
1776 		}
1777 	}
1778 	return scmd;
1779 }
1780 
1781 /**
1782  * scsih_change_queue_depth - setting device queue depth
1783  * @sdev: scsi device struct
1784  * @qdepth: requested queue depth
1785  *
1786  * Return: queue depth.
1787  */
1788 static int
1789 scsih_change_queue_depth(struct scsi_device *sdev, int qdepth)
1790 {
1791 	struct Scsi_Host *shost = sdev->host;
1792 	int max_depth;
1793 	struct MPT3SAS_ADAPTER *ioc = shost_priv(shost);
1794 	struct MPT3SAS_DEVICE *sas_device_priv_data;
1795 	struct MPT3SAS_TARGET *sas_target_priv_data;
1796 	struct _sas_device *sas_device;
1797 	unsigned long flags;
1798 
1799 	max_depth = shost->can_queue;
1800 
1801 	/*
1802 	 * limit max device queue for SATA to 32 if enable_sdev_max_qd
1803 	 * is disabled.
1804 	 */
1805 	if (ioc->enable_sdev_max_qd || ioc->is_gen35_ioc)
1806 		goto not_sata;
1807 
1808 	sas_device_priv_data = sdev->hostdata;
1809 	if (!sas_device_priv_data)
1810 		goto not_sata;
1811 	sas_target_priv_data = sas_device_priv_data->sas_target;
1812 	if (!sas_target_priv_data)
1813 		goto not_sata;
1814 	if ((sas_target_priv_data->flags & MPT_TARGET_FLAGS_VOLUME))
1815 		goto not_sata;
1816 
1817 	spin_lock_irqsave(&ioc->sas_device_lock, flags);
1818 	sas_device = __mpt3sas_get_sdev_from_target(ioc, sas_target_priv_data);
1819 	if (sas_device) {
1820 		if (sas_device->device_info & MPI2_SAS_DEVICE_INFO_SATA_DEVICE)
1821 			max_depth = MPT3SAS_SATA_QUEUE_DEPTH;
1822 
1823 		sas_device_put(sas_device);
1824 	}
1825 	spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
1826 
1827  not_sata:
1828 
1829 	if (!sdev->tagged_supported)
1830 		max_depth = 1;
1831 	if (qdepth > max_depth)
1832 		qdepth = max_depth;
1833 	scsi_change_queue_depth(sdev, qdepth);
1834 	sdev_printk(KERN_INFO, sdev,
1835 	    "qdepth(%d), tagged(%d), scsi_level(%d), cmd_que(%d)\n",
1836 	    sdev->queue_depth, sdev->tagged_supported,
1837 	    sdev->scsi_level, ((sdev->inquiry[7] & 2) >> 1));
1838 	return sdev->queue_depth;
1839 }
1840 
1841 /**
1842  * mpt3sas_scsih_change_queue_depth - setting device queue depth
1843  * @sdev: scsi device struct
1844  * @qdepth: requested queue depth
1845  *
1846  * Returns nothing.
1847  */
1848 void
1849 mpt3sas_scsih_change_queue_depth(struct scsi_device *sdev, int qdepth)
1850 {
1851 	struct Scsi_Host *shost = sdev->host;
1852 	struct MPT3SAS_ADAPTER *ioc = shost_priv(shost);
1853 
1854 	if (ioc->enable_sdev_max_qd)
1855 		qdepth = shost->can_queue;
1856 
1857 	scsih_change_queue_depth(sdev, qdepth);
1858 }
1859 
1860 /**
1861  * scsih_target_alloc - target add routine
1862  * @starget: scsi target struct
1863  *
1864  * Return: 0 if ok. Any other return is assumed to be an error and
1865  * the device is ignored.
1866  */
1867 static int
1868 scsih_target_alloc(struct scsi_target *starget)
1869 {
1870 	struct Scsi_Host *shost = dev_to_shost(&starget->dev);
1871 	struct MPT3SAS_ADAPTER *ioc = shost_priv(shost);
1872 	struct MPT3SAS_TARGET *sas_target_priv_data;
1873 	struct _sas_device *sas_device;
1874 	struct _raid_device *raid_device;
1875 	struct _pcie_device *pcie_device;
1876 	unsigned long flags;
1877 	struct sas_rphy *rphy;
1878 
1879 	sas_target_priv_data = kzalloc(sizeof(*sas_target_priv_data),
1880 				       GFP_KERNEL);
1881 	if (!sas_target_priv_data)
1882 		return -ENOMEM;
1883 
1884 	starget->hostdata = sas_target_priv_data;
1885 	sas_target_priv_data->starget = starget;
1886 	sas_target_priv_data->handle = MPT3SAS_INVALID_DEVICE_HANDLE;
1887 
1888 	/* RAID volumes */
1889 	if (starget->channel == RAID_CHANNEL) {
1890 		spin_lock_irqsave(&ioc->raid_device_lock, flags);
1891 		raid_device = _scsih_raid_device_find_by_id(ioc, starget->id,
1892 		    starget->channel);
1893 		if (raid_device) {
1894 			sas_target_priv_data->handle = raid_device->handle;
1895 			sas_target_priv_data->sas_address = raid_device->wwid;
1896 			sas_target_priv_data->flags |= MPT_TARGET_FLAGS_VOLUME;
1897 			if (ioc->is_warpdrive)
1898 				sas_target_priv_data->raid_device = raid_device;
1899 			raid_device->starget = starget;
1900 		}
1901 		spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
1902 		return 0;
1903 	}
1904 
1905 	/* PCIe devices */
1906 	if (starget->channel == PCIE_CHANNEL) {
1907 		spin_lock_irqsave(&ioc->pcie_device_lock, flags);
1908 		pcie_device = __mpt3sas_get_pdev_by_idchannel(ioc, starget->id,
1909 			starget->channel);
1910 		if (pcie_device) {
1911 			sas_target_priv_data->handle = pcie_device->handle;
1912 			sas_target_priv_data->sas_address = pcie_device->wwid;
1913 			sas_target_priv_data->port = NULL;
1914 			sas_target_priv_data->pcie_dev = pcie_device;
1915 			pcie_device->starget = starget;
1916 			pcie_device->id = starget->id;
1917 			pcie_device->channel = starget->channel;
1918 			sas_target_priv_data->flags |=
1919 				MPT_TARGET_FLAGS_PCIE_DEVICE;
1920 			if (pcie_device->fast_path)
1921 				sas_target_priv_data->flags |=
1922 					MPT_TARGET_FASTPATH_IO;
1923 		}
1924 		spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
1925 		return 0;
1926 	}
1927 
1928 	/* sas/sata devices */
1929 	spin_lock_irqsave(&ioc->sas_device_lock, flags);
1930 	rphy = dev_to_rphy(starget->dev.parent);
1931 	sas_device = __mpt3sas_get_sdev_by_rphy(ioc, rphy);
1932 
1933 	if (sas_device) {
1934 		sas_target_priv_data->handle = sas_device->handle;
1935 		sas_target_priv_data->sas_address = sas_device->sas_address;
1936 		sas_target_priv_data->port = sas_device->port;
1937 		sas_target_priv_data->sas_dev = sas_device;
1938 		sas_device->starget = starget;
1939 		sas_device->id = starget->id;
1940 		sas_device->channel = starget->channel;
1941 		if (test_bit(sas_device->handle, ioc->pd_handles))
1942 			sas_target_priv_data->flags |=
1943 			    MPT_TARGET_FLAGS_RAID_COMPONENT;
1944 		if (sas_device->fast_path)
1945 			sas_target_priv_data->flags |=
1946 					MPT_TARGET_FASTPATH_IO;
1947 	}
1948 	spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
1949 
1950 	return 0;
1951 }
1952 
1953 /**
1954  * scsih_target_destroy - target destroy routine
1955  * @starget: scsi target struct
1956  */
1957 static void
1958 scsih_target_destroy(struct scsi_target *starget)
1959 {
1960 	struct Scsi_Host *shost = dev_to_shost(&starget->dev);
1961 	struct MPT3SAS_ADAPTER *ioc = shost_priv(shost);
1962 	struct MPT3SAS_TARGET *sas_target_priv_data;
1963 	struct _sas_device *sas_device;
1964 	struct _raid_device *raid_device;
1965 	struct _pcie_device *pcie_device;
1966 	unsigned long flags;
1967 
1968 	sas_target_priv_data = starget->hostdata;
1969 	if (!sas_target_priv_data)
1970 		return;
1971 
1972 	if (starget->channel == RAID_CHANNEL) {
1973 		spin_lock_irqsave(&ioc->raid_device_lock, flags);
1974 		raid_device = _scsih_raid_device_find_by_id(ioc, starget->id,
1975 		    starget->channel);
1976 		if (raid_device) {
1977 			raid_device->starget = NULL;
1978 			raid_device->sdev = NULL;
1979 		}
1980 		spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
1981 		goto out;
1982 	}
1983 
1984 	if (starget->channel == PCIE_CHANNEL) {
1985 		spin_lock_irqsave(&ioc->pcie_device_lock, flags);
1986 		pcie_device = __mpt3sas_get_pdev_from_target(ioc,
1987 							sas_target_priv_data);
1988 		if (pcie_device && (pcie_device->starget == starget) &&
1989 			(pcie_device->id == starget->id) &&
1990 			(pcie_device->channel == starget->channel))
1991 			pcie_device->starget = NULL;
1992 
1993 		if (pcie_device) {
1994 			/*
1995 			 * Corresponding get() is in _scsih_target_alloc()
1996 			 */
1997 			sas_target_priv_data->pcie_dev = NULL;
1998 			pcie_device_put(pcie_device);
1999 			pcie_device_put(pcie_device);
2000 		}
2001 		spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
2002 		goto out;
2003 	}
2004 
2005 	spin_lock_irqsave(&ioc->sas_device_lock, flags);
2006 	sas_device = __mpt3sas_get_sdev_from_target(ioc, sas_target_priv_data);
2007 	if (sas_device && (sas_device->starget == starget) &&
2008 	    (sas_device->id == starget->id) &&
2009 	    (sas_device->channel == starget->channel))
2010 		sas_device->starget = NULL;
2011 
2012 	if (sas_device) {
2013 		/*
2014 		 * Corresponding get() is in _scsih_target_alloc()
2015 		 */
2016 		sas_target_priv_data->sas_dev = NULL;
2017 		sas_device_put(sas_device);
2018 
2019 		sas_device_put(sas_device);
2020 	}
2021 	spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
2022 
2023  out:
2024 	kfree(sas_target_priv_data);
2025 	starget->hostdata = NULL;
2026 }
2027 
2028 /**
2029  * scsih_slave_alloc - device add routine
2030  * @sdev: scsi device struct
2031  *
2032  * Return: 0 if ok. Any other return is assumed to be an error and
2033  * the device is ignored.
2034  */
2035 static int
2036 scsih_slave_alloc(struct scsi_device *sdev)
2037 {
2038 	struct Scsi_Host *shost;
2039 	struct MPT3SAS_ADAPTER *ioc;
2040 	struct MPT3SAS_TARGET *sas_target_priv_data;
2041 	struct MPT3SAS_DEVICE *sas_device_priv_data;
2042 	struct scsi_target *starget;
2043 	struct _raid_device *raid_device;
2044 	struct _sas_device *sas_device;
2045 	struct _pcie_device *pcie_device;
2046 	unsigned long flags;
2047 
2048 	sas_device_priv_data = kzalloc(sizeof(*sas_device_priv_data),
2049 				       GFP_KERNEL);
2050 	if (!sas_device_priv_data)
2051 		return -ENOMEM;
2052 
2053 	sas_device_priv_data->lun = sdev->lun;
2054 	sas_device_priv_data->flags = MPT_DEVICE_FLAGS_INIT;
2055 
2056 	starget = scsi_target(sdev);
2057 	sas_target_priv_data = starget->hostdata;
2058 	sas_target_priv_data->num_luns++;
2059 	sas_device_priv_data->sas_target = sas_target_priv_data;
2060 	sdev->hostdata = sas_device_priv_data;
2061 	if ((sas_target_priv_data->flags & MPT_TARGET_FLAGS_RAID_COMPONENT))
2062 		sdev->no_uld_attach = 1;
2063 
2064 	shost = dev_to_shost(&starget->dev);
2065 	ioc = shost_priv(shost);
2066 	if (starget->channel == RAID_CHANNEL) {
2067 		spin_lock_irqsave(&ioc->raid_device_lock, flags);
2068 		raid_device = _scsih_raid_device_find_by_id(ioc,
2069 		    starget->id, starget->channel);
2070 		if (raid_device)
2071 			raid_device->sdev = sdev; /* raid is single lun */
2072 		spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
2073 	}
2074 	if (starget->channel == PCIE_CHANNEL) {
2075 		spin_lock_irqsave(&ioc->pcie_device_lock, flags);
2076 		pcie_device = __mpt3sas_get_pdev_by_wwid(ioc,
2077 				sas_target_priv_data->sas_address);
2078 		if (pcie_device && (pcie_device->starget == NULL)) {
2079 			sdev_printk(KERN_INFO, sdev,
2080 			    "%s : pcie_device->starget set to starget @ %d\n",
2081 			    __func__, __LINE__);
2082 			pcie_device->starget = starget;
2083 		}
2084 
2085 		if (pcie_device)
2086 			pcie_device_put(pcie_device);
2087 		spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
2088 
2089 	} else  if (!(sas_target_priv_data->flags & MPT_TARGET_FLAGS_VOLUME)) {
2090 		spin_lock_irqsave(&ioc->sas_device_lock, flags);
2091 		sas_device = __mpt3sas_get_sdev_by_addr(ioc,
2092 		    sas_target_priv_data->sas_address,
2093 		    sas_target_priv_data->port);
2094 		if (sas_device && (sas_device->starget == NULL)) {
2095 			sdev_printk(KERN_INFO, sdev,
2096 			"%s : sas_device->starget set to starget @ %d\n",
2097 			     __func__, __LINE__);
2098 			sas_device->starget = starget;
2099 		}
2100 
2101 		if (sas_device)
2102 			sas_device_put(sas_device);
2103 
2104 		spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
2105 	}
2106 
2107 	return 0;
2108 }
2109 
2110 /**
2111  * scsih_slave_destroy - device destroy routine
2112  * @sdev: scsi device struct
2113  */
2114 static void
2115 scsih_slave_destroy(struct scsi_device *sdev)
2116 {
2117 	struct MPT3SAS_TARGET *sas_target_priv_data;
2118 	struct scsi_target *starget;
2119 	struct Scsi_Host *shost;
2120 	struct MPT3SAS_ADAPTER *ioc;
2121 	struct _sas_device *sas_device;
2122 	struct _pcie_device *pcie_device;
2123 	unsigned long flags;
2124 
2125 	if (!sdev->hostdata)
2126 		return;
2127 
2128 	starget = scsi_target(sdev);
2129 	sas_target_priv_data = starget->hostdata;
2130 	sas_target_priv_data->num_luns--;
2131 
2132 	shost = dev_to_shost(&starget->dev);
2133 	ioc = shost_priv(shost);
2134 
2135 	if (sas_target_priv_data->flags & MPT_TARGET_FLAGS_PCIE_DEVICE) {
2136 		spin_lock_irqsave(&ioc->pcie_device_lock, flags);
2137 		pcie_device = __mpt3sas_get_pdev_from_target(ioc,
2138 				sas_target_priv_data);
2139 		if (pcie_device && !sas_target_priv_data->num_luns)
2140 			pcie_device->starget = NULL;
2141 
2142 		if (pcie_device)
2143 			pcie_device_put(pcie_device);
2144 
2145 		spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
2146 
2147 	} else if (!(sas_target_priv_data->flags & MPT_TARGET_FLAGS_VOLUME)) {
2148 		spin_lock_irqsave(&ioc->sas_device_lock, flags);
2149 		sas_device = __mpt3sas_get_sdev_from_target(ioc,
2150 				sas_target_priv_data);
2151 		if (sas_device && !sas_target_priv_data->num_luns)
2152 			sas_device->starget = NULL;
2153 
2154 		if (sas_device)
2155 			sas_device_put(sas_device);
2156 		spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
2157 	}
2158 
2159 	kfree(sdev->hostdata);
2160 	sdev->hostdata = NULL;
2161 }
2162 
2163 /**
2164  * _scsih_display_sata_capabilities - sata capabilities
2165  * @ioc: per adapter object
2166  * @handle: device handle
2167  * @sdev: scsi device struct
2168  */
2169 static void
2170 _scsih_display_sata_capabilities(struct MPT3SAS_ADAPTER *ioc,
2171 	u16 handle, struct scsi_device *sdev)
2172 {
2173 	Mpi2ConfigReply_t mpi_reply;
2174 	Mpi2SasDevicePage0_t sas_device_pg0;
2175 	u32 ioc_status;
2176 	u16 flags;
2177 	u32 device_info;
2178 
2179 	if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0,
2180 	    MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) {
2181 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
2182 			__FILE__, __LINE__, __func__);
2183 		return;
2184 	}
2185 
2186 	ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
2187 	    MPI2_IOCSTATUS_MASK;
2188 	if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
2189 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
2190 			__FILE__, __LINE__, __func__);
2191 		return;
2192 	}
2193 
2194 	flags = le16_to_cpu(sas_device_pg0.Flags);
2195 	device_info = le32_to_cpu(sas_device_pg0.DeviceInfo);
2196 
2197 	sdev_printk(KERN_INFO, sdev,
2198 	    "atapi(%s), ncq(%s), asyn_notify(%s), smart(%s), fua(%s), "
2199 	    "sw_preserve(%s)\n",
2200 	    (device_info & MPI2_SAS_DEVICE_INFO_ATAPI_DEVICE) ? "y" : "n",
2201 	    (flags & MPI2_SAS_DEVICE0_FLAGS_SATA_NCQ_SUPPORTED) ? "y" : "n",
2202 	    (flags & MPI2_SAS_DEVICE0_FLAGS_SATA_ASYNCHRONOUS_NOTIFY) ? "y" :
2203 	    "n",
2204 	    (flags & MPI2_SAS_DEVICE0_FLAGS_SATA_SMART_SUPPORTED) ? "y" : "n",
2205 	    (flags & MPI2_SAS_DEVICE0_FLAGS_SATA_FUA_SUPPORTED) ? "y" : "n",
2206 	    (flags & MPI2_SAS_DEVICE0_FLAGS_SATA_SW_PRESERVE) ? "y" : "n");
2207 }
2208 
2209 /*
2210  * raid transport support -
2211  * Enabled for SLES11 and newer, in older kernels the driver will panic when
2212  * unloading the driver followed by a load - I believe that the subroutine
2213  * raid_class_release() is not cleaning up properly.
2214  */
2215 
2216 /**
2217  * scsih_is_raid - return boolean indicating device is raid volume
2218  * @dev: the device struct object
2219  */
2220 static int
2221 scsih_is_raid(struct device *dev)
2222 {
2223 	struct scsi_device *sdev = to_scsi_device(dev);
2224 	struct MPT3SAS_ADAPTER *ioc = shost_priv(sdev->host);
2225 
2226 	if (ioc->is_warpdrive)
2227 		return 0;
2228 	return (sdev->channel == RAID_CHANNEL) ? 1 : 0;
2229 }
2230 
2231 static int
2232 scsih_is_nvme(struct device *dev)
2233 {
2234 	struct scsi_device *sdev = to_scsi_device(dev);
2235 
2236 	return (sdev->channel == PCIE_CHANNEL) ? 1 : 0;
2237 }
2238 
2239 /**
2240  * scsih_get_resync - get raid volume resync percent complete
2241  * @dev: the device struct object
2242  */
2243 static void
2244 scsih_get_resync(struct device *dev)
2245 {
2246 	struct scsi_device *sdev = to_scsi_device(dev);
2247 	struct MPT3SAS_ADAPTER *ioc = shost_priv(sdev->host);
2248 	static struct _raid_device *raid_device;
2249 	unsigned long flags;
2250 	Mpi2RaidVolPage0_t vol_pg0;
2251 	Mpi2ConfigReply_t mpi_reply;
2252 	u32 volume_status_flags;
2253 	u8 percent_complete;
2254 	u16 handle;
2255 
2256 	percent_complete = 0;
2257 	handle = 0;
2258 	if (ioc->is_warpdrive)
2259 		goto out;
2260 
2261 	spin_lock_irqsave(&ioc->raid_device_lock, flags);
2262 	raid_device = _scsih_raid_device_find_by_id(ioc, sdev->id,
2263 	    sdev->channel);
2264 	if (raid_device) {
2265 		handle = raid_device->handle;
2266 		percent_complete = raid_device->percent_complete;
2267 	}
2268 	spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
2269 
2270 	if (!handle)
2271 		goto out;
2272 
2273 	if (mpt3sas_config_get_raid_volume_pg0(ioc, &mpi_reply, &vol_pg0,
2274 	     MPI2_RAID_VOLUME_PGAD_FORM_HANDLE, handle,
2275 	     sizeof(Mpi2RaidVolPage0_t))) {
2276 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
2277 			__FILE__, __LINE__, __func__);
2278 		percent_complete = 0;
2279 		goto out;
2280 	}
2281 
2282 	volume_status_flags = le32_to_cpu(vol_pg0.VolumeStatusFlags);
2283 	if (!(volume_status_flags &
2284 	    MPI2_RAIDVOL0_STATUS_FLAG_RESYNC_IN_PROGRESS))
2285 		percent_complete = 0;
2286 
2287  out:
2288 
2289 	switch (ioc->hba_mpi_version_belonged) {
2290 	case MPI2_VERSION:
2291 		raid_set_resync(mpt2sas_raid_template, dev, percent_complete);
2292 		break;
2293 	case MPI25_VERSION:
2294 	case MPI26_VERSION:
2295 		raid_set_resync(mpt3sas_raid_template, dev, percent_complete);
2296 		break;
2297 	}
2298 }
2299 
2300 /**
2301  * scsih_get_state - get raid volume level
2302  * @dev: the device struct object
2303  */
2304 static void
2305 scsih_get_state(struct device *dev)
2306 {
2307 	struct scsi_device *sdev = to_scsi_device(dev);
2308 	struct MPT3SAS_ADAPTER *ioc = shost_priv(sdev->host);
2309 	static struct _raid_device *raid_device;
2310 	unsigned long flags;
2311 	Mpi2RaidVolPage0_t vol_pg0;
2312 	Mpi2ConfigReply_t mpi_reply;
2313 	u32 volstate;
2314 	enum raid_state state = RAID_STATE_UNKNOWN;
2315 	u16 handle = 0;
2316 
2317 	spin_lock_irqsave(&ioc->raid_device_lock, flags);
2318 	raid_device = _scsih_raid_device_find_by_id(ioc, sdev->id,
2319 	    sdev->channel);
2320 	if (raid_device)
2321 		handle = raid_device->handle;
2322 	spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
2323 
2324 	if (!raid_device)
2325 		goto out;
2326 
2327 	if (mpt3sas_config_get_raid_volume_pg0(ioc, &mpi_reply, &vol_pg0,
2328 	     MPI2_RAID_VOLUME_PGAD_FORM_HANDLE, handle,
2329 	     sizeof(Mpi2RaidVolPage0_t))) {
2330 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
2331 			__FILE__, __LINE__, __func__);
2332 		goto out;
2333 	}
2334 
2335 	volstate = le32_to_cpu(vol_pg0.VolumeStatusFlags);
2336 	if (volstate & MPI2_RAIDVOL0_STATUS_FLAG_RESYNC_IN_PROGRESS) {
2337 		state = RAID_STATE_RESYNCING;
2338 		goto out;
2339 	}
2340 
2341 	switch (vol_pg0.VolumeState) {
2342 	case MPI2_RAID_VOL_STATE_OPTIMAL:
2343 	case MPI2_RAID_VOL_STATE_ONLINE:
2344 		state = RAID_STATE_ACTIVE;
2345 		break;
2346 	case  MPI2_RAID_VOL_STATE_DEGRADED:
2347 		state = RAID_STATE_DEGRADED;
2348 		break;
2349 	case MPI2_RAID_VOL_STATE_FAILED:
2350 	case MPI2_RAID_VOL_STATE_MISSING:
2351 		state = RAID_STATE_OFFLINE;
2352 		break;
2353 	}
2354  out:
2355 	switch (ioc->hba_mpi_version_belonged) {
2356 	case MPI2_VERSION:
2357 		raid_set_state(mpt2sas_raid_template, dev, state);
2358 		break;
2359 	case MPI25_VERSION:
2360 	case MPI26_VERSION:
2361 		raid_set_state(mpt3sas_raid_template, dev, state);
2362 		break;
2363 	}
2364 }
2365 
2366 /**
2367  * _scsih_set_level - set raid level
2368  * @ioc: ?
2369  * @sdev: scsi device struct
2370  * @volume_type: volume type
2371  */
2372 static void
2373 _scsih_set_level(struct MPT3SAS_ADAPTER *ioc,
2374 	struct scsi_device *sdev, u8 volume_type)
2375 {
2376 	enum raid_level level = RAID_LEVEL_UNKNOWN;
2377 
2378 	switch (volume_type) {
2379 	case MPI2_RAID_VOL_TYPE_RAID0:
2380 		level = RAID_LEVEL_0;
2381 		break;
2382 	case MPI2_RAID_VOL_TYPE_RAID10:
2383 		level = RAID_LEVEL_10;
2384 		break;
2385 	case MPI2_RAID_VOL_TYPE_RAID1E:
2386 		level = RAID_LEVEL_1E;
2387 		break;
2388 	case MPI2_RAID_VOL_TYPE_RAID1:
2389 		level = RAID_LEVEL_1;
2390 		break;
2391 	}
2392 
2393 	switch (ioc->hba_mpi_version_belonged) {
2394 	case MPI2_VERSION:
2395 		raid_set_level(mpt2sas_raid_template,
2396 			&sdev->sdev_gendev, level);
2397 		break;
2398 	case MPI25_VERSION:
2399 	case MPI26_VERSION:
2400 		raid_set_level(mpt3sas_raid_template,
2401 			&sdev->sdev_gendev, level);
2402 		break;
2403 	}
2404 }
2405 
2406 
2407 /**
2408  * _scsih_get_volume_capabilities - volume capabilities
2409  * @ioc: per adapter object
2410  * @raid_device: the raid_device object
2411  *
2412  * Return: 0 for success, else 1
2413  */
2414 static int
2415 _scsih_get_volume_capabilities(struct MPT3SAS_ADAPTER *ioc,
2416 	struct _raid_device *raid_device)
2417 {
2418 	Mpi2RaidVolPage0_t *vol_pg0;
2419 	Mpi2RaidPhysDiskPage0_t pd_pg0;
2420 	Mpi2SasDevicePage0_t sas_device_pg0;
2421 	Mpi2ConfigReply_t mpi_reply;
2422 	u16 sz;
2423 	u8 num_pds;
2424 
2425 	if ((mpt3sas_config_get_number_pds(ioc, raid_device->handle,
2426 	    &num_pds)) || !num_pds) {
2427 		dfailprintk(ioc,
2428 			    ioc_warn(ioc, "failure at %s:%d/%s()!\n",
2429 				     __FILE__, __LINE__, __func__));
2430 		return 1;
2431 	}
2432 
2433 	raid_device->num_pds = num_pds;
2434 	sz = offsetof(Mpi2RaidVolPage0_t, PhysDisk) + (num_pds *
2435 	    sizeof(Mpi2RaidVol0PhysDisk_t));
2436 	vol_pg0 = kzalloc(sz, GFP_KERNEL);
2437 	if (!vol_pg0) {
2438 		dfailprintk(ioc,
2439 			    ioc_warn(ioc, "failure at %s:%d/%s()!\n",
2440 				     __FILE__, __LINE__, __func__));
2441 		return 1;
2442 	}
2443 
2444 	if ((mpt3sas_config_get_raid_volume_pg0(ioc, &mpi_reply, vol_pg0,
2445 	     MPI2_RAID_VOLUME_PGAD_FORM_HANDLE, raid_device->handle, sz))) {
2446 		dfailprintk(ioc,
2447 			    ioc_warn(ioc, "failure at %s:%d/%s()!\n",
2448 				     __FILE__, __LINE__, __func__));
2449 		kfree(vol_pg0);
2450 		return 1;
2451 	}
2452 
2453 	raid_device->volume_type = vol_pg0->VolumeType;
2454 
2455 	/* figure out what the underlying devices are by
2456 	 * obtaining the device_info bits for the 1st device
2457 	 */
2458 	if (!(mpt3sas_config_get_phys_disk_pg0(ioc, &mpi_reply,
2459 	    &pd_pg0, MPI2_PHYSDISK_PGAD_FORM_PHYSDISKNUM,
2460 	    vol_pg0->PhysDisk[0].PhysDiskNum))) {
2461 		if (!(mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply,
2462 		    &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE,
2463 		    le16_to_cpu(pd_pg0.DevHandle)))) {
2464 			raid_device->device_info =
2465 			    le32_to_cpu(sas_device_pg0.DeviceInfo);
2466 		}
2467 	}
2468 
2469 	kfree(vol_pg0);
2470 	return 0;
2471 }
2472 
2473 /**
2474  * _scsih_enable_tlr - setting TLR flags
2475  * @ioc: per adapter object
2476  * @sdev: scsi device struct
2477  *
2478  * Enabling Transaction Layer Retries for tape devices when
2479  * vpd page 0x90 is present
2480  *
2481  */
2482 static void
2483 _scsih_enable_tlr(struct MPT3SAS_ADAPTER *ioc, struct scsi_device *sdev)
2484 {
2485 
2486 	/* only for TAPE */
2487 	if (sdev->type != TYPE_TAPE)
2488 		return;
2489 
2490 	if (!(ioc->facts.IOCCapabilities & MPI2_IOCFACTS_CAPABILITY_TLR))
2491 		return;
2492 
2493 	sas_enable_tlr(sdev);
2494 	sdev_printk(KERN_INFO, sdev, "TLR %s\n",
2495 	    sas_is_tlr_enabled(sdev) ? "Enabled" : "Disabled");
2496 	return;
2497 
2498 }
2499 
2500 /**
2501  * scsih_slave_configure - device configure routine.
2502  * @sdev: scsi device struct
2503  *
2504  * Return: 0 if ok. Any other return is assumed to be an error and
2505  * the device is ignored.
2506  */
2507 static int
2508 scsih_slave_configure(struct scsi_device *sdev)
2509 {
2510 	struct Scsi_Host *shost = sdev->host;
2511 	struct MPT3SAS_ADAPTER *ioc = shost_priv(shost);
2512 	struct MPT3SAS_DEVICE *sas_device_priv_data;
2513 	struct MPT3SAS_TARGET *sas_target_priv_data;
2514 	struct _sas_device *sas_device;
2515 	struct _pcie_device *pcie_device;
2516 	struct _raid_device *raid_device;
2517 	unsigned long flags;
2518 	int qdepth;
2519 	u8 ssp_target = 0;
2520 	char *ds = "";
2521 	char *r_level = "";
2522 	u16 handle, volume_handle = 0;
2523 	u64 volume_wwid = 0;
2524 
2525 	qdepth = 1;
2526 	sas_device_priv_data = sdev->hostdata;
2527 	sas_device_priv_data->configured_lun = 1;
2528 	sas_device_priv_data->flags &= ~MPT_DEVICE_FLAGS_INIT;
2529 	sas_target_priv_data = sas_device_priv_data->sas_target;
2530 	handle = sas_target_priv_data->handle;
2531 
2532 	/* raid volume handling */
2533 	if (sas_target_priv_data->flags & MPT_TARGET_FLAGS_VOLUME) {
2534 
2535 		spin_lock_irqsave(&ioc->raid_device_lock, flags);
2536 		raid_device = mpt3sas_raid_device_find_by_handle(ioc, handle);
2537 		spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
2538 		if (!raid_device) {
2539 			dfailprintk(ioc,
2540 				    ioc_warn(ioc, "failure at %s:%d/%s()!\n",
2541 					     __FILE__, __LINE__, __func__));
2542 			return 1;
2543 		}
2544 
2545 		if (_scsih_get_volume_capabilities(ioc, raid_device)) {
2546 			dfailprintk(ioc,
2547 				    ioc_warn(ioc, "failure at %s:%d/%s()!\n",
2548 					     __FILE__, __LINE__, __func__));
2549 			return 1;
2550 		}
2551 
2552 		/*
2553 		 * WARPDRIVE: Initialize the required data for Direct IO
2554 		 */
2555 		mpt3sas_init_warpdrive_properties(ioc, raid_device);
2556 
2557 		/* RAID Queue Depth Support
2558 		 * IS volume = underlying qdepth of drive type, either
2559 		 *    MPT3SAS_SAS_QUEUE_DEPTH or MPT3SAS_SATA_QUEUE_DEPTH
2560 		 * IM/IME/R10 = 128 (MPT3SAS_RAID_QUEUE_DEPTH)
2561 		 */
2562 		if (raid_device->device_info &
2563 		    MPI2_SAS_DEVICE_INFO_SSP_TARGET) {
2564 			qdepth = MPT3SAS_SAS_QUEUE_DEPTH;
2565 			ds = "SSP";
2566 		} else {
2567 			qdepth = MPT3SAS_SATA_QUEUE_DEPTH;
2568 			if (raid_device->device_info &
2569 			    MPI2_SAS_DEVICE_INFO_SATA_DEVICE)
2570 				ds = "SATA";
2571 			else
2572 				ds = "STP";
2573 		}
2574 
2575 		switch (raid_device->volume_type) {
2576 		case MPI2_RAID_VOL_TYPE_RAID0:
2577 			r_level = "RAID0";
2578 			break;
2579 		case MPI2_RAID_VOL_TYPE_RAID1E:
2580 			qdepth = MPT3SAS_RAID_QUEUE_DEPTH;
2581 			if (ioc->manu_pg10.OEMIdentifier &&
2582 			    (le32_to_cpu(ioc->manu_pg10.GenericFlags0) &
2583 			    MFG10_GF0_R10_DISPLAY) &&
2584 			    !(raid_device->num_pds % 2))
2585 				r_level = "RAID10";
2586 			else
2587 				r_level = "RAID1E";
2588 			break;
2589 		case MPI2_RAID_VOL_TYPE_RAID1:
2590 			qdepth = MPT3SAS_RAID_QUEUE_DEPTH;
2591 			r_level = "RAID1";
2592 			break;
2593 		case MPI2_RAID_VOL_TYPE_RAID10:
2594 			qdepth = MPT3SAS_RAID_QUEUE_DEPTH;
2595 			r_level = "RAID10";
2596 			break;
2597 		case MPI2_RAID_VOL_TYPE_UNKNOWN:
2598 		default:
2599 			qdepth = MPT3SAS_RAID_QUEUE_DEPTH;
2600 			r_level = "RAIDX";
2601 			break;
2602 		}
2603 
2604 		if (!ioc->hide_ir_msg)
2605 			sdev_printk(KERN_INFO, sdev,
2606 			   "%s: handle(0x%04x), wwid(0x%016llx),"
2607 			    " pd_count(%d), type(%s)\n",
2608 			    r_level, raid_device->handle,
2609 			    (unsigned long long)raid_device->wwid,
2610 			    raid_device->num_pds, ds);
2611 
2612 		if (shost->max_sectors > MPT3SAS_RAID_MAX_SECTORS) {
2613 			blk_queue_max_hw_sectors(sdev->request_queue,
2614 						MPT3SAS_RAID_MAX_SECTORS);
2615 			sdev_printk(KERN_INFO, sdev,
2616 					"Set queue's max_sector to: %u\n",
2617 						MPT3SAS_RAID_MAX_SECTORS);
2618 		}
2619 
2620 		mpt3sas_scsih_change_queue_depth(sdev, qdepth);
2621 
2622 		/* raid transport support */
2623 		if (!ioc->is_warpdrive)
2624 			_scsih_set_level(ioc, sdev, raid_device->volume_type);
2625 		return 0;
2626 	}
2627 
2628 	/* non-raid handling */
2629 	if (sas_target_priv_data->flags & MPT_TARGET_FLAGS_RAID_COMPONENT) {
2630 		if (mpt3sas_config_get_volume_handle(ioc, handle,
2631 		    &volume_handle)) {
2632 			dfailprintk(ioc,
2633 				    ioc_warn(ioc, "failure at %s:%d/%s()!\n",
2634 					     __FILE__, __LINE__, __func__));
2635 			return 1;
2636 		}
2637 		if (volume_handle && mpt3sas_config_get_volume_wwid(ioc,
2638 		    volume_handle, &volume_wwid)) {
2639 			dfailprintk(ioc,
2640 				    ioc_warn(ioc, "failure at %s:%d/%s()!\n",
2641 					     __FILE__, __LINE__, __func__));
2642 			return 1;
2643 		}
2644 	}
2645 
2646 	/* PCIe handling */
2647 	if (sas_target_priv_data->flags & MPT_TARGET_FLAGS_PCIE_DEVICE) {
2648 		spin_lock_irqsave(&ioc->pcie_device_lock, flags);
2649 		pcie_device = __mpt3sas_get_pdev_by_wwid(ioc,
2650 				sas_device_priv_data->sas_target->sas_address);
2651 		if (!pcie_device) {
2652 			spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
2653 			dfailprintk(ioc,
2654 				    ioc_warn(ioc, "failure at %s:%d/%s()!\n",
2655 					     __FILE__, __LINE__, __func__));
2656 			return 1;
2657 		}
2658 
2659 		qdepth = ioc->max_nvme_qd;
2660 		ds = "NVMe";
2661 		sdev_printk(KERN_INFO, sdev,
2662 			"%s: handle(0x%04x), wwid(0x%016llx), port(%d)\n",
2663 			ds, handle, (unsigned long long)pcie_device->wwid,
2664 			pcie_device->port_num);
2665 		if (pcie_device->enclosure_handle != 0)
2666 			sdev_printk(KERN_INFO, sdev,
2667 			"%s: enclosure logical id(0x%016llx), slot(%d)\n",
2668 			ds,
2669 			(unsigned long long)pcie_device->enclosure_logical_id,
2670 			pcie_device->slot);
2671 		if (pcie_device->connector_name[0] != '\0')
2672 			sdev_printk(KERN_INFO, sdev,
2673 				"%s: enclosure level(0x%04x),"
2674 				"connector name( %s)\n", ds,
2675 				pcie_device->enclosure_level,
2676 				pcie_device->connector_name);
2677 
2678 		if (pcie_device->nvme_mdts)
2679 			blk_queue_max_hw_sectors(sdev->request_queue,
2680 					pcie_device->nvme_mdts/512);
2681 
2682 		pcie_device_put(pcie_device);
2683 		spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
2684 		mpt3sas_scsih_change_queue_depth(sdev, qdepth);
2685 		/* Enable QUEUE_FLAG_NOMERGES flag, so that IOs won't be
2686 		 ** merged and can eliminate holes created during merging
2687 		 ** operation.
2688 		 **/
2689 		blk_queue_flag_set(QUEUE_FLAG_NOMERGES,
2690 				sdev->request_queue);
2691 		blk_queue_virt_boundary(sdev->request_queue,
2692 				ioc->page_size - 1);
2693 		return 0;
2694 	}
2695 
2696 	spin_lock_irqsave(&ioc->sas_device_lock, flags);
2697 	sas_device = __mpt3sas_get_sdev_by_addr(ioc,
2698 	   sas_device_priv_data->sas_target->sas_address,
2699 	   sas_device_priv_data->sas_target->port);
2700 	if (!sas_device) {
2701 		spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
2702 		dfailprintk(ioc,
2703 			    ioc_warn(ioc, "failure at %s:%d/%s()!\n",
2704 				     __FILE__, __LINE__, __func__));
2705 		return 1;
2706 	}
2707 
2708 	sas_device->volume_handle = volume_handle;
2709 	sas_device->volume_wwid = volume_wwid;
2710 	if (sas_device->device_info & MPI2_SAS_DEVICE_INFO_SSP_TARGET) {
2711 		qdepth = (sas_device->port_type > 1) ?
2712 			ioc->max_wideport_qd : ioc->max_narrowport_qd;
2713 		ssp_target = 1;
2714 		if (sas_device->device_info &
2715 				MPI2_SAS_DEVICE_INFO_SEP) {
2716 			sdev_printk(KERN_WARNING, sdev,
2717 			"set ignore_delay_remove for handle(0x%04x)\n",
2718 			sas_device_priv_data->sas_target->handle);
2719 			sas_device_priv_data->ignore_delay_remove = 1;
2720 			ds = "SES";
2721 		} else
2722 			ds = "SSP";
2723 	} else {
2724 		qdepth = ioc->max_sata_qd;
2725 		if (sas_device->device_info & MPI2_SAS_DEVICE_INFO_STP_TARGET)
2726 			ds = "STP";
2727 		else if (sas_device->device_info &
2728 		    MPI2_SAS_DEVICE_INFO_SATA_DEVICE)
2729 			ds = "SATA";
2730 	}
2731 
2732 	sdev_printk(KERN_INFO, sdev, "%s: handle(0x%04x), " \
2733 	    "sas_addr(0x%016llx), phy(%d), device_name(0x%016llx)\n",
2734 	    ds, handle, (unsigned long long)sas_device->sas_address,
2735 	    sas_device->phy, (unsigned long long)sas_device->device_name);
2736 
2737 	_scsih_display_enclosure_chassis_info(NULL, sas_device, sdev, NULL);
2738 
2739 	sas_device_put(sas_device);
2740 	spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
2741 
2742 	if (!ssp_target)
2743 		_scsih_display_sata_capabilities(ioc, handle, sdev);
2744 
2745 
2746 	mpt3sas_scsih_change_queue_depth(sdev, qdepth);
2747 
2748 	if (ssp_target) {
2749 		sas_read_port_mode_page(sdev);
2750 		_scsih_enable_tlr(ioc, sdev);
2751 	}
2752 
2753 	return 0;
2754 }
2755 
2756 /**
2757  * scsih_bios_param - fetch head, sector, cylinder info for a disk
2758  * @sdev: scsi device struct
2759  * @bdev: pointer to block device context
2760  * @capacity: device size (in 512 byte sectors)
2761  * @params: three element array to place output:
2762  *              params[0] number of heads (max 255)
2763  *              params[1] number of sectors (max 63)
2764  *              params[2] number of cylinders
2765  */
2766 static int
2767 scsih_bios_param(struct scsi_device *sdev, struct block_device *bdev,
2768 	sector_t capacity, int params[])
2769 {
2770 	int		heads;
2771 	int		sectors;
2772 	sector_t	cylinders;
2773 	ulong		dummy;
2774 
2775 	heads = 64;
2776 	sectors = 32;
2777 
2778 	dummy = heads * sectors;
2779 	cylinders = capacity;
2780 	sector_div(cylinders, dummy);
2781 
2782 	/*
2783 	 * Handle extended translation size for logical drives
2784 	 * > 1Gb
2785 	 */
2786 	if ((ulong)capacity >= 0x200000) {
2787 		heads = 255;
2788 		sectors = 63;
2789 		dummy = heads * sectors;
2790 		cylinders = capacity;
2791 		sector_div(cylinders, dummy);
2792 	}
2793 
2794 	/* return result */
2795 	params[0] = heads;
2796 	params[1] = sectors;
2797 	params[2] = cylinders;
2798 
2799 	return 0;
2800 }
2801 
2802 /**
2803  * _scsih_response_code - translation of device response code
2804  * @ioc: per adapter object
2805  * @response_code: response code returned by the device
2806  */
2807 static void
2808 _scsih_response_code(struct MPT3SAS_ADAPTER *ioc, u8 response_code)
2809 {
2810 	char *desc;
2811 
2812 	switch (response_code) {
2813 	case MPI2_SCSITASKMGMT_RSP_TM_COMPLETE:
2814 		desc = "task management request completed";
2815 		break;
2816 	case MPI2_SCSITASKMGMT_RSP_INVALID_FRAME:
2817 		desc = "invalid frame";
2818 		break;
2819 	case MPI2_SCSITASKMGMT_RSP_TM_NOT_SUPPORTED:
2820 		desc = "task management request not supported";
2821 		break;
2822 	case MPI2_SCSITASKMGMT_RSP_TM_FAILED:
2823 		desc = "task management request failed";
2824 		break;
2825 	case MPI2_SCSITASKMGMT_RSP_TM_SUCCEEDED:
2826 		desc = "task management request succeeded";
2827 		break;
2828 	case MPI2_SCSITASKMGMT_RSP_TM_INVALID_LUN:
2829 		desc = "invalid lun";
2830 		break;
2831 	case 0xA:
2832 		desc = "overlapped tag attempted";
2833 		break;
2834 	case MPI2_SCSITASKMGMT_RSP_IO_QUEUED_ON_IOC:
2835 		desc = "task queued, however not sent to target";
2836 		break;
2837 	default:
2838 		desc = "unknown";
2839 		break;
2840 	}
2841 	ioc_warn(ioc, "response_code(0x%01x): %s\n", response_code, desc);
2842 }
2843 
2844 /**
2845  * _scsih_tm_done - tm completion routine
2846  * @ioc: per adapter object
2847  * @smid: system request message index
2848  * @msix_index: MSIX table index supplied by the OS
2849  * @reply: reply message frame(lower 32bit addr)
2850  * Context: none.
2851  *
2852  * The callback handler when using scsih_issue_tm.
2853  *
2854  * Return: 1 meaning mf should be freed from _base_interrupt
2855  *         0 means the mf is freed from this function.
2856  */
2857 static u8
2858 _scsih_tm_done(struct MPT3SAS_ADAPTER *ioc, u16 smid, u8 msix_index, u32 reply)
2859 {
2860 	MPI2DefaultReply_t *mpi_reply;
2861 
2862 	if (ioc->tm_cmds.status == MPT3_CMD_NOT_USED)
2863 		return 1;
2864 	if (ioc->tm_cmds.smid != smid)
2865 		return 1;
2866 	ioc->tm_cmds.status |= MPT3_CMD_COMPLETE;
2867 	mpi_reply =  mpt3sas_base_get_reply_virt_addr(ioc, reply);
2868 	if (mpi_reply) {
2869 		memcpy(ioc->tm_cmds.reply, mpi_reply, mpi_reply->MsgLength*4);
2870 		ioc->tm_cmds.status |= MPT3_CMD_REPLY_VALID;
2871 	}
2872 	ioc->tm_cmds.status &= ~MPT3_CMD_PENDING;
2873 	complete(&ioc->tm_cmds.done);
2874 	return 1;
2875 }
2876 
2877 /**
2878  * mpt3sas_scsih_set_tm_flag - set per target tm_busy
2879  * @ioc: per adapter object
2880  * @handle: device handle
2881  *
2882  * During taskmangement request, we need to freeze the device queue.
2883  */
2884 void
2885 mpt3sas_scsih_set_tm_flag(struct MPT3SAS_ADAPTER *ioc, u16 handle)
2886 {
2887 	struct MPT3SAS_DEVICE *sas_device_priv_data;
2888 	struct scsi_device *sdev;
2889 	u8 skip = 0;
2890 
2891 	shost_for_each_device(sdev, ioc->shost) {
2892 		if (skip)
2893 			continue;
2894 		sas_device_priv_data = sdev->hostdata;
2895 		if (!sas_device_priv_data)
2896 			continue;
2897 		if (sas_device_priv_data->sas_target->handle == handle) {
2898 			sas_device_priv_data->sas_target->tm_busy = 1;
2899 			skip = 1;
2900 			ioc->ignore_loginfos = 1;
2901 		}
2902 	}
2903 }
2904 
2905 /**
2906  * mpt3sas_scsih_clear_tm_flag - clear per target tm_busy
2907  * @ioc: per adapter object
2908  * @handle: device handle
2909  *
2910  * During taskmangement request, we need to freeze the device queue.
2911  */
2912 void
2913 mpt3sas_scsih_clear_tm_flag(struct MPT3SAS_ADAPTER *ioc, u16 handle)
2914 {
2915 	struct MPT3SAS_DEVICE *sas_device_priv_data;
2916 	struct scsi_device *sdev;
2917 	u8 skip = 0;
2918 
2919 	shost_for_each_device(sdev, ioc->shost) {
2920 		if (skip)
2921 			continue;
2922 		sas_device_priv_data = sdev->hostdata;
2923 		if (!sas_device_priv_data)
2924 			continue;
2925 		if (sas_device_priv_data->sas_target->handle == handle) {
2926 			sas_device_priv_data->sas_target->tm_busy = 0;
2927 			skip = 1;
2928 			ioc->ignore_loginfos = 0;
2929 		}
2930 	}
2931 }
2932 
2933 /**
2934  * scsih_tm_cmd_map_status - map the target reset & LUN reset TM status
2935  * @ioc: per adapter object
2936  * @channel: the channel assigned by the OS
2937  * @id: the id assigned by the OS
2938  * @lun: lun number
2939  * @type: MPI2_SCSITASKMGMT_TASKTYPE__XXX (defined in mpi2_init.h)
2940  * @smid_task: smid assigned to the task
2941  *
2942  * Look whether TM has aborted the timed out SCSI command, if
2943  * TM has aborted the IO then return SUCCESS else return FAILED.
2944  */
2945 static int
2946 scsih_tm_cmd_map_status(struct MPT3SAS_ADAPTER *ioc, uint channel,
2947 	uint id, uint lun, u8 type, u16 smid_task)
2948 {
2949 
2950 	if (smid_task <= ioc->shost->can_queue) {
2951 		switch (type) {
2952 		case MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET:
2953 			if (!(_scsih_scsi_lookup_find_by_target(ioc,
2954 			    id, channel)))
2955 				return SUCCESS;
2956 			break;
2957 		case MPI2_SCSITASKMGMT_TASKTYPE_ABRT_TASK_SET:
2958 		case MPI2_SCSITASKMGMT_TASKTYPE_LOGICAL_UNIT_RESET:
2959 			if (!(_scsih_scsi_lookup_find_by_lun(ioc, id,
2960 			    lun, channel)))
2961 				return SUCCESS;
2962 			break;
2963 		default:
2964 			return SUCCESS;
2965 		}
2966 	} else if (smid_task == ioc->scsih_cmds.smid) {
2967 		if ((ioc->scsih_cmds.status & MPT3_CMD_COMPLETE) ||
2968 		    (ioc->scsih_cmds.status & MPT3_CMD_NOT_USED))
2969 			return SUCCESS;
2970 	} else if (smid_task == ioc->ctl_cmds.smid) {
2971 		if ((ioc->ctl_cmds.status & MPT3_CMD_COMPLETE) ||
2972 		    (ioc->ctl_cmds.status & MPT3_CMD_NOT_USED))
2973 			return SUCCESS;
2974 	}
2975 
2976 	return FAILED;
2977 }
2978 
2979 /**
2980  * scsih_tm_post_processing - post processing of target & LUN reset
2981  * @ioc: per adapter object
2982  * @handle: device handle
2983  * @channel: the channel assigned by the OS
2984  * @id: the id assigned by the OS
2985  * @lun: lun number
2986  * @type: MPI2_SCSITASKMGMT_TASKTYPE__XXX (defined in mpi2_init.h)
2987  * @smid_task: smid assigned to the task
2988  *
2989  * Post processing of target & LUN reset. Due to interrupt latency
2990  * issue it possible that interrupt for aborted IO might not be
2991  * received yet. So before returning failure status, poll the
2992  * reply descriptor pools for the reply of timed out SCSI command.
2993  * Return FAILED status if reply for timed out is not received
2994  * otherwise return SUCCESS.
2995  */
2996 static int
2997 scsih_tm_post_processing(struct MPT3SAS_ADAPTER *ioc, u16 handle,
2998 	uint channel, uint id, uint lun, u8 type, u16 smid_task)
2999 {
3000 	int rc;
3001 
3002 	rc = scsih_tm_cmd_map_status(ioc, channel, id, lun, type, smid_task);
3003 	if (rc == SUCCESS)
3004 		return rc;
3005 
3006 	ioc_info(ioc,
3007 	    "Poll ReplyDescriptor queues for completion of"
3008 	    " smid(%d), task_type(0x%02x), handle(0x%04x)\n",
3009 	    smid_task, type, handle);
3010 
3011 	/*
3012 	 * Due to interrupt latency issues, driver may receive interrupt for
3013 	 * TM first and then for aborted SCSI IO command. So, poll all the
3014 	 * ReplyDescriptor pools before returning the FAILED status to SML.
3015 	 */
3016 	mpt3sas_base_mask_interrupts(ioc);
3017 	mpt3sas_base_sync_reply_irqs(ioc, 1);
3018 	mpt3sas_base_unmask_interrupts(ioc);
3019 
3020 	return scsih_tm_cmd_map_status(ioc, channel, id, lun, type, smid_task);
3021 }
3022 
3023 /**
3024  * mpt3sas_scsih_issue_tm - main routine for sending tm requests
3025  * @ioc: per adapter struct
3026  * @handle: device handle
3027  * @channel: the channel assigned by the OS
3028  * @id: the id assigned by the OS
3029  * @lun: lun number
3030  * @type: MPI2_SCSITASKMGMT_TASKTYPE__XXX (defined in mpi2_init.h)
3031  * @smid_task: smid assigned to the task
3032  * @msix_task: MSIX table index supplied by the OS
3033  * @timeout: timeout in seconds
3034  * @tr_method: Target Reset Method
3035  * Context: user
3036  *
3037  * A generic API for sending task management requests to firmware.
3038  *
3039  * The callback index is set inside `ioc->tm_cb_idx`.
3040  * The caller is responsible to check for outstanding commands.
3041  *
3042  * Return: SUCCESS or FAILED.
3043  */
3044 int
3045 mpt3sas_scsih_issue_tm(struct MPT3SAS_ADAPTER *ioc, u16 handle, uint channel,
3046 	uint id, u64 lun, u8 type, u16 smid_task, u16 msix_task,
3047 	u8 timeout, u8 tr_method)
3048 {
3049 	Mpi2SCSITaskManagementRequest_t *mpi_request;
3050 	Mpi2SCSITaskManagementReply_t *mpi_reply;
3051 	Mpi25SCSIIORequest_t *request;
3052 	u16 smid = 0;
3053 	u32 ioc_state;
3054 	int rc;
3055 	u8 issue_reset = 0;
3056 
3057 	lockdep_assert_held(&ioc->tm_cmds.mutex);
3058 
3059 	if (ioc->tm_cmds.status != MPT3_CMD_NOT_USED) {
3060 		ioc_info(ioc, "%s: tm_cmd busy!!!\n", __func__);
3061 		return FAILED;
3062 	}
3063 
3064 	if (ioc->shost_recovery || ioc->remove_host ||
3065 	    ioc->pci_error_recovery) {
3066 		ioc_info(ioc, "%s: host reset in progress!\n", __func__);
3067 		return FAILED;
3068 	}
3069 
3070 	ioc_state = mpt3sas_base_get_iocstate(ioc, 0);
3071 	if (ioc_state & MPI2_DOORBELL_USED) {
3072 		dhsprintk(ioc, ioc_info(ioc, "unexpected doorbell active!\n"));
3073 		rc = mpt3sas_base_hard_reset_handler(ioc, FORCE_BIG_HAMMER);
3074 		return (!rc) ? SUCCESS : FAILED;
3075 	}
3076 
3077 	if ((ioc_state & MPI2_IOC_STATE_MASK) == MPI2_IOC_STATE_FAULT) {
3078 		mpt3sas_print_fault_code(ioc, ioc_state &
3079 		    MPI2_DOORBELL_DATA_MASK);
3080 		rc = mpt3sas_base_hard_reset_handler(ioc, FORCE_BIG_HAMMER);
3081 		return (!rc) ? SUCCESS : FAILED;
3082 	} else if ((ioc_state & MPI2_IOC_STATE_MASK) ==
3083 	    MPI2_IOC_STATE_COREDUMP) {
3084 		mpt3sas_print_coredump_info(ioc, ioc_state &
3085 		    MPI2_DOORBELL_DATA_MASK);
3086 		rc = mpt3sas_base_hard_reset_handler(ioc, FORCE_BIG_HAMMER);
3087 		return (!rc) ? SUCCESS : FAILED;
3088 	}
3089 
3090 	smid = mpt3sas_base_get_smid_hpr(ioc, ioc->tm_cb_idx);
3091 	if (!smid) {
3092 		ioc_err(ioc, "%s: failed obtaining a smid\n", __func__);
3093 		return FAILED;
3094 	}
3095 
3096 	dtmprintk(ioc,
3097 		  ioc_info(ioc, "sending tm: handle(0x%04x), task_type(0x%02x), smid(%d), timeout(%d), tr_method(0x%x)\n",
3098 			   handle, type, smid_task, timeout, tr_method));
3099 	ioc->tm_cmds.status = MPT3_CMD_PENDING;
3100 	mpi_request = mpt3sas_base_get_msg_frame(ioc, smid);
3101 	ioc->tm_cmds.smid = smid;
3102 	memset(mpi_request, 0, sizeof(Mpi2SCSITaskManagementRequest_t));
3103 	memset(ioc->tm_cmds.reply, 0, sizeof(Mpi2SCSITaskManagementReply_t));
3104 	mpi_request->Function = MPI2_FUNCTION_SCSI_TASK_MGMT;
3105 	mpi_request->DevHandle = cpu_to_le16(handle);
3106 	mpi_request->TaskType = type;
3107 	if (type == MPI2_SCSITASKMGMT_TASKTYPE_ABORT_TASK ||
3108 	    type == MPI2_SCSITASKMGMT_TASKTYPE_QUERY_TASK)
3109 		mpi_request->MsgFlags = tr_method;
3110 	mpi_request->TaskMID = cpu_to_le16(smid_task);
3111 	int_to_scsilun(lun, (struct scsi_lun *)mpi_request->LUN);
3112 	mpt3sas_scsih_set_tm_flag(ioc, handle);
3113 	init_completion(&ioc->tm_cmds.done);
3114 	ioc->put_smid_hi_priority(ioc, smid, msix_task);
3115 	wait_for_completion_timeout(&ioc->tm_cmds.done, timeout*HZ);
3116 	if (!(ioc->tm_cmds.status & MPT3_CMD_COMPLETE)) {
3117 		mpt3sas_check_cmd_timeout(ioc,
3118 		    ioc->tm_cmds.status, mpi_request,
3119 		    sizeof(Mpi2SCSITaskManagementRequest_t)/4, issue_reset);
3120 		if (issue_reset) {
3121 			rc = mpt3sas_base_hard_reset_handler(ioc,
3122 					FORCE_BIG_HAMMER);
3123 			rc = (!rc) ? SUCCESS : FAILED;
3124 			goto out;
3125 		}
3126 	}
3127 
3128 	/* sync IRQs in case those were busy during flush. */
3129 	mpt3sas_base_sync_reply_irqs(ioc, 0);
3130 
3131 	if (ioc->tm_cmds.status & MPT3_CMD_REPLY_VALID) {
3132 		mpt3sas_trigger_master(ioc, MASTER_TRIGGER_TASK_MANAGMENT);
3133 		mpi_reply = ioc->tm_cmds.reply;
3134 		dtmprintk(ioc,
3135 			  ioc_info(ioc, "complete tm: ioc_status(0x%04x), loginfo(0x%08x), term_count(0x%08x)\n",
3136 				   le16_to_cpu(mpi_reply->IOCStatus),
3137 				   le32_to_cpu(mpi_reply->IOCLogInfo),
3138 				   le32_to_cpu(mpi_reply->TerminationCount)));
3139 		if (ioc->logging_level & MPT_DEBUG_TM) {
3140 			_scsih_response_code(ioc, mpi_reply->ResponseCode);
3141 			if (mpi_reply->IOCStatus)
3142 				_debug_dump_mf(mpi_request,
3143 				    sizeof(Mpi2SCSITaskManagementRequest_t)/4);
3144 		}
3145 	}
3146 
3147 	switch (type) {
3148 	case MPI2_SCSITASKMGMT_TASKTYPE_ABORT_TASK:
3149 		rc = SUCCESS;
3150 		/*
3151 		 * If DevHandle filed in smid_task's entry of request pool
3152 		 * doesn't match with device handle on which this task abort
3153 		 * TM is received then it means that TM has successfully
3154 		 * aborted the timed out command. Since smid_task's entry in
3155 		 * request pool will be memset to zero once the timed out
3156 		 * command is returned to the SML. If the command is not
3157 		 * aborted then smid_task’s entry won’t be cleared and it
3158 		 * will have same DevHandle value on which this task abort TM
3159 		 * is received and driver will return the TM status as FAILED.
3160 		 */
3161 		request = mpt3sas_base_get_msg_frame(ioc, smid_task);
3162 		if (le16_to_cpu(request->DevHandle) != handle)
3163 			break;
3164 
3165 		ioc_info(ioc, "Task abort tm failed: handle(0x%04x),"
3166 		    "timeout(%d) tr_method(0x%x) smid(%d) msix_index(%d)\n",
3167 		    handle, timeout, tr_method, smid_task, msix_task);
3168 		rc = FAILED;
3169 		break;
3170 
3171 	case MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET:
3172 	case MPI2_SCSITASKMGMT_TASKTYPE_ABRT_TASK_SET:
3173 	case MPI2_SCSITASKMGMT_TASKTYPE_LOGICAL_UNIT_RESET:
3174 		rc = scsih_tm_post_processing(ioc, handle, channel, id, lun,
3175 		    type, smid_task);
3176 		break;
3177 	case MPI2_SCSITASKMGMT_TASKTYPE_QUERY_TASK:
3178 		rc = SUCCESS;
3179 		break;
3180 	default:
3181 		rc = FAILED;
3182 		break;
3183 	}
3184 
3185 out:
3186 	mpt3sas_scsih_clear_tm_flag(ioc, handle);
3187 	ioc->tm_cmds.status = MPT3_CMD_NOT_USED;
3188 	return rc;
3189 }
3190 
3191 int mpt3sas_scsih_issue_locked_tm(struct MPT3SAS_ADAPTER *ioc, u16 handle,
3192 		uint channel, uint id, u64 lun, u8 type, u16 smid_task,
3193 		u16 msix_task, u8 timeout, u8 tr_method)
3194 {
3195 	int ret;
3196 
3197 	mutex_lock(&ioc->tm_cmds.mutex);
3198 	ret = mpt3sas_scsih_issue_tm(ioc, handle, channel, id, lun, type,
3199 			smid_task, msix_task, timeout, tr_method);
3200 	mutex_unlock(&ioc->tm_cmds.mutex);
3201 
3202 	return ret;
3203 }
3204 
3205 /**
3206  * _scsih_tm_display_info - displays info about the device
3207  * @ioc: per adapter struct
3208  * @scmd: pointer to scsi command object
3209  *
3210  * Called by task management callback handlers.
3211  */
3212 static void
3213 _scsih_tm_display_info(struct MPT3SAS_ADAPTER *ioc, struct scsi_cmnd *scmd)
3214 {
3215 	struct scsi_target *starget = scmd->device->sdev_target;
3216 	struct MPT3SAS_TARGET *priv_target = starget->hostdata;
3217 	struct _sas_device *sas_device = NULL;
3218 	struct _pcie_device *pcie_device = NULL;
3219 	unsigned long flags;
3220 	char *device_str = NULL;
3221 
3222 	if (!priv_target)
3223 		return;
3224 	if (ioc->hide_ir_msg)
3225 		device_str = "WarpDrive";
3226 	else
3227 		device_str = "volume";
3228 
3229 	scsi_print_command(scmd);
3230 	if (priv_target->flags & MPT_TARGET_FLAGS_VOLUME) {
3231 		starget_printk(KERN_INFO, starget,
3232 			"%s handle(0x%04x), %s wwid(0x%016llx)\n",
3233 			device_str, priv_target->handle,
3234 		    device_str, (unsigned long long)priv_target->sas_address);
3235 
3236 	} else if (priv_target->flags & MPT_TARGET_FLAGS_PCIE_DEVICE) {
3237 		spin_lock_irqsave(&ioc->pcie_device_lock, flags);
3238 		pcie_device = __mpt3sas_get_pdev_from_target(ioc, priv_target);
3239 		if (pcie_device) {
3240 			starget_printk(KERN_INFO, starget,
3241 				"handle(0x%04x), wwid(0x%016llx), port(%d)\n",
3242 				pcie_device->handle,
3243 				(unsigned long long)pcie_device->wwid,
3244 				pcie_device->port_num);
3245 			if (pcie_device->enclosure_handle != 0)
3246 				starget_printk(KERN_INFO, starget,
3247 					"enclosure logical id(0x%016llx), slot(%d)\n",
3248 					(unsigned long long)
3249 					pcie_device->enclosure_logical_id,
3250 					pcie_device->slot);
3251 			if (pcie_device->connector_name[0] != '\0')
3252 				starget_printk(KERN_INFO, starget,
3253 					"enclosure level(0x%04x), connector name( %s)\n",
3254 					pcie_device->enclosure_level,
3255 					pcie_device->connector_name);
3256 			pcie_device_put(pcie_device);
3257 		}
3258 		spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
3259 
3260 	} else {
3261 		spin_lock_irqsave(&ioc->sas_device_lock, flags);
3262 		sas_device = __mpt3sas_get_sdev_from_target(ioc, priv_target);
3263 		if (sas_device) {
3264 			if (priv_target->flags &
3265 			    MPT_TARGET_FLAGS_RAID_COMPONENT) {
3266 				starget_printk(KERN_INFO, starget,
3267 				    "volume handle(0x%04x), "
3268 				    "volume wwid(0x%016llx)\n",
3269 				    sas_device->volume_handle,
3270 				   (unsigned long long)sas_device->volume_wwid);
3271 			}
3272 			starget_printk(KERN_INFO, starget,
3273 			    "handle(0x%04x), sas_address(0x%016llx), phy(%d)\n",
3274 			    sas_device->handle,
3275 			    (unsigned long long)sas_device->sas_address,
3276 			    sas_device->phy);
3277 
3278 			_scsih_display_enclosure_chassis_info(NULL, sas_device,
3279 			    NULL, starget);
3280 
3281 			sas_device_put(sas_device);
3282 		}
3283 		spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
3284 	}
3285 }
3286 
3287 /**
3288  * scsih_abort - eh threads main abort routine
3289  * @scmd: pointer to scsi command object
3290  *
3291  * Return: SUCCESS if command aborted else FAILED
3292  */
3293 static int
3294 scsih_abort(struct scsi_cmnd *scmd)
3295 {
3296 	struct MPT3SAS_ADAPTER *ioc = shost_priv(scmd->device->host);
3297 	struct MPT3SAS_DEVICE *sas_device_priv_data;
3298 	struct scsiio_tracker *st = scsi_cmd_priv(scmd);
3299 	u16 handle;
3300 	int r;
3301 
3302 	u8 timeout = 30;
3303 	struct _pcie_device *pcie_device = NULL;
3304 	sdev_printk(KERN_INFO, scmd->device, "attempting task abort!"
3305 	    "scmd(0x%p), outstanding for %u ms & timeout %u ms\n",
3306 	    scmd, jiffies_to_msecs(jiffies - scmd->jiffies_at_alloc),
3307 	    (scsi_cmd_to_rq(scmd)->timeout / HZ) * 1000);
3308 	_scsih_tm_display_info(ioc, scmd);
3309 
3310 	sas_device_priv_data = scmd->device->hostdata;
3311 	if (!sas_device_priv_data || !sas_device_priv_data->sas_target ||
3312 	    ioc->remove_host) {
3313 		sdev_printk(KERN_INFO, scmd->device,
3314 		    "device been deleted! scmd(0x%p)\n", scmd);
3315 		scmd->result = DID_NO_CONNECT << 16;
3316 		scsi_done(scmd);
3317 		r = SUCCESS;
3318 		goto out;
3319 	}
3320 
3321 	/* check for completed command */
3322 	if (st == NULL || st->cb_idx == 0xFF) {
3323 		sdev_printk(KERN_INFO, scmd->device, "No reference found at "
3324 		    "driver, assuming scmd(0x%p) might have completed\n", scmd);
3325 		scmd->result = DID_RESET << 16;
3326 		r = SUCCESS;
3327 		goto out;
3328 	}
3329 
3330 	/* for hidden raid components and volumes this is not supported */
3331 	if (sas_device_priv_data->sas_target->flags &
3332 	    MPT_TARGET_FLAGS_RAID_COMPONENT ||
3333 	    sas_device_priv_data->sas_target->flags & MPT_TARGET_FLAGS_VOLUME) {
3334 		scmd->result = DID_RESET << 16;
3335 		r = FAILED;
3336 		goto out;
3337 	}
3338 
3339 	mpt3sas_halt_firmware(ioc);
3340 
3341 	handle = sas_device_priv_data->sas_target->handle;
3342 	pcie_device = mpt3sas_get_pdev_by_handle(ioc, handle);
3343 	if (pcie_device && (!ioc->tm_custom_handling) &&
3344 	    (!(mpt3sas_scsih_is_pcie_scsi_device(pcie_device->device_info))))
3345 		timeout = ioc->nvme_abort_timeout;
3346 	r = mpt3sas_scsih_issue_locked_tm(ioc, handle, scmd->device->channel,
3347 		scmd->device->id, scmd->device->lun,
3348 		MPI2_SCSITASKMGMT_TASKTYPE_ABORT_TASK,
3349 		st->smid, st->msix_io, timeout, 0);
3350 	/* Command must be cleared after abort */
3351 	if (r == SUCCESS && st->cb_idx != 0xFF)
3352 		r = FAILED;
3353  out:
3354 	sdev_printk(KERN_INFO, scmd->device, "task abort: %s scmd(0x%p)\n",
3355 	    ((r == SUCCESS) ? "SUCCESS" : "FAILED"), scmd);
3356 	if (pcie_device)
3357 		pcie_device_put(pcie_device);
3358 	return r;
3359 }
3360 
3361 /**
3362  * scsih_dev_reset - eh threads main device reset routine
3363  * @scmd: pointer to scsi command object
3364  *
3365  * Return: SUCCESS if command aborted else FAILED
3366  */
3367 static int
3368 scsih_dev_reset(struct scsi_cmnd *scmd)
3369 {
3370 	struct MPT3SAS_ADAPTER *ioc = shost_priv(scmd->device->host);
3371 	struct MPT3SAS_DEVICE *sas_device_priv_data;
3372 	struct _sas_device *sas_device = NULL;
3373 	struct _pcie_device *pcie_device = NULL;
3374 	u16	handle;
3375 	u8	tr_method = 0;
3376 	u8	tr_timeout = 30;
3377 	int r;
3378 
3379 	struct scsi_target *starget = scmd->device->sdev_target;
3380 	struct MPT3SAS_TARGET *target_priv_data = starget->hostdata;
3381 
3382 	sdev_printk(KERN_INFO, scmd->device,
3383 	    "attempting device reset! scmd(0x%p)\n", scmd);
3384 	_scsih_tm_display_info(ioc, scmd);
3385 
3386 	sas_device_priv_data = scmd->device->hostdata;
3387 	if (!sas_device_priv_data || !sas_device_priv_data->sas_target ||
3388 	    ioc->remove_host) {
3389 		sdev_printk(KERN_INFO, scmd->device,
3390 		    "device been deleted! scmd(0x%p)\n", scmd);
3391 		scmd->result = DID_NO_CONNECT << 16;
3392 		scsi_done(scmd);
3393 		r = SUCCESS;
3394 		goto out;
3395 	}
3396 
3397 	/* for hidden raid components obtain the volume_handle */
3398 	handle = 0;
3399 	if (sas_device_priv_data->sas_target->flags &
3400 	    MPT_TARGET_FLAGS_RAID_COMPONENT) {
3401 		sas_device = mpt3sas_get_sdev_from_target(ioc,
3402 				target_priv_data);
3403 		if (sas_device)
3404 			handle = sas_device->volume_handle;
3405 	} else
3406 		handle = sas_device_priv_data->sas_target->handle;
3407 
3408 	if (!handle) {
3409 		scmd->result = DID_RESET << 16;
3410 		r = FAILED;
3411 		goto out;
3412 	}
3413 
3414 	pcie_device = mpt3sas_get_pdev_by_handle(ioc, handle);
3415 
3416 	if (pcie_device && (!ioc->tm_custom_handling) &&
3417 	    (!(mpt3sas_scsih_is_pcie_scsi_device(pcie_device->device_info)))) {
3418 		tr_timeout = pcie_device->reset_timeout;
3419 		tr_method = MPI26_SCSITASKMGMT_MSGFLAGS_PROTOCOL_LVL_RST_PCIE;
3420 	} else
3421 		tr_method = MPI2_SCSITASKMGMT_MSGFLAGS_LINK_RESET;
3422 
3423 	r = mpt3sas_scsih_issue_locked_tm(ioc, handle, scmd->device->channel,
3424 		scmd->device->id, scmd->device->lun,
3425 		MPI2_SCSITASKMGMT_TASKTYPE_LOGICAL_UNIT_RESET, 0, 0,
3426 		tr_timeout, tr_method);
3427 	/* Check for busy commands after reset */
3428 	if (r == SUCCESS && scsi_device_busy(scmd->device))
3429 		r = FAILED;
3430  out:
3431 	sdev_printk(KERN_INFO, scmd->device, "device reset: %s scmd(0x%p)\n",
3432 	    ((r == SUCCESS) ? "SUCCESS" : "FAILED"), scmd);
3433 
3434 	if (sas_device)
3435 		sas_device_put(sas_device);
3436 	if (pcie_device)
3437 		pcie_device_put(pcie_device);
3438 
3439 	return r;
3440 }
3441 
3442 /**
3443  * scsih_target_reset - eh threads main target reset routine
3444  * @scmd: pointer to scsi command object
3445  *
3446  * Return: SUCCESS if command aborted else FAILED
3447  */
3448 static int
3449 scsih_target_reset(struct scsi_cmnd *scmd)
3450 {
3451 	struct MPT3SAS_ADAPTER *ioc = shost_priv(scmd->device->host);
3452 	struct MPT3SAS_DEVICE *sas_device_priv_data;
3453 	struct _sas_device *sas_device = NULL;
3454 	struct _pcie_device *pcie_device = NULL;
3455 	u16	handle;
3456 	u8	tr_method = 0;
3457 	u8	tr_timeout = 30;
3458 	int r;
3459 	struct scsi_target *starget = scmd->device->sdev_target;
3460 	struct MPT3SAS_TARGET *target_priv_data = starget->hostdata;
3461 
3462 	starget_printk(KERN_INFO, starget,
3463 	    "attempting target reset! scmd(0x%p)\n", scmd);
3464 	_scsih_tm_display_info(ioc, scmd);
3465 
3466 	sas_device_priv_data = scmd->device->hostdata;
3467 	if (!sas_device_priv_data || !sas_device_priv_data->sas_target ||
3468 	    ioc->remove_host) {
3469 		starget_printk(KERN_INFO, starget,
3470 		    "target been deleted! scmd(0x%p)\n", scmd);
3471 		scmd->result = DID_NO_CONNECT << 16;
3472 		scsi_done(scmd);
3473 		r = SUCCESS;
3474 		goto out;
3475 	}
3476 
3477 	/* for hidden raid components obtain the volume_handle */
3478 	handle = 0;
3479 	if (sas_device_priv_data->sas_target->flags &
3480 	    MPT_TARGET_FLAGS_RAID_COMPONENT) {
3481 		sas_device = mpt3sas_get_sdev_from_target(ioc,
3482 				target_priv_data);
3483 		if (sas_device)
3484 			handle = sas_device->volume_handle;
3485 	} else
3486 		handle = sas_device_priv_data->sas_target->handle;
3487 
3488 	if (!handle) {
3489 		scmd->result = DID_RESET << 16;
3490 		r = FAILED;
3491 		goto out;
3492 	}
3493 
3494 	pcie_device = mpt3sas_get_pdev_by_handle(ioc, handle);
3495 
3496 	if (pcie_device && (!ioc->tm_custom_handling) &&
3497 	    (!(mpt3sas_scsih_is_pcie_scsi_device(pcie_device->device_info)))) {
3498 		tr_timeout = pcie_device->reset_timeout;
3499 		tr_method = MPI26_SCSITASKMGMT_MSGFLAGS_PROTOCOL_LVL_RST_PCIE;
3500 	} else
3501 		tr_method = MPI2_SCSITASKMGMT_MSGFLAGS_LINK_RESET;
3502 	r = mpt3sas_scsih_issue_locked_tm(ioc, handle, scmd->device->channel,
3503 		scmd->device->id, 0,
3504 		MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET, 0, 0,
3505 	    tr_timeout, tr_method);
3506 	/* Check for busy commands after reset */
3507 	if (r == SUCCESS && atomic_read(&starget->target_busy))
3508 		r = FAILED;
3509  out:
3510 	starget_printk(KERN_INFO, starget, "target reset: %s scmd(0x%p)\n",
3511 	    ((r == SUCCESS) ? "SUCCESS" : "FAILED"), scmd);
3512 
3513 	if (sas_device)
3514 		sas_device_put(sas_device);
3515 	if (pcie_device)
3516 		pcie_device_put(pcie_device);
3517 	return r;
3518 }
3519 
3520 
3521 /**
3522  * scsih_host_reset - eh threads main host reset routine
3523  * @scmd: pointer to scsi command object
3524  *
3525  * Return: SUCCESS if command aborted else FAILED
3526  */
3527 static int
3528 scsih_host_reset(struct scsi_cmnd *scmd)
3529 {
3530 	struct MPT3SAS_ADAPTER *ioc = shost_priv(scmd->device->host);
3531 	int r, retval;
3532 
3533 	ioc_info(ioc, "attempting host reset! scmd(0x%p)\n", scmd);
3534 	scsi_print_command(scmd);
3535 
3536 	if (ioc->is_driver_loading || ioc->remove_host) {
3537 		ioc_info(ioc, "Blocking the host reset\n");
3538 		r = FAILED;
3539 		goto out;
3540 	}
3541 
3542 	retval = mpt3sas_base_hard_reset_handler(ioc, FORCE_BIG_HAMMER);
3543 	r = (retval < 0) ? FAILED : SUCCESS;
3544 out:
3545 	ioc_info(ioc, "host reset: %s scmd(0x%p)\n",
3546 		 r == SUCCESS ? "SUCCESS" : "FAILED", scmd);
3547 
3548 	return r;
3549 }
3550 
3551 /**
3552  * _scsih_fw_event_add - insert and queue up fw_event
3553  * @ioc: per adapter object
3554  * @fw_event: object describing the event
3555  * Context: This function will acquire ioc->fw_event_lock.
3556  *
3557  * This adds the firmware event object into link list, then queues it up to
3558  * be processed from user context.
3559  */
3560 static void
3561 _scsih_fw_event_add(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work *fw_event)
3562 {
3563 	unsigned long flags;
3564 
3565 	if (ioc->firmware_event_thread == NULL)
3566 		return;
3567 
3568 	spin_lock_irqsave(&ioc->fw_event_lock, flags);
3569 	fw_event_work_get(fw_event);
3570 	INIT_LIST_HEAD(&fw_event->list);
3571 	list_add_tail(&fw_event->list, &ioc->fw_event_list);
3572 	INIT_WORK(&fw_event->work, _firmware_event_work);
3573 	fw_event_work_get(fw_event);
3574 	queue_work(ioc->firmware_event_thread, &fw_event->work);
3575 	spin_unlock_irqrestore(&ioc->fw_event_lock, flags);
3576 }
3577 
3578 /**
3579  * _scsih_fw_event_del_from_list - delete fw_event from the list
3580  * @ioc: per adapter object
3581  * @fw_event: object describing the event
3582  * Context: This function will acquire ioc->fw_event_lock.
3583  *
3584  * If the fw_event is on the fw_event_list, remove it and do a put.
3585  */
3586 static void
3587 _scsih_fw_event_del_from_list(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work
3588 	*fw_event)
3589 {
3590 	unsigned long flags;
3591 
3592 	spin_lock_irqsave(&ioc->fw_event_lock, flags);
3593 	if (!list_empty(&fw_event->list)) {
3594 		list_del_init(&fw_event->list);
3595 		fw_event_work_put(fw_event);
3596 	}
3597 	spin_unlock_irqrestore(&ioc->fw_event_lock, flags);
3598 }
3599 
3600 
3601  /**
3602  * mpt3sas_send_trigger_data_event - send event for processing trigger data
3603  * @ioc: per adapter object
3604  * @event_data: trigger event data
3605  */
3606 void
3607 mpt3sas_send_trigger_data_event(struct MPT3SAS_ADAPTER *ioc,
3608 	struct SL_WH_TRIGGERS_EVENT_DATA_T *event_data)
3609 {
3610 	struct fw_event_work *fw_event;
3611 	u16 sz;
3612 
3613 	if (ioc->is_driver_loading)
3614 		return;
3615 	sz = sizeof(*event_data);
3616 	fw_event = alloc_fw_event_work(sz);
3617 	if (!fw_event)
3618 		return;
3619 	fw_event->event = MPT3SAS_PROCESS_TRIGGER_DIAG;
3620 	fw_event->ioc = ioc;
3621 	memcpy(fw_event->event_data, event_data, sizeof(*event_data));
3622 	_scsih_fw_event_add(ioc, fw_event);
3623 	fw_event_work_put(fw_event);
3624 }
3625 
3626 /**
3627  * _scsih_error_recovery_delete_devices - remove devices not responding
3628  * @ioc: per adapter object
3629  */
3630 static void
3631 _scsih_error_recovery_delete_devices(struct MPT3SAS_ADAPTER *ioc)
3632 {
3633 	struct fw_event_work *fw_event;
3634 
3635 	fw_event = alloc_fw_event_work(0);
3636 	if (!fw_event)
3637 		return;
3638 	fw_event->event = MPT3SAS_REMOVE_UNRESPONDING_DEVICES;
3639 	fw_event->ioc = ioc;
3640 	_scsih_fw_event_add(ioc, fw_event);
3641 	fw_event_work_put(fw_event);
3642 }
3643 
3644 /**
3645  * mpt3sas_port_enable_complete - port enable completed (fake event)
3646  * @ioc: per adapter object
3647  */
3648 void
3649 mpt3sas_port_enable_complete(struct MPT3SAS_ADAPTER *ioc)
3650 {
3651 	struct fw_event_work *fw_event;
3652 
3653 	fw_event = alloc_fw_event_work(0);
3654 	if (!fw_event)
3655 		return;
3656 	fw_event->event = MPT3SAS_PORT_ENABLE_COMPLETE;
3657 	fw_event->ioc = ioc;
3658 	_scsih_fw_event_add(ioc, fw_event);
3659 	fw_event_work_put(fw_event);
3660 }
3661 
3662 static struct fw_event_work *dequeue_next_fw_event(struct MPT3SAS_ADAPTER *ioc)
3663 {
3664 	unsigned long flags;
3665 	struct fw_event_work *fw_event = NULL;
3666 
3667 	spin_lock_irqsave(&ioc->fw_event_lock, flags);
3668 	if (!list_empty(&ioc->fw_event_list)) {
3669 		fw_event = list_first_entry(&ioc->fw_event_list,
3670 				struct fw_event_work, list);
3671 		list_del_init(&fw_event->list);
3672 		fw_event_work_put(fw_event);
3673 	}
3674 	spin_unlock_irqrestore(&ioc->fw_event_lock, flags);
3675 
3676 	return fw_event;
3677 }
3678 
3679 /**
3680  * _scsih_fw_event_cleanup_queue - cleanup event queue
3681  * @ioc: per adapter object
3682  *
3683  * Walk the firmware event queue, either killing timers, or waiting
3684  * for outstanding events to complete
3685  *
3686  * Context: task, can sleep
3687  */
3688 static void
3689 _scsih_fw_event_cleanup_queue(struct MPT3SAS_ADAPTER *ioc)
3690 {
3691 	struct fw_event_work *fw_event;
3692 
3693 	if ((list_empty(&ioc->fw_event_list) && !ioc->current_event) ||
3694 	    !ioc->firmware_event_thread)
3695 		return;
3696 	/*
3697 	 * Set current running event as ignore, so that
3698 	 * current running event will exit quickly.
3699 	 * As diag reset has occurred it is of no use
3700 	 * to process remaining stale event data entries.
3701 	 */
3702 	if (ioc->shost_recovery && ioc->current_event)
3703 		ioc->current_event->ignore = 1;
3704 
3705 	ioc->fw_events_cleanup = 1;
3706 	while ((fw_event = dequeue_next_fw_event(ioc)) ||
3707 	     (fw_event = ioc->current_event)) {
3708 
3709 		/*
3710 		 * Don't call cancel_work_sync() for current_event
3711 		 * other than MPT3SAS_REMOVE_UNRESPONDING_DEVICES;
3712 		 * otherwise we may observe deadlock if current
3713 		 * hard reset issued as part of processing the current_event.
3714 		 *
3715 		 * Orginal logic of cleaning the current_event is added
3716 		 * for handling the back to back host reset issued by the user.
3717 		 * i.e. during back to back host reset, driver use to process
3718 		 * the two instances of MPT3SAS_REMOVE_UNRESPONDING_DEVICES
3719 		 * event back to back and this made the drives to unregister
3720 		 * the devices from SML.
3721 		 */
3722 
3723 		if (fw_event == ioc->current_event &&
3724 		    ioc->current_event->event !=
3725 		    MPT3SAS_REMOVE_UNRESPONDING_DEVICES) {
3726 			ioc->current_event = NULL;
3727 			continue;
3728 		}
3729 
3730 		/*
3731 		 * Driver has to clear ioc->start_scan flag when
3732 		 * it is cleaning up MPT3SAS_PORT_ENABLE_COMPLETE,
3733 		 * otherwise scsi_scan_host() API waits for the
3734 		 * 5 minute timer to expire. If we exit from
3735 		 * scsi_scan_host() early then we can issue the
3736 		 * new port enable request as part of current diag reset.
3737 		 */
3738 		if (fw_event->event == MPT3SAS_PORT_ENABLE_COMPLETE) {
3739 			ioc->port_enable_cmds.status |= MPT3_CMD_RESET;
3740 			ioc->start_scan = 0;
3741 		}
3742 
3743 		/*
3744 		 * Wait on the fw_event to complete. If this returns 1, then
3745 		 * the event was never executed, and we need a put for the
3746 		 * reference the work had on the fw_event.
3747 		 *
3748 		 * If it did execute, we wait for it to finish, and the put will
3749 		 * happen from _firmware_event_work()
3750 		 */
3751 		if (cancel_work_sync(&fw_event->work))
3752 			fw_event_work_put(fw_event);
3753 
3754 	}
3755 	ioc->fw_events_cleanup = 0;
3756 }
3757 
3758 /**
3759  * _scsih_internal_device_block - block the sdev device
3760  * @sdev: per device object
3761  * @sas_device_priv_data : per device driver private data
3762  *
3763  * make sure device is blocked without error, if not
3764  * print an error
3765  */
3766 static void
3767 _scsih_internal_device_block(struct scsi_device *sdev,
3768 			struct MPT3SAS_DEVICE *sas_device_priv_data)
3769 {
3770 	int r = 0;
3771 
3772 	sdev_printk(KERN_INFO, sdev, "device_block, handle(0x%04x)\n",
3773 	    sas_device_priv_data->sas_target->handle);
3774 	sas_device_priv_data->block = 1;
3775 
3776 	r = scsi_internal_device_block_nowait(sdev);
3777 	if (r == -EINVAL)
3778 		sdev_printk(KERN_WARNING, sdev,
3779 		    "device_block failed with return(%d) for handle(0x%04x)\n",
3780 		    r, sas_device_priv_data->sas_target->handle);
3781 }
3782 
3783 /**
3784  * _scsih_internal_device_unblock - unblock the sdev device
3785  * @sdev: per device object
3786  * @sas_device_priv_data : per device driver private data
3787  * make sure device is unblocked without error, if not retry
3788  * by blocking and then unblocking
3789  */
3790 
3791 static void
3792 _scsih_internal_device_unblock(struct scsi_device *sdev,
3793 			struct MPT3SAS_DEVICE *sas_device_priv_data)
3794 {
3795 	int r = 0;
3796 
3797 	sdev_printk(KERN_WARNING, sdev, "device_unblock and setting to running, "
3798 	    "handle(0x%04x)\n", sas_device_priv_data->sas_target->handle);
3799 	sas_device_priv_data->block = 0;
3800 	r = scsi_internal_device_unblock_nowait(sdev, SDEV_RUNNING);
3801 	if (r == -EINVAL) {
3802 		/* The device has been set to SDEV_RUNNING by SD layer during
3803 		 * device addition but the request queue is still stopped by
3804 		 * our earlier block call. We need to perform a block again
3805 		 * to get the device to SDEV_BLOCK and then to SDEV_RUNNING */
3806 
3807 		sdev_printk(KERN_WARNING, sdev,
3808 		    "device_unblock failed with return(%d) for handle(0x%04x) "
3809 		    "performing a block followed by an unblock\n",
3810 		    r, sas_device_priv_data->sas_target->handle);
3811 		sas_device_priv_data->block = 1;
3812 		r = scsi_internal_device_block_nowait(sdev);
3813 		if (r)
3814 			sdev_printk(KERN_WARNING, sdev, "retried device_block "
3815 			    "failed with return(%d) for handle(0x%04x)\n",
3816 			    r, sas_device_priv_data->sas_target->handle);
3817 
3818 		sas_device_priv_data->block = 0;
3819 		r = scsi_internal_device_unblock_nowait(sdev, SDEV_RUNNING);
3820 		if (r)
3821 			sdev_printk(KERN_WARNING, sdev, "retried device_unblock"
3822 			    " failed with return(%d) for handle(0x%04x)\n",
3823 			    r, sas_device_priv_data->sas_target->handle);
3824 	}
3825 }
3826 
3827 /**
3828  * _scsih_ublock_io_all_device - unblock every device
3829  * @ioc: per adapter object
3830  *
3831  * change the device state from block to running
3832  */
3833 static void
3834 _scsih_ublock_io_all_device(struct MPT3SAS_ADAPTER *ioc)
3835 {
3836 	struct MPT3SAS_DEVICE *sas_device_priv_data;
3837 	struct scsi_device *sdev;
3838 
3839 	shost_for_each_device(sdev, ioc->shost) {
3840 		sas_device_priv_data = sdev->hostdata;
3841 		if (!sas_device_priv_data)
3842 			continue;
3843 		if (!sas_device_priv_data->block)
3844 			continue;
3845 
3846 		dewtprintk(ioc, sdev_printk(KERN_INFO, sdev,
3847 			"device_running, handle(0x%04x)\n",
3848 		    sas_device_priv_data->sas_target->handle));
3849 		_scsih_internal_device_unblock(sdev, sas_device_priv_data);
3850 	}
3851 }
3852 
3853 
3854 /**
3855  * _scsih_ublock_io_device - prepare device to be deleted
3856  * @ioc: per adapter object
3857  * @sas_address: sas address
3858  * @port: hba port entry
3859  *
3860  * unblock then put device in offline state
3861  */
3862 static void
3863 _scsih_ublock_io_device(struct MPT3SAS_ADAPTER *ioc,
3864 	u64 sas_address, struct hba_port *port)
3865 {
3866 	struct MPT3SAS_DEVICE *sas_device_priv_data;
3867 	struct scsi_device *sdev;
3868 
3869 	shost_for_each_device(sdev, ioc->shost) {
3870 		sas_device_priv_data = sdev->hostdata;
3871 		if (!sas_device_priv_data || !sas_device_priv_data->sas_target)
3872 			continue;
3873 		if (sas_device_priv_data->sas_target->sas_address
3874 		    != sas_address)
3875 			continue;
3876 		if (sas_device_priv_data->sas_target->port != port)
3877 			continue;
3878 		if (sas_device_priv_data->block)
3879 			_scsih_internal_device_unblock(sdev,
3880 				sas_device_priv_data);
3881 	}
3882 }
3883 
3884 /**
3885  * _scsih_block_io_all_device - set the device state to SDEV_BLOCK
3886  * @ioc: per adapter object
3887  *
3888  * During device pull we need to appropriately set the sdev state.
3889  */
3890 static void
3891 _scsih_block_io_all_device(struct MPT3SAS_ADAPTER *ioc)
3892 {
3893 	struct MPT3SAS_DEVICE *sas_device_priv_data;
3894 	struct scsi_device *sdev;
3895 
3896 	shost_for_each_device(sdev, ioc->shost) {
3897 		sas_device_priv_data = sdev->hostdata;
3898 		if (!sas_device_priv_data)
3899 			continue;
3900 		if (sas_device_priv_data->block)
3901 			continue;
3902 		if (sas_device_priv_data->ignore_delay_remove) {
3903 			sdev_printk(KERN_INFO, sdev,
3904 			"%s skip device_block for SES handle(0x%04x)\n",
3905 			__func__, sas_device_priv_data->sas_target->handle);
3906 			continue;
3907 		}
3908 		_scsih_internal_device_block(sdev, sas_device_priv_data);
3909 	}
3910 }
3911 
3912 /**
3913  * _scsih_block_io_device - set the device state to SDEV_BLOCK
3914  * @ioc: per adapter object
3915  * @handle: device handle
3916  *
3917  * During device pull we need to appropriately set the sdev state.
3918  */
3919 static void
3920 _scsih_block_io_device(struct MPT3SAS_ADAPTER *ioc, u16 handle)
3921 {
3922 	struct MPT3SAS_DEVICE *sas_device_priv_data;
3923 	struct scsi_device *sdev;
3924 	struct _sas_device *sas_device;
3925 
3926 	sas_device = mpt3sas_get_sdev_by_handle(ioc, handle);
3927 
3928 	shost_for_each_device(sdev, ioc->shost) {
3929 		sas_device_priv_data = sdev->hostdata;
3930 		if (!sas_device_priv_data)
3931 			continue;
3932 		if (sas_device_priv_data->sas_target->handle != handle)
3933 			continue;
3934 		if (sas_device_priv_data->block)
3935 			continue;
3936 		if (sas_device && sas_device->pend_sas_rphy_add)
3937 			continue;
3938 		if (sas_device_priv_data->ignore_delay_remove) {
3939 			sdev_printk(KERN_INFO, sdev,
3940 			"%s skip device_block for SES handle(0x%04x)\n",
3941 			__func__, sas_device_priv_data->sas_target->handle);
3942 			continue;
3943 		}
3944 		_scsih_internal_device_block(sdev, sas_device_priv_data);
3945 	}
3946 
3947 	if (sas_device)
3948 		sas_device_put(sas_device);
3949 }
3950 
3951 /**
3952  * _scsih_block_io_to_children_attached_to_ex
3953  * @ioc: per adapter object
3954  * @sas_expander: the sas_device object
3955  *
3956  * This routine set sdev state to SDEV_BLOCK for all devices
3957  * attached to this expander. This function called when expander is
3958  * pulled.
3959  */
3960 static void
3961 _scsih_block_io_to_children_attached_to_ex(struct MPT3SAS_ADAPTER *ioc,
3962 	struct _sas_node *sas_expander)
3963 {
3964 	struct _sas_port *mpt3sas_port;
3965 	struct _sas_device *sas_device;
3966 	struct _sas_node *expander_sibling;
3967 	unsigned long flags;
3968 
3969 	if (!sas_expander)
3970 		return;
3971 
3972 	list_for_each_entry(mpt3sas_port,
3973 	   &sas_expander->sas_port_list, port_list) {
3974 		if (mpt3sas_port->remote_identify.device_type ==
3975 		    SAS_END_DEVICE) {
3976 			spin_lock_irqsave(&ioc->sas_device_lock, flags);
3977 			sas_device = __mpt3sas_get_sdev_by_addr(ioc,
3978 			    mpt3sas_port->remote_identify.sas_address,
3979 			    mpt3sas_port->hba_port);
3980 			if (sas_device) {
3981 				set_bit(sas_device->handle,
3982 						ioc->blocking_handles);
3983 				sas_device_put(sas_device);
3984 			}
3985 			spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
3986 		}
3987 	}
3988 
3989 	list_for_each_entry(mpt3sas_port,
3990 	   &sas_expander->sas_port_list, port_list) {
3991 
3992 		if (mpt3sas_port->remote_identify.device_type ==
3993 		    SAS_EDGE_EXPANDER_DEVICE ||
3994 		    mpt3sas_port->remote_identify.device_type ==
3995 		    SAS_FANOUT_EXPANDER_DEVICE) {
3996 			expander_sibling =
3997 			    mpt3sas_scsih_expander_find_by_sas_address(
3998 			    ioc, mpt3sas_port->remote_identify.sas_address,
3999 			    mpt3sas_port->hba_port);
4000 			_scsih_block_io_to_children_attached_to_ex(ioc,
4001 			    expander_sibling);
4002 		}
4003 	}
4004 }
4005 
4006 /**
4007  * _scsih_block_io_to_children_attached_directly
4008  * @ioc: per adapter object
4009  * @event_data: topology change event data
4010  *
4011  * This routine set sdev state to SDEV_BLOCK for all devices
4012  * direct attached during device pull.
4013  */
4014 static void
4015 _scsih_block_io_to_children_attached_directly(struct MPT3SAS_ADAPTER *ioc,
4016 	Mpi2EventDataSasTopologyChangeList_t *event_data)
4017 {
4018 	int i;
4019 	u16 handle;
4020 	u16 reason_code;
4021 
4022 	for (i = 0; i < event_data->NumEntries; i++) {
4023 		handle = le16_to_cpu(event_data->PHY[i].AttachedDevHandle);
4024 		if (!handle)
4025 			continue;
4026 		reason_code = event_data->PHY[i].PhyStatus &
4027 		    MPI2_EVENT_SAS_TOPO_RC_MASK;
4028 		if (reason_code == MPI2_EVENT_SAS_TOPO_RC_DELAY_NOT_RESPONDING)
4029 			_scsih_block_io_device(ioc, handle);
4030 	}
4031 }
4032 
4033 /**
4034  * _scsih_block_io_to_pcie_children_attached_directly
4035  * @ioc: per adapter object
4036  * @event_data: topology change event data
4037  *
4038  * This routine set sdev state to SDEV_BLOCK for all devices
4039  * direct attached during device pull/reconnect.
4040  */
4041 static void
4042 _scsih_block_io_to_pcie_children_attached_directly(struct MPT3SAS_ADAPTER *ioc,
4043 		Mpi26EventDataPCIeTopologyChangeList_t *event_data)
4044 {
4045 	int i;
4046 	u16 handle;
4047 	u16 reason_code;
4048 
4049 	for (i = 0; i < event_data->NumEntries; i++) {
4050 		handle =
4051 			le16_to_cpu(event_data->PortEntry[i].AttachedDevHandle);
4052 		if (!handle)
4053 			continue;
4054 		reason_code = event_data->PortEntry[i].PortStatus;
4055 		if (reason_code ==
4056 				MPI26_EVENT_PCIE_TOPO_PS_DELAY_NOT_RESPONDING)
4057 			_scsih_block_io_device(ioc, handle);
4058 	}
4059 }
4060 /**
4061  * _scsih_tm_tr_send - send task management request
4062  * @ioc: per adapter object
4063  * @handle: device handle
4064  * Context: interrupt time.
4065  *
4066  * This code is to initiate the device removal handshake protocol
4067  * with controller firmware.  This function will issue target reset
4068  * using high priority request queue.  It will send a sas iounit
4069  * control request (MPI2_SAS_OP_REMOVE_DEVICE) from this completion.
4070  *
4071  * This is designed to send muliple task management request at the same
4072  * time to the fifo. If the fifo is full, we will append the request,
4073  * and process it in a future completion.
4074  */
4075 static void
4076 _scsih_tm_tr_send(struct MPT3SAS_ADAPTER *ioc, u16 handle)
4077 {
4078 	Mpi2SCSITaskManagementRequest_t *mpi_request;
4079 	u16 smid;
4080 	struct _sas_device *sas_device = NULL;
4081 	struct _pcie_device *pcie_device = NULL;
4082 	struct MPT3SAS_TARGET *sas_target_priv_data = NULL;
4083 	u64 sas_address = 0;
4084 	unsigned long flags;
4085 	struct _tr_list *delayed_tr;
4086 	u32 ioc_state;
4087 	u8 tr_method = 0;
4088 	struct hba_port *port = NULL;
4089 
4090 	if (ioc->pci_error_recovery) {
4091 		dewtprintk(ioc,
4092 			   ioc_info(ioc, "%s: host in pci error recovery: handle(0x%04x)\n",
4093 				    __func__, handle));
4094 		return;
4095 	}
4096 	ioc_state = mpt3sas_base_get_iocstate(ioc, 1);
4097 	if (ioc_state != MPI2_IOC_STATE_OPERATIONAL) {
4098 		dewtprintk(ioc,
4099 			   ioc_info(ioc, "%s: host is not operational: handle(0x%04x)\n",
4100 				    __func__, handle));
4101 		return;
4102 	}
4103 
4104 	/* if PD, then return */
4105 	if (test_bit(handle, ioc->pd_handles))
4106 		return;
4107 
4108 	clear_bit(handle, ioc->pend_os_device_add);
4109 
4110 	spin_lock_irqsave(&ioc->sas_device_lock, flags);
4111 	sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle);
4112 	if (sas_device && sas_device->starget &&
4113 	    sas_device->starget->hostdata) {
4114 		sas_target_priv_data = sas_device->starget->hostdata;
4115 		sas_target_priv_data->deleted = 1;
4116 		sas_address = sas_device->sas_address;
4117 		port = sas_device->port;
4118 	}
4119 	spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
4120 	if (!sas_device) {
4121 		spin_lock_irqsave(&ioc->pcie_device_lock, flags);
4122 		pcie_device = __mpt3sas_get_pdev_by_handle(ioc, handle);
4123 		if (pcie_device && pcie_device->starget &&
4124 			pcie_device->starget->hostdata) {
4125 			sas_target_priv_data = pcie_device->starget->hostdata;
4126 			sas_target_priv_data->deleted = 1;
4127 			sas_address = pcie_device->wwid;
4128 		}
4129 		spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
4130 		if (pcie_device && (!ioc->tm_custom_handling) &&
4131 		    (!(mpt3sas_scsih_is_pcie_scsi_device(
4132 		    pcie_device->device_info))))
4133 			tr_method =
4134 			    MPI26_SCSITASKMGMT_MSGFLAGS_PROTOCOL_LVL_RST_PCIE;
4135 		else
4136 			tr_method = MPI2_SCSITASKMGMT_MSGFLAGS_LINK_RESET;
4137 	}
4138 	if (sas_target_priv_data) {
4139 		dewtprintk(ioc,
4140 			   ioc_info(ioc, "setting delete flag: handle(0x%04x), sas_addr(0x%016llx)\n",
4141 				    handle, (u64)sas_address));
4142 		if (sas_device) {
4143 			if (sas_device->enclosure_handle != 0)
4144 				dewtprintk(ioc,
4145 					   ioc_info(ioc, "setting delete flag:enclosure logical id(0x%016llx), slot(%d)\n",
4146 						    (u64)sas_device->enclosure_logical_id,
4147 						    sas_device->slot));
4148 			if (sas_device->connector_name[0] != '\0')
4149 				dewtprintk(ioc,
4150 					   ioc_info(ioc, "setting delete flag: enclosure level(0x%04x), connector name( %s)\n",
4151 						    sas_device->enclosure_level,
4152 						    sas_device->connector_name));
4153 		} else if (pcie_device) {
4154 			if (pcie_device->enclosure_handle != 0)
4155 				dewtprintk(ioc,
4156 					   ioc_info(ioc, "setting delete flag: logical id(0x%016llx), slot(%d)\n",
4157 						    (u64)pcie_device->enclosure_logical_id,
4158 						    pcie_device->slot));
4159 			if (pcie_device->connector_name[0] != '\0')
4160 				dewtprintk(ioc,
4161 					   ioc_info(ioc, "setting delete flag:, enclosure level(0x%04x), connector name( %s)\n",
4162 						    pcie_device->enclosure_level,
4163 						    pcie_device->connector_name));
4164 		}
4165 		_scsih_ublock_io_device(ioc, sas_address, port);
4166 		sas_target_priv_data->handle = MPT3SAS_INVALID_DEVICE_HANDLE;
4167 	}
4168 
4169 	smid = mpt3sas_base_get_smid_hpr(ioc, ioc->tm_tr_cb_idx);
4170 	if (!smid) {
4171 		delayed_tr = kzalloc(sizeof(*delayed_tr), GFP_ATOMIC);
4172 		if (!delayed_tr)
4173 			goto out;
4174 		INIT_LIST_HEAD(&delayed_tr->list);
4175 		delayed_tr->handle = handle;
4176 		list_add_tail(&delayed_tr->list, &ioc->delayed_tr_list);
4177 		dewtprintk(ioc,
4178 			   ioc_info(ioc, "DELAYED:tr:handle(0x%04x), (open)\n",
4179 				    handle));
4180 		goto out;
4181 	}
4182 
4183 	dewtprintk(ioc,
4184 		   ioc_info(ioc, "tr_send:handle(0x%04x), (open), smid(%d), cb(%d)\n",
4185 			    handle, smid, ioc->tm_tr_cb_idx));
4186 	mpi_request = mpt3sas_base_get_msg_frame(ioc, smid);
4187 	memset(mpi_request, 0, sizeof(Mpi2SCSITaskManagementRequest_t));
4188 	mpi_request->Function = MPI2_FUNCTION_SCSI_TASK_MGMT;
4189 	mpi_request->DevHandle = cpu_to_le16(handle);
4190 	mpi_request->TaskType = MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET;
4191 	mpi_request->MsgFlags = tr_method;
4192 	set_bit(handle, ioc->device_remove_in_progress);
4193 	ioc->put_smid_hi_priority(ioc, smid, 0);
4194 	mpt3sas_trigger_master(ioc, MASTER_TRIGGER_DEVICE_REMOVAL);
4195 
4196 out:
4197 	if (sas_device)
4198 		sas_device_put(sas_device);
4199 	if (pcie_device)
4200 		pcie_device_put(pcie_device);
4201 }
4202 
4203 /**
4204  * _scsih_tm_tr_complete -
4205  * @ioc: per adapter object
4206  * @smid: system request message index
4207  * @msix_index: MSIX table index supplied by the OS
4208  * @reply: reply message frame(lower 32bit addr)
4209  * Context: interrupt time.
4210  *
4211  * This is the target reset completion routine.
4212  * This code is part of the code to initiate the device removal
4213  * handshake protocol with controller firmware.
4214  * It will send a sas iounit control request (MPI2_SAS_OP_REMOVE_DEVICE)
4215  *
4216  * Return: 1 meaning mf should be freed from _base_interrupt
4217  *         0 means the mf is freed from this function.
4218  */
4219 static u8
4220 _scsih_tm_tr_complete(struct MPT3SAS_ADAPTER *ioc, u16 smid, u8 msix_index,
4221 	u32 reply)
4222 {
4223 	u16 handle;
4224 	Mpi2SCSITaskManagementRequest_t *mpi_request_tm;
4225 	Mpi2SCSITaskManagementReply_t *mpi_reply =
4226 	    mpt3sas_base_get_reply_virt_addr(ioc, reply);
4227 	Mpi2SasIoUnitControlRequest_t *mpi_request;
4228 	u16 smid_sas_ctrl;
4229 	u32 ioc_state;
4230 	struct _sc_list *delayed_sc;
4231 
4232 	if (ioc->pci_error_recovery) {
4233 		dewtprintk(ioc,
4234 			   ioc_info(ioc, "%s: host in pci error recovery\n",
4235 				    __func__));
4236 		return 1;
4237 	}
4238 	ioc_state = mpt3sas_base_get_iocstate(ioc, 1);
4239 	if (ioc_state != MPI2_IOC_STATE_OPERATIONAL) {
4240 		dewtprintk(ioc,
4241 			   ioc_info(ioc, "%s: host is not operational\n",
4242 				    __func__));
4243 		return 1;
4244 	}
4245 	if (unlikely(!mpi_reply)) {
4246 		ioc_err(ioc, "mpi_reply not valid at %s:%d/%s()!\n",
4247 			__FILE__, __LINE__, __func__);
4248 		return 1;
4249 	}
4250 	mpi_request_tm = mpt3sas_base_get_msg_frame(ioc, smid);
4251 	handle = le16_to_cpu(mpi_request_tm->DevHandle);
4252 	if (handle != le16_to_cpu(mpi_reply->DevHandle)) {
4253 		dewtprintk(ioc,
4254 			   ioc_err(ioc, "spurious interrupt: handle(0x%04x:0x%04x), smid(%d)!!!\n",
4255 				   handle,
4256 				   le16_to_cpu(mpi_reply->DevHandle), smid));
4257 		return 0;
4258 	}
4259 
4260 	mpt3sas_trigger_master(ioc, MASTER_TRIGGER_TASK_MANAGMENT);
4261 	dewtprintk(ioc,
4262 		   ioc_info(ioc, "tr_complete:handle(0x%04x), (open) smid(%d), ioc_status(0x%04x), loginfo(0x%08x), completed(%d)\n",
4263 			    handle, smid, le16_to_cpu(mpi_reply->IOCStatus),
4264 			    le32_to_cpu(mpi_reply->IOCLogInfo),
4265 			    le32_to_cpu(mpi_reply->TerminationCount)));
4266 
4267 	smid_sas_ctrl = mpt3sas_base_get_smid(ioc, ioc->tm_sas_control_cb_idx);
4268 	if (!smid_sas_ctrl) {
4269 		delayed_sc = kzalloc(sizeof(*delayed_sc), GFP_ATOMIC);
4270 		if (!delayed_sc)
4271 			return _scsih_check_for_pending_tm(ioc, smid);
4272 		INIT_LIST_HEAD(&delayed_sc->list);
4273 		delayed_sc->handle = le16_to_cpu(mpi_request_tm->DevHandle);
4274 		list_add_tail(&delayed_sc->list, &ioc->delayed_sc_list);
4275 		dewtprintk(ioc,
4276 			   ioc_info(ioc, "DELAYED:sc:handle(0x%04x), (open)\n",
4277 				    handle));
4278 		return _scsih_check_for_pending_tm(ioc, smid);
4279 	}
4280 
4281 	dewtprintk(ioc,
4282 		   ioc_info(ioc, "sc_send:handle(0x%04x), (open), smid(%d), cb(%d)\n",
4283 			    handle, smid_sas_ctrl, ioc->tm_sas_control_cb_idx));
4284 	mpi_request = mpt3sas_base_get_msg_frame(ioc, smid_sas_ctrl);
4285 	memset(mpi_request, 0, sizeof(Mpi2SasIoUnitControlRequest_t));
4286 	mpi_request->Function = MPI2_FUNCTION_SAS_IO_UNIT_CONTROL;
4287 	mpi_request->Operation = MPI2_SAS_OP_REMOVE_DEVICE;
4288 	mpi_request->DevHandle = mpi_request_tm->DevHandle;
4289 	ioc->put_smid_default(ioc, smid_sas_ctrl);
4290 
4291 	return _scsih_check_for_pending_tm(ioc, smid);
4292 }
4293 
4294 /** _scsih_allow_scmd_to_device - check whether scmd needs to
4295  *				 issue to IOC or not.
4296  * @ioc: per adapter object
4297  * @scmd: pointer to scsi command object
4298  *
4299  * Returns true if scmd can be issued to IOC otherwise returns false.
4300  */
4301 inline bool _scsih_allow_scmd_to_device(struct MPT3SAS_ADAPTER *ioc,
4302 	struct scsi_cmnd *scmd)
4303 {
4304 
4305 	if (ioc->pci_error_recovery)
4306 		return false;
4307 
4308 	if (ioc->hba_mpi_version_belonged == MPI2_VERSION) {
4309 		if (ioc->remove_host)
4310 			return false;
4311 
4312 		return true;
4313 	}
4314 
4315 	if (ioc->remove_host) {
4316 
4317 		switch (scmd->cmnd[0]) {
4318 		case SYNCHRONIZE_CACHE:
4319 		case START_STOP:
4320 			return true;
4321 		default:
4322 			return false;
4323 		}
4324 	}
4325 
4326 	return true;
4327 }
4328 
4329 /**
4330  * _scsih_sas_control_complete - completion routine
4331  * @ioc: per adapter object
4332  * @smid: system request message index
4333  * @msix_index: MSIX table index supplied by the OS
4334  * @reply: reply message frame(lower 32bit addr)
4335  * Context: interrupt time.
4336  *
4337  * This is the sas iounit control completion routine.
4338  * This code is part of the code to initiate the device removal
4339  * handshake protocol with controller firmware.
4340  *
4341  * Return: 1 meaning mf should be freed from _base_interrupt
4342  *         0 means the mf is freed from this function.
4343  */
4344 static u8
4345 _scsih_sas_control_complete(struct MPT3SAS_ADAPTER *ioc, u16 smid,
4346 	u8 msix_index, u32 reply)
4347 {
4348 	Mpi2SasIoUnitControlReply_t *mpi_reply =
4349 	    mpt3sas_base_get_reply_virt_addr(ioc, reply);
4350 
4351 	if (likely(mpi_reply)) {
4352 		dewtprintk(ioc,
4353 			   ioc_info(ioc, "sc_complete:handle(0x%04x), (open) smid(%d), ioc_status(0x%04x), loginfo(0x%08x)\n",
4354 				    le16_to_cpu(mpi_reply->DevHandle), smid,
4355 				    le16_to_cpu(mpi_reply->IOCStatus),
4356 				    le32_to_cpu(mpi_reply->IOCLogInfo)));
4357 		if (le16_to_cpu(mpi_reply->IOCStatus) ==
4358 		     MPI2_IOCSTATUS_SUCCESS) {
4359 			clear_bit(le16_to_cpu(mpi_reply->DevHandle),
4360 			    ioc->device_remove_in_progress);
4361 		}
4362 	} else {
4363 		ioc_err(ioc, "mpi_reply not valid at %s:%d/%s()!\n",
4364 			__FILE__, __LINE__, __func__);
4365 	}
4366 	return mpt3sas_check_for_pending_internal_cmds(ioc, smid);
4367 }
4368 
4369 /**
4370  * _scsih_tm_tr_volume_send - send target reset request for volumes
4371  * @ioc: per adapter object
4372  * @handle: device handle
4373  * Context: interrupt time.
4374  *
4375  * This is designed to send muliple task management request at the same
4376  * time to the fifo. If the fifo is full, we will append the request,
4377  * and process it in a future completion.
4378  */
4379 static void
4380 _scsih_tm_tr_volume_send(struct MPT3SAS_ADAPTER *ioc, u16 handle)
4381 {
4382 	Mpi2SCSITaskManagementRequest_t *mpi_request;
4383 	u16 smid;
4384 	struct _tr_list *delayed_tr;
4385 
4386 	if (ioc->pci_error_recovery) {
4387 		dewtprintk(ioc,
4388 			   ioc_info(ioc, "%s: host reset in progress!\n",
4389 				    __func__));
4390 		return;
4391 	}
4392 
4393 	smid = mpt3sas_base_get_smid_hpr(ioc, ioc->tm_tr_volume_cb_idx);
4394 	if (!smid) {
4395 		delayed_tr = kzalloc(sizeof(*delayed_tr), GFP_ATOMIC);
4396 		if (!delayed_tr)
4397 			return;
4398 		INIT_LIST_HEAD(&delayed_tr->list);
4399 		delayed_tr->handle = handle;
4400 		list_add_tail(&delayed_tr->list, &ioc->delayed_tr_volume_list);
4401 		dewtprintk(ioc,
4402 			   ioc_info(ioc, "DELAYED:tr:handle(0x%04x), (open)\n",
4403 				    handle));
4404 		return;
4405 	}
4406 
4407 	dewtprintk(ioc,
4408 		   ioc_info(ioc, "tr_send:handle(0x%04x), (open), smid(%d), cb(%d)\n",
4409 			    handle, smid, ioc->tm_tr_volume_cb_idx));
4410 	mpi_request = mpt3sas_base_get_msg_frame(ioc, smid);
4411 	memset(mpi_request, 0, sizeof(Mpi2SCSITaskManagementRequest_t));
4412 	mpi_request->Function = MPI2_FUNCTION_SCSI_TASK_MGMT;
4413 	mpi_request->DevHandle = cpu_to_le16(handle);
4414 	mpi_request->TaskType = MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET;
4415 	ioc->put_smid_hi_priority(ioc, smid, 0);
4416 }
4417 
4418 /**
4419  * _scsih_tm_volume_tr_complete - target reset completion
4420  * @ioc: per adapter object
4421  * @smid: system request message index
4422  * @msix_index: MSIX table index supplied by the OS
4423  * @reply: reply message frame(lower 32bit addr)
4424  * Context: interrupt time.
4425  *
4426  * Return: 1 meaning mf should be freed from _base_interrupt
4427  *         0 means the mf is freed from this function.
4428  */
4429 static u8
4430 _scsih_tm_volume_tr_complete(struct MPT3SAS_ADAPTER *ioc, u16 smid,
4431 	u8 msix_index, u32 reply)
4432 {
4433 	u16 handle;
4434 	Mpi2SCSITaskManagementRequest_t *mpi_request_tm;
4435 	Mpi2SCSITaskManagementReply_t *mpi_reply =
4436 	    mpt3sas_base_get_reply_virt_addr(ioc, reply);
4437 
4438 	if (ioc->shost_recovery || ioc->pci_error_recovery) {
4439 		dewtprintk(ioc,
4440 			   ioc_info(ioc, "%s: host reset in progress!\n",
4441 				    __func__));
4442 		return 1;
4443 	}
4444 	if (unlikely(!mpi_reply)) {
4445 		ioc_err(ioc, "mpi_reply not valid at %s:%d/%s()!\n",
4446 			__FILE__, __LINE__, __func__);
4447 		return 1;
4448 	}
4449 
4450 	mpi_request_tm = mpt3sas_base_get_msg_frame(ioc, smid);
4451 	handle = le16_to_cpu(mpi_request_tm->DevHandle);
4452 	if (handle != le16_to_cpu(mpi_reply->DevHandle)) {
4453 		dewtprintk(ioc,
4454 			   ioc_err(ioc, "spurious interrupt: handle(0x%04x:0x%04x), smid(%d)!!!\n",
4455 				   handle, le16_to_cpu(mpi_reply->DevHandle),
4456 				   smid));
4457 		return 0;
4458 	}
4459 
4460 	dewtprintk(ioc,
4461 		   ioc_info(ioc, "tr_complete:handle(0x%04x), (open) smid(%d), ioc_status(0x%04x), loginfo(0x%08x), completed(%d)\n",
4462 			    handle, smid, le16_to_cpu(mpi_reply->IOCStatus),
4463 			    le32_to_cpu(mpi_reply->IOCLogInfo),
4464 			    le32_to_cpu(mpi_reply->TerminationCount)));
4465 
4466 	return _scsih_check_for_pending_tm(ioc, smid);
4467 }
4468 
4469 /**
4470  * _scsih_issue_delayed_event_ack - issue delayed Event ACK messages
4471  * @ioc: per adapter object
4472  * @smid: system request message index
4473  * @event: Event ID
4474  * @event_context: used to track events uniquely
4475  *
4476  * Context - processed in interrupt context.
4477  */
4478 static void
4479 _scsih_issue_delayed_event_ack(struct MPT3SAS_ADAPTER *ioc, u16 smid, U16 event,
4480 				U32 event_context)
4481 {
4482 	Mpi2EventAckRequest_t *ack_request;
4483 	int i = smid - ioc->internal_smid;
4484 	unsigned long flags;
4485 
4486 	/* Without releasing the smid just update the
4487 	 * call back index and reuse the same smid for
4488 	 * processing this delayed request
4489 	 */
4490 	spin_lock_irqsave(&ioc->scsi_lookup_lock, flags);
4491 	ioc->internal_lookup[i].cb_idx = ioc->base_cb_idx;
4492 	spin_unlock_irqrestore(&ioc->scsi_lookup_lock, flags);
4493 
4494 	dewtprintk(ioc,
4495 		   ioc_info(ioc, "EVENT ACK: event(0x%04x), smid(%d), cb(%d)\n",
4496 			    le16_to_cpu(event), smid, ioc->base_cb_idx));
4497 	ack_request = mpt3sas_base_get_msg_frame(ioc, smid);
4498 	memset(ack_request, 0, sizeof(Mpi2EventAckRequest_t));
4499 	ack_request->Function = MPI2_FUNCTION_EVENT_ACK;
4500 	ack_request->Event = event;
4501 	ack_request->EventContext = event_context;
4502 	ack_request->VF_ID = 0;  /* TODO */
4503 	ack_request->VP_ID = 0;
4504 	ioc->put_smid_default(ioc, smid);
4505 }
4506 
4507 /**
4508  * _scsih_issue_delayed_sas_io_unit_ctrl - issue delayed
4509  *				sas_io_unit_ctrl messages
4510  * @ioc: per adapter object
4511  * @smid: system request message index
4512  * @handle: device handle
4513  *
4514  * Context - processed in interrupt context.
4515  */
4516 static void
4517 _scsih_issue_delayed_sas_io_unit_ctrl(struct MPT3SAS_ADAPTER *ioc,
4518 					u16 smid, u16 handle)
4519 {
4520 	Mpi2SasIoUnitControlRequest_t *mpi_request;
4521 	u32 ioc_state;
4522 	int i = smid - ioc->internal_smid;
4523 	unsigned long flags;
4524 
4525 	if (ioc->remove_host) {
4526 		dewtprintk(ioc,
4527 			   ioc_info(ioc, "%s: host has been removed\n",
4528 				    __func__));
4529 		return;
4530 	} else if (ioc->pci_error_recovery) {
4531 		dewtprintk(ioc,
4532 			   ioc_info(ioc, "%s: host in pci error recovery\n",
4533 				    __func__));
4534 		return;
4535 	}
4536 	ioc_state = mpt3sas_base_get_iocstate(ioc, 1);
4537 	if (ioc_state != MPI2_IOC_STATE_OPERATIONAL) {
4538 		dewtprintk(ioc,
4539 			   ioc_info(ioc, "%s: host is not operational\n",
4540 				    __func__));
4541 		return;
4542 	}
4543 
4544 	/* Without releasing the smid just update the
4545 	 * call back index and reuse the same smid for
4546 	 * processing this delayed request
4547 	 */
4548 	spin_lock_irqsave(&ioc->scsi_lookup_lock, flags);
4549 	ioc->internal_lookup[i].cb_idx = ioc->tm_sas_control_cb_idx;
4550 	spin_unlock_irqrestore(&ioc->scsi_lookup_lock, flags);
4551 
4552 	dewtprintk(ioc,
4553 		   ioc_info(ioc, "sc_send:handle(0x%04x), (open), smid(%d), cb(%d)\n",
4554 			    handle, smid, ioc->tm_sas_control_cb_idx));
4555 	mpi_request = mpt3sas_base_get_msg_frame(ioc, smid);
4556 	memset(mpi_request, 0, sizeof(Mpi2SasIoUnitControlRequest_t));
4557 	mpi_request->Function = MPI2_FUNCTION_SAS_IO_UNIT_CONTROL;
4558 	mpi_request->Operation = MPI2_SAS_OP_REMOVE_DEVICE;
4559 	mpi_request->DevHandle = cpu_to_le16(handle);
4560 	ioc->put_smid_default(ioc, smid);
4561 }
4562 
4563 /**
4564  * mpt3sas_check_for_pending_internal_cmds - check for pending internal messages
4565  * @ioc: per adapter object
4566  * @smid: system request message index
4567  *
4568  * Context: Executed in interrupt context
4569  *
4570  * This will check delayed internal messages list, and process the
4571  * next request.
4572  *
4573  * Return: 1 meaning mf should be freed from _base_interrupt
4574  *         0 means the mf is freed from this function.
4575  */
4576 u8
4577 mpt3sas_check_for_pending_internal_cmds(struct MPT3SAS_ADAPTER *ioc, u16 smid)
4578 {
4579 	struct _sc_list *delayed_sc;
4580 	struct _event_ack_list *delayed_event_ack;
4581 
4582 	if (!list_empty(&ioc->delayed_event_ack_list)) {
4583 		delayed_event_ack = list_entry(ioc->delayed_event_ack_list.next,
4584 						struct _event_ack_list, list);
4585 		_scsih_issue_delayed_event_ack(ioc, smid,
4586 		  delayed_event_ack->Event, delayed_event_ack->EventContext);
4587 		list_del(&delayed_event_ack->list);
4588 		kfree(delayed_event_ack);
4589 		return 0;
4590 	}
4591 
4592 	if (!list_empty(&ioc->delayed_sc_list)) {
4593 		delayed_sc = list_entry(ioc->delayed_sc_list.next,
4594 						struct _sc_list, list);
4595 		_scsih_issue_delayed_sas_io_unit_ctrl(ioc, smid,
4596 						 delayed_sc->handle);
4597 		list_del(&delayed_sc->list);
4598 		kfree(delayed_sc);
4599 		return 0;
4600 	}
4601 	return 1;
4602 }
4603 
4604 /**
4605  * _scsih_check_for_pending_tm - check for pending task management
4606  * @ioc: per adapter object
4607  * @smid: system request message index
4608  *
4609  * This will check delayed target reset list, and feed the
4610  * next reqeust.
4611  *
4612  * Return: 1 meaning mf should be freed from _base_interrupt
4613  *         0 means the mf is freed from this function.
4614  */
4615 static u8
4616 _scsih_check_for_pending_tm(struct MPT3SAS_ADAPTER *ioc, u16 smid)
4617 {
4618 	struct _tr_list *delayed_tr;
4619 
4620 	if (!list_empty(&ioc->delayed_tr_volume_list)) {
4621 		delayed_tr = list_entry(ioc->delayed_tr_volume_list.next,
4622 		    struct _tr_list, list);
4623 		mpt3sas_base_free_smid(ioc, smid);
4624 		_scsih_tm_tr_volume_send(ioc, delayed_tr->handle);
4625 		list_del(&delayed_tr->list);
4626 		kfree(delayed_tr);
4627 		return 0;
4628 	}
4629 
4630 	if (!list_empty(&ioc->delayed_tr_list)) {
4631 		delayed_tr = list_entry(ioc->delayed_tr_list.next,
4632 		    struct _tr_list, list);
4633 		mpt3sas_base_free_smid(ioc, smid);
4634 		_scsih_tm_tr_send(ioc, delayed_tr->handle);
4635 		list_del(&delayed_tr->list);
4636 		kfree(delayed_tr);
4637 		return 0;
4638 	}
4639 
4640 	return 1;
4641 }
4642 
4643 /**
4644  * _scsih_check_topo_delete_events - sanity check on topo events
4645  * @ioc: per adapter object
4646  * @event_data: the event data payload
4647  *
4648  * This routine added to better handle cable breaker.
4649  *
4650  * This handles the case where driver receives multiple expander
4651  * add and delete events in a single shot.  When there is a delete event
4652  * the routine will void any pending add events waiting in the event queue.
4653  */
4654 static void
4655 _scsih_check_topo_delete_events(struct MPT3SAS_ADAPTER *ioc,
4656 	Mpi2EventDataSasTopologyChangeList_t *event_data)
4657 {
4658 	struct fw_event_work *fw_event;
4659 	Mpi2EventDataSasTopologyChangeList_t *local_event_data;
4660 	u16 expander_handle;
4661 	struct _sas_node *sas_expander;
4662 	unsigned long flags;
4663 	int i, reason_code;
4664 	u16 handle;
4665 
4666 	for (i = 0 ; i < event_data->NumEntries; i++) {
4667 		handle = le16_to_cpu(event_data->PHY[i].AttachedDevHandle);
4668 		if (!handle)
4669 			continue;
4670 		reason_code = event_data->PHY[i].PhyStatus &
4671 		    MPI2_EVENT_SAS_TOPO_RC_MASK;
4672 		if (reason_code == MPI2_EVENT_SAS_TOPO_RC_TARG_NOT_RESPONDING)
4673 			_scsih_tm_tr_send(ioc, handle);
4674 	}
4675 
4676 	expander_handle = le16_to_cpu(event_data->ExpanderDevHandle);
4677 	if (expander_handle < ioc->sas_hba.num_phys) {
4678 		_scsih_block_io_to_children_attached_directly(ioc, event_data);
4679 		return;
4680 	}
4681 	if (event_data->ExpStatus ==
4682 	    MPI2_EVENT_SAS_TOPO_ES_DELAY_NOT_RESPONDING) {
4683 		/* put expander attached devices into blocking state */
4684 		spin_lock_irqsave(&ioc->sas_node_lock, flags);
4685 		sas_expander = mpt3sas_scsih_expander_find_by_handle(ioc,
4686 		    expander_handle);
4687 		_scsih_block_io_to_children_attached_to_ex(ioc, sas_expander);
4688 		spin_unlock_irqrestore(&ioc->sas_node_lock, flags);
4689 		do {
4690 			handle = find_first_bit(ioc->blocking_handles,
4691 			    ioc->facts.MaxDevHandle);
4692 			if (handle < ioc->facts.MaxDevHandle)
4693 				_scsih_block_io_device(ioc, handle);
4694 		} while (test_and_clear_bit(handle, ioc->blocking_handles));
4695 	} else if (event_data->ExpStatus == MPI2_EVENT_SAS_TOPO_ES_RESPONDING)
4696 		_scsih_block_io_to_children_attached_directly(ioc, event_data);
4697 
4698 	if (event_data->ExpStatus != MPI2_EVENT_SAS_TOPO_ES_NOT_RESPONDING)
4699 		return;
4700 
4701 	/* mark ignore flag for pending events */
4702 	spin_lock_irqsave(&ioc->fw_event_lock, flags);
4703 	list_for_each_entry(fw_event, &ioc->fw_event_list, list) {
4704 		if (fw_event->event != MPI2_EVENT_SAS_TOPOLOGY_CHANGE_LIST ||
4705 		    fw_event->ignore)
4706 			continue;
4707 		local_event_data = (Mpi2EventDataSasTopologyChangeList_t *)
4708 				   fw_event->event_data;
4709 		if (local_event_data->ExpStatus ==
4710 		    MPI2_EVENT_SAS_TOPO_ES_ADDED ||
4711 		    local_event_data->ExpStatus ==
4712 		    MPI2_EVENT_SAS_TOPO_ES_RESPONDING) {
4713 			if (le16_to_cpu(local_event_data->ExpanderDevHandle) ==
4714 			    expander_handle) {
4715 				dewtprintk(ioc,
4716 					   ioc_info(ioc, "setting ignoring flag\n"));
4717 				fw_event->ignore = 1;
4718 			}
4719 		}
4720 	}
4721 	spin_unlock_irqrestore(&ioc->fw_event_lock, flags);
4722 }
4723 
4724 /**
4725  * _scsih_check_pcie_topo_remove_events - sanity check on topo
4726  * events
4727  * @ioc: per adapter object
4728  * @event_data: the event data payload
4729  *
4730  * This handles the case where driver receives multiple switch
4731  * or device add and delete events in a single shot.  When there
4732  * is a delete event the routine will void any pending add
4733  * events waiting in the event queue.
4734  */
4735 static void
4736 _scsih_check_pcie_topo_remove_events(struct MPT3SAS_ADAPTER *ioc,
4737 	Mpi26EventDataPCIeTopologyChangeList_t *event_data)
4738 {
4739 	struct fw_event_work *fw_event;
4740 	Mpi26EventDataPCIeTopologyChangeList_t *local_event_data;
4741 	unsigned long flags;
4742 	int i, reason_code;
4743 	u16 handle, switch_handle;
4744 
4745 	for (i = 0; i < event_data->NumEntries; i++) {
4746 		handle =
4747 			le16_to_cpu(event_data->PortEntry[i].AttachedDevHandle);
4748 		if (!handle)
4749 			continue;
4750 		reason_code = event_data->PortEntry[i].PortStatus;
4751 		if (reason_code == MPI26_EVENT_PCIE_TOPO_PS_NOT_RESPONDING)
4752 			_scsih_tm_tr_send(ioc, handle);
4753 	}
4754 
4755 	switch_handle = le16_to_cpu(event_data->SwitchDevHandle);
4756 	if (!switch_handle) {
4757 		_scsih_block_io_to_pcie_children_attached_directly(
4758 							ioc, event_data);
4759 		return;
4760 	}
4761     /* TODO We are not supporting cascaded PCIe Switch removal yet*/
4762 	if ((event_data->SwitchStatus
4763 		== MPI26_EVENT_PCIE_TOPO_SS_DELAY_NOT_RESPONDING) ||
4764 		(event_data->SwitchStatus ==
4765 					MPI26_EVENT_PCIE_TOPO_SS_RESPONDING))
4766 		_scsih_block_io_to_pcie_children_attached_directly(
4767 							ioc, event_data);
4768 
4769 	if (event_data->SwitchStatus != MPI2_EVENT_SAS_TOPO_ES_NOT_RESPONDING)
4770 		return;
4771 
4772 	/* mark ignore flag for pending events */
4773 	spin_lock_irqsave(&ioc->fw_event_lock, flags);
4774 	list_for_each_entry(fw_event, &ioc->fw_event_list, list) {
4775 		if (fw_event->event != MPI2_EVENT_PCIE_TOPOLOGY_CHANGE_LIST ||
4776 			fw_event->ignore)
4777 			continue;
4778 		local_event_data =
4779 			(Mpi26EventDataPCIeTopologyChangeList_t *)
4780 			fw_event->event_data;
4781 		if (local_event_data->SwitchStatus ==
4782 		    MPI2_EVENT_SAS_TOPO_ES_ADDED ||
4783 		    local_event_data->SwitchStatus ==
4784 		    MPI2_EVENT_SAS_TOPO_ES_RESPONDING) {
4785 			if (le16_to_cpu(local_event_data->SwitchDevHandle) ==
4786 				switch_handle) {
4787 				dewtprintk(ioc,
4788 					   ioc_info(ioc, "setting ignoring flag for switch event\n"));
4789 				fw_event->ignore = 1;
4790 			}
4791 		}
4792 	}
4793 	spin_unlock_irqrestore(&ioc->fw_event_lock, flags);
4794 }
4795 
4796 /**
4797  * _scsih_set_volume_delete_flag - setting volume delete flag
4798  * @ioc: per adapter object
4799  * @handle: device handle
4800  *
4801  * This returns nothing.
4802  */
4803 static void
4804 _scsih_set_volume_delete_flag(struct MPT3SAS_ADAPTER *ioc, u16 handle)
4805 {
4806 	struct _raid_device *raid_device;
4807 	struct MPT3SAS_TARGET *sas_target_priv_data;
4808 	unsigned long flags;
4809 
4810 	spin_lock_irqsave(&ioc->raid_device_lock, flags);
4811 	raid_device = mpt3sas_raid_device_find_by_handle(ioc, handle);
4812 	if (raid_device && raid_device->starget &&
4813 	    raid_device->starget->hostdata) {
4814 		sas_target_priv_data =
4815 		    raid_device->starget->hostdata;
4816 		sas_target_priv_data->deleted = 1;
4817 		dewtprintk(ioc,
4818 			   ioc_info(ioc, "setting delete flag: handle(0x%04x), wwid(0x%016llx)\n",
4819 				    handle, (u64)raid_device->wwid));
4820 	}
4821 	spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
4822 }
4823 
4824 /**
4825  * _scsih_set_volume_handle_for_tr - set handle for target reset to volume
4826  * @handle: input handle
4827  * @a: handle for volume a
4828  * @b: handle for volume b
4829  *
4830  * IR firmware only supports two raid volumes.  The purpose of this
4831  * routine is to set the volume handle in either a or b. When the given
4832  * input handle is non-zero, or when a and b have not been set before.
4833  */
4834 static void
4835 _scsih_set_volume_handle_for_tr(u16 handle, u16 *a, u16 *b)
4836 {
4837 	if (!handle || handle == *a || handle == *b)
4838 		return;
4839 	if (!*a)
4840 		*a = handle;
4841 	else if (!*b)
4842 		*b = handle;
4843 }
4844 
4845 /**
4846  * _scsih_check_ir_config_unhide_events - check for UNHIDE events
4847  * @ioc: per adapter object
4848  * @event_data: the event data payload
4849  * Context: interrupt time.
4850  *
4851  * This routine will send target reset to volume, followed by target
4852  * resets to the PDs. This is called when a PD has been removed, or
4853  * volume has been deleted or removed. When the target reset is sent
4854  * to volume, the PD target resets need to be queued to start upon
4855  * completion of the volume target reset.
4856  */
4857 static void
4858 _scsih_check_ir_config_unhide_events(struct MPT3SAS_ADAPTER *ioc,
4859 	Mpi2EventDataIrConfigChangeList_t *event_data)
4860 {
4861 	Mpi2EventIrConfigElement_t *element;
4862 	int i;
4863 	u16 handle, volume_handle, a, b;
4864 	struct _tr_list *delayed_tr;
4865 
4866 	a = 0;
4867 	b = 0;
4868 
4869 	if (ioc->is_warpdrive)
4870 		return;
4871 
4872 	/* Volume Resets for Deleted or Removed */
4873 	element = (Mpi2EventIrConfigElement_t *)&event_data->ConfigElement[0];
4874 	for (i = 0; i < event_data->NumElements; i++, element++) {
4875 		if (le32_to_cpu(event_data->Flags) &
4876 		    MPI2_EVENT_IR_CHANGE_FLAGS_FOREIGN_CONFIG)
4877 			continue;
4878 		if (element->ReasonCode ==
4879 		    MPI2_EVENT_IR_CHANGE_RC_VOLUME_DELETED ||
4880 		    element->ReasonCode ==
4881 		    MPI2_EVENT_IR_CHANGE_RC_REMOVED) {
4882 			volume_handle = le16_to_cpu(element->VolDevHandle);
4883 			_scsih_set_volume_delete_flag(ioc, volume_handle);
4884 			_scsih_set_volume_handle_for_tr(volume_handle, &a, &b);
4885 		}
4886 	}
4887 
4888 	/* Volume Resets for UNHIDE events */
4889 	element = (Mpi2EventIrConfigElement_t *)&event_data->ConfigElement[0];
4890 	for (i = 0; i < event_data->NumElements; i++, element++) {
4891 		if (le32_to_cpu(event_data->Flags) &
4892 		    MPI2_EVENT_IR_CHANGE_FLAGS_FOREIGN_CONFIG)
4893 			continue;
4894 		if (element->ReasonCode == MPI2_EVENT_IR_CHANGE_RC_UNHIDE) {
4895 			volume_handle = le16_to_cpu(element->VolDevHandle);
4896 			_scsih_set_volume_handle_for_tr(volume_handle, &a, &b);
4897 		}
4898 	}
4899 
4900 	if (a)
4901 		_scsih_tm_tr_volume_send(ioc, a);
4902 	if (b)
4903 		_scsih_tm_tr_volume_send(ioc, b);
4904 
4905 	/* PD target resets */
4906 	element = (Mpi2EventIrConfigElement_t *)&event_data->ConfigElement[0];
4907 	for (i = 0; i < event_data->NumElements; i++, element++) {
4908 		if (element->ReasonCode != MPI2_EVENT_IR_CHANGE_RC_UNHIDE)
4909 			continue;
4910 		handle = le16_to_cpu(element->PhysDiskDevHandle);
4911 		volume_handle = le16_to_cpu(element->VolDevHandle);
4912 		clear_bit(handle, ioc->pd_handles);
4913 		if (!volume_handle)
4914 			_scsih_tm_tr_send(ioc, handle);
4915 		else if (volume_handle == a || volume_handle == b) {
4916 			delayed_tr = kzalloc(sizeof(*delayed_tr), GFP_ATOMIC);
4917 			BUG_ON(!delayed_tr);
4918 			INIT_LIST_HEAD(&delayed_tr->list);
4919 			delayed_tr->handle = handle;
4920 			list_add_tail(&delayed_tr->list, &ioc->delayed_tr_list);
4921 			dewtprintk(ioc,
4922 				   ioc_info(ioc, "DELAYED:tr:handle(0x%04x), (open)\n",
4923 					    handle));
4924 		} else
4925 			_scsih_tm_tr_send(ioc, handle);
4926 	}
4927 }
4928 
4929 
4930 /**
4931  * _scsih_check_volume_delete_events - set delete flag for volumes
4932  * @ioc: per adapter object
4933  * @event_data: the event data payload
4934  * Context: interrupt time.
4935  *
4936  * This will handle the case when the cable connected to entire volume is
4937  * pulled. We will take care of setting the deleted flag so normal IO will
4938  * not be sent.
4939  */
4940 static void
4941 _scsih_check_volume_delete_events(struct MPT3SAS_ADAPTER *ioc,
4942 	Mpi2EventDataIrVolume_t *event_data)
4943 {
4944 	u32 state;
4945 
4946 	if (event_data->ReasonCode != MPI2_EVENT_IR_VOLUME_RC_STATE_CHANGED)
4947 		return;
4948 	state = le32_to_cpu(event_data->NewValue);
4949 	if (state == MPI2_RAID_VOL_STATE_MISSING || state ==
4950 	    MPI2_RAID_VOL_STATE_FAILED)
4951 		_scsih_set_volume_delete_flag(ioc,
4952 		    le16_to_cpu(event_data->VolDevHandle));
4953 }
4954 
4955 /**
4956  * _scsih_temp_threshold_events - display temperature threshold exceeded events
4957  * @ioc: per adapter object
4958  * @event_data: the temp threshold event data
4959  * Context: interrupt time.
4960  */
4961 static void
4962 _scsih_temp_threshold_events(struct MPT3SAS_ADAPTER *ioc,
4963 	Mpi2EventDataTemperature_t *event_data)
4964 {
4965 	u32 doorbell;
4966 	if (ioc->temp_sensors_count >= event_data->SensorNum) {
4967 		ioc_err(ioc, "Temperature Threshold flags %s%s%s%s exceeded for Sensor: %d !!!\n",
4968 			le16_to_cpu(event_data->Status) & 0x1 ? "0 " : " ",
4969 			le16_to_cpu(event_data->Status) & 0x2 ? "1 " : " ",
4970 			le16_to_cpu(event_data->Status) & 0x4 ? "2 " : " ",
4971 			le16_to_cpu(event_data->Status) & 0x8 ? "3 " : " ",
4972 			event_data->SensorNum);
4973 		ioc_err(ioc, "Current Temp In Celsius: %d\n",
4974 			event_data->CurrentTemperature);
4975 		if (ioc->hba_mpi_version_belonged != MPI2_VERSION) {
4976 			doorbell = mpt3sas_base_get_iocstate(ioc, 0);
4977 			if ((doorbell & MPI2_IOC_STATE_MASK) ==
4978 			    MPI2_IOC_STATE_FAULT) {
4979 				mpt3sas_print_fault_code(ioc,
4980 				    doorbell & MPI2_DOORBELL_DATA_MASK);
4981 			} else if ((doorbell & MPI2_IOC_STATE_MASK) ==
4982 			    MPI2_IOC_STATE_COREDUMP) {
4983 				mpt3sas_print_coredump_info(ioc,
4984 				    doorbell & MPI2_DOORBELL_DATA_MASK);
4985 			}
4986 		}
4987 	}
4988 }
4989 
4990 static int _scsih_set_satl_pending(struct scsi_cmnd *scmd, bool pending)
4991 {
4992 	struct MPT3SAS_DEVICE *priv = scmd->device->hostdata;
4993 
4994 	if (scmd->cmnd[0] != ATA_12 && scmd->cmnd[0] != ATA_16)
4995 		return 0;
4996 
4997 	if (pending)
4998 		return test_and_set_bit(0, &priv->ata_command_pending);
4999 
5000 	clear_bit(0, &priv->ata_command_pending);
5001 	return 0;
5002 }
5003 
5004 /**
5005  * _scsih_flush_running_cmds - completing outstanding commands.
5006  * @ioc: per adapter object
5007  *
5008  * The flushing out of all pending scmd commands following host reset,
5009  * where all IO is dropped to the floor.
5010  */
5011 static void
5012 _scsih_flush_running_cmds(struct MPT3SAS_ADAPTER *ioc)
5013 {
5014 	struct scsi_cmnd *scmd;
5015 	struct scsiio_tracker *st;
5016 	u16 smid;
5017 	int count = 0;
5018 
5019 	for (smid = 1; smid <= ioc->scsiio_depth; smid++) {
5020 		scmd = mpt3sas_scsih_scsi_lookup_get(ioc, smid);
5021 		if (!scmd)
5022 			continue;
5023 		count++;
5024 		_scsih_set_satl_pending(scmd, false);
5025 		st = scsi_cmd_priv(scmd);
5026 		mpt3sas_base_clear_st(ioc, st);
5027 		scsi_dma_unmap(scmd);
5028 		if (ioc->pci_error_recovery || ioc->remove_host)
5029 			scmd->result = DID_NO_CONNECT << 16;
5030 		else
5031 			scmd->result = DID_RESET << 16;
5032 		scsi_done(scmd);
5033 	}
5034 	dtmprintk(ioc, ioc_info(ioc, "completing %d cmds\n", count));
5035 }
5036 
5037 /**
5038  * _scsih_setup_eedp - setup MPI request for EEDP transfer
5039  * @ioc: per adapter object
5040  * @scmd: pointer to scsi command object
5041  * @mpi_request: pointer to the SCSI_IO request message frame
5042  *
5043  * Supporting protection 1 and 3.
5044  */
5045 static void
5046 _scsih_setup_eedp(struct MPT3SAS_ADAPTER *ioc, struct scsi_cmnd *scmd,
5047 	Mpi25SCSIIORequest_t *mpi_request)
5048 {
5049 	u16 eedp_flags;
5050 	Mpi25SCSIIORequest_t *mpi_request_3v =
5051 	   (Mpi25SCSIIORequest_t *)mpi_request;
5052 
5053 	switch (scsi_get_prot_op(scmd)) {
5054 	case SCSI_PROT_READ_STRIP:
5055 		eedp_flags = MPI2_SCSIIO_EEDPFLAGS_CHECK_REMOVE_OP;
5056 		break;
5057 	case SCSI_PROT_WRITE_INSERT:
5058 		eedp_flags = MPI2_SCSIIO_EEDPFLAGS_INSERT_OP;
5059 		break;
5060 	default:
5061 		return;
5062 	}
5063 
5064 	if (scmd->prot_flags & SCSI_PROT_GUARD_CHECK)
5065 		eedp_flags |= MPI2_SCSIIO_EEDPFLAGS_CHECK_GUARD;
5066 
5067 	if (scmd->prot_flags & SCSI_PROT_REF_CHECK)
5068 		eedp_flags |= MPI2_SCSIIO_EEDPFLAGS_CHECK_REFTAG;
5069 
5070 	if (scmd->prot_flags & SCSI_PROT_REF_INCREMENT) {
5071 		eedp_flags |= MPI2_SCSIIO_EEDPFLAGS_INC_PRI_REFTAG;
5072 
5073 		mpi_request->CDB.EEDP32.PrimaryReferenceTag =
5074 			cpu_to_be32(scsi_prot_ref_tag(scmd));
5075 	}
5076 
5077 	mpi_request_3v->EEDPBlockSize = cpu_to_le16(scsi_prot_interval(scmd));
5078 
5079 	if (ioc->is_gen35_ioc)
5080 		eedp_flags |= MPI25_SCSIIO_EEDPFLAGS_APPTAG_DISABLE_MODE;
5081 	mpi_request->EEDPFlags = cpu_to_le16(eedp_flags);
5082 }
5083 
5084 /**
5085  * _scsih_eedp_error_handling - return sense code for EEDP errors
5086  * @scmd: pointer to scsi command object
5087  * @ioc_status: ioc status
5088  */
5089 static void
5090 _scsih_eedp_error_handling(struct scsi_cmnd *scmd, u16 ioc_status)
5091 {
5092 	u8 ascq;
5093 
5094 	switch (ioc_status) {
5095 	case MPI2_IOCSTATUS_EEDP_GUARD_ERROR:
5096 		ascq = 0x01;
5097 		break;
5098 	case MPI2_IOCSTATUS_EEDP_APP_TAG_ERROR:
5099 		ascq = 0x02;
5100 		break;
5101 	case MPI2_IOCSTATUS_EEDP_REF_TAG_ERROR:
5102 		ascq = 0x03;
5103 		break;
5104 	default:
5105 		ascq = 0x00;
5106 		break;
5107 	}
5108 	scsi_build_sense(scmd, 0, ILLEGAL_REQUEST, 0x10, ascq);
5109 	set_host_byte(scmd, DID_ABORT);
5110 }
5111 
5112 /**
5113  * scsih_qcmd - main scsi request entry point
5114  * @shost: SCSI host pointer
5115  * @scmd: pointer to scsi command object
5116  *
5117  * The callback index is set inside `ioc->scsi_io_cb_idx`.
5118  *
5119  * Return: 0 on success.  If there's a failure, return either:
5120  * SCSI_MLQUEUE_DEVICE_BUSY if the device queue is full, or
5121  * SCSI_MLQUEUE_HOST_BUSY if the entire host queue is full
5122  */
5123 static int
5124 scsih_qcmd(struct Scsi_Host *shost, struct scsi_cmnd *scmd)
5125 {
5126 	struct MPT3SAS_ADAPTER *ioc = shost_priv(shost);
5127 	struct MPT3SAS_DEVICE *sas_device_priv_data;
5128 	struct MPT3SAS_TARGET *sas_target_priv_data;
5129 	struct _raid_device *raid_device;
5130 	struct request *rq = scsi_cmd_to_rq(scmd);
5131 	int class;
5132 	Mpi25SCSIIORequest_t *mpi_request;
5133 	struct _pcie_device *pcie_device = NULL;
5134 	u32 mpi_control;
5135 	u16 smid;
5136 	u16 handle;
5137 
5138 	if (ioc->logging_level & MPT_DEBUG_SCSI)
5139 		scsi_print_command(scmd);
5140 
5141 	sas_device_priv_data = scmd->device->hostdata;
5142 	if (!sas_device_priv_data || !sas_device_priv_data->sas_target) {
5143 		scmd->result = DID_NO_CONNECT << 16;
5144 		scsi_done(scmd);
5145 		return 0;
5146 	}
5147 
5148 	if (!(_scsih_allow_scmd_to_device(ioc, scmd))) {
5149 		scmd->result = DID_NO_CONNECT << 16;
5150 		scsi_done(scmd);
5151 		return 0;
5152 	}
5153 
5154 	sas_target_priv_data = sas_device_priv_data->sas_target;
5155 
5156 	/* invalid device handle */
5157 	handle = sas_target_priv_data->handle;
5158 
5159 	/*
5160 	 * Avoid error handling escallation when device is disconnected
5161 	 */
5162 	if (handle == MPT3SAS_INVALID_DEVICE_HANDLE || sas_device_priv_data->block) {
5163 		if (scmd->device->host->shost_state == SHOST_RECOVERY &&
5164 		    scmd->cmnd[0] == TEST_UNIT_READY) {
5165 			scsi_build_sense(scmd, 0, UNIT_ATTENTION, 0x29, 0x07);
5166 			scsi_done(scmd);
5167 			return 0;
5168 		}
5169 	}
5170 
5171 	if (handle == MPT3SAS_INVALID_DEVICE_HANDLE) {
5172 		scmd->result = DID_NO_CONNECT << 16;
5173 		scsi_done(scmd);
5174 		return 0;
5175 	}
5176 
5177 
5178 	if (ioc->shost_recovery || ioc->ioc_link_reset_in_progress) {
5179 		/* host recovery or link resets sent via IOCTLs */
5180 		return SCSI_MLQUEUE_HOST_BUSY;
5181 	} else if (sas_target_priv_data->deleted) {
5182 		/* device has been deleted */
5183 		scmd->result = DID_NO_CONNECT << 16;
5184 		scsi_done(scmd);
5185 		return 0;
5186 	} else if (sas_target_priv_data->tm_busy ||
5187 		   sas_device_priv_data->block) {
5188 		/* device busy with task management */
5189 		return SCSI_MLQUEUE_DEVICE_BUSY;
5190 	}
5191 
5192 	/*
5193 	 * Bug work around for firmware SATL handling.  The loop
5194 	 * is based on atomic operations and ensures consistency
5195 	 * since we're lockless at this point
5196 	 */
5197 	do {
5198 		if (test_bit(0, &sas_device_priv_data->ata_command_pending))
5199 			return SCSI_MLQUEUE_DEVICE_BUSY;
5200 	} while (_scsih_set_satl_pending(scmd, true));
5201 
5202 	if (scmd->sc_data_direction == DMA_FROM_DEVICE)
5203 		mpi_control = MPI2_SCSIIO_CONTROL_READ;
5204 	else if (scmd->sc_data_direction == DMA_TO_DEVICE)
5205 		mpi_control = MPI2_SCSIIO_CONTROL_WRITE;
5206 	else
5207 		mpi_control = MPI2_SCSIIO_CONTROL_NODATATRANSFER;
5208 
5209 	/* set tags */
5210 	mpi_control |= MPI2_SCSIIO_CONTROL_SIMPLEQ;
5211 	/* NCQ Prio supported, make sure control indicated high priority */
5212 	if (sas_device_priv_data->ncq_prio_enable) {
5213 		class = IOPRIO_PRIO_CLASS(req_get_ioprio(rq));
5214 		if (class == IOPRIO_CLASS_RT)
5215 			mpi_control |= 1 << MPI2_SCSIIO_CONTROL_CMDPRI_SHIFT;
5216 	}
5217 	/* Make sure Device is not raid volume.
5218 	 * We do not expose raid functionality to upper layer for warpdrive.
5219 	 */
5220 	if (((!ioc->is_warpdrive && !scsih_is_raid(&scmd->device->sdev_gendev))
5221 		&& !scsih_is_nvme(&scmd->device->sdev_gendev))
5222 		&& sas_is_tlr_enabled(scmd->device) && scmd->cmd_len != 32)
5223 		mpi_control |= MPI2_SCSIIO_CONTROL_TLR_ON;
5224 
5225 	smid = mpt3sas_base_get_smid_scsiio(ioc, ioc->scsi_io_cb_idx, scmd);
5226 	if (!smid) {
5227 		ioc_err(ioc, "%s: failed obtaining a smid\n", __func__);
5228 		_scsih_set_satl_pending(scmd, false);
5229 		goto out;
5230 	}
5231 	mpi_request = mpt3sas_base_get_msg_frame(ioc, smid);
5232 	memset(mpi_request, 0, ioc->request_sz);
5233 	_scsih_setup_eedp(ioc, scmd, mpi_request);
5234 
5235 	if (scmd->cmd_len == 32)
5236 		mpi_control |= 4 << MPI2_SCSIIO_CONTROL_ADDCDBLEN_SHIFT;
5237 	mpi_request->Function = MPI2_FUNCTION_SCSI_IO_REQUEST;
5238 	if (sas_device_priv_data->sas_target->flags &
5239 	    MPT_TARGET_FLAGS_RAID_COMPONENT)
5240 		mpi_request->Function = MPI2_FUNCTION_RAID_SCSI_IO_PASSTHROUGH;
5241 	else
5242 		mpi_request->Function = MPI2_FUNCTION_SCSI_IO_REQUEST;
5243 	mpi_request->DevHandle = cpu_to_le16(handle);
5244 	mpi_request->DataLength = cpu_to_le32(scsi_bufflen(scmd));
5245 	mpi_request->Control = cpu_to_le32(mpi_control);
5246 	mpi_request->IoFlags = cpu_to_le16(scmd->cmd_len);
5247 	mpi_request->MsgFlags = MPI2_SCSIIO_MSGFLAGS_SYSTEM_SENSE_ADDR;
5248 	mpi_request->SenseBufferLength = SCSI_SENSE_BUFFERSIZE;
5249 	mpi_request->SenseBufferLowAddress =
5250 	    mpt3sas_base_get_sense_buffer_dma(ioc, smid);
5251 	mpi_request->SGLOffset0 = offsetof(Mpi25SCSIIORequest_t, SGL) / 4;
5252 	int_to_scsilun(sas_device_priv_data->lun, (struct scsi_lun *)
5253 	    mpi_request->LUN);
5254 	memcpy(mpi_request->CDB.CDB32, scmd->cmnd, scmd->cmd_len);
5255 
5256 	if (mpi_request->DataLength) {
5257 		pcie_device = sas_target_priv_data->pcie_dev;
5258 		if (ioc->build_sg_scmd(ioc, scmd, smid, pcie_device)) {
5259 			mpt3sas_base_free_smid(ioc, smid);
5260 			_scsih_set_satl_pending(scmd, false);
5261 			goto out;
5262 		}
5263 	} else
5264 		ioc->build_zero_len_sge(ioc, &mpi_request->SGL);
5265 
5266 	raid_device = sas_target_priv_data->raid_device;
5267 	if (raid_device && raid_device->direct_io_enabled)
5268 		mpt3sas_setup_direct_io(ioc, scmd,
5269 			raid_device, mpi_request);
5270 
5271 	if (likely(mpi_request->Function == MPI2_FUNCTION_SCSI_IO_REQUEST)) {
5272 		if (sas_target_priv_data->flags & MPT_TARGET_FASTPATH_IO) {
5273 			mpi_request->IoFlags = cpu_to_le16(scmd->cmd_len |
5274 			    MPI25_SCSIIO_IOFLAGS_FAST_PATH);
5275 			ioc->put_smid_fast_path(ioc, smid, handle);
5276 		} else
5277 			ioc->put_smid_scsi_io(ioc, smid,
5278 			    le16_to_cpu(mpi_request->DevHandle));
5279 	} else
5280 		ioc->put_smid_default(ioc, smid);
5281 	return 0;
5282 
5283  out:
5284 	return SCSI_MLQUEUE_HOST_BUSY;
5285 }
5286 
5287 /**
5288  * _scsih_normalize_sense - normalize descriptor and fixed format sense data
5289  * @sense_buffer: sense data returned by target
5290  * @data: normalized skey/asc/ascq
5291  */
5292 static void
5293 _scsih_normalize_sense(char *sense_buffer, struct sense_info *data)
5294 {
5295 	if ((sense_buffer[0] & 0x7F) >= 0x72) {
5296 		/* descriptor format */
5297 		data->skey = sense_buffer[1] & 0x0F;
5298 		data->asc = sense_buffer[2];
5299 		data->ascq = sense_buffer[3];
5300 	} else {
5301 		/* fixed format */
5302 		data->skey = sense_buffer[2] & 0x0F;
5303 		data->asc = sense_buffer[12];
5304 		data->ascq = sense_buffer[13];
5305 	}
5306 }
5307 
5308 /**
5309  * _scsih_scsi_ioc_info - translated non-successful SCSI_IO request
5310  * @ioc: per adapter object
5311  * @scmd: pointer to scsi command object
5312  * @mpi_reply: reply mf payload returned from firmware
5313  * @smid: ?
5314  *
5315  * scsi_status - SCSI Status code returned from target device
5316  * scsi_state - state info associated with SCSI_IO determined by ioc
5317  * ioc_status - ioc supplied status info
5318  */
5319 static void
5320 _scsih_scsi_ioc_info(struct MPT3SAS_ADAPTER *ioc, struct scsi_cmnd *scmd,
5321 	Mpi2SCSIIOReply_t *mpi_reply, u16 smid)
5322 {
5323 	u32 response_info;
5324 	u8 *response_bytes;
5325 	u16 ioc_status = le16_to_cpu(mpi_reply->IOCStatus) &
5326 	    MPI2_IOCSTATUS_MASK;
5327 	u8 scsi_state = mpi_reply->SCSIState;
5328 	u8 scsi_status = mpi_reply->SCSIStatus;
5329 	char *desc_ioc_state = NULL;
5330 	char *desc_scsi_status = NULL;
5331 	char *desc_scsi_state = ioc->tmp_string;
5332 	u32 log_info = le32_to_cpu(mpi_reply->IOCLogInfo);
5333 	struct _sas_device *sas_device = NULL;
5334 	struct _pcie_device *pcie_device = NULL;
5335 	struct scsi_target *starget = scmd->device->sdev_target;
5336 	struct MPT3SAS_TARGET *priv_target = starget->hostdata;
5337 	char *device_str = NULL;
5338 
5339 	if (!priv_target)
5340 		return;
5341 	if (ioc->hide_ir_msg)
5342 		device_str = "WarpDrive";
5343 	else
5344 		device_str = "volume";
5345 
5346 	if (log_info == 0x31170000)
5347 		return;
5348 
5349 	switch (ioc_status) {
5350 	case MPI2_IOCSTATUS_SUCCESS:
5351 		desc_ioc_state = "success";
5352 		break;
5353 	case MPI2_IOCSTATUS_INVALID_FUNCTION:
5354 		desc_ioc_state = "invalid function";
5355 		break;
5356 	case MPI2_IOCSTATUS_SCSI_RECOVERED_ERROR:
5357 		desc_ioc_state = "scsi recovered error";
5358 		break;
5359 	case MPI2_IOCSTATUS_SCSI_INVALID_DEVHANDLE:
5360 		desc_ioc_state = "scsi invalid dev handle";
5361 		break;
5362 	case MPI2_IOCSTATUS_SCSI_DEVICE_NOT_THERE:
5363 		desc_ioc_state = "scsi device not there";
5364 		break;
5365 	case MPI2_IOCSTATUS_SCSI_DATA_OVERRUN:
5366 		desc_ioc_state = "scsi data overrun";
5367 		break;
5368 	case MPI2_IOCSTATUS_SCSI_DATA_UNDERRUN:
5369 		desc_ioc_state = "scsi data underrun";
5370 		break;
5371 	case MPI2_IOCSTATUS_SCSI_IO_DATA_ERROR:
5372 		desc_ioc_state = "scsi io data error";
5373 		break;
5374 	case MPI2_IOCSTATUS_SCSI_PROTOCOL_ERROR:
5375 		desc_ioc_state = "scsi protocol error";
5376 		break;
5377 	case MPI2_IOCSTATUS_SCSI_TASK_TERMINATED:
5378 		desc_ioc_state = "scsi task terminated";
5379 		break;
5380 	case MPI2_IOCSTATUS_SCSI_RESIDUAL_MISMATCH:
5381 		desc_ioc_state = "scsi residual mismatch";
5382 		break;
5383 	case MPI2_IOCSTATUS_SCSI_TASK_MGMT_FAILED:
5384 		desc_ioc_state = "scsi task mgmt failed";
5385 		break;
5386 	case MPI2_IOCSTATUS_SCSI_IOC_TERMINATED:
5387 		desc_ioc_state = "scsi ioc terminated";
5388 		break;
5389 	case MPI2_IOCSTATUS_SCSI_EXT_TERMINATED:
5390 		desc_ioc_state = "scsi ext terminated";
5391 		break;
5392 	case MPI2_IOCSTATUS_EEDP_GUARD_ERROR:
5393 		desc_ioc_state = "eedp guard error";
5394 		break;
5395 	case MPI2_IOCSTATUS_EEDP_REF_TAG_ERROR:
5396 		desc_ioc_state = "eedp ref tag error";
5397 		break;
5398 	case MPI2_IOCSTATUS_EEDP_APP_TAG_ERROR:
5399 		desc_ioc_state = "eedp app tag error";
5400 		break;
5401 	case MPI2_IOCSTATUS_INSUFFICIENT_POWER:
5402 		desc_ioc_state = "insufficient power";
5403 		break;
5404 	default:
5405 		desc_ioc_state = "unknown";
5406 		break;
5407 	}
5408 
5409 	switch (scsi_status) {
5410 	case MPI2_SCSI_STATUS_GOOD:
5411 		desc_scsi_status = "good";
5412 		break;
5413 	case MPI2_SCSI_STATUS_CHECK_CONDITION:
5414 		desc_scsi_status = "check condition";
5415 		break;
5416 	case MPI2_SCSI_STATUS_CONDITION_MET:
5417 		desc_scsi_status = "condition met";
5418 		break;
5419 	case MPI2_SCSI_STATUS_BUSY:
5420 		desc_scsi_status = "busy";
5421 		break;
5422 	case MPI2_SCSI_STATUS_INTERMEDIATE:
5423 		desc_scsi_status = "intermediate";
5424 		break;
5425 	case MPI2_SCSI_STATUS_INTERMEDIATE_CONDMET:
5426 		desc_scsi_status = "intermediate condmet";
5427 		break;
5428 	case MPI2_SCSI_STATUS_RESERVATION_CONFLICT:
5429 		desc_scsi_status = "reservation conflict";
5430 		break;
5431 	case MPI2_SCSI_STATUS_COMMAND_TERMINATED:
5432 		desc_scsi_status = "command terminated";
5433 		break;
5434 	case MPI2_SCSI_STATUS_TASK_SET_FULL:
5435 		desc_scsi_status = "task set full";
5436 		break;
5437 	case MPI2_SCSI_STATUS_ACA_ACTIVE:
5438 		desc_scsi_status = "aca active";
5439 		break;
5440 	case MPI2_SCSI_STATUS_TASK_ABORTED:
5441 		desc_scsi_status = "task aborted";
5442 		break;
5443 	default:
5444 		desc_scsi_status = "unknown";
5445 		break;
5446 	}
5447 
5448 	desc_scsi_state[0] = '\0';
5449 	if (!scsi_state)
5450 		desc_scsi_state = " ";
5451 	if (scsi_state & MPI2_SCSI_STATE_RESPONSE_INFO_VALID)
5452 		strcat(desc_scsi_state, "response info ");
5453 	if (scsi_state & MPI2_SCSI_STATE_TERMINATED)
5454 		strcat(desc_scsi_state, "state terminated ");
5455 	if (scsi_state & MPI2_SCSI_STATE_NO_SCSI_STATUS)
5456 		strcat(desc_scsi_state, "no status ");
5457 	if (scsi_state & MPI2_SCSI_STATE_AUTOSENSE_FAILED)
5458 		strcat(desc_scsi_state, "autosense failed ");
5459 	if (scsi_state & MPI2_SCSI_STATE_AUTOSENSE_VALID)
5460 		strcat(desc_scsi_state, "autosense valid ");
5461 
5462 	scsi_print_command(scmd);
5463 
5464 	if (priv_target->flags & MPT_TARGET_FLAGS_VOLUME) {
5465 		ioc_warn(ioc, "\t%s wwid(0x%016llx)\n",
5466 			 device_str, (u64)priv_target->sas_address);
5467 	} else if (priv_target->flags & MPT_TARGET_FLAGS_PCIE_DEVICE) {
5468 		pcie_device = mpt3sas_get_pdev_from_target(ioc, priv_target);
5469 		if (pcie_device) {
5470 			ioc_info(ioc, "\twwid(0x%016llx), port(%d)\n",
5471 				 (u64)pcie_device->wwid, pcie_device->port_num);
5472 			if (pcie_device->enclosure_handle != 0)
5473 				ioc_info(ioc, "\tenclosure logical id(0x%016llx), slot(%d)\n",
5474 					 (u64)pcie_device->enclosure_logical_id,
5475 					 pcie_device->slot);
5476 			if (pcie_device->connector_name[0])
5477 				ioc_info(ioc, "\tenclosure level(0x%04x), connector name( %s)\n",
5478 					 pcie_device->enclosure_level,
5479 					 pcie_device->connector_name);
5480 			pcie_device_put(pcie_device);
5481 		}
5482 	} else {
5483 		sas_device = mpt3sas_get_sdev_from_target(ioc, priv_target);
5484 		if (sas_device) {
5485 			ioc_warn(ioc, "\tsas_address(0x%016llx), phy(%d)\n",
5486 				 (u64)sas_device->sas_address, sas_device->phy);
5487 
5488 			_scsih_display_enclosure_chassis_info(ioc, sas_device,
5489 			    NULL, NULL);
5490 
5491 			sas_device_put(sas_device);
5492 		}
5493 	}
5494 
5495 	ioc_warn(ioc, "\thandle(0x%04x), ioc_status(%s)(0x%04x), smid(%d)\n",
5496 		 le16_to_cpu(mpi_reply->DevHandle),
5497 		 desc_ioc_state, ioc_status, smid);
5498 	ioc_warn(ioc, "\trequest_len(%d), underflow(%d), resid(%d)\n",
5499 		 scsi_bufflen(scmd), scmd->underflow, scsi_get_resid(scmd));
5500 	ioc_warn(ioc, "\ttag(%d), transfer_count(%d), sc->result(0x%08x)\n",
5501 		 le16_to_cpu(mpi_reply->TaskTag),
5502 		 le32_to_cpu(mpi_reply->TransferCount), scmd->result);
5503 	ioc_warn(ioc, "\tscsi_status(%s)(0x%02x), scsi_state(%s)(0x%02x)\n",
5504 		 desc_scsi_status, scsi_status, desc_scsi_state, scsi_state);
5505 
5506 	if (scsi_state & MPI2_SCSI_STATE_AUTOSENSE_VALID) {
5507 		struct sense_info data;
5508 		_scsih_normalize_sense(scmd->sense_buffer, &data);
5509 		ioc_warn(ioc, "\t[sense_key,asc,ascq]: [0x%02x,0x%02x,0x%02x], count(%d)\n",
5510 			 data.skey, data.asc, data.ascq,
5511 			 le32_to_cpu(mpi_reply->SenseCount));
5512 	}
5513 	if (scsi_state & MPI2_SCSI_STATE_RESPONSE_INFO_VALID) {
5514 		response_info = le32_to_cpu(mpi_reply->ResponseInfo);
5515 		response_bytes = (u8 *)&response_info;
5516 		_scsih_response_code(ioc, response_bytes[0]);
5517 	}
5518 }
5519 
5520 /**
5521  * _scsih_turn_on_pfa_led - illuminate PFA LED
5522  * @ioc: per adapter object
5523  * @handle: device handle
5524  * Context: process
5525  */
5526 static void
5527 _scsih_turn_on_pfa_led(struct MPT3SAS_ADAPTER *ioc, u16 handle)
5528 {
5529 	Mpi2SepReply_t mpi_reply;
5530 	Mpi2SepRequest_t mpi_request;
5531 	struct _sas_device *sas_device;
5532 
5533 	sas_device = mpt3sas_get_sdev_by_handle(ioc, handle);
5534 	if (!sas_device)
5535 		return;
5536 
5537 	memset(&mpi_request, 0, sizeof(Mpi2SepRequest_t));
5538 	mpi_request.Function = MPI2_FUNCTION_SCSI_ENCLOSURE_PROCESSOR;
5539 	mpi_request.Action = MPI2_SEP_REQ_ACTION_WRITE_STATUS;
5540 	mpi_request.SlotStatus =
5541 	    cpu_to_le32(MPI2_SEP_REQ_SLOTSTATUS_PREDICTED_FAULT);
5542 	mpi_request.DevHandle = cpu_to_le16(handle);
5543 	mpi_request.Flags = MPI2_SEP_REQ_FLAGS_DEVHANDLE_ADDRESS;
5544 	if ((mpt3sas_base_scsi_enclosure_processor(ioc, &mpi_reply,
5545 	    &mpi_request)) != 0) {
5546 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
5547 			__FILE__, __LINE__, __func__);
5548 		goto out;
5549 	}
5550 	sas_device->pfa_led_on = 1;
5551 
5552 	if (mpi_reply.IOCStatus || mpi_reply.IOCLogInfo) {
5553 		dewtprintk(ioc,
5554 			   ioc_info(ioc, "enclosure_processor: ioc_status (0x%04x), loginfo(0x%08x)\n",
5555 				    le16_to_cpu(mpi_reply.IOCStatus),
5556 				    le32_to_cpu(mpi_reply.IOCLogInfo)));
5557 		goto out;
5558 	}
5559 out:
5560 	sas_device_put(sas_device);
5561 }
5562 
5563 /**
5564  * _scsih_turn_off_pfa_led - turn off Fault LED
5565  * @ioc: per adapter object
5566  * @sas_device: sas device whose PFA LED has to turned off
5567  * Context: process
5568  */
5569 static void
5570 _scsih_turn_off_pfa_led(struct MPT3SAS_ADAPTER *ioc,
5571 	struct _sas_device *sas_device)
5572 {
5573 	Mpi2SepReply_t mpi_reply;
5574 	Mpi2SepRequest_t mpi_request;
5575 
5576 	memset(&mpi_request, 0, sizeof(Mpi2SepRequest_t));
5577 	mpi_request.Function = MPI2_FUNCTION_SCSI_ENCLOSURE_PROCESSOR;
5578 	mpi_request.Action = MPI2_SEP_REQ_ACTION_WRITE_STATUS;
5579 	mpi_request.SlotStatus = 0;
5580 	mpi_request.Slot = cpu_to_le16(sas_device->slot);
5581 	mpi_request.DevHandle = 0;
5582 	mpi_request.EnclosureHandle = cpu_to_le16(sas_device->enclosure_handle);
5583 	mpi_request.Flags = MPI2_SEP_REQ_FLAGS_ENCLOSURE_SLOT_ADDRESS;
5584 	if ((mpt3sas_base_scsi_enclosure_processor(ioc, &mpi_reply,
5585 		&mpi_request)) != 0) {
5586 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
5587 			__FILE__, __LINE__, __func__);
5588 		return;
5589 	}
5590 
5591 	if (mpi_reply.IOCStatus || mpi_reply.IOCLogInfo) {
5592 		dewtprintk(ioc,
5593 			   ioc_info(ioc, "enclosure_processor: ioc_status (0x%04x), loginfo(0x%08x)\n",
5594 				    le16_to_cpu(mpi_reply.IOCStatus),
5595 				    le32_to_cpu(mpi_reply.IOCLogInfo)));
5596 		return;
5597 	}
5598 }
5599 
5600 /**
5601  * _scsih_send_event_to_turn_on_pfa_led - fire delayed event
5602  * @ioc: per adapter object
5603  * @handle: device handle
5604  * Context: interrupt.
5605  */
5606 static void
5607 _scsih_send_event_to_turn_on_pfa_led(struct MPT3SAS_ADAPTER *ioc, u16 handle)
5608 {
5609 	struct fw_event_work *fw_event;
5610 
5611 	fw_event = alloc_fw_event_work(0);
5612 	if (!fw_event)
5613 		return;
5614 	fw_event->event = MPT3SAS_TURN_ON_PFA_LED;
5615 	fw_event->device_handle = handle;
5616 	fw_event->ioc = ioc;
5617 	_scsih_fw_event_add(ioc, fw_event);
5618 	fw_event_work_put(fw_event);
5619 }
5620 
5621 /**
5622  * _scsih_smart_predicted_fault - process smart errors
5623  * @ioc: per adapter object
5624  * @handle: device handle
5625  * Context: interrupt.
5626  */
5627 static void
5628 _scsih_smart_predicted_fault(struct MPT3SAS_ADAPTER *ioc, u16 handle)
5629 {
5630 	struct scsi_target *starget;
5631 	struct MPT3SAS_TARGET *sas_target_priv_data;
5632 	Mpi2EventNotificationReply_t *event_reply;
5633 	Mpi2EventDataSasDeviceStatusChange_t *event_data;
5634 	struct _sas_device *sas_device;
5635 	ssize_t sz;
5636 	unsigned long flags;
5637 
5638 	/* only handle non-raid devices */
5639 	spin_lock_irqsave(&ioc->sas_device_lock, flags);
5640 	sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle);
5641 	if (!sas_device)
5642 		goto out_unlock;
5643 
5644 	starget = sas_device->starget;
5645 	sas_target_priv_data = starget->hostdata;
5646 
5647 	if ((sas_target_priv_data->flags & MPT_TARGET_FLAGS_RAID_COMPONENT) ||
5648 	   ((sas_target_priv_data->flags & MPT_TARGET_FLAGS_VOLUME)))
5649 		goto out_unlock;
5650 
5651 	_scsih_display_enclosure_chassis_info(NULL, sas_device, NULL, starget);
5652 
5653 	spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
5654 
5655 	if (ioc->pdev->subsystem_vendor == PCI_VENDOR_ID_IBM)
5656 		_scsih_send_event_to_turn_on_pfa_led(ioc, handle);
5657 
5658 	/* insert into event log */
5659 	sz = offsetof(Mpi2EventNotificationReply_t, EventData) +
5660 	     sizeof(Mpi2EventDataSasDeviceStatusChange_t);
5661 	event_reply = kzalloc(sz, GFP_ATOMIC);
5662 	if (!event_reply) {
5663 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
5664 			__FILE__, __LINE__, __func__);
5665 		goto out;
5666 	}
5667 
5668 	event_reply->Function = MPI2_FUNCTION_EVENT_NOTIFICATION;
5669 	event_reply->Event =
5670 	    cpu_to_le16(MPI2_EVENT_SAS_DEVICE_STATUS_CHANGE);
5671 	event_reply->MsgLength = sz/4;
5672 	event_reply->EventDataLength =
5673 	    cpu_to_le16(sizeof(Mpi2EventDataSasDeviceStatusChange_t)/4);
5674 	event_data = (Mpi2EventDataSasDeviceStatusChange_t *)
5675 	    event_reply->EventData;
5676 	event_data->ReasonCode = MPI2_EVENT_SAS_DEV_STAT_RC_SMART_DATA;
5677 	event_data->ASC = 0x5D;
5678 	event_data->DevHandle = cpu_to_le16(handle);
5679 	event_data->SASAddress = cpu_to_le64(sas_target_priv_data->sas_address);
5680 	mpt3sas_ctl_add_to_event_log(ioc, event_reply);
5681 	kfree(event_reply);
5682 out:
5683 	if (sas_device)
5684 		sas_device_put(sas_device);
5685 	return;
5686 
5687 out_unlock:
5688 	spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
5689 	goto out;
5690 }
5691 
5692 /**
5693  * _scsih_io_done - scsi request callback
5694  * @ioc: per adapter object
5695  * @smid: system request message index
5696  * @msix_index: MSIX table index supplied by the OS
5697  * @reply: reply message frame(lower 32bit addr)
5698  *
5699  * Callback handler when using _scsih_qcmd.
5700  *
5701  * Return: 1 meaning mf should be freed from _base_interrupt
5702  *         0 means the mf is freed from this function.
5703  */
5704 static u8
5705 _scsih_io_done(struct MPT3SAS_ADAPTER *ioc, u16 smid, u8 msix_index, u32 reply)
5706 {
5707 	Mpi25SCSIIORequest_t *mpi_request;
5708 	Mpi2SCSIIOReply_t *mpi_reply;
5709 	struct scsi_cmnd *scmd;
5710 	struct scsiio_tracker *st;
5711 	u16 ioc_status;
5712 	u32 xfer_cnt;
5713 	u8 scsi_state;
5714 	u8 scsi_status;
5715 	u32 log_info;
5716 	struct MPT3SAS_DEVICE *sas_device_priv_data;
5717 	u32 response_code = 0;
5718 
5719 	mpi_reply = mpt3sas_base_get_reply_virt_addr(ioc, reply);
5720 
5721 	scmd = mpt3sas_scsih_scsi_lookup_get(ioc, smid);
5722 	if (scmd == NULL)
5723 		return 1;
5724 
5725 	_scsih_set_satl_pending(scmd, false);
5726 
5727 	mpi_request = mpt3sas_base_get_msg_frame(ioc, smid);
5728 
5729 	if (mpi_reply == NULL) {
5730 		scmd->result = DID_OK << 16;
5731 		goto out;
5732 	}
5733 
5734 	sas_device_priv_data = scmd->device->hostdata;
5735 	if (!sas_device_priv_data || !sas_device_priv_data->sas_target ||
5736 	     sas_device_priv_data->sas_target->deleted) {
5737 		scmd->result = DID_NO_CONNECT << 16;
5738 		goto out;
5739 	}
5740 	ioc_status = le16_to_cpu(mpi_reply->IOCStatus);
5741 
5742 	/*
5743 	 * WARPDRIVE: If direct_io is set then it is directIO,
5744 	 * the failed direct I/O should be redirected to volume
5745 	 */
5746 	st = scsi_cmd_priv(scmd);
5747 	if (st->direct_io &&
5748 	     ((ioc_status & MPI2_IOCSTATUS_MASK)
5749 	      != MPI2_IOCSTATUS_SCSI_TASK_TERMINATED)) {
5750 		st->direct_io = 0;
5751 		st->scmd = scmd;
5752 		memcpy(mpi_request->CDB.CDB32, scmd->cmnd, scmd->cmd_len);
5753 		mpi_request->DevHandle =
5754 		    cpu_to_le16(sas_device_priv_data->sas_target->handle);
5755 		ioc->put_smid_scsi_io(ioc, smid,
5756 		    sas_device_priv_data->sas_target->handle);
5757 		return 0;
5758 	}
5759 	/* turning off TLR */
5760 	scsi_state = mpi_reply->SCSIState;
5761 	if (scsi_state & MPI2_SCSI_STATE_RESPONSE_INFO_VALID)
5762 		response_code =
5763 		    le32_to_cpu(mpi_reply->ResponseInfo) & 0xFF;
5764 	if (!sas_device_priv_data->tlr_snoop_check) {
5765 		sas_device_priv_data->tlr_snoop_check++;
5766 		if ((!ioc->is_warpdrive &&
5767 		    !scsih_is_raid(&scmd->device->sdev_gendev) &&
5768 		    !scsih_is_nvme(&scmd->device->sdev_gendev))
5769 		    && sas_is_tlr_enabled(scmd->device) &&
5770 		    response_code == MPI2_SCSITASKMGMT_RSP_INVALID_FRAME) {
5771 			sas_disable_tlr(scmd->device);
5772 			sdev_printk(KERN_INFO, scmd->device, "TLR disabled\n");
5773 		}
5774 	}
5775 
5776 	xfer_cnt = le32_to_cpu(mpi_reply->TransferCount);
5777 	scsi_set_resid(scmd, scsi_bufflen(scmd) - xfer_cnt);
5778 	if (ioc_status & MPI2_IOCSTATUS_FLAG_LOG_INFO_AVAILABLE)
5779 		log_info =  le32_to_cpu(mpi_reply->IOCLogInfo);
5780 	else
5781 		log_info = 0;
5782 	ioc_status &= MPI2_IOCSTATUS_MASK;
5783 	scsi_status = mpi_reply->SCSIStatus;
5784 
5785 	if (ioc_status == MPI2_IOCSTATUS_SCSI_DATA_UNDERRUN && xfer_cnt == 0 &&
5786 	    (scsi_status == MPI2_SCSI_STATUS_BUSY ||
5787 	     scsi_status == MPI2_SCSI_STATUS_RESERVATION_CONFLICT ||
5788 	     scsi_status == MPI2_SCSI_STATUS_TASK_SET_FULL)) {
5789 		ioc_status = MPI2_IOCSTATUS_SUCCESS;
5790 	}
5791 
5792 	if (scsi_state & MPI2_SCSI_STATE_AUTOSENSE_VALID) {
5793 		struct sense_info data;
5794 		const void *sense_data = mpt3sas_base_get_sense_buffer(ioc,
5795 		    smid);
5796 		u32 sz = min_t(u32, SCSI_SENSE_BUFFERSIZE,
5797 		    le32_to_cpu(mpi_reply->SenseCount));
5798 		memcpy(scmd->sense_buffer, sense_data, sz);
5799 		_scsih_normalize_sense(scmd->sense_buffer, &data);
5800 		/* failure prediction threshold exceeded */
5801 		if (data.asc == 0x5D)
5802 			_scsih_smart_predicted_fault(ioc,
5803 			    le16_to_cpu(mpi_reply->DevHandle));
5804 		mpt3sas_trigger_scsi(ioc, data.skey, data.asc, data.ascq);
5805 
5806 		if ((ioc->logging_level & MPT_DEBUG_REPLY) &&
5807 		     ((scmd->sense_buffer[2] == UNIT_ATTENTION) ||
5808 		     (scmd->sense_buffer[2] == MEDIUM_ERROR) ||
5809 		     (scmd->sense_buffer[2] == HARDWARE_ERROR)))
5810 			_scsih_scsi_ioc_info(ioc, scmd, mpi_reply, smid);
5811 	}
5812 	switch (ioc_status) {
5813 	case MPI2_IOCSTATUS_BUSY:
5814 	case MPI2_IOCSTATUS_INSUFFICIENT_RESOURCES:
5815 		scmd->result = SAM_STAT_BUSY;
5816 		break;
5817 
5818 	case MPI2_IOCSTATUS_SCSI_DEVICE_NOT_THERE:
5819 		scmd->result = DID_NO_CONNECT << 16;
5820 		break;
5821 
5822 	case MPI2_IOCSTATUS_SCSI_IOC_TERMINATED:
5823 		if (sas_device_priv_data->block) {
5824 			scmd->result = DID_TRANSPORT_DISRUPTED << 16;
5825 			goto out;
5826 		}
5827 		if (log_info == 0x31110630) {
5828 			if (scmd->retries > 2) {
5829 				scmd->result = DID_NO_CONNECT << 16;
5830 				scsi_device_set_state(scmd->device,
5831 				    SDEV_OFFLINE);
5832 			} else {
5833 				scmd->result = DID_SOFT_ERROR << 16;
5834 				scmd->device->expecting_cc_ua = 1;
5835 			}
5836 			break;
5837 		} else if (log_info == VIRTUAL_IO_FAILED_RETRY) {
5838 			scmd->result = DID_RESET << 16;
5839 			break;
5840 		} else if ((scmd->device->channel == RAID_CHANNEL) &&
5841 		   (scsi_state == (MPI2_SCSI_STATE_TERMINATED |
5842 		   MPI2_SCSI_STATE_NO_SCSI_STATUS))) {
5843 			scmd->result = DID_RESET << 16;
5844 			break;
5845 		}
5846 		scmd->result = DID_SOFT_ERROR << 16;
5847 		break;
5848 	case MPI2_IOCSTATUS_SCSI_TASK_TERMINATED:
5849 	case MPI2_IOCSTATUS_SCSI_EXT_TERMINATED:
5850 		scmd->result = DID_RESET << 16;
5851 		break;
5852 
5853 	case MPI2_IOCSTATUS_SCSI_RESIDUAL_MISMATCH:
5854 		if ((xfer_cnt == 0) || (scmd->underflow > xfer_cnt))
5855 			scmd->result = DID_SOFT_ERROR << 16;
5856 		else
5857 			scmd->result = (DID_OK << 16) | scsi_status;
5858 		break;
5859 
5860 	case MPI2_IOCSTATUS_SCSI_DATA_UNDERRUN:
5861 		scmd->result = (DID_OK << 16) | scsi_status;
5862 
5863 		if ((scsi_state & MPI2_SCSI_STATE_AUTOSENSE_VALID))
5864 			break;
5865 
5866 		if (xfer_cnt < scmd->underflow) {
5867 			if (scsi_status == SAM_STAT_BUSY)
5868 				scmd->result = SAM_STAT_BUSY;
5869 			else
5870 				scmd->result = DID_SOFT_ERROR << 16;
5871 		} else if (scsi_state & (MPI2_SCSI_STATE_AUTOSENSE_FAILED |
5872 		     MPI2_SCSI_STATE_NO_SCSI_STATUS))
5873 			scmd->result = DID_SOFT_ERROR << 16;
5874 		else if (scsi_state & MPI2_SCSI_STATE_TERMINATED)
5875 			scmd->result = DID_RESET << 16;
5876 		else if (!xfer_cnt && scmd->cmnd[0] == REPORT_LUNS) {
5877 			mpi_reply->SCSIState = MPI2_SCSI_STATE_AUTOSENSE_VALID;
5878 			mpi_reply->SCSIStatus = SAM_STAT_CHECK_CONDITION;
5879 			scsi_build_sense(scmd, 0, ILLEGAL_REQUEST,
5880 					 0x20, 0);
5881 		}
5882 		break;
5883 
5884 	case MPI2_IOCSTATUS_SCSI_DATA_OVERRUN:
5885 		scsi_set_resid(scmd, 0);
5886 		fallthrough;
5887 	case MPI2_IOCSTATUS_SCSI_RECOVERED_ERROR:
5888 	case MPI2_IOCSTATUS_SUCCESS:
5889 		scmd->result = (DID_OK << 16) | scsi_status;
5890 		if (response_code ==
5891 		    MPI2_SCSITASKMGMT_RSP_INVALID_FRAME ||
5892 		    (scsi_state & (MPI2_SCSI_STATE_AUTOSENSE_FAILED |
5893 		     MPI2_SCSI_STATE_NO_SCSI_STATUS)))
5894 			scmd->result = DID_SOFT_ERROR << 16;
5895 		else if (scsi_state & MPI2_SCSI_STATE_TERMINATED)
5896 			scmd->result = DID_RESET << 16;
5897 		break;
5898 
5899 	case MPI2_IOCSTATUS_EEDP_GUARD_ERROR:
5900 	case MPI2_IOCSTATUS_EEDP_REF_TAG_ERROR:
5901 	case MPI2_IOCSTATUS_EEDP_APP_TAG_ERROR:
5902 		_scsih_eedp_error_handling(scmd, ioc_status);
5903 		break;
5904 
5905 	case MPI2_IOCSTATUS_SCSI_PROTOCOL_ERROR:
5906 	case MPI2_IOCSTATUS_INVALID_FUNCTION:
5907 	case MPI2_IOCSTATUS_INVALID_SGL:
5908 	case MPI2_IOCSTATUS_INTERNAL_ERROR:
5909 	case MPI2_IOCSTATUS_INVALID_FIELD:
5910 	case MPI2_IOCSTATUS_INVALID_STATE:
5911 	case MPI2_IOCSTATUS_SCSI_IO_DATA_ERROR:
5912 	case MPI2_IOCSTATUS_SCSI_TASK_MGMT_FAILED:
5913 	case MPI2_IOCSTATUS_INSUFFICIENT_POWER:
5914 	default:
5915 		scmd->result = DID_SOFT_ERROR << 16;
5916 		break;
5917 
5918 	}
5919 
5920 	if (scmd->result && (ioc->logging_level & MPT_DEBUG_REPLY))
5921 		_scsih_scsi_ioc_info(ioc , scmd, mpi_reply, smid);
5922 
5923  out:
5924 
5925 	scsi_dma_unmap(scmd);
5926 	mpt3sas_base_free_smid(ioc, smid);
5927 	scsi_done(scmd);
5928 	return 0;
5929 }
5930 
5931 /**
5932  * _scsih_update_vphys_after_reset - update the Port's
5933  *			vphys_list after reset
5934  * @ioc: per adapter object
5935  *
5936  * Returns nothing.
5937  */
5938 static void
5939 _scsih_update_vphys_after_reset(struct MPT3SAS_ADAPTER *ioc)
5940 {
5941 	u16 sz, ioc_status;
5942 	int i;
5943 	Mpi2ConfigReply_t mpi_reply;
5944 	Mpi2SasIOUnitPage0_t *sas_iounit_pg0 = NULL;
5945 	u16 attached_handle;
5946 	u64 attached_sas_addr;
5947 	u8 found = 0, port_id;
5948 	Mpi2SasPhyPage0_t phy_pg0;
5949 	struct hba_port *port, *port_next, *mport;
5950 	struct virtual_phy *vphy, *vphy_next;
5951 	struct _sas_device *sas_device;
5952 
5953 	/*
5954 	 * Mark all the vphys objects as dirty.
5955 	 */
5956 	list_for_each_entry_safe(port, port_next,
5957 	    &ioc->port_table_list, list) {
5958 		if (!port->vphys_mask)
5959 			continue;
5960 		list_for_each_entry_safe(vphy, vphy_next,
5961 		    &port->vphys_list, list) {
5962 			vphy->flags |= MPT_VPHY_FLAG_DIRTY_PHY;
5963 		}
5964 	}
5965 
5966 	/*
5967 	 * Read SASIOUnitPage0 to get each HBA Phy's data.
5968 	 */
5969 	sz = offsetof(Mpi2SasIOUnitPage0_t, PhyData) +
5970 	    (ioc->sas_hba.num_phys * sizeof(Mpi2SasIOUnit0PhyData_t));
5971 	sas_iounit_pg0 = kzalloc(sz, GFP_KERNEL);
5972 	if (!sas_iounit_pg0) {
5973 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
5974 		    __FILE__, __LINE__, __func__);
5975 		return;
5976 	}
5977 	if ((mpt3sas_config_get_sas_iounit_pg0(ioc, &mpi_reply,
5978 	    sas_iounit_pg0, sz)) != 0)
5979 		goto out;
5980 	ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK;
5981 	if (ioc_status != MPI2_IOCSTATUS_SUCCESS)
5982 		goto out;
5983 	/*
5984 	 * Loop over each HBA Phy.
5985 	 */
5986 	for (i = 0; i < ioc->sas_hba.num_phys; i++) {
5987 		/*
5988 		 * Check whether Phy's Negotiation Link Rate is > 1.5G or not.
5989 		 */
5990 		if ((sas_iounit_pg0->PhyData[i].NegotiatedLinkRate >> 4) <
5991 		    MPI2_SAS_NEG_LINK_RATE_1_5)
5992 			continue;
5993 		/*
5994 		 * Check whether Phy is connected to SEP device or not,
5995 		 * if it is SEP device then read the Phy's SASPHYPage0 data to
5996 		 * determine whether Phy is a virtual Phy or not. if it is
5997 		 * virtual phy then it is conformed that the attached remote
5998 		 * device is a HBA's vSES device.
5999 		 */
6000 		if (!(le32_to_cpu(
6001 		    sas_iounit_pg0->PhyData[i].ControllerPhyDeviceInfo) &
6002 		    MPI2_SAS_DEVICE_INFO_SEP))
6003 			continue;
6004 
6005 		if ((mpt3sas_config_get_phy_pg0(ioc, &mpi_reply, &phy_pg0,
6006 		    i))) {
6007 			ioc_err(ioc, "failure at %s:%d/%s()!\n",
6008 			    __FILE__, __LINE__, __func__);
6009 			continue;
6010 		}
6011 
6012 		if (!(le32_to_cpu(phy_pg0.PhyInfo) &
6013 		    MPI2_SAS_PHYINFO_VIRTUAL_PHY))
6014 			continue;
6015 		/*
6016 		 * Get the vSES device's SAS Address.
6017 		 */
6018 		attached_handle = le16_to_cpu(
6019 		    sas_iounit_pg0->PhyData[i].AttachedDevHandle);
6020 		if (_scsih_get_sas_address(ioc, attached_handle,
6021 		    &attached_sas_addr) != 0) {
6022 			ioc_err(ioc, "failure at %s:%d/%s()!\n",
6023 			    __FILE__, __LINE__, __func__);
6024 			continue;
6025 		}
6026 
6027 		found = 0;
6028 		port = port_next = NULL;
6029 		/*
6030 		 * Loop over each virtual_phy object from
6031 		 * each port's vphys_list.
6032 		 */
6033 		list_for_each_entry_safe(port,
6034 		    port_next, &ioc->port_table_list, list) {
6035 			if (!port->vphys_mask)
6036 				continue;
6037 			list_for_each_entry_safe(vphy, vphy_next,
6038 			    &port->vphys_list, list) {
6039 				/*
6040 				 * Continue with next virtual_phy object
6041 				 * if the object is not marked as dirty.
6042 				 */
6043 				if (!(vphy->flags & MPT_VPHY_FLAG_DIRTY_PHY))
6044 					continue;
6045 
6046 				/*
6047 				 * Continue with next virtual_phy object
6048 				 * if the object's SAS Address is not equals
6049 				 * to current Phy's vSES device SAS Address.
6050 				 */
6051 				if (vphy->sas_address != attached_sas_addr)
6052 					continue;
6053 				/*
6054 				 * Enable current Phy number bit in object's
6055 				 * phy_mask field.
6056 				 */
6057 				if (!(vphy->phy_mask & (1 << i)))
6058 					vphy->phy_mask = (1 << i);
6059 				/*
6060 				 * Get hba_port object from hba_port table
6061 				 * corresponding to current phy's Port ID.
6062 				 * if there is no hba_port object corresponding
6063 				 * to Phy's Port ID then create a new hba_port
6064 				 * object & add to hba_port table.
6065 				 */
6066 				port_id = sas_iounit_pg0->PhyData[i].Port;
6067 				mport = mpt3sas_get_port_by_id(ioc, port_id, 1);
6068 				if (!mport) {
6069 					mport = kzalloc(
6070 					    sizeof(struct hba_port), GFP_KERNEL);
6071 					if (!mport)
6072 						break;
6073 					mport->port_id = port_id;
6074 					ioc_info(ioc,
6075 					    "%s: hba_port entry: %p, port: %d is added to hba_port list\n",
6076 					    __func__, mport, mport->port_id);
6077 					list_add_tail(&mport->list,
6078 						&ioc->port_table_list);
6079 				}
6080 				/*
6081 				 * If mport & port pointers are not pointing to
6082 				 * same hba_port object then it means that vSES
6083 				 * device's Port ID got changed after reset and
6084 				 * hence move current virtual_phy object from
6085 				 * port's vphys_list to mport's vphys_list.
6086 				 */
6087 				if (port != mport) {
6088 					if (!mport->vphys_mask)
6089 						INIT_LIST_HEAD(
6090 						    &mport->vphys_list);
6091 					mport->vphys_mask |= (1 << i);
6092 					port->vphys_mask &= ~(1 << i);
6093 					list_move(&vphy->list,
6094 					    &mport->vphys_list);
6095 					sas_device = mpt3sas_get_sdev_by_addr(
6096 					    ioc, attached_sas_addr, port);
6097 					if (sas_device)
6098 						sas_device->port = mport;
6099 				}
6100 				/*
6101 				 * Earlier while updating the hba_port table,
6102 				 * it is determined that there is no other
6103 				 * direct attached device with mport's Port ID,
6104 				 * Hence mport was marked as dirty. Only vSES
6105 				 * device has this Port ID, so unmark the mport
6106 				 * as dirt.
6107 				 */
6108 				if (mport->flags & HBA_PORT_FLAG_DIRTY_PORT) {
6109 					mport->sas_address = 0;
6110 					mport->phy_mask = 0;
6111 					mport->flags &=
6112 					    ~HBA_PORT_FLAG_DIRTY_PORT;
6113 				}
6114 				/*
6115 				 * Unmark current virtual_phy object as dirty.
6116 				 */
6117 				vphy->flags &= ~MPT_VPHY_FLAG_DIRTY_PHY;
6118 				found = 1;
6119 				break;
6120 			}
6121 			if (found)
6122 				break;
6123 		}
6124 	}
6125 out:
6126 	kfree(sas_iounit_pg0);
6127 }
6128 
6129 /**
6130  * _scsih_get_port_table_after_reset - Construct temporary port table
6131  * @ioc: per adapter object
6132  * @port_table: address where port table needs to be constructed
6133  *
6134  * return number of HBA port entries available after reset.
6135  */
6136 static int
6137 _scsih_get_port_table_after_reset(struct MPT3SAS_ADAPTER *ioc,
6138 	struct hba_port *port_table)
6139 {
6140 	u16 sz, ioc_status;
6141 	int i, j;
6142 	Mpi2ConfigReply_t mpi_reply;
6143 	Mpi2SasIOUnitPage0_t *sas_iounit_pg0 = NULL;
6144 	u16 attached_handle;
6145 	u64 attached_sas_addr;
6146 	u8 found = 0, port_count = 0, port_id;
6147 
6148 	sz = offsetof(Mpi2SasIOUnitPage0_t, PhyData) + (ioc->sas_hba.num_phys
6149 	    * sizeof(Mpi2SasIOUnit0PhyData_t));
6150 	sas_iounit_pg0 = kzalloc(sz, GFP_KERNEL);
6151 	if (!sas_iounit_pg0) {
6152 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
6153 		    __FILE__, __LINE__, __func__);
6154 		return port_count;
6155 	}
6156 
6157 	if ((mpt3sas_config_get_sas_iounit_pg0(ioc, &mpi_reply,
6158 	    sas_iounit_pg0, sz)) != 0)
6159 		goto out;
6160 	ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK;
6161 	if (ioc_status != MPI2_IOCSTATUS_SUCCESS)
6162 		goto out;
6163 	for (i = 0; i < ioc->sas_hba.num_phys; i++) {
6164 		found = 0;
6165 		if ((sas_iounit_pg0->PhyData[i].NegotiatedLinkRate >> 4) <
6166 		    MPI2_SAS_NEG_LINK_RATE_1_5)
6167 			continue;
6168 		attached_handle =
6169 		    le16_to_cpu(sas_iounit_pg0->PhyData[i].AttachedDevHandle);
6170 		if (_scsih_get_sas_address(
6171 		    ioc, attached_handle, &attached_sas_addr) != 0) {
6172 			ioc_err(ioc, "failure at %s:%d/%s()!\n",
6173 			    __FILE__, __LINE__, __func__);
6174 			continue;
6175 		}
6176 
6177 		for (j = 0; j < port_count; j++) {
6178 			port_id = sas_iounit_pg0->PhyData[i].Port;
6179 			if (port_table[j].port_id == port_id &&
6180 			    port_table[j].sas_address == attached_sas_addr) {
6181 				port_table[j].phy_mask |= (1 << i);
6182 				found = 1;
6183 				break;
6184 			}
6185 		}
6186 
6187 		if (found)
6188 			continue;
6189 
6190 		port_id = sas_iounit_pg0->PhyData[i].Port;
6191 		port_table[port_count].port_id = port_id;
6192 		port_table[port_count].phy_mask = (1 << i);
6193 		port_table[port_count].sas_address = attached_sas_addr;
6194 		port_count++;
6195 	}
6196 out:
6197 	kfree(sas_iounit_pg0);
6198 	return port_count;
6199 }
6200 
6201 enum hba_port_matched_codes {
6202 	NOT_MATCHED = 0,
6203 	MATCHED_WITH_ADDR_AND_PHYMASK,
6204 	MATCHED_WITH_ADDR_SUBPHYMASK_AND_PORT,
6205 	MATCHED_WITH_ADDR_AND_SUBPHYMASK,
6206 	MATCHED_WITH_ADDR,
6207 };
6208 
6209 /**
6210  * _scsih_look_and_get_matched_port_entry - Get matched hba port entry
6211  *					from HBA port table
6212  * @ioc: per adapter object
6213  * @port_entry: hba port entry from temporary port table which needs to be
6214  *		searched for matched entry in the HBA port table
6215  * @matched_port_entry: save matched hba port entry here
6216  * @count: count of matched entries
6217  *
6218  * return type of matched entry found.
6219  */
6220 static enum hba_port_matched_codes
6221 _scsih_look_and_get_matched_port_entry(struct MPT3SAS_ADAPTER *ioc,
6222 	struct hba_port *port_entry,
6223 	struct hba_port **matched_port_entry, int *count)
6224 {
6225 	struct hba_port *port_table_entry, *matched_port = NULL;
6226 	enum hba_port_matched_codes matched_code = NOT_MATCHED;
6227 	int lcount = 0;
6228 	*matched_port_entry = NULL;
6229 
6230 	list_for_each_entry(port_table_entry, &ioc->port_table_list, list) {
6231 		if (!(port_table_entry->flags & HBA_PORT_FLAG_DIRTY_PORT))
6232 			continue;
6233 
6234 		if ((port_table_entry->sas_address == port_entry->sas_address)
6235 		    && (port_table_entry->phy_mask == port_entry->phy_mask)) {
6236 			matched_code = MATCHED_WITH_ADDR_AND_PHYMASK;
6237 			matched_port = port_table_entry;
6238 			break;
6239 		}
6240 
6241 		if ((port_table_entry->sas_address == port_entry->sas_address)
6242 		    && (port_table_entry->phy_mask & port_entry->phy_mask)
6243 		    && (port_table_entry->port_id == port_entry->port_id)) {
6244 			matched_code = MATCHED_WITH_ADDR_SUBPHYMASK_AND_PORT;
6245 			matched_port = port_table_entry;
6246 			continue;
6247 		}
6248 
6249 		if ((port_table_entry->sas_address == port_entry->sas_address)
6250 		    && (port_table_entry->phy_mask & port_entry->phy_mask)) {
6251 			if (matched_code ==
6252 			    MATCHED_WITH_ADDR_SUBPHYMASK_AND_PORT)
6253 				continue;
6254 			matched_code = MATCHED_WITH_ADDR_AND_SUBPHYMASK;
6255 			matched_port = port_table_entry;
6256 			continue;
6257 		}
6258 
6259 		if (port_table_entry->sas_address == port_entry->sas_address) {
6260 			if (matched_code ==
6261 			    MATCHED_WITH_ADDR_SUBPHYMASK_AND_PORT)
6262 				continue;
6263 			if (matched_code == MATCHED_WITH_ADDR_AND_SUBPHYMASK)
6264 				continue;
6265 			matched_code = MATCHED_WITH_ADDR;
6266 			matched_port = port_table_entry;
6267 			lcount++;
6268 		}
6269 	}
6270 
6271 	*matched_port_entry = matched_port;
6272 	if (matched_code ==  MATCHED_WITH_ADDR)
6273 		*count = lcount;
6274 	return matched_code;
6275 }
6276 
6277 /**
6278  * _scsih_del_phy_part_of_anther_port - remove phy if it
6279  *				is a part of anther port
6280  *@ioc: per adapter object
6281  *@port_table: port table after reset
6282  *@index: hba port entry index
6283  *@port_count: number of ports available after host reset
6284  *@offset: HBA phy bit offset
6285  *
6286  */
6287 static void
6288 _scsih_del_phy_part_of_anther_port(struct MPT3SAS_ADAPTER *ioc,
6289 	struct hba_port *port_table,
6290 	int index, u8 port_count, int offset)
6291 {
6292 	struct _sas_node *sas_node = &ioc->sas_hba;
6293 	u32 i, found = 0;
6294 
6295 	for (i = 0; i < port_count; i++) {
6296 		if (i == index)
6297 			continue;
6298 
6299 		if (port_table[i].phy_mask & (1 << offset)) {
6300 			mpt3sas_transport_del_phy_from_an_existing_port(
6301 			    ioc, sas_node, &sas_node->phy[offset]);
6302 			found = 1;
6303 			break;
6304 		}
6305 	}
6306 	if (!found)
6307 		port_table[index].phy_mask |= (1 << offset);
6308 }
6309 
6310 /**
6311  * _scsih_add_or_del_phys_from_existing_port - add/remove phy to/from
6312  *						right port
6313  *@ioc: per adapter object
6314  *@hba_port_entry: hba port table entry
6315  *@port_table: temporary port table
6316  *@index: hba port entry index
6317  *@port_count: number of ports available after host reset
6318  *
6319  */
6320 static void
6321 _scsih_add_or_del_phys_from_existing_port(struct MPT3SAS_ADAPTER *ioc,
6322 	struct hba_port *hba_port_entry, struct hba_port *port_table,
6323 	int index, int port_count)
6324 {
6325 	u32 phy_mask, offset = 0;
6326 	struct _sas_node *sas_node = &ioc->sas_hba;
6327 
6328 	phy_mask = hba_port_entry->phy_mask ^ port_table[index].phy_mask;
6329 
6330 	for (offset = 0; offset < ioc->sas_hba.num_phys; offset++) {
6331 		if (phy_mask & (1 << offset)) {
6332 			if (!(port_table[index].phy_mask & (1 << offset))) {
6333 				_scsih_del_phy_part_of_anther_port(
6334 				    ioc, port_table, index, port_count,
6335 				    offset);
6336 				continue;
6337 			}
6338 			if (sas_node->phy[offset].phy_belongs_to_port)
6339 				mpt3sas_transport_del_phy_from_an_existing_port(
6340 				    ioc, sas_node, &sas_node->phy[offset]);
6341 			mpt3sas_transport_add_phy_to_an_existing_port(
6342 			    ioc, sas_node, &sas_node->phy[offset],
6343 			    hba_port_entry->sas_address,
6344 			    hba_port_entry);
6345 		}
6346 	}
6347 }
6348 
6349 /**
6350  * _scsih_del_dirty_vphy - delete virtual_phy objects marked as dirty.
6351  * @ioc: per adapter object
6352  *
6353  * Returns nothing.
6354  */
6355 static void
6356 _scsih_del_dirty_vphy(struct MPT3SAS_ADAPTER *ioc)
6357 {
6358 	struct hba_port *port, *port_next;
6359 	struct virtual_phy *vphy, *vphy_next;
6360 
6361 	list_for_each_entry_safe(port, port_next,
6362 	    &ioc->port_table_list, list) {
6363 		if (!port->vphys_mask)
6364 			continue;
6365 		list_for_each_entry_safe(vphy, vphy_next,
6366 		    &port->vphys_list, list) {
6367 			if (vphy->flags & MPT_VPHY_FLAG_DIRTY_PHY) {
6368 				drsprintk(ioc, ioc_info(ioc,
6369 				    "Deleting vphy %p entry from port id: %d\t, Phy_mask 0x%08x\n",
6370 				    vphy, port->port_id,
6371 				    vphy->phy_mask));
6372 				port->vphys_mask &= ~vphy->phy_mask;
6373 				list_del(&vphy->list);
6374 				kfree(vphy);
6375 			}
6376 		}
6377 		if (!port->vphys_mask && !port->sas_address)
6378 			port->flags |= HBA_PORT_FLAG_DIRTY_PORT;
6379 	}
6380 }
6381 
6382 /**
6383  * _scsih_del_dirty_port_entries - delete dirty port entries from port list
6384  *					after host reset
6385  *@ioc: per adapter object
6386  *
6387  */
6388 static void
6389 _scsih_del_dirty_port_entries(struct MPT3SAS_ADAPTER *ioc)
6390 {
6391 	struct hba_port *port, *port_next;
6392 
6393 	list_for_each_entry_safe(port, port_next,
6394 	    &ioc->port_table_list, list) {
6395 		if (!(port->flags & HBA_PORT_FLAG_DIRTY_PORT) ||
6396 		    port->flags & HBA_PORT_FLAG_NEW_PORT)
6397 			continue;
6398 
6399 		drsprintk(ioc, ioc_info(ioc,
6400 		    "Deleting port table entry %p having Port: %d\t Phy_mask 0x%08x\n",
6401 		    port, port->port_id, port->phy_mask));
6402 		list_del(&port->list);
6403 		kfree(port);
6404 	}
6405 }
6406 
6407 /**
6408  * _scsih_sas_port_refresh - Update HBA port table after host reset
6409  * @ioc: per adapter object
6410  */
6411 static void
6412 _scsih_sas_port_refresh(struct MPT3SAS_ADAPTER *ioc)
6413 {
6414 	u32 port_count = 0;
6415 	struct hba_port *port_table;
6416 	struct hba_port *port_table_entry;
6417 	struct hba_port *port_entry = NULL;
6418 	int i, j, count = 0, lcount = 0;
6419 	int ret;
6420 	u64 sas_addr;
6421 	u8 num_phys;
6422 
6423 	drsprintk(ioc, ioc_info(ioc,
6424 	    "updating ports for sas_host(0x%016llx)\n",
6425 	    (unsigned long long)ioc->sas_hba.sas_address));
6426 
6427 	mpt3sas_config_get_number_hba_phys(ioc, &num_phys);
6428 	if (!num_phys) {
6429 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
6430 		    __FILE__, __LINE__, __func__);
6431 		return;
6432 	}
6433 
6434 	if (num_phys > ioc->sas_hba.nr_phys_allocated) {
6435 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
6436 		   __FILE__, __LINE__, __func__);
6437 		return;
6438 	}
6439 	ioc->sas_hba.num_phys = num_phys;
6440 
6441 	port_table = kcalloc(ioc->sas_hba.num_phys,
6442 	    sizeof(struct hba_port), GFP_KERNEL);
6443 	if (!port_table)
6444 		return;
6445 
6446 	port_count = _scsih_get_port_table_after_reset(ioc, port_table);
6447 	if (!port_count)
6448 		return;
6449 
6450 	drsprintk(ioc, ioc_info(ioc, "New Port table\n"));
6451 	for (j = 0; j < port_count; j++)
6452 		drsprintk(ioc, ioc_info(ioc,
6453 		    "Port: %d\t Phy_mask 0x%08x\t sas_addr(0x%016llx)\n",
6454 		    port_table[j].port_id,
6455 		    port_table[j].phy_mask, port_table[j].sas_address));
6456 
6457 	list_for_each_entry(port_table_entry, &ioc->port_table_list, list)
6458 		port_table_entry->flags |= HBA_PORT_FLAG_DIRTY_PORT;
6459 
6460 	drsprintk(ioc, ioc_info(ioc, "Old Port table\n"));
6461 	port_table_entry = NULL;
6462 	list_for_each_entry(port_table_entry, &ioc->port_table_list, list) {
6463 		drsprintk(ioc, ioc_info(ioc,
6464 		    "Port: %d\t Phy_mask 0x%08x\t sas_addr(0x%016llx)\n",
6465 		    port_table_entry->port_id,
6466 		    port_table_entry->phy_mask,
6467 		    port_table_entry->sas_address));
6468 	}
6469 
6470 	for (j = 0; j < port_count; j++) {
6471 		ret = _scsih_look_and_get_matched_port_entry(ioc,
6472 		    &port_table[j], &port_entry, &count);
6473 		if (!port_entry) {
6474 			drsprintk(ioc, ioc_info(ioc,
6475 			    "No Matched entry for sas_addr(0x%16llx), Port:%d\n",
6476 			    port_table[j].sas_address,
6477 			    port_table[j].port_id));
6478 			continue;
6479 		}
6480 
6481 		switch (ret) {
6482 		case MATCHED_WITH_ADDR_SUBPHYMASK_AND_PORT:
6483 		case MATCHED_WITH_ADDR_AND_SUBPHYMASK:
6484 			_scsih_add_or_del_phys_from_existing_port(ioc,
6485 			    port_entry, port_table, j, port_count);
6486 			break;
6487 		case MATCHED_WITH_ADDR:
6488 			sas_addr = port_table[j].sas_address;
6489 			for (i = 0; i < port_count; i++) {
6490 				if (port_table[i].sas_address == sas_addr)
6491 					lcount++;
6492 			}
6493 
6494 			if (count > 1 || lcount > 1)
6495 				port_entry = NULL;
6496 			else
6497 				_scsih_add_or_del_phys_from_existing_port(ioc,
6498 				    port_entry, port_table, j, port_count);
6499 		}
6500 
6501 		if (!port_entry)
6502 			continue;
6503 
6504 		if (port_entry->port_id != port_table[j].port_id)
6505 			port_entry->port_id = port_table[j].port_id;
6506 		port_entry->flags &= ~HBA_PORT_FLAG_DIRTY_PORT;
6507 		port_entry->phy_mask = port_table[j].phy_mask;
6508 	}
6509 
6510 	port_table_entry = NULL;
6511 }
6512 
6513 /**
6514  * _scsih_alloc_vphy - allocate virtual_phy object
6515  * @ioc: per adapter object
6516  * @port_id: Port ID number
6517  * @phy_num: HBA Phy number
6518  *
6519  * Returns allocated virtual_phy object.
6520  */
6521 static struct virtual_phy *
6522 _scsih_alloc_vphy(struct MPT3SAS_ADAPTER *ioc, u8 port_id, u8 phy_num)
6523 {
6524 	struct virtual_phy *vphy;
6525 	struct hba_port *port;
6526 
6527 	port = mpt3sas_get_port_by_id(ioc, port_id, 0);
6528 	if (!port)
6529 		return NULL;
6530 
6531 	vphy = mpt3sas_get_vphy_by_phy(ioc, port, phy_num);
6532 	if (!vphy) {
6533 		vphy = kzalloc(sizeof(struct virtual_phy), GFP_KERNEL);
6534 		if (!vphy)
6535 			return NULL;
6536 
6537 		if (!port->vphys_mask)
6538 			INIT_LIST_HEAD(&port->vphys_list);
6539 
6540 		/*
6541 		 * Enable bit corresponding to HBA phy number on its
6542 		 * parent hba_port object's vphys_mask field.
6543 		 */
6544 		port->vphys_mask |= (1 << phy_num);
6545 		vphy->phy_mask |= (1 << phy_num);
6546 
6547 		list_add_tail(&vphy->list, &port->vphys_list);
6548 
6549 		ioc_info(ioc,
6550 		    "vphy entry: %p, port id: %d, phy:%d is added to port's vphys_list\n",
6551 		    vphy, port->port_id, phy_num);
6552 	}
6553 	return vphy;
6554 }
6555 
6556 /**
6557  * _scsih_sas_host_refresh - refreshing sas host object contents
6558  * @ioc: per adapter object
6559  * Context: user
6560  *
6561  * During port enable, fw will send topology events for every device. Its
6562  * possible that the handles may change from the previous setting, so this
6563  * code keeping handles updating if changed.
6564  */
6565 static void
6566 _scsih_sas_host_refresh(struct MPT3SAS_ADAPTER *ioc)
6567 {
6568 	u16 sz;
6569 	u16 ioc_status;
6570 	int i;
6571 	Mpi2ConfigReply_t mpi_reply;
6572 	Mpi2SasIOUnitPage0_t *sas_iounit_pg0 = NULL;
6573 	u16 attached_handle;
6574 	u8 link_rate, port_id;
6575 	struct hba_port *port;
6576 	Mpi2SasPhyPage0_t phy_pg0;
6577 
6578 	dtmprintk(ioc,
6579 		  ioc_info(ioc, "updating handles for sas_host(0x%016llx)\n",
6580 			   (u64)ioc->sas_hba.sas_address));
6581 
6582 	sz = offsetof(Mpi2SasIOUnitPage0_t, PhyData) + (ioc->sas_hba.num_phys
6583 	    * sizeof(Mpi2SasIOUnit0PhyData_t));
6584 	sas_iounit_pg0 = kzalloc(sz, GFP_KERNEL);
6585 	if (!sas_iounit_pg0) {
6586 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
6587 			__FILE__, __LINE__, __func__);
6588 		return;
6589 	}
6590 
6591 	if ((mpt3sas_config_get_sas_iounit_pg0(ioc, &mpi_reply,
6592 	    sas_iounit_pg0, sz)) != 0)
6593 		goto out;
6594 	ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK;
6595 	if (ioc_status != MPI2_IOCSTATUS_SUCCESS)
6596 		goto out;
6597 	for (i = 0; i < ioc->sas_hba.num_phys ; i++) {
6598 		link_rate = sas_iounit_pg0->PhyData[i].NegotiatedLinkRate >> 4;
6599 		if (i == 0)
6600 			ioc->sas_hba.handle = le16_to_cpu(
6601 			    sas_iounit_pg0->PhyData[0].ControllerDevHandle);
6602 		port_id = sas_iounit_pg0->PhyData[i].Port;
6603 		if (!(mpt3sas_get_port_by_id(ioc, port_id, 0))) {
6604 			port = kzalloc(sizeof(struct hba_port), GFP_KERNEL);
6605 			if (!port)
6606 				goto out;
6607 
6608 			port->port_id = port_id;
6609 			ioc_info(ioc,
6610 			    "hba_port entry: %p, port: %d is added to hba_port list\n",
6611 			    port, port->port_id);
6612 			if (ioc->shost_recovery)
6613 				port->flags = HBA_PORT_FLAG_NEW_PORT;
6614 			list_add_tail(&port->list, &ioc->port_table_list);
6615 		}
6616 		/*
6617 		 * Check whether current Phy belongs to HBA vSES device or not.
6618 		 */
6619 		if (le32_to_cpu(sas_iounit_pg0->PhyData[i].ControllerPhyDeviceInfo) &
6620 		    MPI2_SAS_DEVICE_INFO_SEP &&
6621 		    (link_rate >=  MPI2_SAS_NEG_LINK_RATE_1_5)) {
6622 			if ((mpt3sas_config_get_phy_pg0(ioc, &mpi_reply,
6623 			    &phy_pg0, i))) {
6624 				ioc_err(ioc,
6625 				    "failure at %s:%d/%s()!\n",
6626 				     __FILE__, __LINE__, __func__);
6627 				goto out;
6628 			}
6629 			if (!(le32_to_cpu(phy_pg0.PhyInfo) &
6630 			    MPI2_SAS_PHYINFO_VIRTUAL_PHY))
6631 				continue;
6632 			/*
6633 			 * Allocate a virtual_phy object for vSES device, if
6634 			 * this vSES device is hot added.
6635 			 */
6636 			if (!_scsih_alloc_vphy(ioc, port_id, i))
6637 				goto out;
6638 			ioc->sas_hba.phy[i].hba_vphy = 1;
6639 		}
6640 
6641 		/*
6642 		 * Add new HBA phys to STL if these new phys got added as part
6643 		 * of HBA Firmware upgrade/downgrade operation.
6644 		 */
6645 		if (!ioc->sas_hba.phy[i].phy) {
6646 			if ((mpt3sas_config_get_phy_pg0(ioc, &mpi_reply,
6647 							&phy_pg0, i))) {
6648 				ioc_err(ioc, "failure at %s:%d/%s()!\n",
6649 					__FILE__, __LINE__, __func__);
6650 				continue;
6651 			}
6652 			ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
6653 				MPI2_IOCSTATUS_MASK;
6654 			if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
6655 				ioc_err(ioc, "failure at %s:%d/%s()!\n",
6656 					__FILE__, __LINE__, __func__);
6657 				continue;
6658 			}
6659 			ioc->sas_hba.phy[i].phy_id = i;
6660 			mpt3sas_transport_add_host_phy(ioc,
6661 				&ioc->sas_hba.phy[i], phy_pg0,
6662 				ioc->sas_hba.parent_dev);
6663 			continue;
6664 		}
6665 		ioc->sas_hba.phy[i].handle = ioc->sas_hba.handle;
6666 		attached_handle = le16_to_cpu(sas_iounit_pg0->PhyData[i].
6667 		    AttachedDevHandle);
6668 		if (attached_handle && link_rate < MPI2_SAS_NEG_LINK_RATE_1_5)
6669 			link_rate = MPI2_SAS_NEG_LINK_RATE_1_5;
6670 		ioc->sas_hba.phy[i].port =
6671 		    mpt3sas_get_port_by_id(ioc, port_id, 0);
6672 		mpt3sas_transport_update_links(ioc, ioc->sas_hba.sas_address,
6673 		    attached_handle, i, link_rate,
6674 		    ioc->sas_hba.phy[i].port);
6675 	}
6676 	/*
6677 	 * Clear the phy details if this phy got disabled as part of
6678 	 * HBA Firmware upgrade/downgrade operation.
6679 	 */
6680 	for (i = ioc->sas_hba.num_phys;
6681 	     i < ioc->sas_hba.nr_phys_allocated; i++) {
6682 		if (ioc->sas_hba.phy[i].phy &&
6683 		    ioc->sas_hba.phy[i].phy->negotiated_linkrate >=
6684 		    SAS_LINK_RATE_1_5_GBPS)
6685 			mpt3sas_transport_update_links(ioc,
6686 				ioc->sas_hba.sas_address, 0, i,
6687 				MPI2_SAS_NEG_LINK_RATE_PHY_DISABLED, NULL);
6688 	}
6689  out:
6690 	kfree(sas_iounit_pg0);
6691 }
6692 
6693 /**
6694  * _scsih_sas_host_add - create sas host object
6695  * @ioc: per adapter object
6696  *
6697  * Creating host side data object, stored in ioc->sas_hba
6698  */
6699 static void
6700 _scsih_sas_host_add(struct MPT3SAS_ADAPTER *ioc)
6701 {
6702 	int i;
6703 	Mpi2ConfigReply_t mpi_reply;
6704 	Mpi2SasIOUnitPage0_t *sas_iounit_pg0 = NULL;
6705 	Mpi2SasIOUnitPage1_t *sas_iounit_pg1 = NULL;
6706 	Mpi2SasPhyPage0_t phy_pg0;
6707 	Mpi2SasDevicePage0_t sas_device_pg0;
6708 	Mpi2SasEnclosurePage0_t enclosure_pg0;
6709 	u16 ioc_status;
6710 	u16 sz;
6711 	u8 device_missing_delay;
6712 	u8 num_phys, port_id;
6713 	struct hba_port *port;
6714 
6715 	mpt3sas_config_get_number_hba_phys(ioc, &num_phys);
6716 	if (!num_phys) {
6717 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
6718 			__FILE__, __LINE__, __func__);
6719 		return;
6720 	}
6721 
6722 	ioc->sas_hba.nr_phys_allocated = max_t(u8,
6723 	    MPT_MAX_HBA_NUM_PHYS, num_phys);
6724 	ioc->sas_hba.phy = kcalloc(ioc->sas_hba.nr_phys_allocated,
6725 	    sizeof(struct _sas_phy), GFP_KERNEL);
6726 	if (!ioc->sas_hba.phy) {
6727 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
6728 			__FILE__, __LINE__, __func__);
6729 		goto out;
6730 	}
6731 	ioc->sas_hba.num_phys = num_phys;
6732 
6733 	/* sas_iounit page 0 */
6734 	sz = offsetof(Mpi2SasIOUnitPage0_t, PhyData) + (ioc->sas_hba.num_phys *
6735 	    sizeof(Mpi2SasIOUnit0PhyData_t));
6736 	sas_iounit_pg0 = kzalloc(sz, GFP_KERNEL);
6737 	if (!sas_iounit_pg0) {
6738 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
6739 			__FILE__, __LINE__, __func__);
6740 		return;
6741 	}
6742 	if ((mpt3sas_config_get_sas_iounit_pg0(ioc, &mpi_reply,
6743 	    sas_iounit_pg0, sz))) {
6744 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
6745 			__FILE__, __LINE__, __func__);
6746 		goto out;
6747 	}
6748 	ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
6749 	    MPI2_IOCSTATUS_MASK;
6750 	if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
6751 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
6752 			__FILE__, __LINE__, __func__);
6753 		goto out;
6754 	}
6755 
6756 	/* sas_iounit page 1 */
6757 	sz = offsetof(Mpi2SasIOUnitPage1_t, PhyData) + (ioc->sas_hba.num_phys *
6758 	    sizeof(Mpi2SasIOUnit1PhyData_t));
6759 	sas_iounit_pg1 = kzalloc(sz, GFP_KERNEL);
6760 	if (!sas_iounit_pg1) {
6761 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
6762 			__FILE__, __LINE__, __func__);
6763 		goto out;
6764 	}
6765 	if ((mpt3sas_config_get_sas_iounit_pg1(ioc, &mpi_reply,
6766 	    sas_iounit_pg1, sz))) {
6767 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
6768 			__FILE__, __LINE__, __func__);
6769 		goto out;
6770 	}
6771 	ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
6772 	    MPI2_IOCSTATUS_MASK;
6773 	if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
6774 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
6775 			__FILE__, __LINE__, __func__);
6776 		goto out;
6777 	}
6778 
6779 	ioc->io_missing_delay =
6780 	    sas_iounit_pg1->IODeviceMissingDelay;
6781 	device_missing_delay =
6782 	    sas_iounit_pg1->ReportDeviceMissingDelay;
6783 	if (device_missing_delay & MPI2_SASIOUNIT1_REPORT_MISSING_UNIT_16)
6784 		ioc->device_missing_delay = (device_missing_delay &
6785 		    MPI2_SASIOUNIT1_REPORT_MISSING_TIMEOUT_MASK) * 16;
6786 	else
6787 		ioc->device_missing_delay = device_missing_delay &
6788 		    MPI2_SASIOUNIT1_REPORT_MISSING_TIMEOUT_MASK;
6789 
6790 	ioc->sas_hba.parent_dev = &ioc->shost->shost_gendev;
6791 	for (i = 0; i < ioc->sas_hba.num_phys ; i++) {
6792 		if ((mpt3sas_config_get_phy_pg0(ioc, &mpi_reply, &phy_pg0,
6793 		    i))) {
6794 			ioc_err(ioc, "failure at %s:%d/%s()!\n",
6795 				__FILE__, __LINE__, __func__);
6796 			goto out;
6797 		}
6798 		ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
6799 		    MPI2_IOCSTATUS_MASK;
6800 		if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
6801 			ioc_err(ioc, "failure at %s:%d/%s()!\n",
6802 				__FILE__, __LINE__, __func__);
6803 			goto out;
6804 		}
6805 
6806 		if (i == 0)
6807 			ioc->sas_hba.handle = le16_to_cpu(sas_iounit_pg0->
6808 			    PhyData[0].ControllerDevHandle);
6809 
6810 		port_id = sas_iounit_pg0->PhyData[i].Port;
6811 		if (!(mpt3sas_get_port_by_id(ioc, port_id, 0))) {
6812 			port = kzalloc(sizeof(struct hba_port), GFP_KERNEL);
6813 			if (!port)
6814 				goto out;
6815 
6816 			port->port_id = port_id;
6817 			ioc_info(ioc,
6818 			   "hba_port entry: %p, port: %d is added to hba_port list\n",
6819 			   port, port->port_id);
6820 			list_add_tail(&port->list,
6821 			    &ioc->port_table_list);
6822 		}
6823 
6824 		/*
6825 		 * Check whether current Phy belongs to HBA vSES device or not.
6826 		 */
6827 		if ((le32_to_cpu(phy_pg0.PhyInfo) &
6828 		    MPI2_SAS_PHYINFO_VIRTUAL_PHY) &&
6829 		    (phy_pg0.NegotiatedLinkRate >> 4) >=
6830 		    MPI2_SAS_NEG_LINK_RATE_1_5) {
6831 			/*
6832 			 * Allocate a virtual_phy object for vSES device.
6833 			 */
6834 			if (!_scsih_alloc_vphy(ioc, port_id, i))
6835 				goto out;
6836 			ioc->sas_hba.phy[i].hba_vphy = 1;
6837 		}
6838 
6839 		ioc->sas_hba.phy[i].handle = ioc->sas_hba.handle;
6840 		ioc->sas_hba.phy[i].phy_id = i;
6841 		ioc->sas_hba.phy[i].port =
6842 		    mpt3sas_get_port_by_id(ioc, port_id, 0);
6843 		mpt3sas_transport_add_host_phy(ioc, &ioc->sas_hba.phy[i],
6844 		    phy_pg0, ioc->sas_hba.parent_dev);
6845 	}
6846 	if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0,
6847 	    MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, ioc->sas_hba.handle))) {
6848 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
6849 			__FILE__, __LINE__, __func__);
6850 		goto out;
6851 	}
6852 	ioc->sas_hba.enclosure_handle =
6853 	    le16_to_cpu(sas_device_pg0.EnclosureHandle);
6854 	ioc->sas_hba.sas_address = le64_to_cpu(sas_device_pg0.SASAddress);
6855 	ioc_info(ioc, "host_add: handle(0x%04x), sas_addr(0x%016llx), phys(%d)\n",
6856 		 ioc->sas_hba.handle,
6857 		 (u64)ioc->sas_hba.sas_address,
6858 		 ioc->sas_hba.num_phys);
6859 
6860 	if (ioc->sas_hba.enclosure_handle) {
6861 		if (!(mpt3sas_config_get_enclosure_pg0(ioc, &mpi_reply,
6862 		    &enclosure_pg0, MPI2_SAS_ENCLOS_PGAD_FORM_HANDLE,
6863 		   ioc->sas_hba.enclosure_handle)))
6864 			ioc->sas_hba.enclosure_logical_id =
6865 			    le64_to_cpu(enclosure_pg0.EnclosureLogicalID);
6866 	}
6867 
6868  out:
6869 	kfree(sas_iounit_pg1);
6870 	kfree(sas_iounit_pg0);
6871 }
6872 
6873 /**
6874  * _scsih_expander_add -  creating expander object
6875  * @ioc: per adapter object
6876  * @handle: expander handle
6877  *
6878  * Creating expander object, stored in ioc->sas_expander_list.
6879  *
6880  * Return: 0 for success, else error.
6881  */
6882 static int
6883 _scsih_expander_add(struct MPT3SAS_ADAPTER *ioc, u16 handle)
6884 {
6885 	struct _sas_node *sas_expander;
6886 	struct _enclosure_node *enclosure_dev;
6887 	Mpi2ConfigReply_t mpi_reply;
6888 	Mpi2ExpanderPage0_t expander_pg0;
6889 	Mpi2ExpanderPage1_t expander_pg1;
6890 	u32 ioc_status;
6891 	u16 parent_handle;
6892 	u64 sas_address, sas_address_parent = 0;
6893 	int i;
6894 	unsigned long flags;
6895 	struct _sas_port *mpt3sas_port = NULL;
6896 	u8 port_id;
6897 
6898 	int rc = 0;
6899 
6900 	if (!handle)
6901 		return -1;
6902 
6903 	if (ioc->shost_recovery || ioc->pci_error_recovery)
6904 		return -1;
6905 
6906 	if ((mpt3sas_config_get_expander_pg0(ioc, &mpi_reply, &expander_pg0,
6907 	    MPI2_SAS_EXPAND_PGAD_FORM_HNDL, handle))) {
6908 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
6909 			__FILE__, __LINE__, __func__);
6910 		return -1;
6911 	}
6912 
6913 	ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
6914 	    MPI2_IOCSTATUS_MASK;
6915 	if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
6916 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
6917 			__FILE__, __LINE__, __func__);
6918 		return -1;
6919 	}
6920 
6921 	/* handle out of order topology events */
6922 	parent_handle = le16_to_cpu(expander_pg0.ParentDevHandle);
6923 	if (_scsih_get_sas_address(ioc, parent_handle, &sas_address_parent)
6924 	    != 0) {
6925 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
6926 			__FILE__, __LINE__, __func__);
6927 		return -1;
6928 	}
6929 
6930 	port_id = expander_pg0.PhysicalPort;
6931 	if (sas_address_parent != ioc->sas_hba.sas_address) {
6932 		spin_lock_irqsave(&ioc->sas_node_lock, flags);
6933 		sas_expander = mpt3sas_scsih_expander_find_by_sas_address(ioc,
6934 		    sas_address_parent,
6935 		    mpt3sas_get_port_by_id(ioc, port_id, 0));
6936 		spin_unlock_irqrestore(&ioc->sas_node_lock, flags);
6937 		if (!sas_expander) {
6938 			rc = _scsih_expander_add(ioc, parent_handle);
6939 			if (rc != 0)
6940 				return rc;
6941 		}
6942 	}
6943 
6944 	spin_lock_irqsave(&ioc->sas_node_lock, flags);
6945 	sas_address = le64_to_cpu(expander_pg0.SASAddress);
6946 	sas_expander = mpt3sas_scsih_expander_find_by_sas_address(ioc,
6947 	    sas_address, mpt3sas_get_port_by_id(ioc, port_id, 0));
6948 	spin_unlock_irqrestore(&ioc->sas_node_lock, flags);
6949 
6950 	if (sas_expander)
6951 		return 0;
6952 
6953 	sas_expander = kzalloc(sizeof(struct _sas_node),
6954 	    GFP_KERNEL);
6955 	if (!sas_expander) {
6956 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
6957 			__FILE__, __LINE__, __func__);
6958 		return -1;
6959 	}
6960 
6961 	sas_expander->handle = handle;
6962 	sas_expander->num_phys = expander_pg0.NumPhys;
6963 	sas_expander->sas_address_parent = sas_address_parent;
6964 	sas_expander->sas_address = sas_address;
6965 	sas_expander->port = mpt3sas_get_port_by_id(ioc, port_id, 0);
6966 	if (!sas_expander->port) {
6967 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
6968 		    __FILE__, __LINE__, __func__);
6969 		rc = -1;
6970 		goto out_fail;
6971 	}
6972 
6973 	ioc_info(ioc, "expander_add: handle(0x%04x), parent(0x%04x), sas_addr(0x%016llx), phys(%d)\n",
6974 		 handle, parent_handle,
6975 		 (u64)sas_expander->sas_address, sas_expander->num_phys);
6976 
6977 	if (!sas_expander->num_phys) {
6978 		rc = -1;
6979 		goto out_fail;
6980 	}
6981 	sas_expander->phy = kcalloc(sas_expander->num_phys,
6982 	    sizeof(struct _sas_phy), GFP_KERNEL);
6983 	if (!sas_expander->phy) {
6984 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
6985 			__FILE__, __LINE__, __func__);
6986 		rc = -1;
6987 		goto out_fail;
6988 	}
6989 
6990 	INIT_LIST_HEAD(&sas_expander->sas_port_list);
6991 	mpt3sas_port = mpt3sas_transport_port_add(ioc, handle,
6992 	    sas_address_parent, sas_expander->port);
6993 	if (!mpt3sas_port) {
6994 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
6995 			__FILE__, __LINE__, __func__);
6996 		rc = -1;
6997 		goto out_fail;
6998 	}
6999 	sas_expander->parent_dev = &mpt3sas_port->rphy->dev;
7000 	sas_expander->rphy = mpt3sas_port->rphy;
7001 
7002 	for (i = 0 ; i < sas_expander->num_phys ; i++) {
7003 		if ((mpt3sas_config_get_expander_pg1(ioc, &mpi_reply,
7004 		    &expander_pg1, i, handle))) {
7005 			ioc_err(ioc, "failure at %s:%d/%s()!\n",
7006 				__FILE__, __LINE__, __func__);
7007 			rc = -1;
7008 			goto out_fail;
7009 		}
7010 		sas_expander->phy[i].handle = handle;
7011 		sas_expander->phy[i].phy_id = i;
7012 		sas_expander->phy[i].port =
7013 		    mpt3sas_get_port_by_id(ioc, port_id, 0);
7014 
7015 		if ((mpt3sas_transport_add_expander_phy(ioc,
7016 		    &sas_expander->phy[i], expander_pg1,
7017 		    sas_expander->parent_dev))) {
7018 			ioc_err(ioc, "failure at %s:%d/%s()!\n",
7019 				__FILE__, __LINE__, __func__);
7020 			rc = -1;
7021 			goto out_fail;
7022 		}
7023 	}
7024 
7025 	if (sas_expander->enclosure_handle) {
7026 		enclosure_dev =
7027 			mpt3sas_scsih_enclosure_find_by_handle(ioc,
7028 						sas_expander->enclosure_handle);
7029 		if (enclosure_dev)
7030 			sas_expander->enclosure_logical_id =
7031 			    le64_to_cpu(enclosure_dev->pg0.EnclosureLogicalID);
7032 	}
7033 
7034 	_scsih_expander_node_add(ioc, sas_expander);
7035 	return 0;
7036 
7037  out_fail:
7038 
7039 	if (mpt3sas_port)
7040 		mpt3sas_transport_port_remove(ioc, sas_expander->sas_address,
7041 		    sas_address_parent, sas_expander->port);
7042 	kfree(sas_expander);
7043 	return rc;
7044 }
7045 
7046 /**
7047  * mpt3sas_expander_remove - removing expander object
7048  * @ioc: per adapter object
7049  * @sas_address: expander sas_address
7050  * @port: hba port entry
7051  */
7052 void
7053 mpt3sas_expander_remove(struct MPT3SAS_ADAPTER *ioc, u64 sas_address,
7054 	struct hba_port *port)
7055 {
7056 	struct _sas_node *sas_expander;
7057 	unsigned long flags;
7058 
7059 	if (ioc->shost_recovery)
7060 		return;
7061 
7062 	if (!port)
7063 		return;
7064 
7065 	spin_lock_irqsave(&ioc->sas_node_lock, flags);
7066 	sas_expander = mpt3sas_scsih_expander_find_by_sas_address(ioc,
7067 	    sas_address, port);
7068 	spin_unlock_irqrestore(&ioc->sas_node_lock, flags);
7069 	if (sas_expander)
7070 		_scsih_expander_node_remove(ioc, sas_expander);
7071 }
7072 
7073 /**
7074  * _scsih_done -  internal SCSI_IO callback handler.
7075  * @ioc: per adapter object
7076  * @smid: system request message index
7077  * @msix_index: MSIX table index supplied by the OS
7078  * @reply: reply message frame(lower 32bit addr)
7079  *
7080  * Callback handler when sending internal generated SCSI_IO.
7081  * The callback index passed is `ioc->scsih_cb_idx`
7082  *
7083  * Return: 1 meaning mf should be freed from _base_interrupt
7084  *         0 means the mf is freed from this function.
7085  */
7086 static u8
7087 _scsih_done(struct MPT3SAS_ADAPTER *ioc, u16 smid, u8 msix_index, u32 reply)
7088 {
7089 	MPI2DefaultReply_t *mpi_reply;
7090 
7091 	mpi_reply =  mpt3sas_base_get_reply_virt_addr(ioc, reply);
7092 	if (ioc->scsih_cmds.status == MPT3_CMD_NOT_USED)
7093 		return 1;
7094 	if (ioc->scsih_cmds.smid != smid)
7095 		return 1;
7096 	ioc->scsih_cmds.status |= MPT3_CMD_COMPLETE;
7097 	if (mpi_reply) {
7098 		memcpy(ioc->scsih_cmds.reply, mpi_reply,
7099 		    mpi_reply->MsgLength*4);
7100 		ioc->scsih_cmds.status |= MPT3_CMD_REPLY_VALID;
7101 	}
7102 	ioc->scsih_cmds.status &= ~MPT3_CMD_PENDING;
7103 	complete(&ioc->scsih_cmds.done);
7104 	return 1;
7105 }
7106 
7107 
7108 
7109 
7110 #define MPT3_MAX_LUNS (255)
7111 
7112 
7113 /**
7114  * _scsih_check_access_status - check access flags
7115  * @ioc: per adapter object
7116  * @sas_address: sas address
7117  * @handle: sas device handle
7118  * @access_status: errors returned during discovery of the device
7119  *
7120  * Return: 0 for success, else failure
7121  */
7122 static u8
7123 _scsih_check_access_status(struct MPT3SAS_ADAPTER *ioc, u64 sas_address,
7124 	u16 handle, u8 access_status)
7125 {
7126 	u8 rc = 1;
7127 	char *desc = NULL;
7128 
7129 	switch (access_status) {
7130 	case MPI2_SAS_DEVICE0_ASTATUS_NO_ERRORS:
7131 	case MPI2_SAS_DEVICE0_ASTATUS_SATA_NEEDS_INITIALIZATION:
7132 		rc = 0;
7133 		break;
7134 	case MPI2_SAS_DEVICE0_ASTATUS_SATA_CAPABILITY_FAILED:
7135 		desc = "sata capability failed";
7136 		break;
7137 	case MPI2_SAS_DEVICE0_ASTATUS_SATA_AFFILIATION_CONFLICT:
7138 		desc = "sata affiliation conflict";
7139 		break;
7140 	case MPI2_SAS_DEVICE0_ASTATUS_ROUTE_NOT_ADDRESSABLE:
7141 		desc = "route not addressable";
7142 		break;
7143 	case MPI2_SAS_DEVICE0_ASTATUS_SMP_ERROR_NOT_ADDRESSABLE:
7144 		desc = "smp error not addressable";
7145 		break;
7146 	case MPI2_SAS_DEVICE0_ASTATUS_DEVICE_BLOCKED:
7147 		desc = "device blocked";
7148 		break;
7149 	case MPI2_SAS_DEVICE0_ASTATUS_SATA_INIT_FAILED:
7150 	case MPI2_SAS_DEVICE0_ASTATUS_SIF_UNKNOWN:
7151 	case MPI2_SAS_DEVICE0_ASTATUS_SIF_AFFILIATION_CONFLICT:
7152 	case MPI2_SAS_DEVICE0_ASTATUS_SIF_DIAG:
7153 	case MPI2_SAS_DEVICE0_ASTATUS_SIF_IDENTIFICATION:
7154 	case MPI2_SAS_DEVICE0_ASTATUS_SIF_CHECK_POWER:
7155 	case MPI2_SAS_DEVICE0_ASTATUS_SIF_PIO_SN:
7156 	case MPI2_SAS_DEVICE0_ASTATUS_SIF_MDMA_SN:
7157 	case MPI2_SAS_DEVICE0_ASTATUS_SIF_UDMA_SN:
7158 	case MPI2_SAS_DEVICE0_ASTATUS_SIF_ZONING_VIOLATION:
7159 	case MPI2_SAS_DEVICE0_ASTATUS_SIF_NOT_ADDRESSABLE:
7160 	case MPI2_SAS_DEVICE0_ASTATUS_SIF_MAX:
7161 		desc = "sata initialization failed";
7162 		break;
7163 	default:
7164 		desc = "unknown";
7165 		break;
7166 	}
7167 
7168 	if (!rc)
7169 		return 0;
7170 
7171 	ioc_err(ioc, "discovery errors(%s): sas_address(0x%016llx), handle(0x%04x)\n",
7172 		desc, (u64)sas_address, handle);
7173 	return rc;
7174 }
7175 
7176 /**
7177  * _scsih_check_device - checking device responsiveness
7178  * @ioc: per adapter object
7179  * @parent_sas_address: sas address of parent expander or sas host
7180  * @handle: attached device handle
7181  * @phy_number: phy number
7182  * @link_rate: new link rate
7183  */
7184 static void
7185 _scsih_check_device(struct MPT3SAS_ADAPTER *ioc,
7186 	u64 parent_sas_address, u16 handle, u8 phy_number, u8 link_rate)
7187 {
7188 	Mpi2ConfigReply_t mpi_reply;
7189 	Mpi2SasDevicePage0_t sas_device_pg0;
7190 	struct _sas_device *sas_device = NULL;
7191 	struct _enclosure_node *enclosure_dev = NULL;
7192 	u32 ioc_status;
7193 	unsigned long flags;
7194 	u64 sas_address;
7195 	struct scsi_target *starget;
7196 	struct MPT3SAS_TARGET *sas_target_priv_data;
7197 	u32 device_info;
7198 	struct hba_port *port;
7199 
7200 	if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0,
7201 	    MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle)))
7202 		return;
7203 
7204 	ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK;
7205 	if (ioc_status != MPI2_IOCSTATUS_SUCCESS)
7206 		return;
7207 
7208 	/* wide port handling ~ we need only handle device once for the phy that
7209 	 * is matched in sas device page zero
7210 	 */
7211 	if (phy_number != sas_device_pg0.PhyNum)
7212 		return;
7213 
7214 	/* check if this is end device */
7215 	device_info = le32_to_cpu(sas_device_pg0.DeviceInfo);
7216 	if (!(_scsih_is_end_device(device_info)))
7217 		return;
7218 
7219 	spin_lock_irqsave(&ioc->sas_device_lock, flags);
7220 	sas_address = le64_to_cpu(sas_device_pg0.SASAddress);
7221 	port = mpt3sas_get_port_by_id(ioc, sas_device_pg0.PhysicalPort, 0);
7222 	if (!port)
7223 		goto out_unlock;
7224 	sas_device = __mpt3sas_get_sdev_by_addr(ioc,
7225 	    sas_address, port);
7226 
7227 	if (!sas_device)
7228 		goto out_unlock;
7229 
7230 	if (unlikely(sas_device->handle != handle)) {
7231 		starget = sas_device->starget;
7232 		sas_target_priv_data = starget->hostdata;
7233 		starget_printk(KERN_INFO, starget,
7234 			"handle changed from(0x%04x) to (0x%04x)!!!\n",
7235 			sas_device->handle, handle);
7236 		sas_target_priv_data->handle = handle;
7237 		sas_device->handle = handle;
7238 		if (le16_to_cpu(sas_device_pg0.Flags) &
7239 		     MPI2_SAS_DEVICE0_FLAGS_ENCL_LEVEL_VALID) {
7240 			sas_device->enclosure_level =
7241 				sas_device_pg0.EnclosureLevel;
7242 			memcpy(sas_device->connector_name,
7243 				sas_device_pg0.ConnectorName, 4);
7244 			sas_device->connector_name[4] = '\0';
7245 		} else {
7246 			sas_device->enclosure_level = 0;
7247 			sas_device->connector_name[0] = '\0';
7248 		}
7249 
7250 		sas_device->enclosure_handle =
7251 				le16_to_cpu(sas_device_pg0.EnclosureHandle);
7252 		sas_device->is_chassis_slot_valid = 0;
7253 		enclosure_dev = mpt3sas_scsih_enclosure_find_by_handle(ioc,
7254 						sas_device->enclosure_handle);
7255 		if (enclosure_dev) {
7256 			sas_device->enclosure_logical_id =
7257 			    le64_to_cpu(enclosure_dev->pg0.EnclosureLogicalID);
7258 			if (le16_to_cpu(enclosure_dev->pg0.Flags) &
7259 			    MPI2_SAS_ENCLS0_FLAGS_CHASSIS_SLOT_VALID) {
7260 				sas_device->is_chassis_slot_valid = 1;
7261 				sas_device->chassis_slot =
7262 					enclosure_dev->pg0.ChassisSlot;
7263 			}
7264 		}
7265 	}
7266 
7267 	/* check if device is present */
7268 	if (!(le16_to_cpu(sas_device_pg0.Flags) &
7269 	    MPI2_SAS_DEVICE0_FLAGS_DEVICE_PRESENT)) {
7270 		ioc_err(ioc, "device is not present handle(0x%04x), flags!!!\n",
7271 			handle);
7272 		goto out_unlock;
7273 	}
7274 
7275 	/* check if there were any issues with discovery */
7276 	if (_scsih_check_access_status(ioc, sas_address, handle,
7277 	    sas_device_pg0.AccessStatus))
7278 		goto out_unlock;
7279 
7280 	spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
7281 	_scsih_ublock_io_device(ioc, sas_address, port);
7282 
7283 	if (sas_device)
7284 		sas_device_put(sas_device);
7285 	return;
7286 
7287 out_unlock:
7288 	spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
7289 	if (sas_device)
7290 		sas_device_put(sas_device);
7291 }
7292 
7293 /**
7294  * _scsih_add_device -  creating sas device object
7295  * @ioc: per adapter object
7296  * @handle: sas device handle
7297  * @phy_num: phy number end device attached to
7298  * @is_pd: is this hidden raid component
7299  *
7300  * Creating end device object, stored in ioc->sas_device_list.
7301  *
7302  * Return: 0 for success, non-zero for failure.
7303  */
7304 static int
7305 _scsih_add_device(struct MPT3SAS_ADAPTER *ioc, u16 handle, u8 phy_num,
7306 	u8 is_pd)
7307 {
7308 	Mpi2ConfigReply_t mpi_reply;
7309 	Mpi2SasDevicePage0_t sas_device_pg0;
7310 	struct _sas_device *sas_device;
7311 	struct _enclosure_node *enclosure_dev = NULL;
7312 	u32 ioc_status;
7313 	u64 sas_address;
7314 	u32 device_info;
7315 	u8 port_id;
7316 
7317 	if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0,
7318 	    MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) {
7319 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
7320 			__FILE__, __LINE__, __func__);
7321 		return -1;
7322 	}
7323 
7324 	ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
7325 	    MPI2_IOCSTATUS_MASK;
7326 	if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
7327 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
7328 			__FILE__, __LINE__, __func__);
7329 		return -1;
7330 	}
7331 
7332 	/* check if this is end device */
7333 	device_info = le32_to_cpu(sas_device_pg0.DeviceInfo);
7334 	if (!(_scsih_is_end_device(device_info)))
7335 		return -1;
7336 	set_bit(handle, ioc->pend_os_device_add);
7337 	sas_address = le64_to_cpu(sas_device_pg0.SASAddress);
7338 
7339 	/* check if device is present */
7340 	if (!(le16_to_cpu(sas_device_pg0.Flags) &
7341 	    MPI2_SAS_DEVICE0_FLAGS_DEVICE_PRESENT)) {
7342 		ioc_err(ioc, "device is not present handle(0x04%x)!!!\n",
7343 			handle);
7344 		return -1;
7345 	}
7346 
7347 	/* check if there were any issues with discovery */
7348 	if (_scsih_check_access_status(ioc, sas_address, handle,
7349 	    sas_device_pg0.AccessStatus))
7350 		return -1;
7351 
7352 	port_id = sas_device_pg0.PhysicalPort;
7353 	sas_device = mpt3sas_get_sdev_by_addr(ioc,
7354 	    sas_address, mpt3sas_get_port_by_id(ioc, port_id, 0));
7355 	if (sas_device) {
7356 		clear_bit(handle, ioc->pend_os_device_add);
7357 		sas_device_put(sas_device);
7358 		return -1;
7359 	}
7360 
7361 	if (sas_device_pg0.EnclosureHandle) {
7362 		enclosure_dev =
7363 			mpt3sas_scsih_enclosure_find_by_handle(ioc,
7364 			    le16_to_cpu(sas_device_pg0.EnclosureHandle));
7365 		if (enclosure_dev == NULL)
7366 			ioc_info(ioc, "Enclosure handle(0x%04x) doesn't match with enclosure device!\n",
7367 				 sas_device_pg0.EnclosureHandle);
7368 	}
7369 
7370 	sas_device = kzalloc(sizeof(struct _sas_device),
7371 	    GFP_KERNEL);
7372 	if (!sas_device) {
7373 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
7374 			__FILE__, __LINE__, __func__);
7375 		return 0;
7376 	}
7377 
7378 	kref_init(&sas_device->refcount);
7379 	sas_device->handle = handle;
7380 	if (_scsih_get_sas_address(ioc,
7381 	    le16_to_cpu(sas_device_pg0.ParentDevHandle),
7382 	    &sas_device->sas_address_parent) != 0)
7383 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
7384 			__FILE__, __LINE__, __func__);
7385 	sas_device->enclosure_handle =
7386 	    le16_to_cpu(sas_device_pg0.EnclosureHandle);
7387 	if (sas_device->enclosure_handle != 0)
7388 		sas_device->slot =
7389 		    le16_to_cpu(sas_device_pg0.Slot);
7390 	sas_device->device_info = device_info;
7391 	sas_device->sas_address = sas_address;
7392 	sas_device->phy = sas_device_pg0.PhyNum;
7393 	sas_device->fast_path = (le16_to_cpu(sas_device_pg0.Flags) &
7394 	    MPI25_SAS_DEVICE0_FLAGS_FAST_PATH_CAPABLE) ? 1 : 0;
7395 	sas_device->port = mpt3sas_get_port_by_id(ioc, port_id, 0);
7396 	if (!sas_device->port) {
7397 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
7398 		    __FILE__, __LINE__, __func__);
7399 		goto out;
7400 	}
7401 
7402 	if (le16_to_cpu(sas_device_pg0.Flags)
7403 		& MPI2_SAS_DEVICE0_FLAGS_ENCL_LEVEL_VALID) {
7404 		sas_device->enclosure_level =
7405 			sas_device_pg0.EnclosureLevel;
7406 		memcpy(sas_device->connector_name,
7407 			sas_device_pg0.ConnectorName, 4);
7408 		sas_device->connector_name[4] = '\0';
7409 	} else {
7410 		sas_device->enclosure_level = 0;
7411 		sas_device->connector_name[0] = '\0';
7412 	}
7413 	/* get enclosure_logical_id & chassis_slot*/
7414 	sas_device->is_chassis_slot_valid = 0;
7415 	if (enclosure_dev) {
7416 		sas_device->enclosure_logical_id =
7417 		    le64_to_cpu(enclosure_dev->pg0.EnclosureLogicalID);
7418 		if (le16_to_cpu(enclosure_dev->pg0.Flags) &
7419 		    MPI2_SAS_ENCLS0_FLAGS_CHASSIS_SLOT_VALID) {
7420 			sas_device->is_chassis_slot_valid = 1;
7421 			sas_device->chassis_slot =
7422 					enclosure_dev->pg0.ChassisSlot;
7423 		}
7424 	}
7425 
7426 	/* get device name */
7427 	sas_device->device_name = le64_to_cpu(sas_device_pg0.DeviceName);
7428 	sas_device->port_type = sas_device_pg0.MaxPortConnections;
7429 	ioc_info(ioc,
7430 	    "handle(0x%0x) sas_address(0x%016llx) port_type(0x%0x)\n",
7431 	    handle, sas_device->sas_address, sas_device->port_type);
7432 
7433 	if (ioc->wait_for_discovery_to_complete)
7434 		_scsih_sas_device_init_add(ioc, sas_device);
7435 	else
7436 		_scsih_sas_device_add(ioc, sas_device);
7437 
7438 out:
7439 	sas_device_put(sas_device);
7440 	return 0;
7441 }
7442 
7443 /**
7444  * _scsih_remove_device -  removing sas device object
7445  * @ioc: per adapter object
7446  * @sas_device: the sas_device object
7447  */
7448 static void
7449 _scsih_remove_device(struct MPT3SAS_ADAPTER *ioc,
7450 	struct _sas_device *sas_device)
7451 {
7452 	struct MPT3SAS_TARGET *sas_target_priv_data;
7453 
7454 	if ((ioc->pdev->subsystem_vendor == PCI_VENDOR_ID_IBM) &&
7455 	     (sas_device->pfa_led_on)) {
7456 		_scsih_turn_off_pfa_led(ioc, sas_device);
7457 		sas_device->pfa_led_on = 0;
7458 	}
7459 
7460 	dewtprintk(ioc,
7461 		   ioc_info(ioc, "%s: enter: handle(0x%04x), sas_addr(0x%016llx)\n",
7462 			    __func__,
7463 			    sas_device->handle, (u64)sas_device->sas_address));
7464 
7465 	dewtprintk(ioc, _scsih_display_enclosure_chassis_info(ioc, sas_device,
7466 	    NULL, NULL));
7467 
7468 	if (sas_device->starget && sas_device->starget->hostdata) {
7469 		sas_target_priv_data = sas_device->starget->hostdata;
7470 		sas_target_priv_data->deleted = 1;
7471 		_scsih_ublock_io_device(ioc, sas_device->sas_address,
7472 		    sas_device->port);
7473 		sas_target_priv_data->handle =
7474 		     MPT3SAS_INVALID_DEVICE_HANDLE;
7475 	}
7476 
7477 	if (!ioc->hide_drives)
7478 		mpt3sas_transport_port_remove(ioc,
7479 		    sas_device->sas_address,
7480 		    sas_device->sas_address_parent,
7481 		    sas_device->port);
7482 
7483 	ioc_info(ioc, "removing handle(0x%04x), sas_addr(0x%016llx)\n",
7484 		 sas_device->handle, (u64)sas_device->sas_address);
7485 
7486 	_scsih_display_enclosure_chassis_info(ioc, sas_device, NULL, NULL);
7487 
7488 	dewtprintk(ioc,
7489 		   ioc_info(ioc, "%s: exit: handle(0x%04x), sas_addr(0x%016llx)\n",
7490 			    __func__,
7491 			    sas_device->handle, (u64)sas_device->sas_address));
7492 	dewtprintk(ioc, _scsih_display_enclosure_chassis_info(ioc, sas_device,
7493 	    NULL, NULL));
7494 }
7495 
7496 /**
7497  * _scsih_sas_topology_change_event_debug - debug for topology event
7498  * @ioc: per adapter object
7499  * @event_data: event data payload
7500  * Context: user.
7501  */
7502 static void
7503 _scsih_sas_topology_change_event_debug(struct MPT3SAS_ADAPTER *ioc,
7504 	Mpi2EventDataSasTopologyChangeList_t *event_data)
7505 {
7506 	int i;
7507 	u16 handle;
7508 	u16 reason_code;
7509 	u8 phy_number;
7510 	char *status_str = NULL;
7511 	u8 link_rate, prev_link_rate;
7512 
7513 	switch (event_data->ExpStatus) {
7514 	case MPI2_EVENT_SAS_TOPO_ES_ADDED:
7515 		status_str = "add";
7516 		break;
7517 	case MPI2_EVENT_SAS_TOPO_ES_NOT_RESPONDING:
7518 		status_str = "remove";
7519 		break;
7520 	case MPI2_EVENT_SAS_TOPO_ES_RESPONDING:
7521 	case 0:
7522 		status_str =  "responding";
7523 		break;
7524 	case MPI2_EVENT_SAS_TOPO_ES_DELAY_NOT_RESPONDING:
7525 		status_str = "remove delay";
7526 		break;
7527 	default:
7528 		status_str = "unknown status";
7529 		break;
7530 	}
7531 	ioc_info(ioc, "sas topology change: (%s)\n", status_str);
7532 	pr_info("\thandle(0x%04x), enclosure_handle(0x%04x) " \
7533 	    "start_phy(%02d), count(%d)\n",
7534 	    le16_to_cpu(event_data->ExpanderDevHandle),
7535 	    le16_to_cpu(event_data->EnclosureHandle),
7536 	    event_data->StartPhyNum, event_data->NumEntries);
7537 	for (i = 0; i < event_data->NumEntries; i++) {
7538 		handle = le16_to_cpu(event_data->PHY[i].AttachedDevHandle);
7539 		if (!handle)
7540 			continue;
7541 		phy_number = event_data->StartPhyNum + i;
7542 		reason_code = event_data->PHY[i].PhyStatus &
7543 		    MPI2_EVENT_SAS_TOPO_RC_MASK;
7544 		switch (reason_code) {
7545 		case MPI2_EVENT_SAS_TOPO_RC_TARG_ADDED:
7546 			status_str = "target add";
7547 			break;
7548 		case MPI2_EVENT_SAS_TOPO_RC_TARG_NOT_RESPONDING:
7549 			status_str = "target remove";
7550 			break;
7551 		case MPI2_EVENT_SAS_TOPO_RC_DELAY_NOT_RESPONDING:
7552 			status_str = "delay target remove";
7553 			break;
7554 		case MPI2_EVENT_SAS_TOPO_RC_PHY_CHANGED:
7555 			status_str = "link rate change";
7556 			break;
7557 		case MPI2_EVENT_SAS_TOPO_RC_NO_CHANGE:
7558 			status_str = "target responding";
7559 			break;
7560 		default:
7561 			status_str = "unknown";
7562 			break;
7563 		}
7564 		link_rate = event_data->PHY[i].LinkRate >> 4;
7565 		prev_link_rate = event_data->PHY[i].LinkRate & 0xF;
7566 		pr_info("\tphy(%02d), attached_handle(0x%04x): %s:" \
7567 		    " link rate: new(0x%02x), old(0x%02x)\n", phy_number,
7568 		    handle, status_str, link_rate, prev_link_rate);
7569 
7570 	}
7571 }
7572 
7573 /**
7574  * _scsih_sas_topology_change_event - handle topology changes
7575  * @ioc: per adapter object
7576  * @fw_event: The fw_event_work object
7577  * Context: user.
7578  *
7579  */
7580 static int
7581 _scsih_sas_topology_change_event(struct MPT3SAS_ADAPTER *ioc,
7582 	struct fw_event_work *fw_event)
7583 {
7584 	int i;
7585 	u16 parent_handle, handle;
7586 	u16 reason_code;
7587 	u8 phy_number, max_phys;
7588 	struct _sas_node *sas_expander;
7589 	u64 sas_address;
7590 	unsigned long flags;
7591 	u8 link_rate, prev_link_rate;
7592 	struct hba_port *port;
7593 	Mpi2EventDataSasTopologyChangeList_t *event_data =
7594 		(Mpi2EventDataSasTopologyChangeList_t *)
7595 		fw_event->event_data;
7596 
7597 	if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK)
7598 		_scsih_sas_topology_change_event_debug(ioc, event_data);
7599 
7600 	if (ioc->shost_recovery || ioc->remove_host || ioc->pci_error_recovery)
7601 		return 0;
7602 
7603 	if (!ioc->sas_hba.num_phys)
7604 		_scsih_sas_host_add(ioc);
7605 	else
7606 		_scsih_sas_host_refresh(ioc);
7607 
7608 	if (fw_event->ignore) {
7609 		dewtprintk(ioc, ioc_info(ioc, "ignoring expander event\n"));
7610 		return 0;
7611 	}
7612 
7613 	parent_handle = le16_to_cpu(event_data->ExpanderDevHandle);
7614 	port = mpt3sas_get_port_by_id(ioc, event_data->PhysicalPort, 0);
7615 
7616 	/* handle expander add */
7617 	if (event_data->ExpStatus == MPI2_EVENT_SAS_TOPO_ES_ADDED)
7618 		if (_scsih_expander_add(ioc, parent_handle) != 0)
7619 			return 0;
7620 
7621 	spin_lock_irqsave(&ioc->sas_node_lock, flags);
7622 	sas_expander = mpt3sas_scsih_expander_find_by_handle(ioc,
7623 	    parent_handle);
7624 	if (sas_expander) {
7625 		sas_address = sas_expander->sas_address;
7626 		max_phys = sas_expander->num_phys;
7627 		port = sas_expander->port;
7628 	} else if (parent_handle < ioc->sas_hba.num_phys) {
7629 		sas_address = ioc->sas_hba.sas_address;
7630 		max_phys = ioc->sas_hba.num_phys;
7631 	} else {
7632 		spin_unlock_irqrestore(&ioc->sas_node_lock, flags);
7633 		return 0;
7634 	}
7635 	spin_unlock_irqrestore(&ioc->sas_node_lock, flags);
7636 
7637 	/* handle siblings events */
7638 	for (i = 0; i < event_data->NumEntries; i++) {
7639 		if (fw_event->ignore) {
7640 			dewtprintk(ioc,
7641 				   ioc_info(ioc, "ignoring expander event\n"));
7642 			return 0;
7643 		}
7644 		if (ioc->remove_host || ioc->pci_error_recovery)
7645 			return 0;
7646 		phy_number = event_data->StartPhyNum + i;
7647 		if (phy_number >= max_phys)
7648 			continue;
7649 		reason_code = event_data->PHY[i].PhyStatus &
7650 		    MPI2_EVENT_SAS_TOPO_RC_MASK;
7651 		if ((event_data->PHY[i].PhyStatus &
7652 		    MPI2_EVENT_SAS_TOPO_PHYSTATUS_VACANT) && (reason_code !=
7653 		    MPI2_EVENT_SAS_TOPO_RC_TARG_NOT_RESPONDING))
7654 				continue;
7655 		handle = le16_to_cpu(event_data->PHY[i].AttachedDevHandle);
7656 		if (!handle)
7657 			continue;
7658 		link_rate = event_data->PHY[i].LinkRate >> 4;
7659 		prev_link_rate = event_data->PHY[i].LinkRate & 0xF;
7660 		switch (reason_code) {
7661 		case MPI2_EVENT_SAS_TOPO_RC_PHY_CHANGED:
7662 
7663 			if (ioc->shost_recovery)
7664 				break;
7665 
7666 			if (link_rate == prev_link_rate)
7667 				break;
7668 
7669 			mpt3sas_transport_update_links(ioc, sas_address,
7670 			    handle, phy_number, link_rate, port);
7671 
7672 			if (link_rate < MPI2_SAS_NEG_LINK_RATE_1_5)
7673 				break;
7674 
7675 			_scsih_check_device(ioc, sas_address, handle,
7676 			    phy_number, link_rate);
7677 
7678 			if (!test_bit(handle, ioc->pend_os_device_add))
7679 				break;
7680 
7681 			fallthrough;
7682 
7683 		case MPI2_EVENT_SAS_TOPO_RC_TARG_ADDED:
7684 
7685 			if (ioc->shost_recovery)
7686 				break;
7687 
7688 			mpt3sas_transport_update_links(ioc, sas_address,
7689 			    handle, phy_number, link_rate, port);
7690 
7691 			_scsih_add_device(ioc, handle, phy_number, 0);
7692 
7693 			break;
7694 		case MPI2_EVENT_SAS_TOPO_RC_TARG_NOT_RESPONDING:
7695 
7696 			_scsih_device_remove_by_handle(ioc, handle);
7697 			break;
7698 		}
7699 	}
7700 
7701 	/* handle expander removal */
7702 	if (event_data->ExpStatus == MPI2_EVENT_SAS_TOPO_ES_NOT_RESPONDING &&
7703 	    sas_expander)
7704 		mpt3sas_expander_remove(ioc, sas_address, port);
7705 
7706 	return 0;
7707 }
7708 
7709 /**
7710  * _scsih_sas_device_status_change_event_debug - debug for device event
7711  * @ioc: ?
7712  * @event_data: event data payload
7713  * Context: user.
7714  */
7715 static void
7716 _scsih_sas_device_status_change_event_debug(struct MPT3SAS_ADAPTER *ioc,
7717 	Mpi2EventDataSasDeviceStatusChange_t *event_data)
7718 {
7719 	char *reason_str = NULL;
7720 
7721 	switch (event_data->ReasonCode) {
7722 	case MPI2_EVENT_SAS_DEV_STAT_RC_SMART_DATA:
7723 		reason_str = "smart data";
7724 		break;
7725 	case MPI2_EVENT_SAS_DEV_STAT_RC_UNSUPPORTED:
7726 		reason_str = "unsupported device discovered";
7727 		break;
7728 	case MPI2_EVENT_SAS_DEV_STAT_RC_INTERNAL_DEVICE_RESET:
7729 		reason_str = "internal device reset";
7730 		break;
7731 	case MPI2_EVENT_SAS_DEV_STAT_RC_TASK_ABORT_INTERNAL:
7732 		reason_str = "internal task abort";
7733 		break;
7734 	case MPI2_EVENT_SAS_DEV_STAT_RC_ABORT_TASK_SET_INTERNAL:
7735 		reason_str = "internal task abort set";
7736 		break;
7737 	case MPI2_EVENT_SAS_DEV_STAT_RC_CLEAR_TASK_SET_INTERNAL:
7738 		reason_str = "internal clear task set";
7739 		break;
7740 	case MPI2_EVENT_SAS_DEV_STAT_RC_QUERY_TASK_INTERNAL:
7741 		reason_str = "internal query task";
7742 		break;
7743 	case MPI2_EVENT_SAS_DEV_STAT_RC_SATA_INIT_FAILURE:
7744 		reason_str = "sata init failure";
7745 		break;
7746 	case MPI2_EVENT_SAS_DEV_STAT_RC_CMP_INTERNAL_DEV_RESET:
7747 		reason_str = "internal device reset complete";
7748 		break;
7749 	case MPI2_EVENT_SAS_DEV_STAT_RC_CMP_TASK_ABORT_INTERNAL:
7750 		reason_str = "internal task abort complete";
7751 		break;
7752 	case MPI2_EVENT_SAS_DEV_STAT_RC_ASYNC_NOTIFICATION:
7753 		reason_str = "internal async notification";
7754 		break;
7755 	case MPI2_EVENT_SAS_DEV_STAT_RC_EXPANDER_REDUCED_FUNCTIONALITY:
7756 		reason_str = "expander reduced functionality";
7757 		break;
7758 	case MPI2_EVENT_SAS_DEV_STAT_RC_CMP_EXPANDER_REDUCED_FUNCTIONALITY:
7759 		reason_str = "expander reduced functionality complete";
7760 		break;
7761 	default:
7762 		reason_str = "unknown reason";
7763 		break;
7764 	}
7765 	ioc_info(ioc, "device status change: (%s)\thandle(0x%04x), sas address(0x%016llx), tag(%d)",
7766 		 reason_str, le16_to_cpu(event_data->DevHandle),
7767 		 (u64)le64_to_cpu(event_data->SASAddress),
7768 		 le16_to_cpu(event_data->TaskTag));
7769 	if (event_data->ReasonCode == MPI2_EVENT_SAS_DEV_STAT_RC_SMART_DATA)
7770 		pr_cont(", ASC(0x%x), ASCQ(0x%x)\n",
7771 			event_data->ASC, event_data->ASCQ);
7772 	pr_cont("\n");
7773 }
7774 
7775 /**
7776  * _scsih_sas_device_status_change_event - handle device status change
7777  * @ioc: per adapter object
7778  * @event_data: The fw event
7779  * Context: user.
7780  */
7781 static void
7782 _scsih_sas_device_status_change_event(struct MPT3SAS_ADAPTER *ioc,
7783 	Mpi2EventDataSasDeviceStatusChange_t *event_data)
7784 {
7785 	struct MPT3SAS_TARGET *target_priv_data;
7786 	struct _sas_device *sas_device;
7787 	u64 sas_address;
7788 	unsigned long flags;
7789 
7790 	/* In MPI Revision K (0xC), the internal device reset complete was
7791 	 * implemented, so avoid setting tm_busy flag for older firmware.
7792 	 */
7793 	if ((ioc->facts.HeaderVersion >> 8) < 0xC)
7794 		return;
7795 
7796 	if (event_data->ReasonCode !=
7797 	    MPI2_EVENT_SAS_DEV_STAT_RC_INTERNAL_DEVICE_RESET &&
7798 	   event_data->ReasonCode !=
7799 	    MPI2_EVENT_SAS_DEV_STAT_RC_CMP_INTERNAL_DEV_RESET)
7800 		return;
7801 
7802 	spin_lock_irqsave(&ioc->sas_device_lock, flags);
7803 	sas_address = le64_to_cpu(event_data->SASAddress);
7804 	sas_device = __mpt3sas_get_sdev_by_addr(ioc,
7805 	    sas_address,
7806 	    mpt3sas_get_port_by_id(ioc, event_data->PhysicalPort, 0));
7807 
7808 	if (!sas_device || !sas_device->starget)
7809 		goto out;
7810 
7811 	target_priv_data = sas_device->starget->hostdata;
7812 	if (!target_priv_data)
7813 		goto out;
7814 
7815 	if (event_data->ReasonCode ==
7816 	    MPI2_EVENT_SAS_DEV_STAT_RC_INTERNAL_DEVICE_RESET)
7817 		target_priv_data->tm_busy = 1;
7818 	else
7819 		target_priv_data->tm_busy = 0;
7820 
7821 	if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK)
7822 		ioc_info(ioc,
7823 		    "%s tm_busy flag for handle(0x%04x)\n",
7824 		    (target_priv_data->tm_busy == 1) ? "Enable" : "Disable",
7825 		    target_priv_data->handle);
7826 
7827 out:
7828 	if (sas_device)
7829 		sas_device_put(sas_device);
7830 
7831 	spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
7832 }
7833 
7834 
7835 /**
7836  * _scsih_check_pcie_access_status - check access flags
7837  * @ioc: per adapter object
7838  * @wwid: wwid
7839  * @handle: sas device handle
7840  * @access_status: errors returned during discovery of the device
7841  *
7842  * Return: 0 for success, else failure
7843  */
7844 static u8
7845 _scsih_check_pcie_access_status(struct MPT3SAS_ADAPTER *ioc, u64 wwid,
7846 	u16 handle, u8 access_status)
7847 {
7848 	u8 rc = 1;
7849 	char *desc = NULL;
7850 
7851 	switch (access_status) {
7852 	case MPI26_PCIEDEV0_ASTATUS_NO_ERRORS:
7853 	case MPI26_PCIEDEV0_ASTATUS_NEEDS_INITIALIZATION:
7854 		rc = 0;
7855 		break;
7856 	case MPI26_PCIEDEV0_ASTATUS_CAPABILITY_FAILED:
7857 		desc = "PCIe device capability failed";
7858 		break;
7859 	case MPI26_PCIEDEV0_ASTATUS_DEVICE_BLOCKED:
7860 		desc = "PCIe device blocked";
7861 		ioc_info(ioc,
7862 		    "Device with Access Status (%s): wwid(0x%016llx), "
7863 		    "handle(0x%04x)\n ll only be added to the internal list",
7864 		    desc, (u64)wwid, handle);
7865 		rc = 0;
7866 		break;
7867 	case MPI26_PCIEDEV0_ASTATUS_MEMORY_SPACE_ACCESS_FAILED:
7868 		desc = "PCIe device mem space access failed";
7869 		break;
7870 	case MPI26_PCIEDEV0_ASTATUS_UNSUPPORTED_DEVICE:
7871 		desc = "PCIe device unsupported";
7872 		break;
7873 	case MPI26_PCIEDEV0_ASTATUS_MSIX_REQUIRED:
7874 		desc = "PCIe device MSIx Required";
7875 		break;
7876 	case MPI26_PCIEDEV0_ASTATUS_INIT_FAIL_MAX:
7877 		desc = "PCIe device init fail max";
7878 		break;
7879 	case MPI26_PCIEDEV0_ASTATUS_UNKNOWN:
7880 		desc = "PCIe device status unknown";
7881 		break;
7882 	case MPI26_PCIEDEV0_ASTATUS_NVME_READY_TIMEOUT:
7883 		desc = "nvme ready timeout";
7884 		break;
7885 	case MPI26_PCIEDEV0_ASTATUS_NVME_DEVCFG_UNSUPPORTED:
7886 		desc = "nvme device configuration unsupported";
7887 		break;
7888 	case MPI26_PCIEDEV0_ASTATUS_NVME_IDENTIFY_FAILED:
7889 		desc = "nvme identify failed";
7890 		break;
7891 	case MPI26_PCIEDEV0_ASTATUS_NVME_QCONFIG_FAILED:
7892 		desc = "nvme qconfig failed";
7893 		break;
7894 	case MPI26_PCIEDEV0_ASTATUS_NVME_QCREATION_FAILED:
7895 		desc = "nvme qcreation failed";
7896 		break;
7897 	case MPI26_PCIEDEV0_ASTATUS_NVME_EVENTCFG_FAILED:
7898 		desc = "nvme eventcfg failed";
7899 		break;
7900 	case MPI26_PCIEDEV0_ASTATUS_NVME_GET_FEATURE_STAT_FAILED:
7901 		desc = "nvme get feature stat failed";
7902 		break;
7903 	case MPI26_PCIEDEV0_ASTATUS_NVME_IDLE_TIMEOUT:
7904 		desc = "nvme idle timeout";
7905 		break;
7906 	case MPI26_PCIEDEV0_ASTATUS_NVME_FAILURE_STATUS:
7907 		desc = "nvme failure status";
7908 		break;
7909 	default:
7910 		ioc_err(ioc, "NVMe discovery error(0x%02x): wwid(0x%016llx), handle(0x%04x)\n",
7911 			access_status, (u64)wwid, handle);
7912 		return rc;
7913 	}
7914 
7915 	if (!rc)
7916 		return rc;
7917 
7918 	ioc_info(ioc, "NVMe discovery error(%s): wwid(0x%016llx), handle(0x%04x)\n",
7919 		 desc, (u64)wwid, handle);
7920 	return rc;
7921 }
7922 
7923 /**
7924  * _scsih_pcie_device_remove_from_sml -  removing pcie device
7925  * from SML and free up associated memory
7926  * @ioc: per adapter object
7927  * @pcie_device: the pcie_device object
7928  */
7929 static void
7930 _scsih_pcie_device_remove_from_sml(struct MPT3SAS_ADAPTER *ioc,
7931 	struct _pcie_device *pcie_device)
7932 {
7933 	struct MPT3SAS_TARGET *sas_target_priv_data;
7934 
7935 	dewtprintk(ioc,
7936 		   ioc_info(ioc, "%s: enter: handle(0x%04x), wwid(0x%016llx)\n",
7937 			    __func__,
7938 			    pcie_device->handle, (u64)pcie_device->wwid));
7939 	if (pcie_device->enclosure_handle != 0)
7940 		dewtprintk(ioc,
7941 			   ioc_info(ioc, "%s: enter: enclosure logical id(0x%016llx), slot(%d)\n",
7942 				    __func__,
7943 				    (u64)pcie_device->enclosure_logical_id,
7944 				    pcie_device->slot));
7945 	if (pcie_device->connector_name[0] != '\0')
7946 		dewtprintk(ioc,
7947 			   ioc_info(ioc, "%s: enter: enclosure level(0x%04x), connector name(%s)\n",
7948 				    __func__,
7949 				    pcie_device->enclosure_level,
7950 				    pcie_device->connector_name));
7951 
7952 	if (pcie_device->starget && pcie_device->starget->hostdata) {
7953 		sas_target_priv_data = pcie_device->starget->hostdata;
7954 		sas_target_priv_data->deleted = 1;
7955 		_scsih_ublock_io_device(ioc, pcie_device->wwid, NULL);
7956 		sas_target_priv_data->handle = MPT3SAS_INVALID_DEVICE_HANDLE;
7957 	}
7958 
7959 	ioc_info(ioc, "removing handle(0x%04x), wwid(0x%016llx)\n",
7960 		 pcie_device->handle, (u64)pcie_device->wwid);
7961 	if (pcie_device->enclosure_handle != 0)
7962 		ioc_info(ioc, "removing : enclosure logical id(0x%016llx), slot(%d)\n",
7963 			 (u64)pcie_device->enclosure_logical_id,
7964 			 pcie_device->slot);
7965 	if (pcie_device->connector_name[0] != '\0')
7966 		ioc_info(ioc, "removing: enclosure level(0x%04x), connector name( %s)\n",
7967 			 pcie_device->enclosure_level,
7968 			 pcie_device->connector_name);
7969 
7970 	if (pcie_device->starget && (pcie_device->access_status !=
7971 				MPI26_PCIEDEV0_ASTATUS_DEVICE_BLOCKED))
7972 		scsi_remove_target(&pcie_device->starget->dev);
7973 	dewtprintk(ioc,
7974 		   ioc_info(ioc, "%s: exit: handle(0x%04x), wwid(0x%016llx)\n",
7975 			    __func__,
7976 			    pcie_device->handle, (u64)pcie_device->wwid));
7977 	if (pcie_device->enclosure_handle != 0)
7978 		dewtprintk(ioc,
7979 			   ioc_info(ioc, "%s: exit: enclosure logical id(0x%016llx), slot(%d)\n",
7980 				    __func__,
7981 				    (u64)pcie_device->enclosure_logical_id,
7982 				    pcie_device->slot));
7983 	if (pcie_device->connector_name[0] != '\0')
7984 		dewtprintk(ioc,
7985 			   ioc_info(ioc, "%s: exit: enclosure level(0x%04x), connector name( %s)\n",
7986 				    __func__,
7987 				    pcie_device->enclosure_level,
7988 				    pcie_device->connector_name));
7989 
7990 	kfree(pcie_device->serial_number);
7991 }
7992 
7993 
7994 /**
7995  * _scsih_pcie_check_device - checking device responsiveness
7996  * @ioc: per adapter object
7997  * @handle: attached device handle
7998  */
7999 static void
8000 _scsih_pcie_check_device(struct MPT3SAS_ADAPTER *ioc, u16 handle)
8001 {
8002 	Mpi2ConfigReply_t mpi_reply;
8003 	Mpi26PCIeDevicePage0_t pcie_device_pg0;
8004 	u32 ioc_status;
8005 	struct _pcie_device *pcie_device;
8006 	u64 wwid;
8007 	unsigned long flags;
8008 	struct scsi_target *starget;
8009 	struct MPT3SAS_TARGET *sas_target_priv_data;
8010 	u32 device_info;
8011 
8012 	if ((mpt3sas_config_get_pcie_device_pg0(ioc, &mpi_reply,
8013 		&pcie_device_pg0, MPI26_PCIE_DEVICE_PGAD_FORM_HANDLE, handle)))
8014 		return;
8015 
8016 	ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK;
8017 	if (ioc_status != MPI2_IOCSTATUS_SUCCESS)
8018 		return;
8019 
8020 	/* check if this is end device */
8021 	device_info = le32_to_cpu(pcie_device_pg0.DeviceInfo);
8022 	if (!(_scsih_is_nvme_pciescsi_device(device_info)))
8023 		return;
8024 
8025 	wwid = le64_to_cpu(pcie_device_pg0.WWID);
8026 	spin_lock_irqsave(&ioc->pcie_device_lock, flags);
8027 	pcie_device = __mpt3sas_get_pdev_by_wwid(ioc, wwid);
8028 
8029 	if (!pcie_device) {
8030 		spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
8031 		return;
8032 	}
8033 
8034 	if (unlikely(pcie_device->handle != handle)) {
8035 		starget = pcie_device->starget;
8036 		sas_target_priv_data = starget->hostdata;
8037 		pcie_device->access_status = pcie_device_pg0.AccessStatus;
8038 		starget_printk(KERN_INFO, starget,
8039 		    "handle changed from(0x%04x) to (0x%04x)!!!\n",
8040 		    pcie_device->handle, handle);
8041 		sas_target_priv_data->handle = handle;
8042 		pcie_device->handle = handle;
8043 
8044 		if (le32_to_cpu(pcie_device_pg0.Flags) &
8045 		    MPI26_PCIEDEV0_FLAGS_ENCL_LEVEL_VALID) {
8046 			pcie_device->enclosure_level =
8047 			    pcie_device_pg0.EnclosureLevel;
8048 			memcpy(&pcie_device->connector_name[0],
8049 			    &pcie_device_pg0.ConnectorName[0], 4);
8050 		} else {
8051 			pcie_device->enclosure_level = 0;
8052 			pcie_device->connector_name[0] = '\0';
8053 		}
8054 	}
8055 
8056 	/* check if device is present */
8057 	if (!(le32_to_cpu(pcie_device_pg0.Flags) &
8058 	    MPI26_PCIEDEV0_FLAGS_DEVICE_PRESENT)) {
8059 		ioc_info(ioc, "device is not present handle(0x%04x), flags!!!\n",
8060 			 handle);
8061 		spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
8062 		pcie_device_put(pcie_device);
8063 		return;
8064 	}
8065 
8066 	/* check if there were any issues with discovery */
8067 	if (_scsih_check_pcie_access_status(ioc, wwid, handle,
8068 	    pcie_device_pg0.AccessStatus)) {
8069 		spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
8070 		pcie_device_put(pcie_device);
8071 		return;
8072 	}
8073 
8074 	spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
8075 	pcie_device_put(pcie_device);
8076 
8077 	_scsih_ublock_io_device(ioc, wwid, NULL);
8078 
8079 	return;
8080 }
8081 
8082 /**
8083  * _scsih_pcie_add_device -  creating pcie device object
8084  * @ioc: per adapter object
8085  * @handle: pcie device handle
8086  *
8087  * Creating end device object, stored in ioc->pcie_device_list.
8088  *
8089  * Return: 1 means queue the event later, 0 means complete the event
8090  */
8091 static int
8092 _scsih_pcie_add_device(struct MPT3SAS_ADAPTER *ioc, u16 handle)
8093 {
8094 	Mpi26PCIeDevicePage0_t pcie_device_pg0;
8095 	Mpi26PCIeDevicePage2_t pcie_device_pg2;
8096 	Mpi2ConfigReply_t mpi_reply;
8097 	struct _pcie_device *pcie_device;
8098 	struct _enclosure_node *enclosure_dev;
8099 	u32 ioc_status;
8100 	u64 wwid;
8101 
8102 	if ((mpt3sas_config_get_pcie_device_pg0(ioc, &mpi_reply,
8103 	    &pcie_device_pg0, MPI26_PCIE_DEVICE_PGAD_FORM_HANDLE, handle))) {
8104 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
8105 			__FILE__, __LINE__, __func__);
8106 		return 0;
8107 	}
8108 	ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
8109 	    MPI2_IOCSTATUS_MASK;
8110 	if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
8111 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
8112 			__FILE__, __LINE__, __func__);
8113 		return 0;
8114 	}
8115 
8116 	set_bit(handle, ioc->pend_os_device_add);
8117 	wwid = le64_to_cpu(pcie_device_pg0.WWID);
8118 
8119 	/* check if device is present */
8120 	if (!(le32_to_cpu(pcie_device_pg0.Flags) &
8121 		MPI26_PCIEDEV0_FLAGS_DEVICE_PRESENT)) {
8122 		ioc_err(ioc, "device is not present handle(0x04%x)!!!\n",
8123 			handle);
8124 		return 0;
8125 	}
8126 
8127 	/* check if there were any issues with discovery */
8128 	if (_scsih_check_pcie_access_status(ioc, wwid, handle,
8129 	    pcie_device_pg0.AccessStatus))
8130 		return 0;
8131 
8132 	if (!(_scsih_is_nvme_pciescsi_device(le32_to_cpu
8133 	    (pcie_device_pg0.DeviceInfo))))
8134 		return 0;
8135 
8136 	pcie_device = mpt3sas_get_pdev_by_wwid(ioc, wwid);
8137 	if (pcie_device) {
8138 		clear_bit(handle, ioc->pend_os_device_add);
8139 		pcie_device_put(pcie_device);
8140 		return 0;
8141 	}
8142 
8143 	/* PCIe Device Page 2 contains read-only information about a
8144 	 * specific NVMe device; therefore, this page is only
8145 	 * valid for NVMe devices and skip for pcie devices of type scsi.
8146 	 */
8147 	if (!(mpt3sas_scsih_is_pcie_scsi_device(
8148 		le32_to_cpu(pcie_device_pg0.DeviceInfo)))) {
8149 		if (mpt3sas_config_get_pcie_device_pg2(ioc, &mpi_reply,
8150 		    &pcie_device_pg2, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE,
8151 		    handle)) {
8152 			ioc_err(ioc,
8153 			    "failure at %s:%d/%s()!\n", __FILE__,
8154 			    __LINE__, __func__);
8155 			return 0;
8156 		}
8157 
8158 		ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
8159 					MPI2_IOCSTATUS_MASK;
8160 		if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
8161 			ioc_err(ioc,
8162 			    "failure at %s:%d/%s()!\n", __FILE__,
8163 			    __LINE__, __func__);
8164 			return 0;
8165 		}
8166 	}
8167 
8168 	pcie_device = kzalloc(sizeof(struct _pcie_device), GFP_KERNEL);
8169 	if (!pcie_device) {
8170 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
8171 			__FILE__, __LINE__, __func__);
8172 		return 0;
8173 	}
8174 
8175 	kref_init(&pcie_device->refcount);
8176 	pcie_device->id = ioc->pcie_target_id++;
8177 	pcie_device->channel = PCIE_CHANNEL;
8178 	pcie_device->handle = handle;
8179 	pcie_device->access_status = pcie_device_pg0.AccessStatus;
8180 	pcie_device->device_info = le32_to_cpu(pcie_device_pg0.DeviceInfo);
8181 	pcie_device->wwid = wwid;
8182 	pcie_device->port_num = pcie_device_pg0.PortNum;
8183 	pcie_device->fast_path = (le32_to_cpu(pcie_device_pg0.Flags) &
8184 	    MPI26_PCIEDEV0_FLAGS_FAST_PATH_CAPABLE) ? 1 : 0;
8185 
8186 	pcie_device->enclosure_handle =
8187 	    le16_to_cpu(pcie_device_pg0.EnclosureHandle);
8188 	if (pcie_device->enclosure_handle != 0)
8189 		pcie_device->slot = le16_to_cpu(pcie_device_pg0.Slot);
8190 
8191 	if (le32_to_cpu(pcie_device_pg0.Flags) &
8192 	    MPI26_PCIEDEV0_FLAGS_ENCL_LEVEL_VALID) {
8193 		pcie_device->enclosure_level = pcie_device_pg0.EnclosureLevel;
8194 		memcpy(&pcie_device->connector_name[0],
8195 		    &pcie_device_pg0.ConnectorName[0], 4);
8196 	} else {
8197 		pcie_device->enclosure_level = 0;
8198 		pcie_device->connector_name[0] = '\0';
8199 	}
8200 
8201 	/* get enclosure_logical_id */
8202 	if (pcie_device->enclosure_handle) {
8203 		enclosure_dev =
8204 			mpt3sas_scsih_enclosure_find_by_handle(ioc,
8205 						pcie_device->enclosure_handle);
8206 		if (enclosure_dev)
8207 			pcie_device->enclosure_logical_id =
8208 			    le64_to_cpu(enclosure_dev->pg0.EnclosureLogicalID);
8209 	}
8210 	/* TODO -- Add device name once FW supports it */
8211 	if (!(mpt3sas_scsih_is_pcie_scsi_device(
8212 	    le32_to_cpu(pcie_device_pg0.DeviceInfo)))) {
8213 		pcie_device->nvme_mdts =
8214 		    le32_to_cpu(pcie_device_pg2.MaximumDataTransferSize);
8215 		pcie_device->shutdown_latency =
8216 			le16_to_cpu(pcie_device_pg2.ShutdownLatency);
8217 		/*
8218 		 * Set IOC's max_shutdown_latency to drive's RTD3 Entry Latency
8219 		 * if drive's RTD3 Entry Latency is greater then IOC's
8220 		 * max_shutdown_latency.
8221 		 */
8222 		if (pcie_device->shutdown_latency > ioc->max_shutdown_latency)
8223 			ioc->max_shutdown_latency =
8224 				pcie_device->shutdown_latency;
8225 		if (pcie_device_pg2.ControllerResetTO)
8226 			pcie_device->reset_timeout =
8227 			    pcie_device_pg2.ControllerResetTO;
8228 		else
8229 			pcie_device->reset_timeout = 30;
8230 	} else
8231 		pcie_device->reset_timeout = 30;
8232 
8233 	if (ioc->wait_for_discovery_to_complete)
8234 		_scsih_pcie_device_init_add(ioc, pcie_device);
8235 	else
8236 		_scsih_pcie_device_add(ioc, pcie_device);
8237 
8238 	pcie_device_put(pcie_device);
8239 	return 0;
8240 }
8241 
8242 /**
8243  * _scsih_pcie_topology_change_event_debug - debug for topology
8244  * event
8245  * @ioc: per adapter object
8246  * @event_data: event data payload
8247  * Context: user.
8248  */
8249 static void
8250 _scsih_pcie_topology_change_event_debug(struct MPT3SAS_ADAPTER *ioc,
8251 	Mpi26EventDataPCIeTopologyChangeList_t *event_data)
8252 {
8253 	int i;
8254 	u16 handle;
8255 	u16 reason_code;
8256 	u8 port_number;
8257 	char *status_str = NULL;
8258 	u8 link_rate, prev_link_rate;
8259 
8260 	switch (event_data->SwitchStatus) {
8261 	case MPI26_EVENT_PCIE_TOPO_SS_ADDED:
8262 		status_str = "add";
8263 		break;
8264 	case MPI26_EVENT_PCIE_TOPO_SS_NOT_RESPONDING:
8265 		status_str = "remove";
8266 		break;
8267 	case MPI26_EVENT_PCIE_TOPO_SS_RESPONDING:
8268 	case 0:
8269 		status_str =  "responding";
8270 		break;
8271 	case MPI26_EVENT_PCIE_TOPO_SS_DELAY_NOT_RESPONDING:
8272 		status_str = "remove delay";
8273 		break;
8274 	default:
8275 		status_str = "unknown status";
8276 		break;
8277 	}
8278 	ioc_info(ioc, "pcie topology change: (%s)\n", status_str);
8279 	pr_info("\tswitch_handle(0x%04x), enclosure_handle(0x%04x)"
8280 		"start_port(%02d), count(%d)\n",
8281 		le16_to_cpu(event_data->SwitchDevHandle),
8282 		le16_to_cpu(event_data->EnclosureHandle),
8283 		event_data->StartPortNum, event_data->NumEntries);
8284 	for (i = 0; i < event_data->NumEntries; i++) {
8285 		handle =
8286 			le16_to_cpu(event_data->PortEntry[i].AttachedDevHandle);
8287 		if (!handle)
8288 			continue;
8289 		port_number = event_data->StartPortNum + i;
8290 		reason_code = event_data->PortEntry[i].PortStatus;
8291 		switch (reason_code) {
8292 		case MPI26_EVENT_PCIE_TOPO_PS_DEV_ADDED:
8293 			status_str = "target add";
8294 			break;
8295 		case MPI26_EVENT_PCIE_TOPO_PS_NOT_RESPONDING:
8296 			status_str = "target remove";
8297 			break;
8298 		case MPI26_EVENT_PCIE_TOPO_PS_DELAY_NOT_RESPONDING:
8299 			status_str = "delay target remove";
8300 			break;
8301 		case MPI26_EVENT_PCIE_TOPO_PS_PORT_CHANGED:
8302 			status_str = "link rate change";
8303 			break;
8304 		case MPI26_EVENT_PCIE_TOPO_PS_NO_CHANGE:
8305 			status_str = "target responding";
8306 			break;
8307 		default:
8308 			status_str = "unknown";
8309 			break;
8310 		}
8311 		link_rate = event_data->PortEntry[i].CurrentPortInfo &
8312 			MPI26_EVENT_PCIE_TOPO_PI_RATE_MASK;
8313 		prev_link_rate = event_data->PortEntry[i].PreviousPortInfo &
8314 			MPI26_EVENT_PCIE_TOPO_PI_RATE_MASK;
8315 		pr_info("\tport(%02d), attached_handle(0x%04x): %s:"
8316 			" link rate: new(0x%02x), old(0x%02x)\n", port_number,
8317 			handle, status_str, link_rate, prev_link_rate);
8318 	}
8319 }
8320 
8321 /**
8322  * _scsih_pcie_topology_change_event - handle PCIe topology
8323  *  changes
8324  * @ioc: per adapter object
8325  * @fw_event: The fw_event_work object
8326  * Context: user.
8327  *
8328  */
8329 static void
8330 _scsih_pcie_topology_change_event(struct MPT3SAS_ADAPTER *ioc,
8331 	struct fw_event_work *fw_event)
8332 {
8333 	int i;
8334 	u16 handle;
8335 	u16 reason_code;
8336 	u8 link_rate, prev_link_rate;
8337 	unsigned long flags;
8338 	int rc;
8339 	Mpi26EventDataPCIeTopologyChangeList_t *event_data =
8340 		(Mpi26EventDataPCIeTopologyChangeList_t *) fw_event->event_data;
8341 	struct _pcie_device *pcie_device;
8342 
8343 	if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK)
8344 		_scsih_pcie_topology_change_event_debug(ioc, event_data);
8345 
8346 	if (ioc->shost_recovery || ioc->remove_host ||
8347 		ioc->pci_error_recovery)
8348 		return;
8349 
8350 	if (fw_event->ignore) {
8351 		dewtprintk(ioc, ioc_info(ioc, "ignoring switch event\n"));
8352 		return;
8353 	}
8354 
8355 	/* handle siblings events */
8356 	for (i = 0; i < event_data->NumEntries; i++) {
8357 		if (fw_event->ignore) {
8358 			dewtprintk(ioc,
8359 				   ioc_info(ioc, "ignoring switch event\n"));
8360 			return;
8361 		}
8362 		if (ioc->remove_host || ioc->pci_error_recovery)
8363 			return;
8364 		reason_code = event_data->PortEntry[i].PortStatus;
8365 		handle =
8366 			le16_to_cpu(event_data->PortEntry[i].AttachedDevHandle);
8367 		if (!handle)
8368 			continue;
8369 
8370 		link_rate = event_data->PortEntry[i].CurrentPortInfo
8371 			& MPI26_EVENT_PCIE_TOPO_PI_RATE_MASK;
8372 		prev_link_rate = event_data->PortEntry[i].PreviousPortInfo
8373 			& MPI26_EVENT_PCIE_TOPO_PI_RATE_MASK;
8374 
8375 		switch (reason_code) {
8376 		case MPI26_EVENT_PCIE_TOPO_PS_PORT_CHANGED:
8377 			if (ioc->shost_recovery)
8378 				break;
8379 			if (link_rate == prev_link_rate)
8380 				break;
8381 			if (link_rate < MPI26_EVENT_PCIE_TOPO_PI_RATE_2_5)
8382 				break;
8383 
8384 			_scsih_pcie_check_device(ioc, handle);
8385 
8386 			/* This code after this point handles the test case
8387 			 * where a device has been added, however its returning
8388 			 * BUSY for sometime.  Then before the Device Missing
8389 			 * Delay expires and the device becomes READY, the
8390 			 * device is removed and added back.
8391 			 */
8392 			spin_lock_irqsave(&ioc->pcie_device_lock, flags);
8393 			pcie_device = __mpt3sas_get_pdev_by_handle(ioc, handle);
8394 			spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
8395 
8396 			if (pcie_device) {
8397 				pcie_device_put(pcie_device);
8398 				break;
8399 			}
8400 
8401 			if (!test_bit(handle, ioc->pend_os_device_add))
8402 				break;
8403 
8404 			dewtprintk(ioc,
8405 				   ioc_info(ioc, "handle(0x%04x) device not found: convert event to a device add\n",
8406 					    handle));
8407 			event_data->PortEntry[i].PortStatus &= 0xF0;
8408 			event_data->PortEntry[i].PortStatus |=
8409 				MPI26_EVENT_PCIE_TOPO_PS_DEV_ADDED;
8410 			fallthrough;
8411 		case MPI26_EVENT_PCIE_TOPO_PS_DEV_ADDED:
8412 			if (ioc->shost_recovery)
8413 				break;
8414 			if (link_rate < MPI26_EVENT_PCIE_TOPO_PI_RATE_2_5)
8415 				break;
8416 
8417 			rc = _scsih_pcie_add_device(ioc, handle);
8418 			if (!rc) {
8419 				/* mark entry vacant */
8420 				/* TODO This needs to be reviewed and fixed,
8421 				 * we dont have an entry
8422 				 * to make an event void like vacant
8423 				 */
8424 				event_data->PortEntry[i].PortStatus |=
8425 					MPI26_EVENT_PCIE_TOPO_PS_NO_CHANGE;
8426 			}
8427 			break;
8428 		case MPI26_EVENT_PCIE_TOPO_PS_NOT_RESPONDING:
8429 			_scsih_pcie_device_remove_by_handle(ioc, handle);
8430 			break;
8431 		}
8432 	}
8433 }
8434 
8435 /**
8436  * _scsih_pcie_device_status_change_event_debug - debug for device event
8437  * @ioc: ?
8438  * @event_data: event data payload
8439  * Context: user.
8440  */
8441 static void
8442 _scsih_pcie_device_status_change_event_debug(struct MPT3SAS_ADAPTER *ioc,
8443 	Mpi26EventDataPCIeDeviceStatusChange_t *event_data)
8444 {
8445 	char *reason_str = NULL;
8446 
8447 	switch (event_data->ReasonCode) {
8448 	case MPI26_EVENT_PCIDEV_STAT_RC_SMART_DATA:
8449 		reason_str = "smart data";
8450 		break;
8451 	case MPI26_EVENT_PCIDEV_STAT_RC_UNSUPPORTED:
8452 		reason_str = "unsupported device discovered";
8453 		break;
8454 	case MPI26_EVENT_PCIDEV_STAT_RC_INTERNAL_DEVICE_RESET:
8455 		reason_str = "internal device reset";
8456 		break;
8457 	case MPI26_EVENT_PCIDEV_STAT_RC_TASK_ABORT_INTERNAL:
8458 		reason_str = "internal task abort";
8459 		break;
8460 	case MPI26_EVENT_PCIDEV_STAT_RC_ABORT_TASK_SET_INTERNAL:
8461 		reason_str = "internal task abort set";
8462 		break;
8463 	case MPI26_EVENT_PCIDEV_STAT_RC_CLEAR_TASK_SET_INTERNAL:
8464 		reason_str = "internal clear task set";
8465 		break;
8466 	case MPI26_EVENT_PCIDEV_STAT_RC_QUERY_TASK_INTERNAL:
8467 		reason_str = "internal query task";
8468 		break;
8469 	case MPI26_EVENT_PCIDEV_STAT_RC_DEV_INIT_FAILURE:
8470 		reason_str = "device init failure";
8471 		break;
8472 	case MPI26_EVENT_PCIDEV_STAT_RC_CMP_INTERNAL_DEV_RESET:
8473 		reason_str = "internal device reset complete";
8474 		break;
8475 	case MPI26_EVENT_PCIDEV_STAT_RC_CMP_TASK_ABORT_INTERNAL:
8476 		reason_str = "internal task abort complete";
8477 		break;
8478 	case MPI26_EVENT_PCIDEV_STAT_RC_ASYNC_NOTIFICATION:
8479 		reason_str = "internal async notification";
8480 		break;
8481 	case MPI26_EVENT_PCIDEV_STAT_RC_PCIE_HOT_RESET_FAILED:
8482 		reason_str = "pcie hot reset failed";
8483 		break;
8484 	default:
8485 		reason_str = "unknown reason";
8486 		break;
8487 	}
8488 
8489 	ioc_info(ioc, "PCIE device status change: (%s)\n"
8490 		 "\thandle(0x%04x), WWID(0x%016llx), tag(%d)",
8491 		 reason_str, le16_to_cpu(event_data->DevHandle),
8492 		 (u64)le64_to_cpu(event_data->WWID),
8493 		 le16_to_cpu(event_data->TaskTag));
8494 	if (event_data->ReasonCode == MPI26_EVENT_PCIDEV_STAT_RC_SMART_DATA)
8495 		pr_cont(", ASC(0x%x), ASCQ(0x%x)\n",
8496 			event_data->ASC, event_data->ASCQ);
8497 	pr_cont("\n");
8498 }
8499 
8500 /**
8501  * _scsih_pcie_device_status_change_event - handle device status
8502  * change
8503  * @ioc: per adapter object
8504  * @fw_event: The fw_event_work object
8505  * Context: user.
8506  */
8507 static void
8508 _scsih_pcie_device_status_change_event(struct MPT3SAS_ADAPTER *ioc,
8509 	struct fw_event_work *fw_event)
8510 {
8511 	struct MPT3SAS_TARGET *target_priv_data;
8512 	struct _pcie_device *pcie_device;
8513 	u64 wwid;
8514 	unsigned long flags;
8515 	Mpi26EventDataPCIeDeviceStatusChange_t *event_data =
8516 		(Mpi26EventDataPCIeDeviceStatusChange_t *)fw_event->event_data;
8517 	if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK)
8518 		_scsih_pcie_device_status_change_event_debug(ioc,
8519 			event_data);
8520 
8521 	if (event_data->ReasonCode !=
8522 		MPI26_EVENT_PCIDEV_STAT_RC_INTERNAL_DEVICE_RESET &&
8523 		event_data->ReasonCode !=
8524 		MPI26_EVENT_PCIDEV_STAT_RC_CMP_INTERNAL_DEV_RESET)
8525 		return;
8526 
8527 	spin_lock_irqsave(&ioc->pcie_device_lock, flags);
8528 	wwid = le64_to_cpu(event_data->WWID);
8529 	pcie_device = __mpt3sas_get_pdev_by_wwid(ioc, wwid);
8530 
8531 	if (!pcie_device || !pcie_device->starget)
8532 		goto out;
8533 
8534 	target_priv_data = pcie_device->starget->hostdata;
8535 	if (!target_priv_data)
8536 		goto out;
8537 
8538 	if (event_data->ReasonCode ==
8539 		MPI26_EVENT_PCIDEV_STAT_RC_INTERNAL_DEVICE_RESET)
8540 		target_priv_data->tm_busy = 1;
8541 	else
8542 		target_priv_data->tm_busy = 0;
8543 out:
8544 	if (pcie_device)
8545 		pcie_device_put(pcie_device);
8546 
8547 	spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
8548 }
8549 
8550 /**
8551  * _scsih_sas_enclosure_dev_status_change_event_debug - debug for enclosure
8552  * event
8553  * @ioc: per adapter object
8554  * @event_data: event data payload
8555  * Context: user.
8556  */
8557 static void
8558 _scsih_sas_enclosure_dev_status_change_event_debug(struct MPT3SAS_ADAPTER *ioc,
8559 	Mpi2EventDataSasEnclDevStatusChange_t *event_data)
8560 {
8561 	char *reason_str = NULL;
8562 
8563 	switch (event_data->ReasonCode) {
8564 	case MPI2_EVENT_SAS_ENCL_RC_ADDED:
8565 		reason_str = "enclosure add";
8566 		break;
8567 	case MPI2_EVENT_SAS_ENCL_RC_NOT_RESPONDING:
8568 		reason_str = "enclosure remove";
8569 		break;
8570 	default:
8571 		reason_str = "unknown reason";
8572 		break;
8573 	}
8574 
8575 	ioc_info(ioc, "enclosure status change: (%s)\n"
8576 		 "\thandle(0x%04x), enclosure logical id(0x%016llx) number slots(%d)\n",
8577 		 reason_str,
8578 		 le16_to_cpu(event_data->EnclosureHandle),
8579 		 (u64)le64_to_cpu(event_data->EnclosureLogicalID),
8580 		 le16_to_cpu(event_data->StartSlot));
8581 }
8582 
8583 /**
8584  * _scsih_sas_enclosure_dev_status_change_event - handle enclosure events
8585  * @ioc: per adapter object
8586  * @fw_event: The fw_event_work object
8587  * Context: user.
8588  */
8589 static void
8590 _scsih_sas_enclosure_dev_status_change_event(struct MPT3SAS_ADAPTER *ioc,
8591 	struct fw_event_work *fw_event)
8592 {
8593 	Mpi2ConfigReply_t mpi_reply;
8594 	struct _enclosure_node *enclosure_dev = NULL;
8595 	Mpi2EventDataSasEnclDevStatusChange_t *event_data =
8596 		(Mpi2EventDataSasEnclDevStatusChange_t *)fw_event->event_data;
8597 	int rc;
8598 	u16 enclosure_handle = le16_to_cpu(event_data->EnclosureHandle);
8599 
8600 	if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK)
8601 		_scsih_sas_enclosure_dev_status_change_event_debug(ioc,
8602 		     (Mpi2EventDataSasEnclDevStatusChange_t *)
8603 		     fw_event->event_data);
8604 	if (ioc->shost_recovery)
8605 		return;
8606 
8607 	if (enclosure_handle)
8608 		enclosure_dev =
8609 			mpt3sas_scsih_enclosure_find_by_handle(ioc,
8610 						enclosure_handle);
8611 	switch (event_data->ReasonCode) {
8612 	case MPI2_EVENT_SAS_ENCL_RC_ADDED:
8613 		if (!enclosure_dev) {
8614 			enclosure_dev =
8615 				kzalloc(sizeof(struct _enclosure_node),
8616 					GFP_KERNEL);
8617 			if (!enclosure_dev) {
8618 				ioc_info(ioc, "failure at %s:%d/%s()!\n",
8619 					 __FILE__, __LINE__, __func__);
8620 				return;
8621 			}
8622 			rc = mpt3sas_config_get_enclosure_pg0(ioc, &mpi_reply,
8623 				&enclosure_dev->pg0,
8624 				MPI2_SAS_ENCLOS_PGAD_FORM_HANDLE,
8625 				enclosure_handle);
8626 
8627 			if (rc || (le16_to_cpu(mpi_reply.IOCStatus) &
8628 						MPI2_IOCSTATUS_MASK)) {
8629 				kfree(enclosure_dev);
8630 				return;
8631 			}
8632 
8633 			list_add_tail(&enclosure_dev->list,
8634 							&ioc->enclosure_list);
8635 		}
8636 		break;
8637 	case MPI2_EVENT_SAS_ENCL_RC_NOT_RESPONDING:
8638 		if (enclosure_dev) {
8639 			list_del(&enclosure_dev->list);
8640 			kfree(enclosure_dev);
8641 		}
8642 		break;
8643 	default:
8644 		break;
8645 	}
8646 }
8647 
8648 /**
8649  * _scsih_sas_broadcast_primitive_event - handle broadcast events
8650  * @ioc: per adapter object
8651  * @fw_event: The fw_event_work object
8652  * Context: user.
8653  */
8654 static void
8655 _scsih_sas_broadcast_primitive_event(struct MPT3SAS_ADAPTER *ioc,
8656 	struct fw_event_work *fw_event)
8657 {
8658 	struct scsi_cmnd *scmd;
8659 	struct scsi_device *sdev;
8660 	struct scsiio_tracker *st;
8661 	u16 smid, handle;
8662 	u32 lun;
8663 	struct MPT3SAS_DEVICE *sas_device_priv_data;
8664 	u32 termination_count;
8665 	u32 query_count;
8666 	Mpi2SCSITaskManagementReply_t *mpi_reply;
8667 	Mpi2EventDataSasBroadcastPrimitive_t *event_data =
8668 		(Mpi2EventDataSasBroadcastPrimitive_t *)
8669 		fw_event->event_data;
8670 	u16 ioc_status;
8671 	unsigned long flags;
8672 	int r;
8673 	u8 max_retries = 0;
8674 	u8 task_abort_retries;
8675 
8676 	mutex_lock(&ioc->tm_cmds.mutex);
8677 	ioc_info(ioc, "%s: enter: phy number(%d), width(%d)\n",
8678 		 __func__, event_data->PhyNum, event_data->PortWidth);
8679 
8680 	_scsih_block_io_all_device(ioc);
8681 
8682 	spin_lock_irqsave(&ioc->scsi_lookup_lock, flags);
8683 	mpi_reply = ioc->tm_cmds.reply;
8684  broadcast_aen_retry:
8685 
8686 	/* sanity checks for retrying this loop */
8687 	if (max_retries++ == 5) {
8688 		dewtprintk(ioc, ioc_info(ioc, "%s: giving up\n", __func__));
8689 		goto out;
8690 	} else if (max_retries > 1)
8691 		dewtprintk(ioc,
8692 			   ioc_info(ioc, "%s: %d retry\n",
8693 				    __func__, max_retries - 1));
8694 
8695 	termination_count = 0;
8696 	query_count = 0;
8697 	for (smid = 1; smid <= ioc->scsiio_depth; smid++) {
8698 		if (ioc->shost_recovery)
8699 			goto out;
8700 		scmd = mpt3sas_scsih_scsi_lookup_get(ioc, smid);
8701 		if (!scmd)
8702 			continue;
8703 		st = scsi_cmd_priv(scmd);
8704 		sdev = scmd->device;
8705 		sas_device_priv_data = sdev->hostdata;
8706 		if (!sas_device_priv_data || !sas_device_priv_data->sas_target)
8707 			continue;
8708 		 /* skip hidden raid components */
8709 		if (sas_device_priv_data->sas_target->flags &
8710 		    MPT_TARGET_FLAGS_RAID_COMPONENT)
8711 			continue;
8712 		 /* skip volumes */
8713 		if (sas_device_priv_data->sas_target->flags &
8714 		    MPT_TARGET_FLAGS_VOLUME)
8715 			continue;
8716 		 /* skip PCIe devices */
8717 		if (sas_device_priv_data->sas_target->flags &
8718 		    MPT_TARGET_FLAGS_PCIE_DEVICE)
8719 			continue;
8720 
8721 		handle = sas_device_priv_data->sas_target->handle;
8722 		lun = sas_device_priv_data->lun;
8723 		query_count++;
8724 
8725 		if (ioc->shost_recovery)
8726 			goto out;
8727 
8728 		spin_unlock_irqrestore(&ioc->scsi_lookup_lock, flags);
8729 		r = mpt3sas_scsih_issue_tm(ioc, handle, 0, 0, lun,
8730 			MPI2_SCSITASKMGMT_TASKTYPE_QUERY_TASK, st->smid,
8731 			st->msix_io, 30, 0);
8732 		if (r == FAILED) {
8733 			sdev_printk(KERN_WARNING, sdev,
8734 			    "mpt3sas_scsih_issue_tm: FAILED when sending "
8735 			    "QUERY_TASK: scmd(%p)\n", scmd);
8736 			spin_lock_irqsave(&ioc->scsi_lookup_lock, flags);
8737 			goto broadcast_aen_retry;
8738 		}
8739 		ioc_status = le16_to_cpu(mpi_reply->IOCStatus)
8740 		    & MPI2_IOCSTATUS_MASK;
8741 		if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
8742 			sdev_printk(KERN_WARNING, sdev,
8743 				"query task: FAILED with IOCSTATUS(0x%04x), scmd(%p)\n",
8744 				ioc_status, scmd);
8745 			spin_lock_irqsave(&ioc->scsi_lookup_lock, flags);
8746 			goto broadcast_aen_retry;
8747 		}
8748 
8749 		/* see if IO is still owned by IOC and target */
8750 		if (mpi_reply->ResponseCode ==
8751 		     MPI2_SCSITASKMGMT_RSP_TM_SUCCEEDED ||
8752 		     mpi_reply->ResponseCode ==
8753 		     MPI2_SCSITASKMGMT_RSP_IO_QUEUED_ON_IOC) {
8754 			spin_lock_irqsave(&ioc->scsi_lookup_lock, flags);
8755 			continue;
8756 		}
8757 		task_abort_retries = 0;
8758  tm_retry:
8759 		if (task_abort_retries++ == 60) {
8760 			dewtprintk(ioc,
8761 				   ioc_info(ioc, "%s: ABORT_TASK: giving up\n",
8762 					    __func__));
8763 			spin_lock_irqsave(&ioc->scsi_lookup_lock, flags);
8764 			goto broadcast_aen_retry;
8765 		}
8766 
8767 		if (ioc->shost_recovery)
8768 			goto out_no_lock;
8769 
8770 		r = mpt3sas_scsih_issue_tm(ioc, handle, sdev->channel, sdev->id,
8771 			sdev->lun, MPI2_SCSITASKMGMT_TASKTYPE_ABORT_TASK,
8772 			st->smid, st->msix_io, 30, 0);
8773 		if (r == FAILED || st->cb_idx != 0xFF) {
8774 			sdev_printk(KERN_WARNING, sdev,
8775 			    "mpt3sas_scsih_issue_tm: ABORT_TASK: FAILED : "
8776 			    "scmd(%p)\n", scmd);
8777 			goto tm_retry;
8778 		}
8779 
8780 		if (task_abort_retries > 1)
8781 			sdev_printk(KERN_WARNING, sdev,
8782 			    "mpt3sas_scsih_issue_tm: ABORT_TASK: RETRIES (%d):"
8783 			    " scmd(%p)\n",
8784 			    task_abort_retries - 1, scmd);
8785 
8786 		termination_count += le32_to_cpu(mpi_reply->TerminationCount);
8787 		spin_lock_irqsave(&ioc->scsi_lookup_lock, flags);
8788 	}
8789 
8790 	if (ioc->broadcast_aen_pending) {
8791 		dewtprintk(ioc,
8792 			   ioc_info(ioc,
8793 				    "%s: loop back due to pending AEN\n",
8794 				    __func__));
8795 		 ioc->broadcast_aen_pending = 0;
8796 		 goto broadcast_aen_retry;
8797 	}
8798 
8799  out:
8800 	spin_unlock_irqrestore(&ioc->scsi_lookup_lock, flags);
8801  out_no_lock:
8802 
8803 	dewtprintk(ioc,
8804 		   ioc_info(ioc, "%s - exit, query_count = %d termination_count = %d\n",
8805 			    __func__, query_count, termination_count));
8806 
8807 	ioc->broadcast_aen_busy = 0;
8808 	if (!ioc->shost_recovery)
8809 		_scsih_ublock_io_all_device(ioc);
8810 	mutex_unlock(&ioc->tm_cmds.mutex);
8811 }
8812 
8813 /**
8814  * _scsih_sas_discovery_event - handle discovery events
8815  * @ioc: per adapter object
8816  * @fw_event: The fw_event_work object
8817  * Context: user.
8818  */
8819 static void
8820 _scsih_sas_discovery_event(struct MPT3SAS_ADAPTER *ioc,
8821 	struct fw_event_work *fw_event)
8822 {
8823 	Mpi2EventDataSasDiscovery_t *event_data =
8824 		(Mpi2EventDataSasDiscovery_t *) fw_event->event_data;
8825 
8826 	if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) {
8827 		ioc_info(ioc, "discovery event: (%s)",
8828 			 event_data->ReasonCode == MPI2_EVENT_SAS_DISC_RC_STARTED ?
8829 			 "start" : "stop");
8830 		if (event_data->DiscoveryStatus)
8831 			pr_cont("discovery_status(0x%08x)",
8832 				le32_to_cpu(event_data->DiscoveryStatus));
8833 		pr_cont("\n");
8834 	}
8835 
8836 	if (event_data->ReasonCode == MPI2_EVENT_SAS_DISC_RC_STARTED &&
8837 	    !ioc->sas_hba.num_phys) {
8838 		if (disable_discovery > 0 && ioc->shost_recovery) {
8839 			/* Wait for the reset to complete */
8840 			while (ioc->shost_recovery)
8841 				ssleep(1);
8842 		}
8843 		_scsih_sas_host_add(ioc);
8844 	}
8845 }
8846 
8847 /**
8848  * _scsih_sas_device_discovery_error_event - display SAS device discovery error
8849  *						events
8850  * @ioc: per adapter object
8851  * @fw_event: The fw_event_work object
8852  * Context: user.
8853  */
8854 static void
8855 _scsih_sas_device_discovery_error_event(struct MPT3SAS_ADAPTER *ioc,
8856 	struct fw_event_work *fw_event)
8857 {
8858 	Mpi25EventDataSasDeviceDiscoveryError_t *event_data =
8859 		(Mpi25EventDataSasDeviceDiscoveryError_t *)fw_event->event_data;
8860 
8861 	switch (event_data->ReasonCode) {
8862 	case MPI25_EVENT_SAS_DISC_ERR_SMP_FAILED:
8863 		ioc_warn(ioc, "SMP command sent to the expander (handle:0x%04x, sas_address:0x%016llx, physical_port:0x%02x) has failed\n",
8864 			 le16_to_cpu(event_data->DevHandle),
8865 			 (u64)le64_to_cpu(event_data->SASAddress),
8866 			 event_data->PhysicalPort);
8867 		break;
8868 	case MPI25_EVENT_SAS_DISC_ERR_SMP_TIMEOUT:
8869 		ioc_warn(ioc, "SMP command sent to the expander (handle:0x%04x, sas_address:0x%016llx, physical_port:0x%02x) has timed out\n",
8870 			 le16_to_cpu(event_data->DevHandle),
8871 			 (u64)le64_to_cpu(event_data->SASAddress),
8872 			 event_data->PhysicalPort);
8873 		break;
8874 	default:
8875 		break;
8876 	}
8877 }
8878 
8879 /**
8880  * _scsih_pcie_enumeration_event - handle enumeration events
8881  * @ioc: per adapter object
8882  * @fw_event: The fw_event_work object
8883  * Context: user.
8884  */
8885 static void
8886 _scsih_pcie_enumeration_event(struct MPT3SAS_ADAPTER *ioc,
8887 	struct fw_event_work *fw_event)
8888 {
8889 	Mpi26EventDataPCIeEnumeration_t *event_data =
8890 		(Mpi26EventDataPCIeEnumeration_t *)fw_event->event_data;
8891 
8892 	if (!(ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK))
8893 		return;
8894 
8895 	ioc_info(ioc, "pcie enumeration event: (%s) Flag 0x%02x",
8896 		 (event_data->ReasonCode == MPI26_EVENT_PCIE_ENUM_RC_STARTED) ?
8897 		 "started" : "completed",
8898 		 event_data->Flags);
8899 	if (event_data->EnumerationStatus)
8900 		pr_cont("enumeration_status(0x%08x)",
8901 			le32_to_cpu(event_data->EnumerationStatus));
8902 	pr_cont("\n");
8903 }
8904 
8905 /**
8906  * _scsih_ir_fastpath - turn on fastpath for IR physdisk
8907  * @ioc: per adapter object
8908  * @handle: device handle for physical disk
8909  * @phys_disk_num: physical disk number
8910  *
8911  * Return: 0 for success, else failure.
8912  */
8913 static int
8914 _scsih_ir_fastpath(struct MPT3SAS_ADAPTER *ioc, u16 handle, u8 phys_disk_num)
8915 {
8916 	Mpi2RaidActionRequest_t *mpi_request;
8917 	Mpi2RaidActionReply_t *mpi_reply;
8918 	u16 smid;
8919 	u8 issue_reset = 0;
8920 	int rc = 0;
8921 	u16 ioc_status;
8922 	u32 log_info;
8923 
8924 	if (ioc->hba_mpi_version_belonged == MPI2_VERSION)
8925 		return rc;
8926 
8927 	mutex_lock(&ioc->scsih_cmds.mutex);
8928 
8929 	if (ioc->scsih_cmds.status != MPT3_CMD_NOT_USED) {
8930 		ioc_err(ioc, "%s: scsih_cmd in use\n", __func__);
8931 		rc = -EAGAIN;
8932 		goto out;
8933 	}
8934 	ioc->scsih_cmds.status = MPT3_CMD_PENDING;
8935 
8936 	smid = mpt3sas_base_get_smid(ioc, ioc->scsih_cb_idx);
8937 	if (!smid) {
8938 		ioc_err(ioc, "%s: failed obtaining a smid\n", __func__);
8939 		ioc->scsih_cmds.status = MPT3_CMD_NOT_USED;
8940 		rc = -EAGAIN;
8941 		goto out;
8942 	}
8943 
8944 	mpi_request = mpt3sas_base_get_msg_frame(ioc, smid);
8945 	ioc->scsih_cmds.smid = smid;
8946 	memset(mpi_request, 0, sizeof(Mpi2RaidActionRequest_t));
8947 
8948 	mpi_request->Function = MPI2_FUNCTION_RAID_ACTION;
8949 	mpi_request->Action = MPI2_RAID_ACTION_PHYSDISK_HIDDEN;
8950 	mpi_request->PhysDiskNum = phys_disk_num;
8951 
8952 	dewtprintk(ioc,
8953 		   ioc_info(ioc, "IR RAID_ACTION: turning fast path on for handle(0x%04x), phys_disk_num (0x%02x)\n",
8954 			    handle, phys_disk_num));
8955 
8956 	init_completion(&ioc->scsih_cmds.done);
8957 	ioc->put_smid_default(ioc, smid);
8958 	wait_for_completion_timeout(&ioc->scsih_cmds.done, 10*HZ);
8959 
8960 	if (!(ioc->scsih_cmds.status & MPT3_CMD_COMPLETE)) {
8961 		mpt3sas_check_cmd_timeout(ioc,
8962 		    ioc->scsih_cmds.status, mpi_request,
8963 		    sizeof(Mpi2RaidActionRequest_t)/4, issue_reset);
8964 		rc = -EFAULT;
8965 		goto out;
8966 	}
8967 
8968 	if (ioc->scsih_cmds.status & MPT3_CMD_REPLY_VALID) {
8969 
8970 		mpi_reply = ioc->scsih_cmds.reply;
8971 		ioc_status = le16_to_cpu(mpi_reply->IOCStatus);
8972 		if (ioc_status & MPI2_IOCSTATUS_FLAG_LOG_INFO_AVAILABLE)
8973 			log_info =  le32_to_cpu(mpi_reply->IOCLogInfo);
8974 		else
8975 			log_info = 0;
8976 		ioc_status &= MPI2_IOCSTATUS_MASK;
8977 		if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
8978 			dewtprintk(ioc,
8979 				   ioc_info(ioc, "IR RAID_ACTION: failed: ioc_status(0x%04x), loginfo(0x%08x)!!!\n",
8980 					    ioc_status, log_info));
8981 			rc = -EFAULT;
8982 		} else
8983 			dewtprintk(ioc,
8984 				   ioc_info(ioc, "IR RAID_ACTION: completed successfully\n"));
8985 	}
8986 
8987  out:
8988 	ioc->scsih_cmds.status = MPT3_CMD_NOT_USED;
8989 	mutex_unlock(&ioc->scsih_cmds.mutex);
8990 
8991 	if (issue_reset)
8992 		mpt3sas_base_hard_reset_handler(ioc, FORCE_BIG_HAMMER);
8993 	return rc;
8994 }
8995 
8996 /**
8997  * _scsih_reprobe_lun - reprobing lun
8998  * @sdev: scsi device struct
8999  * @no_uld_attach: sdev->no_uld_attach flag setting
9000  *
9001  **/
9002 static void
9003 _scsih_reprobe_lun(struct scsi_device *sdev, void *no_uld_attach)
9004 {
9005 	sdev->no_uld_attach = no_uld_attach ? 1 : 0;
9006 	sdev_printk(KERN_INFO, sdev, "%s raid component\n",
9007 	    sdev->no_uld_attach ? "hiding" : "exposing");
9008 	WARN_ON(scsi_device_reprobe(sdev));
9009 }
9010 
9011 /**
9012  * _scsih_sas_volume_add - add new volume
9013  * @ioc: per adapter object
9014  * @element: IR config element data
9015  * Context: user.
9016  */
9017 static void
9018 _scsih_sas_volume_add(struct MPT3SAS_ADAPTER *ioc,
9019 	Mpi2EventIrConfigElement_t *element)
9020 {
9021 	struct _raid_device *raid_device;
9022 	unsigned long flags;
9023 	u64 wwid;
9024 	u16 handle = le16_to_cpu(element->VolDevHandle);
9025 	int rc;
9026 
9027 	mpt3sas_config_get_volume_wwid(ioc, handle, &wwid);
9028 	if (!wwid) {
9029 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
9030 			__FILE__, __LINE__, __func__);
9031 		return;
9032 	}
9033 
9034 	spin_lock_irqsave(&ioc->raid_device_lock, flags);
9035 	raid_device = _scsih_raid_device_find_by_wwid(ioc, wwid);
9036 	spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
9037 
9038 	if (raid_device)
9039 		return;
9040 
9041 	raid_device = kzalloc(sizeof(struct _raid_device), GFP_KERNEL);
9042 	if (!raid_device) {
9043 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
9044 			__FILE__, __LINE__, __func__);
9045 		return;
9046 	}
9047 
9048 	raid_device->id = ioc->sas_id++;
9049 	raid_device->channel = RAID_CHANNEL;
9050 	raid_device->handle = handle;
9051 	raid_device->wwid = wwid;
9052 	_scsih_raid_device_add(ioc, raid_device);
9053 	if (!ioc->wait_for_discovery_to_complete) {
9054 		rc = scsi_add_device(ioc->shost, RAID_CHANNEL,
9055 		    raid_device->id, 0);
9056 		if (rc)
9057 			_scsih_raid_device_remove(ioc, raid_device);
9058 	} else {
9059 		spin_lock_irqsave(&ioc->raid_device_lock, flags);
9060 		_scsih_determine_boot_device(ioc, raid_device, 1);
9061 		spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
9062 	}
9063 }
9064 
9065 /**
9066  * _scsih_sas_volume_delete - delete volume
9067  * @ioc: per adapter object
9068  * @handle: volume device handle
9069  * Context: user.
9070  */
9071 static void
9072 _scsih_sas_volume_delete(struct MPT3SAS_ADAPTER *ioc, u16 handle)
9073 {
9074 	struct _raid_device *raid_device;
9075 	unsigned long flags;
9076 	struct MPT3SAS_TARGET *sas_target_priv_data;
9077 	struct scsi_target *starget = NULL;
9078 
9079 	spin_lock_irqsave(&ioc->raid_device_lock, flags);
9080 	raid_device = mpt3sas_raid_device_find_by_handle(ioc, handle);
9081 	if (raid_device) {
9082 		if (raid_device->starget) {
9083 			starget = raid_device->starget;
9084 			sas_target_priv_data = starget->hostdata;
9085 			sas_target_priv_data->deleted = 1;
9086 		}
9087 		ioc_info(ioc, "removing handle(0x%04x), wwid(0x%016llx)\n",
9088 			 raid_device->handle, (u64)raid_device->wwid);
9089 		list_del(&raid_device->list);
9090 		kfree(raid_device);
9091 	}
9092 	spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
9093 	if (starget)
9094 		scsi_remove_target(&starget->dev);
9095 }
9096 
9097 /**
9098  * _scsih_sas_pd_expose - expose pd component to /dev/sdX
9099  * @ioc: per adapter object
9100  * @element: IR config element data
9101  * Context: user.
9102  */
9103 static void
9104 _scsih_sas_pd_expose(struct MPT3SAS_ADAPTER *ioc,
9105 	Mpi2EventIrConfigElement_t *element)
9106 {
9107 	struct _sas_device *sas_device;
9108 	struct scsi_target *starget = NULL;
9109 	struct MPT3SAS_TARGET *sas_target_priv_data;
9110 	unsigned long flags;
9111 	u16 handle = le16_to_cpu(element->PhysDiskDevHandle);
9112 
9113 	spin_lock_irqsave(&ioc->sas_device_lock, flags);
9114 	sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle);
9115 	if (sas_device) {
9116 		sas_device->volume_handle = 0;
9117 		sas_device->volume_wwid = 0;
9118 		clear_bit(handle, ioc->pd_handles);
9119 		if (sas_device->starget && sas_device->starget->hostdata) {
9120 			starget = sas_device->starget;
9121 			sas_target_priv_data = starget->hostdata;
9122 			sas_target_priv_data->flags &=
9123 			    ~MPT_TARGET_FLAGS_RAID_COMPONENT;
9124 		}
9125 	}
9126 	spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
9127 	if (!sas_device)
9128 		return;
9129 
9130 	/* exposing raid component */
9131 	if (starget)
9132 		starget_for_each_device(starget, NULL, _scsih_reprobe_lun);
9133 
9134 	sas_device_put(sas_device);
9135 }
9136 
9137 /**
9138  * _scsih_sas_pd_hide - hide pd component from /dev/sdX
9139  * @ioc: per adapter object
9140  * @element: IR config element data
9141  * Context: user.
9142  */
9143 static void
9144 _scsih_sas_pd_hide(struct MPT3SAS_ADAPTER *ioc,
9145 	Mpi2EventIrConfigElement_t *element)
9146 {
9147 	struct _sas_device *sas_device;
9148 	struct scsi_target *starget = NULL;
9149 	struct MPT3SAS_TARGET *sas_target_priv_data;
9150 	unsigned long flags;
9151 	u16 handle = le16_to_cpu(element->PhysDiskDevHandle);
9152 	u16 volume_handle = 0;
9153 	u64 volume_wwid = 0;
9154 
9155 	mpt3sas_config_get_volume_handle(ioc, handle, &volume_handle);
9156 	if (volume_handle)
9157 		mpt3sas_config_get_volume_wwid(ioc, volume_handle,
9158 		    &volume_wwid);
9159 
9160 	spin_lock_irqsave(&ioc->sas_device_lock, flags);
9161 	sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle);
9162 	if (sas_device) {
9163 		set_bit(handle, ioc->pd_handles);
9164 		if (sas_device->starget && sas_device->starget->hostdata) {
9165 			starget = sas_device->starget;
9166 			sas_target_priv_data = starget->hostdata;
9167 			sas_target_priv_data->flags |=
9168 			    MPT_TARGET_FLAGS_RAID_COMPONENT;
9169 			sas_device->volume_handle = volume_handle;
9170 			sas_device->volume_wwid = volume_wwid;
9171 		}
9172 	}
9173 	spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
9174 	if (!sas_device)
9175 		return;
9176 
9177 	/* hiding raid component */
9178 	_scsih_ir_fastpath(ioc, handle, element->PhysDiskNum);
9179 
9180 	if (starget)
9181 		starget_for_each_device(starget, (void *)1, _scsih_reprobe_lun);
9182 
9183 	sas_device_put(sas_device);
9184 }
9185 
9186 /**
9187  * _scsih_sas_pd_delete - delete pd component
9188  * @ioc: per adapter object
9189  * @element: IR config element data
9190  * Context: user.
9191  */
9192 static void
9193 _scsih_sas_pd_delete(struct MPT3SAS_ADAPTER *ioc,
9194 	Mpi2EventIrConfigElement_t *element)
9195 {
9196 	u16 handle = le16_to_cpu(element->PhysDiskDevHandle);
9197 
9198 	_scsih_device_remove_by_handle(ioc, handle);
9199 }
9200 
9201 /**
9202  * _scsih_sas_pd_add - remove pd component
9203  * @ioc: per adapter object
9204  * @element: IR config element data
9205  * Context: user.
9206  */
9207 static void
9208 _scsih_sas_pd_add(struct MPT3SAS_ADAPTER *ioc,
9209 	Mpi2EventIrConfigElement_t *element)
9210 {
9211 	struct _sas_device *sas_device;
9212 	u16 handle = le16_to_cpu(element->PhysDiskDevHandle);
9213 	Mpi2ConfigReply_t mpi_reply;
9214 	Mpi2SasDevicePage0_t sas_device_pg0;
9215 	u32 ioc_status;
9216 	u64 sas_address;
9217 	u16 parent_handle;
9218 
9219 	set_bit(handle, ioc->pd_handles);
9220 
9221 	sas_device = mpt3sas_get_sdev_by_handle(ioc, handle);
9222 	if (sas_device) {
9223 		_scsih_ir_fastpath(ioc, handle, element->PhysDiskNum);
9224 		sas_device_put(sas_device);
9225 		return;
9226 	}
9227 
9228 	if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply, &sas_device_pg0,
9229 	    MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) {
9230 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
9231 			__FILE__, __LINE__, __func__);
9232 		return;
9233 	}
9234 
9235 	ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
9236 	    MPI2_IOCSTATUS_MASK;
9237 	if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
9238 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
9239 			__FILE__, __LINE__, __func__);
9240 		return;
9241 	}
9242 
9243 	parent_handle = le16_to_cpu(sas_device_pg0.ParentDevHandle);
9244 	if (!_scsih_get_sas_address(ioc, parent_handle, &sas_address))
9245 		mpt3sas_transport_update_links(ioc, sas_address, handle,
9246 		    sas_device_pg0.PhyNum, MPI2_SAS_NEG_LINK_RATE_1_5,
9247 		    mpt3sas_get_port_by_id(ioc,
9248 		    sas_device_pg0.PhysicalPort, 0));
9249 
9250 	_scsih_ir_fastpath(ioc, handle, element->PhysDiskNum);
9251 	_scsih_add_device(ioc, handle, 0, 1);
9252 }
9253 
9254 /**
9255  * _scsih_sas_ir_config_change_event_debug - debug for IR Config Change events
9256  * @ioc: per adapter object
9257  * @event_data: event data payload
9258  * Context: user.
9259  */
9260 static void
9261 _scsih_sas_ir_config_change_event_debug(struct MPT3SAS_ADAPTER *ioc,
9262 	Mpi2EventDataIrConfigChangeList_t *event_data)
9263 {
9264 	Mpi2EventIrConfigElement_t *element;
9265 	u8 element_type;
9266 	int i;
9267 	char *reason_str = NULL, *element_str = NULL;
9268 
9269 	element = (Mpi2EventIrConfigElement_t *)&event_data->ConfigElement[0];
9270 
9271 	ioc_info(ioc, "raid config change: (%s), elements(%d)\n",
9272 		 le32_to_cpu(event_data->Flags) & MPI2_EVENT_IR_CHANGE_FLAGS_FOREIGN_CONFIG ?
9273 		 "foreign" : "native",
9274 		 event_data->NumElements);
9275 	for (i = 0; i < event_data->NumElements; i++, element++) {
9276 		switch (element->ReasonCode) {
9277 		case MPI2_EVENT_IR_CHANGE_RC_ADDED:
9278 			reason_str = "add";
9279 			break;
9280 		case MPI2_EVENT_IR_CHANGE_RC_REMOVED:
9281 			reason_str = "remove";
9282 			break;
9283 		case MPI2_EVENT_IR_CHANGE_RC_NO_CHANGE:
9284 			reason_str = "no change";
9285 			break;
9286 		case MPI2_EVENT_IR_CHANGE_RC_HIDE:
9287 			reason_str = "hide";
9288 			break;
9289 		case MPI2_EVENT_IR_CHANGE_RC_UNHIDE:
9290 			reason_str = "unhide";
9291 			break;
9292 		case MPI2_EVENT_IR_CHANGE_RC_VOLUME_CREATED:
9293 			reason_str = "volume_created";
9294 			break;
9295 		case MPI2_EVENT_IR_CHANGE_RC_VOLUME_DELETED:
9296 			reason_str = "volume_deleted";
9297 			break;
9298 		case MPI2_EVENT_IR_CHANGE_RC_PD_CREATED:
9299 			reason_str = "pd_created";
9300 			break;
9301 		case MPI2_EVENT_IR_CHANGE_RC_PD_DELETED:
9302 			reason_str = "pd_deleted";
9303 			break;
9304 		default:
9305 			reason_str = "unknown reason";
9306 			break;
9307 		}
9308 		element_type = le16_to_cpu(element->ElementFlags) &
9309 		    MPI2_EVENT_IR_CHANGE_EFLAGS_ELEMENT_TYPE_MASK;
9310 		switch (element_type) {
9311 		case MPI2_EVENT_IR_CHANGE_EFLAGS_VOLUME_ELEMENT:
9312 			element_str = "volume";
9313 			break;
9314 		case MPI2_EVENT_IR_CHANGE_EFLAGS_VOLPHYSDISK_ELEMENT:
9315 			element_str = "phys disk";
9316 			break;
9317 		case MPI2_EVENT_IR_CHANGE_EFLAGS_HOTSPARE_ELEMENT:
9318 			element_str = "hot spare";
9319 			break;
9320 		default:
9321 			element_str = "unknown element";
9322 			break;
9323 		}
9324 		pr_info("\t(%s:%s), vol handle(0x%04x), " \
9325 		    "pd handle(0x%04x), pd num(0x%02x)\n", element_str,
9326 		    reason_str, le16_to_cpu(element->VolDevHandle),
9327 		    le16_to_cpu(element->PhysDiskDevHandle),
9328 		    element->PhysDiskNum);
9329 	}
9330 }
9331 
9332 /**
9333  * _scsih_sas_ir_config_change_event - handle ir configuration change events
9334  * @ioc: per adapter object
9335  * @fw_event: The fw_event_work object
9336  * Context: user.
9337  */
9338 static void
9339 _scsih_sas_ir_config_change_event(struct MPT3SAS_ADAPTER *ioc,
9340 	struct fw_event_work *fw_event)
9341 {
9342 	Mpi2EventIrConfigElement_t *element;
9343 	int i;
9344 	u8 foreign_config;
9345 	Mpi2EventDataIrConfigChangeList_t *event_data =
9346 		(Mpi2EventDataIrConfigChangeList_t *)
9347 		fw_event->event_data;
9348 
9349 	if ((ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) &&
9350 	     (!ioc->hide_ir_msg))
9351 		_scsih_sas_ir_config_change_event_debug(ioc, event_data);
9352 
9353 	foreign_config = (le32_to_cpu(event_data->Flags) &
9354 	    MPI2_EVENT_IR_CHANGE_FLAGS_FOREIGN_CONFIG) ? 1 : 0;
9355 
9356 	element = (Mpi2EventIrConfigElement_t *)&event_data->ConfigElement[0];
9357 	if (ioc->shost_recovery &&
9358 	    ioc->hba_mpi_version_belonged != MPI2_VERSION) {
9359 		for (i = 0; i < event_data->NumElements; i++, element++) {
9360 			if (element->ReasonCode == MPI2_EVENT_IR_CHANGE_RC_HIDE)
9361 				_scsih_ir_fastpath(ioc,
9362 					le16_to_cpu(element->PhysDiskDevHandle),
9363 					element->PhysDiskNum);
9364 		}
9365 		return;
9366 	}
9367 
9368 	for (i = 0; i < event_data->NumElements; i++, element++) {
9369 
9370 		switch (element->ReasonCode) {
9371 		case MPI2_EVENT_IR_CHANGE_RC_VOLUME_CREATED:
9372 		case MPI2_EVENT_IR_CHANGE_RC_ADDED:
9373 			if (!foreign_config)
9374 				_scsih_sas_volume_add(ioc, element);
9375 			break;
9376 		case MPI2_EVENT_IR_CHANGE_RC_VOLUME_DELETED:
9377 		case MPI2_EVENT_IR_CHANGE_RC_REMOVED:
9378 			if (!foreign_config)
9379 				_scsih_sas_volume_delete(ioc,
9380 				    le16_to_cpu(element->VolDevHandle));
9381 			break;
9382 		case MPI2_EVENT_IR_CHANGE_RC_PD_CREATED:
9383 			if (!ioc->is_warpdrive)
9384 				_scsih_sas_pd_hide(ioc, element);
9385 			break;
9386 		case MPI2_EVENT_IR_CHANGE_RC_PD_DELETED:
9387 			if (!ioc->is_warpdrive)
9388 				_scsih_sas_pd_expose(ioc, element);
9389 			break;
9390 		case MPI2_EVENT_IR_CHANGE_RC_HIDE:
9391 			if (!ioc->is_warpdrive)
9392 				_scsih_sas_pd_add(ioc, element);
9393 			break;
9394 		case MPI2_EVENT_IR_CHANGE_RC_UNHIDE:
9395 			if (!ioc->is_warpdrive)
9396 				_scsih_sas_pd_delete(ioc, element);
9397 			break;
9398 		}
9399 	}
9400 }
9401 
9402 /**
9403  * _scsih_sas_ir_volume_event - IR volume event
9404  * @ioc: per adapter object
9405  * @fw_event: The fw_event_work object
9406  * Context: user.
9407  */
9408 static void
9409 _scsih_sas_ir_volume_event(struct MPT3SAS_ADAPTER *ioc,
9410 	struct fw_event_work *fw_event)
9411 {
9412 	u64 wwid;
9413 	unsigned long flags;
9414 	struct _raid_device *raid_device;
9415 	u16 handle;
9416 	u32 state;
9417 	int rc;
9418 	Mpi2EventDataIrVolume_t *event_data =
9419 		(Mpi2EventDataIrVolume_t *) fw_event->event_data;
9420 
9421 	if (ioc->shost_recovery)
9422 		return;
9423 
9424 	if (event_data->ReasonCode != MPI2_EVENT_IR_VOLUME_RC_STATE_CHANGED)
9425 		return;
9426 
9427 	handle = le16_to_cpu(event_data->VolDevHandle);
9428 	state = le32_to_cpu(event_data->NewValue);
9429 	if (!ioc->hide_ir_msg)
9430 		dewtprintk(ioc,
9431 			   ioc_info(ioc, "%s: handle(0x%04x), old(0x%08x), new(0x%08x)\n",
9432 				    __func__, handle,
9433 				    le32_to_cpu(event_data->PreviousValue),
9434 				    state));
9435 	switch (state) {
9436 	case MPI2_RAID_VOL_STATE_MISSING:
9437 	case MPI2_RAID_VOL_STATE_FAILED:
9438 		_scsih_sas_volume_delete(ioc, handle);
9439 		break;
9440 
9441 	case MPI2_RAID_VOL_STATE_ONLINE:
9442 	case MPI2_RAID_VOL_STATE_DEGRADED:
9443 	case MPI2_RAID_VOL_STATE_OPTIMAL:
9444 
9445 		spin_lock_irqsave(&ioc->raid_device_lock, flags);
9446 		raid_device = mpt3sas_raid_device_find_by_handle(ioc, handle);
9447 		spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
9448 
9449 		if (raid_device)
9450 			break;
9451 
9452 		mpt3sas_config_get_volume_wwid(ioc, handle, &wwid);
9453 		if (!wwid) {
9454 			ioc_err(ioc, "failure at %s:%d/%s()!\n",
9455 				__FILE__, __LINE__, __func__);
9456 			break;
9457 		}
9458 
9459 		raid_device = kzalloc(sizeof(struct _raid_device), GFP_KERNEL);
9460 		if (!raid_device) {
9461 			ioc_err(ioc, "failure at %s:%d/%s()!\n",
9462 				__FILE__, __LINE__, __func__);
9463 			break;
9464 		}
9465 
9466 		raid_device->id = ioc->sas_id++;
9467 		raid_device->channel = RAID_CHANNEL;
9468 		raid_device->handle = handle;
9469 		raid_device->wwid = wwid;
9470 		_scsih_raid_device_add(ioc, raid_device);
9471 		rc = scsi_add_device(ioc->shost, RAID_CHANNEL,
9472 		    raid_device->id, 0);
9473 		if (rc)
9474 			_scsih_raid_device_remove(ioc, raid_device);
9475 		break;
9476 
9477 	case MPI2_RAID_VOL_STATE_INITIALIZING:
9478 	default:
9479 		break;
9480 	}
9481 }
9482 
9483 /**
9484  * _scsih_sas_ir_physical_disk_event - PD event
9485  * @ioc: per adapter object
9486  * @fw_event: The fw_event_work object
9487  * Context: user.
9488  */
9489 static void
9490 _scsih_sas_ir_physical_disk_event(struct MPT3SAS_ADAPTER *ioc,
9491 	struct fw_event_work *fw_event)
9492 {
9493 	u16 handle, parent_handle;
9494 	u32 state;
9495 	struct _sas_device *sas_device;
9496 	Mpi2ConfigReply_t mpi_reply;
9497 	Mpi2SasDevicePage0_t sas_device_pg0;
9498 	u32 ioc_status;
9499 	Mpi2EventDataIrPhysicalDisk_t *event_data =
9500 		(Mpi2EventDataIrPhysicalDisk_t *) fw_event->event_data;
9501 	u64 sas_address;
9502 
9503 	if (ioc->shost_recovery)
9504 		return;
9505 
9506 	if (event_data->ReasonCode != MPI2_EVENT_IR_PHYSDISK_RC_STATE_CHANGED)
9507 		return;
9508 
9509 	handle = le16_to_cpu(event_data->PhysDiskDevHandle);
9510 	state = le32_to_cpu(event_data->NewValue);
9511 
9512 	if (!ioc->hide_ir_msg)
9513 		dewtprintk(ioc,
9514 			   ioc_info(ioc, "%s: handle(0x%04x), old(0x%08x), new(0x%08x)\n",
9515 				    __func__, handle,
9516 				    le32_to_cpu(event_data->PreviousValue),
9517 				    state));
9518 
9519 	switch (state) {
9520 	case MPI2_RAID_PD_STATE_ONLINE:
9521 	case MPI2_RAID_PD_STATE_DEGRADED:
9522 	case MPI2_RAID_PD_STATE_REBUILDING:
9523 	case MPI2_RAID_PD_STATE_OPTIMAL:
9524 	case MPI2_RAID_PD_STATE_HOT_SPARE:
9525 
9526 		if (!ioc->is_warpdrive)
9527 			set_bit(handle, ioc->pd_handles);
9528 
9529 		sas_device = mpt3sas_get_sdev_by_handle(ioc, handle);
9530 		if (sas_device) {
9531 			sas_device_put(sas_device);
9532 			return;
9533 		}
9534 
9535 		if ((mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply,
9536 		    &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE,
9537 		    handle))) {
9538 			ioc_err(ioc, "failure at %s:%d/%s()!\n",
9539 				__FILE__, __LINE__, __func__);
9540 			return;
9541 		}
9542 
9543 		ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
9544 		    MPI2_IOCSTATUS_MASK;
9545 		if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
9546 			ioc_err(ioc, "failure at %s:%d/%s()!\n",
9547 				__FILE__, __LINE__, __func__);
9548 			return;
9549 		}
9550 
9551 		parent_handle = le16_to_cpu(sas_device_pg0.ParentDevHandle);
9552 		if (!_scsih_get_sas_address(ioc, parent_handle, &sas_address))
9553 			mpt3sas_transport_update_links(ioc, sas_address, handle,
9554 			    sas_device_pg0.PhyNum, MPI2_SAS_NEG_LINK_RATE_1_5,
9555 			    mpt3sas_get_port_by_id(ioc,
9556 			    sas_device_pg0.PhysicalPort, 0));
9557 
9558 		_scsih_add_device(ioc, handle, 0, 1);
9559 
9560 		break;
9561 
9562 	case MPI2_RAID_PD_STATE_OFFLINE:
9563 	case MPI2_RAID_PD_STATE_NOT_CONFIGURED:
9564 	case MPI2_RAID_PD_STATE_NOT_COMPATIBLE:
9565 	default:
9566 		break;
9567 	}
9568 }
9569 
9570 /**
9571  * _scsih_sas_ir_operation_status_event_debug - debug for IR op event
9572  * @ioc: per adapter object
9573  * @event_data: event data payload
9574  * Context: user.
9575  */
9576 static void
9577 _scsih_sas_ir_operation_status_event_debug(struct MPT3SAS_ADAPTER *ioc,
9578 	Mpi2EventDataIrOperationStatus_t *event_data)
9579 {
9580 	char *reason_str = NULL;
9581 
9582 	switch (event_data->RAIDOperation) {
9583 	case MPI2_EVENT_IR_RAIDOP_RESYNC:
9584 		reason_str = "resync";
9585 		break;
9586 	case MPI2_EVENT_IR_RAIDOP_ONLINE_CAP_EXPANSION:
9587 		reason_str = "online capacity expansion";
9588 		break;
9589 	case MPI2_EVENT_IR_RAIDOP_CONSISTENCY_CHECK:
9590 		reason_str = "consistency check";
9591 		break;
9592 	case MPI2_EVENT_IR_RAIDOP_BACKGROUND_INIT:
9593 		reason_str = "background init";
9594 		break;
9595 	case MPI2_EVENT_IR_RAIDOP_MAKE_DATA_CONSISTENT:
9596 		reason_str = "make data consistent";
9597 		break;
9598 	}
9599 
9600 	if (!reason_str)
9601 		return;
9602 
9603 	ioc_info(ioc, "raid operational status: (%s)\thandle(0x%04x), percent complete(%d)\n",
9604 		 reason_str,
9605 		 le16_to_cpu(event_data->VolDevHandle),
9606 		 event_data->PercentComplete);
9607 }
9608 
9609 /**
9610  * _scsih_sas_ir_operation_status_event - handle RAID operation events
9611  * @ioc: per adapter object
9612  * @fw_event: The fw_event_work object
9613  * Context: user.
9614  */
9615 static void
9616 _scsih_sas_ir_operation_status_event(struct MPT3SAS_ADAPTER *ioc,
9617 	struct fw_event_work *fw_event)
9618 {
9619 	Mpi2EventDataIrOperationStatus_t *event_data =
9620 		(Mpi2EventDataIrOperationStatus_t *)
9621 		fw_event->event_data;
9622 	static struct _raid_device *raid_device;
9623 	unsigned long flags;
9624 	u16 handle;
9625 
9626 	if ((ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) &&
9627 	    (!ioc->hide_ir_msg))
9628 		_scsih_sas_ir_operation_status_event_debug(ioc,
9629 		     event_data);
9630 
9631 	/* code added for raid transport support */
9632 	if (event_data->RAIDOperation == MPI2_EVENT_IR_RAIDOP_RESYNC) {
9633 
9634 		spin_lock_irqsave(&ioc->raid_device_lock, flags);
9635 		handle = le16_to_cpu(event_data->VolDevHandle);
9636 		raid_device = mpt3sas_raid_device_find_by_handle(ioc, handle);
9637 		if (raid_device)
9638 			raid_device->percent_complete =
9639 			    event_data->PercentComplete;
9640 		spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
9641 	}
9642 }
9643 
9644 /**
9645  * _scsih_prep_device_scan - initialize parameters prior to device scan
9646  * @ioc: per adapter object
9647  *
9648  * Set the deleted flag prior to device scan.  If the device is found during
9649  * the scan, then we clear the deleted flag.
9650  */
9651 static void
9652 _scsih_prep_device_scan(struct MPT3SAS_ADAPTER *ioc)
9653 {
9654 	struct MPT3SAS_DEVICE *sas_device_priv_data;
9655 	struct scsi_device *sdev;
9656 
9657 	shost_for_each_device(sdev, ioc->shost) {
9658 		sas_device_priv_data = sdev->hostdata;
9659 		if (sas_device_priv_data && sas_device_priv_data->sas_target)
9660 			sas_device_priv_data->sas_target->deleted = 1;
9661 	}
9662 }
9663 
9664 /**
9665  * _scsih_update_device_qdepth - Update QD during Reset.
9666  * @ioc: per adapter object
9667  *
9668  */
9669 static void
9670 _scsih_update_device_qdepth(struct MPT3SAS_ADAPTER *ioc)
9671 {
9672 	struct MPT3SAS_DEVICE *sas_device_priv_data;
9673 	struct MPT3SAS_TARGET *sas_target_priv_data;
9674 	struct _sas_device *sas_device;
9675 	struct scsi_device *sdev;
9676 	u16 qdepth;
9677 
9678 	ioc_info(ioc, "Update devices with firmware reported queue depth\n");
9679 	shost_for_each_device(sdev, ioc->shost) {
9680 		sas_device_priv_data = sdev->hostdata;
9681 		if (sas_device_priv_data && sas_device_priv_data->sas_target) {
9682 			sas_target_priv_data = sas_device_priv_data->sas_target;
9683 			sas_device = sas_device_priv_data->sas_target->sas_dev;
9684 			if (sas_target_priv_data->flags & MPT_TARGET_FLAGS_PCIE_DEVICE)
9685 				qdepth = ioc->max_nvme_qd;
9686 			else if (sas_device &&
9687 			    sas_device->device_info & MPI2_SAS_DEVICE_INFO_SSP_TARGET)
9688 				qdepth = (sas_device->port_type > 1) ?
9689 				    ioc->max_wideport_qd : ioc->max_narrowport_qd;
9690 			else if (sas_device &&
9691 			    sas_device->device_info & MPI2_SAS_DEVICE_INFO_SATA_DEVICE)
9692 				qdepth = ioc->max_sata_qd;
9693 			else
9694 				continue;
9695 			mpt3sas_scsih_change_queue_depth(sdev, qdepth);
9696 		}
9697 	}
9698 }
9699 
9700 /**
9701  * _scsih_mark_responding_sas_device - mark a sas_devices as responding
9702  * @ioc: per adapter object
9703  * @sas_device_pg0: SAS Device page 0
9704  *
9705  * After host reset, find out whether devices are still responding.
9706  * Used in _scsih_remove_unresponsive_sas_devices.
9707  */
9708 static void
9709 _scsih_mark_responding_sas_device(struct MPT3SAS_ADAPTER *ioc,
9710 Mpi2SasDevicePage0_t *sas_device_pg0)
9711 {
9712 	struct MPT3SAS_TARGET *sas_target_priv_data = NULL;
9713 	struct scsi_target *starget;
9714 	struct _sas_device *sas_device = NULL;
9715 	struct _enclosure_node *enclosure_dev = NULL;
9716 	unsigned long flags;
9717 	struct hba_port *port = mpt3sas_get_port_by_id(
9718 	    ioc, sas_device_pg0->PhysicalPort, 0);
9719 
9720 	if (sas_device_pg0->EnclosureHandle) {
9721 		enclosure_dev =
9722 			mpt3sas_scsih_enclosure_find_by_handle(ioc,
9723 				le16_to_cpu(sas_device_pg0->EnclosureHandle));
9724 		if (enclosure_dev == NULL)
9725 			ioc_info(ioc, "Enclosure handle(0x%04x) doesn't match with enclosure device!\n",
9726 				 sas_device_pg0->EnclosureHandle);
9727 	}
9728 	spin_lock_irqsave(&ioc->sas_device_lock, flags);
9729 	list_for_each_entry(sas_device, &ioc->sas_device_list, list) {
9730 		if (sas_device->sas_address != le64_to_cpu(
9731 		    sas_device_pg0->SASAddress))
9732 			continue;
9733 		if (sas_device->slot != le16_to_cpu(sas_device_pg0->Slot))
9734 			continue;
9735 		if (sas_device->port != port)
9736 			continue;
9737 		sas_device->responding = 1;
9738 		starget = sas_device->starget;
9739 		if (starget && starget->hostdata) {
9740 			sas_target_priv_data = starget->hostdata;
9741 			sas_target_priv_data->tm_busy = 0;
9742 			sas_target_priv_data->deleted = 0;
9743 		} else
9744 			sas_target_priv_data = NULL;
9745 		if (starget) {
9746 			starget_printk(KERN_INFO, starget,
9747 			    "handle(0x%04x), sas_addr(0x%016llx)\n",
9748 			    le16_to_cpu(sas_device_pg0->DevHandle),
9749 			    (unsigned long long)
9750 			    sas_device->sas_address);
9751 
9752 			if (sas_device->enclosure_handle != 0)
9753 				starget_printk(KERN_INFO, starget,
9754 				 "enclosure logical id(0x%016llx), slot(%d)\n",
9755 				 (unsigned long long)
9756 				 sas_device->enclosure_logical_id,
9757 				 sas_device->slot);
9758 		}
9759 		if (le16_to_cpu(sas_device_pg0->Flags) &
9760 		      MPI2_SAS_DEVICE0_FLAGS_ENCL_LEVEL_VALID) {
9761 			sas_device->enclosure_level =
9762 			   sas_device_pg0->EnclosureLevel;
9763 			memcpy(&sas_device->connector_name[0],
9764 				&sas_device_pg0->ConnectorName[0], 4);
9765 		} else {
9766 			sas_device->enclosure_level = 0;
9767 			sas_device->connector_name[0] = '\0';
9768 		}
9769 
9770 		sas_device->enclosure_handle =
9771 			le16_to_cpu(sas_device_pg0->EnclosureHandle);
9772 		sas_device->is_chassis_slot_valid = 0;
9773 		if (enclosure_dev) {
9774 			sas_device->enclosure_logical_id = le64_to_cpu(
9775 				enclosure_dev->pg0.EnclosureLogicalID);
9776 			if (le16_to_cpu(enclosure_dev->pg0.Flags) &
9777 			    MPI2_SAS_ENCLS0_FLAGS_CHASSIS_SLOT_VALID) {
9778 				sas_device->is_chassis_slot_valid = 1;
9779 				sas_device->chassis_slot =
9780 					enclosure_dev->pg0.ChassisSlot;
9781 			}
9782 		}
9783 
9784 		if (sas_device->handle == le16_to_cpu(
9785 		    sas_device_pg0->DevHandle))
9786 			goto out;
9787 		pr_info("\thandle changed from(0x%04x)!!!\n",
9788 		    sas_device->handle);
9789 		sas_device->handle = le16_to_cpu(
9790 		    sas_device_pg0->DevHandle);
9791 		if (sas_target_priv_data)
9792 			sas_target_priv_data->handle =
9793 			    le16_to_cpu(sas_device_pg0->DevHandle);
9794 		goto out;
9795 	}
9796  out:
9797 	spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
9798 }
9799 
9800 /**
9801  * _scsih_create_enclosure_list_after_reset - Free Existing list,
9802  *	And create enclosure list by scanning all Enclosure Page(0)s
9803  * @ioc: per adapter object
9804  */
9805 static void
9806 _scsih_create_enclosure_list_after_reset(struct MPT3SAS_ADAPTER *ioc)
9807 {
9808 	struct _enclosure_node *enclosure_dev;
9809 	Mpi2ConfigReply_t mpi_reply;
9810 	u16 enclosure_handle;
9811 	int rc;
9812 
9813 	/* Free existing enclosure list */
9814 	mpt3sas_free_enclosure_list(ioc);
9815 
9816 	/* Re constructing enclosure list after reset*/
9817 	enclosure_handle = 0xFFFF;
9818 	do {
9819 		enclosure_dev =
9820 			kzalloc(sizeof(struct _enclosure_node), GFP_KERNEL);
9821 		if (!enclosure_dev) {
9822 			ioc_err(ioc, "failure at %s:%d/%s()!\n",
9823 				__FILE__, __LINE__, __func__);
9824 			return;
9825 		}
9826 		rc = mpt3sas_config_get_enclosure_pg0(ioc, &mpi_reply,
9827 				&enclosure_dev->pg0,
9828 				MPI2_SAS_ENCLOS_PGAD_FORM_GET_NEXT_HANDLE,
9829 				enclosure_handle);
9830 
9831 		if (rc || (le16_to_cpu(mpi_reply.IOCStatus) &
9832 						MPI2_IOCSTATUS_MASK)) {
9833 			kfree(enclosure_dev);
9834 			return;
9835 		}
9836 		list_add_tail(&enclosure_dev->list,
9837 						&ioc->enclosure_list);
9838 		enclosure_handle =
9839 			le16_to_cpu(enclosure_dev->pg0.EnclosureHandle);
9840 	} while (1);
9841 }
9842 
9843 /**
9844  * _scsih_search_responding_sas_devices -
9845  * @ioc: per adapter object
9846  *
9847  * After host reset, find out whether devices are still responding.
9848  * If not remove.
9849  */
9850 static void
9851 _scsih_search_responding_sas_devices(struct MPT3SAS_ADAPTER *ioc)
9852 {
9853 	Mpi2SasDevicePage0_t sas_device_pg0;
9854 	Mpi2ConfigReply_t mpi_reply;
9855 	u16 ioc_status;
9856 	u16 handle;
9857 	u32 device_info;
9858 
9859 	ioc_info(ioc, "search for end-devices: start\n");
9860 
9861 	if (list_empty(&ioc->sas_device_list))
9862 		goto out;
9863 
9864 	handle = 0xFFFF;
9865 	while (!(mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply,
9866 	    &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_GET_NEXT_HANDLE,
9867 	    handle))) {
9868 		ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
9869 		    MPI2_IOCSTATUS_MASK;
9870 		if (ioc_status != MPI2_IOCSTATUS_SUCCESS)
9871 			break;
9872 		handle = le16_to_cpu(sas_device_pg0.DevHandle);
9873 		device_info = le32_to_cpu(sas_device_pg0.DeviceInfo);
9874 		if (!(_scsih_is_end_device(device_info)))
9875 			continue;
9876 		_scsih_mark_responding_sas_device(ioc, &sas_device_pg0);
9877 	}
9878 
9879  out:
9880 	ioc_info(ioc, "search for end-devices: complete\n");
9881 }
9882 
9883 /**
9884  * _scsih_mark_responding_pcie_device - mark a pcie_device as responding
9885  * @ioc: per adapter object
9886  * @pcie_device_pg0: PCIe Device page 0
9887  *
9888  * After host reset, find out whether devices are still responding.
9889  * Used in _scsih_remove_unresponding_devices.
9890  */
9891 static void
9892 _scsih_mark_responding_pcie_device(struct MPT3SAS_ADAPTER *ioc,
9893 	Mpi26PCIeDevicePage0_t *pcie_device_pg0)
9894 {
9895 	struct MPT3SAS_TARGET *sas_target_priv_data = NULL;
9896 	struct scsi_target *starget;
9897 	struct _pcie_device *pcie_device;
9898 	unsigned long flags;
9899 
9900 	spin_lock_irqsave(&ioc->pcie_device_lock, flags);
9901 	list_for_each_entry(pcie_device, &ioc->pcie_device_list, list) {
9902 		if ((pcie_device->wwid == le64_to_cpu(pcie_device_pg0->WWID))
9903 		    && (pcie_device->slot == le16_to_cpu(
9904 		    pcie_device_pg0->Slot))) {
9905 			pcie_device->access_status =
9906 					pcie_device_pg0->AccessStatus;
9907 			pcie_device->responding = 1;
9908 			starget = pcie_device->starget;
9909 			if (starget && starget->hostdata) {
9910 				sas_target_priv_data = starget->hostdata;
9911 				sas_target_priv_data->tm_busy = 0;
9912 				sas_target_priv_data->deleted = 0;
9913 			} else
9914 				sas_target_priv_data = NULL;
9915 			if (starget) {
9916 				starget_printk(KERN_INFO, starget,
9917 				    "handle(0x%04x), wwid(0x%016llx) ",
9918 				    pcie_device->handle,
9919 				    (unsigned long long)pcie_device->wwid);
9920 				if (pcie_device->enclosure_handle != 0)
9921 					starget_printk(KERN_INFO, starget,
9922 					    "enclosure logical id(0x%016llx), "
9923 					    "slot(%d)\n",
9924 					    (unsigned long long)
9925 					    pcie_device->enclosure_logical_id,
9926 					    pcie_device->slot);
9927 			}
9928 
9929 			if (((le32_to_cpu(pcie_device_pg0->Flags)) &
9930 			    MPI26_PCIEDEV0_FLAGS_ENCL_LEVEL_VALID) &&
9931 			    (ioc->hba_mpi_version_belonged != MPI2_VERSION)) {
9932 				pcie_device->enclosure_level =
9933 				    pcie_device_pg0->EnclosureLevel;
9934 				memcpy(&pcie_device->connector_name[0],
9935 				    &pcie_device_pg0->ConnectorName[0], 4);
9936 			} else {
9937 				pcie_device->enclosure_level = 0;
9938 				pcie_device->connector_name[0] = '\0';
9939 			}
9940 
9941 			if (pcie_device->handle == le16_to_cpu(
9942 			    pcie_device_pg0->DevHandle))
9943 				goto out;
9944 			pr_info("\thandle changed from(0x%04x)!!!\n",
9945 			    pcie_device->handle);
9946 			pcie_device->handle = le16_to_cpu(
9947 			    pcie_device_pg0->DevHandle);
9948 			if (sas_target_priv_data)
9949 				sas_target_priv_data->handle =
9950 				    le16_to_cpu(pcie_device_pg0->DevHandle);
9951 			goto out;
9952 		}
9953 	}
9954 
9955  out:
9956 	spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
9957 }
9958 
9959 /**
9960  * _scsih_search_responding_pcie_devices -
9961  * @ioc: per adapter object
9962  *
9963  * After host reset, find out whether devices are still responding.
9964  * If not remove.
9965  */
9966 static void
9967 _scsih_search_responding_pcie_devices(struct MPT3SAS_ADAPTER *ioc)
9968 {
9969 	Mpi26PCIeDevicePage0_t pcie_device_pg0;
9970 	Mpi2ConfigReply_t mpi_reply;
9971 	u16 ioc_status;
9972 	u16 handle;
9973 	u32 device_info;
9974 
9975 	ioc_info(ioc, "search for end-devices: start\n");
9976 
9977 	if (list_empty(&ioc->pcie_device_list))
9978 		goto out;
9979 
9980 	handle = 0xFFFF;
9981 	while (!(mpt3sas_config_get_pcie_device_pg0(ioc, &mpi_reply,
9982 		&pcie_device_pg0, MPI26_PCIE_DEVICE_PGAD_FORM_GET_NEXT_HANDLE,
9983 		handle))) {
9984 		ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
9985 		    MPI2_IOCSTATUS_MASK;
9986 		if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
9987 			ioc_info(ioc, "\tbreak from %s: ioc_status(0x%04x), loginfo(0x%08x)\n",
9988 				 __func__, ioc_status,
9989 				 le32_to_cpu(mpi_reply.IOCLogInfo));
9990 			break;
9991 		}
9992 		handle = le16_to_cpu(pcie_device_pg0.DevHandle);
9993 		device_info = le32_to_cpu(pcie_device_pg0.DeviceInfo);
9994 		if (!(_scsih_is_nvme_pciescsi_device(device_info)))
9995 			continue;
9996 		_scsih_mark_responding_pcie_device(ioc, &pcie_device_pg0);
9997 	}
9998 out:
9999 	ioc_info(ioc, "search for PCIe end-devices: complete\n");
10000 }
10001 
10002 /**
10003  * _scsih_mark_responding_raid_device - mark a raid_device as responding
10004  * @ioc: per adapter object
10005  * @wwid: world wide identifier for raid volume
10006  * @handle: device handle
10007  *
10008  * After host reset, find out whether devices are still responding.
10009  * Used in _scsih_remove_unresponsive_raid_devices.
10010  */
10011 static void
10012 _scsih_mark_responding_raid_device(struct MPT3SAS_ADAPTER *ioc, u64 wwid,
10013 	u16 handle)
10014 {
10015 	struct MPT3SAS_TARGET *sas_target_priv_data = NULL;
10016 	struct scsi_target *starget;
10017 	struct _raid_device *raid_device;
10018 	unsigned long flags;
10019 
10020 	spin_lock_irqsave(&ioc->raid_device_lock, flags);
10021 	list_for_each_entry(raid_device, &ioc->raid_device_list, list) {
10022 		if (raid_device->wwid == wwid && raid_device->starget) {
10023 			starget = raid_device->starget;
10024 			if (starget && starget->hostdata) {
10025 				sas_target_priv_data = starget->hostdata;
10026 				sas_target_priv_data->deleted = 0;
10027 			} else
10028 				sas_target_priv_data = NULL;
10029 			raid_device->responding = 1;
10030 			spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
10031 			starget_printk(KERN_INFO, raid_device->starget,
10032 			    "handle(0x%04x), wwid(0x%016llx)\n", handle,
10033 			    (unsigned long long)raid_device->wwid);
10034 
10035 			/*
10036 			 * WARPDRIVE: The handles of the PDs might have changed
10037 			 * across the host reset so re-initialize the
10038 			 * required data for Direct IO
10039 			 */
10040 			mpt3sas_init_warpdrive_properties(ioc, raid_device);
10041 			spin_lock_irqsave(&ioc->raid_device_lock, flags);
10042 			if (raid_device->handle == handle) {
10043 				spin_unlock_irqrestore(&ioc->raid_device_lock,
10044 				    flags);
10045 				return;
10046 			}
10047 			pr_info("\thandle changed from(0x%04x)!!!\n",
10048 			    raid_device->handle);
10049 			raid_device->handle = handle;
10050 			if (sas_target_priv_data)
10051 				sas_target_priv_data->handle = handle;
10052 			spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
10053 			return;
10054 		}
10055 	}
10056 	spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
10057 }
10058 
10059 /**
10060  * _scsih_search_responding_raid_devices -
10061  * @ioc: per adapter object
10062  *
10063  * After host reset, find out whether devices are still responding.
10064  * If not remove.
10065  */
10066 static void
10067 _scsih_search_responding_raid_devices(struct MPT3SAS_ADAPTER *ioc)
10068 {
10069 	Mpi2RaidVolPage1_t volume_pg1;
10070 	Mpi2RaidVolPage0_t volume_pg0;
10071 	Mpi2RaidPhysDiskPage0_t pd_pg0;
10072 	Mpi2ConfigReply_t mpi_reply;
10073 	u16 ioc_status;
10074 	u16 handle;
10075 	u8 phys_disk_num;
10076 
10077 	if (!ioc->ir_firmware)
10078 		return;
10079 
10080 	ioc_info(ioc, "search for raid volumes: start\n");
10081 
10082 	if (list_empty(&ioc->raid_device_list))
10083 		goto out;
10084 
10085 	handle = 0xFFFF;
10086 	while (!(mpt3sas_config_get_raid_volume_pg1(ioc, &mpi_reply,
10087 	    &volume_pg1, MPI2_RAID_VOLUME_PGAD_FORM_GET_NEXT_HANDLE, handle))) {
10088 		ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
10089 		    MPI2_IOCSTATUS_MASK;
10090 		if (ioc_status != MPI2_IOCSTATUS_SUCCESS)
10091 			break;
10092 		handle = le16_to_cpu(volume_pg1.DevHandle);
10093 
10094 		if (mpt3sas_config_get_raid_volume_pg0(ioc, &mpi_reply,
10095 		    &volume_pg0, MPI2_RAID_VOLUME_PGAD_FORM_HANDLE, handle,
10096 		     sizeof(Mpi2RaidVolPage0_t)))
10097 			continue;
10098 
10099 		if (volume_pg0.VolumeState == MPI2_RAID_VOL_STATE_OPTIMAL ||
10100 		    volume_pg0.VolumeState == MPI2_RAID_VOL_STATE_ONLINE ||
10101 		    volume_pg0.VolumeState == MPI2_RAID_VOL_STATE_DEGRADED)
10102 			_scsih_mark_responding_raid_device(ioc,
10103 			    le64_to_cpu(volume_pg1.WWID), handle);
10104 	}
10105 
10106 	/* refresh the pd_handles */
10107 	if (!ioc->is_warpdrive) {
10108 		phys_disk_num = 0xFF;
10109 		memset(ioc->pd_handles, 0, ioc->pd_handles_sz);
10110 		while (!(mpt3sas_config_get_phys_disk_pg0(ioc, &mpi_reply,
10111 		    &pd_pg0, MPI2_PHYSDISK_PGAD_FORM_GET_NEXT_PHYSDISKNUM,
10112 		    phys_disk_num))) {
10113 			ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
10114 			    MPI2_IOCSTATUS_MASK;
10115 			if (ioc_status != MPI2_IOCSTATUS_SUCCESS)
10116 				break;
10117 			phys_disk_num = pd_pg0.PhysDiskNum;
10118 			handle = le16_to_cpu(pd_pg0.DevHandle);
10119 			set_bit(handle, ioc->pd_handles);
10120 		}
10121 	}
10122  out:
10123 	ioc_info(ioc, "search for responding raid volumes: complete\n");
10124 }
10125 
10126 /**
10127  * _scsih_mark_responding_expander - mark a expander as responding
10128  * @ioc: per adapter object
10129  * @expander_pg0:SAS Expander Config Page0
10130  *
10131  * After host reset, find out whether devices are still responding.
10132  * Used in _scsih_remove_unresponsive_expanders.
10133  */
10134 static void
10135 _scsih_mark_responding_expander(struct MPT3SAS_ADAPTER *ioc,
10136 	Mpi2ExpanderPage0_t *expander_pg0)
10137 {
10138 	struct _sas_node *sas_expander = NULL;
10139 	unsigned long flags;
10140 	int i;
10141 	struct _enclosure_node *enclosure_dev = NULL;
10142 	u16 handle = le16_to_cpu(expander_pg0->DevHandle);
10143 	u16 enclosure_handle = le16_to_cpu(expander_pg0->EnclosureHandle);
10144 	u64 sas_address = le64_to_cpu(expander_pg0->SASAddress);
10145 	struct hba_port *port = mpt3sas_get_port_by_id(
10146 	    ioc, expander_pg0->PhysicalPort, 0);
10147 
10148 	if (enclosure_handle)
10149 		enclosure_dev =
10150 			mpt3sas_scsih_enclosure_find_by_handle(ioc,
10151 							enclosure_handle);
10152 
10153 	spin_lock_irqsave(&ioc->sas_node_lock, flags);
10154 	list_for_each_entry(sas_expander, &ioc->sas_expander_list, list) {
10155 		if (sas_expander->sas_address != sas_address)
10156 			continue;
10157 		if (sas_expander->port != port)
10158 			continue;
10159 		sas_expander->responding = 1;
10160 
10161 		if (enclosure_dev) {
10162 			sas_expander->enclosure_logical_id =
10163 			    le64_to_cpu(enclosure_dev->pg0.EnclosureLogicalID);
10164 			sas_expander->enclosure_handle =
10165 			    le16_to_cpu(expander_pg0->EnclosureHandle);
10166 		}
10167 
10168 		if (sas_expander->handle == handle)
10169 			goto out;
10170 		pr_info("\texpander(0x%016llx): handle changed" \
10171 		    " from(0x%04x) to (0x%04x)!!!\n",
10172 		    (unsigned long long)sas_expander->sas_address,
10173 		    sas_expander->handle, handle);
10174 		sas_expander->handle = handle;
10175 		for (i = 0 ; i < sas_expander->num_phys ; i++)
10176 			sas_expander->phy[i].handle = handle;
10177 		goto out;
10178 	}
10179  out:
10180 	spin_unlock_irqrestore(&ioc->sas_node_lock, flags);
10181 }
10182 
10183 /**
10184  * _scsih_search_responding_expanders -
10185  * @ioc: per adapter object
10186  *
10187  * After host reset, find out whether devices are still responding.
10188  * If not remove.
10189  */
10190 static void
10191 _scsih_search_responding_expanders(struct MPT3SAS_ADAPTER *ioc)
10192 {
10193 	Mpi2ExpanderPage0_t expander_pg0;
10194 	Mpi2ConfigReply_t mpi_reply;
10195 	u16 ioc_status;
10196 	u64 sas_address;
10197 	u16 handle;
10198 	u8 port;
10199 
10200 	ioc_info(ioc, "search for expanders: start\n");
10201 
10202 	if (list_empty(&ioc->sas_expander_list))
10203 		goto out;
10204 
10205 	handle = 0xFFFF;
10206 	while (!(mpt3sas_config_get_expander_pg0(ioc, &mpi_reply, &expander_pg0,
10207 	    MPI2_SAS_EXPAND_PGAD_FORM_GET_NEXT_HNDL, handle))) {
10208 
10209 		ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
10210 		    MPI2_IOCSTATUS_MASK;
10211 		if (ioc_status != MPI2_IOCSTATUS_SUCCESS)
10212 			break;
10213 
10214 		handle = le16_to_cpu(expander_pg0.DevHandle);
10215 		sas_address = le64_to_cpu(expander_pg0.SASAddress);
10216 		port = expander_pg0.PhysicalPort;
10217 		pr_info(
10218 		    "\texpander present: handle(0x%04x), sas_addr(0x%016llx), port:%d\n",
10219 		    handle, (unsigned long long)sas_address,
10220 		    (ioc->multipath_on_hba ?
10221 		    port : MULTIPATH_DISABLED_PORT_ID));
10222 		_scsih_mark_responding_expander(ioc, &expander_pg0);
10223 	}
10224 
10225  out:
10226 	ioc_info(ioc, "search for expanders: complete\n");
10227 }
10228 
10229 /**
10230  * _scsih_remove_unresponding_devices - removing unresponding devices
10231  * @ioc: per adapter object
10232  */
10233 static void
10234 _scsih_remove_unresponding_devices(struct MPT3SAS_ADAPTER *ioc)
10235 {
10236 	struct _sas_device *sas_device, *sas_device_next;
10237 	struct _sas_node *sas_expander, *sas_expander_next;
10238 	struct _raid_device *raid_device, *raid_device_next;
10239 	struct _pcie_device *pcie_device, *pcie_device_next;
10240 	struct list_head tmp_list;
10241 	unsigned long flags;
10242 	LIST_HEAD(head);
10243 
10244 	ioc_info(ioc, "removing unresponding devices: start\n");
10245 
10246 	/* removing unresponding end devices */
10247 	ioc_info(ioc, "removing unresponding devices: end-devices\n");
10248 	/*
10249 	 * Iterate, pulling off devices marked as non-responding. We become the
10250 	 * owner for the reference the list had on any object we prune.
10251 	 */
10252 	spin_lock_irqsave(&ioc->sas_device_lock, flags);
10253 
10254 	/*
10255 	 * Clean up the sas_device_init_list list as
10256 	 * driver goes for fresh scan as part of diag reset.
10257 	 */
10258 	list_for_each_entry_safe(sas_device, sas_device_next,
10259 	    &ioc->sas_device_init_list, list) {
10260 		list_del_init(&sas_device->list);
10261 		sas_device_put(sas_device);
10262 	}
10263 
10264 	list_for_each_entry_safe(sas_device, sas_device_next,
10265 	    &ioc->sas_device_list, list) {
10266 		if (!sas_device->responding)
10267 			list_move_tail(&sas_device->list, &head);
10268 		else
10269 			sas_device->responding = 0;
10270 	}
10271 	spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
10272 
10273 	/*
10274 	 * Now, uninitialize and remove the unresponding devices we pruned.
10275 	 */
10276 	list_for_each_entry_safe(sas_device, sas_device_next, &head, list) {
10277 		_scsih_remove_device(ioc, sas_device);
10278 		list_del_init(&sas_device->list);
10279 		sas_device_put(sas_device);
10280 	}
10281 
10282 	ioc_info(ioc, "Removing unresponding devices: pcie end-devices\n");
10283 	INIT_LIST_HEAD(&head);
10284 	spin_lock_irqsave(&ioc->pcie_device_lock, flags);
10285 	/*
10286 	 * Clean up the pcie_device_init_list list as
10287 	 * driver goes for fresh scan as part of diag reset.
10288 	 */
10289 	list_for_each_entry_safe(pcie_device, pcie_device_next,
10290 	    &ioc->pcie_device_init_list, list) {
10291 		list_del_init(&pcie_device->list);
10292 		pcie_device_put(pcie_device);
10293 	}
10294 
10295 	list_for_each_entry_safe(pcie_device, pcie_device_next,
10296 	    &ioc->pcie_device_list, list) {
10297 		if (!pcie_device->responding)
10298 			list_move_tail(&pcie_device->list, &head);
10299 		else
10300 			pcie_device->responding = 0;
10301 	}
10302 	spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
10303 
10304 	list_for_each_entry_safe(pcie_device, pcie_device_next, &head, list) {
10305 		_scsih_pcie_device_remove_from_sml(ioc, pcie_device);
10306 		list_del_init(&pcie_device->list);
10307 		pcie_device_put(pcie_device);
10308 	}
10309 
10310 	/* removing unresponding volumes */
10311 	if (ioc->ir_firmware) {
10312 		ioc_info(ioc, "removing unresponding devices: volumes\n");
10313 		list_for_each_entry_safe(raid_device, raid_device_next,
10314 		    &ioc->raid_device_list, list) {
10315 			if (!raid_device->responding)
10316 				_scsih_sas_volume_delete(ioc,
10317 				    raid_device->handle);
10318 			else
10319 				raid_device->responding = 0;
10320 		}
10321 	}
10322 
10323 	/* removing unresponding expanders */
10324 	ioc_info(ioc, "removing unresponding devices: expanders\n");
10325 	spin_lock_irqsave(&ioc->sas_node_lock, flags);
10326 	INIT_LIST_HEAD(&tmp_list);
10327 	list_for_each_entry_safe(sas_expander, sas_expander_next,
10328 	    &ioc->sas_expander_list, list) {
10329 		if (!sas_expander->responding)
10330 			list_move_tail(&sas_expander->list, &tmp_list);
10331 		else
10332 			sas_expander->responding = 0;
10333 	}
10334 	spin_unlock_irqrestore(&ioc->sas_node_lock, flags);
10335 	list_for_each_entry_safe(sas_expander, sas_expander_next, &tmp_list,
10336 	    list) {
10337 		_scsih_expander_node_remove(ioc, sas_expander);
10338 	}
10339 
10340 	ioc_info(ioc, "removing unresponding devices: complete\n");
10341 
10342 	/* unblock devices */
10343 	_scsih_ublock_io_all_device(ioc);
10344 }
10345 
10346 static void
10347 _scsih_refresh_expander_links(struct MPT3SAS_ADAPTER *ioc,
10348 	struct _sas_node *sas_expander, u16 handle)
10349 {
10350 	Mpi2ExpanderPage1_t expander_pg1;
10351 	Mpi2ConfigReply_t mpi_reply;
10352 	int i;
10353 
10354 	for (i = 0 ; i < sas_expander->num_phys ; i++) {
10355 		if ((mpt3sas_config_get_expander_pg1(ioc, &mpi_reply,
10356 		    &expander_pg1, i, handle))) {
10357 			ioc_err(ioc, "failure at %s:%d/%s()!\n",
10358 				__FILE__, __LINE__, __func__);
10359 			return;
10360 		}
10361 
10362 		mpt3sas_transport_update_links(ioc, sas_expander->sas_address,
10363 		    le16_to_cpu(expander_pg1.AttachedDevHandle), i,
10364 		    expander_pg1.NegotiatedLinkRate >> 4,
10365 		    sas_expander->port);
10366 	}
10367 }
10368 
10369 /**
10370  * _scsih_scan_for_devices_after_reset - scan for devices after host reset
10371  * @ioc: per adapter object
10372  */
10373 static void
10374 _scsih_scan_for_devices_after_reset(struct MPT3SAS_ADAPTER *ioc)
10375 {
10376 	Mpi2ExpanderPage0_t expander_pg0;
10377 	Mpi2SasDevicePage0_t sas_device_pg0;
10378 	Mpi26PCIeDevicePage0_t pcie_device_pg0;
10379 	Mpi2RaidVolPage1_t *volume_pg1;
10380 	Mpi2RaidVolPage0_t *volume_pg0;
10381 	Mpi2RaidPhysDiskPage0_t pd_pg0;
10382 	Mpi2EventIrConfigElement_t element;
10383 	Mpi2ConfigReply_t mpi_reply;
10384 	u8 phys_disk_num, port_id;
10385 	u16 ioc_status;
10386 	u16 handle, parent_handle;
10387 	u64 sas_address;
10388 	struct _sas_device *sas_device;
10389 	struct _pcie_device *pcie_device;
10390 	struct _sas_node *expander_device;
10391 	static struct _raid_device *raid_device;
10392 	u8 retry_count;
10393 	unsigned long flags;
10394 
10395 	volume_pg0 = kzalloc(sizeof(*volume_pg0), GFP_KERNEL);
10396 	if (!volume_pg0)
10397 		return;
10398 
10399 	volume_pg1 = kzalloc(sizeof(*volume_pg1), GFP_KERNEL);
10400 	if (!volume_pg1) {
10401 		kfree(volume_pg0);
10402 		return;
10403 	}
10404 
10405 	ioc_info(ioc, "scan devices: start\n");
10406 
10407 	_scsih_sas_host_refresh(ioc);
10408 
10409 	ioc_info(ioc, "\tscan devices: expanders start\n");
10410 
10411 	/* expanders */
10412 	handle = 0xFFFF;
10413 	while (!(mpt3sas_config_get_expander_pg0(ioc, &mpi_reply, &expander_pg0,
10414 	    MPI2_SAS_EXPAND_PGAD_FORM_GET_NEXT_HNDL, handle))) {
10415 		ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
10416 		    MPI2_IOCSTATUS_MASK;
10417 		if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
10418 			ioc_info(ioc, "\tbreak from expander scan: ioc_status(0x%04x), loginfo(0x%08x)\n",
10419 				 ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo));
10420 			break;
10421 		}
10422 		handle = le16_to_cpu(expander_pg0.DevHandle);
10423 		spin_lock_irqsave(&ioc->sas_node_lock, flags);
10424 		port_id = expander_pg0.PhysicalPort;
10425 		expander_device = mpt3sas_scsih_expander_find_by_sas_address(
10426 		    ioc, le64_to_cpu(expander_pg0.SASAddress),
10427 		    mpt3sas_get_port_by_id(ioc, port_id, 0));
10428 		spin_unlock_irqrestore(&ioc->sas_node_lock, flags);
10429 		if (expander_device)
10430 			_scsih_refresh_expander_links(ioc, expander_device,
10431 			    handle);
10432 		else {
10433 			ioc_info(ioc, "\tBEFORE adding expander: handle (0x%04x), sas_addr(0x%016llx)\n",
10434 				 handle,
10435 				 (u64)le64_to_cpu(expander_pg0.SASAddress));
10436 			_scsih_expander_add(ioc, handle);
10437 			ioc_info(ioc, "\tAFTER adding expander: handle (0x%04x), sas_addr(0x%016llx)\n",
10438 				 handle,
10439 				 (u64)le64_to_cpu(expander_pg0.SASAddress));
10440 		}
10441 	}
10442 
10443 	ioc_info(ioc, "\tscan devices: expanders complete\n");
10444 
10445 	if (!ioc->ir_firmware)
10446 		goto skip_to_sas;
10447 
10448 	ioc_info(ioc, "\tscan devices: phys disk start\n");
10449 
10450 	/* phys disk */
10451 	phys_disk_num = 0xFF;
10452 	while (!(mpt3sas_config_get_phys_disk_pg0(ioc, &mpi_reply,
10453 	    &pd_pg0, MPI2_PHYSDISK_PGAD_FORM_GET_NEXT_PHYSDISKNUM,
10454 	    phys_disk_num))) {
10455 		ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
10456 		    MPI2_IOCSTATUS_MASK;
10457 		if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
10458 			ioc_info(ioc, "\tbreak from phys disk scan: ioc_status(0x%04x), loginfo(0x%08x)\n",
10459 				 ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo));
10460 			break;
10461 		}
10462 		phys_disk_num = pd_pg0.PhysDiskNum;
10463 		handle = le16_to_cpu(pd_pg0.DevHandle);
10464 		sas_device = mpt3sas_get_sdev_by_handle(ioc, handle);
10465 		if (sas_device) {
10466 			sas_device_put(sas_device);
10467 			continue;
10468 		}
10469 		if (mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply,
10470 		    &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE,
10471 		    handle) != 0)
10472 			continue;
10473 		ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
10474 		    MPI2_IOCSTATUS_MASK;
10475 		if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
10476 			ioc_info(ioc, "\tbreak from phys disk scan ioc_status(0x%04x), loginfo(0x%08x)\n",
10477 				 ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo));
10478 			break;
10479 		}
10480 		parent_handle = le16_to_cpu(sas_device_pg0.ParentDevHandle);
10481 		if (!_scsih_get_sas_address(ioc, parent_handle,
10482 		    &sas_address)) {
10483 			ioc_info(ioc, "\tBEFORE adding phys disk: handle (0x%04x), sas_addr(0x%016llx)\n",
10484 				 handle,
10485 				 (u64)le64_to_cpu(sas_device_pg0.SASAddress));
10486 			port_id = sas_device_pg0.PhysicalPort;
10487 			mpt3sas_transport_update_links(ioc, sas_address,
10488 			    handle, sas_device_pg0.PhyNum,
10489 			    MPI2_SAS_NEG_LINK_RATE_1_5,
10490 			    mpt3sas_get_port_by_id(ioc, port_id, 0));
10491 			set_bit(handle, ioc->pd_handles);
10492 			retry_count = 0;
10493 			/* This will retry adding the end device.
10494 			 * _scsih_add_device() will decide on retries and
10495 			 * return "1" when it should be retried
10496 			 */
10497 			while (_scsih_add_device(ioc, handle, retry_count++,
10498 			    1)) {
10499 				ssleep(1);
10500 			}
10501 			ioc_info(ioc, "\tAFTER adding phys disk: handle (0x%04x), sas_addr(0x%016llx)\n",
10502 				 handle,
10503 				 (u64)le64_to_cpu(sas_device_pg0.SASAddress));
10504 		}
10505 	}
10506 
10507 	ioc_info(ioc, "\tscan devices: phys disk complete\n");
10508 
10509 	ioc_info(ioc, "\tscan devices: volumes start\n");
10510 
10511 	/* volumes */
10512 	handle = 0xFFFF;
10513 	while (!(mpt3sas_config_get_raid_volume_pg1(ioc, &mpi_reply,
10514 	    volume_pg1, MPI2_RAID_VOLUME_PGAD_FORM_GET_NEXT_HANDLE, handle))) {
10515 		ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
10516 		    MPI2_IOCSTATUS_MASK;
10517 		if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
10518 			ioc_info(ioc, "\tbreak from volume scan: ioc_status(0x%04x), loginfo(0x%08x)\n",
10519 				 ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo));
10520 			break;
10521 		}
10522 		handle = le16_to_cpu(volume_pg1->DevHandle);
10523 		spin_lock_irqsave(&ioc->raid_device_lock, flags);
10524 		raid_device = _scsih_raid_device_find_by_wwid(ioc,
10525 		    le64_to_cpu(volume_pg1->WWID));
10526 		spin_unlock_irqrestore(&ioc->raid_device_lock, flags);
10527 		if (raid_device)
10528 			continue;
10529 		if (mpt3sas_config_get_raid_volume_pg0(ioc, &mpi_reply,
10530 		    volume_pg0, MPI2_RAID_VOLUME_PGAD_FORM_HANDLE, handle,
10531 		     sizeof(Mpi2RaidVolPage0_t)))
10532 			continue;
10533 		ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
10534 		    MPI2_IOCSTATUS_MASK;
10535 		if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
10536 			ioc_info(ioc, "\tbreak from volume scan: ioc_status(0x%04x), loginfo(0x%08x)\n",
10537 				 ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo));
10538 			break;
10539 		}
10540 		if (volume_pg0->VolumeState == MPI2_RAID_VOL_STATE_OPTIMAL ||
10541 		    volume_pg0->VolumeState == MPI2_RAID_VOL_STATE_ONLINE ||
10542 		    volume_pg0->VolumeState == MPI2_RAID_VOL_STATE_DEGRADED) {
10543 			memset(&element, 0, sizeof(Mpi2EventIrConfigElement_t));
10544 			element.ReasonCode = MPI2_EVENT_IR_CHANGE_RC_ADDED;
10545 			element.VolDevHandle = volume_pg1->DevHandle;
10546 			ioc_info(ioc, "\tBEFORE adding volume: handle (0x%04x)\n",
10547 				 volume_pg1->DevHandle);
10548 			_scsih_sas_volume_add(ioc, &element);
10549 			ioc_info(ioc, "\tAFTER adding volume: handle (0x%04x)\n",
10550 				 volume_pg1->DevHandle);
10551 		}
10552 	}
10553 
10554 	ioc_info(ioc, "\tscan devices: volumes complete\n");
10555 
10556  skip_to_sas:
10557 
10558 	ioc_info(ioc, "\tscan devices: end devices start\n");
10559 
10560 	/* sas devices */
10561 	handle = 0xFFFF;
10562 	while (!(mpt3sas_config_get_sas_device_pg0(ioc, &mpi_reply,
10563 	    &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_GET_NEXT_HANDLE,
10564 	    handle))) {
10565 		ioc_status = le16_to_cpu(mpi_reply.IOCStatus) &
10566 		    MPI2_IOCSTATUS_MASK;
10567 		if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
10568 			ioc_info(ioc, "\tbreak from end device scan: ioc_status(0x%04x), loginfo(0x%08x)\n",
10569 				 ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo));
10570 			break;
10571 		}
10572 		handle = le16_to_cpu(sas_device_pg0.DevHandle);
10573 		if (!(_scsih_is_end_device(
10574 		    le32_to_cpu(sas_device_pg0.DeviceInfo))))
10575 			continue;
10576 		port_id = sas_device_pg0.PhysicalPort;
10577 		sas_device = mpt3sas_get_sdev_by_addr(ioc,
10578 		    le64_to_cpu(sas_device_pg0.SASAddress),
10579 		    mpt3sas_get_port_by_id(ioc, port_id, 0));
10580 		if (sas_device) {
10581 			sas_device_put(sas_device);
10582 			continue;
10583 		}
10584 		parent_handle = le16_to_cpu(sas_device_pg0.ParentDevHandle);
10585 		if (!_scsih_get_sas_address(ioc, parent_handle, &sas_address)) {
10586 			ioc_info(ioc, "\tBEFORE adding end device: handle (0x%04x), sas_addr(0x%016llx)\n",
10587 				 handle,
10588 				 (u64)le64_to_cpu(sas_device_pg0.SASAddress));
10589 			mpt3sas_transport_update_links(ioc, sas_address, handle,
10590 			    sas_device_pg0.PhyNum, MPI2_SAS_NEG_LINK_RATE_1_5,
10591 			    mpt3sas_get_port_by_id(ioc, port_id, 0));
10592 			retry_count = 0;
10593 			/* This will retry adding the end device.
10594 			 * _scsih_add_device() will decide on retries and
10595 			 * return "1" when it should be retried
10596 			 */
10597 			while (_scsih_add_device(ioc, handle, retry_count++,
10598 			    0)) {
10599 				ssleep(1);
10600 			}
10601 			ioc_info(ioc, "\tAFTER adding end device: handle (0x%04x), sas_addr(0x%016llx)\n",
10602 				 handle,
10603 				 (u64)le64_to_cpu(sas_device_pg0.SASAddress));
10604 		}
10605 	}
10606 	ioc_info(ioc, "\tscan devices: end devices complete\n");
10607 	ioc_info(ioc, "\tscan devices: pcie end devices start\n");
10608 
10609 	/* pcie devices */
10610 	handle = 0xFFFF;
10611 	while (!(mpt3sas_config_get_pcie_device_pg0(ioc, &mpi_reply,
10612 		&pcie_device_pg0, MPI26_PCIE_DEVICE_PGAD_FORM_GET_NEXT_HANDLE,
10613 		handle))) {
10614 		ioc_status = le16_to_cpu(mpi_reply.IOCStatus)
10615 				& MPI2_IOCSTATUS_MASK;
10616 		if (ioc_status != MPI2_IOCSTATUS_SUCCESS) {
10617 			ioc_info(ioc, "\tbreak from pcie end device scan: ioc_status(0x%04x), loginfo(0x%08x)\n",
10618 				 ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo));
10619 			break;
10620 		}
10621 		handle = le16_to_cpu(pcie_device_pg0.DevHandle);
10622 		if (!(_scsih_is_nvme_pciescsi_device(
10623 			le32_to_cpu(pcie_device_pg0.DeviceInfo))))
10624 			continue;
10625 		pcie_device = mpt3sas_get_pdev_by_wwid(ioc,
10626 				le64_to_cpu(pcie_device_pg0.WWID));
10627 		if (pcie_device) {
10628 			pcie_device_put(pcie_device);
10629 			continue;
10630 		}
10631 		retry_count = 0;
10632 		parent_handle = le16_to_cpu(pcie_device_pg0.ParentDevHandle);
10633 		_scsih_pcie_add_device(ioc, handle);
10634 
10635 		ioc_info(ioc, "\tAFTER adding pcie end device: handle (0x%04x), wwid(0x%016llx)\n",
10636 			 handle, (u64)le64_to_cpu(pcie_device_pg0.WWID));
10637 	}
10638 
10639 	kfree(volume_pg0);
10640 	kfree(volume_pg1);
10641 
10642 	ioc_info(ioc, "\tpcie devices: pcie end devices complete\n");
10643 	ioc_info(ioc, "scan devices: complete\n");
10644 }
10645 
10646 /**
10647  * mpt3sas_scsih_pre_reset_handler - reset callback handler (for scsih)
10648  * @ioc: per adapter object
10649  *
10650  * The handler for doing any required cleanup or initialization.
10651  */
10652 void mpt3sas_scsih_pre_reset_handler(struct MPT3SAS_ADAPTER *ioc)
10653 {
10654 	dtmprintk(ioc, ioc_info(ioc, "%s: MPT3_IOC_PRE_RESET\n", __func__));
10655 }
10656 
10657 /**
10658  * mpt3sas_scsih_clear_outstanding_scsi_tm_commands - clears outstanding
10659  *							scsi & tm cmds.
10660  * @ioc: per adapter object
10661  *
10662  * The handler for doing any required cleanup or initialization.
10663  */
10664 void
10665 mpt3sas_scsih_clear_outstanding_scsi_tm_commands(struct MPT3SAS_ADAPTER *ioc)
10666 {
10667 	dtmprintk(ioc,
10668 	    ioc_info(ioc, "%s: clear outstanding scsi & tm cmds\n", __func__));
10669 	if (ioc->scsih_cmds.status & MPT3_CMD_PENDING) {
10670 		ioc->scsih_cmds.status |= MPT3_CMD_RESET;
10671 		mpt3sas_base_free_smid(ioc, ioc->scsih_cmds.smid);
10672 		complete(&ioc->scsih_cmds.done);
10673 	}
10674 	if (ioc->tm_cmds.status & MPT3_CMD_PENDING) {
10675 		ioc->tm_cmds.status |= MPT3_CMD_RESET;
10676 		mpt3sas_base_free_smid(ioc, ioc->tm_cmds.smid);
10677 		complete(&ioc->tm_cmds.done);
10678 	}
10679 
10680 	memset(ioc->pend_os_device_add, 0, ioc->pend_os_device_add_sz);
10681 	memset(ioc->device_remove_in_progress, 0,
10682 	       ioc->device_remove_in_progress_sz);
10683 	_scsih_fw_event_cleanup_queue(ioc);
10684 	_scsih_flush_running_cmds(ioc);
10685 }
10686 
10687 /**
10688  * mpt3sas_scsih_reset_done_handler - reset callback handler (for scsih)
10689  * @ioc: per adapter object
10690  *
10691  * The handler for doing any required cleanup or initialization.
10692  */
10693 void
10694 mpt3sas_scsih_reset_done_handler(struct MPT3SAS_ADAPTER *ioc)
10695 {
10696 	dtmprintk(ioc, ioc_info(ioc, "%s: MPT3_IOC_DONE_RESET\n", __func__));
10697 	if (!(disable_discovery > 0 && !ioc->sas_hba.num_phys)) {
10698 		if (ioc->multipath_on_hba) {
10699 			_scsih_sas_port_refresh(ioc);
10700 			_scsih_update_vphys_after_reset(ioc);
10701 		}
10702 		_scsih_prep_device_scan(ioc);
10703 		_scsih_create_enclosure_list_after_reset(ioc);
10704 		_scsih_search_responding_sas_devices(ioc);
10705 		_scsih_search_responding_pcie_devices(ioc);
10706 		_scsih_search_responding_raid_devices(ioc);
10707 		_scsih_search_responding_expanders(ioc);
10708 		_scsih_error_recovery_delete_devices(ioc);
10709 	}
10710 }
10711 
10712 /**
10713  * _mpt3sas_fw_work - delayed task for processing firmware events
10714  * @ioc: per adapter object
10715  * @fw_event: The fw_event_work object
10716  * Context: user.
10717  */
10718 static void
10719 _mpt3sas_fw_work(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work *fw_event)
10720 {
10721 	ioc->current_event = fw_event;
10722 	_scsih_fw_event_del_from_list(ioc, fw_event);
10723 
10724 	/* the queue is being flushed so ignore this event */
10725 	if (ioc->remove_host || ioc->pci_error_recovery) {
10726 		fw_event_work_put(fw_event);
10727 		ioc->current_event = NULL;
10728 		return;
10729 	}
10730 
10731 	switch (fw_event->event) {
10732 	case MPT3SAS_PROCESS_TRIGGER_DIAG:
10733 		mpt3sas_process_trigger_data(ioc,
10734 			(struct SL_WH_TRIGGERS_EVENT_DATA_T *)
10735 			fw_event->event_data);
10736 		break;
10737 	case MPT3SAS_REMOVE_UNRESPONDING_DEVICES:
10738 		while (scsi_host_in_recovery(ioc->shost) ||
10739 					 ioc->shost_recovery) {
10740 			/*
10741 			 * If we're unloading or cancelling the work, bail.
10742 			 * Otherwise, this can become an infinite loop.
10743 			 */
10744 			if (ioc->remove_host || ioc->fw_events_cleanup)
10745 				goto out;
10746 			ssleep(1);
10747 		}
10748 		_scsih_remove_unresponding_devices(ioc);
10749 		_scsih_del_dirty_vphy(ioc);
10750 		_scsih_del_dirty_port_entries(ioc);
10751 		if (ioc->is_gen35_ioc)
10752 			_scsih_update_device_qdepth(ioc);
10753 		_scsih_scan_for_devices_after_reset(ioc);
10754 		/*
10755 		 * If diag reset has occurred during the driver load
10756 		 * then driver has to complete the driver load operation
10757 		 * by executing the following items:
10758 		 *- Register the devices from sas_device_init_list to SML
10759 		 *- clear is_driver_loading flag,
10760 		 *- start the watchdog thread.
10761 		 * In happy driver load path, above things are taken care of when
10762 		 * driver executes scsih_scan_finished().
10763 		 */
10764 		if (ioc->is_driver_loading)
10765 			_scsih_complete_devices_scanning(ioc);
10766 		_scsih_set_nvme_max_shutdown_latency(ioc);
10767 		break;
10768 	case MPT3SAS_PORT_ENABLE_COMPLETE:
10769 		ioc->start_scan = 0;
10770 		if (missing_delay[0] != -1 && missing_delay[1] != -1)
10771 			mpt3sas_base_update_missing_delay(ioc, missing_delay[0],
10772 			    missing_delay[1]);
10773 		dewtprintk(ioc,
10774 			   ioc_info(ioc, "port enable: complete from worker thread\n"));
10775 		break;
10776 	case MPT3SAS_TURN_ON_PFA_LED:
10777 		_scsih_turn_on_pfa_led(ioc, fw_event->device_handle);
10778 		break;
10779 	case MPI2_EVENT_SAS_TOPOLOGY_CHANGE_LIST:
10780 		_scsih_sas_topology_change_event(ioc, fw_event);
10781 		break;
10782 	case MPI2_EVENT_SAS_DEVICE_STATUS_CHANGE:
10783 		if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK)
10784 			_scsih_sas_device_status_change_event_debug(ioc,
10785 			    (Mpi2EventDataSasDeviceStatusChange_t *)
10786 			    fw_event->event_data);
10787 		break;
10788 	case MPI2_EVENT_SAS_DISCOVERY:
10789 		_scsih_sas_discovery_event(ioc, fw_event);
10790 		break;
10791 	case MPI2_EVENT_SAS_DEVICE_DISCOVERY_ERROR:
10792 		_scsih_sas_device_discovery_error_event(ioc, fw_event);
10793 		break;
10794 	case MPI2_EVENT_SAS_BROADCAST_PRIMITIVE:
10795 		_scsih_sas_broadcast_primitive_event(ioc, fw_event);
10796 		break;
10797 	case MPI2_EVENT_SAS_ENCL_DEVICE_STATUS_CHANGE:
10798 		_scsih_sas_enclosure_dev_status_change_event(ioc,
10799 		    fw_event);
10800 		break;
10801 	case MPI2_EVENT_IR_CONFIGURATION_CHANGE_LIST:
10802 		_scsih_sas_ir_config_change_event(ioc, fw_event);
10803 		break;
10804 	case MPI2_EVENT_IR_VOLUME:
10805 		_scsih_sas_ir_volume_event(ioc, fw_event);
10806 		break;
10807 	case MPI2_EVENT_IR_PHYSICAL_DISK:
10808 		_scsih_sas_ir_physical_disk_event(ioc, fw_event);
10809 		break;
10810 	case MPI2_EVENT_IR_OPERATION_STATUS:
10811 		_scsih_sas_ir_operation_status_event(ioc, fw_event);
10812 		break;
10813 	case MPI2_EVENT_PCIE_DEVICE_STATUS_CHANGE:
10814 		_scsih_pcie_device_status_change_event(ioc, fw_event);
10815 		break;
10816 	case MPI2_EVENT_PCIE_ENUMERATION:
10817 		_scsih_pcie_enumeration_event(ioc, fw_event);
10818 		break;
10819 	case MPI2_EVENT_PCIE_TOPOLOGY_CHANGE_LIST:
10820 		_scsih_pcie_topology_change_event(ioc, fw_event);
10821 		ioc->current_event = NULL;
10822 		return;
10823 	}
10824 out:
10825 	fw_event_work_put(fw_event);
10826 	ioc->current_event = NULL;
10827 }
10828 
10829 /**
10830  * _firmware_event_work
10831  * @work: The fw_event_work object
10832  * Context: user.
10833  *
10834  * wrappers for the work thread handling firmware events
10835  */
10836 
10837 static void
10838 _firmware_event_work(struct work_struct *work)
10839 {
10840 	struct fw_event_work *fw_event = container_of(work,
10841 	    struct fw_event_work, work);
10842 
10843 	_mpt3sas_fw_work(fw_event->ioc, fw_event);
10844 }
10845 
10846 /**
10847  * mpt3sas_scsih_event_callback - firmware event handler (called at ISR time)
10848  * @ioc: per adapter object
10849  * @msix_index: MSIX table index supplied by the OS
10850  * @reply: reply message frame(lower 32bit addr)
10851  * Context: interrupt.
10852  *
10853  * This function merely adds a new work task into ioc->firmware_event_thread.
10854  * The tasks are worked from _firmware_event_work in user context.
10855  *
10856  * Return: 1 meaning mf should be freed from _base_interrupt
10857  *         0 means the mf is freed from this function.
10858  */
10859 u8
10860 mpt3sas_scsih_event_callback(struct MPT3SAS_ADAPTER *ioc, u8 msix_index,
10861 	u32 reply)
10862 {
10863 	struct fw_event_work *fw_event;
10864 	Mpi2EventNotificationReply_t *mpi_reply;
10865 	u16 event;
10866 	u16 sz;
10867 	Mpi26EventDataActiveCableExcept_t *ActiveCableEventData;
10868 
10869 	/* events turned off due to host reset */
10870 	if (ioc->pci_error_recovery)
10871 		return 1;
10872 
10873 	mpi_reply = mpt3sas_base_get_reply_virt_addr(ioc, reply);
10874 
10875 	if (unlikely(!mpi_reply)) {
10876 		ioc_err(ioc, "mpi_reply not valid at %s:%d/%s()!\n",
10877 			__FILE__, __LINE__, __func__);
10878 		return 1;
10879 	}
10880 
10881 	event = le16_to_cpu(mpi_reply->Event);
10882 
10883 	if (event != MPI2_EVENT_LOG_ENTRY_ADDED)
10884 		mpt3sas_trigger_event(ioc, event, 0);
10885 
10886 	switch (event) {
10887 	/* handle these */
10888 	case MPI2_EVENT_SAS_BROADCAST_PRIMITIVE:
10889 	{
10890 		Mpi2EventDataSasBroadcastPrimitive_t *baen_data =
10891 		    (Mpi2EventDataSasBroadcastPrimitive_t *)
10892 		    mpi_reply->EventData;
10893 
10894 		if (baen_data->Primitive !=
10895 		    MPI2_EVENT_PRIMITIVE_ASYNCHRONOUS_EVENT)
10896 			return 1;
10897 
10898 		if (ioc->broadcast_aen_busy) {
10899 			ioc->broadcast_aen_pending++;
10900 			return 1;
10901 		} else
10902 			ioc->broadcast_aen_busy = 1;
10903 		break;
10904 	}
10905 
10906 	case MPI2_EVENT_SAS_TOPOLOGY_CHANGE_LIST:
10907 		_scsih_check_topo_delete_events(ioc,
10908 		    (Mpi2EventDataSasTopologyChangeList_t *)
10909 		    mpi_reply->EventData);
10910 		/*
10911 		 * No need to add the topology change list
10912 		 * event to fw event work queue when
10913 		 * diag reset is going on. Since during diag
10914 		 * reset driver scan the devices by reading
10915 		 * sas device page0's not by processing the
10916 		 * events.
10917 		 */
10918 		if (ioc->shost_recovery)
10919 			return 1;
10920 		break;
10921 	case MPI2_EVENT_PCIE_TOPOLOGY_CHANGE_LIST:
10922 	_scsih_check_pcie_topo_remove_events(ioc,
10923 		    (Mpi26EventDataPCIeTopologyChangeList_t *)
10924 		    mpi_reply->EventData);
10925 		if (ioc->shost_recovery)
10926 			return 1;
10927 		break;
10928 	case MPI2_EVENT_IR_CONFIGURATION_CHANGE_LIST:
10929 		_scsih_check_ir_config_unhide_events(ioc,
10930 		    (Mpi2EventDataIrConfigChangeList_t *)
10931 		    mpi_reply->EventData);
10932 		break;
10933 	case MPI2_EVENT_IR_VOLUME:
10934 		_scsih_check_volume_delete_events(ioc,
10935 		    (Mpi2EventDataIrVolume_t *)
10936 		    mpi_reply->EventData);
10937 		break;
10938 	case MPI2_EVENT_LOG_ENTRY_ADDED:
10939 	{
10940 		Mpi2EventDataLogEntryAdded_t *log_entry;
10941 		u32 log_code;
10942 
10943 		if (!ioc->is_warpdrive)
10944 			break;
10945 
10946 		log_entry = (Mpi2EventDataLogEntryAdded_t *)
10947 		    mpi_reply->EventData;
10948 		log_code = le32_to_cpu(*(__le32 *)log_entry->LogData);
10949 
10950 		if (le16_to_cpu(log_entry->LogEntryQualifier)
10951 		    != MPT2_WARPDRIVE_LOGENTRY)
10952 			break;
10953 
10954 		switch (log_code) {
10955 		case MPT2_WARPDRIVE_LC_SSDT:
10956 			ioc_warn(ioc, "WarpDrive Warning: IO Throttling has occurred in the WarpDrive subsystem. Check WarpDrive documentation for additional details.\n");
10957 			break;
10958 		case MPT2_WARPDRIVE_LC_SSDLW:
10959 			ioc_warn(ioc, "WarpDrive Warning: Program/Erase Cycles for the WarpDrive subsystem in degraded range. Check WarpDrive documentation for additional details.\n");
10960 			break;
10961 		case MPT2_WARPDRIVE_LC_SSDLF:
10962 			ioc_err(ioc, "WarpDrive Fatal Error: There are no Program/Erase Cycles for the WarpDrive subsystem. The storage device will be in read-only mode. Check WarpDrive documentation for additional details.\n");
10963 			break;
10964 		case MPT2_WARPDRIVE_LC_BRMF:
10965 			ioc_err(ioc, "WarpDrive Fatal Error: The Backup Rail Monitor has failed on the WarpDrive subsystem. Check WarpDrive documentation for additional details.\n");
10966 			break;
10967 		}
10968 
10969 		break;
10970 	}
10971 	case MPI2_EVENT_SAS_DEVICE_STATUS_CHANGE:
10972 		_scsih_sas_device_status_change_event(ioc,
10973 		    (Mpi2EventDataSasDeviceStatusChange_t *)
10974 		    mpi_reply->EventData);
10975 		break;
10976 	case MPI2_EVENT_IR_OPERATION_STATUS:
10977 	case MPI2_EVENT_SAS_DISCOVERY:
10978 	case MPI2_EVENT_SAS_DEVICE_DISCOVERY_ERROR:
10979 	case MPI2_EVENT_SAS_ENCL_DEVICE_STATUS_CHANGE:
10980 	case MPI2_EVENT_IR_PHYSICAL_DISK:
10981 	case MPI2_EVENT_PCIE_ENUMERATION:
10982 	case MPI2_EVENT_PCIE_DEVICE_STATUS_CHANGE:
10983 		break;
10984 
10985 	case MPI2_EVENT_TEMP_THRESHOLD:
10986 		_scsih_temp_threshold_events(ioc,
10987 			(Mpi2EventDataTemperature_t *)
10988 			mpi_reply->EventData);
10989 		break;
10990 	case MPI2_EVENT_ACTIVE_CABLE_EXCEPTION:
10991 		ActiveCableEventData =
10992 		    (Mpi26EventDataActiveCableExcept_t *) mpi_reply->EventData;
10993 		switch (ActiveCableEventData->ReasonCode) {
10994 		case MPI26_EVENT_ACTIVE_CABLE_INSUFFICIENT_POWER:
10995 			ioc_notice(ioc, "Currently an active cable with ReceptacleID %d\n",
10996 				   ActiveCableEventData->ReceptacleID);
10997 			pr_notice("cannot be powered and devices connected\n");
10998 			pr_notice("to this active cable will not be seen\n");
10999 			pr_notice("This active cable requires %d mW of power\n",
11000 			    le32_to_cpu(
11001 			    ActiveCableEventData->ActiveCablePowerRequirement));
11002 			break;
11003 
11004 		case MPI26_EVENT_ACTIVE_CABLE_DEGRADED:
11005 			ioc_notice(ioc, "Currently a cable with ReceptacleID %d\n",
11006 				   ActiveCableEventData->ReceptacleID);
11007 			pr_notice(
11008 			    "is not running at optimal speed(12 Gb/s rate)\n");
11009 			break;
11010 		}
11011 
11012 		break;
11013 
11014 	default: /* ignore the rest */
11015 		return 1;
11016 	}
11017 
11018 	sz = le16_to_cpu(mpi_reply->EventDataLength) * 4;
11019 	fw_event = alloc_fw_event_work(sz);
11020 	if (!fw_event) {
11021 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
11022 			__FILE__, __LINE__, __func__);
11023 		return 1;
11024 	}
11025 
11026 	memcpy(fw_event->event_data, mpi_reply->EventData, sz);
11027 	fw_event->ioc = ioc;
11028 	fw_event->VF_ID = mpi_reply->VF_ID;
11029 	fw_event->VP_ID = mpi_reply->VP_ID;
11030 	fw_event->event = event;
11031 	_scsih_fw_event_add(ioc, fw_event);
11032 	fw_event_work_put(fw_event);
11033 	return 1;
11034 }
11035 
11036 /**
11037  * _scsih_expander_node_remove - removing expander device from list.
11038  * @ioc: per adapter object
11039  * @sas_expander: the sas_device object
11040  *
11041  * Removing object and freeing associated memory from the
11042  * ioc->sas_expander_list.
11043  */
11044 static void
11045 _scsih_expander_node_remove(struct MPT3SAS_ADAPTER *ioc,
11046 	struct _sas_node *sas_expander)
11047 {
11048 	struct _sas_port *mpt3sas_port, *next;
11049 	unsigned long flags;
11050 	int port_id;
11051 
11052 	/* remove sibling ports attached to this expander */
11053 	list_for_each_entry_safe(mpt3sas_port, next,
11054 	   &sas_expander->sas_port_list, port_list) {
11055 		if (ioc->shost_recovery)
11056 			return;
11057 		if (mpt3sas_port->remote_identify.device_type ==
11058 		    SAS_END_DEVICE)
11059 			mpt3sas_device_remove_by_sas_address(ioc,
11060 			    mpt3sas_port->remote_identify.sas_address,
11061 			    mpt3sas_port->hba_port);
11062 		else if (mpt3sas_port->remote_identify.device_type ==
11063 		    SAS_EDGE_EXPANDER_DEVICE ||
11064 		    mpt3sas_port->remote_identify.device_type ==
11065 		    SAS_FANOUT_EXPANDER_DEVICE)
11066 			mpt3sas_expander_remove(ioc,
11067 			    mpt3sas_port->remote_identify.sas_address,
11068 			    mpt3sas_port->hba_port);
11069 	}
11070 
11071 	port_id = sas_expander->port->port_id;
11072 
11073 	mpt3sas_transport_port_remove(ioc, sas_expander->sas_address,
11074 	    sas_expander->sas_address_parent, sas_expander->port);
11075 
11076 	ioc_info(ioc,
11077 	    "expander_remove: handle(0x%04x), sas_addr(0x%016llx), port:%d\n",
11078 	    sas_expander->handle, (unsigned long long)
11079 	    sas_expander->sas_address,
11080 	    port_id);
11081 
11082 	spin_lock_irqsave(&ioc->sas_node_lock, flags);
11083 	list_del(&sas_expander->list);
11084 	spin_unlock_irqrestore(&ioc->sas_node_lock, flags);
11085 
11086 	kfree(sas_expander->phy);
11087 	kfree(sas_expander);
11088 }
11089 
11090 /**
11091  * _scsih_nvme_shutdown - NVMe shutdown notification
11092  * @ioc: per adapter object
11093  *
11094  * Sending IoUnitControl request with shutdown operation code to alert IOC that
11095  * the host system is shutting down so that IOC can issue NVMe shutdown to
11096  * NVMe drives attached to it.
11097  */
11098 static void
11099 _scsih_nvme_shutdown(struct MPT3SAS_ADAPTER *ioc)
11100 {
11101 	Mpi26IoUnitControlRequest_t *mpi_request;
11102 	Mpi26IoUnitControlReply_t *mpi_reply;
11103 	u16 smid;
11104 
11105 	/* are there any NVMe devices ? */
11106 	if (list_empty(&ioc->pcie_device_list))
11107 		return;
11108 
11109 	mutex_lock(&ioc->scsih_cmds.mutex);
11110 
11111 	if (ioc->scsih_cmds.status != MPT3_CMD_NOT_USED) {
11112 		ioc_err(ioc, "%s: scsih_cmd in use\n", __func__);
11113 		goto out;
11114 	}
11115 
11116 	ioc->scsih_cmds.status = MPT3_CMD_PENDING;
11117 
11118 	smid = mpt3sas_base_get_smid(ioc, ioc->scsih_cb_idx);
11119 	if (!smid) {
11120 		ioc_err(ioc,
11121 		    "%s: failed obtaining a smid\n", __func__);
11122 		ioc->scsih_cmds.status = MPT3_CMD_NOT_USED;
11123 		goto out;
11124 	}
11125 
11126 	mpi_request = mpt3sas_base_get_msg_frame(ioc, smid);
11127 	ioc->scsih_cmds.smid = smid;
11128 	memset(mpi_request, 0, sizeof(Mpi26IoUnitControlRequest_t));
11129 	mpi_request->Function = MPI2_FUNCTION_IO_UNIT_CONTROL;
11130 	mpi_request->Operation = MPI26_CTRL_OP_SHUTDOWN;
11131 
11132 	init_completion(&ioc->scsih_cmds.done);
11133 	ioc->put_smid_default(ioc, smid);
11134 	/* Wait for max_shutdown_latency seconds */
11135 	ioc_info(ioc,
11136 		"Io Unit Control shutdown (sending), Shutdown latency %d sec\n",
11137 		ioc->max_shutdown_latency);
11138 	wait_for_completion_timeout(&ioc->scsih_cmds.done,
11139 			ioc->max_shutdown_latency*HZ);
11140 
11141 	if (!(ioc->scsih_cmds.status & MPT3_CMD_COMPLETE)) {
11142 		ioc_err(ioc, "%s: timeout\n", __func__);
11143 		goto out;
11144 	}
11145 
11146 	if (ioc->scsih_cmds.status & MPT3_CMD_REPLY_VALID) {
11147 		mpi_reply = ioc->scsih_cmds.reply;
11148 		ioc_info(ioc, "Io Unit Control shutdown (complete):"
11149 			"ioc_status(0x%04x), loginfo(0x%08x)\n",
11150 			le16_to_cpu(mpi_reply->IOCStatus),
11151 			le32_to_cpu(mpi_reply->IOCLogInfo));
11152 	}
11153  out:
11154 	ioc->scsih_cmds.status = MPT3_CMD_NOT_USED;
11155 	mutex_unlock(&ioc->scsih_cmds.mutex);
11156 }
11157 
11158 
11159 /**
11160  * _scsih_ir_shutdown - IR shutdown notification
11161  * @ioc: per adapter object
11162  *
11163  * Sending RAID Action to alert the Integrated RAID subsystem of the IOC that
11164  * the host system is shutting down.
11165  */
11166 static void
11167 _scsih_ir_shutdown(struct MPT3SAS_ADAPTER *ioc)
11168 {
11169 	Mpi2RaidActionRequest_t *mpi_request;
11170 	Mpi2RaidActionReply_t *mpi_reply;
11171 	u16 smid;
11172 
11173 	/* is IR firmware build loaded ? */
11174 	if (!ioc->ir_firmware)
11175 		return;
11176 
11177 	/* are there any volumes ? */
11178 	if (list_empty(&ioc->raid_device_list))
11179 		return;
11180 
11181 	mutex_lock(&ioc->scsih_cmds.mutex);
11182 
11183 	if (ioc->scsih_cmds.status != MPT3_CMD_NOT_USED) {
11184 		ioc_err(ioc, "%s: scsih_cmd in use\n", __func__);
11185 		goto out;
11186 	}
11187 	ioc->scsih_cmds.status = MPT3_CMD_PENDING;
11188 
11189 	smid = mpt3sas_base_get_smid(ioc, ioc->scsih_cb_idx);
11190 	if (!smid) {
11191 		ioc_err(ioc, "%s: failed obtaining a smid\n", __func__);
11192 		ioc->scsih_cmds.status = MPT3_CMD_NOT_USED;
11193 		goto out;
11194 	}
11195 
11196 	mpi_request = mpt3sas_base_get_msg_frame(ioc, smid);
11197 	ioc->scsih_cmds.smid = smid;
11198 	memset(mpi_request, 0, sizeof(Mpi2RaidActionRequest_t));
11199 
11200 	mpi_request->Function = MPI2_FUNCTION_RAID_ACTION;
11201 	mpi_request->Action = MPI2_RAID_ACTION_SYSTEM_SHUTDOWN_INITIATED;
11202 
11203 	if (!ioc->hide_ir_msg)
11204 		ioc_info(ioc, "IR shutdown (sending)\n");
11205 	init_completion(&ioc->scsih_cmds.done);
11206 	ioc->put_smid_default(ioc, smid);
11207 	wait_for_completion_timeout(&ioc->scsih_cmds.done, 10*HZ);
11208 
11209 	if (!(ioc->scsih_cmds.status & MPT3_CMD_COMPLETE)) {
11210 		ioc_err(ioc, "%s: timeout\n", __func__);
11211 		goto out;
11212 	}
11213 
11214 	if (ioc->scsih_cmds.status & MPT3_CMD_REPLY_VALID) {
11215 		mpi_reply = ioc->scsih_cmds.reply;
11216 		if (!ioc->hide_ir_msg)
11217 			ioc_info(ioc, "IR shutdown (complete): ioc_status(0x%04x), loginfo(0x%08x)\n",
11218 				 le16_to_cpu(mpi_reply->IOCStatus),
11219 				 le32_to_cpu(mpi_reply->IOCLogInfo));
11220 	}
11221 
11222  out:
11223 	ioc->scsih_cmds.status = MPT3_CMD_NOT_USED;
11224 	mutex_unlock(&ioc->scsih_cmds.mutex);
11225 }
11226 
11227 /**
11228  * _scsih_get_shost_and_ioc - get shost and ioc
11229  *			and verify whether they are NULL or not
11230  * @pdev: PCI device struct
11231  * @shost: address of scsi host pointer
11232  * @ioc: address of HBA adapter pointer
11233  *
11234  * Return zero if *shost and *ioc are not NULL otherwise return error number.
11235  */
11236 static int
11237 _scsih_get_shost_and_ioc(struct pci_dev *pdev,
11238 	struct Scsi_Host **shost, struct MPT3SAS_ADAPTER **ioc)
11239 {
11240 	*shost = pci_get_drvdata(pdev);
11241 	if (*shost == NULL) {
11242 		dev_err(&pdev->dev, "pdev's driver data is null\n");
11243 		return -ENXIO;
11244 	}
11245 
11246 	*ioc = shost_priv(*shost);
11247 	if (*ioc == NULL) {
11248 		dev_err(&pdev->dev, "shost's private data is null\n");
11249 		return -ENXIO;
11250 	}
11251 
11252 	return 0;
11253 }
11254 
11255 /**
11256  * scsih_remove - detach and remove add host
11257  * @pdev: PCI device struct
11258  *
11259  * Routine called when unloading the driver.
11260  */
11261 static void scsih_remove(struct pci_dev *pdev)
11262 {
11263 	struct Scsi_Host *shost;
11264 	struct MPT3SAS_ADAPTER *ioc;
11265 	struct _sas_port *mpt3sas_port, *next_port;
11266 	struct _raid_device *raid_device, *next;
11267 	struct MPT3SAS_TARGET *sas_target_priv_data;
11268 	struct _pcie_device *pcie_device, *pcienext;
11269 	struct workqueue_struct	*wq;
11270 	unsigned long flags;
11271 	Mpi2ConfigReply_t mpi_reply;
11272 	struct hba_port *port, *port_next;
11273 
11274 	if (_scsih_get_shost_and_ioc(pdev, &shost, &ioc))
11275 		return;
11276 
11277 	ioc->remove_host = 1;
11278 
11279 	if (!pci_device_is_present(pdev)) {
11280 		mpt3sas_base_pause_mq_polling(ioc);
11281 		_scsih_flush_running_cmds(ioc);
11282 	}
11283 
11284 	_scsih_fw_event_cleanup_queue(ioc);
11285 
11286 	spin_lock_irqsave(&ioc->fw_event_lock, flags);
11287 	wq = ioc->firmware_event_thread;
11288 	ioc->firmware_event_thread = NULL;
11289 	spin_unlock_irqrestore(&ioc->fw_event_lock, flags);
11290 	if (wq)
11291 		destroy_workqueue(wq);
11292 	/*
11293 	 * Copy back the unmodified ioc page1. so that on next driver load,
11294 	 * current modified changes on ioc page1 won't take effect.
11295 	 */
11296 	if (ioc->is_aero_ioc)
11297 		mpt3sas_config_set_ioc_pg1(ioc, &mpi_reply,
11298 				&ioc->ioc_pg1_copy);
11299 	/* release all the volumes */
11300 	_scsih_ir_shutdown(ioc);
11301 	mpt3sas_destroy_debugfs(ioc);
11302 	sas_remove_host(shost);
11303 	list_for_each_entry_safe(raid_device, next, &ioc->raid_device_list,
11304 	    list) {
11305 		if (raid_device->starget) {
11306 			sas_target_priv_data =
11307 			    raid_device->starget->hostdata;
11308 			sas_target_priv_data->deleted = 1;
11309 			scsi_remove_target(&raid_device->starget->dev);
11310 		}
11311 		ioc_info(ioc, "removing handle(0x%04x), wwid(0x%016llx)\n",
11312 			 raid_device->handle, (u64)raid_device->wwid);
11313 		_scsih_raid_device_remove(ioc, raid_device);
11314 	}
11315 	list_for_each_entry_safe(pcie_device, pcienext, &ioc->pcie_device_list,
11316 		list) {
11317 		_scsih_pcie_device_remove_from_sml(ioc, pcie_device);
11318 		list_del_init(&pcie_device->list);
11319 		pcie_device_put(pcie_device);
11320 	}
11321 
11322 	/* free ports attached to the sas_host */
11323 	list_for_each_entry_safe(mpt3sas_port, next_port,
11324 	   &ioc->sas_hba.sas_port_list, port_list) {
11325 		if (mpt3sas_port->remote_identify.device_type ==
11326 		    SAS_END_DEVICE)
11327 			mpt3sas_device_remove_by_sas_address(ioc,
11328 			    mpt3sas_port->remote_identify.sas_address,
11329 			    mpt3sas_port->hba_port);
11330 		else if (mpt3sas_port->remote_identify.device_type ==
11331 		    SAS_EDGE_EXPANDER_DEVICE ||
11332 		    mpt3sas_port->remote_identify.device_type ==
11333 		    SAS_FANOUT_EXPANDER_DEVICE)
11334 			mpt3sas_expander_remove(ioc,
11335 			    mpt3sas_port->remote_identify.sas_address,
11336 			    mpt3sas_port->hba_port);
11337 	}
11338 
11339 	list_for_each_entry_safe(port, port_next,
11340 	    &ioc->port_table_list, list) {
11341 		list_del(&port->list);
11342 		kfree(port);
11343 	}
11344 
11345 	/* free phys attached to the sas_host */
11346 	if (ioc->sas_hba.num_phys) {
11347 		kfree(ioc->sas_hba.phy);
11348 		ioc->sas_hba.phy = NULL;
11349 		ioc->sas_hba.num_phys = 0;
11350 	}
11351 
11352 	mpt3sas_base_detach(ioc);
11353 	spin_lock(&gioc_lock);
11354 	list_del(&ioc->list);
11355 	spin_unlock(&gioc_lock);
11356 	scsi_host_put(shost);
11357 }
11358 
11359 /**
11360  * scsih_shutdown - routine call during system shutdown
11361  * @pdev: PCI device struct
11362  */
11363 static void
11364 scsih_shutdown(struct pci_dev *pdev)
11365 {
11366 	struct Scsi_Host *shost;
11367 	struct MPT3SAS_ADAPTER *ioc;
11368 	struct workqueue_struct	*wq;
11369 	unsigned long flags;
11370 	Mpi2ConfigReply_t mpi_reply;
11371 
11372 	if (_scsih_get_shost_and_ioc(pdev, &shost, &ioc))
11373 		return;
11374 
11375 	ioc->remove_host = 1;
11376 
11377 	if (!pci_device_is_present(pdev)) {
11378 		mpt3sas_base_pause_mq_polling(ioc);
11379 		_scsih_flush_running_cmds(ioc);
11380 	}
11381 
11382 	_scsih_fw_event_cleanup_queue(ioc);
11383 
11384 	spin_lock_irqsave(&ioc->fw_event_lock, flags);
11385 	wq = ioc->firmware_event_thread;
11386 	ioc->firmware_event_thread = NULL;
11387 	spin_unlock_irqrestore(&ioc->fw_event_lock, flags);
11388 	if (wq)
11389 		destroy_workqueue(wq);
11390 	/*
11391 	 * Copy back the unmodified ioc page1 so that on next driver load,
11392 	 * current modified changes on ioc page1 won't take effect.
11393 	 */
11394 	if (ioc->is_aero_ioc)
11395 		mpt3sas_config_set_ioc_pg1(ioc, &mpi_reply,
11396 				&ioc->ioc_pg1_copy);
11397 
11398 	_scsih_ir_shutdown(ioc);
11399 	_scsih_nvme_shutdown(ioc);
11400 	mpt3sas_base_mask_interrupts(ioc);
11401 	mpt3sas_base_stop_watchdog(ioc);
11402 	ioc->shost_recovery = 1;
11403 	mpt3sas_base_make_ioc_ready(ioc, SOFT_RESET);
11404 	ioc->shost_recovery = 0;
11405 	mpt3sas_base_free_irq(ioc);
11406 	mpt3sas_base_disable_msix(ioc);
11407 }
11408 
11409 
11410 /**
11411  * _scsih_probe_boot_devices - reports 1st device
11412  * @ioc: per adapter object
11413  *
11414  * If specified in bios page 2, this routine reports the 1st
11415  * device scsi-ml or sas transport for persistent boot device
11416  * purposes.  Please refer to function _scsih_determine_boot_device()
11417  */
11418 static void
11419 _scsih_probe_boot_devices(struct MPT3SAS_ADAPTER *ioc)
11420 {
11421 	u32 channel;
11422 	void *device;
11423 	struct _sas_device *sas_device;
11424 	struct _raid_device *raid_device;
11425 	struct _pcie_device *pcie_device;
11426 	u16 handle;
11427 	u64 sas_address_parent;
11428 	u64 sas_address;
11429 	unsigned long flags;
11430 	int rc;
11431 	int tid;
11432 	struct hba_port *port;
11433 
11434 	 /* no Bios, return immediately */
11435 	if (!ioc->bios_pg3.BiosVersion)
11436 		return;
11437 
11438 	device = NULL;
11439 	if (ioc->req_boot_device.device) {
11440 		device =  ioc->req_boot_device.device;
11441 		channel = ioc->req_boot_device.channel;
11442 	} else if (ioc->req_alt_boot_device.device) {
11443 		device =  ioc->req_alt_boot_device.device;
11444 		channel = ioc->req_alt_boot_device.channel;
11445 	} else if (ioc->current_boot_device.device) {
11446 		device =  ioc->current_boot_device.device;
11447 		channel = ioc->current_boot_device.channel;
11448 	}
11449 
11450 	if (!device)
11451 		return;
11452 
11453 	if (channel == RAID_CHANNEL) {
11454 		raid_device = device;
11455 		/*
11456 		 * If this boot vd is already registered with SML then
11457 		 * no need to register it again as part of device scanning
11458 		 * after diag reset during driver load operation.
11459 		 */
11460 		if (raid_device->starget)
11461 			return;
11462 		rc = scsi_add_device(ioc->shost, RAID_CHANNEL,
11463 		    raid_device->id, 0);
11464 		if (rc)
11465 			_scsih_raid_device_remove(ioc, raid_device);
11466 	} else if (channel == PCIE_CHANNEL) {
11467 		pcie_device = device;
11468 		/*
11469 		 * If this boot NVMe device is already registered with SML then
11470 		 * no need to register it again as part of device scanning
11471 		 * after diag reset during driver load operation.
11472 		 */
11473 		if (pcie_device->starget)
11474 			return;
11475 		spin_lock_irqsave(&ioc->pcie_device_lock, flags);
11476 		tid = pcie_device->id;
11477 		list_move_tail(&pcie_device->list, &ioc->pcie_device_list);
11478 		spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
11479 		rc = scsi_add_device(ioc->shost, PCIE_CHANNEL, tid, 0);
11480 		if (rc)
11481 			_scsih_pcie_device_remove(ioc, pcie_device);
11482 	} else {
11483 		sas_device = device;
11484 		/*
11485 		 * If this boot sas/sata device is already registered with SML
11486 		 * then no need to register it again as part of device scanning
11487 		 * after diag reset during driver load operation.
11488 		 */
11489 		if (sas_device->starget)
11490 			return;
11491 		spin_lock_irqsave(&ioc->sas_device_lock, flags);
11492 		handle = sas_device->handle;
11493 		sas_address_parent = sas_device->sas_address_parent;
11494 		sas_address = sas_device->sas_address;
11495 		port = sas_device->port;
11496 		list_move_tail(&sas_device->list, &ioc->sas_device_list);
11497 		spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
11498 
11499 		if (ioc->hide_drives)
11500 			return;
11501 
11502 		if (!port)
11503 			return;
11504 
11505 		if (!mpt3sas_transport_port_add(ioc, handle,
11506 		    sas_address_parent, port)) {
11507 			_scsih_sas_device_remove(ioc, sas_device);
11508 		} else if (!sas_device->starget) {
11509 			if (!ioc->is_driver_loading) {
11510 				mpt3sas_transport_port_remove(ioc,
11511 				    sas_address,
11512 				    sas_address_parent, port);
11513 				_scsih_sas_device_remove(ioc, sas_device);
11514 			}
11515 		}
11516 	}
11517 }
11518 
11519 /**
11520  * _scsih_probe_raid - reporting raid volumes to scsi-ml
11521  * @ioc: per adapter object
11522  *
11523  * Called during initial loading of the driver.
11524  */
11525 static void
11526 _scsih_probe_raid(struct MPT3SAS_ADAPTER *ioc)
11527 {
11528 	struct _raid_device *raid_device, *raid_next;
11529 	int rc;
11530 
11531 	list_for_each_entry_safe(raid_device, raid_next,
11532 	    &ioc->raid_device_list, list) {
11533 		if (raid_device->starget)
11534 			continue;
11535 		rc = scsi_add_device(ioc->shost, RAID_CHANNEL,
11536 		    raid_device->id, 0);
11537 		if (rc)
11538 			_scsih_raid_device_remove(ioc, raid_device);
11539 	}
11540 }
11541 
11542 static struct _sas_device *get_next_sas_device(struct MPT3SAS_ADAPTER *ioc)
11543 {
11544 	struct _sas_device *sas_device = NULL;
11545 	unsigned long flags;
11546 
11547 	spin_lock_irqsave(&ioc->sas_device_lock, flags);
11548 	if (!list_empty(&ioc->sas_device_init_list)) {
11549 		sas_device = list_first_entry(&ioc->sas_device_init_list,
11550 				struct _sas_device, list);
11551 		sas_device_get(sas_device);
11552 	}
11553 	spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
11554 
11555 	return sas_device;
11556 }
11557 
11558 static void sas_device_make_active(struct MPT3SAS_ADAPTER *ioc,
11559 		struct _sas_device *sas_device)
11560 {
11561 	unsigned long flags;
11562 
11563 	spin_lock_irqsave(&ioc->sas_device_lock, flags);
11564 
11565 	/*
11566 	 * Since we dropped the lock during the call to port_add(), we need to
11567 	 * be careful here that somebody else didn't move or delete this item
11568 	 * while we were busy with other things.
11569 	 *
11570 	 * If it was on the list, we need a put() for the reference the list
11571 	 * had. Either way, we need a get() for the destination list.
11572 	 */
11573 	if (!list_empty(&sas_device->list)) {
11574 		list_del_init(&sas_device->list);
11575 		sas_device_put(sas_device);
11576 	}
11577 
11578 	sas_device_get(sas_device);
11579 	list_add_tail(&sas_device->list, &ioc->sas_device_list);
11580 
11581 	spin_unlock_irqrestore(&ioc->sas_device_lock, flags);
11582 }
11583 
11584 /**
11585  * _scsih_probe_sas - reporting sas devices to sas transport
11586  * @ioc: per adapter object
11587  *
11588  * Called during initial loading of the driver.
11589  */
11590 static void
11591 _scsih_probe_sas(struct MPT3SAS_ADAPTER *ioc)
11592 {
11593 	struct _sas_device *sas_device;
11594 
11595 	if (ioc->hide_drives)
11596 		return;
11597 
11598 	while ((sas_device = get_next_sas_device(ioc))) {
11599 		if (!mpt3sas_transport_port_add(ioc, sas_device->handle,
11600 		    sas_device->sas_address_parent, sas_device->port)) {
11601 			_scsih_sas_device_remove(ioc, sas_device);
11602 			sas_device_put(sas_device);
11603 			continue;
11604 		} else if (!sas_device->starget) {
11605 			/*
11606 			 * When asyn scanning is enabled, its not possible to
11607 			 * remove devices while scanning is turned on due to an
11608 			 * oops in scsi_sysfs_add_sdev()->add_device()->
11609 			 * sysfs_addrm_start()
11610 			 */
11611 			if (!ioc->is_driver_loading) {
11612 				mpt3sas_transport_port_remove(ioc,
11613 				    sas_device->sas_address,
11614 				    sas_device->sas_address_parent,
11615 				    sas_device->port);
11616 				_scsih_sas_device_remove(ioc, sas_device);
11617 				sas_device_put(sas_device);
11618 				continue;
11619 			}
11620 		}
11621 		sas_device_make_active(ioc, sas_device);
11622 		sas_device_put(sas_device);
11623 	}
11624 }
11625 
11626 /**
11627  * get_next_pcie_device - Get the next pcie device
11628  * @ioc: per adapter object
11629  *
11630  * Get the next pcie device from pcie_device_init_list list.
11631  *
11632  * Return: pcie device structure if pcie_device_init_list list is not empty
11633  * otherwise returns NULL
11634  */
11635 static struct _pcie_device *get_next_pcie_device(struct MPT3SAS_ADAPTER *ioc)
11636 {
11637 	struct _pcie_device *pcie_device = NULL;
11638 	unsigned long flags;
11639 
11640 	spin_lock_irqsave(&ioc->pcie_device_lock, flags);
11641 	if (!list_empty(&ioc->pcie_device_init_list)) {
11642 		pcie_device = list_first_entry(&ioc->pcie_device_init_list,
11643 				struct _pcie_device, list);
11644 		pcie_device_get(pcie_device);
11645 	}
11646 	spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
11647 
11648 	return pcie_device;
11649 }
11650 
11651 /**
11652  * pcie_device_make_active - Add pcie device to pcie_device_list list
11653  * @ioc: per adapter object
11654  * @pcie_device: pcie device object
11655  *
11656  * Add the pcie device which has registered with SCSI Transport Later to
11657  * pcie_device_list list
11658  */
11659 static void pcie_device_make_active(struct MPT3SAS_ADAPTER *ioc,
11660 		struct _pcie_device *pcie_device)
11661 {
11662 	unsigned long flags;
11663 
11664 	spin_lock_irqsave(&ioc->pcie_device_lock, flags);
11665 
11666 	if (!list_empty(&pcie_device->list)) {
11667 		list_del_init(&pcie_device->list);
11668 		pcie_device_put(pcie_device);
11669 	}
11670 	pcie_device_get(pcie_device);
11671 	list_add_tail(&pcie_device->list, &ioc->pcie_device_list);
11672 
11673 	spin_unlock_irqrestore(&ioc->pcie_device_lock, flags);
11674 }
11675 
11676 /**
11677  * _scsih_probe_pcie - reporting PCIe devices to scsi-ml
11678  * @ioc: per adapter object
11679  *
11680  * Called during initial loading of the driver.
11681  */
11682 static void
11683 _scsih_probe_pcie(struct MPT3SAS_ADAPTER *ioc)
11684 {
11685 	struct _pcie_device *pcie_device;
11686 	int rc;
11687 
11688 	/* PCIe Device List */
11689 	while ((pcie_device = get_next_pcie_device(ioc))) {
11690 		if (pcie_device->starget) {
11691 			pcie_device_put(pcie_device);
11692 			continue;
11693 		}
11694 		if (pcie_device->access_status ==
11695 		    MPI26_PCIEDEV0_ASTATUS_DEVICE_BLOCKED) {
11696 			pcie_device_make_active(ioc, pcie_device);
11697 			pcie_device_put(pcie_device);
11698 			continue;
11699 		}
11700 		rc = scsi_add_device(ioc->shost, PCIE_CHANNEL,
11701 			pcie_device->id, 0);
11702 		if (rc) {
11703 			_scsih_pcie_device_remove(ioc, pcie_device);
11704 			pcie_device_put(pcie_device);
11705 			continue;
11706 		} else if (!pcie_device->starget) {
11707 			/*
11708 			 * When async scanning is enabled, its not possible to
11709 			 * remove devices while scanning is turned on due to an
11710 			 * oops in scsi_sysfs_add_sdev()->add_device()->
11711 			 * sysfs_addrm_start()
11712 			 */
11713 			if (!ioc->is_driver_loading) {
11714 			/* TODO-- Need to find out whether this condition will
11715 			 * occur or not
11716 			 */
11717 				_scsih_pcie_device_remove(ioc, pcie_device);
11718 				pcie_device_put(pcie_device);
11719 				continue;
11720 			}
11721 		}
11722 		pcie_device_make_active(ioc, pcie_device);
11723 		pcie_device_put(pcie_device);
11724 	}
11725 }
11726 
11727 /**
11728  * _scsih_probe_devices - probing for devices
11729  * @ioc: per adapter object
11730  *
11731  * Called during initial loading of the driver.
11732  */
11733 static void
11734 _scsih_probe_devices(struct MPT3SAS_ADAPTER *ioc)
11735 {
11736 	u16 volume_mapping_flags;
11737 
11738 	if (!(ioc->facts.ProtocolFlags & MPI2_IOCFACTS_PROTOCOL_SCSI_INITIATOR))
11739 		return;  /* return when IOC doesn't support initiator mode */
11740 
11741 	_scsih_probe_boot_devices(ioc);
11742 
11743 	if (ioc->ir_firmware) {
11744 		volume_mapping_flags =
11745 		    le16_to_cpu(ioc->ioc_pg8.IRVolumeMappingFlags) &
11746 		    MPI2_IOCPAGE8_IRFLAGS_MASK_VOLUME_MAPPING_MODE;
11747 		if (volume_mapping_flags ==
11748 		    MPI2_IOCPAGE8_IRFLAGS_LOW_VOLUME_MAPPING) {
11749 			_scsih_probe_raid(ioc);
11750 			_scsih_probe_sas(ioc);
11751 		} else {
11752 			_scsih_probe_sas(ioc);
11753 			_scsih_probe_raid(ioc);
11754 		}
11755 	} else {
11756 		_scsih_probe_sas(ioc);
11757 		_scsih_probe_pcie(ioc);
11758 	}
11759 }
11760 
11761 /**
11762  * scsih_scan_start - scsi lld callback for .scan_start
11763  * @shost: SCSI host pointer
11764  *
11765  * The shost has the ability to discover targets on its own instead
11766  * of scanning the entire bus.  In our implemention, we will kick off
11767  * firmware discovery.
11768  */
11769 static void
11770 scsih_scan_start(struct Scsi_Host *shost)
11771 {
11772 	struct MPT3SAS_ADAPTER *ioc = shost_priv(shost);
11773 	int rc;
11774 	if (diag_buffer_enable != -1 && diag_buffer_enable != 0)
11775 		mpt3sas_enable_diag_buffer(ioc, diag_buffer_enable);
11776 	else if (ioc->manu_pg11.HostTraceBufferMaxSizeKB != 0)
11777 		mpt3sas_enable_diag_buffer(ioc, 1);
11778 
11779 	if (disable_discovery > 0)
11780 		return;
11781 
11782 	ioc->start_scan = 1;
11783 	rc = mpt3sas_port_enable(ioc);
11784 
11785 	if (rc != 0)
11786 		ioc_info(ioc, "port enable: FAILED\n");
11787 }
11788 
11789 /**
11790  * _scsih_complete_devices_scanning - add the devices to sml and
11791  * complete ioc initialization.
11792  * @ioc: per adapter object
11793  *
11794  * Return nothing.
11795  */
11796 static void _scsih_complete_devices_scanning(struct MPT3SAS_ADAPTER *ioc)
11797 {
11798 
11799 	if (ioc->wait_for_discovery_to_complete) {
11800 		ioc->wait_for_discovery_to_complete = 0;
11801 		_scsih_probe_devices(ioc);
11802 	}
11803 
11804 	mpt3sas_base_start_watchdog(ioc);
11805 	ioc->is_driver_loading = 0;
11806 }
11807 
11808 /**
11809  * scsih_scan_finished - scsi lld callback for .scan_finished
11810  * @shost: SCSI host pointer
11811  * @time: elapsed time of the scan in jiffies
11812  *
11813  * This function will be called periodicallyn until it returns 1 with the
11814  * scsi_host and the elapsed time of the scan in jiffies. In our implemention,
11815  * we wait for firmware discovery to complete, then return 1.
11816  */
11817 static int
11818 scsih_scan_finished(struct Scsi_Host *shost, unsigned long time)
11819 {
11820 	struct MPT3SAS_ADAPTER *ioc = shost_priv(shost);
11821 	u32 ioc_state;
11822 	int issue_hard_reset = 0;
11823 
11824 	if (disable_discovery > 0) {
11825 		ioc->is_driver_loading = 0;
11826 		ioc->wait_for_discovery_to_complete = 0;
11827 		return 1;
11828 	}
11829 
11830 	if (time >= (300 * HZ)) {
11831 		ioc->port_enable_cmds.status = MPT3_CMD_NOT_USED;
11832 		ioc_info(ioc, "port enable: FAILED with timeout (timeout=300s)\n");
11833 		ioc->is_driver_loading = 0;
11834 		return 1;
11835 	}
11836 
11837 	if (ioc->start_scan) {
11838 		ioc_state = mpt3sas_base_get_iocstate(ioc, 0);
11839 		if ((ioc_state & MPI2_IOC_STATE_MASK) == MPI2_IOC_STATE_FAULT) {
11840 			mpt3sas_print_fault_code(ioc, ioc_state &
11841 			    MPI2_DOORBELL_DATA_MASK);
11842 			issue_hard_reset = 1;
11843 			goto out;
11844 		} else if ((ioc_state & MPI2_IOC_STATE_MASK) ==
11845 				MPI2_IOC_STATE_COREDUMP) {
11846 			mpt3sas_base_coredump_info(ioc, ioc_state &
11847 			    MPI2_DOORBELL_DATA_MASK);
11848 			mpt3sas_base_wait_for_coredump_completion(ioc, __func__);
11849 			issue_hard_reset = 1;
11850 			goto out;
11851 		}
11852 		return 0;
11853 	}
11854 
11855 	if (ioc->port_enable_cmds.status & MPT3_CMD_RESET) {
11856 		ioc_info(ioc,
11857 		    "port enable: aborted due to diag reset\n");
11858 		ioc->port_enable_cmds.status = MPT3_CMD_NOT_USED;
11859 		goto out;
11860 	}
11861 	if (ioc->start_scan_failed) {
11862 		ioc_info(ioc, "port enable: FAILED with (ioc_status=0x%08x)\n",
11863 			 ioc->start_scan_failed);
11864 		ioc->is_driver_loading = 0;
11865 		ioc->wait_for_discovery_to_complete = 0;
11866 		ioc->remove_host = 1;
11867 		return 1;
11868 	}
11869 
11870 	ioc_info(ioc, "port enable: SUCCESS\n");
11871 	ioc->port_enable_cmds.status = MPT3_CMD_NOT_USED;
11872 	_scsih_complete_devices_scanning(ioc);
11873 
11874 out:
11875 	if (issue_hard_reset) {
11876 		ioc->port_enable_cmds.status = MPT3_CMD_NOT_USED;
11877 		if (mpt3sas_base_hard_reset_handler(ioc, SOFT_RESET))
11878 			ioc->is_driver_loading = 0;
11879 	}
11880 	return 1;
11881 }
11882 
11883 /**
11884  * scsih_map_queues - map reply queues with request queues
11885  * @shost: SCSI host pointer
11886  */
11887 static void scsih_map_queues(struct Scsi_Host *shost)
11888 {
11889 	struct MPT3SAS_ADAPTER *ioc =
11890 	    (struct MPT3SAS_ADAPTER *)shost->hostdata;
11891 	struct blk_mq_queue_map *map;
11892 	int i, qoff, offset;
11893 	int nr_msix_vectors = ioc->iopoll_q_start_index;
11894 	int iopoll_q_count = ioc->reply_queue_count - nr_msix_vectors;
11895 
11896 	if (shost->nr_hw_queues == 1)
11897 		return;
11898 
11899 	for (i = 0, qoff = 0; i < shost->nr_maps; i++) {
11900 		map = &shost->tag_set.map[i];
11901 		map->nr_queues = 0;
11902 		offset = 0;
11903 		if (i == HCTX_TYPE_DEFAULT) {
11904 			map->nr_queues =
11905 			    nr_msix_vectors - ioc->high_iops_queues;
11906 			offset = ioc->high_iops_queues;
11907 		} else if (i == HCTX_TYPE_POLL)
11908 			map->nr_queues = iopoll_q_count;
11909 
11910 		if (!map->nr_queues)
11911 			BUG_ON(i == HCTX_TYPE_DEFAULT);
11912 
11913 		/*
11914 		 * The poll queue(s) doesn't have an IRQ (and hence IRQ
11915 		 * affinity), so use the regular blk-mq cpu mapping
11916 		 */
11917 		map->queue_offset = qoff;
11918 		if (i != HCTX_TYPE_POLL)
11919 			blk_mq_pci_map_queues(map, ioc->pdev, offset);
11920 		else
11921 			blk_mq_map_queues(map);
11922 
11923 		qoff += map->nr_queues;
11924 	}
11925 }
11926 
11927 /* shost template for SAS 2.0 HBA devices */
11928 static const struct scsi_host_template mpt2sas_driver_template = {
11929 	.module				= THIS_MODULE,
11930 	.name				= "Fusion MPT SAS Host",
11931 	.proc_name			= MPT2SAS_DRIVER_NAME,
11932 	.queuecommand			= scsih_qcmd,
11933 	.target_alloc			= scsih_target_alloc,
11934 	.slave_alloc			= scsih_slave_alloc,
11935 	.slave_configure		= scsih_slave_configure,
11936 	.target_destroy			= scsih_target_destroy,
11937 	.slave_destroy			= scsih_slave_destroy,
11938 	.scan_finished			= scsih_scan_finished,
11939 	.scan_start			= scsih_scan_start,
11940 	.change_queue_depth		= scsih_change_queue_depth,
11941 	.eh_abort_handler		= scsih_abort,
11942 	.eh_device_reset_handler	= scsih_dev_reset,
11943 	.eh_target_reset_handler	= scsih_target_reset,
11944 	.eh_host_reset_handler		= scsih_host_reset,
11945 	.bios_param			= scsih_bios_param,
11946 	.can_queue			= 1,
11947 	.this_id			= -1,
11948 	.sg_tablesize			= MPT2SAS_SG_DEPTH,
11949 	.max_sectors			= 32767,
11950 	.cmd_per_lun			= 7,
11951 	.shost_groups			= mpt3sas_host_groups,
11952 	.sdev_groups			= mpt3sas_dev_groups,
11953 	.track_queue_depth		= 1,
11954 	.cmd_size			= sizeof(struct scsiio_tracker),
11955 };
11956 
11957 /* raid transport support for SAS 2.0 HBA devices */
11958 static struct raid_function_template mpt2sas_raid_functions = {
11959 	.cookie		= &mpt2sas_driver_template,
11960 	.is_raid	= scsih_is_raid,
11961 	.get_resync	= scsih_get_resync,
11962 	.get_state	= scsih_get_state,
11963 };
11964 
11965 /* shost template for SAS 3.0 HBA devices */
11966 static const struct scsi_host_template mpt3sas_driver_template = {
11967 	.module				= THIS_MODULE,
11968 	.name				= "Fusion MPT SAS Host",
11969 	.proc_name			= MPT3SAS_DRIVER_NAME,
11970 	.queuecommand			= scsih_qcmd,
11971 	.target_alloc			= scsih_target_alloc,
11972 	.slave_alloc			= scsih_slave_alloc,
11973 	.slave_configure		= scsih_slave_configure,
11974 	.target_destroy			= scsih_target_destroy,
11975 	.slave_destroy			= scsih_slave_destroy,
11976 	.scan_finished			= scsih_scan_finished,
11977 	.scan_start			= scsih_scan_start,
11978 	.change_queue_depth		= scsih_change_queue_depth,
11979 	.eh_abort_handler		= scsih_abort,
11980 	.eh_device_reset_handler	= scsih_dev_reset,
11981 	.eh_target_reset_handler	= scsih_target_reset,
11982 	.eh_host_reset_handler		= scsih_host_reset,
11983 	.bios_param			= scsih_bios_param,
11984 	.can_queue			= 1,
11985 	.this_id			= -1,
11986 	.sg_tablesize			= MPT3SAS_SG_DEPTH,
11987 	.max_sectors			= 32767,
11988 	.max_segment_size		= 0xffffffff,
11989 	.cmd_per_lun			= 128,
11990 	.shost_groups			= mpt3sas_host_groups,
11991 	.sdev_groups			= mpt3sas_dev_groups,
11992 	.track_queue_depth		= 1,
11993 	.cmd_size			= sizeof(struct scsiio_tracker),
11994 	.map_queues			= scsih_map_queues,
11995 	.mq_poll			= mpt3sas_blk_mq_poll,
11996 };
11997 
11998 /* raid transport support for SAS 3.0 HBA devices */
11999 static struct raid_function_template mpt3sas_raid_functions = {
12000 	.cookie		= &mpt3sas_driver_template,
12001 	.is_raid	= scsih_is_raid,
12002 	.get_resync	= scsih_get_resync,
12003 	.get_state	= scsih_get_state,
12004 };
12005 
12006 /**
12007  * _scsih_determine_hba_mpi_version - determine in which MPI version class
12008  *					this device belongs to.
12009  * @pdev: PCI device struct
12010  *
12011  * return MPI2_VERSION for SAS 2.0 HBA devices,
12012  *	MPI25_VERSION for SAS 3.0 HBA devices, and
12013  *	MPI26 VERSION for Cutlass & Invader SAS 3.0 HBA devices
12014  */
12015 static u16
12016 _scsih_determine_hba_mpi_version(struct pci_dev *pdev)
12017 {
12018 
12019 	switch (pdev->device) {
12020 	case MPI2_MFGPAGE_DEVID_SSS6200:
12021 	case MPI2_MFGPAGE_DEVID_SAS2004:
12022 	case MPI2_MFGPAGE_DEVID_SAS2008:
12023 	case MPI2_MFGPAGE_DEVID_SAS2108_1:
12024 	case MPI2_MFGPAGE_DEVID_SAS2108_2:
12025 	case MPI2_MFGPAGE_DEVID_SAS2108_3:
12026 	case MPI2_MFGPAGE_DEVID_SAS2116_1:
12027 	case MPI2_MFGPAGE_DEVID_SAS2116_2:
12028 	case MPI2_MFGPAGE_DEVID_SAS2208_1:
12029 	case MPI2_MFGPAGE_DEVID_SAS2208_2:
12030 	case MPI2_MFGPAGE_DEVID_SAS2208_3:
12031 	case MPI2_MFGPAGE_DEVID_SAS2208_4:
12032 	case MPI2_MFGPAGE_DEVID_SAS2208_5:
12033 	case MPI2_MFGPAGE_DEVID_SAS2208_6:
12034 	case MPI2_MFGPAGE_DEVID_SAS2308_1:
12035 	case MPI2_MFGPAGE_DEVID_SAS2308_2:
12036 	case MPI2_MFGPAGE_DEVID_SAS2308_3:
12037 	case MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP:
12038 	case MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP_1:
12039 		return MPI2_VERSION;
12040 	case MPI25_MFGPAGE_DEVID_SAS3004:
12041 	case MPI25_MFGPAGE_DEVID_SAS3008:
12042 	case MPI25_MFGPAGE_DEVID_SAS3108_1:
12043 	case MPI25_MFGPAGE_DEVID_SAS3108_2:
12044 	case MPI25_MFGPAGE_DEVID_SAS3108_5:
12045 	case MPI25_MFGPAGE_DEVID_SAS3108_6:
12046 		return MPI25_VERSION;
12047 	case MPI26_MFGPAGE_DEVID_SAS3216:
12048 	case MPI26_MFGPAGE_DEVID_SAS3224:
12049 	case MPI26_MFGPAGE_DEVID_SAS3316_1:
12050 	case MPI26_MFGPAGE_DEVID_SAS3316_2:
12051 	case MPI26_MFGPAGE_DEVID_SAS3316_3:
12052 	case MPI26_MFGPAGE_DEVID_SAS3316_4:
12053 	case MPI26_MFGPAGE_DEVID_SAS3324_1:
12054 	case MPI26_MFGPAGE_DEVID_SAS3324_2:
12055 	case MPI26_MFGPAGE_DEVID_SAS3324_3:
12056 	case MPI26_MFGPAGE_DEVID_SAS3324_4:
12057 	case MPI26_MFGPAGE_DEVID_SAS3508:
12058 	case MPI26_MFGPAGE_DEVID_SAS3508_1:
12059 	case MPI26_MFGPAGE_DEVID_SAS3408:
12060 	case MPI26_MFGPAGE_DEVID_SAS3516:
12061 	case MPI26_MFGPAGE_DEVID_SAS3516_1:
12062 	case MPI26_MFGPAGE_DEVID_SAS3416:
12063 	case MPI26_MFGPAGE_DEVID_SAS3616:
12064 	case MPI26_ATLAS_PCIe_SWITCH_DEVID:
12065 	case MPI26_MFGPAGE_DEVID_CFG_SEC_3916:
12066 	case MPI26_MFGPAGE_DEVID_HARD_SEC_3916:
12067 	case MPI26_MFGPAGE_DEVID_CFG_SEC_3816:
12068 	case MPI26_MFGPAGE_DEVID_HARD_SEC_3816:
12069 	case MPI26_MFGPAGE_DEVID_INVALID0_3916:
12070 	case MPI26_MFGPAGE_DEVID_INVALID1_3916:
12071 	case MPI26_MFGPAGE_DEVID_INVALID0_3816:
12072 	case MPI26_MFGPAGE_DEVID_INVALID1_3816:
12073 		return MPI26_VERSION;
12074 	}
12075 	return 0;
12076 }
12077 
12078 /**
12079  * _scsih_probe - attach and add scsi host
12080  * @pdev: PCI device struct
12081  * @id: pci device id
12082  *
12083  * Return: 0 success, anything else error.
12084  */
12085 static int
12086 _scsih_probe(struct pci_dev *pdev, const struct pci_device_id *id)
12087 {
12088 	struct MPT3SAS_ADAPTER *ioc;
12089 	struct Scsi_Host *shost = NULL;
12090 	int rv;
12091 	u16 hba_mpi_version;
12092 	int iopoll_q_count = 0;
12093 
12094 	/* Determine in which MPI version class this pci device belongs */
12095 	hba_mpi_version = _scsih_determine_hba_mpi_version(pdev);
12096 	if (hba_mpi_version == 0)
12097 		return -ENODEV;
12098 
12099 	/* Enumerate only SAS 2.0 HBA's if hbas_to_enumerate is one,
12100 	 * for other generation HBA's return with -ENODEV
12101 	 */
12102 	if ((hbas_to_enumerate == 1) && (hba_mpi_version !=  MPI2_VERSION))
12103 		return -ENODEV;
12104 
12105 	/* Enumerate only SAS 3.0 HBA's if hbas_to_enumerate is two,
12106 	 * for other generation HBA's return with -ENODEV
12107 	 */
12108 	if ((hbas_to_enumerate == 2) && (!(hba_mpi_version ==  MPI25_VERSION
12109 		|| hba_mpi_version ==  MPI26_VERSION)))
12110 		return -ENODEV;
12111 
12112 	switch (hba_mpi_version) {
12113 	case MPI2_VERSION:
12114 		pci_disable_link_state(pdev, PCIE_LINK_STATE_L0S |
12115 			PCIE_LINK_STATE_L1 | PCIE_LINK_STATE_CLKPM);
12116 		/* Use mpt2sas driver host template for SAS 2.0 HBA's */
12117 		shost = scsi_host_alloc(&mpt2sas_driver_template,
12118 		  sizeof(struct MPT3SAS_ADAPTER));
12119 		if (!shost)
12120 			return -ENODEV;
12121 		ioc = shost_priv(shost);
12122 		memset(ioc, 0, sizeof(struct MPT3SAS_ADAPTER));
12123 		ioc->hba_mpi_version_belonged = hba_mpi_version;
12124 		ioc->id = mpt2_ids++;
12125 		sprintf(ioc->driver_name, "%s", MPT2SAS_DRIVER_NAME);
12126 		switch (pdev->device) {
12127 		case MPI2_MFGPAGE_DEVID_SSS6200:
12128 			ioc->is_warpdrive = 1;
12129 			ioc->hide_ir_msg = 1;
12130 			break;
12131 		case MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP:
12132 		case MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP_1:
12133 			ioc->is_mcpu_endpoint = 1;
12134 			break;
12135 		default:
12136 			ioc->mfg_pg10_hide_flag = MFG_PAGE10_EXPOSE_ALL_DISKS;
12137 			break;
12138 		}
12139 
12140 		if (multipath_on_hba == -1 || multipath_on_hba == 0)
12141 			ioc->multipath_on_hba = 0;
12142 		else
12143 			ioc->multipath_on_hba = 1;
12144 
12145 		break;
12146 	case MPI25_VERSION:
12147 	case MPI26_VERSION:
12148 		/* Use mpt3sas driver host template for SAS 3.0 HBA's */
12149 		shost = scsi_host_alloc(&mpt3sas_driver_template,
12150 		  sizeof(struct MPT3SAS_ADAPTER));
12151 		if (!shost)
12152 			return -ENODEV;
12153 		ioc = shost_priv(shost);
12154 		memset(ioc, 0, sizeof(struct MPT3SAS_ADAPTER));
12155 		ioc->hba_mpi_version_belonged = hba_mpi_version;
12156 		ioc->id = mpt3_ids++;
12157 		sprintf(ioc->driver_name, "%s", MPT3SAS_DRIVER_NAME);
12158 		switch (pdev->device) {
12159 		case MPI26_MFGPAGE_DEVID_SAS3508:
12160 		case MPI26_MFGPAGE_DEVID_SAS3508_1:
12161 		case MPI26_MFGPAGE_DEVID_SAS3408:
12162 		case MPI26_MFGPAGE_DEVID_SAS3516:
12163 		case MPI26_MFGPAGE_DEVID_SAS3516_1:
12164 		case MPI26_MFGPAGE_DEVID_SAS3416:
12165 		case MPI26_MFGPAGE_DEVID_SAS3616:
12166 		case MPI26_ATLAS_PCIe_SWITCH_DEVID:
12167 			ioc->is_gen35_ioc = 1;
12168 			break;
12169 		case MPI26_MFGPAGE_DEVID_INVALID0_3816:
12170 		case MPI26_MFGPAGE_DEVID_INVALID0_3916:
12171 			dev_err(&pdev->dev,
12172 			    "HBA with DeviceId 0x%04x, sub VendorId 0x%04x, sub DeviceId 0x%04x is Invalid",
12173 			    pdev->device, pdev->subsystem_vendor,
12174 			    pdev->subsystem_device);
12175 			return 1;
12176 		case MPI26_MFGPAGE_DEVID_INVALID1_3816:
12177 		case MPI26_MFGPAGE_DEVID_INVALID1_3916:
12178 			dev_err(&pdev->dev,
12179 			    "HBA with DeviceId 0x%04x, sub VendorId 0x%04x, sub DeviceId 0x%04x is Tampered",
12180 			    pdev->device, pdev->subsystem_vendor,
12181 			    pdev->subsystem_device);
12182 			return 1;
12183 		case MPI26_MFGPAGE_DEVID_CFG_SEC_3816:
12184 		case MPI26_MFGPAGE_DEVID_CFG_SEC_3916:
12185 			dev_info(&pdev->dev,
12186 			    "HBA is in Configurable Secure mode\n");
12187 			fallthrough;
12188 		case MPI26_MFGPAGE_DEVID_HARD_SEC_3816:
12189 		case MPI26_MFGPAGE_DEVID_HARD_SEC_3916:
12190 			ioc->is_aero_ioc = ioc->is_gen35_ioc = 1;
12191 			break;
12192 		default:
12193 			ioc->is_gen35_ioc = ioc->is_aero_ioc = 0;
12194 		}
12195 		if ((ioc->hba_mpi_version_belonged == MPI25_VERSION &&
12196 			pdev->revision >= SAS3_PCI_DEVICE_C0_REVISION) ||
12197 			(ioc->hba_mpi_version_belonged == MPI26_VERSION)) {
12198 			ioc->combined_reply_queue = 1;
12199 			if (ioc->is_gen35_ioc)
12200 				ioc->combined_reply_index_count =
12201 				 MPT3_SUP_REPLY_POST_HOST_INDEX_REG_COUNT_G35;
12202 			else
12203 				ioc->combined_reply_index_count =
12204 				 MPT3_SUP_REPLY_POST_HOST_INDEX_REG_COUNT_G3;
12205 		}
12206 
12207 		switch (ioc->is_gen35_ioc) {
12208 		case 0:
12209 			if (multipath_on_hba == -1 || multipath_on_hba == 0)
12210 				ioc->multipath_on_hba = 0;
12211 			else
12212 				ioc->multipath_on_hba = 1;
12213 			break;
12214 		case 1:
12215 			if (multipath_on_hba == -1 || multipath_on_hba > 0)
12216 				ioc->multipath_on_hba = 1;
12217 			else
12218 				ioc->multipath_on_hba = 0;
12219 			break;
12220 		default:
12221 			break;
12222 		}
12223 
12224 		break;
12225 	default:
12226 		return -ENODEV;
12227 	}
12228 
12229 	INIT_LIST_HEAD(&ioc->list);
12230 	spin_lock(&gioc_lock);
12231 	list_add_tail(&ioc->list, &mpt3sas_ioc_list);
12232 	spin_unlock(&gioc_lock);
12233 	ioc->shost = shost;
12234 	ioc->pdev = pdev;
12235 	ioc->scsi_io_cb_idx = scsi_io_cb_idx;
12236 	ioc->tm_cb_idx = tm_cb_idx;
12237 	ioc->ctl_cb_idx = ctl_cb_idx;
12238 	ioc->base_cb_idx = base_cb_idx;
12239 	ioc->port_enable_cb_idx = port_enable_cb_idx;
12240 	ioc->transport_cb_idx = transport_cb_idx;
12241 	ioc->scsih_cb_idx = scsih_cb_idx;
12242 	ioc->config_cb_idx = config_cb_idx;
12243 	ioc->tm_tr_cb_idx = tm_tr_cb_idx;
12244 	ioc->tm_tr_volume_cb_idx = tm_tr_volume_cb_idx;
12245 	ioc->tm_sas_control_cb_idx = tm_sas_control_cb_idx;
12246 	ioc->logging_level = logging_level;
12247 	ioc->schedule_dead_ioc_flush_running_cmds = &_scsih_flush_running_cmds;
12248 	/* Host waits for minimum of six seconds */
12249 	ioc->max_shutdown_latency = IO_UNIT_CONTROL_SHUTDOWN_TIMEOUT;
12250 	/*
12251 	 * Enable MEMORY MOVE support flag.
12252 	 */
12253 	ioc->drv_support_bitmap |= MPT_DRV_SUPPORT_BITMAP_MEMMOVE;
12254 	/* Enable ADDITIONAL QUERY support flag. */
12255 	ioc->drv_support_bitmap |= MPT_DRV_SUPPORT_BITMAP_ADDNLQUERY;
12256 
12257 	ioc->enable_sdev_max_qd = enable_sdev_max_qd;
12258 
12259 	/* misc semaphores and spin locks */
12260 	mutex_init(&ioc->reset_in_progress_mutex);
12261 	/* initializing pci_access_mutex lock */
12262 	mutex_init(&ioc->pci_access_mutex);
12263 	spin_lock_init(&ioc->ioc_reset_in_progress_lock);
12264 	spin_lock_init(&ioc->scsi_lookup_lock);
12265 	spin_lock_init(&ioc->sas_device_lock);
12266 	spin_lock_init(&ioc->sas_node_lock);
12267 	spin_lock_init(&ioc->fw_event_lock);
12268 	spin_lock_init(&ioc->raid_device_lock);
12269 	spin_lock_init(&ioc->pcie_device_lock);
12270 	spin_lock_init(&ioc->diag_trigger_lock);
12271 
12272 	INIT_LIST_HEAD(&ioc->sas_device_list);
12273 	INIT_LIST_HEAD(&ioc->sas_device_init_list);
12274 	INIT_LIST_HEAD(&ioc->sas_expander_list);
12275 	INIT_LIST_HEAD(&ioc->enclosure_list);
12276 	INIT_LIST_HEAD(&ioc->pcie_device_list);
12277 	INIT_LIST_HEAD(&ioc->pcie_device_init_list);
12278 	INIT_LIST_HEAD(&ioc->fw_event_list);
12279 	INIT_LIST_HEAD(&ioc->raid_device_list);
12280 	INIT_LIST_HEAD(&ioc->sas_hba.sas_port_list);
12281 	INIT_LIST_HEAD(&ioc->delayed_tr_list);
12282 	INIT_LIST_HEAD(&ioc->delayed_sc_list);
12283 	INIT_LIST_HEAD(&ioc->delayed_event_ack_list);
12284 	INIT_LIST_HEAD(&ioc->delayed_tr_volume_list);
12285 	INIT_LIST_HEAD(&ioc->reply_queue_list);
12286 	INIT_LIST_HEAD(&ioc->port_table_list);
12287 
12288 	sprintf(ioc->name, "%s_cm%d", ioc->driver_name, ioc->id);
12289 
12290 	/* init shost parameters */
12291 	shost->max_cmd_len = 32;
12292 	shost->max_lun = max_lun;
12293 	shost->transportt = mpt3sas_transport_template;
12294 	shost->unique_id = ioc->id;
12295 
12296 	if (ioc->is_mcpu_endpoint) {
12297 		/* mCPU MPI support 64K max IO */
12298 		shost->max_sectors = 128;
12299 		ioc_info(ioc, "The max_sectors value is set to %d\n",
12300 			 shost->max_sectors);
12301 	} else {
12302 		if (max_sectors != 0xFFFF) {
12303 			if (max_sectors < 64) {
12304 				shost->max_sectors = 64;
12305 				ioc_warn(ioc, "Invalid value %d passed for max_sectors, range is 64 to 32767. Assigning value of 64.\n",
12306 					 max_sectors);
12307 			} else if (max_sectors > 32767) {
12308 				shost->max_sectors = 32767;
12309 				ioc_warn(ioc, "Invalid value %d passed for max_sectors, range is 64 to 32767.Assigning default value of 32767.\n",
12310 					 max_sectors);
12311 			} else {
12312 				shost->max_sectors = max_sectors & 0xFFFE;
12313 				ioc_info(ioc, "The max_sectors value is set to %d\n",
12314 					 shost->max_sectors);
12315 			}
12316 		}
12317 	}
12318 	/* register EEDP capabilities with SCSI layer */
12319 	if (prot_mask >= 0)
12320 		scsi_host_set_prot(shost, (prot_mask & 0x07));
12321 	else
12322 		scsi_host_set_prot(shost, SHOST_DIF_TYPE1_PROTECTION
12323 				   | SHOST_DIF_TYPE2_PROTECTION
12324 				   | SHOST_DIF_TYPE3_PROTECTION);
12325 
12326 	scsi_host_set_guard(shost, SHOST_DIX_GUARD_CRC);
12327 
12328 	/* event thread */
12329 	snprintf(ioc->firmware_event_name, sizeof(ioc->firmware_event_name),
12330 	    "fw_event_%s%d", ioc->driver_name, ioc->id);
12331 	ioc->firmware_event_thread = alloc_ordered_workqueue(
12332 	    ioc->firmware_event_name, 0);
12333 	if (!ioc->firmware_event_thread) {
12334 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
12335 			__FILE__, __LINE__, __func__);
12336 		rv = -ENODEV;
12337 		goto out_thread_fail;
12338 	}
12339 
12340 	shost->host_tagset = 0;
12341 
12342 	if (ioc->is_gen35_ioc && host_tagset_enable)
12343 		shost->host_tagset = 1;
12344 
12345 	ioc->is_driver_loading = 1;
12346 	if ((mpt3sas_base_attach(ioc))) {
12347 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
12348 			__FILE__, __LINE__, __func__);
12349 		rv = -ENODEV;
12350 		goto out_attach_fail;
12351 	}
12352 
12353 	if (ioc->is_warpdrive) {
12354 		if (ioc->mfg_pg10_hide_flag ==  MFG_PAGE10_EXPOSE_ALL_DISKS)
12355 			ioc->hide_drives = 0;
12356 		else if (ioc->mfg_pg10_hide_flag ==  MFG_PAGE10_HIDE_ALL_DISKS)
12357 			ioc->hide_drives = 1;
12358 		else {
12359 			if (mpt3sas_get_num_volumes(ioc))
12360 				ioc->hide_drives = 1;
12361 			else
12362 				ioc->hide_drives = 0;
12363 		}
12364 	} else
12365 		ioc->hide_drives = 0;
12366 
12367 	shost->nr_hw_queues = 1;
12368 
12369 	if (shost->host_tagset) {
12370 		shost->nr_hw_queues =
12371 		    ioc->reply_queue_count - ioc->high_iops_queues;
12372 
12373 		iopoll_q_count =
12374 		    ioc->reply_queue_count - ioc->iopoll_q_start_index;
12375 
12376 		shost->nr_maps = iopoll_q_count ? 3 : 1;
12377 
12378 		dev_info(&ioc->pdev->dev,
12379 		    "Max SCSIIO MPT commands: %d shared with nr_hw_queues = %d\n",
12380 		    shost->can_queue, shost->nr_hw_queues);
12381 	}
12382 
12383 	rv = scsi_add_host(shost, &pdev->dev);
12384 	if (rv) {
12385 		ioc_err(ioc, "failure at %s:%d/%s()!\n",
12386 			__FILE__, __LINE__, __func__);
12387 		goto out_add_shost_fail;
12388 	}
12389 
12390 	scsi_scan_host(shost);
12391 	mpt3sas_setup_debugfs(ioc);
12392 	return 0;
12393 out_add_shost_fail:
12394 	mpt3sas_base_detach(ioc);
12395  out_attach_fail:
12396 	destroy_workqueue(ioc->firmware_event_thread);
12397  out_thread_fail:
12398 	spin_lock(&gioc_lock);
12399 	list_del(&ioc->list);
12400 	spin_unlock(&gioc_lock);
12401 	scsi_host_put(shost);
12402 	return rv;
12403 }
12404 
12405 /**
12406  * scsih_suspend - power management suspend main entry point
12407  * @dev: Device struct
12408  *
12409  * Return: 0 success, anything else error.
12410  */
12411 static int __maybe_unused
12412 scsih_suspend(struct device *dev)
12413 {
12414 	struct pci_dev *pdev = to_pci_dev(dev);
12415 	struct Scsi_Host *shost;
12416 	struct MPT3SAS_ADAPTER *ioc;
12417 	int rc;
12418 
12419 	rc = _scsih_get_shost_and_ioc(pdev, &shost, &ioc);
12420 	if (rc)
12421 		return rc;
12422 
12423 	mpt3sas_base_stop_watchdog(ioc);
12424 	scsi_block_requests(shost);
12425 	_scsih_nvme_shutdown(ioc);
12426 	ioc_info(ioc, "pdev=0x%p, slot=%s, entering operating state\n",
12427 		 pdev, pci_name(pdev));
12428 
12429 	mpt3sas_base_free_resources(ioc);
12430 	return 0;
12431 }
12432 
12433 /**
12434  * scsih_resume - power management resume main entry point
12435  * @dev: Device struct
12436  *
12437  * Return: 0 success, anything else error.
12438  */
12439 static int __maybe_unused
12440 scsih_resume(struct device *dev)
12441 {
12442 	struct pci_dev *pdev = to_pci_dev(dev);
12443 	struct Scsi_Host *shost;
12444 	struct MPT3SAS_ADAPTER *ioc;
12445 	pci_power_t device_state = pdev->current_state;
12446 	int r;
12447 
12448 	r = _scsih_get_shost_and_ioc(pdev, &shost, &ioc);
12449 	if (r)
12450 		return r;
12451 
12452 	ioc_info(ioc, "pdev=0x%p, slot=%s, previous operating state [D%d]\n",
12453 		 pdev, pci_name(pdev), device_state);
12454 
12455 	ioc->pdev = pdev;
12456 	r = mpt3sas_base_map_resources(ioc);
12457 	if (r)
12458 		return r;
12459 	ioc_info(ioc, "Issuing Hard Reset as part of OS Resume\n");
12460 	mpt3sas_base_hard_reset_handler(ioc, SOFT_RESET);
12461 	scsi_unblock_requests(shost);
12462 	mpt3sas_base_start_watchdog(ioc);
12463 	return 0;
12464 }
12465 
12466 /**
12467  * scsih_pci_error_detected - Called when a PCI error is detected.
12468  * @pdev: PCI device struct
12469  * @state: PCI channel state
12470  *
12471  * Description: Called when a PCI error is detected.
12472  *
12473  * Return: PCI_ERS_RESULT_NEED_RESET or PCI_ERS_RESULT_DISCONNECT.
12474  */
12475 static pci_ers_result_t
12476 scsih_pci_error_detected(struct pci_dev *pdev, pci_channel_state_t state)
12477 {
12478 	struct Scsi_Host *shost;
12479 	struct MPT3SAS_ADAPTER *ioc;
12480 
12481 	if (_scsih_get_shost_and_ioc(pdev, &shost, &ioc))
12482 		return PCI_ERS_RESULT_DISCONNECT;
12483 
12484 	ioc_info(ioc, "PCI error: detected callback, state(%d)!!\n", state);
12485 
12486 	switch (state) {
12487 	case pci_channel_io_normal:
12488 		return PCI_ERS_RESULT_CAN_RECOVER;
12489 	case pci_channel_io_frozen:
12490 		/* Fatal error, prepare for slot reset */
12491 		ioc->pci_error_recovery = 1;
12492 		scsi_block_requests(ioc->shost);
12493 		mpt3sas_base_stop_watchdog(ioc);
12494 		mpt3sas_base_free_resources(ioc);
12495 		return PCI_ERS_RESULT_NEED_RESET;
12496 	case pci_channel_io_perm_failure:
12497 		/* Permanent error, prepare for device removal */
12498 		ioc->pci_error_recovery = 1;
12499 		mpt3sas_base_stop_watchdog(ioc);
12500 		mpt3sas_base_pause_mq_polling(ioc);
12501 		_scsih_flush_running_cmds(ioc);
12502 		return PCI_ERS_RESULT_DISCONNECT;
12503 	}
12504 	return PCI_ERS_RESULT_NEED_RESET;
12505 }
12506 
12507 /**
12508  * scsih_pci_slot_reset - Called when PCI slot has been reset.
12509  * @pdev: PCI device struct
12510  *
12511  * Description: This routine is called by the pci error recovery
12512  * code after the PCI slot has been reset, just before we
12513  * should resume normal operations.
12514  */
12515 static pci_ers_result_t
12516 scsih_pci_slot_reset(struct pci_dev *pdev)
12517 {
12518 	struct Scsi_Host *shost;
12519 	struct MPT3SAS_ADAPTER *ioc;
12520 	int rc;
12521 
12522 	if (_scsih_get_shost_and_ioc(pdev, &shost, &ioc))
12523 		return PCI_ERS_RESULT_DISCONNECT;
12524 
12525 	ioc_info(ioc, "PCI error: slot reset callback!!\n");
12526 
12527 	ioc->pci_error_recovery = 0;
12528 	ioc->pdev = pdev;
12529 	pci_restore_state(pdev);
12530 	rc = mpt3sas_base_map_resources(ioc);
12531 	if (rc)
12532 		return PCI_ERS_RESULT_DISCONNECT;
12533 
12534 	ioc_info(ioc, "Issuing Hard Reset as part of PCI Slot Reset\n");
12535 	rc = mpt3sas_base_hard_reset_handler(ioc, FORCE_BIG_HAMMER);
12536 
12537 	ioc_warn(ioc, "hard reset: %s\n",
12538 		 (rc == 0) ? "success" : "failed");
12539 
12540 	if (!rc)
12541 		return PCI_ERS_RESULT_RECOVERED;
12542 	else
12543 		return PCI_ERS_RESULT_DISCONNECT;
12544 }
12545 
12546 /**
12547  * scsih_pci_resume() - resume normal ops after PCI reset
12548  * @pdev: pointer to PCI device
12549  *
12550  * Called when the error recovery driver tells us that its
12551  * OK to resume normal operation. Use completion to allow
12552  * halted scsi ops to resume.
12553  */
12554 static void
12555 scsih_pci_resume(struct pci_dev *pdev)
12556 {
12557 	struct Scsi_Host *shost;
12558 	struct MPT3SAS_ADAPTER *ioc;
12559 
12560 	if (_scsih_get_shost_and_ioc(pdev, &shost, &ioc))
12561 		return;
12562 
12563 	ioc_info(ioc, "PCI error: resume callback!!\n");
12564 
12565 	mpt3sas_base_start_watchdog(ioc);
12566 	scsi_unblock_requests(ioc->shost);
12567 }
12568 
12569 /**
12570  * scsih_pci_mmio_enabled - Enable MMIO and dump debug registers
12571  * @pdev: pointer to PCI device
12572  */
12573 static pci_ers_result_t
12574 scsih_pci_mmio_enabled(struct pci_dev *pdev)
12575 {
12576 	struct Scsi_Host *shost;
12577 	struct MPT3SAS_ADAPTER *ioc;
12578 
12579 	if (_scsih_get_shost_and_ioc(pdev, &shost, &ioc))
12580 		return PCI_ERS_RESULT_DISCONNECT;
12581 
12582 	ioc_info(ioc, "PCI error: mmio enabled callback!!\n");
12583 
12584 	/* TODO - dump whatever for debugging purposes */
12585 
12586 	/* This called only if scsih_pci_error_detected returns
12587 	 * PCI_ERS_RESULT_CAN_RECOVER. Read/write to the device still
12588 	 * works, no need to reset slot.
12589 	 */
12590 	return PCI_ERS_RESULT_RECOVERED;
12591 }
12592 
12593 /**
12594  * scsih_ncq_prio_supp - Check for NCQ command priority support
12595  * @sdev: scsi device struct
12596  *
12597  * This is called when a user indicates they would like to enable
12598  * ncq command priorities. This works only on SATA devices.
12599  */
12600 bool scsih_ncq_prio_supp(struct scsi_device *sdev)
12601 {
12602 	struct scsi_vpd *vpd;
12603 	bool ncq_prio_supp = false;
12604 
12605 	rcu_read_lock();
12606 	vpd = rcu_dereference(sdev->vpd_pg89);
12607 	if (!vpd || vpd->len < 214)
12608 		goto out;
12609 
12610 	ncq_prio_supp = (vpd->data[213] >> 4) & 1;
12611 out:
12612 	rcu_read_unlock();
12613 
12614 	return ncq_prio_supp;
12615 }
12616 /*
12617  * The pci device ids are defined in mpi/mpi2_cnfg.h.
12618  */
12619 static const struct pci_device_id mpt3sas_pci_table[] = {
12620 	/* Spitfire ~ 2004 */
12621 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2004,
12622 		PCI_ANY_ID, PCI_ANY_ID },
12623 	/* Falcon ~ 2008 */
12624 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2008,
12625 		PCI_ANY_ID, PCI_ANY_ID },
12626 	/* Liberator ~ 2108 */
12627 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2108_1,
12628 		PCI_ANY_ID, PCI_ANY_ID },
12629 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2108_2,
12630 		PCI_ANY_ID, PCI_ANY_ID },
12631 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2108_3,
12632 		PCI_ANY_ID, PCI_ANY_ID },
12633 	/* Meteor ~ 2116 */
12634 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2116_1,
12635 		PCI_ANY_ID, PCI_ANY_ID },
12636 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2116_2,
12637 		PCI_ANY_ID, PCI_ANY_ID },
12638 	/* Thunderbolt ~ 2208 */
12639 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_1,
12640 		PCI_ANY_ID, PCI_ANY_ID },
12641 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_2,
12642 		PCI_ANY_ID, PCI_ANY_ID },
12643 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_3,
12644 		PCI_ANY_ID, PCI_ANY_ID },
12645 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_4,
12646 		PCI_ANY_ID, PCI_ANY_ID },
12647 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_5,
12648 		PCI_ANY_ID, PCI_ANY_ID },
12649 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_6,
12650 		PCI_ANY_ID, PCI_ANY_ID },
12651 	/* Mustang ~ 2308 */
12652 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2308_1,
12653 		PCI_ANY_ID, PCI_ANY_ID },
12654 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2308_2,
12655 		PCI_ANY_ID, PCI_ANY_ID },
12656 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2308_3,
12657 		PCI_ANY_ID, PCI_ANY_ID },
12658 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP,
12659 		PCI_ANY_ID, PCI_ANY_ID },
12660 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP_1,
12661 		PCI_ANY_ID, PCI_ANY_ID },
12662 	/* SSS6200 */
12663 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SSS6200,
12664 		PCI_ANY_ID, PCI_ANY_ID },
12665 	/* Fury ~ 3004 and 3008 */
12666 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3004,
12667 		PCI_ANY_ID, PCI_ANY_ID },
12668 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3008,
12669 		PCI_ANY_ID, PCI_ANY_ID },
12670 	/* Invader ~ 3108 */
12671 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3108_1,
12672 		PCI_ANY_ID, PCI_ANY_ID },
12673 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3108_2,
12674 		PCI_ANY_ID, PCI_ANY_ID },
12675 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3108_5,
12676 		PCI_ANY_ID, PCI_ANY_ID },
12677 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3108_6,
12678 		PCI_ANY_ID, PCI_ANY_ID },
12679 	/* Cutlass ~ 3216 and 3224 */
12680 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3216,
12681 		PCI_ANY_ID, PCI_ANY_ID },
12682 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3224,
12683 		PCI_ANY_ID, PCI_ANY_ID },
12684 	/* Intruder ~ 3316 and 3324 */
12685 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3316_1,
12686 		PCI_ANY_ID, PCI_ANY_ID },
12687 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3316_2,
12688 		PCI_ANY_ID, PCI_ANY_ID },
12689 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3316_3,
12690 		PCI_ANY_ID, PCI_ANY_ID },
12691 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3316_4,
12692 		PCI_ANY_ID, PCI_ANY_ID },
12693 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3324_1,
12694 		PCI_ANY_ID, PCI_ANY_ID },
12695 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3324_2,
12696 		PCI_ANY_ID, PCI_ANY_ID },
12697 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3324_3,
12698 		PCI_ANY_ID, PCI_ANY_ID },
12699 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3324_4,
12700 		PCI_ANY_ID, PCI_ANY_ID },
12701 	/* Ventura, Crusader, Harpoon & Tomcat ~ 3516, 3416, 3508 & 3408*/
12702 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3508,
12703 		PCI_ANY_ID, PCI_ANY_ID },
12704 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3508_1,
12705 		PCI_ANY_ID, PCI_ANY_ID },
12706 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3408,
12707 		PCI_ANY_ID, PCI_ANY_ID },
12708 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3516,
12709 		PCI_ANY_ID, PCI_ANY_ID },
12710 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3516_1,
12711 		PCI_ANY_ID, PCI_ANY_ID },
12712 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3416,
12713 		PCI_ANY_ID, PCI_ANY_ID },
12714 	/* Mercator ~ 3616*/
12715 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3616,
12716 		PCI_ANY_ID, PCI_ANY_ID },
12717 
12718 	/* Aero SI 0x00E1 Configurable Secure
12719 	 * 0x00E2 Hard Secure
12720 	 */
12721 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_CFG_SEC_3916,
12722 		PCI_ANY_ID, PCI_ANY_ID },
12723 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_HARD_SEC_3916,
12724 		PCI_ANY_ID, PCI_ANY_ID },
12725 
12726 	/*
12727 	 *  Aero SI –> 0x00E0 Invalid, 0x00E3 Tampered
12728 	 */
12729 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_INVALID0_3916,
12730 		PCI_ANY_ID, PCI_ANY_ID },
12731 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_INVALID1_3916,
12732 		PCI_ANY_ID, PCI_ANY_ID },
12733 
12734 	/* Atlas PCIe Switch Management Port */
12735 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_ATLAS_PCIe_SWITCH_DEVID,
12736 		PCI_ANY_ID, PCI_ANY_ID },
12737 
12738 	/* Sea SI 0x00E5 Configurable Secure
12739 	 * 0x00E6 Hard Secure
12740 	 */
12741 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_CFG_SEC_3816,
12742 		PCI_ANY_ID, PCI_ANY_ID },
12743 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_HARD_SEC_3816,
12744 		PCI_ANY_ID, PCI_ANY_ID },
12745 
12746 	/*
12747 	 * ATTO Branded ExpressSAS H12xx GT
12748 	 */
12749 	{ MPI2_MFGPAGE_VENDORID_ATTO, MPI26_MFGPAGE_DEVID_HARD_SEC_3816,
12750 		PCI_ANY_ID, PCI_ANY_ID },
12751 
12752 	/*
12753 	 *  Sea SI –> 0x00E4 Invalid, 0x00E7 Tampered
12754 	 */
12755 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_INVALID0_3816,
12756 		PCI_ANY_ID, PCI_ANY_ID },
12757 	{ MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_INVALID1_3816,
12758 		PCI_ANY_ID, PCI_ANY_ID },
12759 
12760 	{0}     /* Terminating entry */
12761 };
12762 MODULE_DEVICE_TABLE(pci, mpt3sas_pci_table);
12763 
12764 static struct pci_error_handlers _mpt3sas_err_handler = {
12765 	.error_detected	= scsih_pci_error_detected,
12766 	.mmio_enabled	= scsih_pci_mmio_enabled,
12767 	.slot_reset	= scsih_pci_slot_reset,
12768 	.resume		= scsih_pci_resume,
12769 };
12770 
12771 static SIMPLE_DEV_PM_OPS(scsih_pm_ops, scsih_suspend, scsih_resume);
12772 
12773 static struct pci_driver mpt3sas_driver = {
12774 	.name		= MPT3SAS_DRIVER_NAME,
12775 	.id_table	= mpt3sas_pci_table,
12776 	.probe		= _scsih_probe,
12777 	.remove		= scsih_remove,
12778 	.shutdown	= scsih_shutdown,
12779 	.err_handler	= &_mpt3sas_err_handler,
12780 	.driver.pm	= &scsih_pm_ops,
12781 };
12782 
12783 /**
12784  * scsih_init - main entry point for this driver.
12785  *
12786  * Return: 0 success, anything else error.
12787  */
12788 static int
12789 scsih_init(void)
12790 {
12791 	mpt2_ids = 0;
12792 	mpt3_ids = 0;
12793 
12794 	mpt3sas_base_initialize_callback_handler();
12795 
12796 	 /* queuecommand callback hander */
12797 	scsi_io_cb_idx = mpt3sas_base_register_callback_handler(_scsih_io_done);
12798 
12799 	/* task management callback handler */
12800 	tm_cb_idx = mpt3sas_base_register_callback_handler(_scsih_tm_done);
12801 
12802 	/* base internal commands callback handler */
12803 	base_cb_idx = mpt3sas_base_register_callback_handler(mpt3sas_base_done);
12804 	port_enable_cb_idx = mpt3sas_base_register_callback_handler(
12805 	    mpt3sas_port_enable_done);
12806 
12807 	/* transport internal commands callback handler */
12808 	transport_cb_idx = mpt3sas_base_register_callback_handler(
12809 	    mpt3sas_transport_done);
12810 
12811 	/* scsih internal commands callback handler */
12812 	scsih_cb_idx = mpt3sas_base_register_callback_handler(_scsih_done);
12813 
12814 	/* configuration page API internal commands callback handler */
12815 	config_cb_idx = mpt3sas_base_register_callback_handler(
12816 	    mpt3sas_config_done);
12817 
12818 	/* ctl module callback handler */
12819 	ctl_cb_idx = mpt3sas_base_register_callback_handler(mpt3sas_ctl_done);
12820 
12821 	tm_tr_cb_idx = mpt3sas_base_register_callback_handler(
12822 	    _scsih_tm_tr_complete);
12823 
12824 	tm_tr_volume_cb_idx = mpt3sas_base_register_callback_handler(
12825 	    _scsih_tm_volume_tr_complete);
12826 
12827 	tm_sas_control_cb_idx = mpt3sas_base_register_callback_handler(
12828 	    _scsih_sas_control_complete);
12829 
12830 	mpt3sas_init_debugfs();
12831 	return 0;
12832 }
12833 
12834 /**
12835  * scsih_exit - exit point for this driver (when it is a module).
12836  *
12837  * Return: 0 success, anything else error.
12838  */
12839 static void
12840 scsih_exit(void)
12841 {
12842 
12843 	mpt3sas_base_release_callback_handler(scsi_io_cb_idx);
12844 	mpt3sas_base_release_callback_handler(tm_cb_idx);
12845 	mpt3sas_base_release_callback_handler(base_cb_idx);
12846 	mpt3sas_base_release_callback_handler(port_enable_cb_idx);
12847 	mpt3sas_base_release_callback_handler(transport_cb_idx);
12848 	mpt3sas_base_release_callback_handler(scsih_cb_idx);
12849 	mpt3sas_base_release_callback_handler(config_cb_idx);
12850 	mpt3sas_base_release_callback_handler(ctl_cb_idx);
12851 
12852 	mpt3sas_base_release_callback_handler(tm_tr_cb_idx);
12853 	mpt3sas_base_release_callback_handler(tm_tr_volume_cb_idx);
12854 	mpt3sas_base_release_callback_handler(tm_sas_control_cb_idx);
12855 
12856 /* raid transport support */
12857 	if (hbas_to_enumerate != 1)
12858 		raid_class_release(mpt3sas_raid_template);
12859 	if (hbas_to_enumerate != 2)
12860 		raid_class_release(mpt2sas_raid_template);
12861 	sas_release_transport(mpt3sas_transport_template);
12862 	mpt3sas_exit_debugfs();
12863 }
12864 
12865 /**
12866  * _mpt3sas_init - main entry point for this driver.
12867  *
12868  * Return: 0 success, anything else error.
12869  */
12870 static int __init
12871 _mpt3sas_init(void)
12872 {
12873 	int error;
12874 
12875 	pr_info("%s version %s loaded\n", MPT3SAS_DRIVER_NAME,
12876 					MPT3SAS_DRIVER_VERSION);
12877 
12878 	mpt3sas_transport_template =
12879 	    sas_attach_transport(&mpt3sas_transport_functions);
12880 	if (!mpt3sas_transport_template)
12881 		return -ENODEV;
12882 
12883 	/* No need attach mpt3sas raid functions template
12884 	 * if hbas_to_enumarate value is one.
12885 	 */
12886 	if (hbas_to_enumerate != 1) {
12887 		mpt3sas_raid_template =
12888 				raid_class_attach(&mpt3sas_raid_functions);
12889 		if (!mpt3sas_raid_template) {
12890 			sas_release_transport(mpt3sas_transport_template);
12891 			return -ENODEV;
12892 		}
12893 	}
12894 
12895 	/* No need to attach mpt2sas raid functions template
12896 	 * if hbas_to_enumarate value is two
12897 	 */
12898 	if (hbas_to_enumerate != 2) {
12899 		mpt2sas_raid_template =
12900 				raid_class_attach(&mpt2sas_raid_functions);
12901 		if (!mpt2sas_raid_template) {
12902 			sas_release_transport(mpt3sas_transport_template);
12903 			return -ENODEV;
12904 		}
12905 	}
12906 
12907 	error = scsih_init();
12908 	if (error) {
12909 		scsih_exit();
12910 		return error;
12911 	}
12912 
12913 	mpt3sas_ctl_init(hbas_to_enumerate);
12914 
12915 	error = pci_register_driver(&mpt3sas_driver);
12916 	if (error) {
12917 		mpt3sas_ctl_exit(hbas_to_enumerate);
12918 		scsih_exit();
12919 	}
12920 
12921 	return error;
12922 }
12923 
12924 /**
12925  * _mpt3sas_exit - exit point for this driver (when it is a module).
12926  *
12927  */
12928 static void __exit
12929 _mpt3sas_exit(void)
12930 {
12931 	pr_info("mpt3sas version %s unloading\n",
12932 				MPT3SAS_DRIVER_VERSION);
12933 
12934 	pci_unregister_driver(&mpt3sas_driver);
12935 
12936 	mpt3sas_ctl_exit(hbas_to_enumerate);
12937 
12938 	scsih_exit();
12939 }
12940 
12941 module_init(_mpt3sas_init);
12942 module_exit(_mpt3sas_exit);
12943