1 | /* |
2 | * Scsi Host Layer for MPT (Message Passing Technology) based controllers |
3 | * |
4 | * This code is based on drivers/scsi/mpt3sas/mpt3sas_scsih.c |
5 | * Copyright (C) 2012-2014 LSI Corporation |
6 | * Copyright (C) 2013-2014 Avago Technologies |
7 | * (mailto: MPT-FusionLinux.pdl@avagotech.com) |
8 | * |
9 | * This program is free software; you can redistribute it and/or |
10 | * modify it under the terms of the GNU General Public License |
11 | * as published by the Free Software Foundation; either version 2 |
12 | * of the License, or (at your option) any later version. |
13 | * |
14 | * This program is distributed in the hope that it will be useful, |
15 | * but WITHOUT ANY WARRANTY; without even the implied warranty of |
16 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
17 | * GNU General Public License for more details. |
18 | * |
19 | * NO WARRANTY |
20 | * THE PROGRAM IS PROVIDED ON AN "AS IS" BASIS, WITHOUT WARRANTIES OR |
21 | * CONDITIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED INCLUDING, WITHOUT |
22 | * LIMITATION, ANY WARRANTIES OR CONDITIONS OF TITLE, NON-INFRINGEMENT, |
23 | * MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Each Recipient is |
24 | * solely responsible for determining the appropriateness of using and |
25 | * distributing the Program and assumes all risks associated with its |
26 | * exercise of rights under this Agreement, including but not limited to |
27 | * the risks and costs of program errors, damage to or loss of data, |
28 | * programs or equipment, and unavailability or interruption of operations. |
29 | |
30 | * DISCLAIMER OF LIABILITY |
31 | * NEITHER RECIPIENT NOR ANY CONTRIBUTORS SHALL HAVE ANY LIABILITY FOR ANY |
32 | * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL |
33 | * DAMAGES (INCLUDING WITHOUT LIMITATION LOST PROFITS), HOWEVER CAUSED AND |
34 | * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR |
35 | * TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE |
36 | * USE OR DISTRIBUTION OF THE PROGRAM OR THE EXERCISE OF ANY RIGHTS GRANTED |
37 | * HEREUNDER, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGES |
38 | |
39 | * You should have received a copy of the GNU General Public License |
40 | * along with this program; if not, write to the Free Software |
41 | * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, |
42 | * USA. |
43 | */ |
44 | |
45 | #include <linux/module.h> |
46 | #include <linux/kernel.h> |
47 | #include <linux/init.h> |
48 | #include <linux/errno.h> |
49 | #include <linux/blkdev.h> |
50 | #include <linux/sched.h> |
51 | #include <linux/workqueue.h> |
52 | #include <linux/delay.h> |
53 | #include <linux/pci.h> |
54 | #include <linux/interrupt.h> |
55 | #include <linux/raid_class.h> |
56 | #include <linux/blk-mq-pci.h> |
57 | #include <asm/unaligned.h> |
58 | |
59 | #include "mpt3sas_base.h" |
60 | |
61 | #define RAID_CHANNEL 1 |
62 | |
63 | #define PCIE_CHANNEL 2 |
64 | |
65 | /* forward proto's */ |
66 | static void _scsih_expander_node_remove(struct MPT3SAS_ADAPTER *ioc, |
67 | struct _sas_node *sas_expander); |
68 | static void _firmware_event_work(struct work_struct *work); |
69 | |
70 | static void _scsih_remove_device(struct MPT3SAS_ADAPTER *ioc, |
71 | struct _sas_device *sas_device); |
72 | static int _scsih_add_device(struct MPT3SAS_ADAPTER *ioc, u16 handle, |
73 | u8 retry_count, u8 is_pd); |
74 | static int _scsih_pcie_add_device(struct MPT3SAS_ADAPTER *ioc, u16 handle); |
75 | static void _scsih_pcie_device_remove_from_sml(struct MPT3SAS_ADAPTER *ioc, |
76 | struct _pcie_device *pcie_device); |
77 | static void |
78 | _scsih_pcie_check_device(struct MPT3SAS_ADAPTER *ioc, u16 handle); |
79 | static u8 _scsih_check_for_pending_tm(struct MPT3SAS_ADAPTER *ioc, u16 smid); |
80 | static void _scsih_complete_devices_scanning(struct MPT3SAS_ADAPTER *ioc); |
81 | |
82 | /* global parameters */ |
83 | LIST_HEAD(mpt3sas_ioc_list); |
84 | /* global ioc lock for list operations */ |
85 | DEFINE_SPINLOCK(gioc_lock); |
86 | |
87 | MODULE_AUTHOR(MPT3SAS_AUTHOR); |
88 | MODULE_DESCRIPTION(MPT3SAS_DESCRIPTION); |
89 | MODULE_LICENSE("GPL" ); |
90 | MODULE_VERSION(MPT3SAS_DRIVER_VERSION); |
91 | MODULE_ALIAS("mpt2sas" ); |
92 | |
93 | /* local parameters */ |
94 | static u8 scsi_io_cb_idx = -1; |
95 | static u8 tm_cb_idx = -1; |
96 | static u8 ctl_cb_idx = -1; |
97 | static u8 base_cb_idx = -1; |
98 | static u8 port_enable_cb_idx = -1; |
99 | static u8 transport_cb_idx = -1; |
100 | static u8 scsih_cb_idx = -1; |
101 | static u8 config_cb_idx = -1; |
102 | static int mpt2_ids; |
103 | static int mpt3_ids; |
104 | |
105 | static u8 tm_tr_cb_idx = -1 ; |
106 | static u8 tm_tr_volume_cb_idx = -1 ; |
107 | static u8 tm_sas_control_cb_idx = -1; |
108 | |
109 | /* command line options */ |
110 | static u32 logging_level; |
111 | MODULE_PARM_DESC(logging_level, |
112 | " bits for enabling additional logging info (default=0)" ); |
113 | |
114 | |
115 | static ushort max_sectors = 0xFFFF; |
116 | module_param(max_sectors, ushort, 0444); |
117 | MODULE_PARM_DESC(max_sectors, "max sectors, range 64 to 32767 default=32767" ); |
118 | |
119 | |
120 | static int missing_delay[2] = {-1, -1}; |
121 | module_param_array(missing_delay, int, NULL, 0444); |
122 | MODULE_PARM_DESC(missing_delay, " device missing delay , io missing delay" ); |
123 | |
124 | /* scsi-mid layer global parmeter is max_report_luns, which is 511 */ |
125 | #define MPT3SAS_MAX_LUN (16895) |
126 | static u64 max_lun = MPT3SAS_MAX_LUN; |
127 | module_param(max_lun, ullong, 0444); |
128 | MODULE_PARM_DESC(max_lun, " max lun, default=16895 " ); |
129 | |
130 | static ushort hbas_to_enumerate; |
131 | module_param(hbas_to_enumerate, ushort, 0444); |
132 | MODULE_PARM_DESC(hbas_to_enumerate, |
133 | " 0 - enumerates both SAS 2.0 & SAS 3.0 generation HBAs\n \ |
134 | 1 - enumerates only SAS 2.0 generation HBAs\n \ |
135 | 2 - enumerates only SAS 3.0 generation HBAs (default=0)" ); |
136 | |
137 | /* diag_buffer_enable is bitwise |
138 | * bit 0 set = TRACE |
139 | * bit 1 set = SNAPSHOT |
140 | * bit 2 set = EXTENDED |
141 | * |
142 | * Either bit can be set, or both |
143 | */ |
144 | static int diag_buffer_enable = -1; |
145 | module_param(diag_buffer_enable, int, 0444); |
146 | MODULE_PARM_DESC(diag_buffer_enable, |
147 | " post diag buffers (TRACE=1/SNAPSHOT=2/EXTENDED=4/default=0)" ); |
148 | static int disable_discovery = -1; |
149 | module_param(disable_discovery, int, 0444); |
150 | MODULE_PARM_DESC(disable_discovery, " disable discovery " ); |
151 | |
152 | |
153 | /* permit overriding the host protection capabilities mask (EEDP/T10 PI) */ |
154 | static int prot_mask = -1; |
155 | module_param(prot_mask, int, 0444); |
156 | MODULE_PARM_DESC(prot_mask, " host protection capabilities mask, def=7 " ); |
157 | |
158 | static bool enable_sdev_max_qd; |
159 | module_param(enable_sdev_max_qd, bool, 0444); |
160 | MODULE_PARM_DESC(enable_sdev_max_qd, |
161 | "Enable sdev max qd as can_queue, def=disabled(0)" ); |
162 | |
163 | static int multipath_on_hba = -1; |
164 | module_param(multipath_on_hba, int, 0); |
165 | MODULE_PARM_DESC(multipath_on_hba, |
166 | "Multipath support to add same target device\n\t\t" |
167 | "as many times as it is visible to HBA from various paths\n\t\t" |
168 | "(by default:\n\t\t" |
169 | "\t SAS 2.0 & SAS 3.0 HBA - This will be disabled,\n\t\t" |
170 | "\t SAS 3.5 HBA - This will be enabled)" ); |
171 | |
172 | static int host_tagset_enable = 1; |
173 | module_param(host_tagset_enable, int, 0444); |
174 | MODULE_PARM_DESC(host_tagset_enable, |
175 | "Shared host tagset enable/disable Default: enable(1)" ); |
176 | |
177 | /* raid transport support */ |
178 | static struct raid_template *mpt3sas_raid_template; |
179 | static struct raid_template *mpt2sas_raid_template; |
180 | |
181 | |
182 | /** |
183 | * struct sense_info - common structure for obtaining sense keys |
184 | * @skey: sense key |
185 | * @asc: additional sense code |
186 | * @ascq: additional sense code qualifier |
187 | */ |
188 | struct sense_info { |
189 | u8 skey; |
190 | u8 asc; |
191 | u8 ascq; |
192 | }; |
193 | |
194 | #define MPT3SAS_PROCESS_TRIGGER_DIAG (0xFFFB) |
195 | #define MPT3SAS_TURN_ON_PFA_LED (0xFFFC) |
196 | #define MPT3SAS_PORT_ENABLE_COMPLETE (0xFFFD) |
197 | #define MPT3SAS_ABRT_TASK_SET (0xFFFE) |
198 | #define MPT3SAS_REMOVE_UNRESPONDING_DEVICES (0xFFFF) |
199 | /** |
200 | * struct fw_event_work - firmware event struct |
201 | * @list: link list framework |
202 | * @work: work object (ioc->fault_reset_work_q) |
203 | * @ioc: per adapter object |
204 | * @device_handle: device handle |
205 | * @VF_ID: virtual function id |
206 | * @VP_ID: virtual port id |
207 | * @ignore: flag meaning this event has been marked to ignore |
208 | * @event: firmware event MPI2_EVENT_XXX defined in mpi2_ioc.h |
209 | * @refcount: kref for this event |
210 | * @event_data: reply event data payload follows |
211 | * |
212 | * This object stored on ioc->fw_event_list. |
213 | */ |
214 | struct fw_event_work { |
215 | struct list_head list; |
216 | struct work_struct work; |
217 | |
218 | struct MPT3SAS_ADAPTER *ioc; |
219 | u16 device_handle; |
220 | u8 VF_ID; |
221 | u8 VP_ID; |
222 | u8 ignore; |
223 | u16 event; |
224 | struct kref refcount; |
225 | char event_data[] __aligned(4); |
226 | }; |
227 | |
228 | static void fw_event_work_free(struct kref *r) |
229 | { |
230 | kfree(container_of(r, struct fw_event_work, refcount)); |
231 | } |
232 | |
233 | static void fw_event_work_get(struct fw_event_work *fw_work) |
234 | { |
235 | kref_get(kref: &fw_work->refcount); |
236 | } |
237 | |
238 | static void fw_event_work_put(struct fw_event_work *fw_work) |
239 | { |
240 | kref_put(kref: &fw_work->refcount, release: fw_event_work_free); |
241 | } |
242 | |
243 | static struct fw_event_work *alloc_fw_event_work(int len) |
244 | { |
245 | struct fw_event_work *fw_event; |
246 | |
247 | fw_event = kzalloc(size: sizeof(*fw_event) + len, GFP_ATOMIC); |
248 | if (!fw_event) |
249 | return NULL; |
250 | |
251 | kref_init(kref: &fw_event->refcount); |
252 | return fw_event; |
253 | } |
254 | |
255 | /** |
256 | * struct _scsi_io_transfer - scsi io transfer |
257 | * @handle: sas device handle (assigned by firmware) |
258 | * @is_raid: flag set for hidden raid components |
259 | * @dir: DMA_TO_DEVICE, DMA_FROM_DEVICE, |
260 | * @data_length: data transfer length |
261 | * @data_dma: dma pointer to data |
262 | * @sense: sense data |
263 | * @lun: lun number |
264 | * @cdb_length: cdb length |
265 | * @cdb: cdb contents |
266 | * @timeout: timeout for this command |
267 | * @VF_ID: virtual function id |
268 | * @VP_ID: virtual port id |
269 | * @valid_reply: flag set for reply message |
270 | * @sense_length: sense length |
271 | * @ioc_status: ioc status |
272 | * @scsi_state: scsi state |
273 | * @scsi_status: scsi staus |
274 | * @log_info: log information |
275 | * @transfer_length: data length transfer when there is a reply message |
276 | * |
277 | * Used for sending internal scsi commands to devices within this module. |
278 | * Refer to _scsi_send_scsi_io(). |
279 | */ |
280 | struct _scsi_io_transfer { |
281 | u16 handle; |
282 | u8 is_raid; |
283 | enum dma_data_direction dir; |
284 | u32 data_length; |
285 | dma_addr_t data_dma; |
286 | u8 sense[SCSI_SENSE_BUFFERSIZE]; |
287 | u32 lun; |
288 | u8 cdb_length; |
289 | u8 cdb[32]; |
290 | u8 timeout; |
291 | u8 VF_ID; |
292 | u8 VP_ID; |
293 | u8 valid_reply; |
294 | /* the following bits are only valid when 'valid_reply = 1' */ |
295 | u32 sense_length; |
296 | u16 ioc_status; |
297 | u8 scsi_state; |
298 | u8 scsi_status; |
299 | u32 log_info; |
300 | u32 transfer_length; |
301 | }; |
302 | |
303 | /** |
304 | * _scsih_set_debug_level - global setting of ioc->logging_level. |
305 | * @val: ? |
306 | * @kp: ? |
307 | * |
308 | * Note: The logging levels are defined in mpt3sas_debug.h. |
309 | */ |
310 | static int |
311 | _scsih_set_debug_level(const char *val, const struct kernel_param *kp) |
312 | { |
313 | int ret = param_set_int(val, kp); |
314 | struct MPT3SAS_ADAPTER *ioc; |
315 | |
316 | if (ret) |
317 | return ret; |
318 | |
319 | pr_info("setting logging_level(0x%08x)\n" , logging_level); |
320 | spin_lock(lock: &gioc_lock); |
321 | list_for_each_entry(ioc, &mpt3sas_ioc_list, list) |
322 | ioc->logging_level = logging_level; |
323 | spin_unlock(lock: &gioc_lock); |
324 | return 0; |
325 | } |
326 | module_param_call(logging_level, _scsih_set_debug_level, param_get_int, |
327 | &logging_level, 0644); |
328 | |
329 | /** |
330 | * _scsih_srch_boot_sas_address - search based on sas_address |
331 | * @sas_address: sas address |
332 | * @boot_device: boot device object from bios page 2 |
333 | * |
334 | * Return: 1 when there's a match, 0 means no match. |
335 | */ |
336 | static inline int |
337 | _scsih_srch_boot_sas_address(u64 sas_address, |
338 | Mpi2BootDeviceSasWwid_t *boot_device) |
339 | { |
340 | return (sas_address == le64_to_cpu(boot_device->SASAddress)) ? 1 : 0; |
341 | } |
342 | |
343 | /** |
344 | * _scsih_srch_boot_device_name - search based on device name |
345 | * @device_name: device name specified in INDENTIFY fram |
346 | * @boot_device: boot device object from bios page 2 |
347 | * |
348 | * Return: 1 when there's a match, 0 means no match. |
349 | */ |
350 | static inline int |
351 | _scsih_srch_boot_device_name(u64 device_name, |
352 | Mpi2BootDeviceDeviceName_t *boot_device) |
353 | { |
354 | return (device_name == le64_to_cpu(boot_device->DeviceName)) ? 1 : 0; |
355 | } |
356 | |
357 | /** |
358 | * _scsih_srch_boot_encl_slot - search based on enclosure_logical_id/slot |
359 | * @enclosure_logical_id: enclosure logical id |
360 | * @slot_number: slot number |
361 | * @boot_device: boot device object from bios page 2 |
362 | * |
363 | * Return: 1 when there's a match, 0 means no match. |
364 | */ |
365 | static inline int |
366 | _scsih_srch_boot_encl_slot(u64 enclosure_logical_id, u16 slot_number, |
367 | Mpi2BootDeviceEnclosureSlot_t *boot_device) |
368 | { |
369 | return (enclosure_logical_id == le64_to_cpu(boot_device-> |
370 | EnclosureLogicalID) && slot_number == le16_to_cpu(boot_device-> |
371 | SlotNumber)) ? 1 : 0; |
372 | } |
373 | |
374 | /** |
375 | * mpt3sas_get_port_by_id - get hba port entry corresponding to provided |
376 | * port number from port list |
377 | * @ioc: per adapter object |
378 | * @port_id: port number |
379 | * @bypass_dirty_port_flag: when set look the matching hba port entry even |
380 | * if hba port entry is marked as dirty. |
381 | * |
382 | * Search for hba port entry corresponding to provided port number, |
383 | * if available return port object otherwise return NULL. |
384 | */ |
385 | struct hba_port * |
386 | mpt3sas_get_port_by_id(struct MPT3SAS_ADAPTER *ioc, |
387 | u8 port_id, u8 bypass_dirty_port_flag) |
388 | { |
389 | struct hba_port *port, *port_next; |
390 | |
391 | /* |
392 | * When multipath_on_hba is disabled then |
393 | * search the hba_port entry using default |
394 | * port id i.e. 255 |
395 | */ |
396 | if (!ioc->multipath_on_hba) |
397 | port_id = MULTIPATH_DISABLED_PORT_ID; |
398 | |
399 | list_for_each_entry_safe(port, port_next, |
400 | &ioc->port_table_list, list) { |
401 | if (port->port_id != port_id) |
402 | continue; |
403 | if (bypass_dirty_port_flag) |
404 | return port; |
405 | if (port->flags & HBA_PORT_FLAG_DIRTY_PORT) |
406 | continue; |
407 | return port; |
408 | } |
409 | |
410 | /* |
411 | * Allocate hba_port object for default port id (i.e. 255) |
412 | * when multipath_on_hba is disabled for the HBA. |
413 | * And add this object to port_table_list. |
414 | */ |
415 | if (!ioc->multipath_on_hba) { |
416 | port = kzalloc(size: sizeof(struct hba_port), GFP_ATOMIC); |
417 | if (!port) |
418 | return NULL; |
419 | |
420 | port->port_id = port_id; |
421 | ioc_info(ioc, |
422 | "hba_port entry: %p, port: %d is added to hba_port list\n" , |
423 | port, port->port_id); |
424 | list_add_tail(new: &port->list, |
425 | head: &ioc->port_table_list); |
426 | return port; |
427 | } |
428 | return NULL; |
429 | } |
430 | |
431 | /** |
432 | * mpt3sas_get_vphy_by_phy - get virtual_phy object corresponding to phy number |
433 | * @ioc: per adapter object |
434 | * @port: hba_port object |
435 | * @phy: phy number |
436 | * |
437 | * Return virtual_phy object corresponding to phy number. |
438 | */ |
439 | struct virtual_phy * |
440 | mpt3sas_get_vphy_by_phy(struct MPT3SAS_ADAPTER *ioc, |
441 | struct hba_port *port, u32 phy) |
442 | { |
443 | struct virtual_phy *vphy, *vphy_next; |
444 | |
445 | if (!port->vphys_mask) |
446 | return NULL; |
447 | |
448 | list_for_each_entry_safe(vphy, vphy_next, &port->vphys_list, list) { |
449 | if (vphy->phy_mask & (1 << phy)) |
450 | return vphy; |
451 | } |
452 | return NULL; |
453 | } |
454 | |
455 | /** |
456 | * _scsih_is_boot_device - search for matching boot device. |
457 | * @sas_address: sas address |
458 | * @device_name: device name specified in INDENTIFY fram |
459 | * @enclosure_logical_id: enclosure logical id |
460 | * @slot: slot number |
461 | * @form: specifies boot device form |
462 | * @boot_device: boot device object from bios page 2 |
463 | * |
464 | * Return: 1 when there's a match, 0 means no match. |
465 | */ |
466 | static int |
467 | _scsih_is_boot_device(u64 sas_address, u64 device_name, |
468 | u64 enclosure_logical_id, u16 slot, u8 form, |
469 | Mpi2BiosPage2BootDevice_t *boot_device) |
470 | { |
471 | int rc = 0; |
472 | |
473 | switch (form) { |
474 | case MPI2_BIOSPAGE2_FORM_SAS_WWID: |
475 | if (!sas_address) |
476 | break; |
477 | rc = _scsih_srch_boot_sas_address( |
478 | sas_address, boot_device: &boot_device->SasWwid); |
479 | break; |
480 | case MPI2_BIOSPAGE2_FORM_ENCLOSURE_SLOT: |
481 | if (!enclosure_logical_id) |
482 | break; |
483 | rc = _scsih_srch_boot_encl_slot( |
484 | enclosure_logical_id, |
485 | slot_number: slot, boot_device: &boot_device->EnclosureSlot); |
486 | break; |
487 | case MPI2_BIOSPAGE2_FORM_DEVICE_NAME: |
488 | if (!device_name) |
489 | break; |
490 | rc = _scsih_srch_boot_device_name( |
491 | device_name, boot_device: &boot_device->DeviceName); |
492 | break; |
493 | case MPI2_BIOSPAGE2_FORM_NO_DEVICE_SPECIFIED: |
494 | break; |
495 | } |
496 | |
497 | return rc; |
498 | } |
499 | |
500 | /** |
501 | * _scsih_get_sas_address - set the sas_address for given device handle |
502 | * @ioc: ? |
503 | * @handle: device handle |
504 | * @sas_address: sas address |
505 | * |
506 | * Return: 0 success, non-zero when failure |
507 | */ |
508 | static int |
509 | _scsih_get_sas_address(struct MPT3SAS_ADAPTER *ioc, u16 handle, |
510 | u64 *sas_address) |
511 | { |
512 | Mpi2SasDevicePage0_t sas_device_pg0; |
513 | Mpi2ConfigReply_t mpi_reply; |
514 | u32 ioc_status; |
515 | |
516 | *sas_address = 0; |
517 | |
518 | if ((mpt3sas_config_get_sas_device_pg0(ioc, mpi_reply: &mpi_reply, config_page: &sas_device_pg0, |
519 | MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) { |
520 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
521 | __FILE__, __LINE__, __func__); |
522 | return -ENXIO; |
523 | } |
524 | |
525 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; |
526 | if (ioc_status == MPI2_IOCSTATUS_SUCCESS) { |
527 | /* For HBA, vSES doesn't return HBA SAS address. Instead return |
528 | * vSES's sas address. |
529 | */ |
530 | if ((handle <= ioc->sas_hba.num_phys) && |
531 | (!(le32_to_cpu(sas_device_pg0.DeviceInfo) & |
532 | MPI2_SAS_DEVICE_INFO_SEP))) |
533 | *sas_address = ioc->sas_hba.sas_address; |
534 | else |
535 | *sas_address = le64_to_cpu(sas_device_pg0.SASAddress); |
536 | return 0; |
537 | } |
538 | |
539 | /* we hit this because the given parent handle doesn't exist */ |
540 | if (ioc_status == MPI2_IOCSTATUS_CONFIG_INVALID_PAGE) |
541 | return -ENXIO; |
542 | |
543 | /* else error case */ |
544 | ioc_err(ioc, "handle(0x%04x), ioc_status(0x%04x), failure at %s:%d/%s()!\n" , |
545 | handle, ioc_status, __FILE__, __LINE__, __func__); |
546 | return -EIO; |
547 | } |
548 | |
549 | /** |
550 | * _scsih_determine_boot_device - determine boot device. |
551 | * @ioc: per adapter object |
552 | * @device: sas_device or pcie_device object |
553 | * @channel: SAS or PCIe channel |
554 | * |
555 | * Determines whether this device should be first reported device to |
556 | * to scsi-ml or sas transport, this purpose is for persistent boot device. |
557 | * There are primary, alternate, and current entries in bios page 2. The order |
558 | * priority is primary, alternate, then current. This routine saves |
559 | * the corresponding device object. |
560 | * The saved data to be used later in _scsih_probe_boot_devices(). |
561 | */ |
562 | static void |
563 | _scsih_determine_boot_device(struct MPT3SAS_ADAPTER *ioc, void *device, |
564 | u32 channel) |
565 | { |
566 | struct _sas_device *sas_device; |
567 | struct _pcie_device *pcie_device; |
568 | struct _raid_device *raid_device; |
569 | u64 sas_address; |
570 | u64 device_name; |
571 | u64 enclosure_logical_id; |
572 | u16 slot; |
573 | |
574 | /* only process this function when driver loads */ |
575 | if (!ioc->is_driver_loading) |
576 | return; |
577 | |
578 | /* no Bios, return immediately */ |
579 | if (!ioc->bios_pg3.BiosVersion) |
580 | return; |
581 | |
582 | if (channel == RAID_CHANNEL) { |
583 | raid_device = device; |
584 | sas_address = raid_device->wwid; |
585 | device_name = 0; |
586 | enclosure_logical_id = 0; |
587 | slot = 0; |
588 | } else if (channel == PCIE_CHANNEL) { |
589 | pcie_device = device; |
590 | sas_address = pcie_device->wwid; |
591 | device_name = 0; |
592 | enclosure_logical_id = 0; |
593 | slot = 0; |
594 | } else { |
595 | sas_device = device; |
596 | sas_address = sas_device->sas_address; |
597 | device_name = sas_device->device_name; |
598 | enclosure_logical_id = sas_device->enclosure_logical_id; |
599 | slot = sas_device->slot; |
600 | } |
601 | |
602 | if (!ioc->req_boot_device.device) { |
603 | if (_scsih_is_boot_device(sas_address, device_name, |
604 | enclosure_logical_id, slot, |
605 | form: (ioc->bios_pg2.ReqBootDeviceForm & |
606 | MPI2_BIOSPAGE2_FORM_MASK), |
607 | boot_device: &ioc->bios_pg2.RequestedBootDevice)) { |
608 | dinitprintk(ioc, |
609 | ioc_info(ioc, "%s: req_boot_device(0x%016llx)\n" , |
610 | __func__, (u64)sas_address)); |
611 | ioc->req_boot_device.device = device; |
612 | ioc->req_boot_device.channel = channel; |
613 | } |
614 | } |
615 | |
616 | if (!ioc->req_alt_boot_device.device) { |
617 | if (_scsih_is_boot_device(sas_address, device_name, |
618 | enclosure_logical_id, slot, |
619 | form: (ioc->bios_pg2.ReqAltBootDeviceForm & |
620 | MPI2_BIOSPAGE2_FORM_MASK), |
621 | boot_device: &ioc->bios_pg2.RequestedAltBootDevice)) { |
622 | dinitprintk(ioc, |
623 | ioc_info(ioc, "%s: req_alt_boot_device(0x%016llx)\n" , |
624 | __func__, (u64)sas_address)); |
625 | ioc->req_alt_boot_device.device = device; |
626 | ioc->req_alt_boot_device.channel = channel; |
627 | } |
628 | } |
629 | |
630 | if (!ioc->current_boot_device.device) { |
631 | if (_scsih_is_boot_device(sas_address, device_name, |
632 | enclosure_logical_id, slot, |
633 | form: (ioc->bios_pg2.CurrentBootDeviceForm & |
634 | MPI2_BIOSPAGE2_FORM_MASK), |
635 | boot_device: &ioc->bios_pg2.CurrentBootDevice)) { |
636 | dinitprintk(ioc, |
637 | ioc_info(ioc, "%s: current_boot_device(0x%016llx)\n" , |
638 | __func__, (u64)sas_address)); |
639 | ioc->current_boot_device.device = device; |
640 | ioc->current_boot_device.channel = channel; |
641 | } |
642 | } |
643 | } |
644 | |
645 | static struct _sas_device * |
646 | __mpt3sas_get_sdev_from_target(struct MPT3SAS_ADAPTER *ioc, |
647 | struct MPT3SAS_TARGET *tgt_priv) |
648 | { |
649 | struct _sas_device *ret; |
650 | |
651 | assert_spin_locked(&ioc->sas_device_lock); |
652 | |
653 | ret = tgt_priv->sas_dev; |
654 | if (ret) |
655 | sas_device_get(s: ret); |
656 | |
657 | return ret; |
658 | } |
659 | |
660 | static struct _sas_device * |
661 | mpt3sas_get_sdev_from_target(struct MPT3SAS_ADAPTER *ioc, |
662 | struct MPT3SAS_TARGET *tgt_priv) |
663 | { |
664 | struct _sas_device *ret; |
665 | unsigned long flags; |
666 | |
667 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
668 | ret = __mpt3sas_get_sdev_from_target(ioc, tgt_priv); |
669 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
670 | |
671 | return ret; |
672 | } |
673 | |
674 | static struct _pcie_device * |
675 | __mpt3sas_get_pdev_from_target(struct MPT3SAS_ADAPTER *ioc, |
676 | struct MPT3SAS_TARGET *tgt_priv) |
677 | { |
678 | struct _pcie_device *ret; |
679 | |
680 | assert_spin_locked(&ioc->pcie_device_lock); |
681 | |
682 | ret = tgt_priv->pcie_dev; |
683 | if (ret) |
684 | pcie_device_get(p: ret); |
685 | |
686 | return ret; |
687 | } |
688 | |
689 | /** |
690 | * mpt3sas_get_pdev_from_target - pcie device search |
691 | * @ioc: per adapter object |
692 | * @tgt_priv: starget private object |
693 | * |
694 | * Context: This function will acquire ioc->pcie_device_lock and will release |
695 | * before returning the pcie_device object. |
696 | * |
697 | * This searches for pcie_device from target, then return pcie_device object. |
698 | */ |
699 | static struct _pcie_device * |
700 | mpt3sas_get_pdev_from_target(struct MPT3SAS_ADAPTER *ioc, |
701 | struct MPT3SAS_TARGET *tgt_priv) |
702 | { |
703 | struct _pcie_device *ret; |
704 | unsigned long flags; |
705 | |
706 | spin_lock_irqsave(&ioc->pcie_device_lock, flags); |
707 | ret = __mpt3sas_get_pdev_from_target(ioc, tgt_priv); |
708 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
709 | |
710 | return ret; |
711 | } |
712 | |
713 | |
714 | /** |
715 | * __mpt3sas_get_sdev_by_rphy - sas device search |
716 | * @ioc: per adapter object |
717 | * @rphy: sas_rphy pointer |
718 | * |
719 | * Context: This function will acquire ioc->sas_device_lock and will release |
720 | * before returning the sas_device object. |
721 | * |
722 | * This searches for sas_device from rphy object |
723 | * then return sas_device object. |
724 | */ |
725 | struct _sas_device * |
726 | __mpt3sas_get_sdev_by_rphy(struct MPT3SAS_ADAPTER *ioc, |
727 | struct sas_rphy *rphy) |
728 | { |
729 | struct _sas_device *sas_device; |
730 | |
731 | assert_spin_locked(&ioc->sas_device_lock); |
732 | |
733 | list_for_each_entry(sas_device, &ioc->sas_device_list, list) { |
734 | if (sas_device->rphy != rphy) |
735 | continue; |
736 | sas_device_get(s: sas_device); |
737 | return sas_device; |
738 | } |
739 | |
740 | sas_device = NULL; |
741 | list_for_each_entry(sas_device, &ioc->sas_device_init_list, list) { |
742 | if (sas_device->rphy != rphy) |
743 | continue; |
744 | sas_device_get(s: sas_device); |
745 | return sas_device; |
746 | } |
747 | |
748 | return NULL; |
749 | } |
750 | |
751 | /** |
752 | * __mpt3sas_get_sdev_by_addr - get _sas_device object corresponding to provided |
753 | * sas address from sas_device_list list |
754 | * @ioc: per adapter object |
755 | * @sas_address: device sas address |
756 | * @port: port number |
757 | * |
758 | * Search for _sas_device object corresponding to provided sas address, |
759 | * if available return _sas_device object address otherwise return NULL. |
760 | */ |
761 | struct _sas_device * |
762 | __mpt3sas_get_sdev_by_addr(struct MPT3SAS_ADAPTER *ioc, |
763 | u64 sas_address, struct hba_port *port) |
764 | { |
765 | struct _sas_device *sas_device; |
766 | |
767 | if (!port) |
768 | return NULL; |
769 | |
770 | assert_spin_locked(&ioc->sas_device_lock); |
771 | |
772 | list_for_each_entry(sas_device, &ioc->sas_device_list, list) { |
773 | if (sas_device->sas_address != sas_address) |
774 | continue; |
775 | if (sas_device->port != port) |
776 | continue; |
777 | sas_device_get(s: sas_device); |
778 | return sas_device; |
779 | } |
780 | |
781 | list_for_each_entry(sas_device, &ioc->sas_device_init_list, list) { |
782 | if (sas_device->sas_address != sas_address) |
783 | continue; |
784 | if (sas_device->port != port) |
785 | continue; |
786 | sas_device_get(s: sas_device); |
787 | return sas_device; |
788 | } |
789 | |
790 | return NULL; |
791 | } |
792 | |
793 | /** |
794 | * mpt3sas_get_sdev_by_addr - sas device search |
795 | * @ioc: per adapter object |
796 | * @sas_address: sas address |
797 | * @port: hba port entry |
798 | * Context: Calling function should acquire ioc->sas_device_lock |
799 | * |
800 | * This searches for sas_device based on sas_address & port number, |
801 | * then return sas_device object. |
802 | */ |
803 | struct _sas_device * |
804 | mpt3sas_get_sdev_by_addr(struct MPT3SAS_ADAPTER *ioc, |
805 | u64 sas_address, struct hba_port *port) |
806 | { |
807 | struct _sas_device *sas_device; |
808 | unsigned long flags; |
809 | |
810 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
811 | sas_device = __mpt3sas_get_sdev_by_addr(ioc, |
812 | sas_address, port); |
813 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
814 | |
815 | return sas_device; |
816 | } |
817 | |
818 | static struct _sas_device * |
819 | __mpt3sas_get_sdev_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) |
820 | { |
821 | struct _sas_device *sas_device; |
822 | |
823 | assert_spin_locked(&ioc->sas_device_lock); |
824 | |
825 | list_for_each_entry(sas_device, &ioc->sas_device_list, list) |
826 | if (sas_device->handle == handle) |
827 | goto found_device; |
828 | |
829 | list_for_each_entry(sas_device, &ioc->sas_device_init_list, list) |
830 | if (sas_device->handle == handle) |
831 | goto found_device; |
832 | |
833 | return NULL; |
834 | |
835 | found_device: |
836 | sas_device_get(s: sas_device); |
837 | return sas_device; |
838 | } |
839 | |
840 | /** |
841 | * mpt3sas_get_sdev_by_handle - sas device search |
842 | * @ioc: per adapter object |
843 | * @handle: sas device handle (assigned by firmware) |
844 | * Context: Calling function should acquire ioc->sas_device_lock |
845 | * |
846 | * This searches for sas_device based on sas_address, then return sas_device |
847 | * object. |
848 | */ |
849 | struct _sas_device * |
850 | mpt3sas_get_sdev_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) |
851 | { |
852 | struct _sas_device *sas_device; |
853 | unsigned long flags; |
854 | |
855 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
856 | sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle); |
857 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
858 | |
859 | return sas_device; |
860 | } |
861 | |
862 | /** |
863 | * _scsih_display_enclosure_chassis_info - display device location info |
864 | * @ioc: per adapter object |
865 | * @sas_device: per sas device object |
866 | * @sdev: scsi device struct |
867 | * @starget: scsi target struct |
868 | */ |
869 | static void |
870 | _scsih_display_enclosure_chassis_info(struct MPT3SAS_ADAPTER *ioc, |
871 | struct _sas_device *sas_device, struct scsi_device *sdev, |
872 | struct scsi_target *starget) |
873 | { |
874 | if (sdev) { |
875 | if (sas_device->enclosure_handle != 0) |
876 | sdev_printk(KERN_INFO, sdev, |
877 | "enclosure logical id (0x%016llx), slot(%d) \n" , |
878 | (unsigned long long) |
879 | sas_device->enclosure_logical_id, |
880 | sas_device->slot); |
881 | if (sas_device->connector_name[0] != '\0') |
882 | sdev_printk(KERN_INFO, sdev, |
883 | "enclosure level(0x%04x), connector name( %s)\n" , |
884 | sas_device->enclosure_level, |
885 | sas_device->connector_name); |
886 | if (sas_device->is_chassis_slot_valid) |
887 | sdev_printk(KERN_INFO, sdev, "chassis slot(0x%04x)\n" , |
888 | sas_device->chassis_slot); |
889 | } else if (starget) { |
890 | if (sas_device->enclosure_handle != 0) |
891 | starget_printk(KERN_INFO, starget, |
892 | "enclosure logical id(0x%016llx), slot(%d) \n" , |
893 | (unsigned long long) |
894 | sas_device->enclosure_logical_id, |
895 | sas_device->slot); |
896 | if (sas_device->connector_name[0] != '\0') |
897 | starget_printk(KERN_INFO, starget, |
898 | "enclosure level(0x%04x), connector name( %s)\n" , |
899 | sas_device->enclosure_level, |
900 | sas_device->connector_name); |
901 | if (sas_device->is_chassis_slot_valid) |
902 | starget_printk(KERN_INFO, starget, |
903 | "chassis slot(0x%04x)\n" , |
904 | sas_device->chassis_slot); |
905 | } else { |
906 | if (sas_device->enclosure_handle != 0) |
907 | ioc_info(ioc, "enclosure logical id(0x%016llx), slot(%d)\n" , |
908 | (u64)sas_device->enclosure_logical_id, |
909 | sas_device->slot); |
910 | if (sas_device->connector_name[0] != '\0') |
911 | ioc_info(ioc, "enclosure level(0x%04x), connector name( %s)\n" , |
912 | sas_device->enclosure_level, |
913 | sas_device->connector_name); |
914 | if (sas_device->is_chassis_slot_valid) |
915 | ioc_info(ioc, "chassis slot(0x%04x)\n" , |
916 | sas_device->chassis_slot); |
917 | } |
918 | } |
919 | |
920 | /** |
921 | * _scsih_sas_device_remove - remove sas_device from list. |
922 | * @ioc: per adapter object |
923 | * @sas_device: the sas_device object |
924 | * Context: This function will acquire ioc->sas_device_lock. |
925 | * |
926 | * If sas_device is on the list, remove it and decrement its reference count. |
927 | */ |
928 | static void |
929 | _scsih_sas_device_remove(struct MPT3SAS_ADAPTER *ioc, |
930 | struct _sas_device *sas_device) |
931 | { |
932 | unsigned long flags; |
933 | |
934 | if (!sas_device) |
935 | return; |
936 | ioc_info(ioc, "removing handle(0x%04x), sas_addr(0x%016llx)\n" , |
937 | sas_device->handle, (u64)sas_device->sas_address); |
938 | |
939 | _scsih_display_enclosure_chassis_info(ioc, sas_device, NULL, NULL); |
940 | |
941 | /* |
942 | * The lock serializes access to the list, but we still need to verify |
943 | * that nobody removed the entry while we were waiting on the lock. |
944 | */ |
945 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
946 | if (!list_empty(head: &sas_device->list)) { |
947 | list_del_init(entry: &sas_device->list); |
948 | sas_device_put(s: sas_device); |
949 | } |
950 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
951 | } |
952 | |
953 | /** |
954 | * _scsih_device_remove_by_handle - removing device object by handle |
955 | * @ioc: per adapter object |
956 | * @handle: device handle |
957 | */ |
958 | static void |
959 | _scsih_device_remove_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) |
960 | { |
961 | struct _sas_device *sas_device; |
962 | unsigned long flags; |
963 | |
964 | if (ioc->shost_recovery) |
965 | return; |
966 | |
967 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
968 | sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle); |
969 | if (sas_device) { |
970 | list_del_init(entry: &sas_device->list); |
971 | sas_device_put(s: sas_device); |
972 | } |
973 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
974 | if (sas_device) { |
975 | _scsih_remove_device(ioc, sas_device); |
976 | sas_device_put(s: sas_device); |
977 | } |
978 | } |
979 | |
980 | /** |
981 | * mpt3sas_device_remove_by_sas_address - removing device object by |
982 | * sas address & port number |
983 | * @ioc: per adapter object |
984 | * @sas_address: device sas_address |
985 | * @port: hba port entry |
986 | * |
987 | * Return nothing. |
988 | */ |
989 | void |
990 | mpt3sas_device_remove_by_sas_address(struct MPT3SAS_ADAPTER *ioc, |
991 | u64 sas_address, struct hba_port *port) |
992 | { |
993 | struct _sas_device *sas_device; |
994 | unsigned long flags; |
995 | |
996 | if (ioc->shost_recovery) |
997 | return; |
998 | |
999 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
1000 | sas_device = __mpt3sas_get_sdev_by_addr(ioc, sas_address, port); |
1001 | if (sas_device) { |
1002 | list_del_init(entry: &sas_device->list); |
1003 | sas_device_put(s: sas_device); |
1004 | } |
1005 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
1006 | if (sas_device) { |
1007 | _scsih_remove_device(ioc, sas_device); |
1008 | sas_device_put(s: sas_device); |
1009 | } |
1010 | } |
1011 | |
1012 | /** |
1013 | * _scsih_sas_device_add - insert sas_device to the list. |
1014 | * @ioc: per adapter object |
1015 | * @sas_device: the sas_device object |
1016 | * Context: This function will acquire ioc->sas_device_lock. |
1017 | * |
1018 | * Adding new object to the ioc->sas_device_list. |
1019 | */ |
1020 | static void |
1021 | _scsih_sas_device_add(struct MPT3SAS_ADAPTER *ioc, |
1022 | struct _sas_device *sas_device) |
1023 | { |
1024 | unsigned long flags; |
1025 | |
1026 | dewtprintk(ioc, |
1027 | ioc_info(ioc, "%s: handle(0x%04x), sas_addr(0x%016llx)\n" , |
1028 | __func__, sas_device->handle, |
1029 | (u64)sas_device->sas_address)); |
1030 | |
1031 | dewtprintk(ioc, _scsih_display_enclosure_chassis_info(ioc, sas_device, |
1032 | NULL, NULL)); |
1033 | |
1034 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
1035 | sas_device_get(s: sas_device); |
1036 | list_add_tail(new: &sas_device->list, head: &ioc->sas_device_list); |
1037 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
1038 | |
1039 | if (ioc->hide_drives) { |
1040 | clear_bit(nr: sas_device->handle, addr: ioc->pend_os_device_add); |
1041 | return; |
1042 | } |
1043 | |
1044 | if (!mpt3sas_transport_port_add(ioc, handle: sas_device->handle, |
1045 | sas_address: sas_device->sas_address_parent, port: sas_device->port)) { |
1046 | _scsih_sas_device_remove(ioc, sas_device); |
1047 | } else if (!sas_device->starget) { |
1048 | /* |
1049 | * When asyn scanning is enabled, its not possible to remove |
1050 | * devices while scanning is turned on due to an oops in |
1051 | * scsi_sysfs_add_sdev()->add_device()->sysfs_addrm_start() |
1052 | */ |
1053 | if (!ioc->is_driver_loading) { |
1054 | mpt3sas_transport_port_remove(ioc, |
1055 | sas_address: sas_device->sas_address, |
1056 | sas_address_parent: sas_device->sas_address_parent, |
1057 | port: sas_device->port); |
1058 | _scsih_sas_device_remove(ioc, sas_device); |
1059 | } |
1060 | } else |
1061 | clear_bit(nr: sas_device->handle, addr: ioc->pend_os_device_add); |
1062 | } |
1063 | |
1064 | /** |
1065 | * _scsih_sas_device_init_add - insert sas_device to the list. |
1066 | * @ioc: per adapter object |
1067 | * @sas_device: the sas_device object |
1068 | * Context: This function will acquire ioc->sas_device_lock. |
1069 | * |
1070 | * Adding new object at driver load time to the ioc->sas_device_init_list. |
1071 | */ |
1072 | static void |
1073 | _scsih_sas_device_init_add(struct MPT3SAS_ADAPTER *ioc, |
1074 | struct _sas_device *sas_device) |
1075 | { |
1076 | unsigned long flags; |
1077 | |
1078 | dewtprintk(ioc, |
1079 | ioc_info(ioc, "%s: handle(0x%04x), sas_addr(0x%016llx)\n" , |
1080 | __func__, sas_device->handle, |
1081 | (u64)sas_device->sas_address)); |
1082 | |
1083 | dewtprintk(ioc, _scsih_display_enclosure_chassis_info(ioc, sas_device, |
1084 | NULL, NULL)); |
1085 | |
1086 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
1087 | sas_device_get(s: sas_device); |
1088 | list_add_tail(new: &sas_device->list, head: &ioc->sas_device_init_list); |
1089 | _scsih_determine_boot_device(ioc, device: sas_device, channel: 0); |
1090 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
1091 | } |
1092 | |
1093 | |
1094 | static struct _pcie_device * |
1095 | __mpt3sas_get_pdev_by_wwid(struct MPT3SAS_ADAPTER *ioc, u64 wwid) |
1096 | { |
1097 | struct _pcie_device *pcie_device; |
1098 | |
1099 | assert_spin_locked(&ioc->pcie_device_lock); |
1100 | |
1101 | list_for_each_entry(pcie_device, &ioc->pcie_device_list, list) |
1102 | if (pcie_device->wwid == wwid) |
1103 | goto found_device; |
1104 | |
1105 | list_for_each_entry(pcie_device, &ioc->pcie_device_init_list, list) |
1106 | if (pcie_device->wwid == wwid) |
1107 | goto found_device; |
1108 | |
1109 | return NULL; |
1110 | |
1111 | found_device: |
1112 | pcie_device_get(p: pcie_device); |
1113 | return pcie_device; |
1114 | } |
1115 | |
1116 | |
1117 | /** |
1118 | * mpt3sas_get_pdev_by_wwid - pcie device search |
1119 | * @ioc: per adapter object |
1120 | * @wwid: wwid |
1121 | * |
1122 | * Context: This function will acquire ioc->pcie_device_lock and will release |
1123 | * before returning the pcie_device object. |
1124 | * |
1125 | * This searches for pcie_device based on wwid, then return pcie_device object. |
1126 | */ |
1127 | static struct _pcie_device * |
1128 | mpt3sas_get_pdev_by_wwid(struct MPT3SAS_ADAPTER *ioc, u64 wwid) |
1129 | { |
1130 | struct _pcie_device *pcie_device; |
1131 | unsigned long flags; |
1132 | |
1133 | spin_lock_irqsave(&ioc->pcie_device_lock, flags); |
1134 | pcie_device = __mpt3sas_get_pdev_by_wwid(ioc, wwid); |
1135 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
1136 | |
1137 | return pcie_device; |
1138 | } |
1139 | |
1140 | |
1141 | static struct _pcie_device * |
1142 | __mpt3sas_get_pdev_by_idchannel(struct MPT3SAS_ADAPTER *ioc, int id, |
1143 | int channel) |
1144 | { |
1145 | struct _pcie_device *pcie_device; |
1146 | |
1147 | assert_spin_locked(&ioc->pcie_device_lock); |
1148 | |
1149 | list_for_each_entry(pcie_device, &ioc->pcie_device_list, list) |
1150 | if (pcie_device->id == id && pcie_device->channel == channel) |
1151 | goto found_device; |
1152 | |
1153 | list_for_each_entry(pcie_device, &ioc->pcie_device_init_list, list) |
1154 | if (pcie_device->id == id && pcie_device->channel == channel) |
1155 | goto found_device; |
1156 | |
1157 | return NULL; |
1158 | |
1159 | found_device: |
1160 | pcie_device_get(p: pcie_device); |
1161 | return pcie_device; |
1162 | } |
1163 | |
1164 | static struct _pcie_device * |
1165 | __mpt3sas_get_pdev_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) |
1166 | { |
1167 | struct _pcie_device *pcie_device; |
1168 | |
1169 | assert_spin_locked(&ioc->pcie_device_lock); |
1170 | |
1171 | list_for_each_entry(pcie_device, &ioc->pcie_device_list, list) |
1172 | if (pcie_device->handle == handle) |
1173 | goto found_device; |
1174 | |
1175 | list_for_each_entry(pcie_device, &ioc->pcie_device_init_list, list) |
1176 | if (pcie_device->handle == handle) |
1177 | goto found_device; |
1178 | |
1179 | return NULL; |
1180 | |
1181 | found_device: |
1182 | pcie_device_get(p: pcie_device); |
1183 | return pcie_device; |
1184 | } |
1185 | |
1186 | |
1187 | /** |
1188 | * mpt3sas_get_pdev_by_handle - pcie device search |
1189 | * @ioc: per adapter object |
1190 | * @handle: Firmware device handle |
1191 | * |
1192 | * Context: This function will acquire ioc->pcie_device_lock and will release |
1193 | * before returning the pcie_device object. |
1194 | * |
1195 | * This searches for pcie_device based on handle, then return pcie_device |
1196 | * object. |
1197 | */ |
1198 | struct _pcie_device * |
1199 | mpt3sas_get_pdev_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) |
1200 | { |
1201 | struct _pcie_device *pcie_device; |
1202 | unsigned long flags; |
1203 | |
1204 | spin_lock_irqsave(&ioc->pcie_device_lock, flags); |
1205 | pcie_device = __mpt3sas_get_pdev_by_handle(ioc, handle); |
1206 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
1207 | |
1208 | return pcie_device; |
1209 | } |
1210 | |
1211 | /** |
1212 | * _scsih_set_nvme_max_shutdown_latency - Update max_shutdown_latency. |
1213 | * @ioc: per adapter object |
1214 | * Context: This function will acquire ioc->pcie_device_lock |
1215 | * |
1216 | * Update ioc->max_shutdown_latency to that NVMe drives RTD3 Entry Latency |
1217 | * which has reported maximum among all available NVMe drives. |
1218 | * Minimum max_shutdown_latency will be six seconds. |
1219 | */ |
1220 | static void |
1221 | _scsih_set_nvme_max_shutdown_latency(struct MPT3SAS_ADAPTER *ioc) |
1222 | { |
1223 | struct _pcie_device *pcie_device; |
1224 | unsigned long flags; |
1225 | u16 shutdown_latency = IO_UNIT_CONTROL_SHUTDOWN_TIMEOUT; |
1226 | |
1227 | spin_lock_irqsave(&ioc->pcie_device_lock, flags); |
1228 | list_for_each_entry(pcie_device, &ioc->pcie_device_list, list) { |
1229 | if (pcie_device->shutdown_latency) { |
1230 | if (shutdown_latency < pcie_device->shutdown_latency) |
1231 | shutdown_latency = |
1232 | pcie_device->shutdown_latency; |
1233 | } |
1234 | } |
1235 | ioc->max_shutdown_latency = shutdown_latency; |
1236 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
1237 | } |
1238 | |
1239 | /** |
1240 | * _scsih_pcie_device_remove - remove pcie_device from list. |
1241 | * @ioc: per adapter object |
1242 | * @pcie_device: the pcie_device object |
1243 | * Context: This function will acquire ioc->pcie_device_lock. |
1244 | * |
1245 | * If pcie_device is on the list, remove it and decrement its reference count. |
1246 | */ |
1247 | static void |
1248 | _scsih_pcie_device_remove(struct MPT3SAS_ADAPTER *ioc, |
1249 | struct _pcie_device *pcie_device) |
1250 | { |
1251 | unsigned long flags; |
1252 | int was_on_pcie_device_list = 0; |
1253 | u8 update_latency = 0; |
1254 | |
1255 | if (!pcie_device) |
1256 | return; |
1257 | ioc_info(ioc, "removing handle(0x%04x), wwid(0x%016llx)\n" , |
1258 | pcie_device->handle, (u64)pcie_device->wwid); |
1259 | if (pcie_device->enclosure_handle != 0) |
1260 | ioc_info(ioc, "removing enclosure logical id(0x%016llx), slot(%d)\n" , |
1261 | (u64)pcie_device->enclosure_logical_id, |
1262 | pcie_device->slot); |
1263 | if (pcie_device->connector_name[0] != '\0') |
1264 | ioc_info(ioc, "removing enclosure level(0x%04x), connector name( %s)\n" , |
1265 | pcie_device->enclosure_level, |
1266 | pcie_device->connector_name); |
1267 | |
1268 | spin_lock_irqsave(&ioc->pcie_device_lock, flags); |
1269 | if (!list_empty(head: &pcie_device->list)) { |
1270 | list_del_init(entry: &pcie_device->list); |
1271 | was_on_pcie_device_list = 1; |
1272 | } |
1273 | if (pcie_device->shutdown_latency == ioc->max_shutdown_latency) |
1274 | update_latency = 1; |
1275 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
1276 | if (was_on_pcie_device_list) { |
1277 | kfree(objp: pcie_device->serial_number); |
1278 | pcie_device_put(p: pcie_device); |
1279 | } |
1280 | |
1281 | /* |
1282 | * This device's RTD3 Entry Latency matches IOC's |
1283 | * max_shutdown_latency. Recalculate IOC's max_shutdown_latency |
1284 | * from the available drives as current drive is getting removed. |
1285 | */ |
1286 | if (update_latency) |
1287 | _scsih_set_nvme_max_shutdown_latency(ioc); |
1288 | } |
1289 | |
1290 | |
1291 | /** |
1292 | * _scsih_pcie_device_remove_by_handle - removing pcie device object by handle |
1293 | * @ioc: per adapter object |
1294 | * @handle: device handle |
1295 | */ |
1296 | static void |
1297 | _scsih_pcie_device_remove_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) |
1298 | { |
1299 | struct _pcie_device *pcie_device; |
1300 | unsigned long flags; |
1301 | int was_on_pcie_device_list = 0; |
1302 | u8 update_latency = 0; |
1303 | |
1304 | if (ioc->shost_recovery) |
1305 | return; |
1306 | |
1307 | spin_lock_irqsave(&ioc->pcie_device_lock, flags); |
1308 | pcie_device = __mpt3sas_get_pdev_by_handle(ioc, handle); |
1309 | if (pcie_device) { |
1310 | if (!list_empty(head: &pcie_device->list)) { |
1311 | list_del_init(entry: &pcie_device->list); |
1312 | was_on_pcie_device_list = 1; |
1313 | pcie_device_put(p: pcie_device); |
1314 | } |
1315 | if (pcie_device->shutdown_latency == ioc->max_shutdown_latency) |
1316 | update_latency = 1; |
1317 | } |
1318 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
1319 | if (was_on_pcie_device_list) { |
1320 | _scsih_pcie_device_remove_from_sml(ioc, pcie_device); |
1321 | pcie_device_put(p: pcie_device); |
1322 | } |
1323 | |
1324 | /* |
1325 | * This device's RTD3 Entry Latency matches IOC's |
1326 | * max_shutdown_latency. Recalculate IOC's max_shutdown_latency |
1327 | * from the available drives as current drive is getting removed. |
1328 | */ |
1329 | if (update_latency) |
1330 | _scsih_set_nvme_max_shutdown_latency(ioc); |
1331 | } |
1332 | |
1333 | /** |
1334 | * _scsih_pcie_device_add - add pcie_device object |
1335 | * @ioc: per adapter object |
1336 | * @pcie_device: pcie_device object |
1337 | * |
1338 | * This is added to the pcie_device_list link list. |
1339 | */ |
1340 | static void |
1341 | _scsih_pcie_device_add(struct MPT3SAS_ADAPTER *ioc, |
1342 | struct _pcie_device *pcie_device) |
1343 | { |
1344 | unsigned long flags; |
1345 | |
1346 | dewtprintk(ioc, |
1347 | ioc_info(ioc, "%s: handle (0x%04x), wwid(0x%016llx)\n" , |
1348 | __func__, |
1349 | pcie_device->handle, (u64)pcie_device->wwid)); |
1350 | if (pcie_device->enclosure_handle != 0) |
1351 | dewtprintk(ioc, |
1352 | ioc_info(ioc, "%s: enclosure logical id(0x%016llx), slot( %d)\n" , |
1353 | __func__, |
1354 | (u64)pcie_device->enclosure_logical_id, |
1355 | pcie_device->slot)); |
1356 | if (pcie_device->connector_name[0] != '\0') |
1357 | dewtprintk(ioc, |
1358 | ioc_info(ioc, "%s: enclosure level(0x%04x), connector name( %s)\n" , |
1359 | __func__, pcie_device->enclosure_level, |
1360 | pcie_device->connector_name)); |
1361 | |
1362 | spin_lock_irqsave(&ioc->pcie_device_lock, flags); |
1363 | pcie_device_get(p: pcie_device); |
1364 | list_add_tail(new: &pcie_device->list, head: &ioc->pcie_device_list); |
1365 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
1366 | |
1367 | if (pcie_device->access_status == |
1368 | MPI26_PCIEDEV0_ASTATUS_DEVICE_BLOCKED) { |
1369 | clear_bit(nr: pcie_device->handle, addr: ioc->pend_os_device_add); |
1370 | return; |
1371 | } |
1372 | if (scsi_add_device(host: ioc->shost, PCIE_CHANNEL, target: pcie_device->id, lun: 0)) { |
1373 | _scsih_pcie_device_remove(ioc, pcie_device); |
1374 | } else if (!pcie_device->starget) { |
1375 | if (!ioc->is_driver_loading) { |
1376 | /*TODO-- Need to find out whether this condition will occur or not*/ |
1377 | clear_bit(nr: pcie_device->handle, addr: ioc->pend_os_device_add); |
1378 | } |
1379 | } else |
1380 | clear_bit(nr: pcie_device->handle, addr: ioc->pend_os_device_add); |
1381 | } |
1382 | |
1383 | /* |
1384 | * _scsih_pcie_device_init_add - insert pcie_device to the init list. |
1385 | * @ioc: per adapter object |
1386 | * @pcie_device: the pcie_device object |
1387 | * Context: This function will acquire ioc->pcie_device_lock. |
1388 | * |
1389 | * Adding new object at driver load time to the ioc->pcie_device_init_list. |
1390 | */ |
1391 | static void |
1392 | _scsih_pcie_device_init_add(struct MPT3SAS_ADAPTER *ioc, |
1393 | struct _pcie_device *pcie_device) |
1394 | { |
1395 | unsigned long flags; |
1396 | |
1397 | dewtprintk(ioc, |
1398 | ioc_info(ioc, "%s: handle (0x%04x), wwid(0x%016llx)\n" , |
1399 | __func__, |
1400 | pcie_device->handle, (u64)pcie_device->wwid)); |
1401 | if (pcie_device->enclosure_handle != 0) |
1402 | dewtprintk(ioc, |
1403 | ioc_info(ioc, "%s: enclosure logical id(0x%016llx), slot( %d)\n" , |
1404 | __func__, |
1405 | (u64)pcie_device->enclosure_logical_id, |
1406 | pcie_device->slot)); |
1407 | if (pcie_device->connector_name[0] != '\0') |
1408 | dewtprintk(ioc, |
1409 | ioc_info(ioc, "%s: enclosure level(0x%04x), connector name( %s)\n" , |
1410 | __func__, pcie_device->enclosure_level, |
1411 | pcie_device->connector_name)); |
1412 | |
1413 | spin_lock_irqsave(&ioc->pcie_device_lock, flags); |
1414 | pcie_device_get(p: pcie_device); |
1415 | list_add_tail(new: &pcie_device->list, head: &ioc->pcie_device_init_list); |
1416 | if (pcie_device->access_status != |
1417 | MPI26_PCIEDEV0_ASTATUS_DEVICE_BLOCKED) |
1418 | _scsih_determine_boot_device(ioc, device: pcie_device, PCIE_CHANNEL); |
1419 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
1420 | } |
1421 | /** |
1422 | * _scsih_raid_device_find_by_id - raid device search |
1423 | * @ioc: per adapter object |
1424 | * @id: sas device target id |
1425 | * @channel: sas device channel |
1426 | * Context: Calling function should acquire ioc->raid_device_lock |
1427 | * |
1428 | * This searches for raid_device based on target id, then return raid_device |
1429 | * object. |
1430 | */ |
1431 | static struct _raid_device * |
1432 | _scsih_raid_device_find_by_id(struct MPT3SAS_ADAPTER *ioc, int id, int channel) |
1433 | { |
1434 | struct _raid_device *raid_device, *r; |
1435 | |
1436 | r = NULL; |
1437 | list_for_each_entry(raid_device, &ioc->raid_device_list, list) { |
1438 | if (raid_device->id == id && raid_device->channel == channel) { |
1439 | r = raid_device; |
1440 | goto out; |
1441 | } |
1442 | } |
1443 | |
1444 | out: |
1445 | return r; |
1446 | } |
1447 | |
1448 | /** |
1449 | * mpt3sas_raid_device_find_by_handle - raid device search |
1450 | * @ioc: per adapter object |
1451 | * @handle: sas device handle (assigned by firmware) |
1452 | * Context: Calling function should acquire ioc->raid_device_lock |
1453 | * |
1454 | * This searches for raid_device based on handle, then return raid_device |
1455 | * object. |
1456 | */ |
1457 | struct _raid_device * |
1458 | mpt3sas_raid_device_find_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) |
1459 | { |
1460 | struct _raid_device *raid_device, *r; |
1461 | |
1462 | r = NULL; |
1463 | list_for_each_entry(raid_device, &ioc->raid_device_list, list) { |
1464 | if (raid_device->handle != handle) |
1465 | continue; |
1466 | r = raid_device; |
1467 | goto out; |
1468 | } |
1469 | |
1470 | out: |
1471 | return r; |
1472 | } |
1473 | |
1474 | /** |
1475 | * _scsih_raid_device_find_by_wwid - raid device search |
1476 | * @ioc: per adapter object |
1477 | * @wwid: ? |
1478 | * Context: Calling function should acquire ioc->raid_device_lock |
1479 | * |
1480 | * This searches for raid_device based on wwid, then return raid_device |
1481 | * object. |
1482 | */ |
1483 | static struct _raid_device * |
1484 | _scsih_raid_device_find_by_wwid(struct MPT3SAS_ADAPTER *ioc, u64 wwid) |
1485 | { |
1486 | struct _raid_device *raid_device, *r; |
1487 | |
1488 | r = NULL; |
1489 | list_for_each_entry(raid_device, &ioc->raid_device_list, list) { |
1490 | if (raid_device->wwid != wwid) |
1491 | continue; |
1492 | r = raid_device; |
1493 | goto out; |
1494 | } |
1495 | |
1496 | out: |
1497 | return r; |
1498 | } |
1499 | |
1500 | /** |
1501 | * _scsih_raid_device_add - add raid_device object |
1502 | * @ioc: per adapter object |
1503 | * @raid_device: raid_device object |
1504 | * |
1505 | * This is added to the raid_device_list link list. |
1506 | */ |
1507 | static void |
1508 | _scsih_raid_device_add(struct MPT3SAS_ADAPTER *ioc, |
1509 | struct _raid_device *raid_device) |
1510 | { |
1511 | unsigned long flags; |
1512 | |
1513 | dewtprintk(ioc, |
1514 | ioc_info(ioc, "%s: handle(0x%04x), wwid(0x%016llx)\n" , |
1515 | __func__, |
1516 | raid_device->handle, (u64)raid_device->wwid)); |
1517 | |
1518 | spin_lock_irqsave(&ioc->raid_device_lock, flags); |
1519 | list_add_tail(new: &raid_device->list, head: &ioc->raid_device_list); |
1520 | spin_unlock_irqrestore(lock: &ioc->raid_device_lock, flags); |
1521 | } |
1522 | |
1523 | /** |
1524 | * _scsih_raid_device_remove - delete raid_device object |
1525 | * @ioc: per adapter object |
1526 | * @raid_device: raid_device object |
1527 | * |
1528 | */ |
1529 | static void |
1530 | _scsih_raid_device_remove(struct MPT3SAS_ADAPTER *ioc, |
1531 | struct _raid_device *raid_device) |
1532 | { |
1533 | unsigned long flags; |
1534 | |
1535 | spin_lock_irqsave(&ioc->raid_device_lock, flags); |
1536 | list_del(entry: &raid_device->list); |
1537 | kfree(objp: raid_device); |
1538 | spin_unlock_irqrestore(lock: &ioc->raid_device_lock, flags); |
1539 | } |
1540 | |
1541 | /** |
1542 | * mpt3sas_scsih_expander_find_by_handle - expander device search |
1543 | * @ioc: per adapter object |
1544 | * @handle: expander handle (assigned by firmware) |
1545 | * Context: Calling function should acquire ioc->sas_device_lock |
1546 | * |
1547 | * This searches for expander device based on handle, then returns the |
1548 | * sas_node object. |
1549 | */ |
1550 | struct _sas_node * |
1551 | mpt3sas_scsih_expander_find_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) |
1552 | { |
1553 | struct _sas_node *sas_expander, *r; |
1554 | |
1555 | r = NULL; |
1556 | list_for_each_entry(sas_expander, &ioc->sas_expander_list, list) { |
1557 | if (sas_expander->handle != handle) |
1558 | continue; |
1559 | r = sas_expander; |
1560 | goto out; |
1561 | } |
1562 | out: |
1563 | return r; |
1564 | } |
1565 | |
1566 | /** |
1567 | * mpt3sas_scsih_enclosure_find_by_handle - exclosure device search |
1568 | * @ioc: per adapter object |
1569 | * @handle: enclosure handle (assigned by firmware) |
1570 | * Context: Calling function should acquire ioc->sas_device_lock |
1571 | * |
1572 | * This searches for enclosure device based on handle, then returns the |
1573 | * enclosure object. |
1574 | */ |
1575 | static struct _enclosure_node * |
1576 | mpt3sas_scsih_enclosure_find_by_handle(struct MPT3SAS_ADAPTER *ioc, u16 handle) |
1577 | { |
1578 | struct _enclosure_node *enclosure_dev, *r; |
1579 | |
1580 | r = NULL; |
1581 | list_for_each_entry(enclosure_dev, &ioc->enclosure_list, list) { |
1582 | if (le16_to_cpu(enclosure_dev->pg0.EnclosureHandle) != handle) |
1583 | continue; |
1584 | r = enclosure_dev; |
1585 | goto out; |
1586 | } |
1587 | out: |
1588 | return r; |
1589 | } |
1590 | /** |
1591 | * mpt3sas_scsih_expander_find_by_sas_address - expander device search |
1592 | * @ioc: per adapter object |
1593 | * @sas_address: sas address |
1594 | * @port: hba port entry |
1595 | * Context: Calling function should acquire ioc->sas_node_lock. |
1596 | * |
1597 | * This searches for expander device based on sas_address & port number, |
1598 | * then returns the sas_node object. |
1599 | */ |
1600 | struct _sas_node * |
1601 | mpt3sas_scsih_expander_find_by_sas_address(struct MPT3SAS_ADAPTER *ioc, |
1602 | u64 sas_address, struct hba_port *port) |
1603 | { |
1604 | struct _sas_node *sas_expander, *r = NULL; |
1605 | |
1606 | if (!port) |
1607 | return r; |
1608 | |
1609 | list_for_each_entry(sas_expander, &ioc->sas_expander_list, list) { |
1610 | if (sas_expander->sas_address != sas_address) |
1611 | continue; |
1612 | if (sas_expander->port != port) |
1613 | continue; |
1614 | r = sas_expander; |
1615 | goto out; |
1616 | } |
1617 | out: |
1618 | return r; |
1619 | } |
1620 | |
1621 | /** |
1622 | * _scsih_expander_node_add - insert expander device to the list. |
1623 | * @ioc: per adapter object |
1624 | * @sas_expander: the sas_device object |
1625 | * Context: This function will acquire ioc->sas_node_lock. |
1626 | * |
1627 | * Adding new object to the ioc->sas_expander_list. |
1628 | */ |
1629 | static void |
1630 | _scsih_expander_node_add(struct MPT3SAS_ADAPTER *ioc, |
1631 | struct _sas_node *sas_expander) |
1632 | { |
1633 | unsigned long flags; |
1634 | |
1635 | spin_lock_irqsave(&ioc->sas_node_lock, flags); |
1636 | list_add_tail(new: &sas_expander->list, head: &ioc->sas_expander_list); |
1637 | spin_unlock_irqrestore(lock: &ioc->sas_node_lock, flags); |
1638 | } |
1639 | |
1640 | /** |
1641 | * _scsih_is_end_device - determines if device is an end device |
1642 | * @device_info: bitfield providing information about the device. |
1643 | * Context: none |
1644 | * |
1645 | * Return: 1 if end device. |
1646 | */ |
1647 | static int |
1648 | _scsih_is_end_device(u32 device_info) |
1649 | { |
1650 | if (device_info & MPI2_SAS_DEVICE_INFO_END_DEVICE && |
1651 | ((device_info & MPI2_SAS_DEVICE_INFO_SSP_TARGET) | |
1652 | (device_info & MPI2_SAS_DEVICE_INFO_STP_TARGET) | |
1653 | (device_info & MPI2_SAS_DEVICE_INFO_SATA_DEVICE))) |
1654 | return 1; |
1655 | else |
1656 | return 0; |
1657 | } |
1658 | |
1659 | /** |
1660 | * _scsih_is_nvme_pciescsi_device - determines if |
1661 | * device is an pcie nvme/scsi device |
1662 | * @device_info: bitfield providing information about the device. |
1663 | * Context: none |
1664 | * |
1665 | * Returns 1 if device is pcie device type nvme/scsi. |
1666 | */ |
1667 | static int |
1668 | _scsih_is_nvme_pciescsi_device(u32 device_info) |
1669 | { |
1670 | if (((device_info & MPI26_PCIE_DEVINFO_MASK_DEVICE_TYPE) |
1671 | == MPI26_PCIE_DEVINFO_NVME) || |
1672 | ((device_info & MPI26_PCIE_DEVINFO_MASK_DEVICE_TYPE) |
1673 | == MPI26_PCIE_DEVINFO_SCSI)) |
1674 | return 1; |
1675 | else |
1676 | return 0; |
1677 | } |
1678 | |
1679 | /** |
1680 | * _scsih_scsi_lookup_find_by_target - search for matching channel:id |
1681 | * @ioc: per adapter object |
1682 | * @id: target id |
1683 | * @channel: channel |
1684 | * Context: This function will acquire ioc->scsi_lookup_lock. |
1685 | * |
1686 | * This will search for a matching channel:id in the scsi_lookup array, |
1687 | * returning 1 if found. |
1688 | */ |
1689 | static u8 |
1690 | _scsih_scsi_lookup_find_by_target(struct MPT3SAS_ADAPTER *ioc, int id, |
1691 | int channel) |
1692 | { |
1693 | int smid; |
1694 | struct scsi_cmnd *scmd; |
1695 | |
1696 | for (smid = 1; |
1697 | smid <= ioc->shost->can_queue; smid++) { |
1698 | scmd = mpt3sas_scsih_scsi_lookup_get(ioc, smid); |
1699 | if (!scmd) |
1700 | continue; |
1701 | if (scmd->device->id == id && |
1702 | scmd->device->channel == channel) |
1703 | return 1; |
1704 | } |
1705 | return 0; |
1706 | } |
1707 | |
1708 | /** |
1709 | * _scsih_scsi_lookup_find_by_lun - search for matching channel:id:lun |
1710 | * @ioc: per adapter object |
1711 | * @id: target id |
1712 | * @lun: lun number |
1713 | * @channel: channel |
1714 | * Context: This function will acquire ioc->scsi_lookup_lock. |
1715 | * |
1716 | * This will search for a matching channel:id:lun in the scsi_lookup array, |
1717 | * returning 1 if found. |
1718 | */ |
1719 | static u8 |
1720 | _scsih_scsi_lookup_find_by_lun(struct MPT3SAS_ADAPTER *ioc, int id, |
1721 | unsigned int lun, int channel) |
1722 | { |
1723 | int smid; |
1724 | struct scsi_cmnd *scmd; |
1725 | |
1726 | for (smid = 1; smid <= ioc->shost->can_queue; smid++) { |
1727 | |
1728 | scmd = mpt3sas_scsih_scsi_lookup_get(ioc, smid); |
1729 | if (!scmd) |
1730 | continue; |
1731 | if (scmd->device->id == id && |
1732 | scmd->device->channel == channel && |
1733 | scmd->device->lun == lun) |
1734 | return 1; |
1735 | } |
1736 | return 0; |
1737 | } |
1738 | |
1739 | /** |
1740 | * mpt3sas_scsih_scsi_lookup_get - returns scmd entry |
1741 | * @ioc: per adapter object |
1742 | * @smid: system request message index |
1743 | * |
1744 | * Return: the smid stored scmd pointer. |
1745 | * Then will dereference the stored scmd pointer. |
1746 | */ |
1747 | struct scsi_cmnd * |
1748 | mpt3sas_scsih_scsi_lookup_get(struct MPT3SAS_ADAPTER *ioc, u16 smid) |
1749 | { |
1750 | struct scsi_cmnd *scmd = NULL; |
1751 | struct scsiio_tracker *st; |
1752 | Mpi25SCSIIORequest_t *mpi_request; |
1753 | u16 tag = smid - 1; |
1754 | |
1755 | if (smid > 0 && |
1756 | smid <= ioc->scsiio_depth - INTERNAL_SCSIIO_CMDS_COUNT) { |
1757 | u32 unique_tag = |
1758 | ioc->io_queue_num[tag] << BLK_MQ_UNIQUE_TAG_BITS | tag; |
1759 | |
1760 | mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); |
1761 | |
1762 | /* |
1763 | * If SCSI IO request is outstanding at driver level then |
1764 | * DevHandle filed must be non-zero. If DevHandle is zero |
1765 | * then it means that this smid is free at driver level, |
1766 | * so return NULL. |
1767 | */ |
1768 | if (!mpi_request->DevHandle) |
1769 | return scmd; |
1770 | |
1771 | scmd = scsi_host_find_tag(shost: ioc->shost, tag: unique_tag); |
1772 | if (scmd) { |
1773 | st = scsi_cmd_priv(cmd: scmd); |
1774 | if (st->cb_idx == 0xFF || st->smid == 0) |
1775 | scmd = NULL; |
1776 | } |
1777 | } |
1778 | return scmd; |
1779 | } |
1780 | |
1781 | /** |
1782 | * scsih_change_queue_depth - setting device queue depth |
1783 | * @sdev: scsi device struct |
1784 | * @qdepth: requested queue depth |
1785 | * |
1786 | * Return: queue depth. |
1787 | */ |
1788 | static int |
1789 | scsih_change_queue_depth(struct scsi_device *sdev, int qdepth) |
1790 | { |
1791 | struct Scsi_Host *shost = sdev->host; |
1792 | int max_depth; |
1793 | struct MPT3SAS_ADAPTER *ioc = shost_priv(shost); |
1794 | struct MPT3SAS_DEVICE *sas_device_priv_data; |
1795 | struct MPT3SAS_TARGET *sas_target_priv_data; |
1796 | struct _sas_device *sas_device; |
1797 | unsigned long flags; |
1798 | |
1799 | max_depth = shost->can_queue; |
1800 | |
1801 | /* |
1802 | * limit max device queue for SATA to 32 if enable_sdev_max_qd |
1803 | * is disabled. |
1804 | */ |
1805 | if (ioc->enable_sdev_max_qd || ioc->is_gen35_ioc) |
1806 | goto not_sata; |
1807 | |
1808 | sas_device_priv_data = sdev->hostdata; |
1809 | if (!sas_device_priv_data) |
1810 | goto not_sata; |
1811 | sas_target_priv_data = sas_device_priv_data->sas_target; |
1812 | if (!sas_target_priv_data) |
1813 | goto not_sata; |
1814 | if ((sas_target_priv_data->flags & MPT_TARGET_FLAGS_VOLUME)) |
1815 | goto not_sata; |
1816 | |
1817 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
1818 | sas_device = __mpt3sas_get_sdev_from_target(ioc, tgt_priv: sas_target_priv_data); |
1819 | if (sas_device) { |
1820 | if (sas_device->device_info & MPI2_SAS_DEVICE_INFO_SATA_DEVICE) |
1821 | max_depth = MPT3SAS_SATA_QUEUE_DEPTH; |
1822 | |
1823 | sas_device_put(s: sas_device); |
1824 | } |
1825 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
1826 | |
1827 | not_sata: |
1828 | |
1829 | if (!sdev->tagged_supported) |
1830 | max_depth = 1; |
1831 | if (qdepth > max_depth) |
1832 | qdepth = max_depth; |
1833 | scsi_change_queue_depth(sdev, qdepth); |
1834 | sdev_printk(KERN_INFO, sdev, |
1835 | "qdepth(%d), tagged(%d), scsi_level(%d), cmd_que(%d)\n" , |
1836 | sdev->queue_depth, sdev->tagged_supported, |
1837 | sdev->scsi_level, ((sdev->inquiry[7] & 2) >> 1)); |
1838 | return sdev->queue_depth; |
1839 | } |
1840 | |
1841 | /** |
1842 | * mpt3sas_scsih_change_queue_depth - setting device queue depth |
1843 | * @sdev: scsi device struct |
1844 | * @qdepth: requested queue depth |
1845 | * |
1846 | * Returns nothing. |
1847 | */ |
1848 | void |
1849 | mpt3sas_scsih_change_queue_depth(struct scsi_device *sdev, int qdepth) |
1850 | { |
1851 | struct Scsi_Host *shost = sdev->host; |
1852 | struct MPT3SAS_ADAPTER *ioc = shost_priv(shost); |
1853 | |
1854 | if (ioc->enable_sdev_max_qd) |
1855 | qdepth = shost->can_queue; |
1856 | |
1857 | scsih_change_queue_depth(sdev, qdepth); |
1858 | } |
1859 | |
1860 | /** |
1861 | * scsih_target_alloc - target add routine |
1862 | * @starget: scsi target struct |
1863 | * |
1864 | * Return: 0 if ok. Any other return is assumed to be an error and |
1865 | * the device is ignored. |
1866 | */ |
1867 | static int |
1868 | scsih_target_alloc(struct scsi_target *starget) |
1869 | { |
1870 | struct Scsi_Host *shost = dev_to_shost(dev: &starget->dev); |
1871 | struct MPT3SAS_ADAPTER *ioc = shost_priv(shost); |
1872 | struct MPT3SAS_TARGET *sas_target_priv_data; |
1873 | struct _sas_device *sas_device; |
1874 | struct _raid_device *raid_device; |
1875 | struct _pcie_device *pcie_device; |
1876 | unsigned long flags; |
1877 | struct sas_rphy *rphy; |
1878 | |
1879 | sas_target_priv_data = kzalloc(size: sizeof(*sas_target_priv_data), |
1880 | GFP_KERNEL); |
1881 | if (!sas_target_priv_data) |
1882 | return -ENOMEM; |
1883 | |
1884 | starget->hostdata = sas_target_priv_data; |
1885 | sas_target_priv_data->starget = starget; |
1886 | sas_target_priv_data->handle = MPT3SAS_INVALID_DEVICE_HANDLE; |
1887 | |
1888 | /* RAID volumes */ |
1889 | if (starget->channel == RAID_CHANNEL) { |
1890 | spin_lock_irqsave(&ioc->raid_device_lock, flags); |
1891 | raid_device = _scsih_raid_device_find_by_id(ioc, id: starget->id, |
1892 | channel: starget->channel); |
1893 | if (raid_device) { |
1894 | sas_target_priv_data->handle = raid_device->handle; |
1895 | sas_target_priv_data->sas_address = raid_device->wwid; |
1896 | sas_target_priv_data->flags |= MPT_TARGET_FLAGS_VOLUME; |
1897 | if (ioc->is_warpdrive) |
1898 | sas_target_priv_data->raid_device = raid_device; |
1899 | raid_device->starget = starget; |
1900 | } |
1901 | spin_unlock_irqrestore(lock: &ioc->raid_device_lock, flags); |
1902 | return 0; |
1903 | } |
1904 | |
1905 | /* PCIe devices */ |
1906 | if (starget->channel == PCIE_CHANNEL) { |
1907 | spin_lock_irqsave(&ioc->pcie_device_lock, flags); |
1908 | pcie_device = __mpt3sas_get_pdev_by_idchannel(ioc, id: starget->id, |
1909 | channel: starget->channel); |
1910 | if (pcie_device) { |
1911 | sas_target_priv_data->handle = pcie_device->handle; |
1912 | sas_target_priv_data->sas_address = pcie_device->wwid; |
1913 | sas_target_priv_data->port = NULL; |
1914 | sas_target_priv_data->pcie_dev = pcie_device; |
1915 | pcie_device->starget = starget; |
1916 | pcie_device->id = starget->id; |
1917 | pcie_device->channel = starget->channel; |
1918 | sas_target_priv_data->flags |= |
1919 | MPT_TARGET_FLAGS_PCIE_DEVICE; |
1920 | if (pcie_device->fast_path) |
1921 | sas_target_priv_data->flags |= |
1922 | MPT_TARGET_FASTPATH_IO; |
1923 | } |
1924 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
1925 | return 0; |
1926 | } |
1927 | |
1928 | /* sas/sata devices */ |
1929 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
1930 | rphy = dev_to_rphy(starget->dev.parent); |
1931 | sas_device = __mpt3sas_get_sdev_by_rphy(ioc, rphy); |
1932 | |
1933 | if (sas_device) { |
1934 | sas_target_priv_data->handle = sas_device->handle; |
1935 | sas_target_priv_data->sas_address = sas_device->sas_address; |
1936 | sas_target_priv_data->port = sas_device->port; |
1937 | sas_target_priv_data->sas_dev = sas_device; |
1938 | sas_device->starget = starget; |
1939 | sas_device->id = starget->id; |
1940 | sas_device->channel = starget->channel; |
1941 | if (test_bit(sas_device->handle, ioc->pd_handles)) |
1942 | sas_target_priv_data->flags |= |
1943 | MPT_TARGET_FLAGS_RAID_COMPONENT; |
1944 | if (sas_device->fast_path) |
1945 | sas_target_priv_data->flags |= |
1946 | MPT_TARGET_FASTPATH_IO; |
1947 | } |
1948 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
1949 | |
1950 | return 0; |
1951 | } |
1952 | |
1953 | /** |
1954 | * scsih_target_destroy - target destroy routine |
1955 | * @starget: scsi target struct |
1956 | */ |
1957 | static void |
1958 | scsih_target_destroy(struct scsi_target *starget) |
1959 | { |
1960 | struct Scsi_Host *shost = dev_to_shost(dev: &starget->dev); |
1961 | struct MPT3SAS_ADAPTER *ioc = shost_priv(shost); |
1962 | struct MPT3SAS_TARGET *sas_target_priv_data; |
1963 | struct _sas_device *sas_device; |
1964 | struct _raid_device *raid_device; |
1965 | struct _pcie_device *pcie_device; |
1966 | unsigned long flags; |
1967 | |
1968 | sas_target_priv_data = starget->hostdata; |
1969 | if (!sas_target_priv_data) |
1970 | return; |
1971 | |
1972 | if (starget->channel == RAID_CHANNEL) { |
1973 | spin_lock_irqsave(&ioc->raid_device_lock, flags); |
1974 | raid_device = _scsih_raid_device_find_by_id(ioc, id: starget->id, |
1975 | channel: starget->channel); |
1976 | if (raid_device) { |
1977 | raid_device->starget = NULL; |
1978 | raid_device->sdev = NULL; |
1979 | } |
1980 | spin_unlock_irqrestore(lock: &ioc->raid_device_lock, flags); |
1981 | goto out; |
1982 | } |
1983 | |
1984 | if (starget->channel == PCIE_CHANNEL) { |
1985 | spin_lock_irqsave(&ioc->pcie_device_lock, flags); |
1986 | pcie_device = __mpt3sas_get_pdev_from_target(ioc, |
1987 | tgt_priv: sas_target_priv_data); |
1988 | if (pcie_device && (pcie_device->starget == starget) && |
1989 | (pcie_device->id == starget->id) && |
1990 | (pcie_device->channel == starget->channel)) |
1991 | pcie_device->starget = NULL; |
1992 | |
1993 | if (pcie_device) { |
1994 | /* |
1995 | * Corresponding get() is in _scsih_target_alloc() |
1996 | */ |
1997 | sas_target_priv_data->pcie_dev = NULL; |
1998 | pcie_device_put(p: pcie_device); |
1999 | pcie_device_put(p: pcie_device); |
2000 | } |
2001 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
2002 | goto out; |
2003 | } |
2004 | |
2005 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
2006 | sas_device = __mpt3sas_get_sdev_from_target(ioc, tgt_priv: sas_target_priv_data); |
2007 | if (sas_device && (sas_device->starget == starget) && |
2008 | (sas_device->id == starget->id) && |
2009 | (sas_device->channel == starget->channel)) |
2010 | sas_device->starget = NULL; |
2011 | |
2012 | if (sas_device) { |
2013 | /* |
2014 | * Corresponding get() is in _scsih_target_alloc() |
2015 | */ |
2016 | sas_target_priv_data->sas_dev = NULL; |
2017 | sas_device_put(s: sas_device); |
2018 | |
2019 | sas_device_put(s: sas_device); |
2020 | } |
2021 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
2022 | |
2023 | out: |
2024 | kfree(objp: sas_target_priv_data); |
2025 | starget->hostdata = NULL; |
2026 | } |
2027 | |
2028 | /** |
2029 | * scsih_slave_alloc - device add routine |
2030 | * @sdev: scsi device struct |
2031 | * |
2032 | * Return: 0 if ok. Any other return is assumed to be an error and |
2033 | * the device is ignored. |
2034 | */ |
2035 | static int |
2036 | scsih_slave_alloc(struct scsi_device *sdev) |
2037 | { |
2038 | struct Scsi_Host *shost; |
2039 | struct MPT3SAS_ADAPTER *ioc; |
2040 | struct MPT3SAS_TARGET *sas_target_priv_data; |
2041 | struct MPT3SAS_DEVICE *sas_device_priv_data; |
2042 | struct scsi_target *starget; |
2043 | struct _raid_device *raid_device; |
2044 | struct _sas_device *sas_device; |
2045 | struct _pcie_device *pcie_device; |
2046 | unsigned long flags; |
2047 | |
2048 | sas_device_priv_data = kzalloc(size: sizeof(*sas_device_priv_data), |
2049 | GFP_KERNEL); |
2050 | if (!sas_device_priv_data) |
2051 | return -ENOMEM; |
2052 | |
2053 | sas_device_priv_data->lun = sdev->lun; |
2054 | sas_device_priv_data->flags = MPT_DEVICE_FLAGS_INIT; |
2055 | |
2056 | starget = scsi_target(sdev); |
2057 | sas_target_priv_data = starget->hostdata; |
2058 | sas_target_priv_data->num_luns++; |
2059 | sas_device_priv_data->sas_target = sas_target_priv_data; |
2060 | sdev->hostdata = sas_device_priv_data; |
2061 | if ((sas_target_priv_data->flags & MPT_TARGET_FLAGS_RAID_COMPONENT)) |
2062 | sdev->no_uld_attach = 1; |
2063 | |
2064 | shost = dev_to_shost(dev: &starget->dev); |
2065 | ioc = shost_priv(shost); |
2066 | if (starget->channel == RAID_CHANNEL) { |
2067 | spin_lock_irqsave(&ioc->raid_device_lock, flags); |
2068 | raid_device = _scsih_raid_device_find_by_id(ioc, |
2069 | id: starget->id, channel: starget->channel); |
2070 | if (raid_device) |
2071 | raid_device->sdev = sdev; /* raid is single lun */ |
2072 | spin_unlock_irqrestore(lock: &ioc->raid_device_lock, flags); |
2073 | } |
2074 | if (starget->channel == PCIE_CHANNEL) { |
2075 | spin_lock_irqsave(&ioc->pcie_device_lock, flags); |
2076 | pcie_device = __mpt3sas_get_pdev_by_wwid(ioc, |
2077 | wwid: sas_target_priv_data->sas_address); |
2078 | if (pcie_device && (pcie_device->starget == NULL)) { |
2079 | sdev_printk(KERN_INFO, sdev, |
2080 | "%s : pcie_device->starget set to starget @ %d\n" , |
2081 | __func__, __LINE__); |
2082 | pcie_device->starget = starget; |
2083 | } |
2084 | |
2085 | if (pcie_device) |
2086 | pcie_device_put(p: pcie_device); |
2087 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
2088 | |
2089 | } else if (!(sas_target_priv_data->flags & MPT_TARGET_FLAGS_VOLUME)) { |
2090 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
2091 | sas_device = __mpt3sas_get_sdev_by_addr(ioc, |
2092 | sas_address: sas_target_priv_data->sas_address, |
2093 | port: sas_target_priv_data->port); |
2094 | if (sas_device && (sas_device->starget == NULL)) { |
2095 | sdev_printk(KERN_INFO, sdev, |
2096 | "%s : sas_device->starget set to starget @ %d\n" , |
2097 | __func__, __LINE__); |
2098 | sas_device->starget = starget; |
2099 | } |
2100 | |
2101 | if (sas_device) |
2102 | sas_device_put(s: sas_device); |
2103 | |
2104 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
2105 | } |
2106 | |
2107 | return 0; |
2108 | } |
2109 | |
2110 | /** |
2111 | * scsih_slave_destroy - device destroy routine |
2112 | * @sdev: scsi device struct |
2113 | */ |
2114 | static void |
2115 | scsih_slave_destroy(struct scsi_device *sdev) |
2116 | { |
2117 | struct MPT3SAS_TARGET *sas_target_priv_data; |
2118 | struct scsi_target *starget; |
2119 | struct Scsi_Host *shost; |
2120 | struct MPT3SAS_ADAPTER *ioc; |
2121 | struct _sas_device *sas_device; |
2122 | struct _pcie_device *pcie_device; |
2123 | unsigned long flags; |
2124 | |
2125 | if (!sdev->hostdata) |
2126 | return; |
2127 | |
2128 | starget = scsi_target(sdev); |
2129 | sas_target_priv_data = starget->hostdata; |
2130 | sas_target_priv_data->num_luns--; |
2131 | |
2132 | shost = dev_to_shost(dev: &starget->dev); |
2133 | ioc = shost_priv(shost); |
2134 | |
2135 | if (sas_target_priv_data->flags & MPT_TARGET_FLAGS_PCIE_DEVICE) { |
2136 | spin_lock_irqsave(&ioc->pcie_device_lock, flags); |
2137 | pcie_device = __mpt3sas_get_pdev_from_target(ioc, |
2138 | tgt_priv: sas_target_priv_data); |
2139 | if (pcie_device && !sas_target_priv_data->num_luns) |
2140 | pcie_device->starget = NULL; |
2141 | |
2142 | if (pcie_device) |
2143 | pcie_device_put(p: pcie_device); |
2144 | |
2145 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
2146 | |
2147 | } else if (!(sas_target_priv_data->flags & MPT_TARGET_FLAGS_VOLUME)) { |
2148 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
2149 | sas_device = __mpt3sas_get_sdev_from_target(ioc, |
2150 | tgt_priv: sas_target_priv_data); |
2151 | if (sas_device && !sas_target_priv_data->num_luns) |
2152 | sas_device->starget = NULL; |
2153 | |
2154 | if (sas_device) |
2155 | sas_device_put(s: sas_device); |
2156 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
2157 | } |
2158 | |
2159 | kfree(objp: sdev->hostdata); |
2160 | sdev->hostdata = NULL; |
2161 | } |
2162 | |
2163 | /** |
2164 | * _scsih_display_sata_capabilities - sata capabilities |
2165 | * @ioc: per adapter object |
2166 | * @handle: device handle |
2167 | * @sdev: scsi device struct |
2168 | */ |
2169 | static void |
2170 | _scsih_display_sata_capabilities(struct MPT3SAS_ADAPTER *ioc, |
2171 | u16 handle, struct scsi_device *sdev) |
2172 | { |
2173 | Mpi2ConfigReply_t mpi_reply; |
2174 | Mpi2SasDevicePage0_t sas_device_pg0; |
2175 | u32 ioc_status; |
2176 | u16 flags; |
2177 | u32 device_info; |
2178 | |
2179 | if ((mpt3sas_config_get_sas_device_pg0(ioc, mpi_reply: &mpi_reply, config_page: &sas_device_pg0, |
2180 | MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) { |
2181 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
2182 | __FILE__, __LINE__, __func__); |
2183 | return; |
2184 | } |
2185 | |
2186 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & |
2187 | MPI2_IOCSTATUS_MASK; |
2188 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { |
2189 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
2190 | __FILE__, __LINE__, __func__); |
2191 | return; |
2192 | } |
2193 | |
2194 | flags = le16_to_cpu(sas_device_pg0.Flags); |
2195 | device_info = le32_to_cpu(sas_device_pg0.DeviceInfo); |
2196 | |
2197 | sdev_printk(KERN_INFO, sdev, |
2198 | "atapi(%s), ncq(%s), asyn_notify(%s), smart(%s), fua(%s), " |
2199 | "sw_preserve(%s)\n" , |
2200 | (device_info & MPI2_SAS_DEVICE_INFO_ATAPI_DEVICE) ? "y" : "n" , |
2201 | (flags & MPI2_SAS_DEVICE0_FLAGS_SATA_NCQ_SUPPORTED) ? "y" : "n" , |
2202 | (flags & MPI2_SAS_DEVICE0_FLAGS_SATA_ASYNCHRONOUS_NOTIFY) ? "y" : |
2203 | "n" , |
2204 | (flags & MPI2_SAS_DEVICE0_FLAGS_SATA_SMART_SUPPORTED) ? "y" : "n" , |
2205 | (flags & MPI2_SAS_DEVICE0_FLAGS_SATA_FUA_SUPPORTED) ? "y" : "n" , |
2206 | (flags & MPI2_SAS_DEVICE0_FLAGS_SATA_SW_PRESERVE) ? "y" : "n" ); |
2207 | } |
2208 | |
2209 | /* |
2210 | * raid transport support - |
2211 | * Enabled for SLES11 and newer, in older kernels the driver will panic when |
2212 | * unloading the driver followed by a load - I believe that the subroutine |
2213 | * raid_class_release() is not cleaning up properly. |
2214 | */ |
2215 | |
2216 | /** |
2217 | * scsih_is_raid - return boolean indicating device is raid volume |
2218 | * @dev: the device struct object |
2219 | */ |
2220 | static int |
2221 | scsih_is_raid(struct device *dev) |
2222 | { |
2223 | struct scsi_device *sdev = to_scsi_device(dev); |
2224 | struct MPT3SAS_ADAPTER *ioc = shost_priv(shost: sdev->host); |
2225 | |
2226 | if (ioc->is_warpdrive) |
2227 | return 0; |
2228 | return (sdev->channel == RAID_CHANNEL) ? 1 : 0; |
2229 | } |
2230 | |
2231 | static int |
2232 | scsih_is_nvme(struct device *dev) |
2233 | { |
2234 | struct scsi_device *sdev = to_scsi_device(dev); |
2235 | |
2236 | return (sdev->channel == PCIE_CHANNEL) ? 1 : 0; |
2237 | } |
2238 | |
2239 | /** |
2240 | * scsih_get_resync - get raid volume resync percent complete |
2241 | * @dev: the device struct object |
2242 | */ |
2243 | static void |
2244 | scsih_get_resync(struct device *dev) |
2245 | { |
2246 | struct scsi_device *sdev = to_scsi_device(dev); |
2247 | struct MPT3SAS_ADAPTER *ioc = shost_priv(shost: sdev->host); |
2248 | static struct _raid_device *raid_device; |
2249 | unsigned long flags; |
2250 | Mpi2RaidVolPage0_t vol_pg0; |
2251 | Mpi2ConfigReply_t mpi_reply; |
2252 | u32 volume_status_flags; |
2253 | u8 percent_complete; |
2254 | u16 handle; |
2255 | |
2256 | percent_complete = 0; |
2257 | handle = 0; |
2258 | if (ioc->is_warpdrive) |
2259 | goto out; |
2260 | |
2261 | spin_lock_irqsave(&ioc->raid_device_lock, flags); |
2262 | raid_device = _scsih_raid_device_find_by_id(ioc, id: sdev->id, |
2263 | channel: sdev->channel); |
2264 | if (raid_device) { |
2265 | handle = raid_device->handle; |
2266 | percent_complete = raid_device->percent_complete; |
2267 | } |
2268 | spin_unlock_irqrestore(lock: &ioc->raid_device_lock, flags); |
2269 | |
2270 | if (!handle) |
2271 | goto out; |
2272 | |
2273 | if (mpt3sas_config_get_raid_volume_pg0(ioc, mpi_reply: &mpi_reply, config_page: &vol_pg0, |
2274 | MPI2_RAID_VOLUME_PGAD_FORM_HANDLE, handle, |
2275 | sz: sizeof(Mpi2RaidVolPage0_t))) { |
2276 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
2277 | __FILE__, __LINE__, __func__); |
2278 | percent_complete = 0; |
2279 | goto out; |
2280 | } |
2281 | |
2282 | volume_status_flags = le32_to_cpu(vol_pg0.VolumeStatusFlags); |
2283 | if (!(volume_status_flags & |
2284 | MPI2_RAIDVOL0_STATUS_FLAG_RESYNC_IN_PROGRESS)) |
2285 | percent_complete = 0; |
2286 | |
2287 | out: |
2288 | |
2289 | switch (ioc->hba_mpi_version_belonged) { |
2290 | case MPI2_VERSION: |
2291 | raid_set_resync(r: mpt2sas_raid_template, dev, value: percent_complete); |
2292 | break; |
2293 | case MPI25_VERSION: |
2294 | case MPI26_VERSION: |
2295 | raid_set_resync(r: mpt3sas_raid_template, dev, value: percent_complete); |
2296 | break; |
2297 | } |
2298 | } |
2299 | |
2300 | /** |
2301 | * scsih_get_state - get raid volume level |
2302 | * @dev: the device struct object |
2303 | */ |
2304 | static void |
2305 | scsih_get_state(struct device *dev) |
2306 | { |
2307 | struct scsi_device *sdev = to_scsi_device(dev); |
2308 | struct MPT3SAS_ADAPTER *ioc = shost_priv(shost: sdev->host); |
2309 | static struct _raid_device *raid_device; |
2310 | unsigned long flags; |
2311 | Mpi2RaidVolPage0_t vol_pg0; |
2312 | Mpi2ConfigReply_t mpi_reply; |
2313 | u32 volstate; |
2314 | enum raid_state state = RAID_STATE_UNKNOWN; |
2315 | u16 handle = 0; |
2316 | |
2317 | spin_lock_irqsave(&ioc->raid_device_lock, flags); |
2318 | raid_device = _scsih_raid_device_find_by_id(ioc, id: sdev->id, |
2319 | channel: sdev->channel); |
2320 | if (raid_device) |
2321 | handle = raid_device->handle; |
2322 | spin_unlock_irqrestore(lock: &ioc->raid_device_lock, flags); |
2323 | |
2324 | if (!raid_device) |
2325 | goto out; |
2326 | |
2327 | if (mpt3sas_config_get_raid_volume_pg0(ioc, mpi_reply: &mpi_reply, config_page: &vol_pg0, |
2328 | MPI2_RAID_VOLUME_PGAD_FORM_HANDLE, handle, |
2329 | sz: sizeof(Mpi2RaidVolPage0_t))) { |
2330 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
2331 | __FILE__, __LINE__, __func__); |
2332 | goto out; |
2333 | } |
2334 | |
2335 | volstate = le32_to_cpu(vol_pg0.VolumeStatusFlags); |
2336 | if (volstate & MPI2_RAIDVOL0_STATUS_FLAG_RESYNC_IN_PROGRESS) { |
2337 | state = RAID_STATE_RESYNCING; |
2338 | goto out; |
2339 | } |
2340 | |
2341 | switch (vol_pg0.VolumeState) { |
2342 | case MPI2_RAID_VOL_STATE_OPTIMAL: |
2343 | case MPI2_RAID_VOL_STATE_ONLINE: |
2344 | state = RAID_STATE_ACTIVE; |
2345 | break; |
2346 | case MPI2_RAID_VOL_STATE_DEGRADED: |
2347 | state = RAID_STATE_DEGRADED; |
2348 | break; |
2349 | case MPI2_RAID_VOL_STATE_FAILED: |
2350 | case MPI2_RAID_VOL_STATE_MISSING: |
2351 | state = RAID_STATE_OFFLINE; |
2352 | break; |
2353 | } |
2354 | out: |
2355 | switch (ioc->hba_mpi_version_belonged) { |
2356 | case MPI2_VERSION: |
2357 | raid_set_state(r: mpt2sas_raid_template, dev, value: state); |
2358 | break; |
2359 | case MPI25_VERSION: |
2360 | case MPI26_VERSION: |
2361 | raid_set_state(r: mpt3sas_raid_template, dev, value: state); |
2362 | break; |
2363 | } |
2364 | } |
2365 | |
2366 | /** |
2367 | * _scsih_set_level - set raid level |
2368 | * @ioc: ? |
2369 | * @sdev: scsi device struct |
2370 | * @volume_type: volume type |
2371 | */ |
2372 | static void |
2373 | _scsih_set_level(struct MPT3SAS_ADAPTER *ioc, |
2374 | struct scsi_device *sdev, u8 volume_type) |
2375 | { |
2376 | enum raid_level level = RAID_LEVEL_UNKNOWN; |
2377 | |
2378 | switch (volume_type) { |
2379 | case MPI2_RAID_VOL_TYPE_RAID0: |
2380 | level = RAID_LEVEL_0; |
2381 | break; |
2382 | case MPI2_RAID_VOL_TYPE_RAID10: |
2383 | level = RAID_LEVEL_10; |
2384 | break; |
2385 | case MPI2_RAID_VOL_TYPE_RAID1E: |
2386 | level = RAID_LEVEL_1E; |
2387 | break; |
2388 | case MPI2_RAID_VOL_TYPE_RAID1: |
2389 | level = RAID_LEVEL_1; |
2390 | break; |
2391 | } |
2392 | |
2393 | switch (ioc->hba_mpi_version_belonged) { |
2394 | case MPI2_VERSION: |
2395 | raid_set_level(r: mpt2sas_raid_template, |
2396 | dev: &sdev->sdev_gendev, value: level); |
2397 | break; |
2398 | case MPI25_VERSION: |
2399 | case MPI26_VERSION: |
2400 | raid_set_level(r: mpt3sas_raid_template, |
2401 | dev: &sdev->sdev_gendev, value: level); |
2402 | break; |
2403 | } |
2404 | } |
2405 | |
2406 | |
2407 | /** |
2408 | * _scsih_get_volume_capabilities - volume capabilities |
2409 | * @ioc: per adapter object |
2410 | * @raid_device: the raid_device object |
2411 | * |
2412 | * Return: 0 for success, else 1 |
2413 | */ |
2414 | static int |
2415 | _scsih_get_volume_capabilities(struct MPT3SAS_ADAPTER *ioc, |
2416 | struct _raid_device *raid_device) |
2417 | { |
2418 | Mpi2RaidVolPage0_t *vol_pg0; |
2419 | Mpi2RaidPhysDiskPage0_t pd_pg0; |
2420 | Mpi2SasDevicePage0_t sas_device_pg0; |
2421 | Mpi2ConfigReply_t mpi_reply; |
2422 | u16 sz; |
2423 | u8 num_pds; |
2424 | |
2425 | if ((mpt3sas_config_get_number_pds(ioc, handle: raid_device->handle, |
2426 | num_pds: &num_pds)) || !num_pds) { |
2427 | dfailprintk(ioc, |
2428 | ioc_warn(ioc, "failure at %s:%d/%s()!\n" , |
2429 | __FILE__, __LINE__, __func__)); |
2430 | return 1; |
2431 | } |
2432 | |
2433 | raid_device->num_pds = num_pds; |
2434 | sz = struct_size(vol_pg0, PhysDisk, num_pds); |
2435 | vol_pg0 = kzalloc(size: sz, GFP_KERNEL); |
2436 | if (!vol_pg0) { |
2437 | dfailprintk(ioc, |
2438 | ioc_warn(ioc, "failure at %s:%d/%s()!\n" , |
2439 | __FILE__, __LINE__, __func__)); |
2440 | return 1; |
2441 | } |
2442 | |
2443 | if ((mpt3sas_config_get_raid_volume_pg0(ioc, mpi_reply: &mpi_reply, config_page: vol_pg0, |
2444 | MPI2_RAID_VOLUME_PGAD_FORM_HANDLE, handle: raid_device->handle, sz))) { |
2445 | dfailprintk(ioc, |
2446 | ioc_warn(ioc, "failure at %s:%d/%s()!\n" , |
2447 | __FILE__, __LINE__, __func__)); |
2448 | kfree(objp: vol_pg0); |
2449 | return 1; |
2450 | } |
2451 | |
2452 | raid_device->volume_type = vol_pg0->VolumeType; |
2453 | |
2454 | /* figure out what the underlying devices are by |
2455 | * obtaining the device_info bits for the 1st device |
2456 | */ |
2457 | if (!(mpt3sas_config_get_phys_disk_pg0(ioc, mpi_reply: &mpi_reply, |
2458 | config_page: &pd_pg0, MPI2_PHYSDISK_PGAD_FORM_PHYSDISKNUM, |
2459 | form_specific: vol_pg0->PhysDisk[0].PhysDiskNum))) { |
2460 | if (!(mpt3sas_config_get_sas_device_pg0(ioc, mpi_reply: &mpi_reply, |
2461 | config_page: &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, |
2462 | le16_to_cpu(pd_pg0.DevHandle)))) { |
2463 | raid_device->device_info = |
2464 | le32_to_cpu(sas_device_pg0.DeviceInfo); |
2465 | } |
2466 | } |
2467 | |
2468 | kfree(objp: vol_pg0); |
2469 | return 0; |
2470 | } |
2471 | |
2472 | /** |
2473 | * _scsih_enable_tlr - setting TLR flags |
2474 | * @ioc: per adapter object |
2475 | * @sdev: scsi device struct |
2476 | * |
2477 | * Enabling Transaction Layer Retries for tape devices when |
2478 | * vpd page 0x90 is present |
2479 | * |
2480 | */ |
2481 | static void |
2482 | _scsih_enable_tlr(struct MPT3SAS_ADAPTER *ioc, struct scsi_device *sdev) |
2483 | { |
2484 | |
2485 | /* only for TAPE */ |
2486 | if (sdev->type != TYPE_TAPE) |
2487 | return; |
2488 | |
2489 | if (!(ioc->facts.IOCCapabilities & MPI2_IOCFACTS_CAPABILITY_TLR)) |
2490 | return; |
2491 | |
2492 | sas_enable_tlr(sdev); |
2493 | sdev_printk(KERN_INFO, sdev, "TLR %s\n" , |
2494 | sas_is_tlr_enabled(sdev) ? "Enabled" : "Disabled" ); |
2495 | return; |
2496 | |
2497 | } |
2498 | |
2499 | /** |
2500 | * scsih_slave_configure - device configure routine. |
2501 | * @sdev: scsi device struct |
2502 | * |
2503 | * Return: 0 if ok. Any other return is assumed to be an error and |
2504 | * the device is ignored. |
2505 | */ |
2506 | static int |
2507 | scsih_slave_configure(struct scsi_device *sdev) |
2508 | { |
2509 | struct Scsi_Host *shost = sdev->host; |
2510 | struct MPT3SAS_ADAPTER *ioc = shost_priv(shost); |
2511 | struct MPT3SAS_DEVICE *sas_device_priv_data; |
2512 | struct MPT3SAS_TARGET *sas_target_priv_data; |
2513 | struct _sas_device *sas_device; |
2514 | struct _pcie_device *pcie_device; |
2515 | struct _raid_device *raid_device; |
2516 | unsigned long flags; |
2517 | int qdepth; |
2518 | u8 ssp_target = 0; |
2519 | char *ds = "" ; |
2520 | char *r_level = "" ; |
2521 | u16 handle, volume_handle = 0; |
2522 | u64 volume_wwid = 0; |
2523 | |
2524 | qdepth = 1; |
2525 | sas_device_priv_data = sdev->hostdata; |
2526 | sas_device_priv_data->configured_lun = 1; |
2527 | sas_device_priv_data->flags &= ~MPT_DEVICE_FLAGS_INIT; |
2528 | sas_target_priv_data = sas_device_priv_data->sas_target; |
2529 | handle = sas_target_priv_data->handle; |
2530 | |
2531 | /* raid volume handling */ |
2532 | if (sas_target_priv_data->flags & MPT_TARGET_FLAGS_VOLUME) { |
2533 | |
2534 | spin_lock_irqsave(&ioc->raid_device_lock, flags); |
2535 | raid_device = mpt3sas_raid_device_find_by_handle(ioc, handle); |
2536 | spin_unlock_irqrestore(lock: &ioc->raid_device_lock, flags); |
2537 | if (!raid_device) { |
2538 | dfailprintk(ioc, |
2539 | ioc_warn(ioc, "failure at %s:%d/%s()!\n" , |
2540 | __FILE__, __LINE__, __func__)); |
2541 | return 1; |
2542 | } |
2543 | |
2544 | if (_scsih_get_volume_capabilities(ioc, raid_device)) { |
2545 | dfailprintk(ioc, |
2546 | ioc_warn(ioc, "failure at %s:%d/%s()!\n" , |
2547 | __FILE__, __LINE__, __func__)); |
2548 | return 1; |
2549 | } |
2550 | |
2551 | /* |
2552 | * WARPDRIVE: Initialize the required data for Direct IO |
2553 | */ |
2554 | mpt3sas_init_warpdrive_properties(ioc, raid_device); |
2555 | |
2556 | /* RAID Queue Depth Support |
2557 | * IS volume = underlying qdepth of drive type, either |
2558 | * MPT3SAS_SAS_QUEUE_DEPTH or MPT3SAS_SATA_QUEUE_DEPTH |
2559 | * IM/IME/R10 = 128 (MPT3SAS_RAID_QUEUE_DEPTH) |
2560 | */ |
2561 | if (raid_device->device_info & |
2562 | MPI2_SAS_DEVICE_INFO_SSP_TARGET) { |
2563 | qdepth = MPT3SAS_SAS_QUEUE_DEPTH; |
2564 | ds = "SSP" ; |
2565 | } else { |
2566 | qdepth = MPT3SAS_SATA_QUEUE_DEPTH; |
2567 | if (raid_device->device_info & |
2568 | MPI2_SAS_DEVICE_INFO_SATA_DEVICE) |
2569 | ds = "SATA" ; |
2570 | else |
2571 | ds = "STP" ; |
2572 | } |
2573 | |
2574 | switch (raid_device->volume_type) { |
2575 | case MPI2_RAID_VOL_TYPE_RAID0: |
2576 | r_level = "RAID0" ; |
2577 | break; |
2578 | case MPI2_RAID_VOL_TYPE_RAID1E: |
2579 | qdepth = MPT3SAS_RAID_QUEUE_DEPTH; |
2580 | if (ioc->manu_pg10.OEMIdentifier && |
2581 | (le32_to_cpu(ioc->manu_pg10.GenericFlags0) & |
2582 | MFG10_GF0_R10_DISPLAY) && |
2583 | !(raid_device->num_pds % 2)) |
2584 | r_level = "RAID10" ; |
2585 | else |
2586 | r_level = "RAID1E" ; |
2587 | break; |
2588 | case MPI2_RAID_VOL_TYPE_RAID1: |
2589 | qdepth = MPT3SAS_RAID_QUEUE_DEPTH; |
2590 | r_level = "RAID1" ; |
2591 | break; |
2592 | case MPI2_RAID_VOL_TYPE_RAID10: |
2593 | qdepth = MPT3SAS_RAID_QUEUE_DEPTH; |
2594 | r_level = "RAID10" ; |
2595 | break; |
2596 | case MPI2_RAID_VOL_TYPE_UNKNOWN: |
2597 | default: |
2598 | qdepth = MPT3SAS_RAID_QUEUE_DEPTH; |
2599 | r_level = "RAIDX" ; |
2600 | break; |
2601 | } |
2602 | |
2603 | if (!ioc->hide_ir_msg) |
2604 | sdev_printk(KERN_INFO, sdev, |
2605 | "%s: handle(0x%04x), wwid(0x%016llx)," |
2606 | " pd_count(%d), type(%s)\n" , |
2607 | r_level, raid_device->handle, |
2608 | (unsigned long long)raid_device->wwid, |
2609 | raid_device->num_pds, ds); |
2610 | |
2611 | if (shost->max_sectors > MPT3SAS_RAID_MAX_SECTORS) { |
2612 | blk_queue_max_hw_sectors(sdev->request_queue, |
2613 | MPT3SAS_RAID_MAX_SECTORS); |
2614 | sdev_printk(KERN_INFO, sdev, |
2615 | "Set queue's max_sector to: %u\n" , |
2616 | MPT3SAS_RAID_MAX_SECTORS); |
2617 | } |
2618 | |
2619 | mpt3sas_scsih_change_queue_depth(sdev, qdepth); |
2620 | |
2621 | /* raid transport support */ |
2622 | if (!ioc->is_warpdrive) |
2623 | _scsih_set_level(ioc, sdev, volume_type: raid_device->volume_type); |
2624 | return 0; |
2625 | } |
2626 | |
2627 | /* non-raid handling */ |
2628 | if (sas_target_priv_data->flags & MPT_TARGET_FLAGS_RAID_COMPONENT) { |
2629 | if (mpt3sas_config_get_volume_handle(ioc, pd_handle: handle, |
2630 | volume_handle: &volume_handle)) { |
2631 | dfailprintk(ioc, |
2632 | ioc_warn(ioc, "failure at %s:%d/%s()!\n" , |
2633 | __FILE__, __LINE__, __func__)); |
2634 | return 1; |
2635 | } |
2636 | if (volume_handle && mpt3sas_config_get_volume_wwid(ioc, |
2637 | volume_handle, wwid: &volume_wwid)) { |
2638 | dfailprintk(ioc, |
2639 | ioc_warn(ioc, "failure at %s:%d/%s()!\n" , |
2640 | __FILE__, __LINE__, __func__)); |
2641 | return 1; |
2642 | } |
2643 | } |
2644 | |
2645 | /* PCIe handling */ |
2646 | if (sas_target_priv_data->flags & MPT_TARGET_FLAGS_PCIE_DEVICE) { |
2647 | spin_lock_irqsave(&ioc->pcie_device_lock, flags); |
2648 | pcie_device = __mpt3sas_get_pdev_by_wwid(ioc, |
2649 | wwid: sas_device_priv_data->sas_target->sas_address); |
2650 | if (!pcie_device) { |
2651 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
2652 | dfailprintk(ioc, |
2653 | ioc_warn(ioc, "failure at %s:%d/%s()!\n" , |
2654 | __FILE__, __LINE__, __func__)); |
2655 | return 1; |
2656 | } |
2657 | |
2658 | qdepth = ioc->max_nvme_qd; |
2659 | ds = "NVMe" ; |
2660 | sdev_printk(KERN_INFO, sdev, |
2661 | "%s: handle(0x%04x), wwid(0x%016llx), port(%d)\n" , |
2662 | ds, handle, (unsigned long long)pcie_device->wwid, |
2663 | pcie_device->port_num); |
2664 | if (pcie_device->enclosure_handle != 0) |
2665 | sdev_printk(KERN_INFO, sdev, |
2666 | "%s: enclosure logical id(0x%016llx), slot(%d)\n" , |
2667 | ds, |
2668 | (unsigned long long)pcie_device->enclosure_logical_id, |
2669 | pcie_device->slot); |
2670 | if (pcie_device->connector_name[0] != '\0') |
2671 | sdev_printk(KERN_INFO, sdev, |
2672 | "%s: enclosure level(0x%04x)," |
2673 | "connector name( %s)\n" , ds, |
2674 | pcie_device->enclosure_level, |
2675 | pcie_device->connector_name); |
2676 | |
2677 | if (pcie_device->nvme_mdts) |
2678 | blk_queue_max_hw_sectors(sdev->request_queue, |
2679 | pcie_device->nvme_mdts/512); |
2680 | |
2681 | pcie_device_put(p: pcie_device); |
2682 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
2683 | mpt3sas_scsih_change_queue_depth(sdev, qdepth); |
2684 | /* Enable QUEUE_FLAG_NOMERGES flag, so that IOs won't be |
2685 | ** merged and can eliminate holes created during merging |
2686 | ** operation. |
2687 | **/ |
2688 | blk_queue_flag_set(QUEUE_FLAG_NOMERGES, |
2689 | q: sdev->request_queue); |
2690 | blk_queue_virt_boundary(sdev->request_queue, |
2691 | ioc->page_size - 1); |
2692 | return 0; |
2693 | } |
2694 | |
2695 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
2696 | sas_device = __mpt3sas_get_sdev_by_addr(ioc, |
2697 | sas_address: sas_device_priv_data->sas_target->sas_address, |
2698 | port: sas_device_priv_data->sas_target->port); |
2699 | if (!sas_device) { |
2700 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
2701 | dfailprintk(ioc, |
2702 | ioc_warn(ioc, "failure at %s:%d/%s()!\n" , |
2703 | __FILE__, __LINE__, __func__)); |
2704 | return 1; |
2705 | } |
2706 | |
2707 | sas_device->volume_handle = volume_handle; |
2708 | sas_device->volume_wwid = volume_wwid; |
2709 | if (sas_device->device_info & MPI2_SAS_DEVICE_INFO_SSP_TARGET) { |
2710 | qdepth = (sas_device->port_type > 1) ? |
2711 | ioc->max_wideport_qd : ioc->max_narrowport_qd; |
2712 | ssp_target = 1; |
2713 | if (sas_device->device_info & |
2714 | MPI2_SAS_DEVICE_INFO_SEP) { |
2715 | sdev_printk(KERN_WARNING, sdev, |
2716 | "set ignore_delay_remove for handle(0x%04x)\n" , |
2717 | sas_device_priv_data->sas_target->handle); |
2718 | sas_device_priv_data->ignore_delay_remove = 1; |
2719 | ds = "SES" ; |
2720 | } else |
2721 | ds = "SSP" ; |
2722 | } else { |
2723 | qdepth = ioc->max_sata_qd; |
2724 | if (sas_device->device_info & MPI2_SAS_DEVICE_INFO_STP_TARGET) |
2725 | ds = "STP" ; |
2726 | else if (sas_device->device_info & |
2727 | MPI2_SAS_DEVICE_INFO_SATA_DEVICE) |
2728 | ds = "SATA" ; |
2729 | } |
2730 | |
2731 | sdev_printk(KERN_INFO, sdev, "%s: handle(0x%04x), " \ |
2732 | "sas_addr(0x%016llx), phy(%d), device_name(0x%016llx)\n" , |
2733 | ds, handle, (unsigned long long)sas_device->sas_address, |
2734 | sas_device->phy, (unsigned long long)sas_device->device_name); |
2735 | |
2736 | _scsih_display_enclosure_chassis_info(NULL, sas_device, sdev, NULL); |
2737 | |
2738 | sas_device_put(s: sas_device); |
2739 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
2740 | |
2741 | if (!ssp_target) |
2742 | _scsih_display_sata_capabilities(ioc, handle, sdev); |
2743 | |
2744 | |
2745 | mpt3sas_scsih_change_queue_depth(sdev, qdepth); |
2746 | |
2747 | if (ssp_target) { |
2748 | sas_read_port_mode_page(sdev); |
2749 | _scsih_enable_tlr(ioc, sdev); |
2750 | } |
2751 | |
2752 | return 0; |
2753 | } |
2754 | |
2755 | /** |
2756 | * scsih_bios_param - fetch head, sector, cylinder info for a disk |
2757 | * @sdev: scsi device struct |
2758 | * @bdev: pointer to block device context |
2759 | * @capacity: device size (in 512 byte sectors) |
2760 | * @params: three element array to place output: |
2761 | * params[0] number of heads (max 255) |
2762 | * params[1] number of sectors (max 63) |
2763 | * params[2] number of cylinders |
2764 | */ |
2765 | static int |
2766 | scsih_bios_param(struct scsi_device *sdev, struct block_device *bdev, |
2767 | sector_t capacity, int params[]) |
2768 | { |
2769 | int heads; |
2770 | int sectors; |
2771 | sector_t cylinders; |
2772 | ulong dummy; |
2773 | |
2774 | heads = 64; |
2775 | sectors = 32; |
2776 | |
2777 | dummy = heads * sectors; |
2778 | cylinders = capacity; |
2779 | sector_div(cylinders, dummy); |
2780 | |
2781 | /* |
2782 | * Handle extended translation size for logical drives |
2783 | * > 1Gb |
2784 | */ |
2785 | if ((ulong)capacity >= 0x200000) { |
2786 | heads = 255; |
2787 | sectors = 63; |
2788 | dummy = heads * sectors; |
2789 | cylinders = capacity; |
2790 | sector_div(cylinders, dummy); |
2791 | } |
2792 | |
2793 | /* return result */ |
2794 | params[0] = heads; |
2795 | params[1] = sectors; |
2796 | params[2] = cylinders; |
2797 | |
2798 | return 0; |
2799 | } |
2800 | |
2801 | /** |
2802 | * _scsih_response_code - translation of device response code |
2803 | * @ioc: per adapter object |
2804 | * @response_code: response code returned by the device |
2805 | */ |
2806 | static void |
2807 | _scsih_response_code(struct MPT3SAS_ADAPTER *ioc, u8 response_code) |
2808 | { |
2809 | char *desc; |
2810 | |
2811 | switch (response_code) { |
2812 | case MPI2_SCSITASKMGMT_RSP_TM_COMPLETE: |
2813 | desc = "task management request completed" ; |
2814 | break; |
2815 | case MPI2_SCSITASKMGMT_RSP_INVALID_FRAME: |
2816 | desc = "invalid frame" ; |
2817 | break; |
2818 | case MPI2_SCSITASKMGMT_RSP_TM_NOT_SUPPORTED: |
2819 | desc = "task management request not supported" ; |
2820 | break; |
2821 | case MPI2_SCSITASKMGMT_RSP_TM_FAILED: |
2822 | desc = "task management request failed" ; |
2823 | break; |
2824 | case MPI2_SCSITASKMGMT_RSP_TM_SUCCEEDED: |
2825 | desc = "task management request succeeded" ; |
2826 | break; |
2827 | case MPI2_SCSITASKMGMT_RSP_TM_INVALID_LUN: |
2828 | desc = "invalid lun" ; |
2829 | break; |
2830 | case 0xA: |
2831 | desc = "overlapped tag attempted" ; |
2832 | break; |
2833 | case MPI2_SCSITASKMGMT_RSP_IO_QUEUED_ON_IOC: |
2834 | desc = "task queued, however not sent to target" ; |
2835 | break; |
2836 | default: |
2837 | desc = "unknown" ; |
2838 | break; |
2839 | } |
2840 | ioc_warn(ioc, "response_code(0x%01x): %s\n" , response_code, desc); |
2841 | } |
2842 | |
2843 | /** |
2844 | * _scsih_tm_done - tm completion routine |
2845 | * @ioc: per adapter object |
2846 | * @smid: system request message index |
2847 | * @msix_index: MSIX table index supplied by the OS |
2848 | * @reply: reply message frame(lower 32bit addr) |
2849 | * Context: none. |
2850 | * |
2851 | * The callback handler when using scsih_issue_tm. |
2852 | * |
2853 | * Return: 1 meaning mf should be freed from _base_interrupt |
2854 | * 0 means the mf is freed from this function. |
2855 | */ |
2856 | static u8 |
2857 | _scsih_tm_done(struct MPT3SAS_ADAPTER *ioc, u16 smid, u8 msix_index, u32 reply) |
2858 | { |
2859 | MPI2DefaultReply_t *mpi_reply; |
2860 | |
2861 | if (ioc->tm_cmds.status == MPT3_CMD_NOT_USED) |
2862 | return 1; |
2863 | if (ioc->tm_cmds.smid != smid) |
2864 | return 1; |
2865 | ioc->tm_cmds.status |= MPT3_CMD_COMPLETE; |
2866 | mpi_reply = mpt3sas_base_get_reply_virt_addr(ioc, phys_addr: reply); |
2867 | if (mpi_reply) { |
2868 | memcpy(ioc->tm_cmds.reply, mpi_reply, mpi_reply->MsgLength*4); |
2869 | ioc->tm_cmds.status |= MPT3_CMD_REPLY_VALID; |
2870 | } |
2871 | ioc->tm_cmds.status &= ~MPT3_CMD_PENDING; |
2872 | complete(&ioc->tm_cmds.done); |
2873 | return 1; |
2874 | } |
2875 | |
2876 | /** |
2877 | * mpt3sas_scsih_set_tm_flag - set per target tm_busy |
2878 | * @ioc: per adapter object |
2879 | * @handle: device handle |
2880 | * |
2881 | * During taskmangement request, we need to freeze the device queue. |
2882 | */ |
2883 | void |
2884 | mpt3sas_scsih_set_tm_flag(struct MPT3SAS_ADAPTER *ioc, u16 handle) |
2885 | { |
2886 | struct MPT3SAS_DEVICE *sas_device_priv_data; |
2887 | struct scsi_device *sdev; |
2888 | u8 skip = 0; |
2889 | |
2890 | shost_for_each_device(sdev, ioc->shost) { |
2891 | if (skip) |
2892 | continue; |
2893 | sas_device_priv_data = sdev->hostdata; |
2894 | if (!sas_device_priv_data) |
2895 | continue; |
2896 | if (sas_device_priv_data->sas_target->handle == handle) { |
2897 | sas_device_priv_data->sas_target->tm_busy = 1; |
2898 | skip = 1; |
2899 | ioc->ignore_loginfos = 1; |
2900 | } |
2901 | } |
2902 | } |
2903 | |
2904 | /** |
2905 | * mpt3sas_scsih_clear_tm_flag - clear per target tm_busy |
2906 | * @ioc: per adapter object |
2907 | * @handle: device handle |
2908 | * |
2909 | * During taskmangement request, we need to freeze the device queue. |
2910 | */ |
2911 | void |
2912 | mpt3sas_scsih_clear_tm_flag(struct MPT3SAS_ADAPTER *ioc, u16 handle) |
2913 | { |
2914 | struct MPT3SAS_DEVICE *sas_device_priv_data; |
2915 | struct scsi_device *sdev; |
2916 | u8 skip = 0; |
2917 | |
2918 | shost_for_each_device(sdev, ioc->shost) { |
2919 | if (skip) |
2920 | continue; |
2921 | sas_device_priv_data = sdev->hostdata; |
2922 | if (!sas_device_priv_data) |
2923 | continue; |
2924 | if (sas_device_priv_data->sas_target->handle == handle) { |
2925 | sas_device_priv_data->sas_target->tm_busy = 0; |
2926 | skip = 1; |
2927 | ioc->ignore_loginfos = 0; |
2928 | } |
2929 | } |
2930 | } |
2931 | |
2932 | /** |
2933 | * scsih_tm_cmd_map_status - map the target reset & LUN reset TM status |
2934 | * @ioc: per adapter object |
2935 | * @channel: the channel assigned by the OS |
2936 | * @id: the id assigned by the OS |
2937 | * @lun: lun number |
2938 | * @type: MPI2_SCSITASKMGMT_TASKTYPE__XXX (defined in mpi2_init.h) |
2939 | * @smid_task: smid assigned to the task |
2940 | * |
2941 | * Look whether TM has aborted the timed out SCSI command, if |
2942 | * TM has aborted the IO then return SUCCESS else return FAILED. |
2943 | */ |
2944 | static int |
2945 | scsih_tm_cmd_map_status(struct MPT3SAS_ADAPTER *ioc, uint channel, |
2946 | uint id, uint lun, u8 type, u16 smid_task) |
2947 | { |
2948 | |
2949 | if (smid_task <= ioc->shost->can_queue) { |
2950 | switch (type) { |
2951 | case MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET: |
2952 | if (!(_scsih_scsi_lookup_find_by_target(ioc, |
2953 | id, channel))) |
2954 | return SUCCESS; |
2955 | break; |
2956 | case MPI2_SCSITASKMGMT_TASKTYPE_ABRT_TASK_SET: |
2957 | case MPI2_SCSITASKMGMT_TASKTYPE_LOGICAL_UNIT_RESET: |
2958 | if (!(_scsih_scsi_lookup_find_by_lun(ioc, id, |
2959 | lun, channel))) |
2960 | return SUCCESS; |
2961 | break; |
2962 | default: |
2963 | return SUCCESS; |
2964 | } |
2965 | } else if (smid_task == ioc->scsih_cmds.smid) { |
2966 | if ((ioc->scsih_cmds.status & MPT3_CMD_COMPLETE) || |
2967 | (ioc->scsih_cmds.status & MPT3_CMD_NOT_USED)) |
2968 | return SUCCESS; |
2969 | } else if (smid_task == ioc->ctl_cmds.smid) { |
2970 | if ((ioc->ctl_cmds.status & MPT3_CMD_COMPLETE) || |
2971 | (ioc->ctl_cmds.status & MPT3_CMD_NOT_USED)) |
2972 | return SUCCESS; |
2973 | } |
2974 | |
2975 | return FAILED; |
2976 | } |
2977 | |
2978 | /** |
2979 | * scsih_tm_post_processing - post processing of target & LUN reset |
2980 | * @ioc: per adapter object |
2981 | * @handle: device handle |
2982 | * @channel: the channel assigned by the OS |
2983 | * @id: the id assigned by the OS |
2984 | * @lun: lun number |
2985 | * @type: MPI2_SCSITASKMGMT_TASKTYPE__XXX (defined in mpi2_init.h) |
2986 | * @smid_task: smid assigned to the task |
2987 | * |
2988 | * Post processing of target & LUN reset. Due to interrupt latency |
2989 | * issue it possible that interrupt for aborted IO might not be |
2990 | * received yet. So before returning failure status, poll the |
2991 | * reply descriptor pools for the reply of timed out SCSI command. |
2992 | * Return FAILED status if reply for timed out is not received |
2993 | * otherwise return SUCCESS. |
2994 | */ |
2995 | static int |
2996 | scsih_tm_post_processing(struct MPT3SAS_ADAPTER *ioc, u16 handle, |
2997 | uint channel, uint id, uint lun, u8 type, u16 smid_task) |
2998 | { |
2999 | int rc; |
3000 | |
3001 | rc = scsih_tm_cmd_map_status(ioc, channel, id, lun, type, smid_task); |
3002 | if (rc == SUCCESS) |
3003 | return rc; |
3004 | |
3005 | ioc_info(ioc, |
3006 | "Poll ReplyDescriptor queues for completion of" |
3007 | " smid(%d), task_type(0x%02x), handle(0x%04x)\n" , |
3008 | smid_task, type, handle); |
3009 | |
3010 | /* |
3011 | * Due to interrupt latency issues, driver may receive interrupt for |
3012 | * TM first and then for aborted SCSI IO command. So, poll all the |
3013 | * ReplyDescriptor pools before returning the FAILED status to SML. |
3014 | */ |
3015 | mpt3sas_base_mask_interrupts(ioc); |
3016 | mpt3sas_base_sync_reply_irqs(ioc, poll: 1); |
3017 | mpt3sas_base_unmask_interrupts(ioc); |
3018 | |
3019 | return scsih_tm_cmd_map_status(ioc, channel, id, lun, type, smid_task); |
3020 | } |
3021 | |
3022 | /** |
3023 | * mpt3sas_scsih_issue_tm - main routine for sending tm requests |
3024 | * @ioc: per adapter struct |
3025 | * @handle: device handle |
3026 | * @channel: the channel assigned by the OS |
3027 | * @id: the id assigned by the OS |
3028 | * @lun: lun number |
3029 | * @type: MPI2_SCSITASKMGMT_TASKTYPE__XXX (defined in mpi2_init.h) |
3030 | * @smid_task: smid assigned to the task |
3031 | * @msix_task: MSIX table index supplied by the OS |
3032 | * @timeout: timeout in seconds |
3033 | * @tr_method: Target Reset Method |
3034 | * Context: user |
3035 | * |
3036 | * A generic API for sending task management requests to firmware. |
3037 | * |
3038 | * The callback index is set inside `ioc->tm_cb_idx`. |
3039 | * The caller is responsible to check for outstanding commands. |
3040 | * |
3041 | * Return: SUCCESS or FAILED. |
3042 | */ |
3043 | int |
3044 | mpt3sas_scsih_issue_tm(struct MPT3SAS_ADAPTER *ioc, u16 handle, uint channel, |
3045 | uint id, u64 lun, u8 type, u16 smid_task, u16 msix_task, |
3046 | u8 timeout, u8 tr_method) |
3047 | { |
3048 | Mpi2SCSITaskManagementRequest_t *mpi_request; |
3049 | Mpi2SCSITaskManagementReply_t *mpi_reply; |
3050 | Mpi25SCSIIORequest_t *request; |
3051 | u16 smid = 0; |
3052 | u32 ioc_state; |
3053 | int rc; |
3054 | u8 issue_reset = 0; |
3055 | |
3056 | lockdep_assert_held(&ioc->tm_cmds.mutex); |
3057 | |
3058 | if (ioc->tm_cmds.status != MPT3_CMD_NOT_USED) { |
3059 | ioc_info(ioc, "%s: tm_cmd busy!!!\n" , __func__); |
3060 | return FAILED; |
3061 | } |
3062 | |
3063 | if (ioc->shost_recovery || ioc->remove_host || |
3064 | ioc->pci_error_recovery) { |
3065 | ioc_info(ioc, "%s: host reset in progress!\n" , __func__); |
3066 | return FAILED; |
3067 | } |
3068 | |
3069 | ioc_state = mpt3sas_base_get_iocstate(ioc, cooked: 0); |
3070 | if (ioc_state & MPI2_DOORBELL_USED) { |
3071 | dhsprintk(ioc, ioc_info(ioc, "unexpected doorbell active!\n" )); |
3072 | rc = mpt3sas_base_hard_reset_handler(ioc, type: FORCE_BIG_HAMMER); |
3073 | return (!rc) ? SUCCESS : FAILED; |
3074 | } |
3075 | |
3076 | if ((ioc_state & MPI2_IOC_STATE_MASK) == MPI2_IOC_STATE_FAULT) { |
3077 | mpt3sas_print_fault_code(ioc, ioc_state & |
3078 | MPI2_DOORBELL_DATA_MASK); |
3079 | rc = mpt3sas_base_hard_reset_handler(ioc, type: FORCE_BIG_HAMMER); |
3080 | return (!rc) ? SUCCESS : FAILED; |
3081 | } else if ((ioc_state & MPI2_IOC_STATE_MASK) == |
3082 | MPI2_IOC_STATE_COREDUMP) { |
3083 | mpt3sas_print_coredump_info(ioc, ioc_state & |
3084 | MPI2_DOORBELL_DATA_MASK); |
3085 | rc = mpt3sas_base_hard_reset_handler(ioc, type: FORCE_BIG_HAMMER); |
3086 | return (!rc) ? SUCCESS : FAILED; |
3087 | } |
3088 | |
3089 | smid = mpt3sas_base_get_smid_hpr(ioc, cb_idx: ioc->tm_cb_idx); |
3090 | if (!smid) { |
3091 | ioc_err(ioc, "%s: failed obtaining a smid\n" , __func__); |
3092 | return FAILED; |
3093 | } |
3094 | |
3095 | dtmprintk(ioc, |
3096 | ioc_info(ioc, "sending tm: handle(0x%04x), task_type(0x%02x), smid(%d), timeout(%d), tr_method(0x%x)\n" , |
3097 | handle, type, smid_task, timeout, tr_method)); |
3098 | ioc->tm_cmds.status = MPT3_CMD_PENDING; |
3099 | mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); |
3100 | ioc->tm_cmds.smid = smid; |
3101 | memset(mpi_request, 0, sizeof(Mpi2SCSITaskManagementRequest_t)); |
3102 | memset(ioc->tm_cmds.reply, 0, sizeof(Mpi2SCSITaskManagementReply_t)); |
3103 | mpi_request->Function = MPI2_FUNCTION_SCSI_TASK_MGMT; |
3104 | mpi_request->DevHandle = cpu_to_le16(handle); |
3105 | mpi_request->TaskType = type; |
3106 | if (type == MPI2_SCSITASKMGMT_TASKTYPE_ABORT_TASK || |
3107 | type == MPI2_SCSITASKMGMT_TASKTYPE_QUERY_TASK) |
3108 | mpi_request->MsgFlags = tr_method; |
3109 | mpi_request->TaskMID = cpu_to_le16(smid_task); |
3110 | int_to_scsilun(lun, (struct scsi_lun *)mpi_request->LUN); |
3111 | mpt3sas_scsih_set_tm_flag(ioc, handle); |
3112 | init_completion(x: &ioc->tm_cmds.done); |
3113 | ioc->put_smid_hi_priority(ioc, smid, msix_task); |
3114 | wait_for_completion_timeout(x: &ioc->tm_cmds.done, timeout: timeout*HZ); |
3115 | if (!(ioc->tm_cmds.status & MPT3_CMD_COMPLETE)) { |
3116 | mpt3sas_check_cmd_timeout(ioc, |
3117 | ioc->tm_cmds.status, mpi_request, |
3118 | sizeof(Mpi2SCSITaskManagementRequest_t)/4, issue_reset); |
3119 | if (issue_reset) { |
3120 | rc = mpt3sas_base_hard_reset_handler(ioc, |
3121 | type: FORCE_BIG_HAMMER); |
3122 | rc = (!rc) ? SUCCESS : FAILED; |
3123 | goto out; |
3124 | } |
3125 | } |
3126 | |
3127 | /* sync IRQs in case those were busy during flush. */ |
3128 | mpt3sas_base_sync_reply_irqs(ioc, poll: 0); |
3129 | |
3130 | if (ioc->tm_cmds.status & MPT3_CMD_REPLY_VALID) { |
3131 | mpt3sas_trigger_master(ioc, MASTER_TRIGGER_TASK_MANAGMENT); |
3132 | mpi_reply = ioc->tm_cmds.reply; |
3133 | dtmprintk(ioc, |
3134 | ioc_info(ioc, "complete tm: ioc_status(0x%04x), loginfo(0x%08x), term_count(0x%08x)\n" , |
3135 | le16_to_cpu(mpi_reply->IOCStatus), |
3136 | le32_to_cpu(mpi_reply->IOCLogInfo), |
3137 | le32_to_cpu(mpi_reply->TerminationCount))); |
3138 | if (ioc->logging_level & MPT_DEBUG_TM) { |
3139 | _scsih_response_code(ioc, response_code: mpi_reply->ResponseCode); |
3140 | if (mpi_reply->IOCStatus) |
3141 | _debug_dump_mf(mpi_request, |
3142 | sz: sizeof(Mpi2SCSITaskManagementRequest_t)/4); |
3143 | } |
3144 | } |
3145 | |
3146 | switch (type) { |
3147 | case MPI2_SCSITASKMGMT_TASKTYPE_ABORT_TASK: |
3148 | rc = SUCCESS; |
3149 | /* |
3150 | * If DevHandle filed in smid_task's entry of request pool |
3151 | * doesn't match with device handle on which this task abort |
3152 | * TM is received then it means that TM has successfully |
3153 | * aborted the timed out command. Since smid_task's entry in |
3154 | * request pool will be memset to zero once the timed out |
3155 | * command is returned to the SML. If the command is not |
3156 | * aborted then smid_task’s entry won’t be cleared and it |
3157 | * will have same DevHandle value on which this task abort TM |
3158 | * is received and driver will return the TM status as FAILED. |
3159 | */ |
3160 | request = mpt3sas_base_get_msg_frame(ioc, smid: smid_task); |
3161 | if (le16_to_cpu(request->DevHandle) != handle) |
3162 | break; |
3163 | |
3164 | ioc_info(ioc, "Task abort tm failed: handle(0x%04x)," |
3165 | "timeout(%d) tr_method(0x%x) smid(%d) msix_index(%d)\n" , |
3166 | handle, timeout, tr_method, smid_task, msix_task); |
3167 | rc = FAILED; |
3168 | break; |
3169 | |
3170 | case MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET: |
3171 | case MPI2_SCSITASKMGMT_TASKTYPE_ABRT_TASK_SET: |
3172 | case MPI2_SCSITASKMGMT_TASKTYPE_LOGICAL_UNIT_RESET: |
3173 | rc = scsih_tm_post_processing(ioc, handle, channel, id, lun, |
3174 | type, smid_task); |
3175 | break; |
3176 | case MPI2_SCSITASKMGMT_TASKTYPE_QUERY_TASK: |
3177 | rc = SUCCESS; |
3178 | break; |
3179 | default: |
3180 | rc = FAILED; |
3181 | break; |
3182 | } |
3183 | |
3184 | out: |
3185 | mpt3sas_scsih_clear_tm_flag(ioc, handle); |
3186 | ioc->tm_cmds.status = MPT3_CMD_NOT_USED; |
3187 | return rc; |
3188 | } |
3189 | |
3190 | int mpt3sas_scsih_issue_locked_tm(struct MPT3SAS_ADAPTER *ioc, u16 handle, |
3191 | uint channel, uint id, u64 lun, u8 type, u16 smid_task, |
3192 | u16 msix_task, u8 timeout, u8 tr_method) |
3193 | { |
3194 | int ret; |
3195 | |
3196 | mutex_lock(&ioc->tm_cmds.mutex); |
3197 | ret = mpt3sas_scsih_issue_tm(ioc, handle, channel, id, lun, type, |
3198 | smid_task, msix_task, timeout, tr_method); |
3199 | mutex_unlock(lock: &ioc->tm_cmds.mutex); |
3200 | |
3201 | return ret; |
3202 | } |
3203 | |
3204 | /** |
3205 | * _scsih_tm_display_info - displays info about the device |
3206 | * @ioc: per adapter struct |
3207 | * @scmd: pointer to scsi command object |
3208 | * |
3209 | * Called by task management callback handlers. |
3210 | */ |
3211 | static void |
3212 | _scsih_tm_display_info(struct MPT3SAS_ADAPTER *ioc, struct scsi_cmnd *scmd) |
3213 | { |
3214 | struct scsi_target *starget = scmd->device->sdev_target; |
3215 | struct MPT3SAS_TARGET *priv_target = starget->hostdata; |
3216 | struct _sas_device *sas_device = NULL; |
3217 | struct _pcie_device *pcie_device = NULL; |
3218 | unsigned long flags; |
3219 | char *device_str = NULL; |
3220 | |
3221 | if (!priv_target) |
3222 | return; |
3223 | if (ioc->hide_ir_msg) |
3224 | device_str = "WarpDrive" ; |
3225 | else |
3226 | device_str = "volume" ; |
3227 | |
3228 | scsi_print_command(scmd); |
3229 | if (priv_target->flags & MPT_TARGET_FLAGS_VOLUME) { |
3230 | starget_printk(KERN_INFO, starget, |
3231 | "%s handle(0x%04x), %s wwid(0x%016llx)\n" , |
3232 | device_str, priv_target->handle, |
3233 | device_str, (unsigned long long)priv_target->sas_address); |
3234 | |
3235 | } else if (priv_target->flags & MPT_TARGET_FLAGS_PCIE_DEVICE) { |
3236 | spin_lock_irqsave(&ioc->pcie_device_lock, flags); |
3237 | pcie_device = __mpt3sas_get_pdev_from_target(ioc, tgt_priv: priv_target); |
3238 | if (pcie_device) { |
3239 | starget_printk(KERN_INFO, starget, |
3240 | "handle(0x%04x), wwid(0x%016llx), port(%d)\n" , |
3241 | pcie_device->handle, |
3242 | (unsigned long long)pcie_device->wwid, |
3243 | pcie_device->port_num); |
3244 | if (pcie_device->enclosure_handle != 0) |
3245 | starget_printk(KERN_INFO, starget, |
3246 | "enclosure logical id(0x%016llx), slot(%d)\n" , |
3247 | (unsigned long long) |
3248 | pcie_device->enclosure_logical_id, |
3249 | pcie_device->slot); |
3250 | if (pcie_device->connector_name[0] != '\0') |
3251 | starget_printk(KERN_INFO, starget, |
3252 | "enclosure level(0x%04x), connector name( %s)\n" , |
3253 | pcie_device->enclosure_level, |
3254 | pcie_device->connector_name); |
3255 | pcie_device_put(p: pcie_device); |
3256 | } |
3257 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
3258 | |
3259 | } else { |
3260 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
3261 | sas_device = __mpt3sas_get_sdev_from_target(ioc, tgt_priv: priv_target); |
3262 | if (sas_device) { |
3263 | if (priv_target->flags & |
3264 | MPT_TARGET_FLAGS_RAID_COMPONENT) { |
3265 | starget_printk(KERN_INFO, starget, |
3266 | "volume handle(0x%04x), " |
3267 | "volume wwid(0x%016llx)\n" , |
3268 | sas_device->volume_handle, |
3269 | (unsigned long long)sas_device->volume_wwid); |
3270 | } |
3271 | starget_printk(KERN_INFO, starget, |
3272 | "handle(0x%04x), sas_address(0x%016llx), phy(%d)\n" , |
3273 | sas_device->handle, |
3274 | (unsigned long long)sas_device->sas_address, |
3275 | sas_device->phy); |
3276 | |
3277 | _scsih_display_enclosure_chassis_info(NULL, sas_device, |
3278 | NULL, starget); |
3279 | |
3280 | sas_device_put(s: sas_device); |
3281 | } |
3282 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
3283 | } |
3284 | } |
3285 | |
3286 | /** |
3287 | * scsih_abort - eh threads main abort routine |
3288 | * @scmd: pointer to scsi command object |
3289 | * |
3290 | * Return: SUCCESS if command aborted else FAILED |
3291 | */ |
3292 | static int |
3293 | scsih_abort(struct scsi_cmnd *scmd) |
3294 | { |
3295 | struct MPT3SAS_ADAPTER *ioc = shost_priv(shost: scmd->device->host); |
3296 | struct MPT3SAS_DEVICE *sas_device_priv_data; |
3297 | struct scsiio_tracker *st = scsi_cmd_priv(cmd: scmd); |
3298 | u16 handle; |
3299 | int r; |
3300 | |
3301 | u8 timeout = 30; |
3302 | struct _pcie_device *pcie_device = NULL; |
3303 | sdev_printk(KERN_INFO, scmd->device, "attempting task abort!" |
3304 | "scmd(0x%p), outstanding for %u ms & timeout %u ms\n" , |
3305 | scmd, jiffies_to_msecs(jiffies - scmd->jiffies_at_alloc), |
3306 | (scsi_cmd_to_rq(scmd)->timeout / HZ) * 1000); |
3307 | _scsih_tm_display_info(ioc, scmd); |
3308 | |
3309 | sas_device_priv_data = scmd->device->hostdata; |
3310 | if (!sas_device_priv_data || !sas_device_priv_data->sas_target || |
3311 | ioc->remove_host) { |
3312 | sdev_printk(KERN_INFO, scmd->device, |
3313 | "device been deleted! scmd(0x%p)\n" , scmd); |
3314 | scmd->result = DID_NO_CONNECT << 16; |
3315 | scsi_done(cmd: scmd); |
3316 | r = SUCCESS; |
3317 | goto out; |
3318 | } |
3319 | |
3320 | /* check for completed command */ |
3321 | if (st == NULL || st->cb_idx == 0xFF) { |
3322 | sdev_printk(KERN_INFO, scmd->device, "No reference found at " |
3323 | "driver, assuming scmd(0x%p) might have completed\n" , scmd); |
3324 | scmd->result = DID_RESET << 16; |
3325 | r = SUCCESS; |
3326 | goto out; |
3327 | } |
3328 | |
3329 | /* for hidden raid components and volumes this is not supported */ |
3330 | if (sas_device_priv_data->sas_target->flags & |
3331 | MPT_TARGET_FLAGS_RAID_COMPONENT || |
3332 | sas_device_priv_data->sas_target->flags & MPT_TARGET_FLAGS_VOLUME) { |
3333 | scmd->result = DID_RESET << 16; |
3334 | r = FAILED; |
3335 | goto out; |
3336 | } |
3337 | |
3338 | mpt3sas_halt_firmware(ioc); |
3339 | |
3340 | handle = sas_device_priv_data->sas_target->handle; |
3341 | pcie_device = mpt3sas_get_pdev_by_handle(ioc, handle); |
3342 | if (pcie_device && (!ioc->tm_custom_handling) && |
3343 | (!(mpt3sas_scsih_is_pcie_scsi_device(device_info: pcie_device->device_info)))) |
3344 | timeout = ioc->nvme_abort_timeout; |
3345 | r = mpt3sas_scsih_issue_locked_tm(ioc, handle, channel: scmd->device->channel, |
3346 | id: scmd->device->id, lun: scmd->device->lun, |
3347 | MPI2_SCSITASKMGMT_TASKTYPE_ABORT_TASK, |
3348 | smid_task: st->smid, msix_task: st->msix_io, timeout, tr_method: 0); |
3349 | /* Command must be cleared after abort */ |
3350 | if (r == SUCCESS && st->cb_idx != 0xFF) |
3351 | r = FAILED; |
3352 | out: |
3353 | sdev_printk(KERN_INFO, scmd->device, "task abort: %s scmd(0x%p)\n" , |
3354 | ((r == SUCCESS) ? "SUCCESS" : "FAILED" ), scmd); |
3355 | if (pcie_device) |
3356 | pcie_device_put(p: pcie_device); |
3357 | return r; |
3358 | } |
3359 | |
3360 | /** |
3361 | * scsih_dev_reset - eh threads main device reset routine |
3362 | * @scmd: pointer to scsi command object |
3363 | * |
3364 | * Return: SUCCESS if command aborted else FAILED |
3365 | */ |
3366 | static int |
3367 | scsih_dev_reset(struct scsi_cmnd *scmd) |
3368 | { |
3369 | struct MPT3SAS_ADAPTER *ioc = shost_priv(shost: scmd->device->host); |
3370 | struct MPT3SAS_DEVICE *sas_device_priv_data; |
3371 | struct _sas_device *sas_device = NULL; |
3372 | struct _pcie_device *pcie_device = NULL; |
3373 | u16 handle; |
3374 | u8 tr_method = 0; |
3375 | u8 tr_timeout = 30; |
3376 | int r; |
3377 | |
3378 | struct scsi_target *starget = scmd->device->sdev_target; |
3379 | struct MPT3SAS_TARGET *target_priv_data = starget->hostdata; |
3380 | |
3381 | sdev_printk(KERN_INFO, scmd->device, |
3382 | "attempting device reset! scmd(0x%p)\n" , scmd); |
3383 | _scsih_tm_display_info(ioc, scmd); |
3384 | |
3385 | sas_device_priv_data = scmd->device->hostdata; |
3386 | if (!sas_device_priv_data || !sas_device_priv_data->sas_target || |
3387 | ioc->remove_host) { |
3388 | sdev_printk(KERN_INFO, scmd->device, |
3389 | "device been deleted! scmd(0x%p)\n" , scmd); |
3390 | scmd->result = DID_NO_CONNECT << 16; |
3391 | scsi_done(cmd: scmd); |
3392 | r = SUCCESS; |
3393 | goto out; |
3394 | } |
3395 | |
3396 | /* for hidden raid components obtain the volume_handle */ |
3397 | handle = 0; |
3398 | if (sas_device_priv_data->sas_target->flags & |
3399 | MPT_TARGET_FLAGS_RAID_COMPONENT) { |
3400 | sas_device = mpt3sas_get_sdev_from_target(ioc, |
3401 | tgt_priv: target_priv_data); |
3402 | if (sas_device) |
3403 | handle = sas_device->volume_handle; |
3404 | } else |
3405 | handle = sas_device_priv_data->sas_target->handle; |
3406 | |
3407 | if (!handle) { |
3408 | scmd->result = DID_RESET << 16; |
3409 | r = FAILED; |
3410 | goto out; |
3411 | } |
3412 | |
3413 | pcie_device = mpt3sas_get_pdev_by_handle(ioc, handle); |
3414 | |
3415 | if (pcie_device && (!ioc->tm_custom_handling) && |
3416 | (!(mpt3sas_scsih_is_pcie_scsi_device(device_info: pcie_device->device_info)))) { |
3417 | tr_timeout = pcie_device->reset_timeout; |
3418 | tr_method = MPI26_SCSITASKMGMT_MSGFLAGS_PROTOCOL_LVL_RST_PCIE; |
3419 | } else |
3420 | tr_method = MPI2_SCSITASKMGMT_MSGFLAGS_LINK_RESET; |
3421 | |
3422 | r = mpt3sas_scsih_issue_locked_tm(ioc, handle, channel: scmd->device->channel, |
3423 | id: scmd->device->id, lun: scmd->device->lun, |
3424 | MPI2_SCSITASKMGMT_TASKTYPE_LOGICAL_UNIT_RESET, smid_task: 0, msix_task: 0, |
3425 | timeout: tr_timeout, tr_method); |
3426 | /* Check for busy commands after reset */ |
3427 | if (r == SUCCESS && scsi_device_busy(sdev: scmd->device)) |
3428 | r = FAILED; |
3429 | out: |
3430 | sdev_printk(KERN_INFO, scmd->device, "device reset: %s scmd(0x%p)\n" , |
3431 | ((r == SUCCESS) ? "SUCCESS" : "FAILED" ), scmd); |
3432 | |
3433 | if (sas_device) |
3434 | sas_device_put(s: sas_device); |
3435 | if (pcie_device) |
3436 | pcie_device_put(p: pcie_device); |
3437 | |
3438 | return r; |
3439 | } |
3440 | |
3441 | /** |
3442 | * scsih_target_reset - eh threads main target reset routine |
3443 | * @scmd: pointer to scsi command object |
3444 | * |
3445 | * Return: SUCCESS if command aborted else FAILED |
3446 | */ |
3447 | static int |
3448 | scsih_target_reset(struct scsi_cmnd *scmd) |
3449 | { |
3450 | struct MPT3SAS_ADAPTER *ioc = shost_priv(shost: scmd->device->host); |
3451 | struct MPT3SAS_DEVICE *sas_device_priv_data; |
3452 | struct _sas_device *sas_device = NULL; |
3453 | struct _pcie_device *pcie_device = NULL; |
3454 | u16 handle; |
3455 | u8 tr_method = 0; |
3456 | u8 tr_timeout = 30; |
3457 | int r; |
3458 | struct scsi_target *starget = scmd->device->sdev_target; |
3459 | struct MPT3SAS_TARGET *target_priv_data = starget->hostdata; |
3460 | |
3461 | starget_printk(KERN_INFO, starget, |
3462 | "attempting target reset! scmd(0x%p)\n" , scmd); |
3463 | _scsih_tm_display_info(ioc, scmd); |
3464 | |
3465 | sas_device_priv_data = scmd->device->hostdata; |
3466 | if (!sas_device_priv_data || !sas_device_priv_data->sas_target || |
3467 | ioc->remove_host) { |
3468 | starget_printk(KERN_INFO, starget, |
3469 | "target been deleted! scmd(0x%p)\n" , scmd); |
3470 | scmd->result = DID_NO_CONNECT << 16; |
3471 | scsi_done(cmd: scmd); |
3472 | r = SUCCESS; |
3473 | goto out; |
3474 | } |
3475 | |
3476 | /* for hidden raid components obtain the volume_handle */ |
3477 | handle = 0; |
3478 | if (sas_device_priv_data->sas_target->flags & |
3479 | MPT_TARGET_FLAGS_RAID_COMPONENT) { |
3480 | sas_device = mpt3sas_get_sdev_from_target(ioc, |
3481 | tgt_priv: target_priv_data); |
3482 | if (sas_device) |
3483 | handle = sas_device->volume_handle; |
3484 | } else |
3485 | handle = sas_device_priv_data->sas_target->handle; |
3486 | |
3487 | if (!handle) { |
3488 | scmd->result = DID_RESET << 16; |
3489 | r = FAILED; |
3490 | goto out; |
3491 | } |
3492 | |
3493 | pcie_device = mpt3sas_get_pdev_by_handle(ioc, handle); |
3494 | |
3495 | if (pcie_device && (!ioc->tm_custom_handling) && |
3496 | (!(mpt3sas_scsih_is_pcie_scsi_device(device_info: pcie_device->device_info)))) { |
3497 | tr_timeout = pcie_device->reset_timeout; |
3498 | tr_method = MPI26_SCSITASKMGMT_MSGFLAGS_PROTOCOL_LVL_RST_PCIE; |
3499 | } else |
3500 | tr_method = MPI2_SCSITASKMGMT_MSGFLAGS_LINK_RESET; |
3501 | r = mpt3sas_scsih_issue_locked_tm(ioc, handle, channel: scmd->device->channel, |
3502 | id: scmd->device->id, lun: 0, |
3503 | MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET, smid_task: 0, msix_task: 0, |
3504 | timeout: tr_timeout, tr_method); |
3505 | /* Check for busy commands after reset */ |
3506 | if (r == SUCCESS && atomic_read(v: &starget->target_busy)) |
3507 | r = FAILED; |
3508 | out: |
3509 | starget_printk(KERN_INFO, starget, "target reset: %s scmd(0x%p)\n" , |
3510 | ((r == SUCCESS) ? "SUCCESS" : "FAILED" ), scmd); |
3511 | |
3512 | if (sas_device) |
3513 | sas_device_put(s: sas_device); |
3514 | if (pcie_device) |
3515 | pcie_device_put(p: pcie_device); |
3516 | return r; |
3517 | } |
3518 | |
3519 | |
3520 | /** |
3521 | * scsih_host_reset - eh threads main host reset routine |
3522 | * @scmd: pointer to scsi command object |
3523 | * |
3524 | * Return: SUCCESS if command aborted else FAILED |
3525 | */ |
3526 | static int |
3527 | scsih_host_reset(struct scsi_cmnd *scmd) |
3528 | { |
3529 | struct MPT3SAS_ADAPTER *ioc = shost_priv(shost: scmd->device->host); |
3530 | int r, retval; |
3531 | |
3532 | ioc_info(ioc, "attempting host reset! scmd(0x%p)\n" , scmd); |
3533 | scsi_print_command(scmd); |
3534 | |
3535 | if (ioc->is_driver_loading || ioc->remove_host) { |
3536 | ioc_info(ioc, "Blocking the host reset\n" ); |
3537 | r = FAILED; |
3538 | goto out; |
3539 | } |
3540 | |
3541 | retval = mpt3sas_base_hard_reset_handler(ioc, type: FORCE_BIG_HAMMER); |
3542 | r = (retval < 0) ? FAILED : SUCCESS; |
3543 | out: |
3544 | ioc_info(ioc, "host reset: %s scmd(0x%p)\n" , |
3545 | r == SUCCESS ? "SUCCESS" : "FAILED" , scmd); |
3546 | |
3547 | return r; |
3548 | } |
3549 | |
3550 | /** |
3551 | * _scsih_fw_event_add - insert and queue up fw_event |
3552 | * @ioc: per adapter object |
3553 | * @fw_event: object describing the event |
3554 | * Context: This function will acquire ioc->fw_event_lock. |
3555 | * |
3556 | * This adds the firmware event object into link list, then queues it up to |
3557 | * be processed from user context. |
3558 | */ |
3559 | static void |
3560 | _scsih_fw_event_add(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work *fw_event) |
3561 | { |
3562 | unsigned long flags; |
3563 | |
3564 | if (ioc->firmware_event_thread == NULL) |
3565 | return; |
3566 | |
3567 | spin_lock_irqsave(&ioc->fw_event_lock, flags); |
3568 | fw_event_work_get(fw_work: fw_event); |
3569 | INIT_LIST_HEAD(list: &fw_event->list); |
3570 | list_add_tail(new: &fw_event->list, head: &ioc->fw_event_list); |
3571 | INIT_WORK(&fw_event->work, _firmware_event_work); |
3572 | fw_event_work_get(fw_work: fw_event); |
3573 | queue_work(wq: ioc->firmware_event_thread, work: &fw_event->work); |
3574 | spin_unlock_irqrestore(lock: &ioc->fw_event_lock, flags); |
3575 | } |
3576 | |
3577 | /** |
3578 | * _scsih_fw_event_del_from_list - delete fw_event from the list |
3579 | * @ioc: per adapter object |
3580 | * @fw_event: object describing the event |
3581 | * Context: This function will acquire ioc->fw_event_lock. |
3582 | * |
3583 | * If the fw_event is on the fw_event_list, remove it and do a put. |
3584 | */ |
3585 | static void |
3586 | _scsih_fw_event_del_from_list(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work |
3587 | *fw_event) |
3588 | { |
3589 | unsigned long flags; |
3590 | |
3591 | spin_lock_irqsave(&ioc->fw_event_lock, flags); |
3592 | if (!list_empty(head: &fw_event->list)) { |
3593 | list_del_init(entry: &fw_event->list); |
3594 | fw_event_work_put(fw_work: fw_event); |
3595 | } |
3596 | spin_unlock_irqrestore(lock: &ioc->fw_event_lock, flags); |
3597 | } |
3598 | |
3599 | |
3600 | /** |
3601 | * mpt3sas_send_trigger_data_event - send event for processing trigger data |
3602 | * @ioc: per adapter object |
3603 | * @event_data: trigger event data |
3604 | */ |
3605 | void |
3606 | mpt3sas_send_trigger_data_event(struct MPT3SAS_ADAPTER *ioc, |
3607 | struct SL_WH_TRIGGERS_EVENT_DATA_T *event_data) |
3608 | { |
3609 | struct fw_event_work *fw_event; |
3610 | u16 sz; |
3611 | |
3612 | if (ioc->is_driver_loading) |
3613 | return; |
3614 | sz = sizeof(*event_data); |
3615 | fw_event = alloc_fw_event_work(len: sz); |
3616 | if (!fw_event) |
3617 | return; |
3618 | fw_event->event = MPT3SAS_PROCESS_TRIGGER_DIAG; |
3619 | fw_event->ioc = ioc; |
3620 | memcpy(fw_event->event_data, event_data, sizeof(*event_data)); |
3621 | _scsih_fw_event_add(ioc, fw_event); |
3622 | fw_event_work_put(fw_work: fw_event); |
3623 | } |
3624 | |
3625 | /** |
3626 | * _scsih_error_recovery_delete_devices - remove devices not responding |
3627 | * @ioc: per adapter object |
3628 | */ |
3629 | static void |
3630 | _scsih_error_recovery_delete_devices(struct MPT3SAS_ADAPTER *ioc) |
3631 | { |
3632 | struct fw_event_work *fw_event; |
3633 | |
3634 | fw_event = alloc_fw_event_work(len: 0); |
3635 | if (!fw_event) |
3636 | return; |
3637 | fw_event->event = MPT3SAS_REMOVE_UNRESPONDING_DEVICES; |
3638 | fw_event->ioc = ioc; |
3639 | _scsih_fw_event_add(ioc, fw_event); |
3640 | fw_event_work_put(fw_work: fw_event); |
3641 | } |
3642 | |
3643 | /** |
3644 | * mpt3sas_port_enable_complete - port enable completed (fake event) |
3645 | * @ioc: per adapter object |
3646 | */ |
3647 | void |
3648 | mpt3sas_port_enable_complete(struct MPT3SAS_ADAPTER *ioc) |
3649 | { |
3650 | struct fw_event_work *fw_event; |
3651 | |
3652 | fw_event = alloc_fw_event_work(len: 0); |
3653 | if (!fw_event) |
3654 | return; |
3655 | fw_event->event = MPT3SAS_PORT_ENABLE_COMPLETE; |
3656 | fw_event->ioc = ioc; |
3657 | _scsih_fw_event_add(ioc, fw_event); |
3658 | fw_event_work_put(fw_work: fw_event); |
3659 | } |
3660 | |
3661 | static struct fw_event_work *dequeue_next_fw_event(struct MPT3SAS_ADAPTER *ioc) |
3662 | { |
3663 | unsigned long flags; |
3664 | struct fw_event_work *fw_event = NULL; |
3665 | |
3666 | spin_lock_irqsave(&ioc->fw_event_lock, flags); |
3667 | if (!list_empty(head: &ioc->fw_event_list)) { |
3668 | fw_event = list_first_entry(&ioc->fw_event_list, |
3669 | struct fw_event_work, list); |
3670 | list_del_init(entry: &fw_event->list); |
3671 | fw_event_work_put(fw_work: fw_event); |
3672 | } |
3673 | spin_unlock_irqrestore(lock: &ioc->fw_event_lock, flags); |
3674 | |
3675 | return fw_event; |
3676 | } |
3677 | |
3678 | /** |
3679 | * _scsih_fw_event_cleanup_queue - cleanup event queue |
3680 | * @ioc: per adapter object |
3681 | * |
3682 | * Walk the firmware event queue, either killing timers, or waiting |
3683 | * for outstanding events to complete |
3684 | * |
3685 | * Context: task, can sleep |
3686 | */ |
3687 | static void |
3688 | _scsih_fw_event_cleanup_queue(struct MPT3SAS_ADAPTER *ioc) |
3689 | { |
3690 | struct fw_event_work *fw_event; |
3691 | |
3692 | if ((list_empty(head: &ioc->fw_event_list) && !ioc->current_event) || |
3693 | !ioc->firmware_event_thread) |
3694 | return; |
3695 | /* |
3696 | * Set current running event as ignore, so that |
3697 | * current running event will exit quickly. |
3698 | * As diag reset has occurred it is of no use |
3699 | * to process remaining stale event data entries. |
3700 | */ |
3701 | if (ioc->shost_recovery && ioc->current_event) |
3702 | ioc->current_event->ignore = 1; |
3703 | |
3704 | ioc->fw_events_cleanup = 1; |
3705 | while ((fw_event = dequeue_next_fw_event(ioc)) || |
3706 | (fw_event = ioc->current_event)) { |
3707 | |
3708 | /* |
3709 | * Don't call cancel_work_sync() for current_event |
3710 | * other than MPT3SAS_REMOVE_UNRESPONDING_DEVICES; |
3711 | * otherwise we may observe deadlock if current |
3712 | * hard reset issued as part of processing the current_event. |
3713 | * |
3714 | * Orginal logic of cleaning the current_event is added |
3715 | * for handling the back to back host reset issued by the user. |
3716 | * i.e. during back to back host reset, driver use to process |
3717 | * the two instances of MPT3SAS_REMOVE_UNRESPONDING_DEVICES |
3718 | * event back to back and this made the drives to unregister |
3719 | * the devices from SML. |
3720 | */ |
3721 | |
3722 | if (fw_event == ioc->current_event && |
3723 | ioc->current_event->event != |
3724 | MPT3SAS_REMOVE_UNRESPONDING_DEVICES) { |
3725 | ioc->current_event = NULL; |
3726 | continue; |
3727 | } |
3728 | |
3729 | /* |
3730 | * Driver has to clear ioc->start_scan flag when |
3731 | * it is cleaning up MPT3SAS_PORT_ENABLE_COMPLETE, |
3732 | * otherwise scsi_scan_host() API waits for the |
3733 | * 5 minute timer to expire. If we exit from |
3734 | * scsi_scan_host() early then we can issue the |
3735 | * new port enable request as part of current diag reset. |
3736 | */ |
3737 | if (fw_event->event == MPT3SAS_PORT_ENABLE_COMPLETE) { |
3738 | ioc->port_enable_cmds.status |= MPT3_CMD_RESET; |
3739 | ioc->start_scan = 0; |
3740 | } |
3741 | |
3742 | /* |
3743 | * Wait on the fw_event to complete. If this returns 1, then |
3744 | * the event was never executed, and we need a put for the |
3745 | * reference the work had on the fw_event. |
3746 | * |
3747 | * If it did execute, we wait for it to finish, and the put will |
3748 | * happen from _firmware_event_work() |
3749 | */ |
3750 | if (cancel_work_sync(work: &fw_event->work)) |
3751 | fw_event_work_put(fw_work: fw_event); |
3752 | |
3753 | } |
3754 | ioc->fw_events_cleanup = 0; |
3755 | } |
3756 | |
3757 | /** |
3758 | * _scsih_internal_device_block - block the sdev device |
3759 | * @sdev: per device object |
3760 | * @sas_device_priv_data : per device driver private data |
3761 | * |
3762 | * make sure device is blocked without error, if not |
3763 | * print an error |
3764 | */ |
3765 | static void |
3766 | _scsih_internal_device_block(struct scsi_device *sdev, |
3767 | struct MPT3SAS_DEVICE *sas_device_priv_data) |
3768 | { |
3769 | int r = 0; |
3770 | |
3771 | sdev_printk(KERN_INFO, sdev, "device_block, handle(0x%04x)\n" , |
3772 | sas_device_priv_data->sas_target->handle); |
3773 | sas_device_priv_data->block = 1; |
3774 | |
3775 | r = scsi_internal_device_block_nowait(sdev); |
3776 | if (r == -EINVAL) |
3777 | sdev_printk(KERN_WARNING, sdev, |
3778 | "device_block failed with return(%d) for handle(0x%04x)\n" , |
3779 | r, sas_device_priv_data->sas_target->handle); |
3780 | } |
3781 | |
3782 | /** |
3783 | * _scsih_internal_device_unblock - unblock the sdev device |
3784 | * @sdev: per device object |
3785 | * @sas_device_priv_data : per device driver private data |
3786 | * make sure device is unblocked without error, if not retry |
3787 | * by blocking and then unblocking |
3788 | */ |
3789 | |
3790 | static void |
3791 | _scsih_internal_device_unblock(struct scsi_device *sdev, |
3792 | struct MPT3SAS_DEVICE *sas_device_priv_data) |
3793 | { |
3794 | int r = 0; |
3795 | |
3796 | sdev_printk(KERN_WARNING, sdev, "device_unblock and setting to running, " |
3797 | "handle(0x%04x)\n" , sas_device_priv_data->sas_target->handle); |
3798 | sas_device_priv_data->block = 0; |
3799 | r = scsi_internal_device_unblock_nowait(sdev, new_state: SDEV_RUNNING); |
3800 | if (r == -EINVAL) { |
3801 | /* The device has been set to SDEV_RUNNING by SD layer during |
3802 | * device addition but the request queue is still stopped by |
3803 | * our earlier block call. We need to perform a block again |
3804 | * to get the device to SDEV_BLOCK and then to SDEV_RUNNING */ |
3805 | |
3806 | sdev_printk(KERN_WARNING, sdev, |
3807 | "device_unblock failed with return(%d) for handle(0x%04x) " |
3808 | "performing a block followed by an unblock\n" , |
3809 | r, sas_device_priv_data->sas_target->handle); |
3810 | sas_device_priv_data->block = 1; |
3811 | r = scsi_internal_device_block_nowait(sdev); |
3812 | if (r) |
3813 | sdev_printk(KERN_WARNING, sdev, "retried device_block " |
3814 | "failed with return(%d) for handle(0x%04x)\n" , |
3815 | r, sas_device_priv_data->sas_target->handle); |
3816 | |
3817 | sas_device_priv_data->block = 0; |
3818 | r = scsi_internal_device_unblock_nowait(sdev, new_state: SDEV_RUNNING); |
3819 | if (r) |
3820 | sdev_printk(KERN_WARNING, sdev, "retried device_unblock" |
3821 | " failed with return(%d) for handle(0x%04x)\n" , |
3822 | r, sas_device_priv_data->sas_target->handle); |
3823 | } |
3824 | } |
3825 | |
3826 | /** |
3827 | * _scsih_ublock_io_all_device - unblock every device |
3828 | * @ioc: per adapter object |
3829 | * |
3830 | * change the device state from block to running |
3831 | */ |
3832 | static void |
3833 | _scsih_ublock_io_all_device(struct MPT3SAS_ADAPTER *ioc) |
3834 | { |
3835 | struct MPT3SAS_DEVICE *sas_device_priv_data; |
3836 | struct scsi_device *sdev; |
3837 | |
3838 | shost_for_each_device(sdev, ioc->shost) { |
3839 | sas_device_priv_data = sdev->hostdata; |
3840 | if (!sas_device_priv_data) |
3841 | continue; |
3842 | if (!sas_device_priv_data->block) |
3843 | continue; |
3844 | |
3845 | dewtprintk(ioc, sdev_printk(KERN_INFO, sdev, |
3846 | "device_running, handle(0x%04x)\n" , |
3847 | sas_device_priv_data->sas_target->handle)); |
3848 | _scsih_internal_device_unblock(sdev, sas_device_priv_data); |
3849 | } |
3850 | } |
3851 | |
3852 | |
3853 | /** |
3854 | * _scsih_ublock_io_device - prepare device to be deleted |
3855 | * @ioc: per adapter object |
3856 | * @sas_address: sas address |
3857 | * @port: hba port entry |
3858 | * |
3859 | * unblock then put device in offline state |
3860 | */ |
3861 | static void |
3862 | _scsih_ublock_io_device(struct MPT3SAS_ADAPTER *ioc, |
3863 | u64 sas_address, struct hba_port *port) |
3864 | { |
3865 | struct MPT3SAS_DEVICE *sas_device_priv_data; |
3866 | struct scsi_device *sdev; |
3867 | |
3868 | shost_for_each_device(sdev, ioc->shost) { |
3869 | sas_device_priv_data = sdev->hostdata; |
3870 | if (!sas_device_priv_data || !sas_device_priv_data->sas_target) |
3871 | continue; |
3872 | if (sas_device_priv_data->sas_target->sas_address |
3873 | != sas_address) |
3874 | continue; |
3875 | if (sas_device_priv_data->sas_target->port != port) |
3876 | continue; |
3877 | if (sas_device_priv_data->block) |
3878 | _scsih_internal_device_unblock(sdev, |
3879 | sas_device_priv_data); |
3880 | } |
3881 | } |
3882 | |
3883 | /** |
3884 | * _scsih_block_io_all_device - set the device state to SDEV_BLOCK |
3885 | * @ioc: per adapter object |
3886 | * |
3887 | * During device pull we need to appropriately set the sdev state. |
3888 | */ |
3889 | static void |
3890 | _scsih_block_io_all_device(struct MPT3SAS_ADAPTER *ioc) |
3891 | { |
3892 | struct MPT3SAS_DEVICE *sas_device_priv_data; |
3893 | struct scsi_device *sdev; |
3894 | |
3895 | shost_for_each_device(sdev, ioc->shost) { |
3896 | sas_device_priv_data = sdev->hostdata; |
3897 | if (!sas_device_priv_data) |
3898 | continue; |
3899 | if (sas_device_priv_data->block) |
3900 | continue; |
3901 | if (sas_device_priv_data->ignore_delay_remove) { |
3902 | sdev_printk(KERN_INFO, sdev, |
3903 | "%s skip device_block for SES handle(0x%04x)\n" , |
3904 | __func__, sas_device_priv_data->sas_target->handle); |
3905 | continue; |
3906 | } |
3907 | _scsih_internal_device_block(sdev, sas_device_priv_data); |
3908 | } |
3909 | } |
3910 | |
3911 | /** |
3912 | * _scsih_block_io_device - set the device state to SDEV_BLOCK |
3913 | * @ioc: per adapter object |
3914 | * @handle: device handle |
3915 | * |
3916 | * During device pull we need to appropriately set the sdev state. |
3917 | */ |
3918 | static void |
3919 | _scsih_block_io_device(struct MPT3SAS_ADAPTER *ioc, u16 handle) |
3920 | { |
3921 | struct MPT3SAS_DEVICE *sas_device_priv_data; |
3922 | struct scsi_device *sdev; |
3923 | struct _sas_device *sas_device; |
3924 | |
3925 | sas_device = mpt3sas_get_sdev_by_handle(ioc, handle); |
3926 | |
3927 | shost_for_each_device(sdev, ioc->shost) { |
3928 | sas_device_priv_data = sdev->hostdata; |
3929 | if (!sas_device_priv_data) |
3930 | continue; |
3931 | if (sas_device_priv_data->sas_target->handle != handle) |
3932 | continue; |
3933 | if (sas_device_priv_data->block) |
3934 | continue; |
3935 | if (sas_device && sas_device->pend_sas_rphy_add) |
3936 | continue; |
3937 | if (sas_device_priv_data->ignore_delay_remove) { |
3938 | sdev_printk(KERN_INFO, sdev, |
3939 | "%s skip device_block for SES handle(0x%04x)\n" , |
3940 | __func__, sas_device_priv_data->sas_target->handle); |
3941 | continue; |
3942 | } |
3943 | _scsih_internal_device_block(sdev, sas_device_priv_data); |
3944 | } |
3945 | |
3946 | if (sas_device) |
3947 | sas_device_put(s: sas_device); |
3948 | } |
3949 | |
3950 | /** |
3951 | * _scsih_block_io_to_children_attached_to_ex |
3952 | * @ioc: per adapter object |
3953 | * @sas_expander: the sas_device object |
3954 | * |
3955 | * This routine set sdev state to SDEV_BLOCK for all devices |
3956 | * attached to this expander. This function called when expander is |
3957 | * pulled. |
3958 | */ |
3959 | static void |
3960 | _scsih_block_io_to_children_attached_to_ex(struct MPT3SAS_ADAPTER *ioc, |
3961 | struct _sas_node *sas_expander) |
3962 | { |
3963 | struct _sas_port *mpt3sas_port; |
3964 | struct _sas_device *sas_device; |
3965 | struct _sas_node *expander_sibling; |
3966 | unsigned long flags; |
3967 | |
3968 | if (!sas_expander) |
3969 | return; |
3970 | |
3971 | list_for_each_entry(mpt3sas_port, |
3972 | &sas_expander->sas_port_list, port_list) { |
3973 | if (mpt3sas_port->remote_identify.device_type == |
3974 | SAS_END_DEVICE) { |
3975 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
3976 | sas_device = __mpt3sas_get_sdev_by_addr(ioc, |
3977 | sas_address: mpt3sas_port->remote_identify.sas_address, |
3978 | port: mpt3sas_port->hba_port); |
3979 | if (sas_device) { |
3980 | set_bit(nr: sas_device->handle, |
3981 | addr: ioc->blocking_handles); |
3982 | sas_device_put(s: sas_device); |
3983 | } |
3984 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
3985 | } |
3986 | } |
3987 | |
3988 | list_for_each_entry(mpt3sas_port, |
3989 | &sas_expander->sas_port_list, port_list) { |
3990 | |
3991 | if (mpt3sas_port->remote_identify.device_type == |
3992 | SAS_EDGE_EXPANDER_DEVICE || |
3993 | mpt3sas_port->remote_identify.device_type == |
3994 | SAS_FANOUT_EXPANDER_DEVICE) { |
3995 | expander_sibling = |
3996 | mpt3sas_scsih_expander_find_by_sas_address( |
3997 | ioc, sas_address: mpt3sas_port->remote_identify.sas_address, |
3998 | port: mpt3sas_port->hba_port); |
3999 | _scsih_block_io_to_children_attached_to_ex(ioc, |
4000 | sas_expander: expander_sibling); |
4001 | } |
4002 | } |
4003 | } |
4004 | |
4005 | /** |
4006 | * _scsih_block_io_to_children_attached_directly |
4007 | * @ioc: per adapter object |
4008 | * @event_data: topology change event data |
4009 | * |
4010 | * This routine set sdev state to SDEV_BLOCK for all devices |
4011 | * direct attached during device pull. |
4012 | */ |
4013 | static void |
4014 | _scsih_block_io_to_children_attached_directly(struct MPT3SAS_ADAPTER *ioc, |
4015 | Mpi2EventDataSasTopologyChangeList_t *event_data) |
4016 | { |
4017 | int i; |
4018 | u16 handle; |
4019 | u16 reason_code; |
4020 | |
4021 | for (i = 0; i < event_data->NumEntries; i++) { |
4022 | handle = le16_to_cpu(event_data->PHY[i].AttachedDevHandle); |
4023 | if (!handle) |
4024 | continue; |
4025 | reason_code = event_data->PHY[i].PhyStatus & |
4026 | MPI2_EVENT_SAS_TOPO_RC_MASK; |
4027 | if (reason_code == MPI2_EVENT_SAS_TOPO_RC_DELAY_NOT_RESPONDING) |
4028 | _scsih_block_io_device(ioc, handle); |
4029 | } |
4030 | } |
4031 | |
4032 | /** |
4033 | * _scsih_block_io_to_pcie_children_attached_directly |
4034 | * @ioc: per adapter object |
4035 | * @event_data: topology change event data |
4036 | * |
4037 | * This routine set sdev state to SDEV_BLOCK for all devices |
4038 | * direct attached during device pull/reconnect. |
4039 | */ |
4040 | static void |
4041 | _scsih_block_io_to_pcie_children_attached_directly(struct MPT3SAS_ADAPTER *ioc, |
4042 | Mpi26EventDataPCIeTopologyChangeList_t *event_data) |
4043 | { |
4044 | int i; |
4045 | u16 handle; |
4046 | u16 reason_code; |
4047 | |
4048 | for (i = 0; i < event_data->NumEntries; i++) { |
4049 | handle = |
4050 | le16_to_cpu(event_data->PortEntry[i].AttachedDevHandle); |
4051 | if (!handle) |
4052 | continue; |
4053 | reason_code = event_data->PortEntry[i].PortStatus; |
4054 | if (reason_code == |
4055 | MPI26_EVENT_PCIE_TOPO_PS_DELAY_NOT_RESPONDING) |
4056 | _scsih_block_io_device(ioc, handle); |
4057 | } |
4058 | } |
4059 | /** |
4060 | * _scsih_tm_tr_send - send task management request |
4061 | * @ioc: per adapter object |
4062 | * @handle: device handle |
4063 | * Context: interrupt time. |
4064 | * |
4065 | * This code is to initiate the device removal handshake protocol |
4066 | * with controller firmware. This function will issue target reset |
4067 | * using high priority request queue. It will send a sas iounit |
4068 | * control request (MPI2_SAS_OP_REMOVE_DEVICE) from this completion. |
4069 | * |
4070 | * This is designed to send muliple task management request at the same |
4071 | * time to the fifo. If the fifo is full, we will append the request, |
4072 | * and process it in a future completion. |
4073 | */ |
4074 | static void |
4075 | _scsih_tm_tr_send(struct MPT3SAS_ADAPTER *ioc, u16 handle) |
4076 | { |
4077 | Mpi2SCSITaskManagementRequest_t *mpi_request; |
4078 | u16 smid; |
4079 | struct _sas_device *sas_device = NULL; |
4080 | struct _pcie_device *pcie_device = NULL; |
4081 | struct MPT3SAS_TARGET *sas_target_priv_data = NULL; |
4082 | u64 sas_address = 0; |
4083 | unsigned long flags; |
4084 | struct _tr_list *delayed_tr; |
4085 | u32 ioc_state; |
4086 | u8 tr_method = 0; |
4087 | struct hba_port *port = NULL; |
4088 | |
4089 | if (ioc->pci_error_recovery) { |
4090 | dewtprintk(ioc, |
4091 | ioc_info(ioc, "%s: host in pci error recovery: handle(0x%04x)\n" , |
4092 | __func__, handle)); |
4093 | return; |
4094 | } |
4095 | ioc_state = mpt3sas_base_get_iocstate(ioc, cooked: 1); |
4096 | if (ioc_state != MPI2_IOC_STATE_OPERATIONAL) { |
4097 | dewtprintk(ioc, |
4098 | ioc_info(ioc, "%s: host is not operational: handle(0x%04x)\n" , |
4099 | __func__, handle)); |
4100 | return; |
4101 | } |
4102 | |
4103 | /* if PD, then return */ |
4104 | if (test_bit(handle, ioc->pd_handles)) |
4105 | return; |
4106 | |
4107 | clear_bit(nr: handle, addr: ioc->pend_os_device_add); |
4108 | |
4109 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
4110 | sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle); |
4111 | if (sas_device && sas_device->starget && |
4112 | sas_device->starget->hostdata) { |
4113 | sas_target_priv_data = sas_device->starget->hostdata; |
4114 | sas_target_priv_data->deleted = 1; |
4115 | sas_address = sas_device->sas_address; |
4116 | port = sas_device->port; |
4117 | } |
4118 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
4119 | if (!sas_device) { |
4120 | spin_lock_irqsave(&ioc->pcie_device_lock, flags); |
4121 | pcie_device = __mpt3sas_get_pdev_by_handle(ioc, handle); |
4122 | if (pcie_device && pcie_device->starget && |
4123 | pcie_device->starget->hostdata) { |
4124 | sas_target_priv_data = pcie_device->starget->hostdata; |
4125 | sas_target_priv_data->deleted = 1; |
4126 | sas_address = pcie_device->wwid; |
4127 | } |
4128 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
4129 | if (pcie_device && (!ioc->tm_custom_handling) && |
4130 | (!(mpt3sas_scsih_is_pcie_scsi_device( |
4131 | device_info: pcie_device->device_info)))) |
4132 | tr_method = |
4133 | MPI26_SCSITASKMGMT_MSGFLAGS_PROTOCOL_LVL_RST_PCIE; |
4134 | else |
4135 | tr_method = MPI2_SCSITASKMGMT_MSGFLAGS_LINK_RESET; |
4136 | } |
4137 | if (sas_target_priv_data) { |
4138 | dewtprintk(ioc, |
4139 | ioc_info(ioc, "setting delete flag: handle(0x%04x), sas_addr(0x%016llx)\n" , |
4140 | handle, (u64)sas_address)); |
4141 | if (sas_device) { |
4142 | if (sas_device->enclosure_handle != 0) |
4143 | dewtprintk(ioc, |
4144 | ioc_info(ioc, "setting delete flag:enclosure logical id(0x%016llx), slot(%d)\n" , |
4145 | (u64)sas_device->enclosure_logical_id, |
4146 | sas_device->slot)); |
4147 | if (sas_device->connector_name[0] != '\0') |
4148 | dewtprintk(ioc, |
4149 | ioc_info(ioc, "setting delete flag: enclosure level(0x%04x), connector name( %s)\n" , |
4150 | sas_device->enclosure_level, |
4151 | sas_device->connector_name)); |
4152 | } else if (pcie_device) { |
4153 | if (pcie_device->enclosure_handle != 0) |
4154 | dewtprintk(ioc, |
4155 | ioc_info(ioc, "setting delete flag: logical id(0x%016llx), slot(%d)\n" , |
4156 | (u64)pcie_device->enclosure_logical_id, |
4157 | pcie_device->slot)); |
4158 | if (pcie_device->connector_name[0] != '\0') |
4159 | dewtprintk(ioc, |
4160 | ioc_info(ioc, "setting delete flag:, enclosure level(0x%04x), connector name( %s)\n" , |
4161 | pcie_device->enclosure_level, |
4162 | pcie_device->connector_name)); |
4163 | } |
4164 | _scsih_ublock_io_device(ioc, sas_address, port); |
4165 | sas_target_priv_data->handle = MPT3SAS_INVALID_DEVICE_HANDLE; |
4166 | } |
4167 | |
4168 | smid = mpt3sas_base_get_smid_hpr(ioc, cb_idx: ioc->tm_tr_cb_idx); |
4169 | if (!smid) { |
4170 | delayed_tr = kzalloc(size: sizeof(*delayed_tr), GFP_ATOMIC); |
4171 | if (!delayed_tr) |
4172 | goto out; |
4173 | INIT_LIST_HEAD(list: &delayed_tr->list); |
4174 | delayed_tr->handle = handle; |
4175 | list_add_tail(new: &delayed_tr->list, head: &ioc->delayed_tr_list); |
4176 | dewtprintk(ioc, |
4177 | ioc_info(ioc, "DELAYED:tr:handle(0x%04x), (open)\n" , |
4178 | handle)); |
4179 | goto out; |
4180 | } |
4181 | |
4182 | dewtprintk(ioc, |
4183 | ioc_info(ioc, "tr_send:handle(0x%04x), (open), smid(%d), cb(%d)\n" , |
4184 | handle, smid, ioc->tm_tr_cb_idx)); |
4185 | mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); |
4186 | memset(mpi_request, 0, sizeof(Mpi2SCSITaskManagementRequest_t)); |
4187 | mpi_request->Function = MPI2_FUNCTION_SCSI_TASK_MGMT; |
4188 | mpi_request->DevHandle = cpu_to_le16(handle); |
4189 | mpi_request->TaskType = MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET; |
4190 | mpi_request->MsgFlags = tr_method; |
4191 | set_bit(nr: handle, addr: ioc->device_remove_in_progress); |
4192 | ioc->put_smid_hi_priority(ioc, smid, 0); |
4193 | mpt3sas_trigger_master(ioc, MASTER_TRIGGER_DEVICE_REMOVAL); |
4194 | |
4195 | out: |
4196 | if (sas_device) |
4197 | sas_device_put(s: sas_device); |
4198 | if (pcie_device) |
4199 | pcie_device_put(p: pcie_device); |
4200 | } |
4201 | |
4202 | /** |
4203 | * _scsih_tm_tr_complete - |
4204 | * @ioc: per adapter object |
4205 | * @smid: system request message index |
4206 | * @msix_index: MSIX table index supplied by the OS |
4207 | * @reply: reply message frame(lower 32bit addr) |
4208 | * Context: interrupt time. |
4209 | * |
4210 | * This is the target reset completion routine. |
4211 | * This code is part of the code to initiate the device removal |
4212 | * handshake protocol with controller firmware. |
4213 | * It will send a sas iounit control request (MPI2_SAS_OP_REMOVE_DEVICE) |
4214 | * |
4215 | * Return: 1 meaning mf should be freed from _base_interrupt |
4216 | * 0 means the mf is freed from this function. |
4217 | */ |
4218 | static u8 |
4219 | _scsih_tm_tr_complete(struct MPT3SAS_ADAPTER *ioc, u16 smid, u8 msix_index, |
4220 | u32 reply) |
4221 | { |
4222 | u16 handle; |
4223 | Mpi2SCSITaskManagementRequest_t *mpi_request_tm; |
4224 | Mpi2SCSITaskManagementReply_t *mpi_reply = |
4225 | mpt3sas_base_get_reply_virt_addr(ioc, phys_addr: reply); |
4226 | Mpi2SasIoUnitControlRequest_t *mpi_request; |
4227 | u16 smid_sas_ctrl; |
4228 | u32 ioc_state; |
4229 | struct _sc_list *delayed_sc; |
4230 | |
4231 | if (ioc->pci_error_recovery) { |
4232 | dewtprintk(ioc, |
4233 | ioc_info(ioc, "%s: host in pci error recovery\n" , |
4234 | __func__)); |
4235 | return 1; |
4236 | } |
4237 | ioc_state = mpt3sas_base_get_iocstate(ioc, cooked: 1); |
4238 | if (ioc_state != MPI2_IOC_STATE_OPERATIONAL) { |
4239 | dewtprintk(ioc, |
4240 | ioc_info(ioc, "%s: host is not operational\n" , |
4241 | __func__)); |
4242 | return 1; |
4243 | } |
4244 | if (unlikely(!mpi_reply)) { |
4245 | ioc_err(ioc, "mpi_reply not valid at %s:%d/%s()!\n" , |
4246 | __FILE__, __LINE__, __func__); |
4247 | return 1; |
4248 | } |
4249 | mpi_request_tm = mpt3sas_base_get_msg_frame(ioc, smid); |
4250 | handle = le16_to_cpu(mpi_request_tm->DevHandle); |
4251 | if (handle != le16_to_cpu(mpi_reply->DevHandle)) { |
4252 | dewtprintk(ioc, |
4253 | ioc_err(ioc, "spurious interrupt: handle(0x%04x:0x%04x), smid(%d)!!!\n" , |
4254 | handle, |
4255 | le16_to_cpu(mpi_reply->DevHandle), smid)); |
4256 | return 0; |
4257 | } |
4258 | |
4259 | mpt3sas_trigger_master(ioc, MASTER_TRIGGER_TASK_MANAGMENT); |
4260 | dewtprintk(ioc, |
4261 | ioc_info(ioc, "tr_complete:handle(0x%04x), (open) smid(%d), ioc_status(0x%04x), loginfo(0x%08x), completed(%d)\n" , |
4262 | handle, smid, le16_to_cpu(mpi_reply->IOCStatus), |
4263 | le32_to_cpu(mpi_reply->IOCLogInfo), |
4264 | le32_to_cpu(mpi_reply->TerminationCount))); |
4265 | |
4266 | smid_sas_ctrl = mpt3sas_base_get_smid(ioc, cb_idx: ioc->tm_sas_control_cb_idx); |
4267 | if (!smid_sas_ctrl) { |
4268 | delayed_sc = kzalloc(size: sizeof(*delayed_sc), GFP_ATOMIC); |
4269 | if (!delayed_sc) |
4270 | return _scsih_check_for_pending_tm(ioc, smid); |
4271 | INIT_LIST_HEAD(list: &delayed_sc->list); |
4272 | delayed_sc->handle = le16_to_cpu(mpi_request_tm->DevHandle); |
4273 | list_add_tail(new: &delayed_sc->list, head: &ioc->delayed_sc_list); |
4274 | dewtprintk(ioc, |
4275 | ioc_info(ioc, "DELAYED:sc:handle(0x%04x), (open)\n" , |
4276 | handle)); |
4277 | return _scsih_check_for_pending_tm(ioc, smid); |
4278 | } |
4279 | |
4280 | dewtprintk(ioc, |
4281 | ioc_info(ioc, "sc_send:handle(0x%04x), (open), smid(%d), cb(%d)\n" , |
4282 | handle, smid_sas_ctrl, ioc->tm_sas_control_cb_idx)); |
4283 | mpi_request = mpt3sas_base_get_msg_frame(ioc, smid: smid_sas_ctrl); |
4284 | memset(mpi_request, 0, sizeof(Mpi2SasIoUnitControlRequest_t)); |
4285 | mpi_request->Function = MPI2_FUNCTION_SAS_IO_UNIT_CONTROL; |
4286 | mpi_request->Operation = MPI2_SAS_OP_REMOVE_DEVICE; |
4287 | mpi_request->DevHandle = mpi_request_tm->DevHandle; |
4288 | ioc->put_smid_default(ioc, smid_sas_ctrl); |
4289 | |
4290 | return _scsih_check_for_pending_tm(ioc, smid); |
4291 | } |
4292 | |
4293 | /** _scsih_allow_scmd_to_device - check whether scmd needs to |
4294 | * issue to IOC or not. |
4295 | * @ioc: per adapter object |
4296 | * @scmd: pointer to scsi command object |
4297 | * |
4298 | * Returns true if scmd can be issued to IOC otherwise returns false. |
4299 | */ |
4300 | inline bool _scsih_allow_scmd_to_device(struct MPT3SAS_ADAPTER *ioc, |
4301 | struct scsi_cmnd *scmd) |
4302 | { |
4303 | |
4304 | if (ioc->pci_error_recovery) |
4305 | return false; |
4306 | |
4307 | if (ioc->hba_mpi_version_belonged == MPI2_VERSION) { |
4308 | if (ioc->remove_host) |
4309 | return false; |
4310 | |
4311 | return true; |
4312 | } |
4313 | |
4314 | if (ioc->remove_host) { |
4315 | |
4316 | switch (scmd->cmnd[0]) { |
4317 | case SYNCHRONIZE_CACHE: |
4318 | case START_STOP: |
4319 | return true; |
4320 | default: |
4321 | return false; |
4322 | } |
4323 | } |
4324 | |
4325 | return true; |
4326 | } |
4327 | |
4328 | /** |
4329 | * _scsih_sas_control_complete - completion routine |
4330 | * @ioc: per adapter object |
4331 | * @smid: system request message index |
4332 | * @msix_index: MSIX table index supplied by the OS |
4333 | * @reply: reply message frame(lower 32bit addr) |
4334 | * Context: interrupt time. |
4335 | * |
4336 | * This is the sas iounit control completion routine. |
4337 | * This code is part of the code to initiate the device removal |
4338 | * handshake protocol with controller firmware. |
4339 | * |
4340 | * Return: 1 meaning mf should be freed from _base_interrupt |
4341 | * 0 means the mf is freed from this function. |
4342 | */ |
4343 | static u8 |
4344 | _scsih_sas_control_complete(struct MPT3SAS_ADAPTER *ioc, u16 smid, |
4345 | u8 msix_index, u32 reply) |
4346 | { |
4347 | Mpi2SasIoUnitControlReply_t *mpi_reply = |
4348 | mpt3sas_base_get_reply_virt_addr(ioc, phys_addr: reply); |
4349 | |
4350 | if (likely(mpi_reply)) { |
4351 | dewtprintk(ioc, |
4352 | ioc_info(ioc, "sc_complete:handle(0x%04x), (open) smid(%d), ioc_status(0x%04x), loginfo(0x%08x)\n" , |
4353 | le16_to_cpu(mpi_reply->DevHandle), smid, |
4354 | le16_to_cpu(mpi_reply->IOCStatus), |
4355 | le32_to_cpu(mpi_reply->IOCLogInfo))); |
4356 | if (le16_to_cpu(mpi_reply->IOCStatus) == |
4357 | MPI2_IOCSTATUS_SUCCESS) { |
4358 | clear_bit(le16_to_cpu(mpi_reply->DevHandle), |
4359 | addr: ioc->device_remove_in_progress); |
4360 | } |
4361 | } else { |
4362 | ioc_err(ioc, "mpi_reply not valid at %s:%d/%s()!\n" , |
4363 | __FILE__, __LINE__, __func__); |
4364 | } |
4365 | return mpt3sas_check_for_pending_internal_cmds(ioc, smid); |
4366 | } |
4367 | |
4368 | /** |
4369 | * _scsih_tm_tr_volume_send - send target reset request for volumes |
4370 | * @ioc: per adapter object |
4371 | * @handle: device handle |
4372 | * Context: interrupt time. |
4373 | * |
4374 | * This is designed to send muliple task management request at the same |
4375 | * time to the fifo. If the fifo is full, we will append the request, |
4376 | * and process it in a future completion. |
4377 | */ |
4378 | static void |
4379 | _scsih_tm_tr_volume_send(struct MPT3SAS_ADAPTER *ioc, u16 handle) |
4380 | { |
4381 | Mpi2SCSITaskManagementRequest_t *mpi_request; |
4382 | u16 smid; |
4383 | struct _tr_list *delayed_tr; |
4384 | |
4385 | if (ioc->pci_error_recovery) { |
4386 | dewtprintk(ioc, |
4387 | ioc_info(ioc, "%s: host reset in progress!\n" , |
4388 | __func__)); |
4389 | return; |
4390 | } |
4391 | |
4392 | smid = mpt3sas_base_get_smid_hpr(ioc, cb_idx: ioc->tm_tr_volume_cb_idx); |
4393 | if (!smid) { |
4394 | delayed_tr = kzalloc(size: sizeof(*delayed_tr), GFP_ATOMIC); |
4395 | if (!delayed_tr) |
4396 | return; |
4397 | INIT_LIST_HEAD(list: &delayed_tr->list); |
4398 | delayed_tr->handle = handle; |
4399 | list_add_tail(new: &delayed_tr->list, head: &ioc->delayed_tr_volume_list); |
4400 | dewtprintk(ioc, |
4401 | ioc_info(ioc, "DELAYED:tr:handle(0x%04x), (open)\n" , |
4402 | handle)); |
4403 | return; |
4404 | } |
4405 | |
4406 | dewtprintk(ioc, |
4407 | ioc_info(ioc, "tr_send:handle(0x%04x), (open), smid(%d), cb(%d)\n" , |
4408 | handle, smid, ioc->tm_tr_volume_cb_idx)); |
4409 | mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); |
4410 | memset(mpi_request, 0, sizeof(Mpi2SCSITaskManagementRequest_t)); |
4411 | mpi_request->Function = MPI2_FUNCTION_SCSI_TASK_MGMT; |
4412 | mpi_request->DevHandle = cpu_to_le16(handle); |
4413 | mpi_request->TaskType = MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET; |
4414 | ioc->put_smid_hi_priority(ioc, smid, 0); |
4415 | } |
4416 | |
4417 | /** |
4418 | * _scsih_tm_volume_tr_complete - target reset completion |
4419 | * @ioc: per adapter object |
4420 | * @smid: system request message index |
4421 | * @msix_index: MSIX table index supplied by the OS |
4422 | * @reply: reply message frame(lower 32bit addr) |
4423 | * Context: interrupt time. |
4424 | * |
4425 | * Return: 1 meaning mf should be freed from _base_interrupt |
4426 | * 0 means the mf is freed from this function. |
4427 | */ |
4428 | static u8 |
4429 | _scsih_tm_volume_tr_complete(struct MPT3SAS_ADAPTER *ioc, u16 smid, |
4430 | u8 msix_index, u32 reply) |
4431 | { |
4432 | u16 handle; |
4433 | Mpi2SCSITaskManagementRequest_t *mpi_request_tm; |
4434 | Mpi2SCSITaskManagementReply_t *mpi_reply = |
4435 | mpt3sas_base_get_reply_virt_addr(ioc, phys_addr: reply); |
4436 | |
4437 | if (ioc->shost_recovery || ioc->pci_error_recovery) { |
4438 | dewtprintk(ioc, |
4439 | ioc_info(ioc, "%s: host reset in progress!\n" , |
4440 | __func__)); |
4441 | return 1; |
4442 | } |
4443 | if (unlikely(!mpi_reply)) { |
4444 | ioc_err(ioc, "mpi_reply not valid at %s:%d/%s()!\n" , |
4445 | __FILE__, __LINE__, __func__); |
4446 | return 1; |
4447 | } |
4448 | |
4449 | mpi_request_tm = mpt3sas_base_get_msg_frame(ioc, smid); |
4450 | handle = le16_to_cpu(mpi_request_tm->DevHandle); |
4451 | if (handle != le16_to_cpu(mpi_reply->DevHandle)) { |
4452 | dewtprintk(ioc, |
4453 | ioc_err(ioc, "spurious interrupt: handle(0x%04x:0x%04x), smid(%d)!!!\n" , |
4454 | handle, le16_to_cpu(mpi_reply->DevHandle), |
4455 | smid)); |
4456 | return 0; |
4457 | } |
4458 | |
4459 | dewtprintk(ioc, |
4460 | ioc_info(ioc, "tr_complete:handle(0x%04x), (open) smid(%d), ioc_status(0x%04x), loginfo(0x%08x), completed(%d)\n" , |
4461 | handle, smid, le16_to_cpu(mpi_reply->IOCStatus), |
4462 | le32_to_cpu(mpi_reply->IOCLogInfo), |
4463 | le32_to_cpu(mpi_reply->TerminationCount))); |
4464 | |
4465 | return _scsih_check_for_pending_tm(ioc, smid); |
4466 | } |
4467 | |
4468 | /** |
4469 | * _scsih_issue_delayed_event_ack - issue delayed Event ACK messages |
4470 | * @ioc: per adapter object |
4471 | * @smid: system request message index |
4472 | * @event: Event ID |
4473 | * @event_context: used to track events uniquely |
4474 | * |
4475 | * Context - processed in interrupt context. |
4476 | */ |
4477 | static void |
4478 | _scsih_issue_delayed_event_ack(struct MPT3SAS_ADAPTER *ioc, u16 smid, U16 event, |
4479 | U32 event_context) |
4480 | { |
4481 | Mpi2EventAckRequest_t *ack_request; |
4482 | int i = smid - ioc->internal_smid; |
4483 | unsigned long flags; |
4484 | |
4485 | /* Without releasing the smid just update the |
4486 | * call back index and reuse the same smid for |
4487 | * processing this delayed request |
4488 | */ |
4489 | spin_lock_irqsave(&ioc->scsi_lookup_lock, flags); |
4490 | ioc->internal_lookup[i].cb_idx = ioc->base_cb_idx; |
4491 | spin_unlock_irqrestore(lock: &ioc->scsi_lookup_lock, flags); |
4492 | |
4493 | dewtprintk(ioc, |
4494 | ioc_info(ioc, "EVENT ACK: event(0x%04x), smid(%d), cb(%d)\n" , |
4495 | le16_to_cpu(event), smid, ioc->base_cb_idx)); |
4496 | ack_request = mpt3sas_base_get_msg_frame(ioc, smid); |
4497 | memset(ack_request, 0, sizeof(Mpi2EventAckRequest_t)); |
4498 | ack_request->Function = MPI2_FUNCTION_EVENT_ACK; |
4499 | ack_request->Event = event; |
4500 | ack_request->EventContext = event_context; |
4501 | ack_request->VF_ID = 0; /* TODO */ |
4502 | ack_request->VP_ID = 0; |
4503 | ioc->put_smid_default(ioc, smid); |
4504 | } |
4505 | |
4506 | /** |
4507 | * _scsih_issue_delayed_sas_io_unit_ctrl - issue delayed |
4508 | * sas_io_unit_ctrl messages |
4509 | * @ioc: per adapter object |
4510 | * @smid: system request message index |
4511 | * @handle: device handle |
4512 | * |
4513 | * Context - processed in interrupt context. |
4514 | */ |
4515 | static void |
4516 | _scsih_issue_delayed_sas_io_unit_ctrl(struct MPT3SAS_ADAPTER *ioc, |
4517 | u16 smid, u16 handle) |
4518 | { |
4519 | Mpi2SasIoUnitControlRequest_t *mpi_request; |
4520 | u32 ioc_state; |
4521 | int i = smid - ioc->internal_smid; |
4522 | unsigned long flags; |
4523 | |
4524 | if (ioc->remove_host) { |
4525 | dewtprintk(ioc, |
4526 | ioc_info(ioc, "%s: host has been removed\n" , |
4527 | __func__)); |
4528 | return; |
4529 | } else if (ioc->pci_error_recovery) { |
4530 | dewtprintk(ioc, |
4531 | ioc_info(ioc, "%s: host in pci error recovery\n" , |
4532 | __func__)); |
4533 | return; |
4534 | } |
4535 | ioc_state = mpt3sas_base_get_iocstate(ioc, cooked: 1); |
4536 | if (ioc_state != MPI2_IOC_STATE_OPERATIONAL) { |
4537 | dewtprintk(ioc, |
4538 | ioc_info(ioc, "%s: host is not operational\n" , |
4539 | __func__)); |
4540 | return; |
4541 | } |
4542 | |
4543 | /* Without releasing the smid just update the |
4544 | * call back index and reuse the same smid for |
4545 | * processing this delayed request |
4546 | */ |
4547 | spin_lock_irqsave(&ioc->scsi_lookup_lock, flags); |
4548 | ioc->internal_lookup[i].cb_idx = ioc->tm_sas_control_cb_idx; |
4549 | spin_unlock_irqrestore(lock: &ioc->scsi_lookup_lock, flags); |
4550 | |
4551 | dewtprintk(ioc, |
4552 | ioc_info(ioc, "sc_send:handle(0x%04x), (open), smid(%d), cb(%d)\n" , |
4553 | handle, smid, ioc->tm_sas_control_cb_idx)); |
4554 | mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); |
4555 | memset(mpi_request, 0, sizeof(Mpi2SasIoUnitControlRequest_t)); |
4556 | mpi_request->Function = MPI2_FUNCTION_SAS_IO_UNIT_CONTROL; |
4557 | mpi_request->Operation = MPI2_SAS_OP_REMOVE_DEVICE; |
4558 | mpi_request->DevHandle = cpu_to_le16(handle); |
4559 | ioc->put_smid_default(ioc, smid); |
4560 | } |
4561 | |
4562 | /** |
4563 | * mpt3sas_check_for_pending_internal_cmds - check for pending internal messages |
4564 | * @ioc: per adapter object |
4565 | * @smid: system request message index |
4566 | * |
4567 | * Context: Executed in interrupt context |
4568 | * |
4569 | * This will check delayed internal messages list, and process the |
4570 | * next request. |
4571 | * |
4572 | * Return: 1 meaning mf should be freed from _base_interrupt |
4573 | * 0 means the mf is freed from this function. |
4574 | */ |
4575 | u8 |
4576 | mpt3sas_check_for_pending_internal_cmds(struct MPT3SAS_ADAPTER *ioc, u16 smid) |
4577 | { |
4578 | struct _sc_list *delayed_sc; |
4579 | struct _event_ack_list *delayed_event_ack; |
4580 | |
4581 | if (!list_empty(head: &ioc->delayed_event_ack_list)) { |
4582 | delayed_event_ack = list_entry(ioc->delayed_event_ack_list.next, |
4583 | struct _event_ack_list, list); |
4584 | _scsih_issue_delayed_event_ack(ioc, smid, |
4585 | event: delayed_event_ack->Event, event_context: delayed_event_ack->EventContext); |
4586 | list_del(entry: &delayed_event_ack->list); |
4587 | kfree(objp: delayed_event_ack); |
4588 | return 0; |
4589 | } |
4590 | |
4591 | if (!list_empty(head: &ioc->delayed_sc_list)) { |
4592 | delayed_sc = list_entry(ioc->delayed_sc_list.next, |
4593 | struct _sc_list, list); |
4594 | _scsih_issue_delayed_sas_io_unit_ctrl(ioc, smid, |
4595 | handle: delayed_sc->handle); |
4596 | list_del(entry: &delayed_sc->list); |
4597 | kfree(objp: delayed_sc); |
4598 | return 0; |
4599 | } |
4600 | return 1; |
4601 | } |
4602 | |
4603 | /** |
4604 | * _scsih_check_for_pending_tm - check for pending task management |
4605 | * @ioc: per adapter object |
4606 | * @smid: system request message index |
4607 | * |
4608 | * This will check delayed target reset list, and feed the |
4609 | * next reqeust. |
4610 | * |
4611 | * Return: 1 meaning mf should be freed from _base_interrupt |
4612 | * 0 means the mf is freed from this function. |
4613 | */ |
4614 | static u8 |
4615 | _scsih_check_for_pending_tm(struct MPT3SAS_ADAPTER *ioc, u16 smid) |
4616 | { |
4617 | struct _tr_list *delayed_tr; |
4618 | |
4619 | if (!list_empty(head: &ioc->delayed_tr_volume_list)) { |
4620 | delayed_tr = list_entry(ioc->delayed_tr_volume_list.next, |
4621 | struct _tr_list, list); |
4622 | mpt3sas_base_free_smid(ioc, smid); |
4623 | _scsih_tm_tr_volume_send(ioc, handle: delayed_tr->handle); |
4624 | list_del(entry: &delayed_tr->list); |
4625 | kfree(objp: delayed_tr); |
4626 | return 0; |
4627 | } |
4628 | |
4629 | if (!list_empty(head: &ioc->delayed_tr_list)) { |
4630 | delayed_tr = list_entry(ioc->delayed_tr_list.next, |
4631 | struct _tr_list, list); |
4632 | mpt3sas_base_free_smid(ioc, smid); |
4633 | _scsih_tm_tr_send(ioc, handle: delayed_tr->handle); |
4634 | list_del(entry: &delayed_tr->list); |
4635 | kfree(objp: delayed_tr); |
4636 | return 0; |
4637 | } |
4638 | |
4639 | return 1; |
4640 | } |
4641 | |
4642 | /** |
4643 | * _scsih_check_topo_delete_events - sanity check on topo events |
4644 | * @ioc: per adapter object |
4645 | * @event_data: the event data payload |
4646 | * |
4647 | * This routine added to better handle cable breaker. |
4648 | * |
4649 | * This handles the case where driver receives multiple expander |
4650 | * add and delete events in a single shot. When there is a delete event |
4651 | * the routine will void any pending add events waiting in the event queue. |
4652 | */ |
4653 | static void |
4654 | _scsih_check_topo_delete_events(struct MPT3SAS_ADAPTER *ioc, |
4655 | Mpi2EventDataSasTopologyChangeList_t *event_data) |
4656 | { |
4657 | struct fw_event_work *fw_event; |
4658 | Mpi2EventDataSasTopologyChangeList_t *local_event_data; |
4659 | u16 expander_handle; |
4660 | struct _sas_node *sas_expander; |
4661 | unsigned long flags; |
4662 | int i, reason_code; |
4663 | u16 handle; |
4664 | |
4665 | for (i = 0 ; i < event_data->NumEntries; i++) { |
4666 | handle = le16_to_cpu(event_data->PHY[i].AttachedDevHandle); |
4667 | if (!handle) |
4668 | continue; |
4669 | reason_code = event_data->PHY[i].PhyStatus & |
4670 | MPI2_EVENT_SAS_TOPO_RC_MASK; |
4671 | if (reason_code == MPI2_EVENT_SAS_TOPO_RC_TARG_NOT_RESPONDING) |
4672 | _scsih_tm_tr_send(ioc, handle); |
4673 | } |
4674 | |
4675 | expander_handle = le16_to_cpu(event_data->ExpanderDevHandle); |
4676 | if (expander_handle < ioc->sas_hba.num_phys) { |
4677 | _scsih_block_io_to_children_attached_directly(ioc, event_data); |
4678 | return; |
4679 | } |
4680 | if (event_data->ExpStatus == |
4681 | MPI2_EVENT_SAS_TOPO_ES_DELAY_NOT_RESPONDING) { |
4682 | /* put expander attached devices into blocking state */ |
4683 | spin_lock_irqsave(&ioc->sas_node_lock, flags); |
4684 | sas_expander = mpt3sas_scsih_expander_find_by_handle(ioc, |
4685 | handle: expander_handle); |
4686 | _scsih_block_io_to_children_attached_to_ex(ioc, sas_expander); |
4687 | spin_unlock_irqrestore(lock: &ioc->sas_node_lock, flags); |
4688 | do { |
4689 | handle = find_first_bit(addr: ioc->blocking_handles, |
4690 | size: ioc->facts.MaxDevHandle); |
4691 | if (handle < ioc->facts.MaxDevHandle) |
4692 | _scsih_block_io_device(ioc, handle); |
4693 | } while (test_and_clear_bit(nr: handle, addr: ioc->blocking_handles)); |
4694 | } else if (event_data->ExpStatus == MPI2_EVENT_SAS_TOPO_ES_RESPONDING) |
4695 | _scsih_block_io_to_children_attached_directly(ioc, event_data); |
4696 | |
4697 | if (event_data->ExpStatus != MPI2_EVENT_SAS_TOPO_ES_NOT_RESPONDING) |
4698 | return; |
4699 | |
4700 | /* mark ignore flag for pending events */ |
4701 | spin_lock_irqsave(&ioc->fw_event_lock, flags); |
4702 | list_for_each_entry(fw_event, &ioc->fw_event_list, list) { |
4703 | if (fw_event->event != MPI2_EVENT_SAS_TOPOLOGY_CHANGE_LIST || |
4704 | fw_event->ignore) |
4705 | continue; |
4706 | local_event_data = (Mpi2EventDataSasTopologyChangeList_t *) |
4707 | fw_event->event_data; |
4708 | if (local_event_data->ExpStatus == |
4709 | MPI2_EVENT_SAS_TOPO_ES_ADDED || |
4710 | local_event_data->ExpStatus == |
4711 | MPI2_EVENT_SAS_TOPO_ES_RESPONDING) { |
4712 | if (le16_to_cpu(local_event_data->ExpanderDevHandle) == |
4713 | expander_handle) { |
4714 | dewtprintk(ioc, |
4715 | ioc_info(ioc, "setting ignoring flag\n" )); |
4716 | fw_event->ignore = 1; |
4717 | } |
4718 | } |
4719 | } |
4720 | spin_unlock_irqrestore(lock: &ioc->fw_event_lock, flags); |
4721 | } |
4722 | |
4723 | /** |
4724 | * _scsih_check_pcie_topo_remove_events - sanity check on topo |
4725 | * events |
4726 | * @ioc: per adapter object |
4727 | * @event_data: the event data payload |
4728 | * |
4729 | * This handles the case where driver receives multiple switch |
4730 | * or device add and delete events in a single shot. When there |
4731 | * is a delete event the routine will void any pending add |
4732 | * events waiting in the event queue. |
4733 | */ |
4734 | static void |
4735 | _scsih_check_pcie_topo_remove_events(struct MPT3SAS_ADAPTER *ioc, |
4736 | Mpi26EventDataPCIeTopologyChangeList_t *event_data) |
4737 | { |
4738 | struct fw_event_work *fw_event; |
4739 | Mpi26EventDataPCIeTopologyChangeList_t *local_event_data; |
4740 | unsigned long flags; |
4741 | int i, reason_code; |
4742 | u16 handle, switch_handle; |
4743 | |
4744 | for (i = 0; i < event_data->NumEntries; i++) { |
4745 | handle = |
4746 | le16_to_cpu(event_data->PortEntry[i].AttachedDevHandle); |
4747 | if (!handle) |
4748 | continue; |
4749 | reason_code = event_data->PortEntry[i].PortStatus; |
4750 | if (reason_code == MPI26_EVENT_PCIE_TOPO_PS_NOT_RESPONDING) |
4751 | _scsih_tm_tr_send(ioc, handle); |
4752 | } |
4753 | |
4754 | switch_handle = le16_to_cpu(event_data->SwitchDevHandle); |
4755 | if (!switch_handle) { |
4756 | _scsih_block_io_to_pcie_children_attached_directly( |
4757 | ioc, event_data); |
4758 | return; |
4759 | } |
4760 | /* TODO We are not supporting cascaded PCIe Switch removal yet*/ |
4761 | if ((event_data->SwitchStatus |
4762 | == MPI26_EVENT_PCIE_TOPO_SS_DELAY_NOT_RESPONDING) || |
4763 | (event_data->SwitchStatus == |
4764 | MPI26_EVENT_PCIE_TOPO_SS_RESPONDING)) |
4765 | _scsih_block_io_to_pcie_children_attached_directly( |
4766 | ioc, event_data); |
4767 | |
4768 | if (event_data->SwitchStatus != MPI2_EVENT_SAS_TOPO_ES_NOT_RESPONDING) |
4769 | return; |
4770 | |
4771 | /* mark ignore flag for pending events */ |
4772 | spin_lock_irqsave(&ioc->fw_event_lock, flags); |
4773 | list_for_each_entry(fw_event, &ioc->fw_event_list, list) { |
4774 | if (fw_event->event != MPI2_EVENT_PCIE_TOPOLOGY_CHANGE_LIST || |
4775 | fw_event->ignore) |
4776 | continue; |
4777 | local_event_data = |
4778 | (Mpi26EventDataPCIeTopologyChangeList_t *) |
4779 | fw_event->event_data; |
4780 | if (local_event_data->SwitchStatus == |
4781 | MPI2_EVENT_SAS_TOPO_ES_ADDED || |
4782 | local_event_data->SwitchStatus == |
4783 | MPI2_EVENT_SAS_TOPO_ES_RESPONDING) { |
4784 | if (le16_to_cpu(local_event_data->SwitchDevHandle) == |
4785 | switch_handle) { |
4786 | dewtprintk(ioc, |
4787 | ioc_info(ioc, "setting ignoring flag for switch event\n" )); |
4788 | fw_event->ignore = 1; |
4789 | } |
4790 | } |
4791 | } |
4792 | spin_unlock_irqrestore(lock: &ioc->fw_event_lock, flags); |
4793 | } |
4794 | |
4795 | /** |
4796 | * _scsih_set_volume_delete_flag - setting volume delete flag |
4797 | * @ioc: per adapter object |
4798 | * @handle: device handle |
4799 | * |
4800 | * This returns nothing. |
4801 | */ |
4802 | static void |
4803 | _scsih_set_volume_delete_flag(struct MPT3SAS_ADAPTER *ioc, u16 handle) |
4804 | { |
4805 | struct _raid_device *raid_device; |
4806 | struct MPT3SAS_TARGET *sas_target_priv_data; |
4807 | unsigned long flags; |
4808 | |
4809 | spin_lock_irqsave(&ioc->raid_device_lock, flags); |
4810 | raid_device = mpt3sas_raid_device_find_by_handle(ioc, handle); |
4811 | if (raid_device && raid_device->starget && |
4812 | raid_device->starget->hostdata) { |
4813 | sas_target_priv_data = |
4814 | raid_device->starget->hostdata; |
4815 | sas_target_priv_data->deleted = 1; |
4816 | dewtprintk(ioc, |
4817 | ioc_info(ioc, "setting delete flag: handle(0x%04x), wwid(0x%016llx)\n" , |
4818 | handle, (u64)raid_device->wwid)); |
4819 | } |
4820 | spin_unlock_irqrestore(lock: &ioc->raid_device_lock, flags); |
4821 | } |
4822 | |
4823 | /** |
4824 | * _scsih_set_volume_handle_for_tr - set handle for target reset to volume |
4825 | * @handle: input handle |
4826 | * @a: handle for volume a |
4827 | * @b: handle for volume b |
4828 | * |
4829 | * IR firmware only supports two raid volumes. The purpose of this |
4830 | * routine is to set the volume handle in either a or b. When the given |
4831 | * input handle is non-zero, or when a and b have not been set before. |
4832 | */ |
4833 | static void |
4834 | _scsih_set_volume_handle_for_tr(u16 handle, u16 *a, u16 *b) |
4835 | { |
4836 | if (!handle || handle == *a || handle == *b) |
4837 | return; |
4838 | if (!*a) |
4839 | *a = handle; |
4840 | else if (!*b) |
4841 | *b = handle; |
4842 | } |
4843 | |
4844 | /** |
4845 | * _scsih_check_ir_config_unhide_events - check for UNHIDE events |
4846 | * @ioc: per adapter object |
4847 | * @event_data: the event data payload |
4848 | * Context: interrupt time. |
4849 | * |
4850 | * This routine will send target reset to volume, followed by target |
4851 | * resets to the PDs. This is called when a PD has been removed, or |
4852 | * volume has been deleted or removed. When the target reset is sent |
4853 | * to volume, the PD target resets need to be queued to start upon |
4854 | * completion of the volume target reset. |
4855 | */ |
4856 | static void |
4857 | _scsih_check_ir_config_unhide_events(struct MPT3SAS_ADAPTER *ioc, |
4858 | Mpi2EventDataIrConfigChangeList_t *event_data) |
4859 | { |
4860 | Mpi2EventIrConfigElement_t *element; |
4861 | int i; |
4862 | u16 handle, volume_handle, a, b; |
4863 | struct _tr_list *delayed_tr; |
4864 | |
4865 | a = 0; |
4866 | b = 0; |
4867 | |
4868 | if (ioc->is_warpdrive) |
4869 | return; |
4870 | |
4871 | /* Volume Resets for Deleted or Removed */ |
4872 | element = (Mpi2EventIrConfigElement_t *)&event_data->ConfigElement[0]; |
4873 | for (i = 0; i < event_data->NumElements; i++, element++) { |
4874 | if (le32_to_cpu(event_data->Flags) & |
4875 | MPI2_EVENT_IR_CHANGE_FLAGS_FOREIGN_CONFIG) |
4876 | continue; |
4877 | if (element->ReasonCode == |
4878 | MPI2_EVENT_IR_CHANGE_RC_VOLUME_DELETED || |
4879 | element->ReasonCode == |
4880 | MPI2_EVENT_IR_CHANGE_RC_REMOVED) { |
4881 | volume_handle = le16_to_cpu(element->VolDevHandle); |
4882 | _scsih_set_volume_delete_flag(ioc, handle: volume_handle); |
4883 | _scsih_set_volume_handle_for_tr(handle: volume_handle, a: &a, b: &b); |
4884 | } |
4885 | } |
4886 | |
4887 | /* Volume Resets for UNHIDE events */ |
4888 | element = (Mpi2EventIrConfigElement_t *)&event_data->ConfigElement[0]; |
4889 | for (i = 0; i < event_data->NumElements; i++, element++) { |
4890 | if (le32_to_cpu(event_data->Flags) & |
4891 | MPI2_EVENT_IR_CHANGE_FLAGS_FOREIGN_CONFIG) |
4892 | continue; |
4893 | if (element->ReasonCode == MPI2_EVENT_IR_CHANGE_RC_UNHIDE) { |
4894 | volume_handle = le16_to_cpu(element->VolDevHandle); |
4895 | _scsih_set_volume_handle_for_tr(handle: volume_handle, a: &a, b: &b); |
4896 | } |
4897 | } |
4898 | |
4899 | if (a) |
4900 | _scsih_tm_tr_volume_send(ioc, handle: a); |
4901 | if (b) |
4902 | _scsih_tm_tr_volume_send(ioc, handle: b); |
4903 | |
4904 | /* PD target resets */ |
4905 | element = (Mpi2EventIrConfigElement_t *)&event_data->ConfigElement[0]; |
4906 | for (i = 0; i < event_data->NumElements; i++, element++) { |
4907 | if (element->ReasonCode != MPI2_EVENT_IR_CHANGE_RC_UNHIDE) |
4908 | continue; |
4909 | handle = le16_to_cpu(element->PhysDiskDevHandle); |
4910 | volume_handle = le16_to_cpu(element->VolDevHandle); |
4911 | clear_bit(nr: handle, addr: ioc->pd_handles); |
4912 | if (!volume_handle) |
4913 | _scsih_tm_tr_send(ioc, handle); |
4914 | else if (volume_handle == a || volume_handle == b) { |
4915 | delayed_tr = kzalloc(size: sizeof(*delayed_tr), GFP_ATOMIC); |
4916 | BUG_ON(!delayed_tr); |
4917 | INIT_LIST_HEAD(list: &delayed_tr->list); |
4918 | delayed_tr->handle = handle; |
4919 | list_add_tail(new: &delayed_tr->list, head: &ioc->delayed_tr_list); |
4920 | dewtprintk(ioc, |
4921 | ioc_info(ioc, "DELAYED:tr:handle(0x%04x), (open)\n" , |
4922 | handle)); |
4923 | } else |
4924 | _scsih_tm_tr_send(ioc, handle); |
4925 | } |
4926 | } |
4927 | |
4928 | |
4929 | /** |
4930 | * _scsih_check_volume_delete_events - set delete flag for volumes |
4931 | * @ioc: per adapter object |
4932 | * @event_data: the event data payload |
4933 | * Context: interrupt time. |
4934 | * |
4935 | * This will handle the case when the cable connected to entire volume is |
4936 | * pulled. We will take care of setting the deleted flag so normal IO will |
4937 | * not be sent. |
4938 | */ |
4939 | static void |
4940 | _scsih_check_volume_delete_events(struct MPT3SAS_ADAPTER *ioc, |
4941 | Mpi2EventDataIrVolume_t *event_data) |
4942 | { |
4943 | u32 state; |
4944 | |
4945 | if (event_data->ReasonCode != MPI2_EVENT_IR_VOLUME_RC_STATE_CHANGED) |
4946 | return; |
4947 | state = le32_to_cpu(event_data->NewValue); |
4948 | if (state == MPI2_RAID_VOL_STATE_MISSING || state == |
4949 | MPI2_RAID_VOL_STATE_FAILED) |
4950 | _scsih_set_volume_delete_flag(ioc, |
4951 | le16_to_cpu(event_data->VolDevHandle)); |
4952 | } |
4953 | |
4954 | /** |
4955 | * _scsih_temp_threshold_events - display temperature threshold exceeded events |
4956 | * @ioc: per adapter object |
4957 | * @event_data: the temp threshold event data |
4958 | * Context: interrupt time. |
4959 | */ |
4960 | static void |
4961 | _scsih_temp_threshold_events(struct MPT3SAS_ADAPTER *ioc, |
4962 | Mpi2EventDataTemperature_t *event_data) |
4963 | { |
4964 | u32 doorbell; |
4965 | if (ioc->temp_sensors_count >= event_data->SensorNum) { |
4966 | ioc_err(ioc, "Temperature Threshold flags %s%s%s%s exceeded for Sensor: %d !!!\n" , |
4967 | le16_to_cpu(event_data->Status) & 0x1 ? "0 " : " " , |
4968 | le16_to_cpu(event_data->Status) & 0x2 ? "1 " : " " , |
4969 | le16_to_cpu(event_data->Status) & 0x4 ? "2 " : " " , |
4970 | le16_to_cpu(event_data->Status) & 0x8 ? "3 " : " " , |
4971 | event_data->SensorNum); |
4972 | ioc_err(ioc, "Current Temp In Celsius: %d\n" , |
4973 | event_data->CurrentTemperature); |
4974 | if (ioc->hba_mpi_version_belonged != MPI2_VERSION) { |
4975 | doorbell = mpt3sas_base_get_iocstate(ioc, cooked: 0); |
4976 | if ((doorbell & MPI2_IOC_STATE_MASK) == |
4977 | MPI2_IOC_STATE_FAULT) { |
4978 | mpt3sas_print_fault_code(ioc, |
4979 | doorbell & MPI2_DOORBELL_DATA_MASK); |
4980 | } else if ((doorbell & MPI2_IOC_STATE_MASK) == |
4981 | MPI2_IOC_STATE_COREDUMP) { |
4982 | mpt3sas_print_coredump_info(ioc, |
4983 | doorbell & MPI2_DOORBELL_DATA_MASK); |
4984 | } |
4985 | } |
4986 | } |
4987 | } |
4988 | |
4989 | static int _scsih_set_satl_pending(struct scsi_cmnd *scmd, bool pending) |
4990 | { |
4991 | struct MPT3SAS_DEVICE *priv = scmd->device->hostdata; |
4992 | |
4993 | if (scmd->cmnd[0] != ATA_12 && scmd->cmnd[0] != ATA_16) |
4994 | return 0; |
4995 | |
4996 | if (pending) |
4997 | return test_and_set_bit(nr: 0, addr: &priv->ata_command_pending); |
4998 | |
4999 | clear_bit(nr: 0, addr: &priv->ata_command_pending); |
5000 | return 0; |
5001 | } |
5002 | |
5003 | /** |
5004 | * _scsih_flush_running_cmds - completing outstanding commands. |
5005 | * @ioc: per adapter object |
5006 | * |
5007 | * The flushing out of all pending scmd commands following host reset, |
5008 | * where all IO is dropped to the floor. |
5009 | */ |
5010 | static void |
5011 | _scsih_flush_running_cmds(struct MPT3SAS_ADAPTER *ioc) |
5012 | { |
5013 | struct scsi_cmnd *scmd; |
5014 | struct scsiio_tracker *st; |
5015 | u16 smid; |
5016 | int count = 0; |
5017 | |
5018 | for (smid = 1; smid <= ioc->scsiio_depth; smid++) { |
5019 | scmd = mpt3sas_scsih_scsi_lookup_get(ioc, smid); |
5020 | if (!scmd) |
5021 | continue; |
5022 | count++; |
5023 | _scsih_set_satl_pending(scmd, pending: false); |
5024 | st = scsi_cmd_priv(cmd: scmd); |
5025 | mpt3sas_base_clear_st(ioc, st); |
5026 | scsi_dma_unmap(cmd: scmd); |
5027 | if (ioc->pci_error_recovery || ioc->remove_host) |
5028 | scmd->result = DID_NO_CONNECT << 16; |
5029 | else |
5030 | scmd->result = DID_RESET << 16; |
5031 | scsi_done(cmd: scmd); |
5032 | } |
5033 | dtmprintk(ioc, ioc_info(ioc, "completing %d cmds\n" , count)); |
5034 | } |
5035 | |
5036 | /** |
5037 | * _scsih_setup_eedp - setup MPI request for EEDP transfer |
5038 | * @ioc: per adapter object |
5039 | * @scmd: pointer to scsi command object |
5040 | * @mpi_request: pointer to the SCSI_IO request message frame |
5041 | * |
5042 | * Supporting protection 1 and 3. |
5043 | */ |
5044 | static void |
5045 | _scsih_setup_eedp(struct MPT3SAS_ADAPTER *ioc, struct scsi_cmnd *scmd, |
5046 | Mpi25SCSIIORequest_t *mpi_request) |
5047 | { |
5048 | u16 eedp_flags; |
5049 | Mpi25SCSIIORequest_t *mpi_request_3v = |
5050 | (Mpi25SCSIIORequest_t *)mpi_request; |
5051 | |
5052 | switch (scsi_get_prot_op(scmd)) { |
5053 | case SCSI_PROT_READ_STRIP: |
5054 | eedp_flags = MPI2_SCSIIO_EEDPFLAGS_CHECK_REMOVE_OP; |
5055 | break; |
5056 | case SCSI_PROT_WRITE_INSERT: |
5057 | eedp_flags = MPI2_SCSIIO_EEDPFLAGS_INSERT_OP; |
5058 | break; |
5059 | default: |
5060 | return; |
5061 | } |
5062 | |
5063 | if (scmd->prot_flags & SCSI_PROT_GUARD_CHECK) |
5064 | eedp_flags |= MPI2_SCSIIO_EEDPFLAGS_CHECK_GUARD; |
5065 | |
5066 | if (scmd->prot_flags & SCSI_PROT_REF_CHECK) |
5067 | eedp_flags |= MPI2_SCSIIO_EEDPFLAGS_CHECK_REFTAG; |
5068 | |
5069 | if (scmd->prot_flags & SCSI_PROT_REF_INCREMENT) { |
5070 | eedp_flags |= MPI2_SCSIIO_EEDPFLAGS_INC_PRI_REFTAG; |
5071 | |
5072 | mpi_request->CDB.EEDP32.PrimaryReferenceTag = |
5073 | cpu_to_be32(scsi_prot_ref_tag(scmd)); |
5074 | } |
5075 | |
5076 | mpi_request_3v->EEDPBlockSize = cpu_to_le16(scsi_prot_interval(scmd)); |
5077 | |
5078 | if (ioc->is_gen35_ioc) |
5079 | eedp_flags |= MPI25_SCSIIO_EEDPFLAGS_APPTAG_DISABLE_MODE; |
5080 | mpi_request->EEDPFlags = cpu_to_le16(eedp_flags); |
5081 | } |
5082 | |
5083 | /** |
5084 | * _scsih_eedp_error_handling - return sense code for EEDP errors |
5085 | * @scmd: pointer to scsi command object |
5086 | * @ioc_status: ioc status |
5087 | */ |
5088 | static void |
5089 | _scsih_eedp_error_handling(struct scsi_cmnd *scmd, u16 ioc_status) |
5090 | { |
5091 | u8 ascq; |
5092 | |
5093 | switch (ioc_status) { |
5094 | case MPI2_IOCSTATUS_EEDP_GUARD_ERROR: |
5095 | ascq = 0x01; |
5096 | break; |
5097 | case MPI2_IOCSTATUS_EEDP_APP_TAG_ERROR: |
5098 | ascq = 0x02; |
5099 | break; |
5100 | case MPI2_IOCSTATUS_EEDP_REF_TAG_ERROR: |
5101 | ascq = 0x03; |
5102 | break; |
5103 | default: |
5104 | ascq = 0x00; |
5105 | break; |
5106 | } |
5107 | scsi_build_sense(scmd, desc: 0, ILLEGAL_REQUEST, asc: 0x10, ascq); |
5108 | set_host_byte(cmd: scmd, status: DID_ABORT); |
5109 | } |
5110 | |
5111 | /** |
5112 | * scsih_qcmd - main scsi request entry point |
5113 | * @shost: SCSI host pointer |
5114 | * @scmd: pointer to scsi command object |
5115 | * |
5116 | * The callback index is set inside `ioc->scsi_io_cb_idx`. |
5117 | * |
5118 | * Return: 0 on success. If there's a failure, return either: |
5119 | * SCSI_MLQUEUE_DEVICE_BUSY if the device queue is full, or |
5120 | * SCSI_MLQUEUE_HOST_BUSY if the entire host queue is full |
5121 | */ |
5122 | static int |
5123 | scsih_qcmd(struct Scsi_Host *shost, struct scsi_cmnd *scmd) |
5124 | { |
5125 | struct MPT3SAS_ADAPTER *ioc = shost_priv(shost); |
5126 | struct MPT3SAS_DEVICE *sas_device_priv_data; |
5127 | struct MPT3SAS_TARGET *sas_target_priv_data; |
5128 | struct _raid_device *raid_device; |
5129 | struct request *rq = scsi_cmd_to_rq(scmd); |
5130 | int class; |
5131 | Mpi25SCSIIORequest_t *mpi_request; |
5132 | struct _pcie_device *pcie_device = NULL; |
5133 | u32 mpi_control; |
5134 | u16 smid; |
5135 | u16 handle; |
5136 | |
5137 | if (ioc->logging_level & MPT_DEBUG_SCSI) |
5138 | scsi_print_command(scmd); |
5139 | |
5140 | sas_device_priv_data = scmd->device->hostdata; |
5141 | if (!sas_device_priv_data || !sas_device_priv_data->sas_target) { |
5142 | scmd->result = DID_NO_CONNECT << 16; |
5143 | scsi_done(cmd: scmd); |
5144 | return 0; |
5145 | } |
5146 | |
5147 | if (!(_scsih_allow_scmd_to_device(ioc, scmd))) { |
5148 | scmd->result = DID_NO_CONNECT << 16; |
5149 | scsi_done(cmd: scmd); |
5150 | return 0; |
5151 | } |
5152 | |
5153 | sas_target_priv_data = sas_device_priv_data->sas_target; |
5154 | |
5155 | /* invalid device handle */ |
5156 | handle = sas_target_priv_data->handle; |
5157 | |
5158 | /* |
5159 | * Avoid error handling escallation when device is disconnected |
5160 | */ |
5161 | if (handle == MPT3SAS_INVALID_DEVICE_HANDLE || sas_device_priv_data->block) { |
5162 | if (scmd->device->host->shost_state == SHOST_RECOVERY && |
5163 | scmd->cmnd[0] == TEST_UNIT_READY) { |
5164 | scsi_build_sense(scmd, desc: 0, UNIT_ATTENTION, asc: 0x29, ascq: 0x07); |
5165 | scsi_done(cmd: scmd); |
5166 | return 0; |
5167 | } |
5168 | } |
5169 | |
5170 | if (handle == MPT3SAS_INVALID_DEVICE_HANDLE) { |
5171 | scmd->result = DID_NO_CONNECT << 16; |
5172 | scsi_done(cmd: scmd); |
5173 | return 0; |
5174 | } |
5175 | |
5176 | |
5177 | if (ioc->shost_recovery || ioc->ioc_link_reset_in_progress) { |
5178 | /* host recovery or link resets sent via IOCTLs */ |
5179 | return SCSI_MLQUEUE_HOST_BUSY; |
5180 | } else if (sas_target_priv_data->deleted) { |
5181 | /* device has been deleted */ |
5182 | scmd->result = DID_NO_CONNECT << 16; |
5183 | scsi_done(cmd: scmd); |
5184 | return 0; |
5185 | } else if (sas_target_priv_data->tm_busy || |
5186 | sas_device_priv_data->block) { |
5187 | /* device busy with task management */ |
5188 | return SCSI_MLQUEUE_DEVICE_BUSY; |
5189 | } |
5190 | |
5191 | /* |
5192 | * Bug work around for firmware SATL handling. The loop |
5193 | * is based on atomic operations and ensures consistency |
5194 | * since we're lockless at this point |
5195 | */ |
5196 | do { |
5197 | if (test_bit(0, &sas_device_priv_data->ata_command_pending)) |
5198 | return SCSI_MLQUEUE_DEVICE_BUSY; |
5199 | } while (_scsih_set_satl_pending(scmd, pending: true)); |
5200 | |
5201 | if (scmd->sc_data_direction == DMA_FROM_DEVICE) |
5202 | mpi_control = MPI2_SCSIIO_CONTROL_READ; |
5203 | else if (scmd->sc_data_direction == DMA_TO_DEVICE) |
5204 | mpi_control = MPI2_SCSIIO_CONTROL_WRITE; |
5205 | else |
5206 | mpi_control = MPI2_SCSIIO_CONTROL_NODATATRANSFER; |
5207 | |
5208 | /* set tags */ |
5209 | mpi_control |= MPI2_SCSIIO_CONTROL_SIMPLEQ; |
5210 | /* NCQ Prio supported, make sure control indicated high priority */ |
5211 | if (sas_device_priv_data->ncq_prio_enable) { |
5212 | class = IOPRIO_PRIO_CLASS(req_get_ioprio(rq)); |
5213 | if (class == IOPRIO_CLASS_RT) |
5214 | mpi_control |= 1 << MPI2_SCSIIO_CONTROL_CMDPRI_SHIFT; |
5215 | } |
5216 | /* Make sure Device is not raid volume. |
5217 | * We do not expose raid functionality to upper layer for warpdrive. |
5218 | */ |
5219 | if (((!ioc->is_warpdrive && !scsih_is_raid(dev: &scmd->device->sdev_gendev)) |
5220 | && !scsih_is_nvme(dev: &scmd->device->sdev_gendev)) |
5221 | && sas_is_tlr_enabled(scmd->device) && scmd->cmd_len != 32) |
5222 | mpi_control |= MPI2_SCSIIO_CONTROL_TLR_ON; |
5223 | |
5224 | smid = mpt3sas_base_get_smid_scsiio(ioc, cb_idx: ioc->scsi_io_cb_idx, scmd); |
5225 | if (!smid) { |
5226 | ioc_err(ioc, "%s: failed obtaining a smid\n" , __func__); |
5227 | _scsih_set_satl_pending(scmd, pending: false); |
5228 | goto out; |
5229 | } |
5230 | mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); |
5231 | memset(mpi_request, 0, ioc->request_sz); |
5232 | _scsih_setup_eedp(ioc, scmd, mpi_request); |
5233 | |
5234 | if (scmd->cmd_len == 32) |
5235 | mpi_control |= 4 << MPI2_SCSIIO_CONTROL_ADDCDBLEN_SHIFT; |
5236 | mpi_request->Function = MPI2_FUNCTION_SCSI_IO_REQUEST; |
5237 | if (sas_device_priv_data->sas_target->flags & |
5238 | MPT_TARGET_FLAGS_RAID_COMPONENT) |
5239 | mpi_request->Function = MPI2_FUNCTION_RAID_SCSI_IO_PASSTHROUGH; |
5240 | else |
5241 | mpi_request->Function = MPI2_FUNCTION_SCSI_IO_REQUEST; |
5242 | mpi_request->DevHandle = cpu_to_le16(handle); |
5243 | mpi_request->DataLength = cpu_to_le32(scsi_bufflen(scmd)); |
5244 | mpi_request->Control = cpu_to_le32(mpi_control); |
5245 | mpi_request->IoFlags = cpu_to_le16(scmd->cmd_len); |
5246 | mpi_request->MsgFlags = MPI2_SCSIIO_MSGFLAGS_SYSTEM_SENSE_ADDR; |
5247 | mpi_request->SenseBufferLength = SCSI_SENSE_BUFFERSIZE; |
5248 | mpi_request->SenseBufferLowAddress = |
5249 | mpt3sas_base_get_sense_buffer_dma(ioc, smid); |
5250 | mpi_request->SGLOffset0 = offsetof(Mpi25SCSIIORequest_t, SGL) / 4; |
5251 | int_to_scsilun(sas_device_priv_data->lun, (struct scsi_lun *) |
5252 | mpi_request->LUN); |
5253 | memcpy(mpi_request->CDB.CDB32, scmd->cmnd, scmd->cmd_len); |
5254 | |
5255 | if (mpi_request->DataLength) { |
5256 | pcie_device = sas_target_priv_data->pcie_dev; |
5257 | if (ioc->build_sg_scmd(ioc, scmd, smid, pcie_device)) { |
5258 | mpt3sas_base_free_smid(ioc, smid); |
5259 | _scsih_set_satl_pending(scmd, pending: false); |
5260 | goto out; |
5261 | } |
5262 | } else |
5263 | ioc->build_zero_len_sge(ioc, &mpi_request->SGL); |
5264 | |
5265 | raid_device = sas_target_priv_data->raid_device; |
5266 | if (raid_device && raid_device->direct_io_enabled) |
5267 | mpt3sas_setup_direct_io(ioc, scmd, |
5268 | raid_device, mpi_request); |
5269 | |
5270 | if (likely(mpi_request->Function == MPI2_FUNCTION_SCSI_IO_REQUEST)) { |
5271 | if (sas_target_priv_data->flags & MPT_TARGET_FASTPATH_IO) { |
5272 | mpi_request->IoFlags = cpu_to_le16(scmd->cmd_len | |
5273 | MPI25_SCSIIO_IOFLAGS_FAST_PATH); |
5274 | ioc->put_smid_fast_path(ioc, smid, handle); |
5275 | } else |
5276 | ioc->put_smid_scsi_io(ioc, smid, |
5277 | le16_to_cpu(mpi_request->DevHandle)); |
5278 | } else |
5279 | ioc->put_smid_default(ioc, smid); |
5280 | return 0; |
5281 | |
5282 | out: |
5283 | return SCSI_MLQUEUE_HOST_BUSY; |
5284 | } |
5285 | |
5286 | /** |
5287 | * _scsih_normalize_sense - normalize descriptor and fixed format sense data |
5288 | * @sense_buffer: sense data returned by target |
5289 | * @data: normalized skey/asc/ascq |
5290 | */ |
5291 | static void |
5292 | _scsih_normalize_sense(char *sense_buffer, struct sense_info *data) |
5293 | { |
5294 | if ((sense_buffer[0] & 0x7F) >= 0x72) { |
5295 | /* descriptor format */ |
5296 | data->skey = sense_buffer[1] & 0x0F; |
5297 | data->asc = sense_buffer[2]; |
5298 | data->ascq = sense_buffer[3]; |
5299 | } else { |
5300 | /* fixed format */ |
5301 | data->skey = sense_buffer[2] & 0x0F; |
5302 | data->asc = sense_buffer[12]; |
5303 | data->ascq = sense_buffer[13]; |
5304 | } |
5305 | } |
5306 | |
5307 | /** |
5308 | * _scsih_scsi_ioc_info - translated non-successful SCSI_IO request |
5309 | * @ioc: per adapter object |
5310 | * @scmd: pointer to scsi command object |
5311 | * @mpi_reply: reply mf payload returned from firmware |
5312 | * @smid: ? |
5313 | * |
5314 | * scsi_status - SCSI Status code returned from target device |
5315 | * scsi_state - state info associated with SCSI_IO determined by ioc |
5316 | * ioc_status - ioc supplied status info |
5317 | */ |
5318 | static void |
5319 | _scsih_scsi_ioc_info(struct MPT3SAS_ADAPTER *ioc, struct scsi_cmnd *scmd, |
5320 | Mpi2SCSIIOReply_t *mpi_reply, u16 smid) |
5321 | { |
5322 | u32 response_info; |
5323 | u8 *response_bytes; |
5324 | u16 ioc_status = le16_to_cpu(mpi_reply->IOCStatus) & |
5325 | MPI2_IOCSTATUS_MASK; |
5326 | u8 scsi_state = mpi_reply->SCSIState; |
5327 | u8 scsi_status = mpi_reply->SCSIStatus; |
5328 | char *desc_ioc_state = NULL; |
5329 | char *desc_scsi_status = NULL; |
5330 | char *desc_scsi_state = ioc->tmp_string; |
5331 | u32 log_info = le32_to_cpu(mpi_reply->IOCLogInfo); |
5332 | struct _sas_device *sas_device = NULL; |
5333 | struct _pcie_device *pcie_device = NULL; |
5334 | struct scsi_target *starget = scmd->device->sdev_target; |
5335 | struct MPT3SAS_TARGET *priv_target = starget->hostdata; |
5336 | char *device_str = NULL; |
5337 | |
5338 | if (!priv_target) |
5339 | return; |
5340 | if (ioc->hide_ir_msg) |
5341 | device_str = "WarpDrive" ; |
5342 | else |
5343 | device_str = "volume" ; |
5344 | |
5345 | if (log_info == 0x31170000) |
5346 | return; |
5347 | |
5348 | switch (ioc_status) { |
5349 | case MPI2_IOCSTATUS_SUCCESS: |
5350 | desc_ioc_state = "success" ; |
5351 | break; |
5352 | case MPI2_IOCSTATUS_INVALID_FUNCTION: |
5353 | desc_ioc_state = "invalid function" ; |
5354 | break; |
5355 | case MPI2_IOCSTATUS_SCSI_RECOVERED_ERROR: |
5356 | desc_ioc_state = "scsi recovered error" ; |
5357 | break; |
5358 | case MPI2_IOCSTATUS_SCSI_INVALID_DEVHANDLE: |
5359 | desc_ioc_state = "scsi invalid dev handle" ; |
5360 | break; |
5361 | case MPI2_IOCSTATUS_SCSI_DEVICE_NOT_THERE: |
5362 | desc_ioc_state = "scsi device not there" ; |
5363 | break; |
5364 | case MPI2_IOCSTATUS_SCSI_DATA_OVERRUN: |
5365 | desc_ioc_state = "scsi data overrun" ; |
5366 | break; |
5367 | case MPI2_IOCSTATUS_SCSI_DATA_UNDERRUN: |
5368 | desc_ioc_state = "scsi data underrun" ; |
5369 | break; |
5370 | case MPI2_IOCSTATUS_SCSI_IO_DATA_ERROR: |
5371 | desc_ioc_state = "scsi io data error" ; |
5372 | break; |
5373 | case MPI2_IOCSTATUS_SCSI_PROTOCOL_ERROR: |
5374 | desc_ioc_state = "scsi protocol error" ; |
5375 | break; |
5376 | case MPI2_IOCSTATUS_SCSI_TASK_TERMINATED: |
5377 | desc_ioc_state = "scsi task terminated" ; |
5378 | break; |
5379 | case MPI2_IOCSTATUS_SCSI_RESIDUAL_MISMATCH: |
5380 | desc_ioc_state = "scsi residual mismatch" ; |
5381 | break; |
5382 | case MPI2_IOCSTATUS_SCSI_TASK_MGMT_FAILED: |
5383 | desc_ioc_state = "scsi task mgmt failed" ; |
5384 | break; |
5385 | case MPI2_IOCSTATUS_SCSI_IOC_TERMINATED: |
5386 | desc_ioc_state = "scsi ioc terminated" ; |
5387 | break; |
5388 | case MPI2_IOCSTATUS_SCSI_EXT_TERMINATED: |
5389 | desc_ioc_state = "scsi ext terminated" ; |
5390 | break; |
5391 | case MPI2_IOCSTATUS_EEDP_GUARD_ERROR: |
5392 | desc_ioc_state = "eedp guard error" ; |
5393 | break; |
5394 | case MPI2_IOCSTATUS_EEDP_REF_TAG_ERROR: |
5395 | desc_ioc_state = "eedp ref tag error" ; |
5396 | break; |
5397 | case MPI2_IOCSTATUS_EEDP_APP_TAG_ERROR: |
5398 | desc_ioc_state = "eedp app tag error" ; |
5399 | break; |
5400 | case MPI2_IOCSTATUS_INSUFFICIENT_POWER: |
5401 | desc_ioc_state = "insufficient power" ; |
5402 | break; |
5403 | default: |
5404 | desc_ioc_state = "unknown" ; |
5405 | break; |
5406 | } |
5407 | |
5408 | switch (scsi_status) { |
5409 | case MPI2_SCSI_STATUS_GOOD: |
5410 | desc_scsi_status = "good" ; |
5411 | break; |
5412 | case MPI2_SCSI_STATUS_CHECK_CONDITION: |
5413 | desc_scsi_status = "check condition" ; |
5414 | break; |
5415 | case MPI2_SCSI_STATUS_CONDITION_MET: |
5416 | desc_scsi_status = "condition met" ; |
5417 | break; |
5418 | case MPI2_SCSI_STATUS_BUSY: |
5419 | desc_scsi_status = "busy" ; |
5420 | break; |
5421 | case MPI2_SCSI_STATUS_INTERMEDIATE: |
5422 | desc_scsi_status = "intermediate" ; |
5423 | break; |
5424 | case MPI2_SCSI_STATUS_INTERMEDIATE_CONDMET: |
5425 | desc_scsi_status = "intermediate condmet" ; |
5426 | break; |
5427 | case MPI2_SCSI_STATUS_RESERVATION_CONFLICT: |
5428 | desc_scsi_status = "reservation conflict" ; |
5429 | break; |
5430 | case MPI2_SCSI_STATUS_COMMAND_TERMINATED: |
5431 | desc_scsi_status = "command terminated" ; |
5432 | break; |
5433 | case MPI2_SCSI_STATUS_TASK_SET_FULL: |
5434 | desc_scsi_status = "task set full" ; |
5435 | break; |
5436 | case MPI2_SCSI_STATUS_ACA_ACTIVE: |
5437 | desc_scsi_status = "aca active" ; |
5438 | break; |
5439 | case MPI2_SCSI_STATUS_TASK_ABORTED: |
5440 | desc_scsi_status = "task aborted" ; |
5441 | break; |
5442 | default: |
5443 | desc_scsi_status = "unknown" ; |
5444 | break; |
5445 | } |
5446 | |
5447 | desc_scsi_state[0] = '\0'; |
5448 | if (!scsi_state) |
5449 | desc_scsi_state = " " ; |
5450 | if (scsi_state & MPI2_SCSI_STATE_RESPONSE_INFO_VALID) |
5451 | strcat(p: desc_scsi_state, q: "response info " ); |
5452 | if (scsi_state & MPI2_SCSI_STATE_TERMINATED) |
5453 | strcat(p: desc_scsi_state, q: "state terminated " ); |
5454 | if (scsi_state & MPI2_SCSI_STATE_NO_SCSI_STATUS) |
5455 | strcat(p: desc_scsi_state, q: "no status " ); |
5456 | if (scsi_state & MPI2_SCSI_STATE_AUTOSENSE_FAILED) |
5457 | strcat(p: desc_scsi_state, q: "autosense failed " ); |
5458 | if (scsi_state & MPI2_SCSI_STATE_AUTOSENSE_VALID) |
5459 | strcat(p: desc_scsi_state, q: "autosense valid " ); |
5460 | |
5461 | scsi_print_command(scmd); |
5462 | |
5463 | if (priv_target->flags & MPT_TARGET_FLAGS_VOLUME) { |
5464 | ioc_warn(ioc, "\t%s wwid(0x%016llx)\n" , |
5465 | device_str, (u64)priv_target->sas_address); |
5466 | } else if (priv_target->flags & MPT_TARGET_FLAGS_PCIE_DEVICE) { |
5467 | pcie_device = mpt3sas_get_pdev_from_target(ioc, tgt_priv: priv_target); |
5468 | if (pcie_device) { |
5469 | ioc_info(ioc, "\twwid(0x%016llx), port(%d)\n" , |
5470 | (u64)pcie_device->wwid, pcie_device->port_num); |
5471 | if (pcie_device->enclosure_handle != 0) |
5472 | ioc_info(ioc, "\tenclosure logical id(0x%016llx), slot(%d)\n" , |
5473 | (u64)pcie_device->enclosure_logical_id, |
5474 | pcie_device->slot); |
5475 | if (pcie_device->connector_name[0]) |
5476 | ioc_info(ioc, "\tenclosure level(0x%04x), connector name( %s)\n" , |
5477 | pcie_device->enclosure_level, |
5478 | pcie_device->connector_name); |
5479 | pcie_device_put(p: pcie_device); |
5480 | } |
5481 | } else { |
5482 | sas_device = mpt3sas_get_sdev_from_target(ioc, tgt_priv: priv_target); |
5483 | if (sas_device) { |
5484 | ioc_warn(ioc, "\tsas_address(0x%016llx), phy(%d)\n" , |
5485 | (u64)sas_device->sas_address, sas_device->phy); |
5486 | |
5487 | _scsih_display_enclosure_chassis_info(ioc, sas_device, |
5488 | NULL, NULL); |
5489 | |
5490 | sas_device_put(s: sas_device); |
5491 | } |
5492 | } |
5493 | |
5494 | ioc_warn(ioc, "\thandle(0x%04x), ioc_status(%s)(0x%04x), smid(%d)\n" , |
5495 | le16_to_cpu(mpi_reply->DevHandle), |
5496 | desc_ioc_state, ioc_status, smid); |
5497 | ioc_warn(ioc, "\trequest_len(%d), underflow(%d), resid(%d)\n" , |
5498 | scsi_bufflen(scmd), scmd->underflow, scsi_get_resid(scmd)); |
5499 | ioc_warn(ioc, "\ttag(%d), transfer_count(%d), sc->result(0x%08x)\n" , |
5500 | le16_to_cpu(mpi_reply->TaskTag), |
5501 | le32_to_cpu(mpi_reply->TransferCount), scmd->result); |
5502 | ioc_warn(ioc, "\tscsi_status(%s)(0x%02x), scsi_state(%s)(0x%02x)\n" , |
5503 | desc_scsi_status, scsi_status, desc_scsi_state, scsi_state); |
5504 | |
5505 | if (scsi_state & MPI2_SCSI_STATE_AUTOSENSE_VALID) { |
5506 | struct sense_info data; |
5507 | _scsih_normalize_sense(sense_buffer: scmd->sense_buffer, data: &data); |
5508 | ioc_warn(ioc, "\t[sense_key,asc,ascq]: [0x%02x,0x%02x,0x%02x], count(%d)\n" , |
5509 | data.skey, data.asc, data.ascq, |
5510 | le32_to_cpu(mpi_reply->SenseCount)); |
5511 | } |
5512 | if (scsi_state & MPI2_SCSI_STATE_RESPONSE_INFO_VALID) { |
5513 | response_info = le32_to_cpu(mpi_reply->ResponseInfo); |
5514 | response_bytes = (u8 *)&response_info; |
5515 | _scsih_response_code(ioc, response_code: response_bytes[0]); |
5516 | } |
5517 | } |
5518 | |
5519 | /** |
5520 | * _scsih_turn_on_pfa_led - illuminate PFA LED |
5521 | * @ioc: per adapter object |
5522 | * @handle: device handle |
5523 | * Context: process |
5524 | */ |
5525 | static void |
5526 | _scsih_turn_on_pfa_led(struct MPT3SAS_ADAPTER *ioc, u16 handle) |
5527 | { |
5528 | Mpi2SepReply_t mpi_reply; |
5529 | Mpi2SepRequest_t mpi_request; |
5530 | struct _sas_device *sas_device; |
5531 | |
5532 | sas_device = mpt3sas_get_sdev_by_handle(ioc, handle); |
5533 | if (!sas_device) |
5534 | return; |
5535 | |
5536 | memset(&mpi_request, 0, sizeof(Mpi2SepRequest_t)); |
5537 | mpi_request.Function = MPI2_FUNCTION_SCSI_ENCLOSURE_PROCESSOR; |
5538 | mpi_request.Action = MPI2_SEP_REQ_ACTION_WRITE_STATUS; |
5539 | mpi_request.SlotStatus = |
5540 | cpu_to_le32(MPI2_SEP_REQ_SLOTSTATUS_PREDICTED_FAULT); |
5541 | mpi_request.DevHandle = cpu_to_le16(handle); |
5542 | mpi_request.Flags = MPI2_SEP_REQ_FLAGS_DEVHANDLE_ADDRESS; |
5543 | if ((mpt3sas_base_scsi_enclosure_processor(ioc, mpi_reply: &mpi_reply, |
5544 | mpi_request: &mpi_request)) != 0) { |
5545 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
5546 | __FILE__, __LINE__, __func__); |
5547 | goto out; |
5548 | } |
5549 | sas_device->pfa_led_on = 1; |
5550 | |
5551 | if (mpi_reply.IOCStatus || mpi_reply.IOCLogInfo) { |
5552 | dewtprintk(ioc, |
5553 | ioc_info(ioc, "enclosure_processor: ioc_status (0x%04x), loginfo(0x%08x)\n" , |
5554 | le16_to_cpu(mpi_reply.IOCStatus), |
5555 | le32_to_cpu(mpi_reply.IOCLogInfo))); |
5556 | goto out; |
5557 | } |
5558 | out: |
5559 | sas_device_put(s: sas_device); |
5560 | } |
5561 | |
5562 | /** |
5563 | * _scsih_turn_off_pfa_led - turn off Fault LED |
5564 | * @ioc: per adapter object |
5565 | * @sas_device: sas device whose PFA LED has to turned off |
5566 | * Context: process |
5567 | */ |
5568 | static void |
5569 | _scsih_turn_off_pfa_led(struct MPT3SAS_ADAPTER *ioc, |
5570 | struct _sas_device *sas_device) |
5571 | { |
5572 | Mpi2SepReply_t mpi_reply; |
5573 | Mpi2SepRequest_t mpi_request; |
5574 | |
5575 | memset(&mpi_request, 0, sizeof(Mpi2SepRequest_t)); |
5576 | mpi_request.Function = MPI2_FUNCTION_SCSI_ENCLOSURE_PROCESSOR; |
5577 | mpi_request.Action = MPI2_SEP_REQ_ACTION_WRITE_STATUS; |
5578 | mpi_request.SlotStatus = 0; |
5579 | mpi_request.Slot = cpu_to_le16(sas_device->slot); |
5580 | mpi_request.DevHandle = 0; |
5581 | mpi_request.EnclosureHandle = cpu_to_le16(sas_device->enclosure_handle); |
5582 | mpi_request.Flags = MPI2_SEP_REQ_FLAGS_ENCLOSURE_SLOT_ADDRESS; |
5583 | if ((mpt3sas_base_scsi_enclosure_processor(ioc, mpi_reply: &mpi_reply, |
5584 | mpi_request: &mpi_request)) != 0) { |
5585 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
5586 | __FILE__, __LINE__, __func__); |
5587 | return; |
5588 | } |
5589 | |
5590 | if (mpi_reply.IOCStatus || mpi_reply.IOCLogInfo) { |
5591 | dewtprintk(ioc, |
5592 | ioc_info(ioc, "enclosure_processor: ioc_status (0x%04x), loginfo(0x%08x)\n" , |
5593 | le16_to_cpu(mpi_reply.IOCStatus), |
5594 | le32_to_cpu(mpi_reply.IOCLogInfo))); |
5595 | return; |
5596 | } |
5597 | } |
5598 | |
5599 | /** |
5600 | * _scsih_send_event_to_turn_on_pfa_led - fire delayed event |
5601 | * @ioc: per adapter object |
5602 | * @handle: device handle |
5603 | * Context: interrupt. |
5604 | */ |
5605 | static void |
5606 | _scsih_send_event_to_turn_on_pfa_led(struct MPT3SAS_ADAPTER *ioc, u16 handle) |
5607 | { |
5608 | struct fw_event_work *fw_event; |
5609 | |
5610 | fw_event = alloc_fw_event_work(len: 0); |
5611 | if (!fw_event) |
5612 | return; |
5613 | fw_event->event = MPT3SAS_TURN_ON_PFA_LED; |
5614 | fw_event->device_handle = handle; |
5615 | fw_event->ioc = ioc; |
5616 | _scsih_fw_event_add(ioc, fw_event); |
5617 | fw_event_work_put(fw_work: fw_event); |
5618 | } |
5619 | |
5620 | /** |
5621 | * _scsih_smart_predicted_fault - process smart errors |
5622 | * @ioc: per adapter object |
5623 | * @handle: device handle |
5624 | * Context: interrupt. |
5625 | */ |
5626 | static void |
5627 | _scsih_smart_predicted_fault(struct MPT3SAS_ADAPTER *ioc, u16 handle) |
5628 | { |
5629 | struct scsi_target *starget; |
5630 | struct MPT3SAS_TARGET *sas_target_priv_data; |
5631 | Mpi2EventNotificationReply_t *event_reply; |
5632 | Mpi2EventDataSasDeviceStatusChange_t *event_data; |
5633 | struct _sas_device *sas_device; |
5634 | ssize_t sz; |
5635 | unsigned long flags; |
5636 | |
5637 | /* only handle non-raid devices */ |
5638 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
5639 | sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle); |
5640 | if (!sas_device) |
5641 | goto out_unlock; |
5642 | |
5643 | starget = sas_device->starget; |
5644 | sas_target_priv_data = starget->hostdata; |
5645 | |
5646 | if ((sas_target_priv_data->flags & MPT_TARGET_FLAGS_RAID_COMPONENT) || |
5647 | ((sas_target_priv_data->flags & MPT_TARGET_FLAGS_VOLUME))) |
5648 | goto out_unlock; |
5649 | |
5650 | _scsih_display_enclosure_chassis_info(NULL, sas_device, NULL, starget); |
5651 | |
5652 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
5653 | |
5654 | if (ioc->pdev->subsystem_vendor == PCI_VENDOR_ID_IBM) |
5655 | _scsih_send_event_to_turn_on_pfa_led(ioc, handle); |
5656 | |
5657 | /* insert into event log */ |
5658 | sz = offsetof(Mpi2EventNotificationReply_t, EventData) + |
5659 | sizeof(Mpi2EventDataSasDeviceStatusChange_t); |
5660 | event_reply = kzalloc(size: sz, GFP_ATOMIC); |
5661 | if (!event_reply) { |
5662 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
5663 | __FILE__, __LINE__, __func__); |
5664 | goto out; |
5665 | } |
5666 | |
5667 | event_reply->Function = MPI2_FUNCTION_EVENT_NOTIFICATION; |
5668 | event_reply->Event = |
5669 | cpu_to_le16(MPI2_EVENT_SAS_DEVICE_STATUS_CHANGE); |
5670 | event_reply->MsgLength = sz/4; |
5671 | event_reply->EventDataLength = |
5672 | cpu_to_le16(sizeof(Mpi2EventDataSasDeviceStatusChange_t)/4); |
5673 | event_data = (Mpi2EventDataSasDeviceStatusChange_t *) |
5674 | event_reply->EventData; |
5675 | event_data->ReasonCode = MPI2_EVENT_SAS_DEV_STAT_RC_SMART_DATA; |
5676 | event_data->ASC = 0x5D; |
5677 | event_data->DevHandle = cpu_to_le16(handle); |
5678 | event_data->SASAddress = cpu_to_le64(sas_target_priv_data->sas_address); |
5679 | mpt3sas_ctl_add_to_event_log(ioc, mpi_reply: event_reply); |
5680 | kfree(objp: event_reply); |
5681 | out: |
5682 | if (sas_device) |
5683 | sas_device_put(s: sas_device); |
5684 | return; |
5685 | |
5686 | out_unlock: |
5687 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
5688 | goto out; |
5689 | } |
5690 | |
5691 | /** |
5692 | * _scsih_io_done - scsi request callback |
5693 | * @ioc: per adapter object |
5694 | * @smid: system request message index |
5695 | * @msix_index: MSIX table index supplied by the OS |
5696 | * @reply: reply message frame(lower 32bit addr) |
5697 | * |
5698 | * Callback handler when using _scsih_qcmd. |
5699 | * |
5700 | * Return: 1 meaning mf should be freed from _base_interrupt |
5701 | * 0 means the mf is freed from this function. |
5702 | */ |
5703 | static u8 |
5704 | _scsih_io_done(struct MPT3SAS_ADAPTER *ioc, u16 smid, u8 msix_index, u32 reply) |
5705 | { |
5706 | Mpi25SCSIIORequest_t *mpi_request; |
5707 | Mpi2SCSIIOReply_t *mpi_reply; |
5708 | struct scsi_cmnd *scmd; |
5709 | struct scsiio_tracker *st; |
5710 | u16 ioc_status; |
5711 | u32 xfer_cnt; |
5712 | u8 scsi_state; |
5713 | u8 scsi_status; |
5714 | u32 log_info; |
5715 | struct MPT3SAS_DEVICE *sas_device_priv_data; |
5716 | u32 response_code = 0; |
5717 | |
5718 | mpi_reply = mpt3sas_base_get_reply_virt_addr(ioc, phys_addr: reply); |
5719 | |
5720 | scmd = mpt3sas_scsih_scsi_lookup_get(ioc, smid); |
5721 | if (scmd == NULL) |
5722 | return 1; |
5723 | |
5724 | _scsih_set_satl_pending(scmd, pending: false); |
5725 | |
5726 | mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); |
5727 | |
5728 | if (mpi_reply == NULL) { |
5729 | scmd->result = DID_OK << 16; |
5730 | goto out; |
5731 | } |
5732 | |
5733 | sas_device_priv_data = scmd->device->hostdata; |
5734 | if (!sas_device_priv_data || !sas_device_priv_data->sas_target || |
5735 | sas_device_priv_data->sas_target->deleted) { |
5736 | scmd->result = DID_NO_CONNECT << 16; |
5737 | goto out; |
5738 | } |
5739 | ioc_status = le16_to_cpu(mpi_reply->IOCStatus); |
5740 | |
5741 | /* |
5742 | * WARPDRIVE: If direct_io is set then it is directIO, |
5743 | * the failed direct I/O should be redirected to volume |
5744 | */ |
5745 | st = scsi_cmd_priv(cmd: scmd); |
5746 | if (st->direct_io && |
5747 | ((ioc_status & MPI2_IOCSTATUS_MASK) |
5748 | != MPI2_IOCSTATUS_SCSI_TASK_TERMINATED)) { |
5749 | st->direct_io = 0; |
5750 | st->scmd = scmd; |
5751 | memcpy(mpi_request->CDB.CDB32, scmd->cmnd, scmd->cmd_len); |
5752 | mpi_request->DevHandle = |
5753 | cpu_to_le16(sas_device_priv_data->sas_target->handle); |
5754 | ioc->put_smid_scsi_io(ioc, smid, |
5755 | sas_device_priv_data->sas_target->handle); |
5756 | return 0; |
5757 | } |
5758 | /* turning off TLR */ |
5759 | scsi_state = mpi_reply->SCSIState; |
5760 | if (scsi_state & MPI2_SCSI_STATE_RESPONSE_INFO_VALID) |
5761 | response_code = |
5762 | le32_to_cpu(mpi_reply->ResponseInfo) & 0xFF; |
5763 | if (!sas_device_priv_data->tlr_snoop_check) { |
5764 | sas_device_priv_data->tlr_snoop_check++; |
5765 | if ((!ioc->is_warpdrive && |
5766 | !scsih_is_raid(dev: &scmd->device->sdev_gendev) && |
5767 | !scsih_is_nvme(dev: &scmd->device->sdev_gendev)) |
5768 | && sas_is_tlr_enabled(scmd->device) && |
5769 | response_code == MPI2_SCSITASKMGMT_RSP_INVALID_FRAME) { |
5770 | sas_disable_tlr(scmd->device); |
5771 | sdev_printk(KERN_INFO, scmd->device, "TLR disabled\n" ); |
5772 | } |
5773 | } |
5774 | |
5775 | xfer_cnt = le32_to_cpu(mpi_reply->TransferCount); |
5776 | scsi_set_resid(cmd: scmd, resid: scsi_bufflen(cmd: scmd) - xfer_cnt); |
5777 | if (ioc_status & MPI2_IOCSTATUS_FLAG_LOG_INFO_AVAILABLE) |
5778 | log_info = le32_to_cpu(mpi_reply->IOCLogInfo); |
5779 | else |
5780 | log_info = 0; |
5781 | ioc_status &= MPI2_IOCSTATUS_MASK; |
5782 | scsi_status = mpi_reply->SCSIStatus; |
5783 | |
5784 | if (ioc_status == MPI2_IOCSTATUS_SCSI_DATA_UNDERRUN && xfer_cnt == 0 && |
5785 | (scsi_status == MPI2_SCSI_STATUS_BUSY || |
5786 | scsi_status == MPI2_SCSI_STATUS_RESERVATION_CONFLICT || |
5787 | scsi_status == MPI2_SCSI_STATUS_TASK_SET_FULL)) { |
5788 | ioc_status = MPI2_IOCSTATUS_SUCCESS; |
5789 | } |
5790 | |
5791 | if (scsi_state & MPI2_SCSI_STATE_AUTOSENSE_VALID) { |
5792 | struct sense_info data; |
5793 | const void *sense_data = mpt3sas_base_get_sense_buffer(ioc, |
5794 | smid); |
5795 | u32 sz = min_t(u32, SCSI_SENSE_BUFFERSIZE, |
5796 | le32_to_cpu(mpi_reply->SenseCount)); |
5797 | memcpy(scmd->sense_buffer, sense_data, sz); |
5798 | _scsih_normalize_sense(sense_buffer: scmd->sense_buffer, data: &data); |
5799 | /* failure prediction threshold exceeded */ |
5800 | if (data.asc == 0x5D) |
5801 | _scsih_smart_predicted_fault(ioc, |
5802 | le16_to_cpu(mpi_reply->DevHandle)); |
5803 | mpt3sas_trigger_scsi(ioc, sense_key: data.skey, asc: data.asc, ascq: data.ascq); |
5804 | |
5805 | if ((ioc->logging_level & MPT_DEBUG_REPLY) && |
5806 | ((scmd->sense_buffer[2] == UNIT_ATTENTION) || |
5807 | (scmd->sense_buffer[2] == MEDIUM_ERROR) || |
5808 | (scmd->sense_buffer[2] == HARDWARE_ERROR))) |
5809 | _scsih_scsi_ioc_info(ioc, scmd, mpi_reply, smid); |
5810 | } |
5811 | switch (ioc_status) { |
5812 | case MPI2_IOCSTATUS_BUSY: |
5813 | case MPI2_IOCSTATUS_INSUFFICIENT_RESOURCES: |
5814 | scmd->result = SAM_STAT_BUSY; |
5815 | break; |
5816 | |
5817 | case MPI2_IOCSTATUS_SCSI_DEVICE_NOT_THERE: |
5818 | scmd->result = DID_NO_CONNECT << 16; |
5819 | break; |
5820 | |
5821 | case MPI2_IOCSTATUS_SCSI_IOC_TERMINATED: |
5822 | if (sas_device_priv_data->block) { |
5823 | scmd->result = DID_TRANSPORT_DISRUPTED << 16; |
5824 | goto out; |
5825 | } |
5826 | if (log_info == 0x31110630) { |
5827 | if (scmd->retries > 2) { |
5828 | scmd->result = DID_NO_CONNECT << 16; |
5829 | scsi_device_set_state(sdev: scmd->device, |
5830 | state: SDEV_OFFLINE); |
5831 | } else { |
5832 | scmd->result = DID_SOFT_ERROR << 16; |
5833 | scmd->device->expecting_cc_ua = 1; |
5834 | } |
5835 | break; |
5836 | } else if (log_info == VIRTUAL_IO_FAILED_RETRY) { |
5837 | scmd->result = DID_RESET << 16; |
5838 | break; |
5839 | } else if ((scmd->device->channel == RAID_CHANNEL) && |
5840 | (scsi_state == (MPI2_SCSI_STATE_TERMINATED | |
5841 | MPI2_SCSI_STATE_NO_SCSI_STATUS))) { |
5842 | scmd->result = DID_RESET << 16; |
5843 | break; |
5844 | } |
5845 | scmd->result = DID_SOFT_ERROR << 16; |
5846 | break; |
5847 | case MPI2_IOCSTATUS_SCSI_TASK_TERMINATED: |
5848 | case MPI2_IOCSTATUS_SCSI_EXT_TERMINATED: |
5849 | scmd->result = DID_RESET << 16; |
5850 | break; |
5851 | |
5852 | case MPI2_IOCSTATUS_SCSI_RESIDUAL_MISMATCH: |
5853 | if ((xfer_cnt == 0) || (scmd->underflow > xfer_cnt)) |
5854 | scmd->result = DID_SOFT_ERROR << 16; |
5855 | else |
5856 | scmd->result = (DID_OK << 16) | scsi_status; |
5857 | break; |
5858 | |
5859 | case MPI2_IOCSTATUS_SCSI_DATA_UNDERRUN: |
5860 | scmd->result = (DID_OK << 16) | scsi_status; |
5861 | |
5862 | if ((scsi_state & MPI2_SCSI_STATE_AUTOSENSE_VALID)) |
5863 | break; |
5864 | |
5865 | if (xfer_cnt < scmd->underflow) { |
5866 | if (scsi_status == SAM_STAT_BUSY) |
5867 | scmd->result = SAM_STAT_BUSY; |
5868 | else |
5869 | scmd->result = DID_SOFT_ERROR << 16; |
5870 | } else if (scsi_state & (MPI2_SCSI_STATE_AUTOSENSE_FAILED | |
5871 | MPI2_SCSI_STATE_NO_SCSI_STATUS)) |
5872 | scmd->result = DID_SOFT_ERROR << 16; |
5873 | else if (scsi_state & MPI2_SCSI_STATE_TERMINATED) |
5874 | scmd->result = DID_RESET << 16; |
5875 | else if (!xfer_cnt && scmd->cmnd[0] == REPORT_LUNS) { |
5876 | mpi_reply->SCSIState = MPI2_SCSI_STATE_AUTOSENSE_VALID; |
5877 | mpi_reply->SCSIStatus = SAM_STAT_CHECK_CONDITION; |
5878 | scsi_build_sense(scmd, desc: 0, ILLEGAL_REQUEST, |
5879 | asc: 0x20, ascq: 0); |
5880 | } |
5881 | break; |
5882 | |
5883 | case MPI2_IOCSTATUS_SCSI_DATA_OVERRUN: |
5884 | scsi_set_resid(cmd: scmd, resid: 0); |
5885 | fallthrough; |
5886 | case MPI2_IOCSTATUS_SCSI_RECOVERED_ERROR: |
5887 | case MPI2_IOCSTATUS_SUCCESS: |
5888 | scmd->result = (DID_OK << 16) | scsi_status; |
5889 | if (response_code == |
5890 | MPI2_SCSITASKMGMT_RSP_INVALID_FRAME || |
5891 | (scsi_state & (MPI2_SCSI_STATE_AUTOSENSE_FAILED | |
5892 | MPI2_SCSI_STATE_NO_SCSI_STATUS))) |
5893 | scmd->result = DID_SOFT_ERROR << 16; |
5894 | else if (scsi_state & MPI2_SCSI_STATE_TERMINATED) |
5895 | scmd->result = DID_RESET << 16; |
5896 | break; |
5897 | |
5898 | case MPI2_IOCSTATUS_EEDP_GUARD_ERROR: |
5899 | case MPI2_IOCSTATUS_EEDP_REF_TAG_ERROR: |
5900 | case MPI2_IOCSTATUS_EEDP_APP_TAG_ERROR: |
5901 | _scsih_eedp_error_handling(scmd, ioc_status); |
5902 | break; |
5903 | |
5904 | case MPI2_IOCSTATUS_SCSI_PROTOCOL_ERROR: |
5905 | case MPI2_IOCSTATUS_INVALID_FUNCTION: |
5906 | case MPI2_IOCSTATUS_INVALID_SGL: |
5907 | case MPI2_IOCSTATUS_INTERNAL_ERROR: |
5908 | case MPI2_IOCSTATUS_INVALID_FIELD: |
5909 | case MPI2_IOCSTATUS_INVALID_STATE: |
5910 | case MPI2_IOCSTATUS_SCSI_IO_DATA_ERROR: |
5911 | case MPI2_IOCSTATUS_SCSI_TASK_MGMT_FAILED: |
5912 | case MPI2_IOCSTATUS_INSUFFICIENT_POWER: |
5913 | default: |
5914 | scmd->result = DID_SOFT_ERROR << 16; |
5915 | break; |
5916 | |
5917 | } |
5918 | |
5919 | if (scmd->result && (ioc->logging_level & MPT_DEBUG_REPLY)) |
5920 | _scsih_scsi_ioc_info(ioc , scmd, mpi_reply, smid); |
5921 | |
5922 | out: |
5923 | |
5924 | scsi_dma_unmap(cmd: scmd); |
5925 | mpt3sas_base_free_smid(ioc, smid); |
5926 | scsi_done(cmd: scmd); |
5927 | return 0; |
5928 | } |
5929 | |
5930 | /** |
5931 | * _scsih_update_vphys_after_reset - update the Port's |
5932 | * vphys_list after reset |
5933 | * @ioc: per adapter object |
5934 | * |
5935 | * Returns nothing. |
5936 | */ |
5937 | static void |
5938 | _scsih_update_vphys_after_reset(struct MPT3SAS_ADAPTER *ioc) |
5939 | { |
5940 | u16 sz, ioc_status; |
5941 | int i; |
5942 | Mpi2ConfigReply_t mpi_reply; |
5943 | Mpi2SasIOUnitPage0_t *sas_iounit_pg0 = NULL; |
5944 | u16 attached_handle; |
5945 | u64 attached_sas_addr; |
5946 | u8 found = 0, port_id; |
5947 | Mpi2SasPhyPage0_t phy_pg0; |
5948 | struct hba_port *port, *port_next, *mport; |
5949 | struct virtual_phy *vphy, *vphy_next; |
5950 | struct _sas_device *sas_device; |
5951 | |
5952 | /* |
5953 | * Mark all the vphys objects as dirty. |
5954 | */ |
5955 | list_for_each_entry_safe(port, port_next, |
5956 | &ioc->port_table_list, list) { |
5957 | if (!port->vphys_mask) |
5958 | continue; |
5959 | list_for_each_entry_safe(vphy, vphy_next, |
5960 | &port->vphys_list, list) { |
5961 | vphy->flags |= MPT_VPHY_FLAG_DIRTY_PHY; |
5962 | } |
5963 | } |
5964 | |
5965 | /* |
5966 | * Read SASIOUnitPage0 to get each HBA Phy's data. |
5967 | */ |
5968 | sz = struct_size(sas_iounit_pg0, PhyData, ioc->sas_hba.num_phys); |
5969 | sas_iounit_pg0 = kzalloc(size: sz, GFP_KERNEL); |
5970 | if (!sas_iounit_pg0) { |
5971 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
5972 | __FILE__, __LINE__, __func__); |
5973 | return; |
5974 | } |
5975 | if ((mpt3sas_config_get_sas_iounit_pg0(ioc, mpi_reply: &mpi_reply, |
5976 | config_page: sas_iounit_pg0, sz)) != 0) |
5977 | goto out; |
5978 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; |
5979 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) |
5980 | goto out; |
5981 | /* |
5982 | * Loop over each HBA Phy. |
5983 | */ |
5984 | for (i = 0; i < ioc->sas_hba.num_phys; i++) { |
5985 | /* |
5986 | * Check whether Phy's Negotiation Link Rate is > 1.5G or not. |
5987 | */ |
5988 | if ((sas_iounit_pg0->PhyData[i].NegotiatedLinkRate >> 4) < |
5989 | MPI2_SAS_NEG_LINK_RATE_1_5) |
5990 | continue; |
5991 | /* |
5992 | * Check whether Phy is connected to SEP device or not, |
5993 | * if it is SEP device then read the Phy's SASPHYPage0 data to |
5994 | * determine whether Phy is a virtual Phy or not. if it is |
5995 | * virtual phy then it is conformed that the attached remote |
5996 | * device is a HBA's vSES device. |
5997 | */ |
5998 | if (!(le32_to_cpu( |
5999 | sas_iounit_pg0->PhyData[i].ControllerPhyDeviceInfo) & |
6000 | MPI2_SAS_DEVICE_INFO_SEP)) |
6001 | continue; |
6002 | |
6003 | if ((mpt3sas_config_get_phy_pg0(ioc, mpi_reply: &mpi_reply, config_page: &phy_pg0, |
6004 | phy_number: i))) { |
6005 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6006 | __FILE__, __LINE__, __func__); |
6007 | continue; |
6008 | } |
6009 | |
6010 | if (!(le32_to_cpu(phy_pg0.PhyInfo) & |
6011 | MPI2_SAS_PHYINFO_VIRTUAL_PHY)) |
6012 | continue; |
6013 | /* |
6014 | * Get the vSES device's SAS Address. |
6015 | */ |
6016 | attached_handle = le16_to_cpu( |
6017 | sas_iounit_pg0->PhyData[i].AttachedDevHandle); |
6018 | if (_scsih_get_sas_address(ioc, handle: attached_handle, |
6019 | sas_address: &attached_sas_addr) != 0) { |
6020 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6021 | __FILE__, __LINE__, __func__); |
6022 | continue; |
6023 | } |
6024 | |
6025 | found = 0; |
6026 | port = port_next = NULL; |
6027 | /* |
6028 | * Loop over each virtual_phy object from |
6029 | * each port's vphys_list. |
6030 | */ |
6031 | list_for_each_entry_safe(port, |
6032 | port_next, &ioc->port_table_list, list) { |
6033 | if (!port->vphys_mask) |
6034 | continue; |
6035 | list_for_each_entry_safe(vphy, vphy_next, |
6036 | &port->vphys_list, list) { |
6037 | /* |
6038 | * Continue with next virtual_phy object |
6039 | * if the object is not marked as dirty. |
6040 | */ |
6041 | if (!(vphy->flags & MPT_VPHY_FLAG_DIRTY_PHY)) |
6042 | continue; |
6043 | |
6044 | /* |
6045 | * Continue with next virtual_phy object |
6046 | * if the object's SAS Address is not equals |
6047 | * to current Phy's vSES device SAS Address. |
6048 | */ |
6049 | if (vphy->sas_address != attached_sas_addr) |
6050 | continue; |
6051 | /* |
6052 | * Enable current Phy number bit in object's |
6053 | * phy_mask field. |
6054 | */ |
6055 | if (!(vphy->phy_mask & (1 << i))) |
6056 | vphy->phy_mask = (1 << i); |
6057 | /* |
6058 | * Get hba_port object from hba_port table |
6059 | * corresponding to current phy's Port ID. |
6060 | * if there is no hba_port object corresponding |
6061 | * to Phy's Port ID then create a new hba_port |
6062 | * object & add to hba_port table. |
6063 | */ |
6064 | port_id = sas_iounit_pg0->PhyData[i].Port; |
6065 | mport = mpt3sas_get_port_by_id(ioc, port_id, bypass_dirty_port_flag: 1); |
6066 | if (!mport) { |
6067 | mport = kzalloc( |
6068 | size: sizeof(struct hba_port), GFP_KERNEL); |
6069 | if (!mport) |
6070 | break; |
6071 | mport->port_id = port_id; |
6072 | ioc_info(ioc, |
6073 | "%s: hba_port entry: %p, port: %d is added to hba_port list\n" , |
6074 | __func__, mport, mport->port_id); |
6075 | list_add_tail(new: &mport->list, |
6076 | head: &ioc->port_table_list); |
6077 | } |
6078 | /* |
6079 | * If mport & port pointers are not pointing to |
6080 | * same hba_port object then it means that vSES |
6081 | * device's Port ID got changed after reset and |
6082 | * hence move current virtual_phy object from |
6083 | * port's vphys_list to mport's vphys_list. |
6084 | */ |
6085 | if (port != mport) { |
6086 | if (!mport->vphys_mask) |
6087 | INIT_LIST_HEAD( |
6088 | list: &mport->vphys_list); |
6089 | mport->vphys_mask |= (1 << i); |
6090 | port->vphys_mask &= ~(1 << i); |
6091 | list_move(list: &vphy->list, |
6092 | head: &mport->vphys_list); |
6093 | sas_device = mpt3sas_get_sdev_by_addr( |
6094 | ioc, sas_address: attached_sas_addr, port); |
6095 | if (sas_device) |
6096 | sas_device->port = mport; |
6097 | } |
6098 | /* |
6099 | * Earlier while updating the hba_port table, |
6100 | * it is determined that there is no other |
6101 | * direct attached device with mport's Port ID, |
6102 | * Hence mport was marked as dirty. Only vSES |
6103 | * device has this Port ID, so unmark the mport |
6104 | * as dirt. |
6105 | */ |
6106 | if (mport->flags & HBA_PORT_FLAG_DIRTY_PORT) { |
6107 | mport->sas_address = 0; |
6108 | mport->phy_mask = 0; |
6109 | mport->flags &= |
6110 | ~HBA_PORT_FLAG_DIRTY_PORT; |
6111 | } |
6112 | /* |
6113 | * Unmark current virtual_phy object as dirty. |
6114 | */ |
6115 | vphy->flags &= ~MPT_VPHY_FLAG_DIRTY_PHY; |
6116 | found = 1; |
6117 | break; |
6118 | } |
6119 | if (found) |
6120 | break; |
6121 | } |
6122 | } |
6123 | out: |
6124 | kfree(objp: sas_iounit_pg0); |
6125 | } |
6126 | |
6127 | /** |
6128 | * _scsih_get_port_table_after_reset - Construct temporary port table |
6129 | * @ioc: per adapter object |
6130 | * @port_table: address where port table needs to be constructed |
6131 | * |
6132 | * return number of HBA port entries available after reset. |
6133 | */ |
6134 | static int |
6135 | _scsih_get_port_table_after_reset(struct MPT3SAS_ADAPTER *ioc, |
6136 | struct hba_port *port_table) |
6137 | { |
6138 | u16 sz, ioc_status; |
6139 | int i, j; |
6140 | Mpi2ConfigReply_t mpi_reply; |
6141 | Mpi2SasIOUnitPage0_t *sas_iounit_pg0 = NULL; |
6142 | u16 attached_handle; |
6143 | u64 attached_sas_addr; |
6144 | u8 found = 0, port_count = 0, port_id; |
6145 | |
6146 | sz = struct_size(sas_iounit_pg0, PhyData, ioc->sas_hba.num_phys); |
6147 | sas_iounit_pg0 = kzalloc(size: sz, GFP_KERNEL); |
6148 | if (!sas_iounit_pg0) { |
6149 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6150 | __FILE__, __LINE__, __func__); |
6151 | return port_count; |
6152 | } |
6153 | |
6154 | if ((mpt3sas_config_get_sas_iounit_pg0(ioc, mpi_reply: &mpi_reply, |
6155 | config_page: sas_iounit_pg0, sz)) != 0) |
6156 | goto out; |
6157 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; |
6158 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) |
6159 | goto out; |
6160 | for (i = 0; i < ioc->sas_hba.num_phys; i++) { |
6161 | found = 0; |
6162 | if ((sas_iounit_pg0->PhyData[i].NegotiatedLinkRate >> 4) < |
6163 | MPI2_SAS_NEG_LINK_RATE_1_5) |
6164 | continue; |
6165 | attached_handle = |
6166 | le16_to_cpu(sas_iounit_pg0->PhyData[i].AttachedDevHandle); |
6167 | if (_scsih_get_sas_address( |
6168 | ioc, handle: attached_handle, sas_address: &attached_sas_addr) != 0) { |
6169 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6170 | __FILE__, __LINE__, __func__); |
6171 | continue; |
6172 | } |
6173 | |
6174 | for (j = 0; j < port_count; j++) { |
6175 | port_id = sas_iounit_pg0->PhyData[i].Port; |
6176 | if (port_table[j].port_id == port_id && |
6177 | port_table[j].sas_address == attached_sas_addr) { |
6178 | port_table[j].phy_mask |= (1 << i); |
6179 | found = 1; |
6180 | break; |
6181 | } |
6182 | } |
6183 | |
6184 | if (found) |
6185 | continue; |
6186 | |
6187 | port_id = sas_iounit_pg0->PhyData[i].Port; |
6188 | port_table[port_count].port_id = port_id; |
6189 | port_table[port_count].phy_mask = (1 << i); |
6190 | port_table[port_count].sas_address = attached_sas_addr; |
6191 | port_count++; |
6192 | } |
6193 | out: |
6194 | kfree(objp: sas_iounit_pg0); |
6195 | return port_count; |
6196 | } |
6197 | |
6198 | enum hba_port_matched_codes { |
6199 | NOT_MATCHED = 0, |
6200 | MATCHED_WITH_ADDR_AND_PHYMASK, |
6201 | MATCHED_WITH_ADDR_SUBPHYMASK_AND_PORT, |
6202 | MATCHED_WITH_ADDR_AND_SUBPHYMASK, |
6203 | MATCHED_WITH_ADDR, |
6204 | }; |
6205 | |
6206 | /** |
6207 | * _scsih_look_and_get_matched_port_entry - Get matched hba port entry |
6208 | * from HBA port table |
6209 | * @ioc: per adapter object |
6210 | * @port_entry: hba port entry from temporary port table which needs to be |
6211 | * searched for matched entry in the HBA port table |
6212 | * @matched_port_entry: save matched hba port entry here |
6213 | * @count: count of matched entries |
6214 | * |
6215 | * return type of matched entry found. |
6216 | */ |
6217 | static enum hba_port_matched_codes |
6218 | _scsih_look_and_get_matched_port_entry(struct MPT3SAS_ADAPTER *ioc, |
6219 | struct hba_port *port_entry, |
6220 | struct hba_port **matched_port_entry, int *count) |
6221 | { |
6222 | struct hba_port *port_table_entry, *matched_port = NULL; |
6223 | enum hba_port_matched_codes matched_code = NOT_MATCHED; |
6224 | int lcount = 0; |
6225 | *matched_port_entry = NULL; |
6226 | |
6227 | list_for_each_entry(port_table_entry, &ioc->port_table_list, list) { |
6228 | if (!(port_table_entry->flags & HBA_PORT_FLAG_DIRTY_PORT)) |
6229 | continue; |
6230 | |
6231 | if ((port_table_entry->sas_address == port_entry->sas_address) |
6232 | && (port_table_entry->phy_mask == port_entry->phy_mask)) { |
6233 | matched_code = MATCHED_WITH_ADDR_AND_PHYMASK; |
6234 | matched_port = port_table_entry; |
6235 | break; |
6236 | } |
6237 | |
6238 | if ((port_table_entry->sas_address == port_entry->sas_address) |
6239 | && (port_table_entry->phy_mask & port_entry->phy_mask) |
6240 | && (port_table_entry->port_id == port_entry->port_id)) { |
6241 | matched_code = MATCHED_WITH_ADDR_SUBPHYMASK_AND_PORT; |
6242 | matched_port = port_table_entry; |
6243 | continue; |
6244 | } |
6245 | |
6246 | if ((port_table_entry->sas_address == port_entry->sas_address) |
6247 | && (port_table_entry->phy_mask & port_entry->phy_mask)) { |
6248 | if (matched_code == |
6249 | MATCHED_WITH_ADDR_SUBPHYMASK_AND_PORT) |
6250 | continue; |
6251 | matched_code = MATCHED_WITH_ADDR_AND_SUBPHYMASK; |
6252 | matched_port = port_table_entry; |
6253 | continue; |
6254 | } |
6255 | |
6256 | if (port_table_entry->sas_address == port_entry->sas_address) { |
6257 | if (matched_code == |
6258 | MATCHED_WITH_ADDR_SUBPHYMASK_AND_PORT) |
6259 | continue; |
6260 | if (matched_code == MATCHED_WITH_ADDR_AND_SUBPHYMASK) |
6261 | continue; |
6262 | matched_code = MATCHED_WITH_ADDR; |
6263 | matched_port = port_table_entry; |
6264 | lcount++; |
6265 | } |
6266 | } |
6267 | |
6268 | *matched_port_entry = matched_port; |
6269 | if (matched_code == MATCHED_WITH_ADDR) |
6270 | *count = lcount; |
6271 | return matched_code; |
6272 | } |
6273 | |
6274 | /** |
6275 | * _scsih_del_phy_part_of_anther_port - remove phy if it |
6276 | * is a part of anther port |
6277 | *@ioc: per adapter object |
6278 | *@port_table: port table after reset |
6279 | *@index: hba port entry index |
6280 | *@port_count: number of ports available after host reset |
6281 | *@offset: HBA phy bit offset |
6282 | * |
6283 | */ |
6284 | static void |
6285 | _scsih_del_phy_part_of_anther_port(struct MPT3SAS_ADAPTER *ioc, |
6286 | struct hba_port *port_table, |
6287 | int index, u8 port_count, int offset) |
6288 | { |
6289 | struct _sas_node *sas_node = &ioc->sas_hba; |
6290 | u32 i, found = 0; |
6291 | |
6292 | for (i = 0; i < port_count; i++) { |
6293 | if (i == index) |
6294 | continue; |
6295 | |
6296 | if (port_table[i].phy_mask & (1 << offset)) { |
6297 | mpt3sas_transport_del_phy_from_an_existing_port( |
6298 | ioc, sas_node, mpt3sas_phy: &sas_node->phy[offset]); |
6299 | found = 1; |
6300 | break; |
6301 | } |
6302 | } |
6303 | if (!found) |
6304 | port_table[index].phy_mask |= (1 << offset); |
6305 | } |
6306 | |
6307 | /** |
6308 | * _scsih_add_or_del_phys_from_existing_port - add/remove phy to/from |
6309 | * right port |
6310 | *@ioc: per adapter object |
6311 | *@hba_port_entry: hba port table entry |
6312 | *@port_table: temporary port table |
6313 | *@index: hba port entry index |
6314 | *@port_count: number of ports available after host reset |
6315 | * |
6316 | */ |
6317 | static void |
6318 | _scsih_add_or_del_phys_from_existing_port(struct MPT3SAS_ADAPTER *ioc, |
6319 | struct hba_port *hba_port_entry, struct hba_port *port_table, |
6320 | int index, int port_count) |
6321 | { |
6322 | u32 phy_mask, offset = 0; |
6323 | struct _sas_node *sas_node = &ioc->sas_hba; |
6324 | |
6325 | phy_mask = hba_port_entry->phy_mask ^ port_table[index].phy_mask; |
6326 | |
6327 | for (offset = 0; offset < ioc->sas_hba.num_phys; offset++) { |
6328 | if (phy_mask & (1 << offset)) { |
6329 | if (!(port_table[index].phy_mask & (1 << offset))) { |
6330 | _scsih_del_phy_part_of_anther_port( |
6331 | ioc, port_table, index, port_count, |
6332 | offset); |
6333 | continue; |
6334 | } |
6335 | if (sas_node->phy[offset].phy_belongs_to_port) |
6336 | mpt3sas_transport_del_phy_from_an_existing_port( |
6337 | ioc, sas_node, mpt3sas_phy: &sas_node->phy[offset]); |
6338 | mpt3sas_transport_add_phy_to_an_existing_port( |
6339 | ioc, sas_node, mpt3sas_phy: &sas_node->phy[offset], |
6340 | sas_address: hba_port_entry->sas_address, |
6341 | port: hba_port_entry); |
6342 | } |
6343 | } |
6344 | } |
6345 | |
6346 | /** |
6347 | * _scsih_del_dirty_vphy - delete virtual_phy objects marked as dirty. |
6348 | * @ioc: per adapter object |
6349 | * |
6350 | * Returns nothing. |
6351 | */ |
6352 | static void |
6353 | _scsih_del_dirty_vphy(struct MPT3SAS_ADAPTER *ioc) |
6354 | { |
6355 | struct hba_port *port, *port_next; |
6356 | struct virtual_phy *vphy, *vphy_next; |
6357 | |
6358 | list_for_each_entry_safe(port, port_next, |
6359 | &ioc->port_table_list, list) { |
6360 | if (!port->vphys_mask) |
6361 | continue; |
6362 | list_for_each_entry_safe(vphy, vphy_next, |
6363 | &port->vphys_list, list) { |
6364 | if (vphy->flags & MPT_VPHY_FLAG_DIRTY_PHY) { |
6365 | drsprintk(ioc, ioc_info(ioc, |
6366 | "Deleting vphy %p entry from port id: %d\t, Phy_mask 0x%08x\n" , |
6367 | vphy, port->port_id, |
6368 | vphy->phy_mask)); |
6369 | port->vphys_mask &= ~vphy->phy_mask; |
6370 | list_del(entry: &vphy->list); |
6371 | kfree(objp: vphy); |
6372 | } |
6373 | } |
6374 | if (!port->vphys_mask && !port->sas_address) |
6375 | port->flags |= HBA_PORT_FLAG_DIRTY_PORT; |
6376 | } |
6377 | } |
6378 | |
6379 | /** |
6380 | * _scsih_del_dirty_port_entries - delete dirty port entries from port list |
6381 | * after host reset |
6382 | *@ioc: per adapter object |
6383 | * |
6384 | */ |
6385 | static void |
6386 | _scsih_del_dirty_port_entries(struct MPT3SAS_ADAPTER *ioc) |
6387 | { |
6388 | struct hba_port *port, *port_next; |
6389 | |
6390 | list_for_each_entry_safe(port, port_next, |
6391 | &ioc->port_table_list, list) { |
6392 | if (!(port->flags & HBA_PORT_FLAG_DIRTY_PORT) || |
6393 | port->flags & HBA_PORT_FLAG_NEW_PORT) |
6394 | continue; |
6395 | |
6396 | drsprintk(ioc, ioc_info(ioc, |
6397 | "Deleting port table entry %p having Port: %d\t Phy_mask 0x%08x\n" , |
6398 | port, port->port_id, port->phy_mask)); |
6399 | list_del(entry: &port->list); |
6400 | kfree(objp: port); |
6401 | } |
6402 | } |
6403 | |
6404 | /** |
6405 | * _scsih_sas_port_refresh - Update HBA port table after host reset |
6406 | * @ioc: per adapter object |
6407 | */ |
6408 | static void |
6409 | _scsih_sas_port_refresh(struct MPT3SAS_ADAPTER *ioc) |
6410 | { |
6411 | u32 port_count = 0; |
6412 | struct hba_port *port_table; |
6413 | struct hba_port *port_table_entry; |
6414 | struct hba_port *port_entry = NULL; |
6415 | int i, j, count = 0, lcount = 0; |
6416 | int ret; |
6417 | u64 sas_addr; |
6418 | u8 num_phys; |
6419 | |
6420 | drsprintk(ioc, ioc_info(ioc, |
6421 | "updating ports for sas_host(0x%016llx)\n" , |
6422 | (unsigned long long)ioc->sas_hba.sas_address)); |
6423 | |
6424 | mpt3sas_config_get_number_hba_phys(ioc, num_phys: &num_phys); |
6425 | if (!num_phys) { |
6426 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6427 | __FILE__, __LINE__, __func__); |
6428 | return; |
6429 | } |
6430 | |
6431 | if (num_phys > ioc->sas_hba.nr_phys_allocated) { |
6432 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6433 | __FILE__, __LINE__, __func__); |
6434 | return; |
6435 | } |
6436 | ioc->sas_hba.num_phys = num_phys; |
6437 | |
6438 | port_table = kcalloc(n: ioc->sas_hba.num_phys, |
6439 | size: sizeof(struct hba_port), GFP_KERNEL); |
6440 | if (!port_table) |
6441 | return; |
6442 | |
6443 | port_count = _scsih_get_port_table_after_reset(ioc, port_table); |
6444 | if (!port_count) |
6445 | return; |
6446 | |
6447 | drsprintk(ioc, ioc_info(ioc, "New Port table\n" )); |
6448 | for (j = 0; j < port_count; j++) |
6449 | drsprintk(ioc, ioc_info(ioc, |
6450 | "Port: %d\t Phy_mask 0x%08x\t sas_addr(0x%016llx)\n" , |
6451 | port_table[j].port_id, |
6452 | port_table[j].phy_mask, port_table[j].sas_address)); |
6453 | |
6454 | list_for_each_entry(port_table_entry, &ioc->port_table_list, list) |
6455 | port_table_entry->flags |= HBA_PORT_FLAG_DIRTY_PORT; |
6456 | |
6457 | drsprintk(ioc, ioc_info(ioc, "Old Port table\n" )); |
6458 | port_table_entry = NULL; |
6459 | list_for_each_entry(port_table_entry, &ioc->port_table_list, list) { |
6460 | drsprintk(ioc, ioc_info(ioc, |
6461 | "Port: %d\t Phy_mask 0x%08x\t sas_addr(0x%016llx)\n" , |
6462 | port_table_entry->port_id, |
6463 | port_table_entry->phy_mask, |
6464 | port_table_entry->sas_address)); |
6465 | } |
6466 | |
6467 | for (j = 0; j < port_count; j++) { |
6468 | ret = _scsih_look_and_get_matched_port_entry(ioc, |
6469 | port_entry: &port_table[j], matched_port_entry: &port_entry, count: &count); |
6470 | if (!port_entry) { |
6471 | drsprintk(ioc, ioc_info(ioc, |
6472 | "No Matched entry for sas_addr(0x%16llx), Port:%d\n" , |
6473 | port_table[j].sas_address, |
6474 | port_table[j].port_id)); |
6475 | continue; |
6476 | } |
6477 | |
6478 | switch (ret) { |
6479 | case MATCHED_WITH_ADDR_SUBPHYMASK_AND_PORT: |
6480 | case MATCHED_WITH_ADDR_AND_SUBPHYMASK: |
6481 | _scsih_add_or_del_phys_from_existing_port(ioc, |
6482 | hba_port_entry: port_entry, port_table, index: j, port_count); |
6483 | break; |
6484 | case MATCHED_WITH_ADDR: |
6485 | sas_addr = port_table[j].sas_address; |
6486 | for (i = 0; i < port_count; i++) { |
6487 | if (port_table[i].sas_address == sas_addr) |
6488 | lcount++; |
6489 | } |
6490 | |
6491 | if (count > 1 || lcount > 1) |
6492 | port_entry = NULL; |
6493 | else |
6494 | _scsih_add_or_del_phys_from_existing_port(ioc, |
6495 | hba_port_entry: port_entry, port_table, index: j, port_count); |
6496 | } |
6497 | |
6498 | if (!port_entry) |
6499 | continue; |
6500 | |
6501 | if (port_entry->port_id != port_table[j].port_id) |
6502 | port_entry->port_id = port_table[j].port_id; |
6503 | port_entry->flags &= ~HBA_PORT_FLAG_DIRTY_PORT; |
6504 | port_entry->phy_mask = port_table[j].phy_mask; |
6505 | } |
6506 | |
6507 | port_table_entry = NULL; |
6508 | } |
6509 | |
6510 | /** |
6511 | * _scsih_alloc_vphy - allocate virtual_phy object |
6512 | * @ioc: per adapter object |
6513 | * @port_id: Port ID number |
6514 | * @phy_num: HBA Phy number |
6515 | * |
6516 | * Returns allocated virtual_phy object. |
6517 | */ |
6518 | static struct virtual_phy * |
6519 | _scsih_alloc_vphy(struct MPT3SAS_ADAPTER *ioc, u8 port_id, u8 phy_num) |
6520 | { |
6521 | struct virtual_phy *vphy; |
6522 | struct hba_port *port; |
6523 | |
6524 | port = mpt3sas_get_port_by_id(ioc, port_id, bypass_dirty_port_flag: 0); |
6525 | if (!port) |
6526 | return NULL; |
6527 | |
6528 | vphy = mpt3sas_get_vphy_by_phy(ioc, port, phy: phy_num); |
6529 | if (!vphy) { |
6530 | vphy = kzalloc(size: sizeof(struct virtual_phy), GFP_KERNEL); |
6531 | if (!vphy) |
6532 | return NULL; |
6533 | |
6534 | if (!port->vphys_mask) |
6535 | INIT_LIST_HEAD(list: &port->vphys_list); |
6536 | |
6537 | /* |
6538 | * Enable bit corresponding to HBA phy number on its |
6539 | * parent hba_port object's vphys_mask field. |
6540 | */ |
6541 | port->vphys_mask |= (1 << phy_num); |
6542 | vphy->phy_mask |= (1 << phy_num); |
6543 | |
6544 | list_add_tail(new: &vphy->list, head: &port->vphys_list); |
6545 | |
6546 | ioc_info(ioc, |
6547 | "vphy entry: %p, port id: %d, phy:%d is added to port's vphys_list\n" , |
6548 | vphy, port->port_id, phy_num); |
6549 | } |
6550 | return vphy; |
6551 | } |
6552 | |
6553 | /** |
6554 | * _scsih_sas_host_refresh - refreshing sas host object contents |
6555 | * @ioc: per adapter object |
6556 | * Context: user |
6557 | * |
6558 | * During port enable, fw will send topology events for every device. Its |
6559 | * possible that the handles may change from the previous setting, so this |
6560 | * code keeping handles updating if changed. |
6561 | */ |
6562 | static void |
6563 | _scsih_sas_host_refresh(struct MPT3SAS_ADAPTER *ioc) |
6564 | { |
6565 | u16 sz; |
6566 | u16 ioc_status; |
6567 | int i; |
6568 | Mpi2ConfigReply_t mpi_reply; |
6569 | Mpi2SasIOUnitPage0_t *sas_iounit_pg0 = NULL; |
6570 | u16 attached_handle; |
6571 | u8 link_rate, port_id; |
6572 | struct hba_port *port; |
6573 | Mpi2SasPhyPage0_t phy_pg0; |
6574 | |
6575 | dtmprintk(ioc, |
6576 | ioc_info(ioc, "updating handles for sas_host(0x%016llx)\n" , |
6577 | (u64)ioc->sas_hba.sas_address)); |
6578 | |
6579 | sz = struct_size(sas_iounit_pg0, PhyData, ioc->sas_hba.num_phys); |
6580 | sas_iounit_pg0 = kzalloc(size: sz, GFP_KERNEL); |
6581 | if (!sas_iounit_pg0) { |
6582 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6583 | __FILE__, __LINE__, __func__); |
6584 | return; |
6585 | } |
6586 | |
6587 | if ((mpt3sas_config_get_sas_iounit_pg0(ioc, mpi_reply: &mpi_reply, |
6588 | config_page: sas_iounit_pg0, sz)) != 0) |
6589 | goto out; |
6590 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; |
6591 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) |
6592 | goto out; |
6593 | for (i = 0; i < ioc->sas_hba.num_phys ; i++) { |
6594 | link_rate = sas_iounit_pg0->PhyData[i].NegotiatedLinkRate >> 4; |
6595 | if (i == 0) |
6596 | ioc->sas_hba.handle = le16_to_cpu( |
6597 | sas_iounit_pg0->PhyData[0].ControllerDevHandle); |
6598 | port_id = sas_iounit_pg0->PhyData[i].Port; |
6599 | if (!(mpt3sas_get_port_by_id(ioc, port_id, bypass_dirty_port_flag: 0))) { |
6600 | port = kzalloc(size: sizeof(struct hba_port), GFP_KERNEL); |
6601 | if (!port) |
6602 | goto out; |
6603 | |
6604 | port->port_id = port_id; |
6605 | ioc_info(ioc, |
6606 | "hba_port entry: %p, port: %d is added to hba_port list\n" , |
6607 | port, port->port_id); |
6608 | if (ioc->shost_recovery) |
6609 | port->flags = HBA_PORT_FLAG_NEW_PORT; |
6610 | list_add_tail(new: &port->list, head: &ioc->port_table_list); |
6611 | } |
6612 | /* |
6613 | * Check whether current Phy belongs to HBA vSES device or not. |
6614 | */ |
6615 | if (le32_to_cpu(sas_iounit_pg0->PhyData[i].ControllerPhyDeviceInfo) & |
6616 | MPI2_SAS_DEVICE_INFO_SEP && |
6617 | (link_rate >= MPI2_SAS_NEG_LINK_RATE_1_5)) { |
6618 | if ((mpt3sas_config_get_phy_pg0(ioc, mpi_reply: &mpi_reply, |
6619 | config_page: &phy_pg0, phy_number: i))) { |
6620 | ioc_err(ioc, |
6621 | "failure at %s:%d/%s()!\n" , |
6622 | __FILE__, __LINE__, __func__); |
6623 | goto out; |
6624 | } |
6625 | if (!(le32_to_cpu(phy_pg0.PhyInfo) & |
6626 | MPI2_SAS_PHYINFO_VIRTUAL_PHY)) |
6627 | continue; |
6628 | /* |
6629 | * Allocate a virtual_phy object for vSES device, if |
6630 | * this vSES device is hot added. |
6631 | */ |
6632 | if (!_scsih_alloc_vphy(ioc, port_id, phy_num: i)) |
6633 | goto out; |
6634 | ioc->sas_hba.phy[i].hba_vphy = 1; |
6635 | } |
6636 | |
6637 | /* |
6638 | * Add new HBA phys to STL if these new phys got added as part |
6639 | * of HBA Firmware upgrade/downgrade operation. |
6640 | */ |
6641 | if (!ioc->sas_hba.phy[i].phy) { |
6642 | if ((mpt3sas_config_get_phy_pg0(ioc, mpi_reply: &mpi_reply, |
6643 | config_page: &phy_pg0, phy_number: i))) { |
6644 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6645 | __FILE__, __LINE__, __func__); |
6646 | continue; |
6647 | } |
6648 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & |
6649 | MPI2_IOCSTATUS_MASK; |
6650 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { |
6651 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6652 | __FILE__, __LINE__, __func__); |
6653 | continue; |
6654 | } |
6655 | ioc->sas_hba.phy[i].phy_id = i; |
6656 | mpt3sas_transport_add_host_phy(ioc, |
6657 | mpt3sas_phy: &ioc->sas_hba.phy[i], phy_pg0, |
6658 | parent_dev: ioc->sas_hba.parent_dev); |
6659 | continue; |
6660 | } |
6661 | ioc->sas_hba.phy[i].handle = ioc->sas_hba.handle; |
6662 | attached_handle = le16_to_cpu(sas_iounit_pg0->PhyData[i]. |
6663 | AttachedDevHandle); |
6664 | if (attached_handle && link_rate < MPI2_SAS_NEG_LINK_RATE_1_5) |
6665 | link_rate = MPI2_SAS_NEG_LINK_RATE_1_5; |
6666 | ioc->sas_hba.phy[i].port = |
6667 | mpt3sas_get_port_by_id(ioc, port_id, bypass_dirty_port_flag: 0); |
6668 | mpt3sas_transport_update_links(ioc, sas_address: ioc->sas_hba.sas_address, |
6669 | handle: attached_handle, phy_number: i, link_rate, |
6670 | port: ioc->sas_hba.phy[i].port); |
6671 | } |
6672 | /* |
6673 | * Clear the phy details if this phy got disabled as part of |
6674 | * HBA Firmware upgrade/downgrade operation. |
6675 | */ |
6676 | for (i = ioc->sas_hba.num_phys; |
6677 | i < ioc->sas_hba.nr_phys_allocated; i++) { |
6678 | if (ioc->sas_hba.phy[i].phy && |
6679 | ioc->sas_hba.phy[i].phy->negotiated_linkrate >= |
6680 | SAS_LINK_RATE_1_5_GBPS) |
6681 | mpt3sas_transport_update_links(ioc, |
6682 | sas_address: ioc->sas_hba.sas_address, handle: 0, phy_number: i, |
6683 | MPI2_SAS_NEG_LINK_RATE_PHY_DISABLED, NULL); |
6684 | } |
6685 | out: |
6686 | kfree(objp: sas_iounit_pg0); |
6687 | } |
6688 | |
6689 | /** |
6690 | * _scsih_sas_host_add - create sas host object |
6691 | * @ioc: per adapter object |
6692 | * |
6693 | * Creating host side data object, stored in ioc->sas_hba |
6694 | */ |
6695 | static void |
6696 | _scsih_sas_host_add(struct MPT3SAS_ADAPTER *ioc) |
6697 | { |
6698 | int i; |
6699 | Mpi2ConfigReply_t mpi_reply; |
6700 | Mpi2SasIOUnitPage0_t *sas_iounit_pg0 = NULL; |
6701 | Mpi2SasIOUnitPage1_t *sas_iounit_pg1 = NULL; |
6702 | Mpi2SasPhyPage0_t phy_pg0; |
6703 | Mpi2SasDevicePage0_t sas_device_pg0; |
6704 | Mpi2SasEnclosurePage0_t enclosure_pg0; |
6705 | u16 ioc_status; |
6706 | u16 sz; |
6707 | u8 device_missing_delay; |
6708 | u8 num_phys, port_id; |
6709 | struct hba_port *port; |
6710 | |
6711 | mpt3sas_config_get_number_hba_phys(ioc, num_phys: &num_phys); |
6712 | if (!num_phys) { |
6713 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6714 | __FILE__, __LINE__, __func__); |
6715 | return; |
6716 | } |
6717 | |
6718 | ioc->sas_hba.nr_phys_allocated = max_t(u8, |
6719 | MPT_MAX_HBA_NUM_PHYS, num_phys); |
6720 | ioc->sas_hba.phy = kcalloc(n: ioc->sas_hba.nr_phys_allocated, |
6721 | size: sizeof(struct _sas_phy), GFP_KERNEL); |
6722 | if (!ioc->sas_hba.phy) { |
6723 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6724 | __FILE__, __LINE__, __func__); |
6725 | goto out; |
6726 | } |
6727 | ioc->sas_hba.num_phys = num_phys; |
6728 | |
6729 | /* sas_iounit page 0 */ |
6730 | sz = struct_size(sas_iounit_pg0, PhyData, ioc->sas_hba.num_phys); |
6731 | sas_iounit_pg0 = kzalloc(size: sz, GFP_KERNEL); |
6732 | if (!sas_iounit_pg0) { |
6733 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6734 | __FILE__, __LINE__, __func__); |
6735 | return; |
6736 | } |
6737 | if ((mpt3sas_config_get_sas_iounit_pg0(ioc, mpi_reply: &mpi_reply, |
6738 | config_page: sas_iounit_pg0, sz))) { |
6739 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6740 | __FILE__, __LINE__, __func__); |
6741 | goto out; |
6742 | } |
6743 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & |
6744 | MPI2_IOCSTATUS_MASK; |
6745 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { |
6746 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6747 | __FILE__, __LINE__, __func__); |
6748 | goto out; |
6749 | } |
6750 | |
6751 | /* sas_iounit page 1 */ |
6752 | sz = struct_size(sas_iounit_pg1, PhyData, ioc->sas_hba.num_phys); |
6753 | sas_iounit_pg1 = kzalloc(size: sz, GFP_KERNEL); |
6754 | if (!sas_iounit_pg1) { |
6755 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6756 | __FILE__, __LINE__, __func__); |
6757 | goto out; |
6758 | } |
6759 | if ((mpt3sas_config_get_sas_iounit_pg1(ioc, mpi_reply: &mpi_reply, |
6760 | config_page: sas_iounit_pg1, sz))) { |
6761 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6762 | __FILE__, __LINE__, __func__); |
6763 | goto out; |
6764 | } |
6765 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & |
6766 | MPI2_IOCSTATUS_MASK; |
6767 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { |
6768 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6769 | __FILE__, __LINE__, __func__); |
6770 | goto out; |
6771 | } |
6772 | |
6773 | ioc->io_missing_delay = |
6774 | sas_iounit_pg1->IODeviceMissingDelay; |
6775 | device_missing_delay = |
6776 | sas_iounit_pg1->ReportDeviceMissingDelay; |
6777 | if (device_missing_delay & MPI2_SASIOUNIT1_REPORT_MISSING_UNIT_16) |
6778 | ioc->device_missing_delay = (device_missing_delay & |
6779 | MPI2_SASIOUNIT1_REPORT_MISSING_TIMEOUT_MASK) * 16; |
6780 | else |
6781 | ioc->device_missing_delay = device_missing_delay & |
6782 | MPI2_SASIOUNIT1_REPORT_MISSING_TIMEOUT_MASK; |
6783 | |
6784 | ioc->sas_hba.parent_dev = &ioc->shost->shost_gendev; |
6785 | for (i = 0; i < ioc->sas_hba.num_phys ; i++) { |
6786 | if ((mpt3sas_config_get_phy_pg0(ioc, mpi_reply: &mpi_reply, config_page: &phy_pg0, |
6787 | phy_number: i))) { |
6788 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6789 | __FILE__, __LINE__, __func__); |
6790 | goto out; |
6791 | } |
6792 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & |
6793 | MPI2_IOCSTATUS_MASK; |
6794 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { |
6795 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6796 | __FILE__, __LINE__, __func__); |
6797 | goto out; |
6798 | } |
6799 | |
6800 | if (i == 0) |
6801 | ioc->sas_hba.handle = le16_to_cpu(sas_iounit_pg0-> |
6802 | PhyData[0].ControllerDevHandle); |
6803 | |
6804 | port_id = sas_iounit_pg0->PhyData[i].Port; |
6805 | if (!(mpt3sas_get_port_by_id(ioc, port_id, bypass_dirty_port_flag: 0))) { |
6806 | port = kzalloc(size: sizeof(struct hba_port), GFP_KERNEL); |
6807 | if (!port) |
6808 | goto out; |
6809 | |
6810 | port->port_id = port_id; |
6811 | ioc_info(ioc, |
6812 | "hba_port entry: %p, port: %d is added to hba_port list\n" , |
6813 | port, port->port_id); |
6814 | list_add_tail(new: &port->list, |
6815 | head: &ioc->port_table_list); |
6816 | } |
6817 | |
6818 | /* |
6819 | * Check whether current Phy belongs to HBA vSES device or not. |
6820 | */ |
6821 | if ((le32_to_cpu(phy_pg0.PhyInfo) & |
6822 | MPI2_SAS_PHYINFO_VIRTUAL_PHY) && |
6823 | (phy_pg0.NegotiatedLinkRate >> 4) >= |
6824 | MPI2_SAS_NEG_LINK_RATE_1_5) { |
6825 | /* |
6826 | * Allocate a virtual_phy object for vSES device. |
6827 | */ |
6828 | if (!_scsih_alloc_vphy(ioc, port_id, phy_num: i)) |
6829 | goto out; |
6830 | ioc->sas_hba.phy[i].hba_vphy = 1; |
6831 | } |
6832 | |
6833 | ioc->sas_hba.phy[i].handle = ioc->sas_hba.handle; |
6834 | ioc->sas_hba.phy[i].phy_id = i; |
6835 | ioc->sas_hba.phy[i].port = |
6836 | mpt3sas_get_port_by_id(ioc, port_id, bypass_dirty_port_flag: 0); |
6837 | mpt3sas_transport_add_host_phy(ioc, mpt3sas_phy: &ioc->sas_hba.phy[i], |
6838 | phy_pg0, parent_dev: ioc->sas_hba.parent_dev); |
6839 | } |
6840 | if ((mpt3sas_config_get_sas_device_pg0(ioc, mpi_reply: &mpi_reply, config_page: &sas_device_pg0, |
6841 | MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle: ioc->sas_hba.handle))) { |
6842 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6843 | __FILE__, __LINE__, __func__); |
6844 | goto out; |
6845 | } |
6846 | ioc->sas_hba.enclosure_handle = |
6847 | le16_to_cpu(sas_device_pg0.EnclosureHandle); |
6848 | ioc->sas_hba.sas_address = le64_to_cpu(sas_device_pg0.SASAddress); |
6849 | ioc_info(ioc, "host_add: handle(0x%04x), sas_addr(0x%016llx), phys(%d)\n" , |
6850 | ioc->sas_hba.handle, |
6851 | (u64)ioc->sas_hba.sas_address, |
6852 | ioc->sas_hba.num_phys); |
6853 | |
6854 | if (ioc->sas_hba.enclosure_handle) { |
6855 | if (!(mpt3sas_config_get_enclosure_pg0(ioc, mpi_reply: &mpi_reply, |
6856 | config_page: &enclosure_pg0, MPI2_SAS_ENCLOS_PGAD_FORM_HANDLE, |
6857 | handle: ioc->sas_hba.enclosure_handle))) |
6858 | ioc->sas_hba.enclosure_logical_id = |
6859 | le64_to_cpu(enclosure_pg0.EnclosureLogicalID); |
6860 | } |
6861 | |
6862 | out: |
6863 | kfree(objp: sas_iounit_pg1); |
6864 | kfree(objp: sas_iounit_pg0); |
6865 | } |
6866 | |
6867 | /** |
6868 | * _scsih_expander_add - creating expander object |
6869 | * @ioc: per adapter object |
6870 | * @handle: expander handle |
6871 | * |
6872 | * Creating expander object, stored in ioc->sas_expander_list. |
6873 | * |
6874 | * Return: 0 for success, else error. |
6875 | */ |
6876 | static int |
6877 | _scsih_expander_add(struct MPT3SAS_ADAPTER *ioc, u16 handle) |
6878 | { |
6879 | struct _sas_node *sas_expander; |
6880 | struct _enclosure_node *enclosure_dev; |
6881 | Mpi2ConfigReply_t mpi_reply; |
6882 | Mpi2ExpanderPage0_t expander_pg0; |
6883 | Mpi2ExpanderPage1_t expander_pg1; |
6884 | u32 ioc_status; |
6885 | u16 parent_handle; |
6886 | u64 sas_address, sas_address_parent = 0; |
6887 | int i; |
6888 | unsigned long flags; |
6889 | struct _sas_port *mpt3sas_port = NULL; |
6890 | u8 port_id; |
6891 | |
6892 | int rc = 0; |
6893 | |
6894 | if (!handle) |
6895 | return -1; |
6896 | |
6897 | if (ioc->shost_recovery || ioc->pci_error_recovery) |
6898 | return -1; |
6899 | |
6900 | if ((mpt3sas_config_get_expander_pg0(ioc, mpi_reply: &mpi_reply, config_page: &expander_pg0, |
6901 | MPI2_SAS_EXPAND_PGAD_FORM_HNDL, handle))) { |
6902 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6903 | __FILE__, __LINE__, __func__); |
6904 | return -1; |
6905 | } |
6906 | |
6907 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & |
6908 | MPI2_IOCSTATUS_MASK; |
6909 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { |
6910 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6911 | __FILE__, __LINE__, __func__); |
6912 | return -1; |
6913 | } |
6914 | |
6915 | /* handle out of order topology events */ |
6916 | parent_handle = le16_to_cpu(expander_pg0.ParentDevHandle); |
6917 | if (_scsih_get_sas_address(ioc, handle: parent_handle, sas_address: &sas_address_parent) |
6918 | != 0) { |
6919 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6920 | __FILE__, __LINE__, __func__); |
6921 | return -1; |
6922 | } |
6923 | |
6924 | port_id = expander_pg0.PhysicalPort; |
6925 | if (sas_address_parent != ioc->sas_hba.sas_address) { |
6926 | spin_lock_irqsave(&ioc->sas_node_lock, flags); |
6927 | sas_expander = mpt3sas_scsih_expander_find_by_sas_address(ioc, |
6928 | sas_address: sas_address_parent, |
6929 | port: mpt3sas_get_port_by_id(ioc, port_id, bypass_dirty_port_flag: 0)); |
6930 | spin_unlock_irqrestore(lock: &ioc->sas_node_lock, flags); |
6931 | if (!sas_expander) { |
6932 | rc = _scsih_expander_add(ioc, handle: parent_handle); |
6933 | if (rc != 0) |
6934 | return rc; |
6935 | } |
6936 | } |
6937 | |
6938 | spin_lock_irqsave(&ioc->sas_node_lock, flags); |
6939 | sas_address = le64_to_cpu(expander_pg0.SASAddress); |
6940 | sas_expander = mpt3sas_scsih_expander_find_by_sas_address(ioc, |
6941 | sas_address, port: mpt3sas_get_port_by_id(ioc, port_id, bypass_dirty_port_flag: 0)); |
6942 | spin_unlock_irqrestore(lock: &ioc->sas_node_lock, flags); |
6943 | |
6944 | if (sas_expander) |
6945 | return 0; |
6946 | |
6947 | sas_expander = kzalloc(size: sizeof(struct _sas_node), |
6948 | GFP_KERNEL); |
6949 | if (!sas_expander) { |
6950 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6951 | __FILE__, __LINE__, __func__); |
6952 | return -1; |
6953 | } |
6954 | |
6955 | sas_expander->handle = handle; |
6956 | sas_expander->num_phys = expander_pg0.NumPhys; |
6957 | sas_expander->sas_address_parent = sas_address_parent; |
6958 | sas_expander->sas_address = sas_address; |
6959 | sas_expander->port = mpt3sas_get_port_by_id(ioc, port_id, bypass_dirty_port_flag: 0); |
6960 | if (!sas_expander->port) { |
6961 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6962 | __FILE__, __LINE__, __func__); |
6963 | rc = -1; |
6964 | goto out_fail; |
6965 | } |
6966 | |
6967 | ioc_info(ioc, "expander_add: handle(0x%04x), parent(0x%04x), sas_addr(0x%016llx), phys(%d)\n" , |
6968 | handle, parent_handle, |
6969 | (u64)sas_expander->sas_address, sas_expander->num_phys); |
6970 | |
6971 | if (!sas_expander->num_phys) { |
6972 | rc = -1; |
6973 | goto out_fail; |
6974 | } |
6975 | sas_expander->phy = kcalloc(n: sas_expander->num_phys, |
6976 | size: sizeof(struct _sas_phy), GFP_KERNEL); |
6977 | if (!sas_expander->phy) { |
6978 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6979 | __FILE__, __LINE__, __func__); |
6980 | rc = -1; |
6981 | goto out_fail; |
6982 | } |
6983 | |
6984 | INIT_LIST_HEAD(list: &sas_expander->sas_port_list); |
6985 | mpt3sas_port = mpt3sas_transport_port_add(ioc, handle, |
6986 | sas_address: sas_address_parent, port: sas_expander->port); |
6987 | if (!mpt3sas_port) { |
6988 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
6989 | __FILE__, __LINE__, __func__); |
6990 | rc = -1; |
6991 | goto out_fail; |
6992 | } |
6993 | sas_expander->parent_dev = &mpt3sas_port->rphy->dev; |
6994 | sas_expander->rphy = mpt3sas_port->rphy; |
6995 | |
6996 | for (i = 0 ; i < sas_expander->num_phys ; i++) { |
6997 | if ((mpt3sas_config_get_expander_pg1(ioc, mpi_reply: &mpi_reply, |
6998 | config_page: &expander_pg1, phy_number: i, handle))) { |
6999 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
7000 | __FILE__, __LINE__, __func__); |
7001 | rc = -1; |
7002 | goto out_fail; |
7003 | } |
7004 | sas_expander->phy[i].handle = handle; |
7005 | sas_expander->phy[i].phy_id = i; |
7006 | sas_expander->phy[i].port = |
7007 | mpt3sas_get_port_by_id(ioc, port_id, bypass_dirty_port_flag: 0); |
7008 | |
7009 | if ((mpt3sas_transport_add_expander_phy(ioc, |
7010 | mpt3sas_phy: &sas_expander->phy[i], expander_pg1, |
7011 | parent_dev: sas_expander->parent_dev))) { |
7012 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
7013 | __FILE__, __LINE__, __func__); |
7014 | rc = -1; |
7015 | goto out_fail; |
7016 | } |
7017 | } |
7018 | |
7019 | if (sas_expander->enclosure_handle) { |
7020 | enclosure_dev = |
7021 | mpt3sas_scsih_enclosure_find_by_handle(ioc, |
7022 | handle: sas_expander->enclosure_handle); |
7023 | if (enclosure_dev) |
7024 | sas_expander->enclosure_logical_id = |
7025 | le64_to_cpu(enclosure_dev->pg0.EnclosureLogicalID); |
7026 | } |
7027 | |
7028 | _scsih_expander_node_add(ioc, sas_expander); |
7029 | return 0; |
7030 | |
7031 | out_fail: |
7032 | |
7033 | if (mpt3sas_port) |
7034 | mpt3sas_transport_port_remove(ioc, sas_address: sas_expander->sas_address, |
7035 | sas_address_parent, port: sas_expander->port); |
7036 | kfree(objp: sas_expander); |
7037 | return rc; |
7038 | } |
7039 | |
7040 | /** |
7041 | * mpt3sas_expander_remove - removing expander object |
7042 | * @ioc: per adapter object |
7043 | * @sas_address: expander sas_address |
7044 | * @port: hba port entry |
7045 | */ |
7046 | void |
7047 | mpt3sas_expander_remove(struct MPT3SAS_ADAPTER *ioc, u64 sas_address, |
7048 | struct hba_port *port) |
7049 | { |
7050 | struct _sas_node *sas_expander; |
7051 | unsigned long flags; |
7052 | |
7053 | if (ioc->shost_recovery) |
7054 | return; |
7055 | |
7056 | if (!port) |
7057 | return; |
7058 | |
7059 | spin_lock_irqsave(&ioc->sas_node_lock, flags); |
7060 | sas_expander = mpt3sas_scsih_expander_find_by_sas_address(ioc, |
7061 | sas_address, port); |
7062 | spin_unlock_irqrestore(lock: &ioc->sas_node_lock, flags); |
7063 | if (sas_expander) |
7064 | _scsih_expander_node_remove(ioc, sas_expander); |
7065 | } |
7066 | |
7067 | /** |
7068 | * _scsih_done - internal SCSI_IO callback handler. |
7069 | * @ioc: per adapter object |
7070 | * @smid: system request message index |
7071 | * @msix_index: MSIX table index supplied by the OS |
7072 | * @reply: reply message frame(lower 32bit addr) |
7073 | * |
7074 | * Callback handler when sending internal generated SCSI_IO. |
7075 | * The callback index passed is `ioc->scsih_cb_idx` |
7076 | * |
7077 | * Return: 1 meaning mf should be freed from _base_interrupt |
7078 | * 0 means the mf is freed from this function. |
7079 | */ |
7080 | static u8 |
7081 | _scsih_done(struct MPT3SAS_ADAPTER *ioc, u16 smid, u8 msix_index, u32 reply) |
7082 | { |
7083 | MPI2DefaultReply_t *mpi_reply; |
7084 | |
7085 | mpi_reply = mpt3sas_base_get_reply_virt_addr(ioc, phys_addr: reply); |
7086 | if (ioc->scsih_cmds.status == MPT3_CMD_NOT_USED) |
7087 | return 1; |
7088 | if (ioc->scsih_cmds.smid != smid) |
7089 | return 1; |
7090 | ioc->scsih_cmds.status |= MPT3_CMD_COMPLETE; |
7091 | if (mpi_reply) { |
7092 | memcpy(ioc->scsih_cmds.reply, mpi_reply, |
7093 | mpi_reply->MsgLength*4); |
7094 | ioc->scsih_cmds.status |= MPT3_CMD_REPLY_VALID; |
7095 | } |
7096 | ioc->scsih_cmds.status &= ~MPT3_CMD_PENDING; |
7097 | complete(&ioc->scsih_cmds.done); |
7098 | return 1; |
7099 | } |
7100 | |
7101 | |
7102 | |
7103 | |
7104 | #define MPT3_MAX_LUNS (255) |
7105 | |
7106 | |
7107 | /** |
7108 | * _scsih_check_access_status - check access flags |
7109 | * @ioc: per adapter object |
7110 | * @sas_address: sas address |
7111 | * @handle: sas device handle |
7112 | * @access_status: errors returned during discovery of the device |
7113 | * |
7114 | * Return: 0 for success, else failure |
7115 | */ |
7116 | static u8 |
7117 | _scsih_check_access_status(struct MPT3SAS_ADAPTER *ioc, u64 sas_address, |
7118 | u16 handle, u8 access_status) |
7119 | { |
7120 | u8 rc = 1; |
7121 | char *desc = NULL; |
7122 | |
7123 | switch (access_status) { |
7124 | case MPI2_SAS_DEVICE0_ASTATUS_NO_ERRORS: |
7125 | case MPI2_SAS_DEVICE0_ASTATUS_SATA_NEEDS_INITIALIZATION: |
7126 | rc = 0; |
7127 | break; |
7128 | case MPI2_SAS_DEVICE0_ASTATUS_SATA_CAPABILITY_FAILED: |
7129 | desc = "sata capability failed" ; |
7130 | break; |
7131 | case MPI2_SAS_DEVICE0_ASTATUS_SATA_AFFILIATION_CONFLICT: |
7132 | desc = "sata affiliation conflict" ; |
7133 | break; |
7134 | case MPI2_SAS_DEVICE0_ASTATUS_ROUTE_NOT_ADDRESSABLE: |
7135 | desc = "route not addressable" ; |
7136 | break; |
7137 | case MPI2_SAS_DEVICE0_ASTATUS_SMP_ERROR_NOT_ADDRESSABLE: |
7138 | desc = "smp error not addressable" ; |
7139 | break; |
7140 | case MPI2_SAS_DEVICE0_ASTATUS_DEVICE_BLOCKED: |
7141 | desc = "device blocked" ; |
7142 | break; |
7143 | case MPI2_SAS_DEVICE0_ASTATUS_SATA_INIT_FAILED: |
7144 | case MPI2_SAS_DEVICE0_ASTATUS_SIF_UNKNOWN: |
7145 | case MPI2_SAS_DEVICE0_ASTATUS_SIF_AFFILIATION_CONFLICT: |
7146 | case MPI2_SAS_DEVICE0_ASTATUS_SIF_DIAG: |
7147 | case MPI2_SAS_DEVICE0_ASTATUS_SIF_IDENTIFICATION: |
7148 | case MPI2_SAS_DEVICE0_ASTATUS_SIF_CHECK_POWER: |
7149 | case MPI2_SAS_DEVICE0_ASTATUS_SIF_PIO_SN: |
7150 | case MPI2_SAS_DEVICE0_ASTATUS_SIF_MDMA_SN: |
7151 | case MPI2_SAS_DEVICE0_ASTATUS_SIF_UDMA_SN: |
7152 | case MPI2_SAS_DEVICE0_ASTATUS_SIF_ZONING_VIOLATION: |
7153 | case MPI2_SAS_DEVICE0_ASTATUS_SIF_NOT_ADDRESSABLE: |
7154 | case MPI2_SAS_DEVICE0_ASTATUS_SIF_MAX: |
7155 | desc = "sata initialization failed" ; |
7156 | break; |
7157 | default: |
7158 | desc = "unknown" ; |
7159 | break; |
7160 | } |
7161 | |
7162 | if (!rc) |
7163 | return 0; |
7164 | |
7165 | ioc_err(ioc, "discovery errors(%s): sas_address(0x%016llx), handle(0x%04x)\n" , |
7166 | desc, (u64)sas_address, handle); |
7167 | return rc; |
7168 | } |
7169 | |
7170 | /** |
7171 | * _scsih_check_device - checking device responsiveness |
7172 | * @ioc: per adapter object |
7173 | * @parent_sas_address: sas address of parent expander or sas host |
7174 | * @handle: attached device handle |
7175 | * @phy_number: phy number |
7176 | * @link_rate: new link rate |
7177 | */ |
7178 | static void |
7179 | _scsih_check_device(struct MPT3SAS_ADAPTER *ioc, |
7180 | u64 parent_sas_address, u16 handle, u8 phy_number, u8 link_rate) |
7181 | { |
7182 | Mpi2ConfigReply_t mpi_reply; |
7183 | Mpi2SasDevicePage0_t sas_device_pg0; |
7184 | struct _sas_device *sas_device = NULL; |
7185 | struct _enclosure_node *enclosure_dev = NULL; |
7186 | u32 ioc_status; |
7187 | unsigned long flags; |
7188 | u64 sas_address; |
7189 | struct scsi_target *starget; |
7190 | struct MPT3SAS_TARGET *sas_target_priv_data; |
7191 | u32 device_info; |
7192 | struct hba_port *port; |
7193 | |
7194 | if ((mpt3sas_config_get_sas_device_pg0(ioc, mpi_reply: &mpi_reply, config_page: &sas_device_pg0, |
7195 | MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) |
7196 | return; |
7197 | |
7198 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; |
7199 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) |
7200 | return; |
7201 | |
7202 | /* wide port handling ~ we need only handle device once for the phy that |
7203 | * is matched in sas device page zero |
7204 | */ |
7205 | if (phy_number != sas_device_pg0.PhyNum) |
7206 | return; |
7207 | |
7208 | /* check if this is end device */ |
7209 | device_info = le32_to_cpu(sas_device_pg0.DeviceInfo); |
7210 | if (!(_scsih_is_end_device(device_info))) |
7211 | return; |
7212 | |
7213 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
7214 | sas_address = le64_to_cpu(sas_device_pg0.SASAddress); |
7215 | port = mpt3sas_get_port_by_id(ioc, port_id: sas_device_pg0.PhysicalPort, bypass_dirty_port_flag: 0); |
7216 | if (!port) |
7217 | goto out_unlock; |
7218 | sas_device = __mpt3sas_get_sdev_by_addr(ioc, |
7219 | sas_address, port); |
7220 | |
7221 | if (!sas_device) |
7222 | goto out_unlock; |
7223 | |
7224 | if (unlikely(sas_device->handle != handle)) { |
7225 | starget = sas_device->starget; |
7226 | sas_target_priv_data = starget->hostdata; |
7227 | starget_printk(KERN_INFO, starget, |
7228 | "handle changed from(0x%04x) to (0x%04x)!!!\n" , |
7229 | sas_device->handle, handle); |
7230 | sas_target_priv_data->handle = handle; |
7231 | sas_device->handle = handle; |
7232 | if (le16_to_cpu(sas_device_pg0.Flags) & |
7233 | MPI2_SAS_DEVICE0_FLAGS_ENCL_LEVEL_VALID) { |
7234 | sas_device->enclosure_level = |
7235 | sas_device_pg0.EnclosureLevel; |
7236 | memcpy(sas_device->connector_name, |
7237 | sas_device_pg0.ConnectorName, 4); |
7238 | sas_device->connector_name[4] = '\0'; |
7239 | } else { |
7240 | sas_device->enclosure_level = 0; |
7241 | sas_device->connector_name[0] = '\0'; |
7242 | } |
7243 | |
7244 | sas_device->enclosure_handle = |
7245 | le16_to_cpu(sas_device_pg0.EnclosureHandle); |
7246 | sas_device->is_chassis_slot_valid = 0; |
7247 | enclosure_dev = mpt3sas_scsih_enclosure_find_by_handle(ioc, |
7248 | handle: sas_device->enclosure_handle); |
7249 | if (enclosure_dev) { |
7250 | sas_device->enclosure_logical_id = |
7251 | le64_to_cpu(enclosure_dev->pg0.EnclosureLogicalID); |
7252 | if (le16_to_cpu(enclosure_dev->pg0.Flags) & |
7253 | MPI2_SAS_ENCLS0_FLAGS_CHASSIS_SLOT_VALID) { |
7254 | sas_device->is_chassis_slot_valid = 1; |
7255 | sas_device->chassis_slot = |
7256 | enclosure_dev->pg0.ChassisSlot; |
7257 | } |
7258 | } |
7259 | } |
7260 | |
7261 | /* check if device is present */ |
7262 | if (!(le16_to_cpu(sas_device_pg0.Flags) & |
7263 | MPI2_SAS_DEVICE0_FLAGS_DEVICE_PRESENT)) { |
7264 | ioc_err(ioc, "device is not present handle(0x%04x), flags!!!\n" , |
7265 | handle); |
7266 | goto out_unlock; |
7267 | } |
7268 | |
7269 | /* check if there were any issues with discovery */ |
7270 | if (_scsih_check_access_status(ioc, sas_address, handle, |
7271 | access_status: sas_device_pg0.AccessStatus)) |
7272 | goto out_unlock; |
7273 | |
7274 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
7275 | _scsih_ublock_io_device(ioc, sas_address, port); |
7276 | |
7277 | if (sas_device) |
7278 | sas_device_put(s: sas_device); |
7279 | return; |
7280 | |
7281 | out_unlock: |
7282 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
7283 | if (sas_device) |
7284 | sas_device_put(s: sas_device); |
7285 | } |
7286 | |
7287 | /** |
7288 | * _scsih_add_device - creating sas device object |
7289 | * @ioc: per adapter object |
7290 | * @handle: sas device handle |
7291 | * @phy_num: phy number end device attached to |
7292 | * @is_pd: is this hidden raid component |
7293 | * |
7294 | * Creating end device object, stored in ioc->sas_device_list. |
7295 | * |
7296 | * Return: 0 for success, non-zero for failure. |
7297 | */ |
7298 | static int |
7299 | _scsih_add_device(struct MPT3SAS_ADAPTER *ioc, u16 handle, u8 phy_num, |
7300 | u8 is_pd) |
7301 | { |
7302 | Mpi2ConfigReply_t mpi_reply; |
7303 | Mpi2SasDevicePage0_t sas_device_pg0; |
7304 | struct _sas_device *sas_device; |
7305 | struct _enclosure_node *enclosure_dev = NULL; |
7306 | u32 ioc_status; |
7307 | u64 sas_address; |
7308 | u32 device_info; |
7309 | u8 port_id; |
7310 | |
7311 | if ((mpt3sas_config_get_sas_device_pg0(ioc, mpi_reply: &mpi_reply, config_page: &sas_device_pg0, |
7312 | MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) { |
7313 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
7314 | __FILE__, __LINE__, __func__); |
7315 | return -1; |
7316 | } |
7317 | |
7318 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & |
7319 | MPI2_IOCSTATUS_MASK; |
7320 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { |
7321 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
7322 | __FILE__, __LINE__, __func__); |
7323 | return -1; |
7324 | } |
7325 | |
7326 | /* check if this is end device */ |
7327 | device_info = le32_to_cpu(sas_device_pg0.DeviceInfo); |
7328 | if (!(_scsih_is_end_device(device_info))) |
7329 | return -1; |
7330 | set_bit(nr: handle, addr: ioc->pend_os_device_add); |
7331 | sas_address = le64_to_cpu(sas_device_pg0.SASAddress); |
7332 | |
7333 | /* check if device is present */ |
7334 | if (!(le16_to_cpu(sas_device_pg0.Flags) & |
7335 | MPI2_SAS_DEVICE0_FLAGS_DEVICE_PRESENT)) { |
7336 | ioc_err(ioc, "device is not present handle(0x04%x)!!!\n" , |
7337 | handle); |
7338 | return -1; |
7339 | } |
7340 | |
7341 | /* check if there were any issues with discovery */ |
7342 | if (_scsih_check_access_status(ioc, sas_address, handle, |
7343 | access_status: sas_device_pg0.AccessStatus)) |
7344 | return -1; |
7345 | |
7346 | port_id = sas_device_pg0.PhysicalPort; |
7347 | sas_device = mpt3sas_get_sdev_by_addr(ioc, |
7348 | sas_address, port: mpt3sas_get_port_by_id(ioc, port_id, bypass_dirty_port_flag: 0)); |
7349 | if (sas_device) { |
7350 | clear_bit(nr: handle, addr: ioc->pend_os_device_add); |
7351 | sas_device_put(s: sas_device); |
7352 | return -1; |
7353 | } |
7354 | |
7355 | if (sas_device_pg0.EnclosureHandle) { |
7356 | enclosure_dev = |
7357 | mpt3sas_scsih_enclosure_find_by_handle(ioc, |
7358 | le16_to_cpu(sas_device_pg0.EnclosureHandle)); |
7359 | if (enclosure_dev == NULL) |
7360 | ioc_info(ioc, "Enclosure handle(0x%04x) doesn't match with enclosure device!\n" , |
7361 | sas_device_pg0.EnclosureHandle); |
7362 | } |
7363 | |
7364 | sas_device = kzalloc(size: sizeof(struct _sas_device), |
7365 | GFP_KERNEL); |
7366 | if (!sas_device) { |
7367 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
7368 | __FILE__, __LINE__, __func__); |
7369 | return 0; |
7370 | } |
7371 | |
7372 | kref_init(kref: &sas_device->refcount); |
7373 | sas_device->handle = handle; |
7374 | if (_scsih_get_sas_address(ioc, |
7375 | le16_to_cpu(sas_device_pg0.ParentDevHandle), |
7376 | sas_address: &sas_device->sas_address_parent) != 0) |
7377 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
7378 | __FILE__, __LINE__, __func__); |
7379 | sas_device->enclosure_handle = |
7380 | le16_to_cpu(sas_device_pg0.EnclosureHandle); |
7381 | if (sas_device->enclosure_handle != 0) |
7382 | sas_device->slot = |
7383 | le16_to_cpu(sas_device_pg0.Slot); |
7384 | sas_device->device_info = device_info; |
7385 | sas_device->sas_address = sas_address; |
7386 | sas_device->phy = sas_device_pg0.PhyNum; |
7387 | sas_device->fast_path = (le16_to_cpu(sas_device_pg0.Flags) & |
7388 | MPI25_SAS_DEVICE0_FLAGS_FAST_PATH_CAPABLE) ? 1 : 0; |
7389 | sas_device->port = mpt3sas_get_port_by_id(ioc, port_id, bypass_dirty_port_flag: 0); |
7390 | if (!sas_device->port) { |
7391 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
7392 | __FILE__, __LINE__, __func__); |
7393 | goto out; |
7394 | } |
7395 | |
7396 | if (le16_to_cpu(sas_device_pg0.Flags) |
7397 | & MPI2_SAS_DEVICE0_FLAGS_ENCL_LEVEL_VALID) { |
7398 | sas_device->enclosure_level = |
7399 | sas_device_pg0.EnclosureLevel; |
7400 | memcpy(sas_device->connector_name, |
7401 | sas_device_pg0.ConnectorName, 4); |
7402 | sas_device->connector_name[4] = '\0'; |
7403 | } else { |
7404 | sas_device->enclosure_level = 0; |
7405 | sas_device->connector_name[0] = '\0'; |
7406 | } |
7407 | /* get enclosure_logical_id & chassis_slot*/ |
7408 | sas_device->is_chassis_slot_valid = 0; |
7409 | if (enclosure_dev) { |
7410 | sas_device->enclosure_logical_id = |
7411 | le64_to_cpu(enclosure_dev->pg0.EnclosureLogicalID); |
7412 | if (le16_to_cpu(enclosure_dev->pg0.Flags) & |
7413 | MPI2_SAS_ENCLS0_FLAGS_CHASSIS_SLOT_VALID) { |
7414 | sas_device->is_chassis_slot_valid = 1; |
7415 | sas_device->chassis_slot = |
7416 | enclosure_dev->pg0.ChassisSlot; |
7417 | } |
7418 | } |
7419 | |
7420 | /* get device name */ |
7421 | sas_device->device_name = le64_to_cpu(sas_device_pg0.DeviceName); |
7422 | sas_device->port_type = sas_device_pg0.MaxPortConnections; |
7423 | ioc_info(ioc, |
7424 | "handle(0x%0x) sas_address(0x%016llx) port_type(0x%0x)\n" , |
7425 | handle, sas_device->sas_address, sas_device->port_type); |
7426 | |
7427 | if (ioc->wait_for_discovery_to_complete) |
7428 | _scsih_sas_device_init_add(ioc, sas_device); |
7429 | else |
7430 | _scsih_sas_device_add(ioc, sas_device); |
7431 | |
7432 | out: |
7433 | sas_device_put(s: sas_device); |
7434 | return 0; |
7435 | } |
7436 | |
7437 | /** |
7438 | * _scsih_remove_device - removing sas device object |
7439 | * @ioc: per adapter object |
7440 | * @sas_device: the sas_device object |
7441 | */ |
7442 | static void |
7443 | _scsih_remove_device(struct MPT3SAS_ADAPTER *ioc, |
7444 | struct _sas_device *sas_device) |
7445 | { |
7446 | struct MPT3SAS_TARGET *sas_target_priv_data; |
7447 | |
7448 | if ((ioc->pdev->subsystem_vendor == PCI_VENDOR_ID_IBM) && |
7449 | (sas_device->pfa_led_on)) { |
7450 | _scsih_turn_off_pfa_led(ioc, sas_device); |
7451 | sas_device->pfa_led_on = 0; |
7452 | } |
7453 | |
7454 | dewtprintk(ioc, |
7455 | ioc_info(ioc, "%s: enter: handle(0x%04x), sas_addr(0x%016llx)\n" , |
7456 | __func__, |
7457 | sas_device->handle, (u64)sas_device->sas_address)); |
7458 | |
7459 | dewtprintk(ioc, _scsih_display_enclosure_chassis_info(ioc, sas_device, |
7460 | NULL, NULL)); |
7461 | |
7462 | if (sas_device->starget && sas_device->starget->hostdata) { |
7463 | sas_target_priv_data = sas_device->starget->hostdata; |
7464 | sas_target_priv_data->deleted = 1; |
7465 | _scsih_ublock_io_device(ioc, sas_address: sas_device->sas_address, |
7466 | port: sas_device->port); |
7467 | sas_target_priv_data->handle = |
7468 | MPT3SAS_INVALID_DEVICE_HANDLE; |
7469 | } |
7470 | |
7471 | if (!ioc->hide_drives) |
7472 | mpt3sas_transport_port_remove(ioc, |
7473 | sas_address: sas_device->sas_address, |
7474 | sas_address_parent: sas_device->sas_address_parent, |
7475 | port: sas_device->port); |
7476 | |
7477 | ioc_info(ioc, "removing handle(0x%04x), sas_addr(0x%016llx)\n" , |
7478 | sas_device->handle, (u64)sas_device->sas_address); |
7479 | |
7480 | _scsih_display_enclosure_chassis_info(ioc, sas_device, NULL, NULL); |
7481 | |
7482 | dewtprintk(ioc, |
7483 | ioc_info(ioc, "%s: exit: handle(0x%04x), sas_addr(0x%016llx)\n" , |
7484 | __func__, |
7485 | sas_device->handle, (u64)sas_device->sas_address)); |
7486 | dewtprintk(ioc, _scsih_display_enclosure_chassis_info(ioc, sas_device, |
7487 | NULL, NULL)); |
7488 | } |
7489 | |
7490 | /** |
7491 | * _scsih_sas_topology_change_event_debug - debug for topology event |
7492 | * @ioc: per adapter object |
7493 | * @event_data: event data payload |
7494 | * Context: user. |
7495 | */ |
7496 | static void |
7497 | _scsih_sas_topology_change_event_debug(struct MPT3SAS_ADAPTER *ioc, |
7498 | Mpi2EventDataSasTopologyChangeList_t *event_data) |
7499 | { |
7500 | int i; |
7501 | u16 handle; |
7502 | u16 reason_code; |
7503 | u8 phy_number; |
7504 | char *status_str = NULL; |
7505 | u8 link_rate, prev_link_rate; |
7506 | |
7507 | switch (event_data->ExpStatus) { |
7508 | case MPI2_EVENT_SAS_TOPO_ES_ADDED: |
7509 | status_str = "add" ; |
7510 | break; |
7511 | case MPI2_EVENT_SAS_TOPO_ES_NOT_RESPONDING: |
7512 | status_str = "remove" ; |
7513 | break; |
7514 | case MPI2_EVENT_SAS_TOPO_ES_RESPONDING: |
7515 | case 0: |
7516 | status_str = "responding" ; |
7517 | break; |
7518 | case MPI2_EVENT_SAS_TOPO_ES_DELAY_NOT_RESPONDING: |
7519 | status_str = "remove delay" ; |
7520 | break; |
7521 | default: |
7522 | status_str = "unknown status" ; |
7523 | break; |
7524 | } |
7525 | ioc_info(ioc, "sas topology change: (%s)\n" , status_str); |
7526 | pr_info("\thandle(0x%04x), enclosure_handle(0x%04x) " \ |
7527 | "start_phy(%02d), count(%d)\n" , |
7528 | le16_to_cpu(event_data->ExpanderDevHandle), |
7529 | le16_to_cpu(event_data->EnclosureHandle), |
7530 | event_data->StartPhyNum, event_data->NumEntries); |
7531 | for (i = 0; i < event_data->NumEntries; i++) { |
7532 | handle = le16_to_cpu(event_data->PHY[i].AttachedDevHandle); |
7533 | if (!handle) |
7534 | continue; |
7535 | phy_number = event_data->StartPhyNum + i; |
7536 | reason_code = event_data->PHY[i].PhyStatus & |
7537 | MPI2_EVENT_SAS_TOPO_RC_MASK; |
7538 | switch (reason_code) { |
7539 | case MPI2_EVENT_SAS_TOPO_RC_TARG_ADDED: |
7540 | status_str = "target add" ; |
7541 | break; |
7542 | case MPI2_EVENT_SAS_TOPO_RC_TARG_NOT_RESPONDING: |
7543 | status_str = "target remove" ; |
7544 | break; |
7545 | case MPI2_EVENT_SAS_TOPO_RC_DELAY_NOT_RESPONDING: |
7546 | status_str = "delay target remove" ; |
7547 | break; |
7548 | case MPI2_EVENT_SAS_TOPO_RC_PHY_CHANGED: |
7549 | status_str = "link rate change" ; |
7550 | break; |
7551 | case MPI2_EVENT_SAS_TOPO_RC_NO_CHANGE: |
7552 | status_str = "target responding" ; |
7553 | break; |
7554 | default: |
7555 | status_str = "unknown" ; |
7556 | break; |
7557 | } |
7558 | link_rate = event_data->PHY[i].LinkRate >> 4; |
7559 | prev_link_rate = event_data->PHY[i].LinkRate & 0xF; |
7560 | pr_info("\tphy(%02d), attached_handle(0x%04x): %s:" \ |
7561 | " link rate: new(0x%02x), old(0x%02x)\n" , phy_number, |
7562 | handle, status_str, link_rate, prev_link_rate); |
7563 | |
7564 | } |
7565 | } |
7566 | |
7567 | /** |
7568 | * _scsih_sas_topology_change_event - handle topology changes |
7569 | * @ioc: per adapter object |
7570 | * @fw_event: The fw_event_work object |
7571 | * Context: user. |
7572 | * |
7573 | */ |
7574 | static int |
7575 | _scsih_sas_topology_change_event(struct MPT3SAS_ADAPTER *ioc, |
7576 | struct fw_event_work *fw_event) |
7577 | { |
7578 | int i; |
7579 | u16 parent_handle, handle; |
7580 | u16 reason_code; |
7581 | u8 phy_number, max_phys; |
7582 | struct _sas_node *sas_expander; |
7583 | u64 sas_address; |
7584 | unsigned long flags; |
7585 | u8 link_rate, prev_link_rate; |
7586 | struct hba_port *port; |
7587 | Mpi2EventDataSasTopologyChangeList_t *event_data = |
7588 | (Mpi2EventDataSasTopologyChangeList_t *) |
7589 | fw_event->event_data; |
7590 | |
7591 | if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) |
7592 | _scsih_sas_topology_change_event_debug(ioc, event_data); |
7593 | |
7594 | if (ioc->shost_recovery || ioc->remove_host || ioc->pci_error_recovery) |
7595 | return 0; |
7596 | |
7597 | if (!ioc->sas_hba.num_phys) |
7598 | _scsih_sas_host_add(ioc); |
7599 | else |
7600 | _scsih_sas_host_refresh(ioc); |
7601 | |
7602 | if (fw_event->ignore) { |
7603 | dewtprintk(ioc, ioc_info(ioc, "ignoring expander event\n" )); |
7604 | return 0; |
7605 | } |
7606 | |
7607 | parent_handle = le16_to_cpu(event_data->ExpanderDevHandle); |
7608 | port = mpt3sas_get_port_by_id(ioc, port_id: event_data->PhysicalPort, bypass_dirty_port_flag: 0); |
7609 | |
7610 | /* handle expander add */ |
7611 | if (event_data->ExpStatus == MPI2_EVENT_SAS_TOPO_ES_ADDED) |
7612 | if (_scsih_expander_add(ioc, handle: parent_handle) != 0) |
7613 | return 0; |
7614 | |
7615 | spin_lock_irqsave(&ioc->sas_node_lock, flags); |
7616 | sas_expander = mpt3sas_scsih_expander_find_by_handle(ioc, |
7617 | handle: parent_handle); |
7618 | if (sas_expander) { |
7619 | sas_address = sas_expander->sas_address; |
7620 | max_phys = sas_expander->num_phys; |
7621 | port = sas_expander->port; |
7622 | } else if (parent_handle < ioc->sas_hba.num_phys) { |
7623 | sas_address = ioc->sas_hba.sas_address; |
7624 | max_phys = ioc->sas_hba.num_phys; |
7625 | } else { |
7626 | spin_unlock_irqrestore(lock: &ioc->sas_node_lock, flags); |
7627 | return 0; |
7628 | } |
7629 | spin_unlock_irqrestore(lock: &ioc->sas_node_lock, flags); |
7630 | |
7631 | /* handle siblings events */ |
7632 | for (i = 0; i < event_data->NumEntries; i++) { |
7633 | if (fw_event->ignore) { |
7634 | dewtprintk(ioc, |
7635 | ioc_info(ioc, "ignoring expander event\n" )); |
7636 | return 0; |
7637 | } |
7638 | if (ioc->remove_host || ioc->pci_error_recovery) |
7639 | return 0; |
7640 | phy_number = event_data->StartPhyNum + i; |
7641 | if (phy_number >= max_phys) |
7642 | continue; |
7643 | reason_code = event_data->PHY[i].PhyStatus & |
7644 | MPI2_EVENT_SAS_TOPO_RC_MASK; |
7645 | if ((event_data->PHY[i].PhyStatus & |
7646 | MPI2_EVENT_SAS_TOPO_PHYSTATUS_VACANT) && (reason_code != |
7647 | MPI2_EVENT_SAS_TOPO_RC_TARG_NOT_RESPONDING)) |
7648 | continue; |
7649 | handle = le16_to_cpu(event_data->PHY[i].AttachedDevHandle); |
7650 | if (!handle) |
7651 | continue; |
7652 | link_rate = event_data->PHY[i].LinkRate >> 4; |
7653 | prev_link_rate = event_data->PHY[i].LinkRate & 0xF; |
7654 | switch (reason_code) { |
7655 | case MPI2_EVENT_SAS_TOPO_RC_PHY_CHANGED: |
7656 | |
7657 | if (ioc->shost_recovery) |
7658 | break; |
7659 | |
7660 | if (link_rate == prev_link_rate) |
7661 | break; |
7662 | |
7663 | mpt3sas_transport_update_links(ioc, sas_address, |
7664 | handle, phy_number, link_rate, port); |
7665 | |
7666 | if (link_rate < MPI2_SAS_NEG_LINK_RATE_1_5) |
7667 | break; |
7668 | |
7669 | _scsih_check_device(ioc, parent_sas_address: sas_address, handle, |
7670 | phy_number, link_rate); |
7671 | |
7672 | if (!test_bit(handle, ioc->pend_os_device_add)) |
7673 | break; |
7674 | |
7675 | fallthrough; |
7676 | |
7677 | case MPI2_EVENT_SAS_TOPO_RC_TARG_ADDED: |
7678 | |
7679 | if (ioc->shost_recovery) |
7680 | break; |
7681 | |
7682 | mpt3sas_transport_update_links(ioc, sas_address, |
7683 | handle, phy_number, link_rate, port); |
7684 | |
7685 | _scsih_add_device(ioc, handle, phy_num: phy_number, is_pd: 0); |
7686 | |
7687 | break; |
7688 | case MPI2_EVENT_SAS_TOPO_RC_TARG_NOT_RESPONDING: |
7689 | |
7690 | _scsih_device_remove_by_handle(ioc, handle); |
7691 | break; |
7692 | } |
7693 | } |
7694 | |
7695 | /* handle expander removal */ |
7696 | if (event_data->ExpStatus == MPI2_EVENT_SAS_TOPO_ES_NOT_RESPONDING && |
7697 | sas_expander) |
7698 | mpt3sas_expander_remove(ioc, sas_address, port); |
7699 | |
7700 | return 0; |
7701 | } |
7702 | |
7703 | /** |
7704 | * _scsih_sas_device_status_change_event_debug - debug for device event |
7705 | * @ioc: ? |
7706 | * @event_data: event data payload |
7707 | * Context: user. |
7708 | */ |
7709 | static void |
7710 | _scsih_sas_device_status_change_event_debug(struct MPT3SAS_ADAPTER *ioc, |
7711 | Mpi2EventDataSasDeviceStatusChange_t *event_data) |
7712 | { |
7713 | char *reason_str = NULL; |
7714 | |
7715 | switch (event_data->ReasonCode) { |
7716 | case MPI2_EVENT_SAS_DEV_STAT_RC_SMART_DATA: |
7717 | reason_str = "smart data" ; |
7718 | break; |
7719 | case MPI2_EVENT_SAS_DEV_STAT_RC_UNSUPPORTED: |
7720 | reason_str = "unsupported device discovered" ; |
7721 | break; |
7722 | case MPI2_EVENT_SAS_DEV_STAT_RC_INTERNAL_DEVICE_RESET: |
7723 | reason_str = "internal device reset" ; |
7724 | break; |
7725 | case MPI2_EVENT_SAS_DEV_STAT_RC_TASK_ABORT_INTERNAL: |
7726 | reason_str = "internal task abort" ; |
7727 | break; |
7728 | case MPI2_EVENT_SAS_DEV_STAT_RC_ABORT_TASK_SET_INTERNAL: |
7729 | reason_str = "internal task abort set" ; |
7730 | break; |
7731 | case MPI2_EVENT_SAS_DEV_STAT_RC_CLEAR_TASK_SET_INTERNAL: |
7732 | reason_str = "internal clear task set" ; |
7733 | break; |
7734 | case MPI2_EVENT_SAS_DEV_STAT_RC_QUERY_TASK_INTERNAL: |
7735 | reason_str = "internal query task" ; |
7736 | break; |
7737 | case MPI2_EVENT_SAS_DEV_STAT_RC_SATA_INIT_FAILURE: |
7738 | reason_str = "sata init failure" ; |
7739 | break; |
7740 | case MPI2_EVENT_SAS_DEV_STAT_RC_CMP_INTERNAL_DEV_RESET: |
7741 | reason_str = "internal device reset complete" ; |
7742 | break; |
7743 | case MPI2_EVENT_SAS_DEV_STAT_RC_CMP_TASK_ABORT_INTERNAL: |
7744 | reason_str = "internal task abort complete" ; |
7745 | break; |
7746 | case MPI2_EVENT_SAS_DEV_STAT_RC_ASYNC_NOTIFICATION: |
7747 | reason_str = "internal async notification" ; |
7748 | break; |
7749 | case MPI2_EVENT_SAS_DEV_STAT_RC_EXPANDER_REDUCED_FUNCTIONALITY: |
7750 | reason_str = "expander reduced functionality" ; |
7751 | break; |
7752 | case MPI2_EVENT_SAS_DEV_STAT_RC_CMP_EXPANDER_REDUCED_FUNCTIONALITY: |
7753 | reason_str = "expander reduced functionality complete" ; |
7754 | break; |
7755 | default: |
7756 | reason_str = "unknown reason" ; |
7757 | break; |
7758 | } |
7759 | ioc_info(ioc, "device status change: (%s)\thandle(0x%04x), sas address(0x%016llx), tag(%d)" , |
7760 | reason_str, le16_to_cpu(event_data->DevHandle), |
7761 | (u64)le64_to_cpu(event_data->SASAddress), |
7762 | le16_to_cpu(event_data->TaskTag)); |
7763 | if (event_data->ReasonCode == MPI2_EVENT_SAS_DEV_STAT_RC_SMART_DATA) |
7764 | pr_cont(", ASC(0x%x), ASCQ(0x%x)\n" , |
7765 | event_data->ASC, event_data->ASCQ); |
7766 | pr_cont("\n" ); |
7767 | } |
7768 | |
7769 | /** |
7770 | * _scsih_sas_device_status_change_event - handle device status change |
7771 | * @ioc: per adapter object |
7772 | * @event_data: The fw event |
7773 | * Context: user. |
7774 | */ |
7775 | static void |
7776 | _scsih_sas_device_status_change_event(struct MPT3SAS_ADAPTER *ioc, |
7777 | Mpi2EventDataSasDeviceStatusChange_t *event_data) |
7778 | { |
7779 | struct MPT3SAS_TARGET *target_priv_data; |
7780 | struct _sas_device *sas_device; |
7781 | u64 sas_address; |
7782 | unsigned long flags; |
7783 | |
7784 | /* In MPI Revision K (0xC), the internal device reset complete was |
7785 | * implemented, so avoid setting tm_busy flag for older firmware. |
7786 | */ |
7787 | if ((ioc->facts.HeaderVersion >> 8) < 0xC) |
7788 | return; |
7789 | |
7790 | if (event_data->ReasonCode != |
7791 | MPI2_EVENT_SAS_DEV_STAT_RC_INTERNAL_DEVICE_RESET && |
7792 | event_data->ReasonCode != |
7793 | MPI2_EVENT_SAS_DEV_STAT_RC_CMP_INTERNAL_DEV_RESET) |
7794 | return; |
7795 | |
7796 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
7797 | sas_address = le64_to_cpu(event_data->SASAddress); |
7798 | sas_device = __mpt3sas_get_sdev_by_addr(ioc, |
7799 | sas_address, |
7800 | port: mpt3sas_get_port_by_id(ioc, port_id: event_data->PhysicalPort, bypass_dirty_port_flag: 0)); |
7801 | |
7802 | if (!sas_device || !sas_device->starget) |
7803 | goto out; |
7804 | |
7805 | target_priv_data = sas_device->starget->hostdata; |
7806 | if (!target_priv_data) |
7807 | goto out; |
7808 | |
7809 | if (event_data->ReasonCode == |
7810 | MPI2_EVENT_SAS_DEV_STAT_RC_INTERNAL_DEVICE_RESET) |
7811 | target_priv_data->tm_busy = 1; |
7812 | else |
7813 | target_priv_data->tm_busy = 0; |
7814 | |
7815 | if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) |
7816 | ioc_info(ioc, |
7817 | "%s tm_busy flag for handle(0x%04x)\n" , |
7818 | (target_priv_data->tm_busy == 1) ? "Enable" : "Disable" , |
7819 | target_priv_data->handle); |
7820 | |
7821 | out: |
7822 | if (sas_device) |
7823 | sas_device_put(s: sas_device); |
7824 | |
7825 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
7826 | } |
7827 | |
7828 | |
7829 | /** |
7830 | * _scsih_check_pcie_access_status - check access flags |
7831 | * @ioc: per adapter object |
7832 | * @wwid: wwid |
7833 | * @handle: sas device handle |
7834 | * @access_status: errors returned during discovery of the device |
7835 | * |
7836 | * Return: 0 for success, else failure |
7837 | */ |
7838 | static u8 |
7839 | _scsih_check_pcie_access_status(struct MPT3SAS_ADAPTER *ioc, u64 wwid, |
7840 | u16 handle, u8 access_status) |
7841 | { |
7842 | u8 rc = 1; |
7843 | char *desc = NULL; |
7844 | |
7845 | switch (access_status) { |
7846 | case MPI26_PCIEDEV0_ASTATUS_NO_ERRORS: |
7847 | case MPI26_PCIEDEV0_ASTATUS_NEEDS_INITIALIZATION: |
7848 | rc = 0; |
7849 | break; |
7850 | case MPI26_PCIEDEV0_ASTATUS_CAPABILITY_FAILED: |
7851 | desc = "PCIe device capability failed" ; |
7852 | break; |
7853 | case MPI26_PCIEDEV0_ASTATUS_DEVICE_BLOCKED: |
7854 | desc = "PCIe device blocked" ; |
7855 | ioc_info(ioc, |
7856 | "Device with Access Status (%s): wwid(0x%016llx), " |
7857 | "handle(0x%04x)\n ll only be added to the internal list" , |
7858 | desc, (u64)wwid, handle); |
7859 | rc = 0; |
7860 | break; |
7861 | case MPI26_PCIEDEV0_ASTATUS_MEMORY_SPACE_ACCESS_FAILED: |
7862 | desc = "PCIe device mem space access failed" ; |
7863 | break; |
7864 | case MPI26_PCIEDEV0_ASTATUS_UNSUPPORTED_DEVICE: |
7865 | desc = "PCIe device unsupported" ; |
7866 | break; |
7867 | case MPI26_PCIEDEV0_ASTATUS_MSIX_REQUIRED: |
7868 | desc = "PCIe device MSIx Required" ; |
7869 | break; |
7870 | case MPI26_PCIEDEV0_ASTATUS_INIT_FAIL_MAX: |
7871 | desc = "PCIe device init fail max" ; |
7872 | break; |
7873 | case MPI26_PCIEDEV0_ASTATUS_UNKNOWN: |
7874 | desc = "PCIe device status unknown" ; |
7875 | break; |
7876 | case MPI26_PCIEDEV0_ASTATUS_NVME_READY_TIMEOUT: |
7877 | desc = "nvme ready timeout" ; |
7878 | break; |
7879 | case MPI26_PCIEDEV0_ASTATUS_NVME_DEVCFG_UNSUPPORTED: |
7880 | desc = "nvme device configuration unsupported" ; |
7881 | break; |
7882 | case MPI26_PCIEDEV0_ASTATUS_NVME_IDENTIFY_FAILED: |
7883 | desc = "nvme identify failed" ; |
7884 | break; |
7885 | case MPI26_PCIEDEV0_ASTATUS_NVME_QCONFIG_FAILED: |
7886 | desc = "nvme qconfig failed" ; |
7887 | break; |
7888 | case MPI26_PCIEDEV0_ASTATUS_NVME_QCREATION_FAILED: |
7889 | desc = "nvme qcreation failed" ; |
7890 | break; |
7891 | case MPI26_PCIEDEV0_ASTATUS_NVME_EVENTCFG_FAILED: |
7892 | desc = "nvme eventcfg failed" ; |
7893 | break; |
7894 | case MPI26_PCIEDEV0_ASTATUS_NVME_GET_FEATURE_STAT_FAILED: |
7895 | desc = "nvme get feature stat failed" ; |
7896 | break; |
7897 | case MPI26_PCIEDEV0_ASTATUS_NVME_IDLE_TIMEOUT: |
7898 | desc = "nvme idle timeout" ; |
7899 | break; |
7900 | case MPI26_PCIEDEV0_ASTATUS_NVME_FAILURE_STATUS: |
7901 | desc = "nvme failure status" ; |
7902 | break; |
7903 | default: |
7904 | ioc_err(ioc, "NVMe discovery error(0x%02x): wwid(0x%016llx), handle(0x%04x)\n" , |
7905 | access_status, (u64)wwid, handle); |
7906 | return rc; |
7907 | } |
7908 | |
7909 | if (!rc) |
7910 | return rc; |
7911 | |
7912 | ioc_info(ioc, "NVMe discovery error(%s): wwid(0x%016llx), handle(0x%04x)\n" , |
7913 | desc, (u64)wwid, handle); |
7914 | return rc; |
7915 | } |
7916 | |
7917 | /** |
7918 | * _scsih_pcie_device_remove_from_sml - removing pcie device |
7919 | * from SML and free up associated memory |
7920 | * @ioc: per adapter object |
7921 | * @pcie_device: the pcie_device object |
7922 | */ |
7923 | static void |
7924 | _scsih_pcie_device_remove_from_sml(struct MPT3SAS_ADAPTER *ioc, |
7925 | struct _pcie_device *pcie_device) |
7926 | { |
7927 | struct MPT3SAS_TARGET *sas_target_priv_data; |
7928 | |
7929 | dewtprintk(ioc, |
7930 | ioc_info(ioc, "%s: enter: handle(0x%04x), wwid(0x%016llx)\n" , |
7931 | __func__, |
7932 | pcie_device->handle, (u64)pcie_device->wwid)); |
7933 | if (pcie_device->enclosure_handle != 0) |
7934 | dewtprintk(ioc, |
7935 | ioc_info(ioc, "%s: enter: enclosure logical id(0x%016llx), slot(%d)\n" , |
7936 | __func__, |
7937 | (u64)pcie_device->enclosure_logical_id, |
7938 | pcie_device->slot)); |
7939 | if (pcie_device->connector_name[0] != '\0') |
7940 | dewtprintk(ioc, |
7941 | ioc_info(ioc, "%s: enter: enclosure level(0x%04x), connector name(%s)\n" , |
7942 | __func__, |
7943 | pcie_device->enclosure_level, |
7944 | pcie_device->connector_name)); |
7945 | |
7946 | if (pcie_device->starget && pcie_device->starget->hostdata) { |
7947 | sas_target_priv_data = pcie_device->starget->hostdata; |
7948 | sas_target_priv_data->deleted = 1; |
7949 | _scsih_ublock_io_device(ioc, sas_address: pcie_device->wwid, NULL); |
7950 | sas_target_priv_data->handle = MPT3SAS_INVALID_DEVICE_HANDLE; |
7951 | } |
7952 | |
7953 | ioc_info(ioc, "removing handle(0x%04x), wwid(0x%016llx)\n" , |
7954 | pcie_device->handle, (u64)pcie_device->wwid); |
7955 | if (pcie_device->enclosure_handle != 0) |
7956 | ioc_info(ioc, "removing : enclosure logical id(0x%016llx), slot(%d)\n" , |
7957 | (u64)pcie_device->enclosure_logical_id, |
7958 | pcie_device->slot); |
7959 | if (pcie_device->connector_name[0] != '\0') |
7960 | ioc_info(ioc, "removing: enclosure level(0x%04x), connector name( %s)\n" , |
7961 | pcie_device->enclosure_level, |
7962 | pcie_device->connector_name); |
7963 | |
7964 | if (pcie_device->starget && (pcie_device->access_status != |
7965 | MPI26_PCIEDEV0_ASTATUS_DEVICE_BLOCKED)) |
7966 | scsi_remove_target(&pcie_device->starget->dev); |
7967 | dewtprintk(ioc, |
7968 | ioc_info(ioc, "%s: exit: handle(0x%04x), wwid(0x%016llx)\n" , |
7969 | __func__, |
7970 | pcie_device->handle, (u64)pcie_device->wwid)); |
7971 | if (pcie_device->enclosure_handle != 0) |
7972 | dewtprintk(ioc, |
7973 | ioc_info(ioc, "%s: exit: enclosure logical id(0x%016llx), slot(%d)\n" , |
7974 | __func__, |
7975 | (u64)pcie_device->enclosure_logical_id, |
7976 | pcie_device->slot)); |
7977 | if (pcie_device->connector_name[0] != '\0') |
7978 | dewtprintk(ioc, |
7979 | ioc_info(ioc, "%s: exit: enclosure level(0x%04x), connector name( %s)\n" , |
7980 | __func__, |
7981 | pcie_device->enclosure_level, |
7982 | pcie_device->connector_name)); |
7983 | |
7984 | kfree(objp: pcie_device->serial_number); |
7985 | } |
7986 | |
7987 | |
7988 | /** |
7989 | * _scsih_pcie_check_device - checking device responsiveness |
7990 | * @ioc: per adapter object |
7991 | * @handle: attached device handle |
7992 | */ |
7993 | static void |
7994 | _scsih_pcie_check_device(struct MPT3SAS_ADAPTER *ioc, u16 handle) |
7995 | { |
7996 | Mpi2ConfigReply_t mpi_reply; |
7997 | Mpi26PCIeDevicePage0_t pcie_device_pg0; |
7998 | u32 ioc_status; |
7999 | struct _pcie_device *pcie_device; |
8000 | u64 wwid; |
8001 | unsigned long flags; |
8002 | struct scsi_target *starget; |
8003 | struct MPT3SAS_TARGET *sas_target_priv_data; |
8004 | u32 device_info; |
8005 | |
8006 | if ((mpt3sas_config_get_pcie_device_pg0(ioc, mpi_reply: &mpi_reply, |
8007 | config_page: &pcie_device_pg0, MPI26_PCIE_DEVICE_PGAD_FORM_HANDLE, handle))) |
8008 | return; |
8009 | |
8010 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & MPI2_IOCSTATUS_MASK; |
8011 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) |
8012 | return; |
8013 | |
8014 | /* check if this is end device */ |
8015 | device_info = le32_to_cpu(pcie_device_pg0.DeviceInfo); |
8016 | if (!(_scsih_is_nvme_pciescsi_device(device_info))) |
8017 | return; |
8018 | |
8019 | wwid = le64_to_cpu(pcie_device_pg0.WWID); |
8020 | spin_lock_irqsave(&ioc->pcie_device_lock, flags); |
8021 | pcie_device = __mpt3sas_get_pdev_by_wwid(ioc, wwid); |
8022 | |
8023 | if (!pcie_device) { |
8024 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
8025 | return; |
8026 | } |
8027 | |
8028 | if (unlikely(pcie_device->handle != handle)) { |
8029 | starget = pcie_device->starget; |
8030 | sas_target_priv_data = starget->hostdata; |
8031 | pcie_device->access_status = pcie_device_pg0.AccessStatus; |
8032 | starget_printk(KERN_INFO, starget, |
8033 | "handle changed from(0x%04x) to (0x%04x)!!!\n" , |
8034 | pcie_device->handle, handle); |
8035 | sas_target_priv_data->handle = handle; |
8036 | pcie_device->handle = handle; |
8037 | |
8038 | if (le32_to_cpu(pcie_device_pg0.Flags) & |
8039 | MPI26_PCIEDEV0_FLAGS_ENCL_LEVEL_VALID) { |
8040 | pcie_device->enclosure_level = |
8041 | pcie_device_pg0.EnclosureLevel; |
8042 | memcpy(&pcie_device->connector_name[0], |
8043 | &pcie_device_pg0.ConnectorName[0], 4); |
8044 | } else { |
8045 | pcie_device->enclosure_level = 0; |
8046 | pcie_device->connector_name[0] = '\0'; |
8047 | } |
8048 | } |
8049 | |
8050 | /* check if device is present */ |
8051 | if (!(le32_to_cpu(pcie_device_pg0.Flags) & |
8052 | MPI26_PCIEDEV0_FLAGS_DEVICE_PRESENT)) { |
8053 | ioc_info(ioc, "device is not present handle(0x%04x), flags!!!\n" , |
8054 | handle); |
8055 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
8056 | pcie_device_put(p: pcie_device); |
8057 | return; |
8058 | } |
8059 | |
8060 | /* check if there were any issues with discovery */ |
8061 | if (_scsih_check_pcie_access_status(ioc, wwid, handle, |
8062 | access_status: pcie_device_pg0.AccessStatus)) { |
8063 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
8064 | pcie_device_put(p: pcie_device); |
8065 | return; |
8066 | } |
8067 | |
8068 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
8069 | pcie_device_put(p: pcie_device); |
8070 | |
8071 | _scsih_ublock_io_device(ioc, sas_address: wwid, NULL); |
8072 | |
8073 | return; |
8074 | } |
8075 | |
8076 | /** |
8077 | * _scsih_pcie_add_device - creating pcie device object |
8078 | * @ioc: per adapter object |
8079 | * @handle: pcie device handle |
8080 | * |
8081 | * Creating end device object, stored in ioc->pcie_device_list. |
8082 | * |
8083 | * Return: 1 means queue the event later, 0 means complete the event |
8084 | */ |
8085 | static int |
8086 | _scsih_pcie_add_device(struct MPT3SAS_ADAPTER *ioc, u16 handle) |
8087 | { |
8088 | Mpi26PCIeDevicePage0_t pcie_device_pg0; |
8089 | Mpi26PCIeDevicePage2_t pcie_device_pg2; |
8090 | Mpi2ConfigReply_t mpi_reply; |
8091 | struct _pcie_device *pcie_device; |
8092 | struct _enclosure_node *enclosure_dev; |
8093 | u32 ioc_status; |
8094 | u64 wwid; |
8095 | |
8096 | if ((mpt3sas_config_get_pcie_device_pg0(ioc, mpi_reply: &mpi_reply, |
8097 | config_page: &pcie_device_pg0, MPI26_PCIE_DEVICE_PGAD_FORM_HANDLE, handle))) { |
8098 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
8099 | __FILE__, __LINE__, __func__); |
8100 | return 0; |
8101 | } |
8102 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & |
8103 | MPI2_IOCSTATUS_MASK; |
8104 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { |
8105 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
8106 | __FILE__, __LINE__, __func__); |
8107 | return 0; |
8108 | } |
8109 | |
8110 | set_bit(nr: handle, addr: ioc->pend_os_device_add); |
8111 | wwid = le64_to_cpu(pcie_device_pg0.WWID); |
8112 | |
8113 | /* check if device is present */ |
8114 | if (!(le32_to_cpu(pcie_device_pg0.Flags) & |
8115 | MPI26_PCIEDEV0_FLAGS_DEVICE_PRESENT)) { |
8116 | ioc_err(ioc, "device is not present handle(0x04%x)!!!\n" , |
8117 | handle); |
8118 | return 0; |
8119 | } |
8120 | |
8121 | /* check if there were any issues with discovery */ |
8122 | if (_scsih_check_pcie_access_status(ioc, wwid, handle, |
8123 | access_status: pcie_device_pg0.AccessStatus)) |
8124 | return 0; |
8125 | |
8126 | if (!(_scsih_is_nvme_pciescsi_device(le32_to_cpu |
8127 | (pcie_device_pg0.DeviceInfo)))) |
8128 | return 0; |
8129 | |
8130 | pcie_device = mpt3sas_get_pdev_by_wwid(ioc, wwid); |
8131 | if (pcie_device) { |
8132 | clear_bit(nr: handle, addr: ioc->pend_os_device_add); |
8133 | pcie_device_put(p: pcie_device); |
8134 | return 0; |
8135 | } |
8136 | |
8137 | /* PCIe Device Page 2 contains read-only information about a |
8138 | * specific NVMe device; therefore, this page is only |
8139 | * valid for NVMe devices and skip for pcie devices of type scsi. |
8140 | */ |
8141 | if (!(mpt3sas_scsih_is_pcie_scsi_device( |
8142 | le32_to_cpu(pcie_device_pg0.DeviceInfo)))) { |
8143 | if (mpt3sas_config_get_pcie_device_pg2(ioc, mpi_reply: &mpi_reply, |
8144 | config_page: &pcie_device_pg2, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, |
8145 | handle)) { |
8146 | ioc_err(ioc, |
8147 | "failure at %s:%d/%s()!\n" , __FILE__, |
8148 | __LINE__, __func__); |
8149 | return 0; |
8150 | } |
8151 | |
8152 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & |
8153 | MPI2_IOCSTATUS_MASK; |
8154 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { |
8155 | ioc_err(ioc, |
8156 | "failure at %s:%d/%s()!\n" , __FILE__, |
8157 | __LINE__, __func__); |
8158 | return 0; |
8159 | } |
8160 | } |
8161 | |
8162 | pcie_device = kzalloc(size: sizeof(struct _pcie_device), GFP_KERNEL); |
8163 | if (!pcie_device) { |
8164 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
8165 | __FILE__, __LINE__, __func__); |
8166 | return 0; |
8167 | } |
8168 | |
8169 | kref_init(kref: &pcie_device->refcount); |
8170 | pcie_device->id = ioc->pcie_target_id++; |
8171 | pcie_device->channel = PCIE_CHANNEL; |
8172 | pcie_device->handle = handle; |
8173 | pcie_device->access_status = pcie_device_pg0.AccessStatus; |
8174 | pcie_device->device_info = le32_to_cpu(pcie_device_pg0.DeviceInfo); |
8175 | pcie_device->wwid = wwid; |
8176 | pcie_device->port_num = pcie_device_pg0.PortNum; |
8177 | pcie_device->fast_path = (le32_to_cpu(pcie_device_pg0.Flags) & |
8178 | MPI26_PCIEDEV0_FLAGS_FAST_PATH_CAPABLE) ? 1 : 0; |
8179 | |
8180 | pcie_device->enclosure_handle = |
8181 | le16_to_cpu(pcie_device_pg0.EnclosureHandle); |
8182 | if (pcie_device->enclosure_handle != 0) |
8183 | pcie_device->slot = le16_to_cpu(pcie_device_pg0.Slot); |
8184 | |
8185 | if (le32_to_cpu(pcie_device_pg0.Flags) & |
8186 | MPI26_PCIEDEV0_FLAGS_ENCL_LEVEL_VALID) { |
8187 | pcie_device->enclosure_level = pcie_device_pg0.EnclosureLevel; |
8188 | memcpy(&pcie_device->connector_name[0], |
8189 | &pcie_device_pg0.ConnectorName[0], 4); |
8190 | } else { |
8191 | pcie_device->enclosure_level = 0; |
8192 | pcie_device->connector_name[0] = '\0'; |
8193 | } |
8194 | |
8195 | /* get enclosure_logical_id */ |
8196 | if (pcie_device->enclosure_handle) { |
8197 | enclosure_dev = |
8198 | mpt3sas_scsih_enclosure_find_by_handle(ioc, |
8199 | handle: pcie_device->enclosure_handle); |
8200 | if (enclosure_dev) |
8201 | pcie_device->enclosure_logical_id = |
8202 | le64_to_cpu(enclosure_dev->pg0.EnclosureLogicalID); |
8203 | } |
8204 | /* TODO -- Add device name once FW supports it */ |
8205 | if (!(mpt3sas_scsih_is_pcie_scsi_device( |
8206 | le32_to_cpu(pcie_device_pg0.DeviceInfo)))) { |
8207 | pcie_device->nvme_mdts = |
8208 | le32_to_cpu(pcie_device_pg2.MaximumDataTransferSize); |
8209 | pcie_device->shutdown_latency = |
8210 | le16_to_cpu(pcie_device_pg2.ShutdownLatency); |
8211 | /* |
8212 | * Set IOC's max_shutdown_latency to drive's RTD3 Entry Latency |
8213 | * if drive's RTD3 Entry Latency is greater then IOC's |
8214 | * max_shutdown_latency. |
8215 | */ |
8216 | if (pcie_device->shutdown_latency > ioc->max_shutdown_latency) |
8217 | ioc->max_shutdown_latency = |
8218 | pcie_device->shutdown_latency; |
8219 | if (pcie_device_pg2.ControllerResetTO) |
8220 | pcie_device->reset_timeout = |
8221 | pcie_device_pg2.ControllerResetTO; |
8222 | else |
8223 | pcie_device->reset_timeout = 30; |
8224 | } else |
8225 | pcie_device->reset_timeout = 30; |
8226 | |
8227 | if (ioc->wait_for_discovery_to_complete) |
8228 | _scsih_pcie_device_init_add(ioc, pcie_device); |
8229 | else |
8230 | _scsih_pcie_device_add(ioc, pcie_device); |
8231 | |
8232 | pcie_device_put(p: pcie_device); |
8233 | return 0; |
8234 | } |
8235 | |
8236 | /** |
8237 | * _scsih_pcie_topology_change_event_debug - debug for topology |
8238 | * event |
8239 | * @ioc: per adapter object |
8240 | * @event_data: event data payload |
8241 | * Context: user. |
8242 | */ |
8243 | static void |
8244 | _scsih_pcie_topology_change_event_debug(struct MPT3SAS_ADAPTER *ioc, |
8245 | Mpi26EventDataPCIeTopologyChangeList_t *event_data) |
8246 | { |
8247 | int i; |
8248 | u16 handle; |
8249 | u16 reason_code; |
8250 | u8 port_number; |
8251 | char *status_str = NULL; |
8252 | u8 link_rate, prev_link_rate; |
8253 | |
8254 | switch (event_data->SwitchStatus) { |
8255 | case MPI26_EVENT_PCIE_TOPO_SS_ADDED: |
8256 | status_str = "add" ; |
8257 | break; |
8258 | case MPI26_EVENT_PCIE_TOPO_SS_NOT_RESPONDING: |
8259 | status_str = "remove" ; |
8260 | break; |
8261 | case MPI26_EVENT_PCIE_TOPO_SS_RESPONDING: |
8262 | case 0: |
8263 | status_str = "responding" ; |
8264 | break; |
8265 | case MPI26_EVENT_PCIE_TOPO_SS_DELAY_NOT_RESPONDING: |
8266 | status_str = "remove delay" ; |
8267 | break; |
8268 | default: |
8269 | status_str = "unknown status" ; |
8270 | break; |
8271 | } |
8272 | ioc_info(ioc, "pcie topology change: (%s)\n" , status_str); |
8273 | pr_info("\tswitch_handle(0x%04x), enclosure_handle(0x%04x)" |
8274 | "start_port(%02d), count(%d)\n" , |
8275 | le16_to_cpu(event_data->SwitchDevHandle), |
8276 | le16_to_cpu(event_data->EnclosureHandle), |
8277 | event_data->StartPortNum, event_data->NumEntries); |
8278 | for (i = 0; i < event_data->NumEntries; i++) { |
8279 | handle = |
8280 | le16_to_cpu(event_data->PortEntry[i].AttachedDevHandle); |
8281 | if (!handle) |
8282 | continue; |
8283 | port_number = event_data->StartPortNum + i; |
8284 | reason_code = event_data->PortEntry[i].PortStatus; |
8285 | switch (reason_code) { |
8286 | case MPI26_EVENT_PCIE_TOPO_PS_DEV_ADDED: |
8287 | status_str = "target add" ; |
8288 | break; |
8289 | case MPI26_EVENT_PCIE_TOPO_PS_NOT_RESPONDING: |
8290 | status_str = "target remove" ; |
8291 | break; |
8292 | case MPI26_EVENT_PCIE_TOPO_PS_DELAY_NOT_RESPONDING: |
8293 | status_str = "delay target remove" ; |
8294 | break; |
8295 | case MPI26_EVENT_PCIE_TOPO_PS_PORT_CHANGED: |
8296 | status_str = "link rate change" ; |
8297 | break; |
8298 | case MPI26_EVENT_PCIE_TOPO_PS_NO_CHANGE: |
8299 | status_str = "target responding" ; |
8300 | break; |
8301 | default: |
8302 | status_str = "unknown" ; |
8303 | break; |
8304 | } |
8305 | link_rate = event_data->PortEntry[i].CurrentPortInfo & |
8306 | MPI26_EVENT_PCIE_TOPO_PI_RATE_MASK; |
8307 | prev_link_rate = event_data->PortEntry[i].PreviousPortInfo & |
8308 | MPI26_EVENT_PCIE_TOPO_PI_RATE_MASK; |
8309 | pr_info("\tport(%02d), attached_handle(0x%04x): %s:" |
8310 | " link rate: new(0x%02x), old(0x%02x)\n" , port_number, |
8311 | handle, status_str, link_rate, prev_link_rate); |
8312 | } |
8313 | } |
8314 | |
8315 | /** |
8316 | * _scsih_pcie_topology_change_event - handle PCIe topology |
8317 | * changes |
8318 | * @ioc: per adapter object |
8319 | * @fw_event: The fw_event_work object |
8320 | * Context: user. |
8321 | * |
8322 | */ |
8323 | static void |
8324 | _scsih_pcie_topology_change_event(struct MPT3SAS_ADAPTER *ioc, |
8325 | struct fw_event_work *fw_event) |
8326 | { |
8327 | int i; |
8328 | u16 handle; |
8329 | u16 reason_code; |
8330 | u8 link_rate, prev_link_rate; |
8331 | unsigned long flags; |
8332 | int rc; |
8333 | Mpi26EventDataPCIeTopologyChangeList_t *event_data = |
8334 | (Mpi26EventDataPCIeTopologyChangeList_t *) fw_event->event_data; |
8335 | struct _pcie_device *pcie_device; |
8336 | |
8337 | if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) |
8338 | _scsih_pcie_topology_change_event_debug(ioc, event_data); |
8339 | |
8340 | if (ioc->shost_recovery || ioc->remove_host || |
8341 | ioc->pci_error_recovery) |
8342 | return; |
8343 | |
8344 | if (fw_event->ignore) { |
8345 | dewtprintk(ioc, ioc_info(ioc, "ignoring switch event\n" )); |
8346 | return; |
8347 | } |
8348 | |
8349 | /* handle siblings events */ |
8350 | for (i = 0; i < event_data->NumEntries; i++) { |
8351 | if (fw_event->ignore) { |
8352 | dewtprintk(ioc, |
8353 | ioc_info(ioc, "ignoring switch event\n" )); |
8354 | return; |
8355 | } |
8356 | if (ioc->remove_host || ioc->pci_error_recovery) |
8357 | return; |
8358 | reason_code = event_data->PortEntry[i].PortStatus; |
8359 | handle = |
8360 | le16_to_cpu(event_data->PortEntry[i].AttachedDevHandle); |
8361 | if (!handle) |
8362 | continue; |
8363 | |
8364 | link_rate = event_data->PortEntry[i].CurrentPortInfo |
8365 | & MPI26_EVENT_PCIE_TOPO_PI_RATE_MASK; |
8366 | prev_link_rate = event_data->PortEntry[i].PreviousPortInfo |
8367 | & MPI26_EVENT_PCIE_TOPO_PI_RATE_MASK; |
8368 | |
8369 | switch (reason_code) { |
8370 | case MPI26_EVENT_PCIE_TOPO_PS_PORT_CHANGED: |
8371 | if (ioc->shost_recovery) |
8372 | break; |
8373 | if (link_rate == prev_link_rate) |
8374 | break; |
8375 | if (link_rate < MPI26_EVENT_PCIE_TOPO_PI_RATE_2_5) |
8376 | break; |
8377 | |
8378 | _scsih_pcie_check_device(ioc, handle); |
8379 | |
8380 | /* This code after this point handles the test case |
8381 | * where a device has been added, however its returning |
8382 | * BUSY for sometime. Then before the Device Missing |
8383 | * Delay expires and the device becomes READY, the |
8384 | * device is removed and added back. |
8385 | */ |
8386 | spin_lock_irqsave(&ioc->pcie_device_lock, flags); |
8387 | pcie_device = __mpt3sas_get_pdev_by_handle(ioc, handle); |
8388 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
8389 | |
8390 | if (pcie_device) { |
8391 | pcie_device_put(p: pcie_device); |
8392 | break; |
8393 | } |
8394 | |
8395 | if (!test_bit(handle, ioc->pend_os_device_add)) |
8396 | break; |
8397 | |
8398 | dewtprintk(ioc, |
8399 | ioc_info(ioc, "handle(0x%04x) device not found: convert event to a device add\n" , |
8400 | handle)); |
8401 | event_data->PortEntry[i].PortStatus &= 0xF0; |
8402 | event_data->PortEntry[i].PortStatus |= |
8403 | MPI26_EVENT_PCIE_TOPO_PS_DEV_ADDED; |
8404 | fallthrough; |
8405 | case MPI26_EVENT_PCIE_TOPO_PS_DEV_ADDED: |
8406 | if (ioc->shost_recovery) |
8407 | break; |
8408 | if (link_rate < MPI26_EVENT_PCIE_TOPO_PI_RATE_2_5) |
8409 | break; |
8410 | |
8411 | rc = _scsih_pcie_add_device(ioc, handle); |
8412 | if (!rc) { |
8413 | /* mark entry vacant */ |
8414 | /* TODO This needs to be reviewed and fixed, |
8415 | * we dont have an entry |
8416 | * to make an event void like vacant |
8417 | */ |
8418 | event_data->PortEntry[i].PortStatus |= |
8419 | MPI26_EVENT_PCIE_TOPO_PS_NO_CHANGE; |
8420 | } |
8421 | break; |
8422 | case MPI26_EVENT_PCIE_TOPO_PS_NOT_RESPONDING: |
8423 | _scsih_pcie_device_remove_by_handle(ioc, handle); |
8424 | break; |
8425 | } |
8426 | } |
8427 | } |
8428 | |
8429 | /** |
8430 | * _scsih_pcie_device_status_change_event_debug - debug for device event |
8431 | * @ioc: ? |
8432 | * @event_data: event data payload |
8433 | * Context: user. |
8434 | */ |
8435 | static void |
8436 | _scsih_pcie_device_status_change_event_debug(struct MPT3SAS_ADAPTER *ioc, |
8437 | Mpi26EventDataPCIeDeviceStatusChange_t *event_data) |
8438 | { |
8439 | char *reason_str = NULL; |
8440 | |
8441 | switch (event_data->ReasonCode) { |
8442 | case MPI26_EVENT_PCIDEV_STAT_RC_SMART_DATA: |
8443 | reason_str = "smart data" ; |
8444 | break; |
8445 | case MPI26_EVENT_PCIDEV_STAT_RC_UNSUPPORTED: |
8446 | reason_str = "unsupported device discovered" ; |
8447 | break; |
8448 | case MPI26_EVENT_PCIDEV_STAT_RC_INTERNAL_DEVICE_RESET: |
8449 | reason_str = "internal device reset" ; |
8450 | break; |
8451 | case MPI26_EVENT_PCIDEV_STAT_RC_TASK_ABORT_INTERNAL: |
8452 | reason_str = "internal task abort" ; |
8453 | break; |
8454 | case MPI26_EVENT_PCIDEV_STAT_RC_ABORT_TASK_SET_INTERNAL: |
8455 | reason_str = "internal task abort set" ; |
8456 | break; |
8457 | case MPI26_EVENT_PCIDEV_STAT_RC_CLEAR_TASK_SET_INTERNAL: |
8458 | reason_str = "internal clear task set" ; |
8459 | break; |
8460 | case MPI26_EVENT_PCIDEV_STAT_RC_QUERY_TASK_INTERNAL: |
8461 | reason_str = "internal query task" ; |
8462 | break; |
8463 | case MPI26_EVENT_PCIDEV_STAT_RC_DEV_INIT_FAILURE: |
8464 | reason_str = "device init failure" ; |
8465 | break; |
8466 | case MPI26_EVENT_PCIDEV_STAT_RC_CMP_INTERNAL_DEV_RESET: |
8467 | reason_str = "internal device reset complete" ; |
8468 | break; |
8469 | case MPI26_EVENT_PCIDEV_STAT_RC_CMP_TASK_ABORT_INTERNAL: |
8470 | reason_str = "internal task abort complete" ; |
8471 | break; |
8472 | case MPI26_EVENT_PCIDEV_STAT_RC_ASYNC_NOTIFICATION: |
8473 | reason_str = "internal async notification" ; |
8474 | break; |
8475 | case MPI26_EVENT_PCIDEV_STAT_RC_PCIE_HOT_RESET_FAILED: |
8476 | reason_str = "pcie hot reset failed" ; |
8477 | break; |
8478 | default: |
8479 | reason_str = "unknown reason" ; |
8480 | break; |
8481 | } |
8482 | |
8483 | ioc_info(ioc, "PCIE device status change: (%s)\n" |
8484 | "\thandle(0x%04x), WWID(0x%016llx), tag(%d)" , |
8485 | reason_str, le16_to_cpu(event_data->DevHandle), |
8486 | (u64)le64_to_cpu(event_data->WWID), |
8487 | le16_to_cpu(event_data->TaskTag)); |
8488 | if (event_data->ReasonCode == MPI26_EVENT_PCIDEV_STAT_RC_SMART_DATA) |
8489 | pr_cont(", ASC(0x%x), ASCQ(0x%x)\n" , |
8490 | event_data->ASC, event_data->ASCQ); |
8491 | pr_cont("\n" ); |
8492 | } |
8493 | |
8494 | /** |
8495 | * _scsih_pcie_device_status_change_event - handle device status |
8496 | * change |
8497 | * @ioc: per adapter object |
8498 | * @fw_event: The fw_event_work object |
8499 | * Context: user. |
8500 | */ |
8501 | static void |
8502 | _scsih_pcie_device_status_change_event(struct MPT3SAS_ADAPTER *ioc, |
8503 | struct fw_event_work *fw_event) |
8504 | { |
8505 | struct MPT3SAS_TARGET *target_priv_data; |
8506 | struct _pcie_device *pcie_device; |
8507 | u64 wwid; |
8508 | unsigned long flags; |
8509 | Mpi26EventDataPCIeDeviceStatusChange_t *event_data = |
8510 | (Mpi26EventDataPCIeDeviceStatusChange_t *)fw_event->event_data; |
8511 | if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) |
8512 | _scsih_pcie_device_status_change_event_debug(ioc, |
8513 | event_data); |
8514 | |
8515 | if (event_data->ReasonCode != |
8516 | MPI26_EVENT_PCIDEV_STAT_RC_INTERNAL_DEVICE_RESET && |
8517 | event_data->ReasonCode != |
8518 | MPI26_EVENT_PCIDEV_STAT_RC_CMP_INTERNAL_DEV_RESET) |
8519 | return; |
8520 | |
8521 | spin_lock_irqsave(&ioc->pcie_device_lock, flags); |
8522 | wwid = le64_to_cpu(event_data->WWID); |
8523 | pcie_device = __mpt3sas_get_pdev_by_wwid(ioc, wwid); |
8524 | |
8525 | if (!pcie_device || !pcie_device->starget) |
8526 | goto out; |
8527 | |
8528 | target_priv_data = pcie_device->starget->hostdata; |
8529 | if (!target_priv_data) |
8530 | goto out; |
8531 | |
8532 | if (event_data->ReasonCode == |
8533 | MPI26_EVENT_PCIDEV_STAT_RC_INTERNAL_DEVICE_RESET) |
8534 | target_priv_data->tm_busy = 1; |
8535 | else |
8536 | target_priv_data->tm_busy = 0; |
8537 | out: |
8538 | if (pcie_device) |
8539 | pcie_device_put(p: pcie_device); |
8540 | |
8541 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
8542 | } |
8543 | |
8544 | /** |
8545 | * _scsih_sas_enclosure_dev_status_change_event_debug - debug for enclosure |
8546 | * event |
8547 | * @ioc: per adapter object |
8548 | * @event_data: event data payload |
8549 | * Context: user. |
8550 | */ |
8551 | static void |
8552 | _scsih_sas_enclosure_dev_status_change_event_debug(struct MPT3SAS_ADAPTER *ioc, |
8553 | Mpi2EventDataSasEnclDevStatusChange_t *event_data) |
8554 | { |
8555 | char *reason_str = NULL; |
8556 | |
8557 | switch (event_data->ReasonCode) { |
8558 | case MPI2_EVENT_SAS_ENCL_RC_ADDED: |
8559 | reason_str = "enclosure add" ; |
8560 | break; |
8561 | case MPI2_EVENT_SAS_ENCL_RC_NOT_RESPONDING: |
8562 | reason_str = "enclosure remove" ; |
8563 | break; |
8564 | default: |
8565 | reason_str = "unknown reason" ; |
8566 | break; |
8567 | } |
8568 | |
8569 | ioc_info(ioc, "enclosure status change: (%s)\n" |
8570 | "\thandle(0x%04x), enclosure logical id(0x%016llx) number slots(%d)\n" , |
8571 | reason_str, |
8572 | le16_to_cpu(event_data->EnclosureHandle), |
8573 | (u64)le64_to_cpu(event_data->EnclosureLogicalID), |
8574 | le16_to_cpu(event_data->StartSlot)); |
8575 | } |
8576 | |
8577 | /** |
8578 | * _scsih_sas_enclosure_dev_status_change_event - handle enclosure events |
8579 | * @ioc: per adapter object |
8580 | * @fw_event: The fw_event_work object |
8581 | * Context: user. |
8582 | */ |
8583 | static void |
8584 | _scsih_sas_enclosure_dev_status_change_event(struct MPT3SAS_ADAPTER *ioc, |
8585 | struct fw_event_work *fw_event) |
8586 | { |
8587 | Mpi2ConfigReply_t mpi_reply; |
8588 | struct _enclosure_node *enclosure_dev = NULL; |
8589 | Mpi2EventDataSasEnclDevStatusChange_t *event_data = |
8590 | (Mpi2EventDataSasEnclDevStatusChange_t *)fw_event->event_data; |
8591 | int rc; |
8592 | u16 enclosure_handle = le16_to_cpu(event_data->EnclosureHandle); |
8593 | |
8594 | if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) |
8595 | _scsih_sas_enclosure_dev_status_change_event_debug(ioc, |
8596 | event_data: (Mpi2EventDataSasEnclDevStatusChange_t *) |
8597 | fw_event->event_data); |
8598 | if (ioc->shost_recovery) |
8599 | return; |
8600 | |
8601 | if (enclosure_handle) |
8602 | enclosure_dev = |
8603 | mpt3sas_scsih_enclosure_find_by_handle(ioc, |
8604 | handle: enclosure_handle); |
8605 | switch (event_data->ReasonCode) { |
8606 | case MPI2_EVENT_SAS_ENCL_RC_ADDED: |
8607 | if (!enclosure_dev) { |
8608 | enclosure_dev = |
8609 | kzalloc(size: sizeof(struct _enclosure_node), |
8610 | GFP_KERNEL); |
8611 | if (!enclosure_dev) { |
8612 | ioc_info(ioc, "failure at %s:%d/%s()!\n" , |
8613 | __FILE__, __LINE__, __func__); |
8614 | return; |
8615 | } |
8616 | rc = mpt3sas_config_get_enclosure_pg0(ioc, mpi_reply: &mpi_reply, |
8617 | config_page: &enclosure_dev->pg0, |
8618 | MPI2_SAS_ENCLOS_PGAD_FORM_HANDLE, |
8619 | handle: enclosure_handle); |
8620 | |
8621 | if (rc || (le16_to_cpu(mpi_reply.IOCStatus) & |
8622 | MPI2_IOCSTATUS_MASK)) { |
8623 | kfree(objp: enclosure_dev); |
8624 | return; |
8625 | } |
8626 | |
8627 | list_add_tail(new: &enclosure_dev->list, |
8628 | head: &ioc->enclosure_list); |
8629 | } |
8630 | break; |
8631 | case MPI2_EVENT_SAS_ENCL_RC_NOT_RESPONDING: |
8632 | if (enclosure_dev) { |
8633 | list_del(entry: &enclosure_dev->list); |
8634 | kfree(objp: enclosure_dev); |
8635 | } |
8636 | break; |
8637 | default: |
8638 | break; |
8639 | } |
8640 | } |
8641 | |
8642 | /** |
8643 | * _scsih_sas_broadcast_primitive_event - handle broadcast events |
8644 | * @ioc: per adapter object |
8645 | * @fw_event: The fw_event_work object |
8646 | * Context: user. |
8647 | */ |
8648 | static void |
8649 | _scsih_sas_broadcast_primitive_event(struct MPT3SAS_ADAPTER *ioc, |
8650 | struct fw_event_work *fw_event) |
8651 | { |
8652 | struct scsi_cmnd *scmd; |
8653 | struct scsi_device *sdev; |
8654 | struct scsiio_tracker *st; |
8655 | u16 smid, handle; |
8656 | u32 lun; |
8657 | struct MPT3SAS_DEVICE *sas_device_priv_data; |
8658 | u32 termination_count; |
8659 | u32 query_count; |
8660 | Mpi2SCSITaskManagementReply_t *mpi_reply; |
8661 | Mpi2EventDataSasBroadcastPrimitive_t *event_data = |
8662 | (Mpi2EventDataSasBroadcastPrimitive_t *) |
8663 | fw_event->event_data; |
8664 | u16 ioc_status; |
8665 | unsigned long flags; |
8666 | int r; |
8667 | u8 max_retries = 0; |
8668 | u8 task_abort_retries; |
8669 | |
8670 | mutex_lock(&ioc->tm_cmds.mutex); |
8671 | ioc_info(ioc, "%s: enter: phy number(%d), width(%d)\n" , |
8672 | __func__, event_data->PhyNum, event_data->PortWidth); |
8673 | |
8674 | _scsih_block_io_all_device(ioc); |
8675 | |
8676 | spin_lock_irqsave(&ioc->scsi_lookup_lock, flags); |
8677 | mpi_reply = ioc->tm_cmds.reply; |
8678 | broadcast_aen_retry: |
8679 | |
8680 | /* sanity checks for retrying this loop */ |
8681 | if (max_retries++ == 5) { |
8682 | dewtprintk(ioc, ioc_info(ioc, "%s: giving up\n" , __func__)); |
8683 | goto out; |
8684 | } else if (max_retries > 1) |
8685 | dewtprintk(ioc, |
8686 | ioc_info(ioc, "%s: %d retry\n" , |
8687 | __func__, max_retries - 1)); |
8688 | |
8689 | termination_count = 0; |
8690 | query_count = 0; |
8691 | for (smid = 1; smid <= ioc->scsiio_depth; smid++) { |
8692 | if (ioc->shost_recovery) |
8693 | goto out; |
8694 | scmd = mpt3sas_scsih_scsi_lookup_get(ioc, smid); |
8695 | if (!scmd) |
8696 | continue; |
8697 | st = scsi_cmd_priv(cmd: scmd); |
8698 | sdev = scmd->device; |
8699 | sas_device_priv_data = sdev->hostdata; |
8700 | if (!sas_device_priv_data || !sas_device_priv_data->sas_target) |
8701 | continue; |
8702 | /* skip hidden raid components */ |
8703 | if (sas_device_priv_data->sas_target->flags & |
8704 | MPT_TARGET_FLAGS_RAID_COMPONENT) |
8705 | continue; |
8706 | /* skip volumes */ |
8707 | if (sas_device_priv_data->sas_target->flags & |
8708 | MPT_TARGET_FLAGS_VOLUME) |
8709 | continue; |
8710 | /* skip PCIe devices */ |
8711 | if (sas_device_priv_data->sas_target->flags & |
8712 | MPT_TARGET_FLAGS_PCIE_DEVICE) |
8713 | continue; |
8714 | |
8715 | handle = sas_device_priv_data->sas_target->handle; |
8716 | lun = sas_device_priv_data->lun; |
8717 | query_count++; |
8718 | |
8719 | if (ioc->shost_recovery) |
8720 | goto out; |
8721 | |
8722 | spin_unlock_irqrestore(lock: &ioc->scsi_lookup_lock, flags); |
8723 | r = mpt3sas_scsih_issue_tm(ioc, handle, channel: 0, id: 0, lun, |
8724 | MPI2_SCSITASKMGMT_TASKTYPE_QUERY_TASK, smid_task: st->smid, |
8725 | msix_task: st->msix_io, timeout: 30, tr_method: 0); |
8726 | if (r == FAILED) { |
8727 | sdev_printk(KERN_WARNING, sdev, |
8728 | "mpt3sas_scsih_issue_tm: FAILED when sending " |
8729 | "QUERY_TASK: scmd(%p)\n" , scmd); |
8730 | spin_lock_irqsave(&ioc->scsi_lookup_lock, flags); |
8731 | goto broadcast_aen_retry; |
8732 | } |
8733 | ioc_status = le16_to_cpu(mpi_reply->IOCStatus) |
8734 | & MPI2_IOCSTATUS_MASK; |
8735 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { |
8736 | sdev_printk(KERN_WARNING, sdev, |
8737 | "query task: FAILED with IOCSTATUS(0x%04x), scmd(%p)\n" , |
8738 | ioc_status, scmd); |
8739 | spin_lock_irqsave(&ioc->scsi_lookup_lock, flags); |
8740 | goto broadcast_aen_retry; |
8741 | } |
8742 | |
8743 | /* see if IO is still owned by IOC and target */ |
8744 | if (mpi_reply->ResponseCode == |
8745 | MPI2_SCSITASKMGMT_RSP_TM_SUCCEEDED || |
8746 | mpi_reply->ResponseCode == |
8747 | MPI2_SCSITASKMGMT_RSP_IO_QUEUED_ON_IOC) { |
8748 | spin_lock_irqsave(&ioc->scsi_lookup_lock, flags); |
8749 | continue; |
8750 | } |
8751 | task_abort_retries = 0; |
8752 | tm_retry: |
8753 | if (task_abort_retries++ == 60) { |
8754 | dewtprintk(ioc, |
8755 | ioc_info(ioc, "%s: ABORT_TASK: giving up\n" , |
8756 | __func__)); |
8757 | spin_lock_irqsave(&ioc->scsi_lookup_lock, flags); |
8758 | goto broadcast_aen_retry; |
8759 | } |
8760 | |
8761 | if (ioc->shost_recovery) |
8762 | goto out_no_lock; |
8763 | |
8764 | r = mpt3sas_scsih_issue_tm(ioc, handle, channel: sdev->channel, id: sdev->id, |
8765 | lun: sdev->lun, MPI2_SCSITASKMGMT_TASKTYPE_ABORT_TASK, |
8766 | smid_task: st->smid, msix_task: st->msix_io, timeout: 30, tr_method: 0); |
8767 | if (r == FAILED || st->cb_idx != 0xFF) { |
8768 | sdev_printk(KERN_WARNING, sdev, |
8769 | "mpt3sas_scsih_issue_tm: ABORT_TASK: FAILED : " |
8770 | "scmd(%p)\n" , scmd); |
8771 | goto tm_retry; |
8772 | } |
8773 | |
8774 | if (task_abort_retries > 1) |
8775 | sdev_printk(KERN_WARNING, sdev, |
8776 | "mpt3sas_scsih_issue_tm: ABORT_TASK: RETRIES (%d):" |
8777 | " scmd(%p)\n" , |
8778 | task_abort_retries - 1, scmd); |
8779 | |
8780 | termination_count += le32_to_cpu(mpi_reply->TerminationCount); |
8781 | spin_lock_irqsave(&ioc->scsi_lookup_lock, flags); |
8782 | } |
8783 | |
8784 | if (ioc->broadcast_aen_pending) { |
8785 | dewtprintk(ioc, |
8786 | ioc_info(ioc, |
8787 | "%s: loop back due to pending AEN\n" , |
8788 | __func__)); |
8789 | ioc->broadcast_aen_pending = 0; |
8790 | goto broadcast_aen_retry; |
8791 | } |
8792 | |
8793 | out: |
8794 | spin_unlock_irqrestore(lock: &ioc->scsi_lookup_lock, flags); |
8795 | out_no_lock: |
8796 | |
8797 | dewtprintk(ioc, |
8798 | ioc_info(ioc, "%s - exit, query_count = %d termination_count = %d\n" , |
8799 | __func__, query_count, termination_count)); |
8800 | |
8801 | ioc->broadcast_aen_busy = 0; |
8802 | if (!ioc->shost_recovery) |
8803 | _scsih_ublock_io_all_device(ioc); |
8804 | mutex_unlock(lock: &ioc->tm_cmds.mutex); |
8805 | } |
8806 | |
8807 | /** |
8808 | * _scsih_sas_discovery_event - handle discovery events |
8809 | * @ioc: per adapter object |
8810 | * @fw_event: The fw_event_work object |
8811 | * Context: user. |
8812 | */ |
8813 | static void |
8814 | _scsih_sas_discovery_event(struct MPT3SAS_ADAPTER *ioc, |
8815 | struct fw_event_work *fw_event) |
8816 | { |
8817 | Mpi2EventDataSasDiscovery_t *event_data = |
8818 | (Mpi2EventDataSasDiscovery_t *) fw_event->event_data; |
8819 | |
8820 | if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) { |
8821 | ioc_info(ioc, "discovery event: (%s)" , |
8822 | event_data->ReasonCode == MPI2_EVENT_SAS_DISC_RC_STARTED ? |
8823 | "start" : "stop" ); |
8824 | if (event_data->DiscoveryStatus) |
8825 | pr_cont("discovery_status(0x%08x)" , |
8826 | le32_to_cpu(event_data->DiscoveryStatus)); |
8827 | pr_cont("\n" ); |
8828 | } |
8829 | |
8830 | if (event_data->ReasonCode == MPI2_EVENT_SAS_DISC_RC_STARTED && |
8831 | !ioc->sas_hba.num_phys) { |
8832 | if (disable_discovery > 0 && ioc->shost_recovery) { |
8833 | /* Wait for the reset to complete */ |
8834 | while (ioc->shost_recovery) |
8835 | ssleep(seconds: 1); |
8836 | } |
8837 | _scsih_sas_host_add(ioc); |
8838 | } |
8839 | } |
8840 | |
8841 | /** |
8842 | * _scsih_sas_device_discovery_error_event - display SAS device discovery error |
8843 | * events |
8844 | * @ioc: per adapter object |
8845 | * @fw_event: The fw_event_work object |
8846 | * Context: user. |
8847 | */ |
8848 | static void |
8849 | _scsih_sas_device_discovery_error_event(struct MPT3SAS_ADAPTER *ioc, |
8850 | struct fw_event_work *fw_event) |
8851 | { |
8852 | Mpi25EventDataSasDeviceDiscoveryError_t *event_data = |
8853 | (Mpi25EventDataSasDeviceDiscoveryError_t *)fw_event->event_data; |
8854 | |
8855 | switch (event_data->ReasonCode) { |
8856 | case MPI25_EVENT_SAS_DISC_ERR_SMP_FAILED: |
8857 | ioc_warn(ioc, "SMP command sent to the expander (handle:0x%04x, sas_address:0x%016llx, physical_port:0x%02x) has failed\n" , |
8858 | le16_to_cpu(event_data->DevHandle), |
8859 | (u64)le64_to_cpu(event_data->SASAddress), |
8860 | event_data->PhysicalPort); |
8861 | break; |
8862 | case MPI25_EVENT_SAS_DISC_ERR_SMP_TIMEOUT: |
8863 | ioc_warn(ioc, "SMP command sent to the expander (handle:0x%04x, sas_address:0x%016llx, physical_port:0x%02x) has timed out\n" , |
8864 | le16_to_cpu(event_data->DevHandle), |
8865 | (u64)le64_to_cpu(event_data->SASAddress), |
8866 | event_data->PhysicalPort); |
8867 | break; |
8868 | default: |
8869 | break; |
8870 | } |
8871 | } |
8872 | |
8873 | /** |
8874 | * _scsih_pcie_enumeration_event - handle enumeration events |
8875 | * @ioc: per adapter object |
8876 | * @fw_event: The fw_event_work object |
8877 | * Context: user. |
8878 | */ |
8879 | static void |
8880 | _scsih_pcie_enumeration_event(struct MPT3SAS_ADAPTER *ioc, |
8881 | struct fw_event_work *fw_event) |
8882 | { |
8883 | Mpi26EventDataPCIeEnumeration_t *event_data = |
8884 | (Mpi26EventDataPCIeEnumeration_t *)fw_event->event_data; |
8885 | |
8886 | if (!(ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK)) |
8887 | return; |
8888 | |
8889 | ioc_info(ioc, "pcie enumeration event: (%s) Flag 0x%02x" , |
8890 | (event_data->ReasonCode == MPI26_EVENT_PCIE_ENUM_RC_STARTED) ? |
8891 | "started" : "completed" , |
8892 | event_data->Flags); |
8893 | if (event_data->EnumerationStatus) |
8894 | pr_cont("enumeration_status(0x%08x)" , |
8895 | le32_to_cpu(event_data->EnumerationStatus)); |
8896 | pr_cont("\n" ); |
8897 | } |
8898 | |
8899 | /** |
8900 | * _scsih_ir_fastpath - turn on fastpath for IR physdisk |
8901 | * @ioc: per adapter object |
8902 | * @handle: device handle for physical disk |
8903 | * @phys_disk_num: physical disk number |
8904 | * |
8905 | * Return: 0 for success, else failure. |
8906 | */ |
8907 | static int |
8908 | _scsih_ir_fastpath(struct MPT3SAS_ADAPTER *ioc, u16 handle, u8 phys_disk_num) |
8909 | { |
8910 | Mpi2RaidActionRequest_t *mpi_request; |
8911 | Mpi2RaidActionReply_t *mpi_reply; |
8912 | u16 smid; |
8913 | u8 issue_reset = 0; |
8914 | int rc = 0; |
8915 | u16 ioc_status; |
8916 | u32 log_info; |
8917 | |
8918 | if (ioc->hba_mpi_version_belonged == MPI2_VERSION) |
8919 | return rc; |
8920 | |
8921 | mutex_lock(&ioc->scsih_cmds.mutex); |
8922 | |
8923 | if (ioc->scsih_cmds.status != MPT3_CMD_NOT_USED) { |
8924 | ioc_err(ioc, "%s: scsih_cmd in use\n" , __func__); |
8925 | rc = -EAGAIN; |
8926 | goto out; |
8927 | } |
8928 | ioc->scsih_cmds.status = MPT3_CMD_PENDING; |
8929 | |
8930 | smid = mpt3sas_base_get_smid(ioc, cb_idx: ioc->scsih_cb_idx); |
8931 | if (!smid) { |
8932 | ioc_err(ioc, "%s: failed obtaining a smid\n" , __func__); |
8933 | ioc->scsih_cmds.status = MPT3_CMD_NOT_USED; |
8934 | rc = -EAGAIN; |
8935 | goto out; |
8936 | } |
8937 | |
8938 | mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); |
8939 | ioc->scsih_cmds.smid = smid; |
8940 | memset(mpi_request, 0, sizeof(Mpi2RaidActionRequest_t)); |
8941 | |
8942 | mpi_request->Function = MPI2_FUNCTION_RAID_ACTION; |
8943 | mpi_request->Action = MPI2_RAID_ACTION_PHYSDISK_HIDDEN; |
8944 | mpi_request->PhysDiskNum = phys_disk_num; |
8945 | |
8946 | dewtprintk(ioc, |
8947 | ioc_info(ioc, "IR RAID_ACTION: turning fast path on for handle(0x%04x), phys_disk_num (0x%02x)\n" , |
8948 | handle, phys_disk_num)); |
8949 | |
8950 | init_completion(x: &ioc->scsih_cmds.done); |
8951 | ioc->put_smid_default(ioc, smid); |
8952 | wait_for_completion_timeout(x: &ioc->scsih_cmds.done, timeout: 10*HZ); |
8953 | |
8954 | if (!(ioc->scsih_cmds.status & MPT3_CMD_COMPLETE)) { |
8955 | mpt3sas_check_cmd_timeout(ioc, |
8956 | ioc->scsih_cmds.status, mpi_request, |
8957 | sizeof(Mpi2RaidActionRequest_t)/4, issue_reset); |
8958 | rc = -EFAULT; |
8959 | goto out; |
8960 | } |
8961 | |
8962 | if (ioc->scsih_cmds.status & MPT3_CMD_REPLY_VALID) { |
8963 | |
8964 | mpi_reply = ioc->scsih_cmds.reply; |
8965 | ioc_status = le16_to_cpu(mpi_reply->IOCStatus); |
8966 | if (ioc_status & MPI2_IOCSTATUS_FLAG_LOG_INFO_AVAILABLE) |
8967 | log_info = le32_to_cpu(mpi_reply->IOCLogInfo); |
8968 | else |
8969 | log_info = 0; |
8970 | ioc_status &= MPI2_IOCSTATUS_MASK; |
8971 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { |
8972 | dewtprintk(ioc, |
8973 | ioc_info(ioc, "IR RAID_ACTION: failed: ioc_status(0x%04x), loginfo(0x%08x)!!!\n" , |
8974 | ioc_status, log_info)); |
8975 | rc = -EFAULT; |
8976 | } else |
8977 | dewtprintk(ioc, |
8978 | ioc_info(ioc, "IR RAID_ACTION: completed successfully\n" )); |
8979 | } |
8980 | |
8981 | out: |
8982 | ioc->scsih_cmds.status = MPT3_CMD_NOT_USED; |
8983 | mutex_unlock(lock: &ioc->scsih_cmds.mutex); |
8984 | |
8985 | if (issue_reset) |
8986 | mpt3sas_base_hard_reset_handler(ioc, type: FORCE_BIG_HAMMER); |
8987 | return rc; |
8988 | } |
8989 | |
8990 | /** |
8991 | * _scsih_reprobe_lun - reprobing lun |
8992 | * @sdev: scsi device struct |
8993 | * @no_uld_attach: sdev->no_uld_attach flag setting |
8994 | * |
8995 | **/ |
8996 | static void |
8997 | _scsih_reprobe_lun(struct scsi_device *sdev, void *no_uld_attach) |
8998 | { |
8999 | sdev->no_uld_attach = no_uld_attach ? 1 : 0; |
9000 | sdev_printk(KERN_INFO, sdev, "%s raid component\n" , |
9001 | sdev->no_uld_attach ? "hiding" : "exposing" ); |
9002 | WARN_ON(scsi_device_reprobe(sdev)); |
9003 | } |
9004 | |
9005 | /** |
9006 | * _scsih_sas_volume_add - add new volume |
9007 | * @ioc: per adapter object |
9008 | * @element: IR config element data |
9009 | * Context: user. |
9010 | */ |
9011 | static void |
9012 | _scsih_sas_volume_add(struct MPT3SAS_ADAPTER *ioc, |
9013 | Mpi2EventIrConfigElement_t *element) |
9014 | { |
9015 | struct _raid_device *raid_device; |
9016 | unsigned long flags; |
9017 | u64 wwid; |
9018 | u16 handle = le16_to_cpu(element->VolDevHandle); |
9019 | int rc; |
9020 | |
9021 | mpt3sas_config_get_volume_wwid(ioc, volume_handle: handle, wwid: &wwid); |
9022 | if (!wwid) { |
9023 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
9024 | __FILE__, __LINE__, __func__); |
9025 | return; |
9026 | } |
9027 | |
9028 | spin_lock_irqsave(&ioc->raid_device_lock, flags); |
9029 | raid_device = _scsih_raid_device_find_by_wwid(ioc, wwid); |
9030 | spin_unlock_irqrestore(lock: &ioc->raid_device_lock, flags); |
9031 | |
9032 | if (raid_device) |
9033 | return; |
9034 | |
9035 | raid_device = kzalloc(size: sizeof(struct _raid_device), GFP_KERNEL); |
9036 | if (!raid_device) { |
9037 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
9038 | __FILE__, __LINE__, __func__); |
9039 | return; |
9040 | } |
9041 | |
9042 | raid_device->id = ioc->sas_id++; |
9043 | raid_device->channel = RAID_CHANNEL; |
9044 | raid_device->handle = handle; |
9045 | raid_device->wwid = wwid; |
9046 | _scsih_raid_device_add(ioc, raid_device); |
9047 | if (!ioc->wait_for_discovery_to_complete) { |
9048 | rc = scsi_add_device(host: ioc->shost, RAID_CHANNEL, |
9049 | target: raid_device->id, lun: 0); |
9050 | if (rc) |
9051 | _scsih_raid_device_remove(ioc, raid_device); |
9052 | } else { |
9053 | spin_lock_irqsave(&ioc->raid_device_lock, flags); |
9054 | _scsih_determine_boot_device(ioc, device: raid_device, channel: 1); |
9055 | spin_unlock_irqrestore(lock: &ioc->raid_device_lock, flags); |
9056 | } |
9057 | } |
9058 | |
9059 | /** |
9060 | * _scsih_sas_volume_delete - delete volume |
9061 | * @ioc: per adapter object |
9062 | * @handle: volume device handle |
9063 | * Context: user. |
9064 | */ |
9065 | static void |
9066 | _scsih_sas_volume_delete(struct MPT3SAS_ADAPTER *ioc, u16 handle) |
9067 | { |
9068 | struct _raid_device *raid_device; |
9069 | unsigned long flags; |
9070 | struct MPT3SAS_TARGET *sas_target_priv_data; |
9071 | struct scsi_target *starget = NULL; |
9072 | |
9073 | spin_lock_irqsave(&ioc->raid_device_lock, flags); |
9074 | raid_device = mpt3sas_raid_device_find_by_handle(ioc, handle); |
9075 | if (raid_device) { |
9076 | if (raid_device->starget) { |
9077 | starget = raid_device->starget; |
9078 | sas_target_priv_data = starget->hostdata; |
9079 | sas_target_priv_data->deleted = 1; |
9080 | } |
9081 | ioc_info(ioc, "removing handle(0x%04x), wwid(0x%016llx)\n" , |
9082 | raid_device->handle, (u64)raid_device->wwid); |
9083 | list_del(entry: &raid_device->list); |
9084 | kfree(objp: raid_device); |
9085 | } |
9086 | spin_unlock_irqrestore(lock: &ioc->raid_device_lock, flags); |
9087 | if (starget) |
9088 | scsi_remove_target(&starget->dev); |
9089 | } |
9090 | |
9091 | /** |
9092 | * _scsih_sas_pd_expose - expose pd component to /dev/sdX |
9093 | * @ioc: per adapter object |
9094 | * @element: IR config element data |
9095 | * Context: user. |
9096 | */ |
9097 | static void |
9098 | _scsih_sas_pd_expose(struct MPT3SAS_ADAPTER *ioc, |
9099 | Mpi2EventIrConfigElement_t *element) |
9100 | { |
9101 | struct _sas_device *sas_device; |
9102 | struct scsi_target *starget = NULL; |
9103 | struct MPT3SAS_TARGET *sas_target_priv_data; |
9104 | unsigned long flags; |
9105 | u16 handle = le16_to_cpu(element->PhysDiskDevHandle); |
9106 | |
9107 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
9108 | sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle); |
9109 | if (sas_device) { |
9110 | sas_device->volume_handle = 0; |
9111 | sas_device->volume_wwid = 0; |
9112 | clear_bit(nr: handle, addr: ioc->pd_handles); |
9113 | if (sas_device->starget && sas_device->starget->hostdata) { |
9114 | starget = sas_device->starget; |
9115 | sas_target_priv_data = starget->hostdata; |
9116 | sas_target_priv_data->flags &= |
9117 | ~MPT_TARGET_FLAGS_RAID_COMPONENT; |
9118 | } |
9119 | } |
9120 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
9121 | if (!sas_device) |
9122 | return; |
9123 | |
9124 | /* exposing raid component */ |
9125 | if (starget) |
9126 | starget_for_each_device(starget, NULL, fn: _scsih_reprobe_lun); |
9127 | |
9128 | sas_device_put(s: sas_device); |
9129 | } |
9130 | |
9131 | /** |
9132 | * _scsih_sas_pd_hide - hide pd component from /dev/sdX |
9133 | * @ioc: per adapter object |
9134 | * @element: IR config element data |
9135 | * Context: user. |
9136 | */ |
9137 | static void |
9138 | _scsih_sas_pd_hide(struct MPT3SAS_ADAPTER *ioc, |
9139 | Mpi2EventIrConfigElement_t *element) |
9140 | { |
9141 | struct _sas_device *sas_device; |
9142 | struct scsi_target *starget = NULL; |
9143 | struct MPT3SAS_TARGET *sas_target_priv_data; |
9144 | unsigned long flags; |
9145 | u16 handle = le16_to_cpu(element->PhysDiskDevHandle); |
9146 | u16 volume_handle = 0; |
9147 | u64 volume_wwid = 0; |
9148 | |
9149 | mpt3sas_config_get_volume_handle(ioc, pd_handle: handle, volume_handle: &volume_handle); |
9150 | if (volume_handle) |
9151 | mpt3sas_config_get_volume_wwid(ioc, volume_handle, |
9152 | wwid: &volume_wwid); |
9153 | |
9154 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
9155 | sas_device = __mpt3sas_get_sdev_by_handle(ioc, handle); |
9156 | if (sas_device) { |
9157 | set_bit(nr: handle, addr: ioc->pd_handles); |
9158 | if (sas_device->starget && sas_device->starget->hostdata) { |
9159 | starget = sas_device->starget; |
9160 | sas_target_priv_data = starget->hostdata; |
9161 | sas_target_priv_data->flags |= |
9162 | MPT_TARGET_FLAGS_RAID_COMPONENT; |
9163 | sas_device->volume_handle = volume_handle; |
9164 | sas_device->volume_wwid = volume_wwid; |
9165 | } |
9166 | } |
9167 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
9168 | if (!sas_device) |
9169 | return; |
9170 | |
9171 | /* hiding raid component */ |
9172 | _scsih_ir_fastpath(ioc, handle, phys_disk_num: element->PhysDiskNum); |
9173 | |
9174 | if (starget) |
9175 | starget_for_each_device(starget, (void *)1, fn: _scsih_reprobe_lun); |
9176 | |
9177 | sas_device_put(s: sas_device); |
9178 | } |
9179 | |
9180 | /** |
9181 | * _scsih_sas_pd_delete - delete pd component |
9182 | * @ioc: per adapter object |
9183 | * @element: IR config element data |
9184 | * Context: user. |
9185 | */ |
9186 | static void |
9187 | _scsih_sas_pd_delete(struct MPT3SAS_ADAPTER *ioc, |
9188 | Mpi2EventIrConfigElement_t *element) |
9189 | { |
9190 | u16 handle = le16_to_cpu(element->PhysDiskDevHandle); |
9191 | |
9192 | _scsih_device_remove_by_handle(ioc, handle); |
9193 | } |
9194 | |
9195 | /** |
9196 | * _scsih_sas_pd_add - remove pd component |
9197 | * @ioc: per adapter object |
9198 | * @element: IR config element data |
9199 | * Context: user. |
9200 | */ |
9201 | static void |
9202 | _scsih_sas_pd_add(struct MPT3SAS_ADAPTER *ioc, |
9203 | Mpi2EventIrConfigElement_t *element) |
9204 | { |
9205 | struct _sas_device *sas_device; |
9206 | u16 handle = le16_to_cpu(element->PhysDiskDevHandle); |
9207 | Mpi2ConfigReply_t mpi_reply; |
9208 | Mpi2SasDevicePage0_t sas_device_pg0; |
9209 | u32 ioc_status; |
9210 | u64 sas_address; |
9211 | u16 parent_handle; |
9212 | |
9213 | set_bit(nr: handle, addr: ioc->pd_handles); |
9214 | |
9215 | sas_device = mpt3sas_get_sdev_by_handle(ioc, handle); |
9216 | if (sas_device) { |
9217 | _scsih_ir_fastpath(ioc, handle, phys_disk_num: element->PhysDiskNum); |
9218 | sas_device_put(s: sas_device); |
9219 | return; |
9220 | } |
9221 | |
9222 | if ((mpt3sas_config_get_sas_device_pg0(ioc, mpi_reply: &mpi_reply, config_page: &sas_device_pg0, |
9223 | MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, handle))) { |
9224 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
9225 | __FILE__, __LINE__, __func__); |
9226 | return; |
9227 | } |
9228 | |
9229 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & |
9230 | MPI2_IOCSTATUS_MASK; |
9231 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { |
9232 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
9233 | __FILE__, __LINE__, __func__); |
9234 | return; |
9235 | } |
9236 | |
9237 | parent_handle = le16_to_cpu(sas_device_pg0.ParentDevHandle); |
9238 | if (!_scsih_get_sas_address(ioc, handle: parent_handle, sas_address: &sas_address)) |
9239 | mpt3sas_transport_update_links(ioc, sas_address, handle, |
9240 | phy_number: sas_device_pg0.PhyNum, MPI2_SAS_NEG_LINK_RATE_1_5, |
9241 | port: mpt3sas_get_port_by_id(ioc, |
9242 | port_id: sas_device_pg0.PhysicalPort, bypass_dirty_port_flag: 0)); |
9243 | |
9244 | _scsih_ir_fastpath(ioc, handle, phys_disk_num: element->PhysDiskNum); |
9245 | _scsih_add_device(ioc, handle, phy_num: 0, is_pd: 1); |
9246 | } |
9247 | |
9248 | /** |
9249 | * _scsih_sas_ir_config_change_event_debug - debug for IR Config Change events |
9250 | * @ioc: per adapter object |
9251 | * @event_data: event data payload |
9252 | * Context: user. |
9253 | */ |
9254 | static void |
9255 | _scsih_sas_ir_config_change_event_debug(struct MPT3SAS_ADAPTER *ioc, |
9256 | Mpi2EventDataIrConfigChangeList_t *event_data) |
9257 | { |
9258 | Mpi2EventIrConfigElement_t *element; |
9259 | u8 element_type; |
9260 | int i; |
9261 | char *reason_str = NULL, *element_str = NULL; |
9262 | |
9263 | element = (Mpi2EventIrConfigElement_t *)&event_data->ConfigElement[0]; |
9264 | |
9265 | ioc_info(ioc, "raid config change: (%s), elements(%d)\n" , |
9266 | le32_to_cpu(event_data->Flags) & MPI2_EVENT_IR_CHANGE_FLAGS_FOREIGN_CONFIG ? |
9267 | "foreign" : "native" , |
9268 | event_data->NumElements); |
9269 | for (i = 0; i < event_data->NumElements; i++, element++) { |
9270 | switch (element->ReasonCode) { |
9271 | case MPI2_EVENT_IR_CHANGE_RC_ADDED: |
9272 | reason_str = "add" ; |
9273 | break; |
9274 | case MPI2_EVENT_IR_CHANGE_RC_REMOVED: |
9275 | reason_str = "remove" ; |
9276 | break; |
9277 | case MPI2_EVENT_IR_CHANGE_RC_NO_CHANGE: |
9278 | reason_str = "no change" ; |
9279 | break; |
9280 | case MPI2_EVENT_IR_CHANGE_RC_HIDE: |
9281 | reason_str = "hide" ; |
9282 | break; |
9283 | case MPI2_EVENT_IR_CHANGE_RC_UNHIDE: |
9284 | reason_str = "unhide" ; |
9285 | break; |
9286 | case MPI2_EVENT_IR_CHANGE_RC_VOLUME_CREATED: |
9287 | reason_str = "volume_created" ; |
9288 | break; |
9289 | case MPI2_EVENT_IR_CHANGE_RC_VOLUME_DELETED: |
9290 | reason_str = "volume_deleted" ; |
9291 | break; |
9292 | case MPI2_EVENT_IR_CHANGE_RC_PD_CREATED: |
9293 | reason_str = "pd_created" ; |
9294 | break; |
9295 | case MPI2_EVENT_IR_CHANGE_RC_PD_DELETED: |
9296 | reason_str = "pd_deleted" ; |
9297 | break; |
9298 | default: |
9299 | reason_str = "unknown reason" ; |
9300 | break; |
9301 | } |
9302 | element_type = le16_to_cpu(element->ElementFlags) & |
9303 | MPI2_EVENT_IR_CHANGE_EFLAGS_ELEMENT_TYPE_MASK; |
9304 | switch (element_type) { |
9305 | case MPI2_EVENT_IR_CHANGE_EFLAGS_VOLUME_ELEMENT: |
9306 | element_str = "volume" ; |
9307 | break; |
9308 | case MPI2_EVENT_IR_CHANGE_EFLAGS_VOLPHYSDISK_ELEMENT: |
9309 | element_str = "phys disk" ; |
9310 | break; |
9311 | case MPI2_EVENT_IR_CHANGE_EFLAGS_HOTSPARE_ELEMENT: |
9312 | element_str = "hot spare" ; |
9313 | break; |
9314 | default: |
9315 | element_str = "unknown element" ; |
9316 | break; |
9317 | } |
9318 | pr_info("\t(%s:%s), vol handle(0x%04x), " \ |
9319 | "pd handle(0x%04x), pd num(0x%02x)\n" , element_str, |
9320 | reason_str, le16_to_cpu(element->VolDevHandle), |
9321 | le16_to_cpu(element->PhysDiskDevHandle), |
9322 | element->PhysDiskNum); |
9323 | } |
9324 | } |
9325 | |
9326 | /** |
9327 | * _scsih_sas_ir_config_change_event - handle ir configuration change events |
9328 | * @ioc: per adapter object |
9329 | * @fw_event: The fw_event_work object |
9330 | * Context: user. |
9331 | */ |
9332 | static void |
9333 | _scsih_sas_ir_config_change_event(struct MPT3SAS_ADAPTER *ioc, |
9334 | struct fw_event_work *fw_event) |
9335 | { |
9336 | Mpi2EventIrConfigElement_t *element; |
9337 | int i; |
9338 | u8 foreign_config; |
9339 | Mpi2EventDataIrConfigChangeList_t *event_data = |
9340 | (Mpi2EventDataIrConfigChangeList_t *) |
9341 | fw_event->event_data; |
9342 | |
9343 | if ((ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) && |
9344 | (!ioc->hide_ir_msg)) |
9345 | _scsih_sas_ir_config_change_event_debug(ioc, event_data); |
9346 | |
9347 | foreign_config = (le32_to_cpu(event_data->Flags) & |
9348 | MPI2_EVENT_IR_CHANGE_FLAGS_FOREIGN_CONFIG) ? 1 : 0; |
9349 | |
9350 | element = (Mpi2EventIrConfigElement_t *)&event_data->ConfigElement[0]; |
9351 | if (ioc->shost_recovery && |
9352 | ioc->hba_mpi_version_belonged != MPI2_VERSION) { |
9353 | for (i = 0; i < event_data->NumElements; i++, element++) { |
9354 | if (element->ReasonCode == MPI2_EVENT_IR_CHANGE_RC_HIDE) |
9355 | _scsih_ir_fastpath(ioc, |
9356 | le16_to_cpu(element->PhysDiskDevHandle), |
9357 | phys_disk_num: element->PhysDiskNum); |
9358 | } |
9359 | return; |
9360 | } |
9361 | |
9362 | for (i = 0; i < event_data->NumElements; i++, element++) { |
9363 | |
9364 | switch (element->ReasonCode) { |
9365 | case MPI2_EVENT_IR_CHANGE_RC_VOLUME_CREATED: |
9366 | case MPI2_EVENT_IR_CHANGE_RC_ADDED: |
9367 | if (!foreign_config) |
9368 | _scsih_sas_volume_add(ioc, element); |
9369 | break; |
9370 | case MPI2_EVENT_IR_CHANGE_RC_VOLUME_DELETED: |
9371 | case MPI2_EVENT_IR_CHANGE_RC_REMOVED: |
9372 | if (!foreign_config) |
9373 | _scsih_sas_volume_delete(ioc, |
9374 | le16_to_cpu(element->VolDevHandle)); |
9375 | break; |
9376 | case MPI2_EVENT_IR_CHANGE_RC_PD_CREATED: |
9377 | if (!ioc->is_warpdrive) |
9378 | _scsih_sas_pd_hide(ioc, element); |
9379 | break; |
9380 | case MPI2_EVENT_IR_CHANGE_RC_PD_DELETED: |
9381 | if (!ioc->is_warpdrive) |
9382 | _scsih_sas_pd_expose(ioc, element); |
9383 | break; |
9384 | case MPI2_EVENT_IR_CHANGE_RC_HIDE: |
9385 | if (!ioc->is_warpdrive) |
9386 | _scsih_sas_pd_add(ioc, element); |
9387 | break; |
9388 | case MPI2_EVENT_IR_CHANGE_RC_UNHIDE: |
9389 | if (!ioc->is_warpdrive) |
9390 | _scsih_sas_pd_delete(ioc, element); |
9391 | break; |
9392 | } |
9393 | } |
9394 | } |
9395 | |
9396 | /** |
9397 | * _scsih_sas_ir_volume_event - IR volume event |
9398 | * @ioc: per adapter object |
9399 | * @fw_event: The fw_event_work object |
9400 | * Context: user. |
9401 | */ |
9402 | static void |
9403 | _scsih_sas_ir_volume_event(struct MPT3SAS_ADAPTER *ioc, |
9404 | struct fw_event_work *fw_event) |
9405 | { |
9406 | u64 wwid; |
9407 | unsigned long flags; |
9408 | struct _raid_device *raid_device; |
9409 | u16 handle; |
9410 | u32 state; |
9411 | int rc; |
9412 | Mpi2EventDataIrVolume_t *event_data = |
9413 | (Mpi2EventDataIrVolume_t *) fw_event->event_data; |
9414 | |
9415 | if (ioc->shost_recovery) |
9416 | return; |
9417 | |
9418 | if (event_data->ReasonCode != MPI2_EVENT_IR_VOLUME_RC_STATE_CHANGED) |
9419 | return; |
9420 | |
9421 | handle = le16_to_cpu(event_data->VolDevHandle); |
9422 | state = le32_to_cpu(event_data->NewValue); |
9423 | if (!ioc->hide_ir_msg) |
9424 | dewtprintk(ioc, |
9425 | ioc_info(ioc, "%s: handle(0x%04x), old(0x%08x), new(0x%08x)\n" , |
9426 | __func__, handle, |
9427 | le32_to_cpu(event_data->PreviousValue), |
9428 | state)); |
9429 | switch (state) { |
9430 | case MPI2_RAID_VOL_STATE_MISSING: |
9431 | case MPI2_RAID_VOL_STATE_FAILED: |
9432 | _scsih_sas_volume_delete(ioc, handle); |
9433 | break; |
9434 | |
9435 | case MPI2_RAID_VOL_STATE_ONLINE: |
9436 | case MPI2_RAID_VOL_STATE_DEGRADED: |
9437 | case MPI2_RAID_VOL_STATE_OPTIMAL: |
9438 | |
9439 | spin_lock_irqsave(&ioc->raid_device_lock, flags); |
9440 | raid_device = mpt3sas_raid_device_find_by_handle(ioc, handle); |
9441 | spin_unlock_irqrestore(lock: &ioc->raid_device_lock, flags); |
9442 | |
9443 | if (raid_device) |
9444 | break; |
9445 | |
9446 | mpt3sas_config_get_volume_wwid(ioc, volume_handle: handle, wwid: &wwid); |
9447 | if (!wwid) { |
9448 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
9449 | __FILE__, __LINE__, __func__); |
9450 | break; |
9451 | } |
9452 | |
9453 | raid_device = kzalloc(size: sizeof(struct _raid_device), GFP_KERNEL); |
9454 | if (!raid_device) { |
9455 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
9456 | __FILE__, __LINE__, __func__); |
9457 | break; |
9458 | } |
9459 | |
9460 | raid_device->id = ioc->sas_id++; |
9461 | raid_device->channel = RAID_CHANNEL; |
9462 | raid_device->handle = handle; |
9463 | raid_device->wwid = wwid; |
9464 | _scsih_raid_device_add(ioc, raid_device); |
9465 | rc = scsi_add_device(host: ioc->shost, RAID_CHANNEL, |
9466 | target: raid_device->id, lun: 0); |
9467 | if (rc) |
9468 | _scsih_raid_device_remove(ioc, raid_device); |
9469 | break; |
9470 | |
9471 | case MPI2_RAID_VOL_STATE_INITIALIZING: |
9472 | default: |
9473 | break; |
9474 | } |
9475 | } |
9476 | |
9477 | /** |
9478 | * _scsih_sas_ir_physical_disk_event - PD event |
9479 | * @ioc: per adapter object |
9480 | * @fw_event: The fw_event_work object |
9481 | * Context: user. |
9482 | */ |
9483 | static void |
9484 | _scsih_sas_ir_physical_disk_event(struct MPT3SAS_ADAPTER *ioc, |
9485 | struct fw_event_work *fw_event) |
9486 | { |
9487 | u16 handle, parent_handle; |
9488 | u32 state; |
9489 | struct _sas_device *sas_device; |
9490 | Mpi2ConfigReply_t mpi_reply; |
9491 | Mpi2SasDevicePage0_t sas_device_pg0; |
9492 | u32 ioc_status; |
9493 | Mpi2EventDataIrPhysicalDisk_t *event_data = |
9494 | (Mpi2EventDataIrPhysicalDisk_t *) fw_event->event_data; |
9495 | u64 sas_address; |
9496 | |
9497 | if (ioc->shost_recovery) |
9498 | return; |
9499 | |
9500 | if (event_data->ReasonCode != MPI2_EVENT_IR_PHYSDISK_RC_STATE_CHANGED) |
9501 | return; |
9502 | |
9503 | handle = le16_to_cpu(event_data->PhysDiskDevHandle); |
9504 | state = le32_to_cpu(event_data->NewValue); |
9505 | |
9506 | if (!ioc->hide_ir_msg) |
9507 | dewtprintk(ioc, |
9508 | ioc_info(ioc, "%s: handle(0x%04x), old(0x%08x), new(0x%08x)\n" , |
9509 | __func__, handle, |
9510 | le32_to_cpu(event_data->PreviousValue), |
9511 | state)); |
9512 | |
9513 | switch (state) { |
9514 | case MPI2_RAID_PD_STATE_ONLINE: |
9515 | case MPI2_RAID_PD_STATE_DEGRADED: |
9516 | case MPI2_RAID_PD_STATE_REBUILDING: |
9517 | case MPI2_RAID_PD_STATE_OPTIMAL: |
9518 | case MPI2_RAID_PD_STATE_HOT_SPARE: |
9519 | |
9520 | if (!ioc->is_warpdrive) |
9521 | set_bit(nr: handle, addr: ioc->pd_handles); |
9522 | |
9523 | sas_device = mpt3sas_get_sdev_by_handle(ioc, handle); |
9524 | if (sas_device) { |
9525 | sas_device_put(s: sas_device); |
9526 | return; |
9527 | } |
9528 | |
9529 | if ((mpt3sas_config_get_sas_device_pg0(ioc, mpi_reply: &mpi_reply, |
9530 | config_page: &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, |
9531 | handle))) { |
9532 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
9533 | __FILE__, __LINE__, __func__); |
9534 | return; |
9535 | } |
9536 | |
9537 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & |
9538 | MPI2_IOCSTATUS_MASK; |
9539 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { |
9540 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
9541 | __FILE__, __LINE__, __func__); |
9542 | return; |
9543 | } |
9544 | |
9545 | parent_handle = le16_to_cpu(sas_device_pg0.ParentDevHandle); |
9546 | if (!_scsih_get_sas_address(ioc, handle: parent_handle, sas_address: &sas_address)) |
9547 | mpt3sas_transport_update_links(ioc, sas_address, handle, |
9548 | phy_number: sas_device_pg0.PhyNum, MPI2_SAS_NEG_LINK_RATE_1_5, |
9549 | port: mpt3sas_get_port_by_id(ioc, |
9550 | port_id: sas_device_pg0.PhysicalPort, bypass_dirty_port_flag: 0)); |
9551 | |
9552 | _scsih_add_device(ioc, handle, phy_num: 0, is_pd: 1); |
9553 | |
9554 | break; |
9555 | |
9556 | case MPI2_RAID_PD_STATE_OFFLINE: |
9557 | case MPI2_RAID_PD_STATE_NOT_CONFIGURED: |
9558 | case MPI2_RAID_PD_STATE_NOT_COMPATIBLE: |
9559 | default: |
9560 | break; |
9561 | } |
9562 | } |
9563 | |
9564 | /** |
9565 | * _scsih_sas_ir_operation_status_event_debug - debug for IR op event |
9566 | * @ioc: per adapter object |
9567 | * @event_data: event data payload |
9568 | * Context: user. |
9569 | */ |
9570 | static void |
9571 | _scsih_sas_ir_operation_status_event_debug(struct MPT3SAS_ADAPTER *ioc, |
9572 | Mpi2EventDataIrOperationStatus_t *event_data) |
9573 | { |
9574 | char *reason_str = NULL; |
9575 | |
9576 | switch (event_data->RAIDOperation) { |
9577 | case MPI2_EVENT_IR_RAIDOP_RESYNC: |
9578 | reason_str = "resync" ; |
9579 | break; |
9580 | case MPI2_EVENT_IR_RAIDOP_ONLINE_CAP_EXPANSION: |
9581 | reason_str = "online capacity expansion" ; |
9582 | break; |
9583 | case MPI2_EVENT_IR_RAIDOP_CONSISTENCY_CHECK: |
9584 | reason_str = "consistency check" ; |
9585 | break; |
9586 | case MPI2_EVENT_IR_RAIDOP_BACKGROUND_INIT: |
9587 | reason_str = "background init" ; |
9588 | break; |
9589 | case MPI2_EVENT_IR_RAIDOP_MAKE_DATA_CONSISTENT: |
9590 | reason_str = "make data consistent" ; |
9591 | break; |
9592 | } |
9593 | |
9594 | if (!reason_str) |
9595 | return; |
9596 | |
9597 | ioc_info(ioc, "raid operational status: (%s)\thandle(0x%04x), percent complete(%d)\n" , |
9598 | reason_str, |
9599 | le16_to_cpu(event_data->VolDevHandle), |
9600 | event_data->PercentComplete); |
9601 | } |
9602 | |
9603 | /** |
9604 | * _scsih_sas_ir_operation_status_event - handle RAID operation events |
9605 | * @ioc: per adapter object |
9606 | * @fw_event: The fw_event_work object |
9607 | * Context: user. |
9608 | */ |
9609 | static void |
9610 | _scsih_sas_ir_operation_status_event(struct MPT3SAS_ADAPTER *ioc, |
9611 | struct fw_event_work *fw_event) |
9612 | { |
9613 | Mpi2EventDataIrOperationStatus_t *event_data = |
9614 | (Mpi2EventDataIrOperationStatus_t *) |
9615 | fw_event->event_data; |
9616 | static struct _raid_device *raid_device; |
9617 | unsigned long flags; |
9618 | u16 handle; |
9619 | |
9620 | if ((ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) && |
9621 | (!ioc->hide_ir_msg)) |
9622 | _scsih_sas_ir_operation_status_event_debug(ioc, |
9623 | event_data); |
9624 | |
9625 | /* code added for raid transport support */ |
9626 | if (event_data->RAIDOperation == MPI2_EVENT_IR_RAIDOP_RESYNC) { |
9627 | |
9628 | spin_lock_irqsave(&ioc->raid_device_lock, flags); |
9629 | handle = le16_to_cpu(event_data->VolDevHandle); |
9630 | raid_device = mpt3sas_raid_device_find_by_handle(ioc, handle); |
9631 | if (raid_device) |
9632 | raid_device->percent_complete = |
9633 | event_data->PercentComplete; |
9634 | spin_unlock_irqrestore(lock: &ioc->raid_device_lock, flags); |
9635 | } |
9636 | } |
9637 | |
9638 | /** |
9639 | * _scsih_prep_device_scan - initialize parameters prior to device scan |
9640 | * @ioc: per adapter object |
9641 | * |
9642 | * Set the deleted flag prior to device scan. If the device is found during |
9643 | * the scan, then we clear the deleted flag. |
9644 | */ |
9645 | static void |
9646 | _scsih_prep_device_scan(struct MPT3SAS_ADAPTER *ioc) |
9647 | { |
9648 | struct MPT3SAS_DEVICE *sas_device_priv_data; |
9649 | struct scsi_device *sdev; |
9650 | |
9651 | shost_for_each_device(sdev, ioc->shost) { |
9652 | sas_device_priv_data = sdev->hostdata; |
9653 | if (sas_device_priv_data && sas_device_priv_data->sas_target) |
9654 | sas_device_priv_data->sas_target->deleted = 1; |
9655 | } |
9656 | } |
9657 | |
9658 | /** |
9659 | * _scsih_update_device_qdepth - Update QD during Reset. |
9660 | * @ioc: per adapter object |
9661 | * |
9662 | */ |
9663 | static void |
9664 | _scsih_update_device_qdepth(struct MPT3SAS_ADAPTER *ioc) |
9665 | { |
9666 | struct MPT3SAS_DEVICE *sas_device_priv_data; |
9667 | struct MPT3SAS_TARGET *sas_target_priv_data; |
9668 | struct _sas_device *sas_device; |
9669 | struct scsi_device *sdev; |
9670 | u16 qdepth; |
9671 | |
9672 | ioc_info(ioc, "Update devices with firmware reported queue depth\n" ); |
9673 | shost_for_each_device(sdev, ioc->shost) { |
9674 | sas_device_priv_data = sdev->hostdata; |
9675 | if (sas_device_priv_data && sas_device_priv_data->sas_target) { |
9676 | sas_target_priv_data = sas_device_priv_data->sas_target; |
9677 | sas_device = sas_device_priv_data->sas_target->sas_dev; |
9678 | if (sas_target_priv_data->flags & MPT_TARGET_FLAGS_PCIE_DEVICE) |
9679 | qdepth = ioc->max_nvme_qd; |
9680 | else if (sas_device && |
9681 | sas_device->device_info & MPI2_SAS_DEVICE_INFO_SSP_TARGET) |
9682 | qdepth = (sas_device->port_type > 1) ? |
9683 | ioc->max_wideport_qd : ioc->max_narrowport_qd; |
9684 | else if (sas_device && |
9685 | sas_device->device_info & MPI2_SAS_DEVICE_INFO_SATA_DEVICE) |
9686 | qdepth = ioc->max_sata_qd; |
9687 | else |
9688 | continue; |
9689 | mpt3sas_scsih_change_queue_depth(sdev, qdepth); |
9690 | } |
9691 | } |
9692 | } |
9693 | |
9694 | /** |
9695 | * _scsih_mark_responding_sas_device - mark a sas_devices as responding |
9696 | * @ioc: per adapter object |
9697 | * @sas_device_pg0: SAS Device page 0 |
9698 | * |
9699 | * After host reset, find out whether devices are still responding. |
9700 | * Used in _scsih_remove_unresponsive_sas_devices. |
9701 | */ |
9702 | static void |
9703 | _scsih_mark_responding_sas_device(struct MPT3SAS_ADAPTER *ioc, |
9704 | Mpi2SasDevicePage0_t *sas_device_pg0) |
9705 | { |
9706 | struct MPT3SAS_TARGET *sas_target_priv_data = NULL; |
9707 | struct scsi_target *starget; |
9708 | struct _sas_device *sas_device = NULL; |
9709 | struct _enclosure_node *enclosure_dev = NULL; |
9710 | unsigned long flags; |
9711 | struct hba_port *port = mpt3sas_get_port_by_id( |
9712 | ioc, port_id: sas_device_pg0->PhysicalPort, bypass_dirty_port_flag: 0); |
9713 | |
9714 | if (sas_device_pg0->EnclosureHandle) { |
9715 | enclosure_dev = |
9716 | mpt3sas_scsih_enclosure_find_by_handle(ioc, |
9717 | le16_to_cpu(sas_device_pg0->EnclosureHandle)); |
9718 | if (enclosure_dev == NULL) |
9719 | ioc_info(ioc, "Enclosure handle(0x%04x) doesn't match with enclosure device!\n" , |
9720 | sas_device_pg0->EnclosureHandle); |
9721 | } |
9722 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
9723 | list_for_each_entry(sas_device, &ioc->sas_device_list, list) { |
9724 | if (sas_device->sas_address != le64_to_cpu( |
9725 | sas_device_pg0->SASAddress)) |
9726 | continue; |
9727 | if (sas_device->slot != le16_to_cpu(sas_device_pg0->Slot)) |
9728 | continue; |
9729 | if (sas_device->port != port) |
9730 | continue; |
9731 | sas_device->responding = 1; |
9732 | starget = sas_device->starget; |
9733 | if (starget && starget->hostdata) { |
9734 | sas_target_priv_data = starget->hostdata; |
9735 | sas_target_priv_data->tm_busy = 0; |
9736 | sas_target_priv_data->deleted = 0; |
9737 | } else |
9738 | sas_target_priv_data = NULL; |
9739 | if (starget) { |
9740 | starget_printk(KERN_INFO, starget, |
9741 | "handle(0x%04x), sas_addr(0x%016llx)\n" , |
9742 | le16_to_cpu(sas_device_pg0->DevHandle), |
9743 | (unsigned long long) |
9744 | sas_device->sas_address); |
9745 | |
9746 | if (sas_device->enclosure_handle != 0) |
9747 | starget_printk(KERN_INFO, starget, |
9748 | "enclosure logical id(0x%016llx), slot(%d)\n" , |
9749 | (unsigned long long) |
9750 | sas_device->enclosure_logical_id, |
9751 | sas_device->slot); |
9752 | } |
9753 | if (le16_to_cpu(sas_device_pg0->Flags) & |
9754 | MPI2_SAS_DEVICE0_FLAGS_ENCL_LEVEL_VALID) { |
9755 | sas_device->enclosure_level = |
9756 | sas_device_pg0->EnclosureLevel; |
9757 | memcpy(&sas_device->connector_name[0], |
9758 | &sas_device_pg0->ConnectorName[0], 4); |
9759 | } else { |
9760 | sas_device->enclosure_level = 0; |
9761 | sas_device->connector_name[0] = '\0'; |
9762 | } |
9763 | |
9764 | sas_device->enclosure_handle = |
9765 | le16_to_cpu(sas_device_pg0->EnclosureHandle); |
9766 | sas_device->is_chassis_slot_valid = 0; |
9767 | if (enclosure_dev) { |
9768 | sas_device->enclosure_logical_id = le64_to_cpu( |
9769 | enclosure_dev->pg0.EnclosureLogicalID); |
9770 | if (le16_to_cpu(enclosure_dev->pg0.Flags) & |
9771 | MPI2_SAS_ENCLS0_FLAGS_CHASSIS_SLOT_VALID) { |
9772 | sas_device->is_chassis_slot_valid = 1; |
9773 | sas_device->chassis_slot = |
9774 | enclosure_dev->pg0.ChassisSlot; |
9775 | } |
9776 | } |
9777 | |
9778 | if (sas_device->handle == le16_to_cpu( |
9779 | sas_device_pg0->DevHandle)) |
9780 | goto out; |
9781 | pr_info("\thandle changed from(0x%04x)!!!\n" , |
9782 | sas_device->handle); |
9783 | sas_device->handle = le16_to_cpu( |
9784 | sas_device_pg0->DevHandle); |
9785 | if (sas_target_priv_data) |
9786 | sas_target_priv_data->handle = |
9787 | le16_to_cpu(sas_device_pg0->DevHandle); |
9788 | goto out; |
9789 | } |
9790 | out: |
9791 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
9792 | } |
9793 | |
9794 | /** |
9795 | * _scsih_create_enclosure_list_after_reset - Free Existing list, |
9796 | * And create enclosure list by scanning all Enclosure Page(0)s |
9797 | * @ioc: per adapter object |
9798 | */ |
9799 | static void |
9800 | _scsih_create_enclosure_list_after_reset(struct MPT3SAS_ADAPTER *ioc) |
9801 | { |
9802 | struct _enclosure_node *enclosure_dev; |
9803 | Mpi2ConfigReply_t mpi_reply; |
9804 | u16 enclosure_handle; |
9805 | int rc; |
9806 | |
9807 | /* Free existing enclosure list */ |
9808 | mpt3sas_free_enclosure_list(ioc); |
9809 | |
9810 | /* Re constructing enclosure list after reset*/ |
9811 | enclosure_handle = 0xFFFF; |
9812 | do { |
9813 | enclosure_dev = |
9814 | kzalloc(size: sizeof(struct _enclosure_node), GFP_KERNEL); |
9815 | if (!enclosure_dev) { |
9816 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
9817 | __FILE__, __LINE__, __func__); |
9818 | return; |
9819 | } |
9820 | rc = mpt3sas_config_get_enclosure_pg0(ioc, mpi_reply: &mpi_reply, |
9821 | config_page: &enclosure_dev->pg0, |
9822 | MPI2_SAS_ENCLOS_PGAD_FORM_GET_NEXT_HANDLE, |
9823 | handle: enclosure_handle); |
9824 | |
9825 | if (rc || (le16_to_cpu(mpi_reply.IOCStatus) & |
9826 | MPI2_IOCSTATUS_MASK)) { |
9827 | kfree(objp: enclosure_dev); |
9828 | return; |
9829 | } |
9830 | list_add_tail(new: &enclosure_dev->list, |
9831 | head: &ioc->enclosure_list); |
9832 | enclosure_handle = |
9833 | le16_to_cpu(enclosure_dev->pg0.EnclosureHandle); |
9834 | } while (1); |
9835 | } |
9836 | |
9837 | /** |
9838 | * _scsih_search_responding_sas_devices - |
9839 | * @ioc: per adapter object |
9840 | * |
9841 | * After host reset, find out whether devices are still responding. |
9842 | * If not remove. |
9843 | */ |
9844 | static void |
9845 | _scsih_search_responding_sas_devices(struct MPT3SAS_ADAPTER *ioc) |
9846 | { |
9847 | Mpi2SasDevicePage0_t sas_device_pg0; |
9848 | Mpi2ConfigReply_t mpi_reply; |
9849 | u16 ioc_status; |
9850 | u16 handle; |
9851 | u32 device_info; |
9852 | |
9853 | ioc_info(ioc, "search for end-devices: start\n" ); |
9854 | |
9855 | if (list_empty(head: &ioc->sas_device_list)) |
9856 | goto out; |
9857 | |
9858 | handle = 0xFFFF; |
9859 | while (!(mpt3sas_config_get_sas_device_pg0(ioc, mpi_reply: &mpi_reply, |
9860 | config_page: &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_GET_NEXT_HANDLE, |
9861 | handle))) { |
9862 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & |
9863 | MPI2_IOCSTATUS_MASK; |
9864 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) |
9865 | break; |
9866 | handle = le16_to_cpu(sas_device_pg0.DevHandle); |
9867 | device_info = le32_to_cpu(sas_device_pg0.DeviceInfo); |
9868 | if (!(_scsih_is_end_device(device_info))) |
9869 | continue; |
9870 | _scsih_mark_responding_sas_device(ioc, sas_device_pg0: &sas_device_pg0); |
9871 | } |
9872 | |
9873 | out: |
9874 | ioc_info(ioc, "search for end-devices: complete\n" ); |
9875 | } |
9876 | |
9877 | /** |
9878 | * _scsih_mark_responding_pcie_device - mark a pcie_device as responding |
9879 | * @ioc: per adapter object |
9880 | * @pcie_device_pg0: PCIe Device page 0 |
9881 | * |
9882 | * After host reset, find out whether devices are still responding. |
9883 | * Used in _scsih_remove_unresponding_devices. |
9884 | */ |
9885 | static void |
9886 | _scsih_mark_responding_pcie_device(struct MPT3SAS_ADAPTER *ioc, |
9887 | Mpi26PCIeDevicePage0_t *pcie_device_pg0) |
9888 | { |
9889 | struct MPT3SAS_TARGET *sas_target_priv_data = NULL; |
9890 | struct scsi_target *starget; |
9891 | struct _pcie_device *pcie_device; |
9892 | unsigned long flags; |
9893 | |
9894 | spin_lock_irqsave(&ioc->pcie_device_lock, flags); |
9895 | list_for_each_entry(pcie_device, &ioc->pcie_device_list, list) { |
9896 | if ((pcie_device->wwid == le64_to_cpu(pcie_device_pg0->WWID)) |
9897 | && (pcie_device->slot == le16_to_cpu( |
9898 | pcie_device_pg0->Slot))) { |
9899 | pcie_device->access_status = |
9900 | pcie_device_pg0->AccessStatus; |
9901 | pcie_device->responding = 1; |
9902 | starget = pcie_device->starget; |
9903 | if (starget && starget->hostdata) { |
9904 | sas_target_priv_data = starget->hostdata; |
9905 | sas_target_priv_data->tm_busy = 0; |
9906 | sas_target_priv_data->deleted = 0; |
9907 | } else |
9908 | sas_target_priv_data = NULL; |
9909 | if (starget) { |
9910 | starget_printk(KERN_INFO, starget, |
9911 | "handle(0x%04x), wwid(0x%016llx) " , |
9912 | pcie_device->handle, |
9913 | (unsigned long long)pcie_device->wwid); |
9914 | if (pcie_device->enclosure_handle != 0) |
9915 | starget_printk(KERN_INFO, starget, |
9916 | "enclosure logical id(0x%016llx), " |
9917 | "slot(%d)\n" , |
9918 | (unsigned long long) |
9919 | pcie_device->enclosure_logical_id, |
9920 | pcie_device->slot); |
9921 | } |
9922 | |
9923 | if (((le32_to_cpu(pcie_device_pg0->Flags)) & |
9924 | MPI26_PCIEDEV0_FLAGS_ENCL_LEVEL_VALID) && |
9925 | (ioc->hba_mpi_version_belonged != MPI2_VERSION)) { |
9926 | pcie_device->enclosure_level = |
9927 | pcie_device_pg0->EnclosureLevel; |
9928 | memcpy(&pcie_device->connector_name[0], |
9929 | &pcie_device_pg0->ConnectorName[0], 4); |
9930 | } else { |
9931 | pcie_device->enclosure_level = 0; |
9932 | pcie_device->connector_name[0] = '\0'; |
9933 | } |
9934 | |
9935 | if (pcie_device->handle == le16_to_cpu( |
9936 | pcie_device_pg0->DevHandle)) |
9937 | goto out; |
9938 | pr_info("\thandle changed from(0x%04x)!!!\n" , |
9939 | pcie_device->handle); |
9940 | pcie_device->handle = le16_to_cpu( |
9941 | pcie_device_pg0->DevHandle); |
9942 | if (sas_target_priv_data) |
9943 | sas_target_priv_data->handle = |
9944 | le16_to_cpu(pcie_device_pg0->DevHandle); |
9945 | goto out; |
9946 | } |
9947 | } |
9948 | |
9949 | out: |
9950 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
9951 | } |
9952 | |
9953 | /** |
9954 | * _scsih_search_responding_pcie_devices - |
9955 | * @ioc: per adapter object |
9956 | * |
9957 | * After host reset, find out whether devices are still responding. |
9958 | * If not remove. |
9959 | */ |
9960 | static void |
9961 | _scsih_search_responding_pcie_devices(struct MPT3SAS_ADAPTER *ioc) |
9962 | { |
9963 | Mpi26PCIeDevicePage0_t pcie_device_pg0; |
9964 | Mpi2ConfigReply_t mpi_reply; |
9965 | u16 ioc_status; |
9966 | u16 handle; |
9967 | u32 device_info; |
9968 | |
9969 | ioc_info(ioc, "search for end-devices: start\n" ); |
9970 | |
9971 | if (list_empty(head: &ioc->pcie_device_list)) |
9972 | goto out; |
9973 | |
9974 | handle = 0xFFFF; |
9975 | while (!(mpt3sas_config_get_pcie_device_pg0(ioc, mpi_reply: &mpi_reply, |
9976 | config_page: &pcie_device_pg0, MPI26_PCIE_DEVICE_PGAD_FORM_GET_NEXT_HANDLE, |
9977 | handle))) { |
9978 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & |
9979 | MPI2_IOCSTATUS_MASK; |
9980 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { |
9981 | ioc_info(ioc, "\tbreak from %s: ioc_status(0x%04x), loginfo(0x%08x)\n" , |
9982 | __func__, ioc_status, |
9983 | le32_to_cpu(mpi_reply.IOCLogInfo)); |
9984 | break; |
9985 | } |
9986 | handle = le16_to_cpu(pcie_device_pg0.DevHandle); |
9987 | device_info = le32_to_cpu(pcie_device_pg0.DeviceInfo); |
9988 | if (!(_scsih_is_nvme_pciescsi_device(device_info))) |
9989 | continue; |
9990 | _scsih_mark_responding_pcie_device(ioc, pcie_device_pg0: &pcie_device_pg0); |
9991 | } |
9992 | out: |
9993 | ioc_info(ioc, "search for PCIe end-devices: complete\n" ); |
9994 | } |
9995 | |
9996 | /** |
9997 | * _scsih_mark_responding_raid_device - mark a raid_device as responding |
9998 | * @ioc: per adapter object |
9999 | * @wwid: world wide identifier for raid volume |
10000 | * @handle: device handle |
10001 | * |
10002 | * After host reset, find out whether devices are still responding. |
10003 | * Used in _scsih_remove_unresponsive_raid_devices. |
10004 | */ |
10005 | static void |
10006 | _scsih_mark_responding_raid_device(struct MPT3SAS_ADAPTER *ioc, u64 wwid, |
10007 | u16 handle) |
10008 | { |
10009 | struct MPT3SAS_TARGET *sas_target_priv_data = NULL; |
10010 | struct scsi_target *starget; |
10011 | struct _raid_device *raid_device; |
10012 | unsigned long flags; |
10013 | |
10014 | spin_lock_irqsave(&ioc->raid_device_lock, flags); |
10015 | list_for_each_entry(raid_device, &ioc->raid_device_list, list) { |
10016 | if (raid_device->wwid == wwid && raid_device->starget) { |
10017 | starget = raid_device->starget; |
10018 | if (starget && starget->hostdata) { |
10019 | sas_target_priv_data = starget->hostdata; |
10020 | sas_target_priv_data->deleted = 0; |
10021 | } else |
10022 | sas_target_priv_data = NULL; |
10023 | raid_device->responding = 1; |
10024 | spin_unlock_irqrestore(lock: &ioc->raid_device_lock, flags); |
10025 | starget_printk(KERN_INFO, raid_device->starget, |
10026 | "handle(0x%04x), wwid(0x%016llx)\n" , handle, |
10027 | (unsigned long long)raid_device->wwid); |
10028 | |
10029 | /* |
10030 | * WARPDRIVE: The handles of the PDs might have changed |
10031 | * across the host reset so re-initialize the |
10032 | * required data for Direct IO |
10033 | */ |
10034 | mpt3sas_init_warpdrive_properties(ioc, raid_device); |
10035 | spin_lock_irqsave(&ioc->raid_device_lock, flags); |
10036 | if (raid_device->handle == handle) { |
10037 | spin_unlock_irqrestore(lock: &ioc->raid_device_lock, |
10038 | flags); |
10039 | return; |
10040 | } |
10041 | pr_info("\thandle changed from(0x%04x)!!!\n" , |
10042 | raid_device->handle); |
10043 | raid_device->handle = handle; |
10044 | if (sas_target_priv_data) |
10045 | sas_target_priv_data->handle = handle; |
10046 | spin_unlock_irqrestore(lock: &ioc->raid_device_lock, flags); |
10047 | return; |
10048 | } |
10049 | } |
10050 | spin_unlock_irqrestore(lock: &ioc->raid_device_lock, flags); |
10051 | } |
10052 | |
10053 | /** |
10054 | * _scsih_search_responding_raid_devices - |
10055 | * @ioc: per adapter object |
10056 | * |
10057 | * After host reset, find out whether devices are still responding. |
10058 | * If not remove. |
10059 | */ |
10060 | static void |
10061 | _scsih_search_responding_raid_devices(struct MPT3SAS_ADAPTER *ioc) |
10062 | { |
10063 | Mpi2RaidVolPage1_t volume_pg1; |
10064 | Mpi2RaidVolPage0_t volume_pg0; |
10065 | Mpi2RaidPhysDiskPage0_t pd_pg0; |
10066 | Mpi2ConfigReply_t mpi_reply; |
10067 | u16 ioc_status; |
10068 | u16 handle; |
10069 | u8 phys_disk_num; |
10070 | |
10071 | if (!ioc->ir_firmware) |
10072 | return; |
10073 | |
10074 | ioc_info(ioc, "search for raid volumes: start\n" ); |
10075 | |
10076 | if (list_empty(head: &ioc->raid_device_list)) |
10077 | goto out; |
10078 | |
10079 | handle = 0xFFFF; |
10080 | while (!(mpt3sas_config_get_raid_volume_pg1(ioc, mpi_reply: &mpi_reply, |
10081 | config_page: &volume_pg1, MPI2_RAID_VOLUME_PGAD_FORM_GET_NEXT_HANDLE, handle))) { |
10082 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & |
10083 | MPI2_IOCSTATUS_MASK; |
10084 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) |
10085 | break; |
10086 | handle = le16_to_cpu(volume_pg1.DevHandle); |
10087 | |
10088 | if (mpt3sas_config_get_raid_volume_pg0(ioc, mpi_reply: &mpi_reply, |
10089 | config_page: &volume_pg0, MPI2_RAID_VOLUME_PGAD_FORM_HANDLE, handle, |
10090 | sz: sizeof(Mpi2RaidVolPage0_t))) |
10091 | continue; |
10092 | |
10093 | if (volume_pg0.VolumeState == MPI2_RAID_VOL_STATE_OPTIMAL || |
10094 | volume_pg0.VolumeState == MPI2_RAID_VOL_STATE_ONLINE || |
10095 | volume_pg0.VolumeState == MPI2_RAID_VOL_STATE_DEGRADED) |
10096 | _scsih_mark_responding_raid_device(ioc, |
10097 | le64_to_cpu(volume_pg1.WWID), handle); |
10098 | } |
10099 | |
10100 | /* refresh the pd_handles */ |
10101 | if (!ioc->is_warpdrive) { |
10102 | phys_disk_num = 0xFF; |
10103 | memset(ioc->pd_handles, 0, ioc->pd_handles_sz); |
10104 | while (!(mpt3sas_config_get_phys_disk_pg0(ioc, mpi_reply: &mpi_reply, |
10105 | config_page: &pd_pg0, MPI2_PHYSDISK_PGAD_FORM_GET_NEXT_PHYSDISKNUM, |
10106 | form_specific: phys_disk_num))) { |
10107 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & |
10108 | MPI2_IOCSTATUS_MASK; |
10109 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) |
10110 | break; |
10111 | phys_disk_num = pd_pg0.PhysDiskNum; |
10112 | handle = le16_to_cpu(pd_pg0.DevHandle); |
10113 | set_bit(nr: handle, addr: ioc->pd_handles); |
10114 | } |
10115 | } |
10116 | out: |
10117 | ioc_info(ioc, "search for responding raid volumes: complete\n" ); |
10118 | } |
10119 | |
10120 | /** |
10121 | * _scsih_mark_responding_expander - mark a expander as responding |
10122 | * @ioc: per adapter object |
10123 | * @expander_pg0:SAS Expander Config Page0 |
10124 | * |
10125 | * After host reset, find out whether devices are still responding. |
10126 | * Used in _scsih_remove_unresponsive_expanders. |
10127 | */ |
10128 | static void |
10129 | _scsih_mark_responding_expander(struct MPT3SAS_ADAPTER *ioc, |
10130 | Mpi2ExpanderPage0_t *expander_pg0) |
10131 | { |
10132 | struct _sas_node *sas_expander = NULL; |
10133 | unsigned long flags; |
10134 | int i; |
10135 | struct _enclosure_node *enclosure_dev = NULL; |
10136 | u16 handle = le16_to_cpu(expander_pg0->DevHandle); |
10137 | u16 enclosure_handle = le16_to_cpu(expander_pg0->EnclosureHandle); |
10138 | u64 sas_address = le64_to_cpu(expander_pg0->SASAddress); |
10139 | struct hba_port *port = mpt3sas_get_port_by_id( |
10140 | ioc, port_id: expander_pg0->PhysicalPort, bypass_dirty_port_flag: 0); |
10141 | |
10142 | if (enclosure_handle) |
10143 | enclosure_dev = |
10144 | mpt3sas_scsih_enclosure_find_by_handle(ioc, |
10145 | handle: enclosure_handle); |
10146 | |
10147 | spin_lock_irqsave(&ioc->sas_node_lock, flags); |
10148 | list_for_each_entry(sas_expander, &ioc->sas_expander_list, list) { |
10149 | if (sas_expander->sas_address != sas_address) |
10150 | continue; |
10151 | if (sas_expander->port != port) |
10152 | continue; |
10153 | sas_expander->responding = 1; |
10154 | |
10155 | if (enclosure_dev) { |
10156 | sas_expander->enclosure_logical_id = |
10157 | le64_to_cpu(enclosure_dev->pg0.EnclosureLogicalID); |
10158 | sas_expander->enclosure_handle = |
10159 | le16_to_cpu(expander_pg0->EnclosureHandle); |
10160 | } |
10161 | |
10162 | if (sas_expander->handle == handle) |
10163 | goto out; |
10164 | pr_info("\texpander(0x%016llx): handle changed" \ |
10165 | " from(0x%04x) to (0x%04x)!!!\n" , |
10166 | (unsigned long long)sas_expander->sas_address, |
10167 | sas_expander->handle, handle); |
10168 | sas_expander->handle = handle; |
10169 | for (i = 0 ; i < sas_expander->num_phys ; i++) |
10170 | sas_expander->phy[i].handle = handle; |
10171 | goto out; |
10172 | } |
10173 | out: |
10174 | spin_unlock_irqrestore(lock: &ioc->sas_node_lock, flags); |
10175 | } |
10176 | |
10177 | /** |
10178 | * _scsih_search_responding_expanders - |
10179 | * @ioc: per adapter object |
10180 | * |
10181 | * After host reset, find out whether devices are still responding. |
10182 | * If not remove. |
10183 | */ |
10184 | static void |
10185 | _scsih_search_responding_expanders(struct MPT3SAS_ADAPTER *ioc) |
10186 | { |
10187 | Mpi2ExpanderPage0_t expander_pg0; |
10188 | Mpi2ConfigReply_t mpi_reply; |
10189 | u16 ioc_status; |
10190 | u64 sas_address; |
10191 | u16 handle; |
10192 | u8 port; |
10193 | |
10194 | ioc_info(ioc, "search for expanders: start\n" ); |
10195 | |
10196 | if (list_empty(head: &ioc->sas_expander_list)) |
10197 | goto out; |
10198 | |
10199 | handle = 0xFFFF; |
10200 | while (!(mpt3sas_config_get_expander_pg0(ioc, mpi_reply: &mpi_reply, config_page: &expander_pg0, |
10201 | MPI2_SAS_EXPAND_PGAD_FORM_GET_NEXT_HNDL, handle))) { |
10202 | |
10203 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & |
10204 | MPI2_IOCSTATUS_MASK; |
10205 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) |
10206 | break; |
10207 | |
10208 | handle = le16_to_cpu(expander_pg0.DevHandle); |
10209 | sas_address = le64_to_cpu(expander_pg0.SASAddress); |
10210 | port = expander_pg0.PhysicalPort; |
10211 | pr_info( |
10212 | "\texpander present: handle(0x%04x), sas_addr(0x%016llx), port:%d\n" , |
10213 | handle, (unsigned long long)sas_address, |
10214 | (ioc->multipath_on_hba ? |
10215 | port : MULTIPATH_DISABLED_PORT_ID)); |
10216 | _scsih_mark_responding_expander(ioc, expander_pg0: &expander_pg0); |
10217 | } |
10218 | |
10219 | out: |
10220 | ioc_info(ioc, "search for expanders: complete\n" ); |
10221 | } |
10222 | |
10223 | /** |
10224 | * _scsih_remove_unresponding_devices - removing unresponding devices |
10225 | * @ioc: per adapter object |
10226 | */ |
10227 | static void |
10228 | _scsih_remove_unresponding_devices(struct MPT3SAS_ADAPTER *ioc) |
10229 | { |
10230 | struct _sas_device *sas_device, *sas_device_next; |
10231 | struct _sas_node *sas_expander, *sas_expander_next; |
10232 | struct _raid_device *raid_device, *raid_device_next; |
10233 | struct _pcie_device *pcie_device, *pcie_device_next; |
10234 | struct list_head tmp_list; |
10235 | unsigned long flags; |
10236 | LIST_HEAD(head); |
10237 | |
10238 | ioc_info(ioc, "removing unresponding devices: start\n" ); |
10239 | |
10240 | /* removing unresponding end devices */ |
10241 | ioc_info(ioc, "removing unresponding devices: end-devices\n" ); |
10242 | /* |
10243 | * Iterate, pulling off devices marked as non-responding. We become the |
10244 | * owner for the reference the list had on any object we prune. |
10245 | */ |
10246 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
10247 | |
10248 | /* |
10249 | * Clean up the sas_device_init_list list as |
10250 | * driver goes for fresh scan as part of diag reset. |
10251 | */ |
10252 | list_for_each_entry_safe(sas_device, sas_device_next, |
10253 | &ioc->sas_device_init_list, list) { |
10254 | list_del_init(entry: &sas_device->list); |
10255 | sas_device_put(s: sas_device); |
10256 | } |
10257 | |
10258 | list_for_each_entry_safe(sas_device, sas_device_next, |
10259 | &ioc->sas_device_list, list) { |
10260 | if (!sas_device->responding) |
10261 | list_move_tail(list: &sas_device->list, head: &head); |
10262 | else |
10263 | sas_device->responding = 0; |
10264 | } |
10265 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
10266 | |
10267 | /* |
10268 | * Now, uninitialize and remove the unresponding devices we pruned. |
10269 | */ |
10270 | list_for_each_entry_safe(sas_device, sas_device_next, &head, list) { |
10271 | _scsih_remove_device(ioc, sas_device); |
10272 | list_del_init(entry: &sas_device->list); |
10273 | sas_device_put(s: sas_device); |
10274 | } |
10275 | |
10276 | ioc_info(ioc, "Removing unresponding devices: pcie end-devices\n" ); |
10277 | INIT_LIST_HEAD(list: &head); |
10278 | spin_lock_irqsave(&ioc->pcie_device_lock, flags); |
10279 | /* |
10280 | * Clean up the pcie_device_init_list list as |
10281 | * driver goes for fresh scan as part of diag reset. |
10282 | */ |
10283 | list_for_each_entry_safe(pcie_device, pcie_device_next, |
10284 | &ioc->pcie_device_init_list, list) { |
10285 | list_del_init(entry: &pcie_device->list); |
10286 | pcie_device_put(p: pcie_device); |
10287 | } |
10288 | |
10289 | list_for_each_entry_safe(pcie_device, pcie_device_next, |
10290 | &ioc->pcie_device_list, list) { |
10291 | if (!pcie_device->responding) |
10292 | list_move_tail(list: &pcie_device->list, head: &head); |
10293 | else |
10294 | pcie_device->responding = 0; |
10295 | } |
10296 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
10297 | |
10298 | list_for_each_entry_safe(pcie_device, pcie_device_next, &head, list) { |
10299 | _scsih_pcie_device_remove_from_sml(ioc, pcie_device); |
10300 | list_del_init(entry: &pcie_device->list); |
10301 | pcie_device_put(p: pcie_device); |
10302 | } |
10303 | |
10304 | /* removing unresponding volumes */ |
10305 | if (ioc->ir_firmware) { |
10306 | ioc_info(ioc, "removing unresponding devices: volumes\n" ); |
10307 | list_for_each_entry_safe(raid_device, raid_device_next, |
10308 | &ioc->raid_device_list, list) { |
10309 | if (!raid_device->responding) |
10310 | _scsih_sas_volume_delete(ioc, |
10311 | handle: raid_device->handle); |
10312 | else |
10313 | raid_device->responding = 0; |
10314 | } |
10315 | } |
10316 | |
10317 | /* removing unresponding expanders */ |
10318 | ioc_info(ioc, "removing unresponding devices: expanders\n" ); |
10319 | spin_lock_irqsave(&ioc->sas_node_lock, flags); |
10320 | INIT_LIST_HEAD(list: &tmp_list); |
10321 | list_for_each_entry_safe(sas_expander, sas_expander_next, |
10322 | &ioc->sas_expander_list, list) { |
10323 | if (!sas_expander->responding) |
10324 | list_move_tail(list: &sas_expander->list, head: &tmp_list); |
10325 | else |
10326 | sas_expander->responding = 0; |
10327 | } |
10328 | spin_unlock_irqrestore(lock: &ioc->sas_node_lock, flags); |
10329 | list_for_each_entry_safe(sas_expander, sas_expander_next, &tmp_list, |
10330 | list) { |
10331 | _scsih_expander_node_remove(ioc, sas_expander); |
10332 | } |
10333 | |
10334 | ioc_info(ioc, "removing unresponding devices: complete\n" ); |
10335 | |
10336 | /* unblock devices */ |
10337 | _scsih_ublock_io_all_device(ioc); |
10338 | } |
10339 | |
10340 | static void |
10341 | _scsih_refresh_expander_links(struct MPT3SAS_ADAPTER *ioc, |
10342 | struct _sas_node *sas_expander, u16 handle) |
10343 | { |
10344 | Mpi2ExpanderPage1_t expander_pg1; |
10345 | Mpi2ConfigReply_t mpi_reply; |
10346 | int i; |
10347 | |
10348 | for (i = 0 ; i < sas_expander->num_phys ; i++) { |
10349 | if ((mpt3sas_config_get_expander_pg1(ioc, mpi_reply: &mpi_reply, |
10350 | config_page: &expander_pg1, phy_number: i, handle))) { |
10351 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
10352 | __FILE__, __LINE__, __func__); |
10353 | return; |
10354 | } |
10355 | |
10356 | mpt3sas_transport_update_links(ioc, sas_address: sas_expander->sas_address, |
10357 | le16_to_cpu(expander_pg1.AttachedDevHandle), phy_number: i, |
10358 | link_rate: expander_pg1.NegotiatedLinkRate >> 4, |
10359 | port: sas_expander->port); |
10360 | } |
10361 | } |
10362 | |
10363 | /** |
10364 | * _scsih_scan_for_devices_after_reset - scan for devices after host reset |
10365 | * @ioc: per adapter object |
10366 | */ |
10367 | static void |
10368 | _scsih_scan_for_devices_after_reset(struct MPT3SAS_ADAPTER *ioc) |
10369 | { |
10370 | Mpi2ExpanderPage0_t expander_pg0; |
10371 | Mpi2SasDevicePage0_t sas_device_pg0; |
10372 | Mpi26PCIeDevicePage0_t pcie_device_pg0; |
10373 | Mpi2RaidVolPage1_t volume_pg1; |
10374 | Mpi2RaidVolPage0_t volume_pg0; |
10375 | Mpi2RaidPhysDiskPage0_t pd_pg0; |
10376 | Mpi2EventIrConfigElement_t element; |
10377 | Mpi2ConfigReply_t mpi_reply; |
10378 | u8 phys_disk_num, port_id; |
10379 | u16 ioc_status; |
10380 | u16 handle, parent_handle; |
10381 | u64 sas_address; |
10382 | struct _sas_device *sas_device; |
10383 | struct _pcie_device *pcie_device; |
10384 | struct _sas_node *expander_device; |
10385 | static struct _raid_device *raid_device; |
10386 | u8 retry_count; |
10387 | unsigned long flags; |
10388 | |
10389 | ioc_info(ioc, "scan devices: start\n" ); |
10390 | |
10391 | _scsih_sas_host_refresh(ioc); |
10392 | |
10393 | ioc_info(ioc, "\tscan devices: expanders start\n" ); |
10394 | |
10395 | /* expanders */ |
10396 | handle = 0xFFFF; |
10397 | while (!(mpt3sas_config_get_expander_pg0(ioc, mpi_reply: &mpi_reply, config_page: &expander_pg0, |
10398 | MPI2_SAS_EXPAND_PGAD_FORM_GET_NEXT_HNDL, handle))) { |
10399 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & |
10400 | MPI2_IOCSTATUS_MASK; |
10401 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { |
10402 | ioc_info(ioc, "\tbreak from expander scan: ioc_status(0x%04x), loginfo(0x%08x)\n" , |
10403 | ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo)); |
10404 | break; |
10405 | } |
10406 | handle = le16_to_cpu(expander_pg0.DevHandle); |
10407 | spin_lock_irqsave(&ioc->sas_node_lock, flags); |
10408 | port_id = expander_pg0.PhysicalPort; |
10409 | expander_device = mpt3sas_scsih_expander_find_by_sas_address( |
10410 | ioc, le64_to_cpu(expander_pg0.SASAddress), |
10411 | port: mpt3sas_get_port_by_id(ioc, port_id, bypass_dirty_port_flag: 0)); |
10412 | spin_unlock_irqrestore(lock: &ioc->sas_node_lock, flags); |
10413 | if (expander_device) |
10414 | _scsih_refresh_expander_links(ioc, sas_expander: expander_device, |
10415 | handle); |
10416 | else { |
10417 | ioc_info(ioc, "\tBEFORE adding expander: handle (0x%04x), sas_addr(0x%016llx)\n" , |
10418 | handle, |
10419 | (u64)le64_to_cpu(expander_pg0.SASAddress)); |
10420 | _scsih_expander_add(ioc, handle); |
10421 | ioc_info(ioc, "\tAFTER adding expander: handle (0x%04x), sas_addr(0x%016llx)\n" , |
10422 | handle, |
10423 | (u64)le64_to_cpu(expander_pg0.SASAddress)); |
10424 | } |
10425 | } |
10426 | |
10427 | ioc_info(ioc, "\tscan devices: expanders complete\n" ); |
10428 | |
10429 | if (!ioc->ir_firmware) |
10430 | goto skip_to_sas; |
10431 | |
10432 | ioc_info(ioc, "\tscan devices: phys disk start\n" ); |
10433 | |
10434 | /* phys disk */ |
10435 | phys_disk_num = 0xFF; |
10436 | while (!(mpt3sas_config_get_phys_disk_pg0(ioc, mpi_reply: &mpi_reply, |
10437 | config_page: &pd_pg0, MPI2_PHYSDISK_PGAD_FORM_GET_NEXT_PHYSDISKNUM, |
10438 | form_specific: phys_disk_num))) { |
10439 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & |
10440 | MPI2_IOCSTATUS_MASK; |
10441 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { |
10442 | ioc_info(ioc, "\tbreak from phys disk scan: ioc_status(0x%04x), loginfo(0x%08x)\n" , |
10443 | ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo)); |
10444 | break; |
10445 | } |
10446 | phys_disk_num = pd_pg0.PhysDiskNum; |
10447 | handle = le16_to_cpu(pd_pg0.DevHandle); |
10448 | sas_device = mpt3sas_get_sdev_by_handle(ioc, handle); |
10449 | if (sas_device) { |
10450 | sas_device_put(s: sas_device); |
10451 | continue; |
10452 | } |
10453 | if (mpt3sas_config_get_sas_device_pg0(ioc, mpi_reply: &mpi_reply, |
10454 | config_page: &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_HANDLE, |
10455 | handle) != 0) |
10456 | continue; |
10457 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & |
10458 | MPI2_IOCSTATUS_MASK; |
10459 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { |
10460 | ioc_info(ioc, "\tbreak from phys disk scan ioc_status(0x%04x), loginfo(0x%08x)\n" , |
10461 | ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo)); |
10462 | break; |
10463 | } |
10464 | parent_handle = le16_to_cpu(sas_device_pg0.ParentDevHandle); |
10465 | if (!_scsih_get_sas_address(ioc, handle: parent_handle, |
10466 | sas_address: &sas_address)) { |
10467 | ioc_info(ioc, "\tBEFORE adding phys disk: handle (0x%04x), sas_addr(0x%016llx)\n" , |
10468 | handle, |
10469 | (u64)le64_to_cpu(sas_device_pg0.SASAddress)); |
10470 | port_id = sas_device_pg0.PhysicalPort; |
10471 | mpt3sas_transport_update_links(ioc, sas_address, |
10472 | handle, phy_number: sas_device_pg0.PhyNum, |
10473 | MPI2_SAS_NEG_LINK_RATE_1_5, |
10474 | port: mpt3sas_get_port_by_id(ioc, port_id, bypass_dirty_port_flag: 0)); |
10475 | set_bit(nr: handle, addr: ioc->pd_handles); |
10476 | retry_count = 0; |
10477 | /* This will retry adding the end device. |
10478 | * _scsih_add_device() will decide on retries and |
10479 | * return "1" when it should be retried |
10480 | */ |
10481 | while (_scsih_add_device(ioc, handle, phy_num: retry_count++, |
10482 | is_pd: 1)) { |
10483 | ssleep(seconds: 1); |
10484 | } |
10485 | ioc_info(ioc, "\tAFTER adding phys disk: handle (0x%04x), sas_addr(0x%016llx)\n" , |
10486 | handle, |
10487 | (u64)le64_to_cpu(sas_device_pg0.SASAddress)); |
10488 | } |
10489 | } |
10490 | |
10491 | ioc_info(ioc, "\tscan devices: phys disk complete\n" ); |
10492 | |
10493 | ioc_info(ioc, "\tscan devices: volumes start\n" ); |
10494 | |
10495 | /* volumes */ |
10496 | handle = 0xFFFF; |
10497 | while (!(mpt3sas_config_get_raid_volume_pg1(ioc, mpi_reply: &mpi_reply, |
10498 | config_page: &volume_pg1, MPI2_RAID_VOLUME_PGAD_FORM_GET_NEXT_HANDLE, handle))) { |
10499 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & |
10500 | MPI2_IOCSTATUS_MASK; |
10501 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { |
10502 | ioc_info(ioc, "\tbreak from volume scan: ioc_status(0x%04x), loginfo(0x%08x)\n" , |
10503 | ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo)); |
10504 | break; |
10505 | } |
10506 | handle = le16_to_cpu(volume_pg1.DevHandle); |
10507 | spin_lock_irqsave(&ioc->raid_device_lock, flags); |
10508 | raid_device = _scsih_raid_device_find_by_wwid(ioc, |
10509 | le64_to_cpu(volume_pg1.WWID)); |
10510 | spin_unlock_irqrestore(lock: &ioc->raid_device_lock, flags); |
10511 | if (raid_device) |
10512 | continue; |
10513 | if (mpt3sas_config_get_raid_volume_pg0(ioc, mpi_reply: &mpi_reply, |
10514 | config_page: &volume_pg0, MPI2_RAID_VOLUME_PGAD_FORM_HANDLE, handle, |
10515 | sz: sizeof(Mpi2RaidVolPage0_t))) |
10516 | continue; |
10517 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & |
10518 | MPI2_IOCSTATUS_MASK; |
10519 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { |
10520 | ioc_info(ioc, "\tbreak from volume scan: ioc_status(0x%04x), loginfo(0x%08x)\n" , |
10521 | ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo)); |
10522 | break; |
10523 | } |
10524 | if (volume_pg0.VolumeState == MPI2_RAID_VOL_STATE_OPTIMAL || |
10525 | volume_pg0.VolumeState == MPI2_RAID_VOL_STATE_ONLINE || |
10526 | volume_pg0.VolumeState == MPI2_RAID_VOL_STATE_DEGRADED) { |
10527 | memset(&element, 0, sizeof(Mpi2EventIrConfigElement_t)); |
10528 | element.ReasonCode = MPI2_EVENT_IR_CHANGE_RC_ADDED; |
10529 | element.VolDevHandle = volume_pg1.DevHandle; |
10530 | ioc_info(ioc, "\tBEFORE adding volume: handle (0x%04x)\n" , |
10531 | volume_pg1.DevHandle); |
10532 | _scsih_sas_volume_add(ioc, element: &element); |
10533 | ioc_info(ioc, "\tAFTER adding volume: handle (0x%04x)\n" , |
10534 | volume_pg1.DevHandle); |
10535 | } |
10536 | } |
10537 | |
10538 | ioc_info(ioc, "\tscan devices: volumes complete\n" ); |
10539 | |
10540 | skip_to_sas: |
10541 | |
10542 | ioc_info(ioc, "\tscan devices: end devices start\n" ); |
10543 | |
10544 | /* sas devices */ |
10545 | handle = 0xFFFF; |
10546 | while (!(mpt3sas_config_get_sas_device_pg0(ioc, mpi_reply: &mpi_reply, |
10547 | config_page: &sas_device_pg0, MPI2_SAS_DEVICE_PGAD_FORM_GET_NEXT_HANDLE, |
10548 | handle))) { |
10549 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) & |
10550 | MPI2_IOCSTATUS_MASK; |
10551 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { |
10552 | ioc_info(ioc, "\tbreak from end device scan: ioc_status(0x%04x), loginfo(0x%08x)\n" , |
10553 | ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo)); |
10554 | break; |
10555 | } |
10556 | handle = le16_to_cpu(sas_device_pg0.DevHandle); |
10557 | if (!(_scsih_is_end_device( |
10558 | le32_to_cpu(sas_device_pg0.DeviceInfo)))) |
10559 | continue; |
10560 | port_id = sas_device_pg0.PhysicalPort; |
10561 | sas_device = mpt3sas_get_sdev_by_addr(ioc, |
10562 | le64_to_cpu(sas_device_pg0.SASAddress), |
10563 | port: mpt3sas_get_port_by_id(ioc, port_id, bypass_dirty_port_flag: 0)); |
10564 | if (sas_device) { |
10565 | sas_device_put(s: sas_device); |
10566 | continue; |
10567 | } |
10568 | parent_handle = le16_to_cpu(sas_device_pg0.ParentDevHandle); |
10569 | if (!_scsih_get_sas_address(ioc, handle: parent_handle, sas_address: &sas_address)) { |
10570 | ioc_info(ioc, "\tBEFORE adding end device: handle (0x%04x), sas_addr(0x%016llx)\n" , |
10571 | handle, |
10572 | (u64)le64_to_cpu(sas_device_pg0.SASAddress)); |
10573 | mpt3sas_transport_update_links(ioc, sas_address, handle, |
10574 | phy_number: sas_device_pg0.PhyNum, MPI2_SAS_NEG_LINK_RATE_1_5, |
10575 | port: mpt3sas_get_port_by_id(ioc, port_id, bypass_dirty_port_flag: 0)); |
10576 | retry_count = 0; |
10577 | /* This will retry adding the end device. |
10578 | * _scsih_add_device() will decide on retries and |
10579 | * return "1" when it should be retried |
10580 | */ |
10581 | while (_scsih_add_device(ioc, handle, phy_num: retry_count++, |
10582 | is_pd: 0)) { |
10583 | ssleep(seconds: 1); |
10584 | } |
10585 | ioc_info(ioc, "\tAFTER adding end device: handle (0x%04x), sas_addr(0x%016llx)\n" , |
10586 | handle, |
10587 | (u64)le64_to_cpu(sas_device_pg0.SASAddress)); |
10588 | } |
10589 | } |
10590 | ioc_info(ioc, "\tscan devices: end devices complete\n" ); |
10591 | ioc_info(ioc, "\tscan devices: pcie end devices start\n" ); |
10592 | |
10593 | /* pcie devices */ |
10594 | handle = 0xFFFF; |
10595 | while (!(mpt3sas_config_get_pcie_device_pg0(ioc, mpi_reply: &mpi_reply, |
10596 | config_page: &pcie_device_pg0, MPI26_PCIE_DEVICE_PGAD_FORM_GET_NEXT_HANDLE, |
10597 | handle))) { |
10598 | ioc_status = le16_to_cpu(mpi_reply.IOCStatus) |
10599 | & MPI2_IOCSTATUS_MASK; |
10600 | if (ioc_status != MPI2_IOCSTATUS_SUCCESS) { |
10601 | ioc_info(ioc, "\tbreak from pcie end device scan: ioc_status(0x%04x), loginfo(0x%08x)\n" , |
10602 | ioc_status, le32_to_cpu(mpi_reply.IOCLogInfo)); |
10603 | break; |
10604 | } |
10605 | handle = le16_to_cpu(pcie_device_pg0.DevHandle); |
10606 | if (!(_scsih_is_nvme_pciescsi_device( |
10607 | le32_to_cpu(pcie_device_pg0.DeviceInfo)))) |
10608 | continue; |
10609 | pcie_device = mpt3sas_get_pdev_by_wwid(ioc, |
10610 | le64_to_cpu(pcie_device_pg0.WWID)); |
10611 | if (pcie_device) { |
10612 | pcie_device_put(p: pcie_device); |
10613 | continue; |
10614 | } |
10615 | retry_count = 0; |
10616 | parent_handle = le16_to_cpu(pcie_device_pg0.ParentDevHandle); |
10617 | _scsih_pcie_add_device(ioc, handle); |
10618 | |
10619 | ioc_info(ioc, "\tAFTER adding pcie end device: handle (0x%04x), wwid(0x%016llx)\n" , |
10620 | handle, (u64)le64_to_cpu(pcie_device_pg0.WWID)); |
10621 | } |
10622 | |
10623 | ioc_info(ioc, "\tpcie devices: pcie end devices complete\n" ); |
10624 | ioc_info(ioc, "scan devices: complete\n" ); |
10625 | } |
10626 | |
10627 | /** |
10628 | * mpt3sas_scsih_pre_reset_handler - reset callback handler (for scsih) |
10629 | * @ioc: per adapter object |
10630 | * |
10631 | * The handler for doing any required cleanup or initialization. |
10632 | */ |
10633 | void mpt3sas_scsih_pre_reset_handler(struct MPT3SAS_ADAPTER *ioc) |
10634 | { |
10635 | dtmprintk(ioc, ioc_info(ioc, "%s: MPT3_IOC_PRE_RESET\n" , __func__)); |
10636 | } |
10637 | |
10638 | /** |
10639 | * mpt3sas_scsih_clear_outstanding_scsi_tm_commands - clears outstanding |
10640 | * scsi & tm cmds. |
10641 | * @ioc: per adapter object |
10642 | * |
10643 | * The handler for doing any required cleanup or initialization. |
10644 | */ |
10645 | void |
10646 | mpt3sas_scsih_clear_outstanding_scsi_tm_commands(struct MPT3SAS_ADAPTER *ioc) |
10647 | { |
10648 | dtmprintk(ioc, |
10649 | ioc_info(ioc, "%s: clear outstanding scsi & tm cmds\n" , __func__)); |
10650 | if (ioc->scsih_cmds.status & MPT3_CMD_PENDING) { |
10651 | ioc->scsih_cmds.status |= MPT3_CMD_RESET; |
10652 | mpt3sas_base_free_smid(ioc, smid: ioc->scsih_cmds.smid); |
10653 | complete(&ioc->scsih_cmds.done); |
10654 | } |
10655 | if (ioc->tm_cmds.status & MPT3_CMD_PENDING) { |
10656 | ioc->tm_cmds.status |= MPT3_CMD_RESET; |
10657 | mpt3sas_base_free_smid(ioc, smid: ioc->tm_cmds.smid); |
10658 | complete(&ioc->tm_cmds.done); |
10659 | } |
10660 | |
10661 | memset(ioc->pend_os_device_add, 0, ioc->pend_os_device_add_sz); |
10662 | memset(ioc->device_remove_in_progress, 0, |
10663 | ioc->device_remove_in_progress_sz); |
10664 | _scsih_fw_event_cleanup_queue(ioc); |
10665 | _scsih_flush_running_cmds(ioc); |
10666 | } |
10667 | |
10668 | /** |
10669 | * mpt3sas_scsih_reset_done_handler - reset callback handler (for scsih) |
10670 | * @ioc: per adapter object |
10671 | * |
10672 | * The handler for doing any required cleanup or initialization. |
10673 | */ |
10674 | void |
10675 | mpt3sas_scsih_reset_done_handler(struct MPT3SAS_ADAPTER *ioc) |
10676 | { |
10677 | dtmprintk(ioc, ioc_info(ioc, "%s: MPT3_IOC_DONE_RESET\n" , __func__)); |
10678 | if (!(disable_discovery > 0 && !ioc->sas_hba.num_phys)) { |
10679 | if (ioc->multipath_on_hba) { |
10680 | _scsih_sas_port_refresh(ioc); |
10681 | _scsih_update_vphys_after_reset(ioc); |
10682 | } |
10683 | _scsih_prep_device_scan(ioc); |
10684 | _scsih_create_enclosure_list_after_reset(ioc); |
10685 | _scsih_search_responding_sas_devices(ioc); |
10686 | _scsih_search_responding_pcie_devices(ioc); |
10687 | _scsih_search_responding_raid_devices(ioc); |
10688 | _scsih_search_responding_expanders(ioc); |
10689 | _scsih_error_recovery_delete_devices(ioc); |
10690 | } |
10691 | } |
10692 | |
10693 | /** |
10694 | * _mpt3sas_fw_work - delayed task for processing firmware events |
10695 | * @ioc: per adapter object |
10696 | * @fw_event: The fw_event_work object |
10697 | * Context: user. |
10698 | */ |
10699 | static void |
10700 | _mpt3sas_fw_work(struct MPT3SAS_ADAPTER *ioc, struct fw_event_work *fw_event) |
10701 | { |
10702 | ioc->current_event = fw_event; |
10703 | _scsih_fw_event_del_from_list(ioc, fw_event); |
10704 | |
10705 | /* the queue is being flushed so ignore this event */ |
10706 | if (ioc->remove_host || ioc->pci_error_recovery) { |
10707 | fw_event_work_put(fw_work: fw_event); |
10708 | ioc->current_event = NULL; |
10709 | return; |
10710 | } |
10711 | |
10712 | switch (fw_event->event) { |
10713 | case MPT3SAS_PROCESS_TRIGGER_DIAG: |
10714 | mpt3sas_process_trigger_data(ioc, |
10715 | event_data: (struct SL_WH_TRIGGERS_EVENT_DATA_T *) |
10716 | fw_event->event_data); |
10717 | break; |
10718 | case MPT3SAS_REMOVE_UNRESPONDING_DEVICES: |
10719 | while (scsi_host_in_recovery(shost: ioc->shost) || |
10720 | ioc->shost_recovery) { |
10721 | /* |
10722 | * If we're unloading or cancelling the work, bail. |
10723 | * Otherwise, this can become an infinite loop. |
10724 | */ |
10725 | if (ioc->remove_host || ioc->fw_events_cleanup) |
10726 | goto out; |
10727 | ssleep(seconds: 1); |
10728 | } |
10729 | _scsih_remove_unresponding_devices(ioc); |
10730 | _scsih_del_dirty_vphy(ioc); |
10731 | _scsih_del_dirty_port_entries(ioc); |
10732 | if (ioc->is_gen35_ioc) |
10733 | _scsih_update_device_qdepth(ioc); |
10734 | _scsih_scan_for_devices_after_reset(ioc); |
10735 | /* |
10736 | * If diag reset has occurred during the driver load |
10737 | * then driver has to complete the driver load operation |
10738 | * by executing the following items: |
10739 | *- Register the devices from sas_device_init_list to SML |
10740 | *- clear is_driver_loading flag, |
10741 | *- start the watchdog thread. |
10742 | * In happy driver load path, above things are taken care of when |
10743 | * driver executes scsih_scan_finished(). |
10744 | */ |
10745 | if (ioc->is_driver_loading) |
10746 | _scsih_complete_devices_scanning(ioc); |
10747 | _scsih_set_nvme_max_shutdown_latency(ioc); |
10748 | break; |
10749 | case MPT3SAS_PORT_ENABLE_COMPLETE: |
10750 | ioc->start_scan = 0; |
10751 | if (missing_delay[0] != -1 && missing_delay[1] != -1) |
10752 | mpt3sas_base_update_missing_delay(ioc, device_missing_delay: missing_delay[0], |
10753 | io_missing_delay: missing_delay[1]); |
10754 | dewtprintk(ioc, |
10755 | ioc_info(ioc, "port enable: complete from worker thread\n" )); |
10756 | break; |
10757 | case MPT3SAS_TURN_ON_PFA_LED: |
10758 | _scsih_turn_on_pfa_led(ioc, handle: fw_event->device_handle); |
10759 | break; |
10760 | case MPI2_EVENT_SAS_TOPOLOGY_CHANGE_LIST: |
10761 | _scsih_sas_topology_change_event(ioc, fw_event); |
10762 | break; |
10763 | case MPI2_EVENT_SAS_DEVICE_STATUS_CHANGE: |
10764 | if (ioc->logging_level & MPT_DEBUG_EVENT_WORK_TASK) |
10765 | _scsih_sas_device_status_change_event_debug(ioc, |
10766 | event_data: (Mpi2EventDataSasDeviceStatusChange_t *) |
10767 | fw_event->event_data); |
10768 | break; |
10769 | case MPI2_EVENT_SAS_DISCOVERY: |
10770 | _scsih_sas_discovery_event(ioc, fw_event); |
10771 | break; |
10772 | case MPI2_EVENT_SAS_DEVICE_DISCOVERY_ERROR: |
10773 | _scsih_sas_device_discovery_error_event(ioc, fw_event); |
10774 | break; |
10775 | case MPI2_EVENT_SAS_BROADCAST_PRIMITIVE: |
10776 | _scsih_sas_broadcast_primitive_event(ioc, fw_event); |
10777 | break; |
10778 | case MPI2_EVENT_SAS_ENCL_DEVICE_STATUS_CHANGE: |
10779 | _scsih_sas_enclosure_dev_status_change_event(ioc, |
10780 | fw_event); |
10781 | break; |
10782 | case MPI2_EVENT_IR_CONFIGURATION_CHANGE_LIST: |
10783 | _scsih_sas_ir_config_change_event(ioc, fw_event); |
10784 | break; |
10785 | case MPI2_EVENT_IR_VOLUME: |
10786 | _scsih_sas_ir_volume_event(ioc, fw_event); |
10787 | break; |
10788 | case MPI2_EVENT_IR_PHYSICAL_DISK: |
10789 | _scsih_sas_ir_physical_disk_event(ioc, fw_event); |
10790 | break; |
10791 | case MPI2_EVENT_IR_OPERATION_STATUS: |
10792 | _scsih_sas_ir_operation_status_event(ioc, fw_event); |
10793 | break; |
10794 | case MPI2_EVENT_PCIE_DEVICE_STATUS_CHANGE: |
10795 | _scsih_pcie_device_status_change_event(ioc, fw_event); |
10796 | break; |
10797 | case MPI2_EVENT_PCIE_ENUMERATION: |
10798 | _scsih_pcie_enumeration_event(ioc, fw_event); |
10799 | break; |
10800 | case MPI2_EVENT_PCIE_TOPOLOGY_CHANGE_LIST: |
10801 | _scsih_pcie_topology_change_event(ioc, fw_event); |
10802 | ioc->current_event = NULL; |
10803 | return; |
10804 | } |
10805 | out: |
10806 | fw_event_work_put(fw_work: fw_event); |
10807 | ioc->current_event = NULL; |
10808 | } |
10809 | |
10810 | /** |
10811 | * _firmware_event_work |
10812 | * @work: The fw_event_work object |
10813 | * Context: user. |
10814 | * |
10815 | * wrappers for the work thread handling firmware events |
10816 | */ |
10817 | |
10818 | static void |
10819 | _firmware_event_work(struct work_struct *work) |
10820 | { |
10821 | struct fw_event_work *fw_event = container_of(work, |
10822 | struct fw_event_work, work); |
10823 | |
10824 | _mpt3sas_fw_work(ioc: fw_event->ioc, fw_event); |
10825 | } |
10826 | |
10827 | /** |
10828 | * mpt3sas_scsih_event_callback - firmware event handler (called at ISR time) |
10829 | * @ioc: per adapter object |
10830 | * @msix_index: MSIX table index supplied by the OS |
10831 | * @reply: reply message frame(lower 32bit addr) |
10832 | * Context: interrupt. |
10833 | * |
10834 | * This function merely adds a new work task into ioc->firmware_event_thread. |
10835 | * The tasks are worked from _firmware_event_work in user context. |
10836 | * |
10837 | * Return: 1 meaning mf should be freed from _base_interrupt |
10838 | * 0 means the mf is freed from this function. |
10839 | */ |
10840 | u8 |
10841 | mpt3sas_scsih_event_callback(struct MPT3SAS_ADAPTER *ioc, u8 msix_index, |
10842 | u32 reply) |
10843 | { |
10844 | struct fw_event_work *fw_event; |
10845 | Mpi2EventNotificationReply_t *mpi_reply; |
10846 | u16 event; |
10847 | u16 sz; |
10848 | Mpi26EventDataActiveCableExcept_t *ActiveCableEventData; |
10849 | |
10850 | /* events turned off due to host reset */ |
10851 | if (ioc->pci_error_recovery) |
10852 | return 1; |
10853 | |
10854 | mpi_reply = mpt3sas_base_get_reply_virt_addr(ioc, phys_addr: reply); |
10855 | |
10856 | if (unlikely(!mpi_reply)) { |
10857 | ioc_err(ioc, "mpi_reply not valid at %s:%d/%s()!\n" , |
10858 | __FILE__, __LINE__, __func__); |
10859 | return 1; |
10860 | } |
10861 | |
10862 | event = le16_to_cpu(mpi_reply->Event); |
10863 | |
10864 | if (event != MPI2_EVENT_LOG_ENTRY_ADDED) |
10865 | mpt3sas_trigger_event(ioc, event, log_entry_qualifier: 0); |
10866 | |
10867 | switch (event) { |
10868 | /* handle these */ |
10869 | case MPI2_EVENT_SAS_BROADCAST_PRIMITIVE: |
10870 | { |
10871 | Mpi2EventDataSasBroadcastPrimitive_t *baen_data = |
10872 | (Mpi2EventDataSasBroadcastPrimitive_t *) |
10873 | mpi_reply->EventData; |
10874 | |
10875 | if (baen_data->Primitive != |
10876 | MPI2_EVENT_PRIMITIVE_ASYNCHRONOUS_EVENT) |
10877 | return 1; |
10878 | |
10879 | if (ioc->broadcast_aen_busy) { |
10880 | ioc->broadcast_aen_pending++; |
10881 | return 1; |
10882 | } else |
10883 | ioc->broadcast_aen_busy = 1; |
10884 | break; |
10885 | } |
10886 | |
10887 | case MPI2_EVENT_SAS_TOPOLOGY_CHANGE_LIST: |
10888 | _scsih_check_topo_delete_events(ioc, |
10889 | event_data: (Mpi2EventDataSasTopologyChangeList_t *) |
10890 | mpi_reply->EventData); |
10891 | /* |
10892 | * No need to add the topology change list |
10893 | * event to fw event work queue when |
10894 | * diag reset is going on. Since during diag |
10895 | * reset driver scan the devices by reading |
10896 | * sas device page0's not by processing the |
10897 | * events. |
10898 | */ |
10899 | if (ioc->shost_recovery) |
10900 | return 1; |
10901 | break; |
10902 | case MPI2_EVENT_PCIE_TOPOLOGY_CHANGE_LIST: |
10903 | _scsih_check_pcie_topo_remove_events(ioc, |
10904 | event_data: (Mpi26EventDataPCIeTopologyChangeList_t *) |
10905 | mpi_reply->EventData); |
10906 | if (ioc->shost_recovery) |
10907 | return 1; |
10908 | break; |
10909 | case MPI2_EVENT_IR_CONFIGURATION_CHANGE_LIST: |
10910 | _scsih_check_ir_config_unhide_events(ioc, |
10911 | event_data: (Mpi2EventDataIrConfigChangeList_t *) |
10912 | mpi_reply->EventData); |
10913 | break; |
10914 | case MPI2_EVENT_IR_VOLUME: |
10915 | _scsih_check_volume_delete_events(ioc, |
10916 | event_data: (Mpi2EventDataIrVolume_t *) |
10917 | mpi_reply->EventData); |
10918 | break; |
10919 | case MPI2_EVENT_LOG_ENTRY_ADDED: |
10920 | { |
10921 | Mpi2EventDataLogEntryAdded_t *log_entry; |
10922 | u32 log_code; |
10923 | |
10924 | if (!ioc->is_warpdrive) |
10925 | break; |
10926 | |
10927 | log_entry = (Mpi2EventDataLogEntryAdded_t *) |
10928 | mpi_reply->EventData; |
10929 | log_code = le32_to_cpu(*(__le32 *)log_entry->LogData); |
10930 | |
10931 | if (le16_to_cpu(log_entry->LogEntryQualifier) |
10932 | != MPT2_WARPDRIVE_LOGENTRY) |
10933 | break; |
10934 | |
10935 | switch (log_code) { |
10936 | case MPT2_WARPDRIVE_LC_SSDT: |
10937 | ioc_warn(ioc, "WarpDrive Warning: IO Throttling has occurred in the WarpDrive subsystem. Check WarpDrive documentation for additional details.\n" ); |
10938 | break; |
10939 | case MPT2_WARPDRIVE_LC_SSDLW: |
10940 | ioc_warn(ioc, "WarpDrive Warning: Program/Erase Cycles for the WarpDrive subsystem in degraded range. Check WarpDrive documentation for additional details.\n" ); |
10941 | break; |
10942 | case MPT2_WARPDRIVE_LC_SSDLF: |
10943 | ioc_err(ioc, "WarpDrive Fatal Error: There are no Program/Erase Cycles for the WarpDrive subsystem. The storage device will be in read-only mode. Check WarpDrive documentation for additional details.\n" ); |
10944 | break; |
10945 | case MPT2_WARPDRIVE_LC_BRMF: |
10946 | ioc_err(ioc, "WarpDrive Fatal Error: The Backup Rail Monitor has failed on the WarpDrive subsystem. Check WarpDrive documentation for additional details.\n" ); |
10947 | break; |
10948 | } |
10949 | |
10950 | break; |
10951 | } |
10952 | case MPI2_EVENT_SAS_DEVICE_STATUS_CHANGE: |
10953 | _scsih_sas_device_status_change_event(ioc, |
10954 | event_data: (Mpi2EventDataSasDeviceStatusChange_t *) |
10955 | mpi_reply->EventData); |
10956 | break; |
10957 | case MPI2_EVENT_IR_OPERATION_STATUS: |
10958 | case MPI2_EVENT_SAS_DISCOVERY: |
10959 | case MPI2_EVENT_SAS_DEVICE_DISCOVERY_ERROR: |
10960 | case MPI2_EVENT_SAS_ENCL_DEVICE_STATUS_CHANGE: |
10961 | case MPI2_EVENT_IR_PHYSICAL_DISK: |
10962 | case MPI2_EVENT_PCIE_ENUMERATION: |
10963 | case MPI2_EVENT_PCIE_DEVICE_STATUS_CHANGE: |
10964 | break; |
10965 | |
10966 | case MPI2_EVENT_TEMP_THRESHOLD: |
10967 | _scsih_temp_threshold_events(ioc, |
10968 | event_data: (Mpi2EventDataTemperature_t *) |
10969 | mpi_reply->EventData); |
10970 | break; |
10971 | case MPI2_EVENT_ACTIVE_CABLE_EXCEPTION: |
10972 | ActiveCableEventData = |
10973 | (Mpi26EventDataActiveCableExcept_t *) mpi_reply->EventData; |
10974 | switch (ActiveCableEventData->ReasonCode) { |
10975 | case MPI26_EVENT_ACTIVE_CABLE_INSUFFICIENT_POWER: |
10976 | ioc_notice(ioc, "Currently an active cable with ReceptacleID %d\n" , |
10977 | ActiveCableEventData->ReceptacleID); |
10978 | pr_notice("cannot be powered and devices connected\n" ); |
10979 | pr_notice("to this active cable will not be seen\n" ); |
10980 | pr_notice("This active cable requires %d mW of power\n" , |
10981 | le32_to_cpu( |
10982 | ActiveCableEventData->ActiveCablePowerRequirement)); |
10983 | break; |
10984 | |
10985 | case MPI26_EVENT_ACTIVE_CABLE_DEGRADED: |
10986 | ioc_notice(ioc, "Currently a cable with ReceptacleID %d\n" , |
10987 | ActiveCableEventData->ReceptacleID); |
10988 | pr_notice( |
10989 | "is not running at optimal speed(12 Gb/s rate)\n" ); |
10990 | break; |
10991 | } |
10992 | |
10993 | break; |
10994 | |
10995 | default: /* ignore the rest */ |
10996 | return 1; |
10997 | } |
10998 | |
10999 | sz = le16_to_cpu(mpi_reply->EventDataLength) * 4; |
11000 | fw_event = alloc_fw_event_work(len: sz); |
11001 | if (!fw_event) { |
11002 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
11003 | __FILE__, __LINE__, __func__); |
11004 | return 1; |
11005 | } |
11006 | |
11007 | memcpy(fw_event->event_data, mpi_reply->EventData, sz); |
11008 | fw_event->ioc = ioc; |
11009 | fw_event->VF_ID = mpi_reply->VF_ID; |
11010 | fw_event->VP_ID = mpi_reply->VP_ID; |
11011 | fw_event->event = event; |
11012 | _scsih_fw_event_add(ioc, fw_event); |
11013 | fw_event_work_put(fw_work: fw_event); |
11014 | return 1; |
11015 | } |
11016 | |
11017 | /** |
11018 | * _scsih_expander_node_remove - removing expander device from list. |
11019 | * @ioc: per adapter object |
11020 | * @sas_expander: the sas_device object |
11021 | * |
11022 | * Removing object and freeing associated memory from the |
11023 | * ioc->sas_expander_list. |
11024 | */ |
11025 | static void |
11026 | _scsih_expander_node_remove(struct MPT3SAS_ADAPTER *ioc, |
11027 | struct _sas_node *sas_expander) |
11028 | { |
11029 | struct _sas_port *mpt3sas_port, *next; |
11030 | unsigned long flags; |
11031 | int port_id; |
11032 | |
11033 | /* remove sibling ports attached to this expander */ |
11034 | list_for_each_entry_safe(mpt3sas_port, next, |
11035 | &sas_expander->sas_port_list, port_list) { |
11036 | if (ioc->shost_recovery) |
11037 | return; |
11038 | if (mpt3sas_port->remote_identify.device_type == |
11039 | SAS_END_DEVICE) |
11040 | mpt3sas_device_remove_by_sas_address(ioc, |
11041 | sas_address: mpt3sas_port->remote_identify.sas_address, |
11042 | port: mpt3sas_port->hba_port); |
11043 | else if (mpt3sas_port->remote_identify.device_type == |
11044 | SAS_EDGE_EXPANDER_DEVICE || |
11045 | mpt3sas_port->remote_identify.device_type == |
11046 | SAS_FANOUT_EXPANDER_DEVICE) |
11047 | mpt3sas_expander_remove(ioc, |
11048 | sas_address: mpt3sas_port->remote_identify.sas_address, |
11049 | port: mpt3sas_port->hba_port); |
11050 | } |
11051 | |
11052 | port_id = sas_expander->port->port_id; |
11053 | |
11054 | mpt3sas_transport_port_remove(ioc, sas_address: sas_expander->sas_address, |
11055 | sas_address_parent: sas_expander->sas_address_parent, port: sas_expander->port); |
11056 | |
11057 | ioc_info(ioc, |
11058 | "expander_remove: handle(0x%04x), sas_addr(0x%016llx), port:%d\n" , |
11059 | sas_expander->handle, (unsigned long long) |
11060 | sas_expander->sas_address, |
11061 | port_id); |
11062 | |
11063 | spin_lock_irqsave(&ioc->sas_node_lock, flags); |
11064 | list_del(entry: &sas_expander->list); |
11065 | spin_unlock_irqrestore(lock: &ioc->sas_node_lock, flags); |
11066 | |
11067 | kfree(objp: sas_expander->phy); |
11068 | kfree(objp: sas_expander); |
11069 | } |
11070 | |
11071 | /** |
11072 | * _scsih_nvme_shutdown - NVMe shutdown notification |
11073 | * @ioc: per adapter object |
11074 | * |
11075 | * Sending IoUnitControl request with shutdown operation code to alert IOC that |
11076 | * the host system is shutting down so that IOC can issue NVMe shutdown to |
11077 | * NVMe drives attached to it. |
11078 | */ |
11079 | static void |
11080 | _scsih_nvme_shutdown(struct MPT3SAS_ADAPTER *ioc) |
11081 | { |
11082 | Mpi26IoUnitControlRequest_t *mpi_request; |
11083 | Mpi26IoUnitControlReply_t *mpi_reply; |
11084 | u16 smid; |
11085 | |
11086 | /* are there any NVMe devices ? */ |
11087 | if (list_empty(head: &ioc->pcie_device_list)) |
11088 | return; |
11089 | |
11090 | mutex_lock(&ioc->scsih_cmds.mutex); |
11091 | |
11092 | if (ioc->scsih_cmds.status != MPT3_CMD_NOT_USED) { |
11093 | ioc_err(ioc, "%s: scsih_cmd in use\n" , __func__); |
11094 | goto out; |
11095 | } |
11096 | |
11097 | ioc->scsih_cmds.status = MPT3_CMD_PENDING; |
11098 | |
11099 | smid = mpt3sas_base_get_smid(ioc, cb_idx: ioc->scsih_cb_idx); |
11100 | if (!smid) { |
11101 | ioc_err(ioc, |
11102 | "%s: failed obtaining a smid\n" , __func__); |
11103 | ioc->scsih_cmds.status = MPT3_CMD_NOT_USED; |
11104 | goto out; |
11105 | } |
11106 | |
11107 | mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); |
11108 | ioc->scsih_cmds.smid = smid; |
11109 | memset(mpi_request, 0, sizeof(Mpi26IoUnitControlRequest_t)); |
11110 | mpi_request->Function = MPI2_FUNCTION_IO_UNIT_CONTROL; |
11111 | mpi_request->Operation = MPI26_CTRL_OP_SHUTDOWN; |
11112 | |
11113 | init_completion(x: &ioc->scsih_cmds.done); |
11114 | ioc->put_smid_default(ioc, smid); |
11115 | /* Wait for max_shutdown_latency seconds */ |
11116 | ioc_info(ioc, |
11117 | "Io Unit Control shutdown (sending), Shutdown latency %d sec\n" , |
11118 | ioc->max_shutdown_latency); |
11119 | wait_for_completion_timeout(x: &ioc->scsih_cmds.done, |
11120 | timeout: ioc->max_shutdown_latency*HZ); |
11121 | |
11122 | if (!(ioc->scsih_cmds.status & MPT3_CMD_COMPLETE)) { |
11123 | ioc_err(ioc, "%s: timeout\n" , __func__); |
11124 | goto out; |
11125 | } |
11126 | |
11127 | if (ioc->scsih_cmds.status & MPT3_CMD_REPLY_VALID) { |
11128 | mpi_reply = ioc->scsih_cmds.reply; |
11129 | ioc_info(ioc, "Io Unit Control shutdown (complete):" |
11130 | "ioc_status(0x%04x), loginfo(0x%08x)\n" , |
11131 | le16_to_cpu(mpi_reply->IOCStatus), |
11132 | le32_to_cpu(mpi_reply->IOCLogInfo)); |
11133 | } |
11134 | out: |
11135 | ioc->scsih_cmds.status = MPT3_CMD_NOT_USED; |
11136 | mutex_unlock(lock: &ioc->scsih_cmds.mutex); |
11137 | } |
11138 | |
11139 | |
11140 | /** |
11141 | * _scsih_ir_shutdown - IR shutdown notification |
11142 | * @ioc: per adapter object |
11143 | * |
11144 | * Sending RAID Action to alert the Integrated RAID subsystem of the IOC that |
11145 | * the host system is shutting down. |
11146 | */ |
11147 | static void |
11148 | _scsih_ir_shutdown(struct MPT3SAS_ADAPTER *ioc) |
11149 | { |
11150 | Mpi2RaidActionRequest_t *mpi_request; |
11151 | Mpi2RaidActionReply_t *mpi_reply; |
11152 | u16 smid; |
11153 | |
11154 | /* is IR firmware build loaded ? */ |
11155 | if (!ioc->ir_firmware) |
11156 | return; |
11157 | |
11158 | /* are there any volumes ? */ |
11159 | if (list_empty(head: &ioc->raid_device_list)) |
11160 | return; |
11161 | |
11162 | mutex_lock(&ioc->scsih_cmds.mutex); |
11163 | |
11164 | if (ioc->scsih_cmds.status != MPT3_CMD_NOT_USED) { |
11165 | ioc_err(ioc, "%s: scsih_cmd in use\n" , __func__); |
11166 | goto out; |
11167 | } |
11168 | ioc->scsih_cmds.status = MPT3_CMD_PENDING; |
11169 | |
11170 | smid = mpt3sas_base_get_smid(ioc, cb_idx: ioc->scsih_cb_idx); |
11171 | if (!smid) { |
11172 | ioc_err(ioc, "%s: failed obtaining a smid\n" , __func__); |
11173 | ioc->scsih_cmds.status = MPT3_CMD_NOT_USED; |
11174 | goto out; |
11175 | } |
11176 | |
11177 | mpi_request = mpt3sas_base_get_msg_frame(ioc, smid); |
11178 | ioc->scsih_cmds.smid = smid; |
11179 | memset(mpi_request, 0, sizeof(Mpi2RaidActionRequest_t)); |
11180 | |
11181 | mpi_request->Function = MPI2_FUNCTION_RAID_ACTION; |
11182 | mpi_request->Action = MPI2_RAID_ACTION_SYSTEM_SHUTDOWN_INITIATED; |
11183 | |
11184 | if (!ioc->hide_ir_msg) |
11185 | ioc_info(ioc, "IR shutdown (sending)\n" ); |
11186 | init_completion(x: &ioc->scsih_cmds.done); |
11187 | ioc->put_smid_default(ioc, smid); |
11188 | wait_for_completion_timeout(x: &ioc->scsih_cmds.done, timeout: 10*HZ); |
11189 | |
11190 | if (!(ioc->scsih_cmds.status & MPT3_CMD_COMPLETE)) { |
11191 | ioc_err(ioc, "%s: timeout\n" , __func__); |
11192 | goto out; |
11193 | } |
11194 | |
11195 | if (ioc->scsih_cmds.status & MPT3_CMD_REPLY_VALID) { |
11196 | mpi_reply = ioc->scsih_cmds.reply; |
11197 | if (!ioc->hide_ir_msg) |
11198 | ioc_info(ioc, "IR shutdown (complete): ioc_status(0x%04x), loginfo(0x%08x)\n" , |
11199 | le16_to_cpu(mpi_reply->IOCStatus), |
11200 | le32_to_cpu(mpi_reply->IOCLogInfo)); |
11201 | } |
11202 | |
11203 | out: |
11204 | ioc->scsih_cmds.status = MPT3_CMD_NOT_USED; |
11205 | mutex_unlock(lock: &ioc->scsih_cmds.mutex); |
11206 | } |
11207 | |
11208 | /** |
11209 | * _scsih_get_shost_and_ioc - get shost and ioc |
11210 | * and verify whether they are NULL or not |
11211 | * @pdev: PCI device struct |
11212 | * @shost: address of scsi host pointer |
11213 | * @ioc: address of HBA adapter pointer |
11214 | * |
11215 | * Return zero if *shost and *ioc are not NULL otherwise return error number. |
11216 | */ |
11217 | static int |
11218 | _scsih_get_shost_and_ioc(struct pci_dev *pdev, |
11219 | struct Scsi_Host **shost, struct MPT3SAS_ADAPTER **ioc) |
11220 | { |
11221 | *shost = pci_get_drvdata(pdev); |
11222 | if (*shost == NULL) { |
11223 | dev_err(&pdev->dev, "pdev's driver data is null\n" ); |
11224 | return -ENXIO; |
11225 | } |
11226 | |
11227 | *ioc = shost_priv(shost: *shost); |
11228 | if (*ioc == NULL) { |
11229 | dev_err(&pdev->dev, "shost's private data is null\n" ); |
11230 | return -ENXIO; |
11231 | } |
11232 | |
11233 | return 0; |
11234 | } |
11235 | |
11236 | /** |
11237 | * scsih_remove - detach and remove add host |
11238 | * @pdev: PCI device struct |
11239 | * |
11240 | * Routine called when unloading the driver. |
11241 | */ |
11242 | static void scsih_remove(struct pci_dev *pdev) |
11243 | { |
11244 | struct Scsi_Host *shost; |
11245 | struct MPT3SAS_ADAPTER *ioc; |
11246 | struct _sas_port *mpt3sas_port, *next_port; |
11247 | struct _raid_device *raid_device, *next; |
11248 | struct MPT3SAS_TARGET *sas_target_priv_data; |
11249 | struct _pcie_device *pcie_device, *pcienext; |
11250 | struct workqueue_struct *wq; |
11251 | unsigned long flags; |
11252 | Mpi2ConfigReply_t mpi_reply; |
11253 | struct hba_port *port, *port_next; |
11254 | |
11255 | if (_scsih_get_shost_and_ioc(pdev, shost: &shost, ioc: &ioc)) |
11256 | return; |
11257 | |
11258 | ioc->remove_host = 1; |
11259 | |
11260 | if (!pci_device_is_present(pdev)) { |
11261 | mpt3sas_base_pause_mq_polling(ioc); |
11262 | _scsih_flush_running_cmds(ioc); |
11263 | } |
11264 | |
11265 | _scsih_fw_event_cleanup_queue(ioc); |
11266 | |
11267 | spin_lock_irqsave(&ioc->fw_event_lock, flags); |
11268 | wq = ioc->firmware_event_thread; |
11269 | ioc->firmware_event_thread = NULL; |
11270 | spin_unlock_irqrestore(lock: &ioc->fw_event_lock, flags); |
11271 | if (wq) |
11272 | destroy_workqueue(wq); |
11273 | /* |
11274 | * Copy back the unmodified ioc page1. so that on next driver load, |
11275 | * current modified changes on ioc page1 won't take effect. |
11276 | */ |
11277 | if (ioc->is_aero_ioc) |
11278 | mpt3sas_config_set_ioc_pg1(ioc, mpi_reply: &mpi_reply, |
11279 | config_page: &ioc->ioc_pg1_copy); |
11280 | /* release all the volumes */ |
11281 | _scsih_ir_shutdown(ioc); |
11282 | mpt3sas_destroy_debugfs(ioc); |
11283 | sas_remove_host(shost); |
11284 | list_for_each_entry_safe(raid_device, next, &ioc->raid_device_list, |
11285 | list) { |
11286 | if (raid_device->starget) { |
11287 | sas_target_priv_data = |
11288 | raid_device->starget->hostdata; |
11289 | sas_target_priv_data->deleted = 1; |
11290 | scsi_remove_target(&raid_device->starget->dev); |
11291 | } |
11292 | ioc_info(ioc, "removing handle(0x%04x), wwid(0x%016llx)\n" , |
11293 | raid_device->handle, (u64)raid_device->wwid); |
11294 | _scsih_raid_device_remove(ioc, raid_device); |
11295 | } |
11296 | list_for_each_entry_safe(pcie_device, pcienext, &ioc->pcie_device_list, |
11297 | list) { |
11298 | _scsih_pcie_device_remove_from_sml(ioc, pcie_device); |
11299 | list_del_init(entry: &pcie_device->list); |
11300 | pcie_device_put(p: pcie_device); |
11301 | } |
11302 | |
11303 | /* free ports attached to the sas_host */ |
11304 | list_for_each_entry_safe(mpt3sas_port, next_port, |
11305 | &ioc->sas_hba.sas_port_list, port_list) { |
11306 | if (mpt3sas_port->remote_identify.device_type == |
11307 | SAS_END_DEVICE) |
11308 | mpt3sas_device_remove_by_sas_address(ioc, |
11309 | sas_address: mpt3sas_port->remote_identify.sas_address, |
11310 | port: mpt3sas_port->hba_port); |
11311 | else if (mpt3sas_port->remote_identify.device_type == |
11312 | SAS_EDGE_EXPANDER_DEVICE || |
11313 | mpt3sas_port->remote_identify.device_type == |
11314 | SAS_FANOUT_EXPANDER_DEVICE) |
11315 | mpt3sas_expander_remove(ioc, |
11316 | sas_address: mpt3sas_port->remote_identify.sas_address, |
11317 | port: mpt3sas_port->hba_port); |
11318 | } |
11319 | |
11320 | list_for_each_entry_safe(port, port_next, |
11321 | &ioc->port_table_list, list) { |
11322 | list_del(entry: &port->list); |
11323 | kfree(objp: port); |
11324 | } |
11325 | |
11326 | /* free phys attached to the sas_host */ |
11327 | if (ioc->sas_hba.num_phys) { |
11328 | kfree(objp: ioc->sas_hba.phy); |
11329 | ioc->sas_hba.phy = NULL; |
11330 | ioc->sas_hba.num_phys = 0; |
11331 | } |
11332 | |
11333 | mpt3sas_base_detach(ioc); |
11334 | mpt3sas_ctl_release(ioc); |
11335 | spin_lock(lock: &gioc_lock); |
11336 | list_del(entry: &ioc->list); |
11337 | spin_unlock(lock: &gioc_lock); |
11338 | scsi_host_put(t: shost); |
11339 | } |
11340 | |
11341 | /** |
11342 | * scsih_shutdown - routine call during system shutdown |
11343 | * @pdev: PCI device struct |
11344 | */ |
11345 | static void |
11346 | scsih_shutdown(struct pci_dev *pdev) |
11347 | { |
11348 | struct Scsi_Host *shost; |
11349 | struct MPT3SAS_ADAPTER *ioc; |
11350 | struct workqueue_struct *wq; |
11351 | unsigned long flags; |
11352 | Mpi2ConfigReply_t mpi_reply; |
11353 | |
11354 | if (_scsih_get_shost_and_ioc(pdev, shost: &shost, ioc: &ioc)) |
11355 | return; |
11356 | |
11357 | ioc->remove_host = 1; |
11358 | |
11359 | if (!pci_device_is_present(pdev)) { |
11360 | mpt3sas_base_pause_mq_polling(ioc); |
11361 | _scsih_flush_running_cmds(ioc); |
11362 | } |
11363 | |
11364 | _scsih_fw_event_cleanup_queue(ioc); |
11365 | |
11366 | spin_lock_irqsave(&ioc->fw_event_lock, flags); |
11367 | wq = ioc->firmware_event_thread; |
11368 | ioc->firmware_event_thread = NULL; |
11369 | spin_unlock_irqrestore(lock: &ioc->fw_event_lock, flags); |
11370 | if (wq) |
11371 | destroy_workqueue(wq); |
11372 | /* |
11373 | * Copy back the unmodified ioc page1 so that on next driver load, |
11374 | * current modified changes on ioc page1 won't take effect. |
11375 | */ |
11376 | if (ioc->is_aero_ioc) |
11377 | mpt3sas_config_set_ioc_pg1(ioc, mpi_reply: &mpi_reply, |
11378 | config_page: &ioc->ioc_pg1_copy); |
11379 | |
11380 | _scsih_ir_shutdown(ioc); |
11381 | _scsih_nvme_shutdown(ioc); |
11382 | mpt3sas_base_mask_interrupts(ioc); |
11383 | mpt3sas_base_stop_watchdog(ioc); |
11384 | ioc->shost_recovery = 1; |
11385 | mpt3sas_base_make_ioc_ready(ioc, type: SOFT_RESET); |
11386 | ioc->shost_recovery = 0; |
11387 | mpt3sas_base_free_irq(ioc); |
11388 | mpt3sas_base_disable_msix(ioc); |
11389 | } |
11390 | |
11391 | |
11392 | /** |
11393 | * _scsih_probe_boot_devices - reports 1st device |
11394 | * @ioc: per adapter object |
11395 | * |
11396 | * If specified in bios page 2, this routine reports the 1st |
11397 | * device scsi-ml or sas transport for persistent boot device |
11398 | * purposes. Please refer to function _scsih_determine_boot_device() |
11399 | */ |
11400 | static void |
11401 | _scsih_probe_boot_devices(struct MPT3SAS_ADAPTER *ioc) |
11402 | { |
11403 | u32 channel; |
11404 | void *device; |
11405 | struct _sas_device *sas_device; |
11406 | struct _raid_device *raid_device; |
11407 | struct _pcie_device *pcie_device; |
11408 | u16 handle; |
11409 | u64 sas_address_parent; |
11410 | u64 sas_address; |
11411 | unsigned long flags; |
11412 | int rc; |
11413 | int tid; |
11414 | struct hba_port *port; |
11415 | |
11416 | /* no Bios, return immediately */ |
11417 | if (!ioc->bios_pg3.BiosVersion) |
11418 | return; |
11419 | |
11420 | device = NULL; |
11421 | if (ioc->req_boot_device.device) { |
11422 | device = ioc->req_boot_device.device; |
11423 | channel = ioc->req_boot_device.channel; |
11424 | } else if (ioc->req_alt_boot_device.device) { |
11425 | device = ioc->req_alt_boot_device.device; |
11426 | channel = ioc->req_alt_boot_device.channel; |
11427 | } else if (ioc->current_boot_device.device) { |
11428 | device = ioc->current_boot_device.device; |
11429 | channel = ioc->current_boot_device.channel; |
11430 | } |
11431 | |
11432 | if (!device) |
11433 | return; |
11434 | |
11435 | if (channel == RAID_CHANNEL) { |
11436 | raid_device = device; |
11437 | /* |
11438 | * If this boot vd is already registered with SML then |
11439 | * no need to register it again as part of device scanning |
11440 | * after diag reset during driver load operation. |
11441 | */ |
11442 | if (raid_device->starget) |
11443 | return; |
11444 | rc = scsi_add_device(host: ioc->shost, RAID_CHANNEL, |
11445 | target: raid_device->id, lun: 0); |
11446 | if (rc) |
11447 | _scsih_raid_device_remove(ioc, raid_device); |
11448 | } else if (channel == PCIE_CHANNEL) { |
11449 | pcie_device = device; |
11450 | /* |
11451 | * If this boot NVMe device is already registered with SML then |
11452 | * no need to register it again as part of device scanning |
11453 | * after diag reset during driver load operation. |
11454 | */ |
11455 | if (pcie_device->starget) |
11456 | return; |
11457 | spin_lock_irqsave(&ioc->pcie_device_lock, flags); |
11458 | tid = pcie_device->id; |
11459 | list_move_tail(list: &pcie_device->list, head: &ioc->pcie_device_list); |
11460 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
11461 | rc = scsi_add_device(host: ioc->shost, PCIE_CHANNEL, target: tid, lun: 0); |
11462 | if (rc) |
11463 | _scsih_pcie_device_remove(ioc, pcie_device); |
11464 | } else { |
11465 | sas_device = device; |
11466 | /* |
11467 | * If this boot sas/sata device is already registered with SML |
11468 | * then no need to register it again as part of device scanning |
11469 | * after diag reset during driver load operation. |
11470 | */ |
11471 | if (sas_device->starget) |
11472 | return; |
11473 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
11474 | handle = sas_device->handle; |
11475 | sas_address_parent = sas_device->sas_address_parent; |
11476 | sas_address = sas_device->sas_address; |
11477 | port = sas_device->port; |
11478 | list_move_tail(list: &sas_device->list, head: &ioc->sas_device_list); |
11479 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
11480 | |
11481 | if (ioc->hide_drives) |
11482 | return; |
11483 | |
11484 | if (!port) |
11485 | return; |
11486 | |
11487 | if (!mpt3sas_transport_port_add(ioc, handle, |
11488 | sas_address: sas_address_parent, port)) { |
11489 | _scsih_sas_device_remove(ioc, sas_device); |
11490 | } else if (!sas_device->starget) { |
11491 | if (!ioc->is_driver_loading) { |
11492 | mpt3sas_transport_port_remove(ioc, |
11493 | sas_address, |
11494 | sas_address_parent, port); |
11495 | _scsih_sas_device_remove(ioc, sas_device); |
11496 | } |
11497 | } |
11498 | } |
11499 | } |
11500 | |
11501 | /** |
11502 | * _scsih_probe_raid - reporting raid volumes to scsi-ml |
11503 | * @ioc: per adapter object |
11504 | * |
11505 | * Called during initial loading of the driver. |
11506 | */ |
11507 | static void |
11508 | _scsih_probe_raid(struct MPT3SAS_ADAPTER *ioc) |
11509 | { |
11510 | struct _raid_device *raid_device, *raid_next; |
11511 | int rc; |
11512 | |
11513 | list_for_each_entry_safe(raid_device, raid_next, |
11514 | &ioc->raid_device_list, list) { |
11515 | if (raid_device->starget) |
11516 | continue; |
11517 | rc = scsi_add_device(host: ioc->shost, RAID_CHANNEL, |
11518 | target: raid_device->id, lun: 0); |
11519 | if (rc) |
11520 | _scsih_raid_device_remove(ioc, raid_device); |
11521 | } |
11522 | } |
11523 | |
11524 | static struct _sas_device *get_next_sas_device(struct MPT3SAS_ADAPTER *ioc) |
11525 | { |
11526 | struct _sas_device *sas_device = NULL; |
11527 | unsigned long flags; |
11528 | |
11529 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
11530 | if (!list_empty(head: &ioc->sas_device_init_list)) { |
11531 | sas_device = list_first_entry(&ioc->sas_device_init_list, |
11532 | struct _sas_device, list); |
11533 | sas_device_get(s: sas_device); |
11534 | } |
11535 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
11536 | |
11537 | return sas_device; |
11538 | } |
11539 | |
11540 | static void sas_device_make_active(struct MPT3SAS_ADAPTER *ioc, |
11541 | struct _sas_device *sas_device) |
11542 | { |
11543 | unsigned long flags; |
11544 | |
11545 | spin_lock_irqsave(&ioc->sas_device_lock, flags); |
11546 | |
11547 | /* |
11548 | * Since we dropped the lock during the call to port_add(), we need to |
11549 | * be careful here that somebody else didn't move or delete this item |
11550 | * while we were busy with other things. |
11551 | * |
11552 | * If it was on the list, we need a put() for the reference the list |
11553 | * had. Either way, we need a get() for the destination list. |
11554 | */ |
11555 | if (!list_empty(head: &sas_device->list)) { |
11556 | list_del_init(entry: &sas_device->list); |
11557 | sas_device_put(s: sas_device); |
11558 | } |
11559 | |
11560 | sas_device_get(s: sas_device); |
11561 | list_add_tail(new: &sas_device->list, head: &ioc->sas_device_list); |
11562 | |
11563 | spin_unlock_irqrestore(lock: &ioc->sas_device_lock, flags); |
11564 | } |
11565 | |
11566 | /** |
11567 | * _scsih_probe_sas - reporting sas devices to sas transport |
11568 | * @ioc: per adapter object |
11569 | * |
11570 | * Called during initial loading of the driver. |
11571 | */ |
11572 | static void |
11573 | _scsih_probe_sas(struct MPT3SAS_ADAPTER *ioc) |
11574 | { |
11575 | struct _sas_device *sas_device; |
11576 | |
11577 | if (ioc->hide_drives) |
11578 | return; |
11579 | |
11580 | while ((sas_device = get_next_sas_device(ioc))) { |
11581 | if (!mpt3sas_transport_port_add(ioc, handle: sas_device->handle, |
11582 | sas_address: sas_device->sas_address_parent, port: sas_device->port)) { |
11583 | _scsih_sas_device_remove(ioc, sas_device); |
11584 | sas_device_put(s: sas_device); |
11585 | continue; |
11586 | } else if (!sas_device->starget) { |
11587 | /* |
11588 | * When asyn scanning is enabled, its not possible to |
11589 | * remove devices while scanning is turned on due to an |
11590 | * oops in scsi_sysfs_add_sdev()->add_device()-> |
11591 | * sysfs_addrm_start() |
11592 | */ |
11593 | if (!ioc->is_driver_loading) { |
11594 | mpt3sas_transport_port_remove(ioc, |
11595 | sas_address: sas_device->sas_address, |
11596 | sas_address_parent: sas_device->sas_address_parent, |
11597 | port: sas_device->port); |
11598 | _scsih_sas_device_remove(ioc, sas_device); |
11599 | sas_device_put(s: sas_device); |
11600 | continue; |
11601 | } |
11602 | } |
11603 | sas_device_make_active(ioc, sas_device); |
11604 | sas_device_put(s: sas_device); |
11605 | } |
11606 | } |
11607 | |
11608 | /** |
11609 | * get_next_pcie_device - Get the next pcie device |
11610 | * @ioc: per adapter object |
11611 | * |
11612 | * Get the next pcie device from pcie_device_init_list list. |
11613 | * |
11614 | * Return: pcie device structure if pcie_device_init_list list is not empty |
11615 | * otherwise returns NULL |
11616 | */ |
11617 | static struct _pcie_device *get_next_pcie_device(struct MPT3SAS_ADAPTER *ioc) |
11618 | { |
11619 | struct _pcie_device *pcie_device = NULL; |
11620 | unsigned long flags; |
11621 | |
11622 | spin_lock_irqsave(&ioc->pcie_device_lock, flags); |
11623 | if (!list_empty(head: &ioc->pcie_device_init_list)) { |
11624 | pcie_device = list_first_entry(&ioc->pcie_device_init_list, |
11625 | struct _pcie_device, list); |
11626 | pcie_device_get(p: pcie_device); |
11627 | } |
11628 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
11629 | |
11630 | return pcie_device; |
11631 | } |
11632 | |
11633 | /** |
11634 | * pcie_device_make_active - Add pcie device to pcie_device_list list |
11635 | * @ioc: per adapter object |
11636 | * @pcie_device: pcie device object |
11637 | * |
11638 | * Add the pcie device which has registered with SCSI Transport Later to |
11639 | * pcie_device_list list |
11640 | */ |
11641 | static void pcie_device_make_active(struct MPT3SAS_ADAPTER *ioc, |
11642 | struct _pcie_device *pcie_device) |
11643 | { |
11644 | unsigned long flags; |
11645 | |
11646 | spin_lock_irqsave(&ioc->pcie_device_lock, flags); |
11647 | |
11648 | if (!list_empty(head: &pcie_device->list)) { |
11649 | list_del_init(entry: &pcie_device->list); |
11650 | pcie_device_put(p: pcie_device); |
11651 | } |
11652 | pcie_device_get(p: pcie_device); |
11653 | list_add_tail(new: &pcie_device->list, head: &ioc->pcie_device_list); |
11654 | |
11655 | spin_unlock_irqrestore(lock: &ioc->pcie_device_lock, flags); |
11656 | } |
11657 | |
11658 | /** |
11659 | * _scsih_probe_pcie - reporting PCIe devices to scsi-ml |
11660 | * @ioc: per adapter object |
11661 | * |
11662 | * Called during initial loading of the driver. |
11663 | */ |
11664 | static void |
11665 | _scsih_probe_pcie(struct MPT3SAS_ADAPTER *ioc) |
11666 | { |
11667 | struct _pcie_device *pcie_device; |
11668 | int rc; |
11669 | |
11670 | /* PCIe Device List */ |
11671 | while ((pcie_device = get_next_pcie_device(ioc))) { |
11672 | if (pcie_device->starget) { |
11673 | pcie_device_put(p: pcie_device); |
11674 | continue; |
11675 | } |
11676 | if (pcie_device->access_status == |
11677 | MPI26_PCIEDEV0_ASTATUS_DEVICE_BLOCKED) { |
11678 | pcie_device_make_active(ioc, pcie_device); |
11679 | pcie_device_put(p: pcie_device); |
11680 | continue; |
11681 | } |
11682 | rc = scsi_add_device(host: ioc->shost, PCIE_CHANNEL, |
11683 | target: pcie_device->id, lun: 0); |
11684 | if (rc) { |
11685 | _scsih_pcie_device_remove(ioc, pcie_device); |
11686 | pcie_device_put(p: pcie_device); |
11687 | continue; |
11688 | } else if (!pcie_device->starget) { |
11689 | /* |
11690 | * When async scanning is enabled, its not possible to |
11691 | * remove devices while scanning is turned on due to an |
11692 | * oops in scsi_sysfs_add_sdev()->add_device()-> |
11693 | * sysfs_addrm_start() |
11694 | */ |
11695 | if (!ioc->is_driver_loading) { |
11696 | /* TODO-- Need to find out whether this condition will |
11697 | * occur or not |
11698 | */ |
11699 | _scsih_pcie_device_remove(ioc, pcie_device); |
11700 | pcie_device_put(p: pcie_device); |
11701 | continue; |
11702 | } |
11703 | } |
11704 | pcie_device_make_active(ioc, pcie_device); |
11705 | pcie_device_put(p: pcie_device); |
11706 | } |
11707 | } |
11708 | |
11709 | /** |
11710 | * _scsih_probe_devices - probing for devices |
11711 | * @ioc: per adapter object |
11712 | * |
11713 | * Called during initial loading of the driver. |
11714 | */ |
11715 | static void |
11716 | _scsih_probe_devices(struct MPT3SAS_ADAPTER *ioc) |
11717 | { |
11718 | u16 volume_mapping_flags; |
11719 | |
11720 | if (!(ioc->facts.ProtocolFlags & MPI2_IOCFACTS_PROTOCOL_SCSI_INITIATOR)) |
11721 | return; /* return when IOC doesn't support initiator mode */ |
11722 | |
11723 | _scsih_probe_boot_devices(ioc); |
11724 | |
11725 | if (ioc->ir_firmware) { |
11726 | volume_mapping_flags = |
11727 | le16_to_cpu(ioc->ioc_pg8.IRVolumeMappingFlags) & |
11728 | MPI2_IOCPAGE8_IRFLAGS_MASK_VOLUME_MAPPING_MODE; |
11729 | if (volume_mapping_flags == |
11730 | MPI2_IOCPAGE8_IRFLAGS_LOW_VOLUME_MAPPING) { |
11731 | _scsih_probe_raid(ioc); |
11732 | _scsih_probe_sas(ioc); |
11733 | } else { |
11734 | _scsih_probe_sas(ioc); |
11735 | _scsih_probe_raid(ioc); |
11736 | } |
11737 | } else { |
11738 | _scsih_probe_sas(ioc); |
11739 | _scsih_probe_pcie(ioc); |
11740 | } |
11741 | } |
11742 | |
11743 | /** |
11744 | * scsih_scan_start - scsi lld callback for .scan_start |
11745 | * @shost: SCSI host pointer |
11746 | * |
11747 | * The shost has the ability to discover targets on its own instead |
11748 | * of scanning the entire bus. In our implemention, we will kick off |
11749 | * firmware discovery. |
11750 | */ |
11751 | static void |
11752 | scsih_scan_start(struct Scsi_Host *shost) |
11753 | { |
11754 | struct MPT3SAS_ADAPTER *ioc = shost_priv(shost); |
11755 | int rc; |
11756 | if (diag_buffer_enable != -1 && diag_buffer_enable != 0) |
11757 | mpt3sas_enable_diag_buffer(ioc, bits_to_register: diag_buffer_enable); |
11758 | else if (ioc->manu_pg11.HostTraceBufferMaxSizeKB != 0) |
11759 | mpt3sas_enable_diag_buffer(ioc, bits_to_register: 1); |
11760 | |
11761 | if (disable_discovery > 0) |
11762 | return; |
11763 | |
11764 | ioc->start_scan = 1; |
11765 | rc = mpt3sas_port_enable(ioc); |
11766 | |
11767 | if (rc != 0) |
11768 | ioc_info(ioc, "port enable: FAILED\n" ); |
11769 | } |
11770 | |
11771 | /** |
11772 | * _scsih_complete_devices_scanning - add the devices to sml and |
11773 | * complete ioc initialization. |
11774 | * @ioc: per adapter object |
11775 | * |
11776 | * Return nothing. |
11777 | */ |
11778 | static void _scsih_complete_devices_scanning(struct MPT3SAS_ADAPTER *ioc) |
11779 | { |
11780 | |
11781 | if (ioc->wait_for_discovery_to_complete) { |
11782 | ioc->wait_for_discovery_to_complete = 0; |
11783 | _scsih_probe_devices(ioc); |
11784 | } |
11785 | |
11786 | mpt3sas_base_start_watchdog(ioc); |
11787 | ioc->is_driver_loading = 0; |
11788 | } |
11789 | |
11790 | /** |
11791 | * scsih_scan_finished - scsi lld callback for .scan_finished |
11792 | * @shost: SCSI host pointer |
11793 | * @time: elapsed time of the scan in jiffies |
11794 | * |
11795 | * This function will be called periodicallyn until it returns 1 with the |
11796 | * scsi_host and the elapsed time of the scan in jiffies. In our implemention, |
11797 | * we wait for firmware discovery to complete, then return 1. |
11798 | */ |
11799 | static int |
11800 | scsih_scan_finished(struct Scsi_Host *shost, unsigned long time) |
11801 | { |
11802 | struct MPT3SAS_ADAPTER *ioc = shost_priv(shost); |
11803 | u32 ioc_state; |
11804 | int issue_hard_reset = 0; |
11805 | |
11806 | if (disable_discovery > 0) { |
11807 | ioc->is_driver_loading = 0; |
11808 | ioc->wait_for_discovery_to_complete = 0; |
11809 | return 1; |
11810 | } |
11811 | |
11812 | if (time >= (300 * HZ)) { |
11813 | ioc->port_enable_cmds.status = MPT3_CMD_NOT_USED; |
11814 | ioc_info(ioc, "port enable: FAILED with timeout (timeout=300s)\n" ); |
11815 | ioc->is_driver_loading = 0; |
11816 | return 1; |
11817 | } |
11818 | |
11819 | if (ioc->start_scan) { |
11820 | ioc_state = mpt3sas_base_get_iocstate(ioc, cooked: 0); |
11821 | if ((ioc_state & MPI2_IOC_STATE_MASK) == MPI2_IOC_STATE_FAULT) { |
11822 | mpt3sas_print_fault_code(ioc, ioc_state & |
11823 | MPI2_DOORBELL_DATA_MASK); |
11824 | issue_hard_reset = 1; |
11825 | goto out; |
11826 | } else if ((ioc_state & MPI2_IOC_STATE_MASK) == |
11827 | MPI2_IOC_STATE_COREDUMP) { |
11828 | mpt3sas_base_coredump_info(ioc, fault_code: ioc_state & |
11829 | MPI2_DOORBELL_DATA_MASK); |
11830 | mpt3sas_base_wait_for_coredump_completion(ioc, caller: __func__); |
11831 | issue_hard_reset = 1; |
11832 | goto out; |
11833 | } |
11834 | return 0; |
11835 | } |
11836 | |
11837 | if (ioc->port_enable_cmds.status & MPT3_CMD_RESET) { |
11838 | ioc_info(ioc, |
11839 | "port enable: aborted due to diag reset\n" ); |
11840 | ioc->port_enable_cmds.status = MPT3_CMD_NOT_USED; |
11841 | goto out; |
11842 | } |
11843 | if (ioc->start_scan_failed) { |
11844 | ioc_info(ioc, "port enable: FAILED with (ioc_status=0x%08x)\n" , |
11845 | ioc->start_scan_failed); |
11846 | ioc->is_driver_loading = 0; |
11847 | ioc->wait_for_discovery_to_complete = 0; |
11848 | ioc->remove_host = 1; |
11849 | return 1; |
11850 | } |
11851 | |
11852 | ioc_info(ioc, "port enable: SUCCESS\n" ); |
11853 | ioc->port_enable_cmds.status = MPT3_CMD_NOT_USED; |
11854 | _scsih_complete_devices_scanning(ioc); |
11855 | |
11856 | out: |
11857 | if (issue_hard_reset) { |
11858 | ioc->port_enable_cmds.status = MPT3_CMD_NOT_USED; |
11859 | if (mpt3sas_base_hard_reset_handler(ioc, type: SOFT_RESET)) |
11860 | ioc->is_driver_loading = 0; |
11861 | } |
11862 | return 1; |
11863 | } |
11864 | |
11865 | /** |
11866 | * scsih_map_queues - map reply queues with request queues |
11867 | * @shost: SCSI host pointer |
11868 | */ |
11869 | static void scsih_map_queues(struct Scsi_Host *shost) |
11870 | { |
11871 | struct MPT3SAS_ADAPTER *ioc = |
11872 | (struct MPT3SAS_ADAPTER *)shost->hostdata; |
11873 | struct blk_mq_queue_map *map; |
11874 | int i, qoff, offset; |
11875 | int nr_msix_vectors = ioc->iopoll_q_start_index; |
11876 | int iopoll_q_count = ioc->reply_queue_count - nr_msix_vectors; |
11877 | |
11878 | if (shost->nr_hw_queues == 1) |
11879 | return; |
11880 | |
11881 | for (i = 0, qoff = 0; i < shost->nr_maps; i++) { |
11882 | map = &shost->tag_set.map[i]; |
11883 | map->nr_queues = 0; |
11884 | offset = 0; |
11885 | if (i == HCTX_TYPE_DEFAULT) { |
11886 | map->nr_queues = |
11887 | nr_msix_vectors - ioc->high_iops_queues; |
11888 | offset = ioc->high_iops_queues; |
11889 | } else if (i == HCTX_TYPE_POLL) |
11890 | map->nr_queues = iopoll_q_count; |
11891 | |
11892 | if (!map->nr_queues) |
11893 | BUG_ON(i == HCTX_TYPE_DEFAULT); |
11894 | |
11895 | /* |
11896 | * The poll queue(s) doesn't have an IRQ (and hence IRQ |
11897 | * affinity), so use the regular blk-mq cpu mapping |
11898 | */ |
11899 | map->queue_offset = qoff; |
11900 | if (i != HCTX_TYPE_POLL) |
11901 | blk_mq_pci_map_queues(qmap: map, pdev: ioc->pdev, offset); |
11902 | else |
11903 | blk_mq_map_queues(qmap: map); |
11904 | |
11905 | qoff += map->nr_queues; |
11906 | } |
11907 | } |
11908 | |
11909 | /* shost template for SAS 2.0 HBA devices */ |
11910 | static const struct scsi_host_template mpt2sas_driver_template = { |
11911 | .module = THIS_MODULE, |
11912 | .name = "Fusion MPT SAS Host" , |
11913 | .proc_name = MPT2SAS_DRIVER_NAME, |
11914 | .queuecommand = scsih_qcmd, |
11915 | .target_alloc = scsih_target_alloc, |
11916 | .slave_alloc = scsih_slave_alloc, |
11917 | .slave_configure = scsih_slave_configure, |
11918 | .target_destroy = scsih_target_destroy, |
11919 | .slave_destroy = scsih_slave_destroy, |
11920 | .scan_finished = scsih_scan_finished, |
11921 | .scan_start = scsih_scan_start, |
11922 | .change_queue_depth = scsih_change_queue_depth, |
11923 | .eh_abort_handler = scsih_abort, |
11924 | .eh_device_reset_handler = scsih_dev_reset, |
11925 | .eh_target_reset_handler = scsih_target_reset, |
11926 | .eh_host_reset_handler = scsih_host_reset, |
11927 | .bios_param = scsih_bios_param, |
11928 | .can_queue = 1, |
11929 | .this_id = -1, |
11930 | .sg_tablesize = MPT2SAS_SG_DEPTH, |
11931 | .max_sectors = 32767, |
11932 | .cmd_per_lun = 7, |
11933 | .shost_groups = mpt3sas_host_groups, |
11934 | .sdev_groups = mpt3sas_dev_groups, |
11935 | .track_queue_depth = 1, |
11936 | .cmd_size = sizeof(struct scsiio_tracker), |
11937 | }; |
11938 | |
11939 | /* raid transport support for SAS 2.0 HBA devices */ |
11940 | static struct raid_function_template mpt2sas_raid_functions = { |
11941 | .cookie = &mpt2sas_driver_template, |
11942 | .is_raid = scsih_is_raid, |
11943 | .get_resync = scsih_get_resync, |
11944 | .get_state = scsih_get_state, |
11945 | }; |
11946 | |
11947 | /* shost template for SAS 3.0 HBA devices */ |
11948 | static const struct scsi_host_template mpt3sas_driver_template = { |
11949 | .module = THIS_MODULE, |
11950 | .name = "Fusion MPT SAS Host" , |
11951 | .proc_name = MPT3SAS_DRIVER_NAME, |
11952 | .queuecommand = scsih_qcmd, |
11953 | .target_alloc = scsih_target_alloc, |
11954 | .slave_alloc = scsih_slave_alloc, |
11955 | .slave_configure = scsih_slave_configure, |
11956 | .target_destroy = scsih_target_destroy, |
11957 | .slave_destroy = scsih_slave_destroy, |
11958 | .scan_finished = scsih_scan_finished, |
11959 | .scan_start = scsih_scan_start, |
11960 | .change_queue_depth = scsih_change_queue_depth, |
11961 | .eh_abort_handler = scsih_abort, |
11962 | .eh_device_reset_handler = scsih_dev_reset, |
11963 | .eh_target_reset_handler = scsih_target_reset, |
11964 | .eh_host_reset_handler = scsih_host_reset, |
11965 | .bios_param = scsih_bios_param, |
11966 | .can_queue = 1, |
11967 | .this_id = -1, |
11968 | .sg_tablesize = MPT3SAS_SG_DEPTH, |
11969 | .max_sectors = 32767, |
11970 | .max_segment_size = 0xffffffff, |
11971 | .cmd_per_lun = 128, |
11972 | .shost_groups = mpt3sas_host_groups, |
11973 | .sdev_groups = mpt3sas_dev_groups, |
11974 | .track_queue_depth = 1, |
11975 | .cmd_size = sizeof(struct scsiio_tracker), |
11976 | .map_queues = scsih_map_queues, |
11977 | .mq_poll = mpt3sas_blk_mq_poll, |
11978 | }; |
11979 | |
11980 | /* raid transport support for SAS 3.0 HBA devices */ |
11981 | static struct raid_function_template mpt3sas_raid_functions = { |
11982 | .cookie = &mpt3sas_driver_template, |
11983 | .is_raid = scsih_is_raid, |
11984 | .get_resync = scsih_get_resync, |
11985 | .get_state = scsih_get_state, |
11986 | }; |
11987 | |
11988 | /** |
11989 | * _scsih_determine_hba_mpi_version - determine in which MPI version class |
11990 | * this device belongs to. |
11991 | * @pdev: PCI device struct |
11992 | * |
11993 | * return MPI2_VERSION for SAS 2.0 HBA devices, |
11994 | * MPI25_VERSION for SAS 3.0 HBA devices, and |
11995 | * MPI26 VERSION for Cutlass & Invader SAS 3.0 HBA devices |
11996 | */ |
11997 | static u16 |
11998 | _scsih_determine_hba_mpi_version(struct pci_dev *pdev) |
11999 | { |
12000 | |
12001 | switch (pdev->device) { |
12002 | case MPI2_MFGPAGE_DEVID_SSS6200: |
12003 | case MPI2_MFGPAGE_DEVID_SAS2004: |
12004 | case MPI2_MFGPAGE_DEVID_SAS2008: |
12005 | case MPI2_MFGPAGE_DEVID_SAS2108_1: |
12006 | case MPI2_MFGPAGE_DEVID_SAS2108_2: |
12007 | case MPI2_MFGPAGE_DEVID_SAS2108_3: |
12008 | case MPI2_MFGPAGE_DEVID_SAS2116_1: |
12009 | case MPI2_MFGPAGE_DEVID_SAS2116_2: |
12010 | case MPI2_MFGPAGE_DEVID_SAS2208_1: |
12011 | case MPI2_MFGPAGE_DEVID_SAS2208_2: |
12012 | case MPI2_MFGPAGE_DEVID_SAS2208_3: |
12013 | case MPI2_MFGPAGE_DEVID_SAS2208_4: |
12014 | case MPI2_MFGPAGE_DEVID_SAS2208_5: |
12015 | case MPI2_MFGPAGE_DEVID_SAS2208_6: |
12016 | case MPI2_MFGPAGE_DEVID_SAS2308_1: |
12017 | case MPI2_MFGPAGE_DEVID_SAS2308_2: |
12018 | case MPI2_MFGPAGE_DEVID_SAS2308_3: |
12019 | case MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP: |
12020 | case MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP_1: |
12021 | return MPI2_VERSION; |
12022 | case MPI25_MFGPAGE_DEVID_SAS3004: |
12023 | case MPI25_MFGPAGE_DEVID_SAS3008: |
12024 | case MPI25_MFGPAGE_DEVID_SAS3108_1: |
12025 | case MPI25_MFGPAGE_DEVID_SAS3108_2: |
12026 | case MPI25_MFGPAGE_DEVID_SAS3108_5: |
12027 | case MPI25_MFGPAGE_DEVID_SAS3108_6: |
12028 | return MPI25_VERSION; |
12029 | case MPI26_MFGPAGE_DEVID_SAS3216: |
12030 | case MPI26_MFGPAGE_DEVID_SAS3224: |
12031 | case MPI26_MFGPAGE_DEVID_SAS3316_1: |
12032 | case MPI26_MFGPAGE_DEVID_SAS3316_2: |
12033 | case MPI26_MFGPAGE_DEVID_SAS3316_3: |
12034 | case MPI26_MFGPAGE_DEVID_SAS3316_4: |
12035 | case MPI26_MFGPAGE_DEVID_SAS3324_1: |
12036 | case MPI26_MFGPAGE_DEVID_SAS3324_2: |
12037 | case MPI26_MFGPAGE_DEVID_SAS3324_3: |
12038 | case MPI26_MFGPAGE_DEVID_SAS3324_4: |
12039 | case MPI26_MFGPAGE_DEVID_SAS3508: |
12040 | case MPI26_MFGPAGE_DEVID_SAS3508_1: |
12041 | case MPI26_MFGPAGE_DEVID_SAS3408: |
12042 | case MPI26_MFGPAGE_DEVID_SAS3516: |
12043 | case MPI26_MFGPAGE_DEVID_SAS3516_1: |
12044 | case MPI26_MFGPAGE_DEVID_SAS3416: |
12045 | case MPI26_MFGPAGE_DEVID_SAS3616: |
12046 | case MPI26_ATLAS_PCIe_SWITCH_DEVID: |
12047 | case MPI26_MFGPAGE_DEVID_CFG_SEC_3916: |
12048 | case MPI26_MFGPAGE_DEVID_HARD_SEC_3916: |
12049 | case MPI26_MFGPAGE_DEVID_CFG_SEC_3816: |
12050 | case MPI26_MFGPAGE_DEVID_HARD_SEC_3816: |
12051 | case MPI26_MFGPAGE_DEVID_INVALID0_3916: |
12052 | case MPI26_MFGPAGE_DEVID_INVALID1_3916: |
12053 | case MPI26_MFGPAGE_DEVID_INVALID0_3816: |
12054 | case MPI26_MFGPAGE_DEVID_INVALID1_3816: |
12055 | return MPI26_VERSION; |
12056 | } |
12057 | return 0; |
12058 | } |
12059 | |
12060 | /** |
12061 | * _scsih_probe - attach and add scsi host |
12062 | * @pdev: PCI device struct |
12063 | * @id: pci device id |
12064 | * |
12065 | * Return: 0 success, anything else error. |
12066 | */ |
12067 | static int |
12068 | _scsih_probe(struct pci_dev *pdev, const struct pci_device_id *id) |
12069 | { |
12070 | struct MPT3SAS_ADAPTER *ioc; |
12071 | struct Scsi_Host *shost = NULL; |
12072 | int rv; |
12073 | u16 hba_mpi_version; |
12074 | int iopoll_q_count = 0; |
12075 | |
12076 | /* Determine in which MPI version class this pci device belongs */ |
12077 | hba_mpi_version = _scsih_determine_hba_mpi_version(pdev); |
12078 | if (hba_mpi_version == 0) |
12079 | return -ENODEV; |
12080 | |
12081 | /* Enumerate only SAS 2.0 HBA's if hbas_to_enumerate is one, |
12082 | * for other generation HBA's return with -ENODEV |
12083 | */ |
12084 | if ((hbas_to_enumerate == 1) && (hba_mpi_version != MPI2_VERSION)) |
12085 | return -ENODEV; |
12086 | |
12087 | /* Enumerate only SAS 3.0 HBA's if hbas_to_enumerate is two, |
12088 | * for other generation HBA's return with -ENODEV |
12089 | */ |
12090 | if ((hbas_to_enumerate == 2) && (!(hba_mpi_version == MPI25_VERSION |
12091 | || hba_mpi_version == MPI26_VERSION))) |
12092 | return -ENODEV; |
12093 | |
12094 | switch (hba_mpi_version) { |
12095 | case MPI2_VERSION: |
12096 | pci_disable_link_state(pdev, PCIE_LINK_STATE_L0S | |
12097 | PCIE_LINK_STATE_L1 | PCIE_LINK_STATE_CLKPM); |
12098 | /* Use mpt2sas driver host template for SAS 2.0 HBA's */ |
12099 | shost = scsi_host_alloc(&mpt2sas_driver_template, |
12100 | sizeof(struct MPT3SAS_ADAPTER)); |
12101 | if (!shost) |
12102 | return -ENODEV; |
12103 | ioc = shost_priv(shost); |
12104 | memset(ioc, 0, sizeof(struct MPT3SAS_ADAPTER)); |
12105 | ioc->hba_mpi_version_belonged = hba_mpi_version; |
12106 | ioc->id = mpt2_ids++; |
12107 | sprintf(buf: ioc->driver_name, fmt: "%s" , MPT2SAS_DRIVER_NAME); |
12108 | switch (pdev->device) { |
12109 | case MPI2_MFGPAGE_DEVID_SSS6200: |
12110 | ioc->is_warpdrive = 1; |
12111 | ioc->hide_ir_msg = 1; |
12112 | break; |
12113 | case MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP: |
12114 | case MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP_1: |
12115 | ioc->is_mcpu_endpoint = 1; |
12116 | break; |
12117 | default: |
12118 | ioc->mfg_pg10_hide_flag = MFG_PAGE10_EXPOSE_ALL_DISKS; |
12119 | break; |
12120 | } |
12121 | |
12122 | if (multipath_on_hba == -1 || multipath_on_hba == 0) |
12123 | ioc->multipath_on_hba = 0; |
12124 | else |
12125 | ioc->multipath_on_hba = 1; |
12126 | |
12127 | break; |
12128 | case MPI25_VERSION: |
12129 | case MPI26_VERSION: |
12130 | /* Use mpt3sas driver host template for SAS 3.0 HBA's */ |
12131 | shost = scsi_host_alloc(&mpt3sas_driver_template, |
12132 | sizeof(struct MPT3SAS_ADAPTER)); |
12133 | if (!shost) |
12134 | return -ENODEV; |
12135 | ioc = shost_priv(shost); |
12136 | memset(ioc, 0, sizeof(struct MPT3SAS_ADAPTER)); |
12137 | ioc->hba_mpi_version_belonged = hba_mpi_version; |
12138 | ioc->id = mpt3_ids++; |
12139 | sprintf(buf: ioc->driver_name, fmt: "%s" , MPT3SAS_DRIVER_NAME); |
12140 | switch (pdev->device) { |
12141 | case MPI26_MFGPAGE_DEVID_SAS3508: |
12142 | case MPI26_MFGPAGE_DEVID_SAS3508_1: |
12143 | case MPI26_MFGPAGE_DEVID_SAS3408: |
12144 | case MPI26_MFGPAGE_DEVID_SAS3516: |
12145 | case MPI26_MFGPAGE_DEVID_SAS3516_1: |
12146 | case MPI26_MFGPAGE_DEVID_SAS3416: |
12147 | case MPI26_MFGPAGE_DEVID_SAS3616: |
12148 | case MPI26_ATLAS_PCIe_SWITCH_DEVID: |
12149 | ioc->is_gen35_ioc = 1; |
12150 | break; |
12151 | case MPI26_MFGPAGE_DEVID_INVALID0_3816: |
12152 | case MPI26_MFGPAGE_DEVID_INVALID0_3916: |
12153 | dev_err(&pdev->dev, |
12154 | "HBA with DeviceId 0x%04x, sub VendorId 0x%04x, sub DeviceId 0x%04x is Invalid" , |
12155 | pdev->device, pdev->subsystem_vendor, |
12156 | pdev->subsystem_device); |
12157 | return 1; |
12158 | case MPI26_MFGPAGE_DEVID_INVALID1_3816: |
12159 | case MPI26_MFGPAGE_DEVID_INVALID1_3916: |
12160 | dev_err(&pdev->dev, |
12161 | "HBA with DeviceId 0x%04x, sub VendorId 0x%04x, sub DeviceId 0x%04x is Tampered" , |
12162 | pdev->device, pdev->subsystem_vendor, |
12163 | pdev->subsystem_device); |
12164 | return 1; |
12165 | case MPI26_MFGPAGE_DEVID_CFG_SEC_3816: |
12166 | case MPI26_MFGPAGE_DEVID_CFG_SEC_3916: |
12167 | dev_info(&pdev->dev, |
12168 | "HBA is in Configurable Secure mode\n" ); |
12169 | fallthrough; |
12170 | case MPI26_MFGPAGE_DEVID_HARD_SEC_3816: |
12171 | case MPI26_MFGPAGE_DEVID_HARD_SEC_3916: |
12172 | ioc->is_aero_ioc = ioc->is_gen35_ioc = 1; |
12173 | break; |
12174 | default: |
12175 | ioc->is_gen35_ioc = ioc->is_aero_ioc = 0; |
12176 | } |
12177 | if ((ioc->hba_mpi_version_belonged == MPI25_VERSION && |
12178 | pdev->revision >= SAS3_PCI_DEVICE_C0_REVISION) || |
12179 | (ioc->hba_mpi_version_belonged == MPI26_VERSION)) { |
12180 | ioc->combined_reply_queue = 1; |
12181 | if (ioc->is_gen35_ioc) |
12182 | ioc->combined_reply_index_count = |
12183 | MPT3_SUP_REPLY_POST_HOST_INDEX_REG_COUNT_G35; |
12184 | else |
12185 | ioc->combined_reply_index_count = |
12186 | MPT3_SUP_REPLY_POST_HOST_INDEX_REG_COUNT_G3; |
12187 | } |
12188 | |
12189 | switch (ioc->is_gen35_ioc) { |
12190 | case 0: |
12191 | if (multipath_on_hba == -1 || multipath_on_hba == 0) |
12192 | ioc->multipath_on_hba = 0; |
12193 | else |
12194 | ioc->multipath_on_hba = 1; |
12195 | break; |
12196 | case 1: |
12197 | if (multipath_on_hba == -1 || multipath_on_hba > 0) |
12198 | ioc->multipath_on_hba = 1; |
12199 | else |
12200 | ioc->multipath_on_hba = 0; |
12201 | break; |
12202 | default: |
12203 | break; |
12204 | } |
12205 | |
12206 | break; |
12207 | default: |
12208 | return -ENODEV; |
12209 | } |
12210 | |
12211 | INIT_LIST_HEAD(list: &ioc->list); |
12212 | spin_lock(lock: &gioc_lock); |
12213 | list_add_tail(new: &ioc->list, head: &mpt3sas_ioc_list); |
12214 | spin_unlock(lock: &gioc_lock); |
12215 | ioc->shost = shost; |
12216 | ioc->pdev = pdev; |
12217 | ioc->scsi_io_cb_idx = scsi_io_cb_idx; |
12218 | ioc->tm_cb_idx = tm_cb_idx; |
12219 | ioc->ctl_cb_idx = ctl_cb_idx; |
12220 | ioc->base_cb_idx = base_cb_idx; |
12221 | ioc->port_enable_cb_idx = port_enable_cb_idx; |
12222 | ioc->transport_cb_idx = transport_cb_idx; |
12223 | ioc->scsih_cb_idx = scsih_cb_idx; |
12224 | ioc->config_cb_idx = config_cb_idx; |
12225 | ioc->tm_tr_cb_idx = tm_tr_cb_idx; |
12226 | ioc->tm_tr_volume_cb_idx = tm_tr_volume_cb_idx; |
12227 | ioc->tm_sas_control_cb_idx = tm_sas_control_cb_idx; |
12228 | ioc->logging_level = logging_level; |
12229 | ioc->schedule_dead_ioc_flush_running_cmds = &_scsih_flush_running_cmds; |
12230 | /* Host waits for minimum of six seconds */ |
12231 | ioc->max_shutdown_latency = IO_UNIT_CONTROL_SHUTDOWN_TIMEOUT; |
12232 | /* |
12233 | * Enable MEMORY MOVE support flag. |
12234 | */ |
12235 | ioc->drv_support_bitmap |= MPT_DRV_SUPPORT_BITMAP_MEMMOVE; |
12236 | /* Enable ADDITIONAL QUERY support flag. */ |
12237 | ioc->drv_support_bitmap |= MPT_DRV_SUPPORT_BITMAP_ADDNLQUERY; |
12238 | |
12239 | ioc->enable_sdev_max_qd = enable_sdev_max_qd; |
12240 | |
12241 | /* misc semaphores and spin locks */ |
12242 | mutex_init(&ioc->reset_in_progress_mutex); |
12243 | mutex_init(&ioc->hostdiag_unlock_mutex); |
12244 | /* initializing pci_access_mutex lock */ |
12245 | mutex_init(&ioc->pci_access_mutex); |
12246 | spin_lock_init(&ioc->ioc_reset_in_progress_lock); |
12247 | spin_lock_init(&ioc->scsi_lookup_lock); |
12248 | spin_lock_init(&ioc->sas_device_lock); |
12249 | spin_lock_init(&ioc->sas_node_lock); |
12250 | spin_lock_init(&ioc->fw_event_lock); |
12251 | spin_lock_init(&ioc->raid_device_lock); |
12252 | spin_lock_init(&ioc->pcie_device_lock); |
12253 | spin_lock_init(&ioc->diag_trigger_lock); |
12254 | |
12255 | INIT_LIST_HEAD(list: &ioc->sas_device_list); |
12256 | INIT_LIST_HEAD(list: &ioc->sas_device_init_list); |
12257 | INIT_LIST_HEAD(list: &ioc->sas_expander_list); |
12258 | INIT_LIST_HEAD(list: &ioc->enclosure_list); |
12259 | INIT_LIST_HEAD(list: &ioc->pcie_device_list); |
12260 | INIT_LIST_HEAD(list: &ioc->pcie_device_init_list); |
12261 | INIT_LIST_HEAD(list: &ioc->fw_event_list); |
12262 | INIT_LIST_HEAD(list: &ioc->raid_device_list); |
12263 | INIT_LIST_HEAD(list: &ioc->sas_hba.sas_port_list); |
12264 | INIT_LIST_HEAD(list: &ioc->delayed_tr_list); |
12265 | INIT_LIST_HEAD(list: &ioc->delayed_sc_list); |
12266 | INIT_LIST_HEAD(list: &ioc->delayed_event_ack_list); |
12267 | INIT_LIST_HEAD(list: &ioc->delayed_tr_volume_list); |
12268 | INIT_LIST_HEAD(list: &ioc->reply_queue_list); |
12269 | INIT_LIST_HEAD(list: &ioc->port_table_list); |
12270 | |
12271 | sprintf(buf: ioc->name, fmt: "%s_cm%d" , ioc->driver_name, ioc->id); |
12272 | |
12273 | /* init shost parameters */ |
12274 | shost->max_cmd_len = 32; |
12275 | shost->max_lun = max_lun; |
12276 | shost->transportt = mpt3sas_transport_template; |
12277 | shost->unique_id = ioc->id; |
12278 | |
12279 | if (ioc->is_mcpu_endpoint) { |
12280 | /* mCPU MPI support 64K max IO */ |
12281 | shost->max_sectors = 128; |
12282 | ioc_info(ioc, "The max_sectors value is set to %d\n" , |
12283 | shost->max_sectors); |
12284 | } else { |
12285 | if (max_sectors != 0xFFFF) { |
12286 | if (max_sectors < 64) { |
12287 | shost->max_sectors = 64; |
12288 | ioc_warn(ioc, "Invalid value %d passed for max_sectors, range is 64 to 32767. Assigning value of 64.\n" , |
12289 | max_sectors); |
12290 | } else if (max_sectors > 32767) { |
12291 | shost->max_sectors = 32767; |
12292 | ioc_warn(ioc, "Invalid value %d passed for max_sectors, range is 64 to 32767.Assigning default value of 32767.\n" , |
12293 | max_sectors); |
12294 | } else { |
12295 | shost->max_sectors = max_sectors & 0xFFFE; |
12296 | ioc_info(ioc, "The max_sectors value is set to %d\n" , |
12297 | shost->max_sectors); |
12298 | } |
12299 | } |
12300 | } |
12301 | /* register EEDP capabilities with SCSI layer */ |
12302 | if (prot_mask >= 0) |
12303 | scsi_host_set_prot(shost, mask: (prot_mask & 0x07)); |
12304 | else |
12305 | scsi_host_set_prot(shost, mask: SHOST_DIF_TYPE1_PROTECTION |
12306 | | SHOST_DIF_TYPE2_PROTECTION |
12307 | | SHOST_DIF_TYPE3_PROTECTION); |
12308 | |
12309 | scsi_host_set_guard(shost, type: SHOST_DIX_GUARD_CRC); |
12310 | |
12311 | /* event thread */ |
12312 | snprintf(buf: ioc->firmware_event_name, size: sizeof(ioc->firmware_event_name), |
12313 | fmt: "fw_event_%s%d" , ioc->driver_name, ioc->id); |
12314 | ioc->firmware_event_thread = alloc_ordered_workqueue( |
12315 | ioc->firmware_event_name, 0); |
12316 | if (!ioc->firmware_event_thread) { |
12317 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
12318 | __FILE__, __LINE__, __func__); |
12319 | rv = -ENODEV; |
12320 | goto out_thread_fail; |
12321 | } |
12322 | |
12323 | shost->host_tagset = 0; |
12324 | |
12325 | if (ioc->is_gen35_ioc && host_tagset_enable) |
12326 | shost->host_tagset = 1; |
12327 | |
12328 | ioc->is_driver_loading = 1; |
12329 | if ((mpt3sas_base_attach(ioc))) { |
12330 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
12331 | __FILE__, __LINE__, __func__); |
12332 | rv = -ENODEV; |
12333 | goto out_attach_fail; |
12334 | } |
12335 | |
12336 | if (ioc->is_warpdrive) { |
12337 | if (ioc->mfg_pg10_hide_flag == MFG_PAGE10_EXPOSE_ALL_DISKS) |
12338 | ioc->hide_drives = 0; |
12339 | else if (ioc->mfg_pg10_hide_flag == MFG_PAGE10_HIDE_ALL_DISKS) |
12340 | ioc->hide_drives = 1; |
12341 | else { |
12342 | if (mpt3sas_get_num_volumes(ioc)) |
12343 | ioc->hide_drives = 1; |
12344 | else |
12345 | ioc->hide_drives = 0; |
12346 | } |
12347 | } else |
12348 | ioc->hide_drives = 0; |
12349 | |
12350 | shost->nr_hw_queues = 1; |
12351 | |
12352 | if (shost->host_tagset) { |
12353 | shost->nr_hw_queues = |
12354 | ioc->reply_queue_count - ioc->high_iops_queues; |
12355 | |
12356 | iopoll_q_count = |
12357 | ioc->reply_queue_count - ioc->iopoll_q_start_index; |
12358 | |
12359 | shost->nr_maps = iopoll_q_count ? 3 : 1; |
12360 | |
12361 | dev_info(&ioc->pdev->dev, |
12362 | "Max SCSIIO MPT commands: %d shared with nr_hw_queues = %d\n" , |
12363 | shost->can_queue, shost->nr_hw_queues); |
12364 | } |
12365 | |
12366 | rv = scsi_add_host(host: shost, dev: &pdev->dev); |
12367 | if (rv) { |
12368 | ioc_err(ioc, "failure at %s:%d/%s()!\n" , |
12369 | __FILE__, __LINE__, __func__); |
12370 | goto out_add_shost_fail; |
12371 | } |
12372 | |
12373 | scsi_scan_host(shost); |
12374 | mpt3sas_setup_debugfs(ioc); |
12375 | return 0; |
12376 | out_add_shost_fail: |
12377 | mpt3sas_base_detach(ioc); |
12378 | out_attach_fail: |
12379 | destroy_workqueue(wq: ioc->firmware_event_thread); |
12380 | out_thread_fail: |
12381 | spin_lock(lock: &gioc_lock); |
12382 | list_del(entry: &ioc->list); |
12383 | spin_unlock(lock: &gioc_lock); |
12384 | scsi_host_put(t: shost); |
12385 | return rv; |
12386 | } |
12387 | |
12388 | /** |
12389 | * scsih_suspend - power management suspend main entry point |
12390 | * @dev: Device struct |
12391 | * |
12392 | * Return: 0 success, anything else error. |
12393 | */ |
12394 | static int __maybe_unused |
12395 | scsih_suspend(struct device *dev) |
12396 | { |
12397 | struct pci_dev *pdev = to_pci_dev(dev); |
12398 | struct Scsi_Host *shost; |
12399 | struct MPT3SAS_ADAPTER *ioc; |
12400 | int rc; |
12401 | |
12402 | rc = _scsih_get_shost_and_ioc(pdev, shost: &shost, ioc: &ioc); |
12403 | if (rc) |
12404 | return rc; |
12405 | |
12406 | mpt3sas_base_stop_watchdog(ioc); |
12407 | scsi_block_requests(shost); |
12408 | _scsih_nvme_shutdown(ioc); |
12409 | ioc_info(ioc, "pdev=0x%p, slot=%s, entering operating state\n" , |
12410 | pdev, pci_name(pdev)); |
12411 | |
12412 | mpt3sas_base_free_resources(ioc); |
12413 | return 0; |
12414 | } |
12415 | |
12416 | /** |
12417 | * scsih_resume - power management resume main entry point |
12418 | * @dev: Device struct |
12419 | * |
12420 | * Return: 0 success, anything else error. |
12421 | */ |
12422 | static int __maybe_unused |
12423 | scsih_resume(struct device *dev) |
12424 | { |
12425 | struct pci_dev *pdev = to_pci_dev(dev); |
12426 | struct Scsi_Host *shost; |
12427 | struct MPT3SAS_ADAPTER *ioc; |
12428 | pci_power_t device_state = pdev->current_state; |
12429 | int r; |
12430 | |
12431 | r = _scsih_get_shost_and_ioc(pdev, shost: &shost, ioc: &ioc); |
12432 | if (r) |
12433 | return r; |
12434 | |
12435 | ioc_info(ioc, "pdev=0x%p, slot=%s, previous operating state [D%d]\n" , |
12436 | pdev, pci_name(pdev), device_state); |
12437 | |
12438 | ioc->pdev = pdev; |
12439 | r = mpt3sas_base_map_resources(ioc); |
12440 | if (r) |
12441 | return r; |
12442 | ioc_info(ioc, "Issuing Hard Reset as part of OS Resume\n" ); |
12443 | mpt3sas_base_hard_reset_handler(ioc, type: SOFT_RESET); |
12444 | scsi_unblock_requests(shost); |
12445 | mpt3sas_base_start_watchdog(ioc); |
12446 | return 0; |
12447 | } |
12448 | |
12449 | /** |
12450 | * scsih_pci_error_detected - Called when a PCI error is detected. |
12451 | * @pdev: PCI device struct |
12452 | * @state: PCI channel state |
12453 | * |
12454 | * Description: Called when a PCI error is detected. |
12455 | * |
12456 | * Return: PCI_ERS_RESULT_NEED_RESET or PCI_ERS_RESULT_DISCONNECT. |
12457 | */ |
12458 | static pci_ers_result_t |
12459 | scsih_pci_error_detected(struct pci_dev *pdev, pci_channel_state_t state) |
12460 | { |
12461 | struct Scsi_Host *shost; |
12462 | struct MPT3SAS_ADAPTER *ioc; |
12463 | |
12464 | if (_scsih_get_shost_and_ioc(pdev, shost: &shost, ioc: &ioc)) |
12465 | return PCI_ERS_RESULT_DISCONNECT; |
12466 | |
12467 | ioc_info(ioc, "PCI error: detected callback, state(%d)!!\n" , state); |
12468 | |
12469 | switch (state) { |
12470 | case pci_channel_io_normal: |
12471 | return PCI_ERS_RESULT_CAN_RECOVER; |
12472 | case pci_channel_io_frozen: |
12473 | /* Fatal error, prepare for slot reset */ |
12474 | ioc->pci_error_recovery = 1; |
12475 | scsi_block_requests(ioc->shost); |
12476 | mpt3sas_base_stop_watchdog(ioc); |
12477 | mpt3sas_base_free_resources(ioc); |
12478 | return PCI_ERS_RESULT_NEED_RESET; |
12479 | case pci_channel_io_perm_failure: |
12480 | /* Permanent error, prepare for device removal */ |
12481 | ioc->pci_error_recovery = 1; |
12482 | mpt3sas_base_stop_watchdog(ioc); |
12483 | mpt3sas_base_pause_mq_polling(ioc); |
12484 | _scsih_flush_running_cmds(ioc); |
12485 | return PCI_ERS_RESULT_DISCONNECT; |
12486 | } |
12487 | return PCI_ERS_RESULT_NEED_RESET; |
12488 | } |
12489 | |
12490 | /** |
12491 | * scsih_pci_slot_reset - Called when PCI slot has been reset. |
12492 | * @pdev: PCI device struct |
12493 | * |
12494 | * Description: This routine is called by the pci error recovery |
12495 | * code after the PCI slot has been reset, just before we |
12496 | * should resume normal operations. |
12497 | */ |
12498 | static pci_ers_result_t |
12499 | scsih_pci_slot_reset(struct pci_dev *pdev) |
12500 | { |
12501 | struct Scsi_Host *shost; |
12502 | struct MPT3SAS_ADAPTER *ioc; |
12503 | int rc; |
12504 | |
12505 | if (_scsih_get_shost_and_ioc(pdev, shost: &shost, ioc: &ioc)) |
12506 | return PCI_ERS_RESULT_DISCONNECT; |
12507 | |
12508 | ioc_info(ioc, "PCI error: slot reset callback!!\n" ); |
12509 | |
12510 | ioc->pci_error_recovery = 0; |
12511 | ioc->pdev = pdev; |
12512 | pci_restore_state(dev: pdev); |
12513 | rc = mpt3sas_base_map_resources(ioc); |
12514 | if (rc) |
12515 | return PCI_ERS_RESULT_DISCONNECT; |
12516 | |
12517 | ioc_info(ioc, "Issuing Hard Reset as part of PCI Slot Reset\n" ); |
12518 | rc = mpt3sas_base_hard_reset_handler(ioc, type: FORCE_BIG_HAMMER); |
12519 | |
12520 | ioc_warn(ioc, "hard reset: %s\n" , |
12521 | (rc == 0) ? "success" : "failed" ); |
12522 | |
12523 | if (!rc) |
12524 | return PCI_ERS_RESULT_RECOVERED; |
12525 | else |
12526 | return PCI_ERS_RESULT_DISCONNECT; |
12527 | } |
12528 | |
12529 | /** |
12530 | * scsih_pci_resume() - resume normal ops after PCI reset |
12531 | * @pdev: pointer to PCI device |
12532 | * |
12533 | * Called when the error recovery driver tells us that its |
12534 | * OK to resume normal operation. Use completion to allow |
12535 | * halted scsi ops to resume. |
12536 | */ |
12537 | static void |
12538 | scsih_pci_resume(struct pci_dev *pdev) |
12539 | { |
12540 | struct Scsi_Host *shost; |
12541 | struct MPT3SAS_ADAPTER *ioc; |
12542 | |
12543 | if (_scsih_get_shost_and_ioc(pdev, shost: &shost, ioc: &ioc)) |
12544 | return; |
12545 | |
12546 | ioc_info(ioc, "PCI error: resume callback!!\n" ); |
12547 | |
12548 | mpt3sas_base_start_watchdog(ioc); |
12549 | scsi_unblock_requests(ioc->shost); |
12550 | } |
12551 | |
12552 | /** |
12553 | * scsih_pci_mmio_enabled - Enable MMIO and dump debug registers |
12554 | * @pdev: pointer to PCI device |
12555 | */ |
12556 | static pci_ers_result_t |
12557 | scsih_pci_mmio_enabled(struct pci_dev *pdev) |
12558 | { |
12559 | struct Scsi_Host *shost; |
12560 | struct MPT3SAS_ADAPTER *ioc; |
12561 | |
12562 | if (_scsih_get_shost_and_ioc(pdev, shost: &shost, ioc: &ioc)) |
12563 | return PCI_ERS_RESULT_DISCONNECT; |
12564 | |
12565 | ioc_info(ioc, "PCI error: mmio enabled callback!!\n" ); |
12566 | |
12567 | /* TODO - dump whatever for debugging purposes */ |
12568 | |
12569 | /* This called only if scsih_pci_error_detected returns |
12570 | * PCI_ERS_RESULT_CAN_RECOVER. Read/write to the device still |
12571 | * works, no need to reset slot. |
12572 | */ |
12573 | return PCI_ERS_RESULT_RECOVERED; |
12574 | } |
12575 | |
12576 | /** |
12577 | * scsih_ncq_prio_supp - Check for NCQ command priority support |
12578 | * @sdev: scsi device struct |
12579 | * |
12580 | * This is called when a user indicates they would like to enable |
12581 | * ncq command priorities. This works only on SATA devices. |
12582 | */ |
12583 | bool scsih_ncq_prio_supp(struct scsi_device *sdev) |
12584 | { |
12585 | struct scsi_vpd *vpd; |
12586 | bool ncq_prio_supp = false; |
12587 | |
12588 | rcu_read_lock(); |
12589 | vpd = rcu_dereference(sdev->vpd_pg89); |
12590 | if (!vpd || vpd->len < 214) |
12591 | goto out; |
12592 | |
12593 | ncq_prio_supp = (vpd->data[213] >> 4) & 1; |
12594 | out: |
12595 | rcu_read_unlock(); |
12596 | |
12597 | return ncq_prio_supp; |
12598 | } |
12599 | /* |
12600 | * The pci device ids are defined in mpi/mpi2_cnfg.h. |
12601 | */ |
12602 | static const struct pci_device_id mpt3sas_pci_table[] = { |
12603 | /* Spitfire ~ 2004 */ |
12604 | { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2004, |
12605 | PCI_ANY_ID, PCI_ANY_ID }, |
12606 | /* Falcon ~ 2008 */ |
12607 | { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2008, |
12608 | PCI_ANY_ID, PCI_ANY_ID }, |
12609 | /* Liberator ~ 2108 */ |
12610 | { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2108_1, |
12611 | PCI_ANY_ID, PCI_ANY_ID }, |
12612 | { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2108_2, |
12613 | PCI_ANY_ID, PCI_ANY_ID }, |
12614 | { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2108_3, |
12615 | PCI_ANY_ID, PCI_ANY_ID }, |
12616 | /* Meteor ~ 2116 */ |
12617 | { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2116_1, |
12618 | PCI_ANY_ID, PCI_ANY_ID }, |
12619 | { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2116_2, |
12620 | PCI_ANY_ID, PCI_ANY_ID }, |
12621 | /* Thunderbolt ~ 2208 */ |
12622 | { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_1, |
12623 | PCI_ANY_ID, PCI_ANY_ID }, |
12624 | { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_2, |
12625 | PCI_ANY_ID, PCI_ANY_ID }, |
12626 | { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_3, |
12627 | PCI_ANY_ID, PCI_ANY_ID }, |
12628 | { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_4, |
12629 | PCI_ANY_ID, PCI_ANY_ID }, |
12630 | { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_5, |
12631 | PCI_ANY_ID, PCI_ANY_ID }, |
12632 | { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2208_6, |
12633 | PCI_ANY_ID, PCI_ANY_ID }, |
12634 | /* Mustang ~ 2308 */ |
12635 | { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2308_1, |
12636 | PCI_ANY_ID, PCI_ANY_ID }, |
12637 | { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2308_2, |
12638 | PCI_ANY_ID, PCI_ANY_ID }, |
12639 | { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SAS2308_3, |
12640 | PCI_ANY_ID, PCI_ANY_ID }, |
12641 | { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP, |
12642 | PCI_ANY_ID, PCI_ANY_ID }, |
12643 | { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SWITCH_MPI_EP_1, |
12644 | PCI_ANY_ID, PCI_ANY_ID }, |
12645 | /* SSS6200 */ |
12646 | { MPI2_MFGPAGE_VENDORID_LSI, MPI2_MFGPAGE_DEVID_SSS6200, |
12647 | PCI_ANY_ID, PCI_ANY_ID }, |
12648 | /* Fury ~ 3004 and 3008 */ |
12649 | { MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3004, |
12650 | PCI_ANY_ID, PCI_ANY_ID }, |
12651 | { MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3008, |
12652 | PCI_ANY_ID, PCI_ANY_ID }, |
12653 | /* Invader ~ 3108 */ |
12654 | { MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3108_1, |
12655 | PCI_ANY_ID, PCI_ANY_ID }, |
12656 | { MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3108_2, |
12657 | PCI_ANY_ID, PCI_ANY_ID }, |
12658 | { MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3108_5, |
12659 | PCI_ANY_ID, PCI_ANY_ID }, |
12660 | { MPI2_MFGPAGE_VENDORID_LSI, MPI25_MFGPAGE_DEVID_SAS3108_6, |
12661 | PCI_ANY_ID, PCI_ANY_ID }, |
12662 | /* Cutlass ~ 3216 and 3224 */ |
12663 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3216, |
12664 | PCI_ANY_ID, PCI_ANY_ID }, |
12665 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3224, |
12666 | PCI_ANY_ID, PCI_ANY_ID }, |
12667 | /* Intruder ~ 3316 and 3324 */ |
12668 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3316_1, |
12669 | PCI_ANY_ID, PCI_ANY_ID }, |
12670 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3316_2, |
12671 | PCI_ANY_ID, PCI_ANY_ID }, |
12672 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3316_3, |
12673 | PCI_ANY_ID, PCI_ANY_ID }, |
12674 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3316_4, |
12675 | PCI_ANY_ID, PCI_ANY_ID }, |
12676 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3324_1, |
12677 | PCI_ANY_ID, PCI_ANY_ID }, |
12678 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3324_2, |
12679 | PCI_ANY_ID, PCI_ANY_ID }, |
12680 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3324_3, |
12681 | PCI_ANY_ID, PCI_ANY_ID }, |
12682 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3324_4, |
12683 | PCI_ANY_ID, PCI_ANY_ID }, |
12684 | /* Ventura, Crusader, Harpoon & Tomcat ~ 3516, 3416, 3508 & 3408*/ |
12685 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3508, |
12686 | PCI_ANY_ID, PCI_ANY_ID }, |
12687 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3508_1, |
12688 | PCI_ANY_ID, PCI_ANY_ID }, |
12689 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3408, |
12690 | PCI_ANY_ID, PCI_ANY_ID }, |
12691 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3516, |
12692 | PCI_ANY_ID, PCI_ANY_ID }, |
12693 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3516_1, |
12694 | PCI_ANY_ID, PCI_ANY_ID }, |
12695 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3416, |
12696 | PCI_ANY_ID, PCI_ANY_ID }, |
12697 | /* Mercator ~ 3616*/ |
12698 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_SAS3616, |
12699 | PCI_ANY_ID, PCI_ANY_ID }, |
12700 | |
12701 | /* Aero SI 0x00E1 Configurable Secure |
12702 | * 0x00E2 Hard Secure |
12703 | */ |
12704 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_CFG_SEC_3916, |
12705 | PCI_ANY_ID, PCI_ANY_ID }, |
12706 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_HARD_SEC_3916, |
12707 | PCI_ANY_ID, PCI_ANY_ID }, |
12708 | |
12709 | /* |
12710 | * Aero SI –> 0x00E0 Invalid, 0x00E3 Tampered |
12711 | */ |
12712 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_INVALID0_3916, |
12713 | PCI_ANY_ID, PCI_ANY_ID }, |
12714 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_INVALID1_3916, |
12715 | PCI_ANY_ID, PCI_ANY_ID }, |
12716 | |
12717 | /* Atlas PCIe Switch Management Port */ |
12718 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_ATLAS_PCIe_SWITCH_DEVID, |
12719 | PCI_ANY_ID, PCI_ANY_ID }, |
12720 | |
12721 | /* Sea SI 0x00E5 Configurable Secure |
12722 | * 0x00E6 Hard Secure |
12723 | */ |
12724 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_CFG_SEC_3816, |
12725 | PCI_ANY_ID, PCI_ANY_ID }, |
12726 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_HARD_SEC_3816, |
12727 | PCI_ANY_ID, PCI_ANY_ID }, |
12728 | |
12729 | /* |
12730 | * ATTO Branded ExpressSAS H12xx GT |
12731 | */ |
12732 | { MPI2_MFGPAGE_VENDORID_ATTO, MPI26_MFGPAGE_DEVID_HARD_SEC_3816, |
12733 | PCI_ANY_ID, PCI_ANY_ID }, |
12734 | |
12735 | /* |
12736 | * Sea SI –> 0x00E4 Invalid, 0x00E7 Tampered |
12737 | */ |
12738 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_INVALID0_3816, |
12739 | PCI_ANY_ID, PCI_ANY_ID }, |
12740 | { MPI2_MFGPAGE_VENDORID_LSI, MPI26_MFGPAGE_DEVID_INVALID1_3816, |
12741 | PCI_ANY_ID, PCI_ANY_ID }, |
12742 | |
12743 | {0} /* Terminating entry */ |
12744 | }; |
12745 | MODULE_DEVICE_TABLE(pci, mpt3sas_pci_table); |
12746 | |
12747 | static struct pci_error_handlers _mpt3sas_err_handler = { |
12748 | .error_detected = scsih_pci_error_detected, |
12749 | .mmio_enabled = scsih_pci_mmio_enabled, |
12750 | .slot_reset = scsih_pci_slot_reset, |
12751 | .resume = scsih_pci_resume, |
12752 | }; |
12753 | |
12754 | static SIMPLE_DEV_PM_OPS(scsih_pm_ops, scsih_suspend, scsih_resume); |
12755 | |
12756 | static struct pci_driver mpt3sas_driver = { |
12757 | .name = MPT3SAS_DRIVER_NAME, |
12758 | .id_table = mpt3sas_pci_table, |
12759 | .probe = _scsih_probe, |
12760 | .remove = scsih_remove, |
12761 | .shutdown = scsih_shutdown, |
12762 | .err_handler = &_mpt3sas_err_handler, |
12763 | .driver.pm = &scsih_pm_ops, |
12764 | }; |
12765 | |
12766 | /** |
12767 | * scsih_init - main entry point for this driver. |
12768 | * |
12769 | * Return: 0 success, anything else error. |
12770 | */ |
12771 | static int |
12772 | scsih_init(void) |
12773 | { |
12774 | mpt2_ids = 0; |
12775 | mpt3_ids = 0; |
12776 | |
12777 | mpt3sas_base_initialize_callback_handler(); |
12778 | |
12779 | /* queuecommand callback hander */ |
12780 | scsi_io_cb_idx = mpt3sas_base_register_callback_handler(cb_func: _scsih_io_done); |
12781 | |
12782 | /* task management callback handler */ |
12783 | tm_cb_idx = mpt3sas_base_register_callback_handler(cb_func: _scsih_tm_done); |
12784 | |
12785 | /* base internal commands callback handler */ |
12786 | base_cb_idx = mpt3sas_base_register_callback_handler(cb_func: mpt3sas_base_done); |
12787 | port_enable_cb_idx = mpt3sas_base_register_callback_handler( |
12788 | cb_func: mpt3sas_port_enable_done); |
12789 | |
12790 | /* transport internal commands callback handler */ |
12791 | transport_cb_idx = mpt3sas_base_register_callback_handler( |
12792 | cb_func: mpt3sas_transport_done); |
12793 | |
12794 | /* scsih internal commands callback handler */ |
12795 | scsih_cb_idx = mpt3sas_base_register_callback_handler(cb_func: _scsih_done); |
12796 | |
12797 | /* configuration page API internal commands callback handler */ |
12798 | config_cb_idx = mpt3sas_base_register_callback_handler( |
12799 | cb_func: mpt3sas_config_done); |
12800 | |
12801 | /* ctl module callback handler */ |
12802 | ctl_cb_idx = mpt3sas_base_register_callback_handler(cb_func: mpt3sas_ctl_done); |
12803 | |
12804 | tm_tr_cb_idx = mpt3sas_base_register_callback_handler( |
12805 | cb_func: _scsih_tm_tr_complete); |
12806 | |
12807 | tm_tr_volume_cb_idx = mpt3sas_base_register_callback_handler( |
12808 | cb_func: _scsih_tm_volume_tr_complete); |
12809 | |
12810 | tm_sas_control_cb_idx = mpt3sas_base_register_callback_handler( |
12811 | cb_func: _scsih_sas_control_complete); |
12812 | |
12813 | mpt3sas_init_debugfs(); |
12814 | return 0; |
12815 | } |
12816 | |
12817 | /** |
12818 | * scsih_exit - exit point for this driver (when it is a module). |
12819 | * |
12820 | * Return: 0 success, anything else error. |
12821 | */ |
12822 | static void |
12823 | scsih_exit(void) |
12824 | { |
12825 | |
12826 | mpt3sas_base_release_callback_handler(cb_idx: scsi_io_cb_idx); |
12827 | mpt3sas_base_release_callback_handler(cb_idx: tm_cb_idx); |
12828 | mpt3sas_base_release_callback_handler(cb_idx: base_cb_idx); |
12829 | mpt3sas_base_release_callback_handler(cb_idx: port_enable_cb_idx); |
12830 | mpt3sas_base_release_callback_handler(cb_idx: transport_cb_idx); |
12831 | mpt3sas_base_release_callback_handler(cb_idx: scsih_cb_idx); |
12832 | mpt3sas_base_release_callback_handler(cb_idx: config_cb_idx); |
12833 | mpt3sas_base_release_callback_handler(cb_idx: ctl_cb_idx); |
12834 | |
12835 | mpt3sas_base_release_callback_handler(cb_idx: tm_tr_cb_idx); |
12836 | mpt3sas_base_release_callback_handler(cb_idx: tm_tr_volume_cb_idx); |
12837 | mpt3sas_base_release_callback_handler(cb_idx: tm_sas_control_cb_idx); |
12838 | |
12839 | /* raid transport support */ |
12840 | if (hbas_to_enumerate != 1) |
12841 | raid_class_release(mpt3sas_raid_template); |
12842 | if (hbas_to_enumerate != 2) |
12843 | raid_class_release(mpt2sas_raid_template); |
12844 | sas_release_transport(mpt3sas_transport_template); |
12845 | mpt3sas_exit_debugfs(); |
12846 | } |
12847 | |
12848 | /** |
12849 | * _mpt3sas_init - main entry point for this driver. |
12850 | * |
12851 | * Return: 0 success, anything else error. |
12852 | */ |
12853 | static int __init |
12854 | _mpt3sas_init(void) |
12855 | { |
12856 | int error; |
12857 | |
12858 | pr_info("%s version %s loaded\n" , MPT3SAS_DRIVER_NAME, |
12859 | MPT3SAS_DRIVER_VERSION); |
12860 | |
12861 | mpt3sas_transport_template = |
12862 | sas_attach_transport(&mpt3sas_transport_functions); |
12863 | if (!mpt3sas_transport_template) |
12864 | return -ENODEV; |
12865 | |
12866 | /* No need attach mpt3sas raid functions template |
12867 | * if hbas_to_enumarate value is one. |
12868 | */ |
12869 | if (hbas_to_enumerate != 1) { |
12870 | mpt3sas_raid_template = |
12871 | raid_class_attach(&mpt3sas_raid_functions); |
12872 | if (!mpt3sas_raid_template) { |
12873 | sas_release_transport(mpt3sas_transport_template); |
12874 | return -ENODEV; |
12875 | } |
12876 | } |
12877 | |
12878 | /* No need to attach mpt2sas raid functions template |
12879 | * if hbas_to_enumarate value is two |
12880 | */ |
12881 | if (hbas_to_enumerate != 2) { |
12882 | mpt2sas_raid_template = |
12883 | raid_class_attach(&mpt2sas_raid_functions); |
12884 | if (!mpt2sas_raid_template) { |
12885 | sas_release_transport(mpt3sas_transport_template); |
12886 | return -ENODEV; |
12887 | } |
12888 | } |
12889 | |
12890 | error = scsih_init(); |
12891 | if (error) { |
12892 | scsih_exit(); |
12893 | return error; |
12894 | } |
12895 | |
12896 | mpt3sas_ctl_init(hbas_to_enumerate); |
12897 | |
12898 | error = pci_register_driver(&mpt3sas_driver); |
12899 | if (error) { |
12900 | mpt3sas_ctl_exit(hbas_to_enumerate); |
12901 | scsih_exit(); |
12902 | } |
12903 | |
12904 | return error; |
12905 | } |
12906 | |
12907 | /** |
12908 | * _mpt3sas_exit - exit point for this driver (when it is a module). |
12909 | * |
12910 | */ |
12911 | static void __exit |
12912 | _mpt3sas_exit(void) |
12913 | { |
12914 | pr_info("mpt3sas version %s unloading\n" , |
12915 | MPT3SAS_DRIVER_VERSION); |
12916 | |
12917 | pci_unregister_driver(dev: &mpt3sas_driver); |
12918 | |
12919 | mpt3sas_ctl_exit(hbas_to_enumerate); |
12920 | |
12921 | scsih_exit(); |
12922 | } |
12923 | |
12924 | module_init(_mpt3sas_init); |
12925 | module_exit(_mpt3sas_exit); |
12926 | |