Lines Matching refs:shost

66 void scsi_eh_wakeup(struct Scsi_Host *shost)  in scsi_eh_wakeup()  argument
68 lockdep_assert_held(shost->host_lock); in scsi_eh_wakeup()
70 if (scsi_host_busy(shost) == shost->host_failed) { in scsi_eh_wakeup()
71 trace_scsi_eh_wakeup(shost); in scsi_eh_wakeup()
72 wake_up_process(shost->ehandler); in scsi_eh_wakeup()
73 SCSI_LOG_ERROR_RECOVERY(5, shost_printk(KERN_INFO, shost, in scsi_eh_wakeup()
84 void scsi_schedule_eh(struct Scsi_Host *shost) in scsi_schedule_eh() argument
88 spin_lock_irqsave(shost->host_lock, flags); in scsi_schedule_eh()
90 if (scsi_host_set_state(shost, SHOST_RECOVERY) == 0 || in scsi_schedule_eh()
91 scsi_host_set_state(shost, SHOST_CANCEL_RECOVERY) == 0) { in scsi_schedule_eh()
92 shost->host_eh_scheduled++; in scsi_schedule_eh()
93 scsi_eh_wakeup(shost); in scsi_schedule_eh()
96 spin_unlock_irqrestore(shost->host_lock, flags); in scsi_schedule_eh()
100 static int scsi_host_eh_past_deadline(struct Scsi_Host *shost) in scsi_host_eh_past_deadline() argument
102 if (!shost->last_reset || shost->eh_deadline == -1) in scsi_host_eh_past_deadline()
112 if (time_before(jiffies, shost->last_reset + shost->eh_deadline) && in scsi_host_eh_past_deadline()
113 shost->eh_deadline > -1) in scsi_host_eh_past_deadline()
197 struct Scsi_Host *shost = sdev->host; in scsi_abort_command() local
211 spin_lock_irqsave(shost->host_lock, flags); in scsi_abort_command()
212 if (shost->eh_deadline != -1 && !shost->last_reset) in scsi_abort_command()
213 shost->last_reset = jiffies; in scsi_abort_command()
214 spin_unlock_irqrestore(shost->host_lock, flags); in scsi_abort_command()
219 queue_delayed_work(shost->tmf_work_q, &scmd->abort_work, HZ / 100); in scsi_abort_command()
243 struct Scsi_Host *shost = scmd->device->host; in scsi_eh_inc_host_failed() local
246 spin_lock_irqsave(shost->host_lock, flags); in scsi_eh_inc_host_failed()
247 shost->host_failed++; in scsi_eh_inc_host_failed()
248 scsi_eh_wakeup(shost); in scsi_eh_inc_host_failed()
249 spin_unlock_irqrestore(shost->host_lock, flags); in scsi_eh_inc_host_failed()
258 struct Scsi_Host *shost = scmd->device->host; in scsi_eh_scmd_add() local
262 WARN_ON_ONCE(!shost->ehandler); in scsi_eh_scmd_add()
264 spin_lock_irqsave(shost->host_lock, flags); in scsi_eh_scmd_add()
265 if (scsi_host_set_state(shost, SHOST_RECOVERY)) { in scsi_eh_scmd_add()
266 ret = scsi_host_set_state(shost, SHOST_CANCEL_RECOVERY); in scsi_eh_scmd_add()
269 if (shost->eh_deadline != -1 && !shost->last_reset) in scsi_eh_scmd_add()
270 shost->last_reset = jiffies; in scsi_eh_scmd_add()
273 list_add_tail(&scmd->eh_entry, &shost->eh_cmd_q); in scsi_eh_scmd_add()
274 spin_unlock_irqrestore(shost->host_lock, flags); in scsi_eh_scmd_add()
360 static inline void scsi_eh_prt_fail_stats(struct Scsi_Host *shost, in scsi_eh_prt_fail_stats() argument
370 shost_for_each_device(sdev, shost) { in scsi_eh_prt_fail_stats()
383 shost_printk(KERN_INFO, shost, in scsi_eh_prt_fail_stats()
393 SCSI_LOG_ERROR_RECOVERY(2, shost_printk(KERN_INFO, shost, in scsi_eh_prt_fail_stats()
1067 struct Scsi_Host *shost = sdev->host; in scsi_send_eh_cmnd() local
1076 shost->eh_action = &done; in scsi_send_eh_cmnd()
1098 rtn = shost->hostt->queuecommand(shost, scmd); in scsi_send_eh_cmnd()
1118 shost->eh_action = NULL; in scsi_send_eh_cmnd()
1228 struct Scsi_Host *shost; in scsi_eh_get_sense() local
1240 shost = scmd->device->host; in scsi_eh_get_sense()
1241 if (scsi_host_eh_past_deadline(shost)) { in scsi_eh_get_sense()
1417 static int scsi_eh_stu(struct Scsi_Host *shost, in scsi_eh_stu() argument
1424 shost_for_each_device(sdev, shost) { in scsi_eh_stu()
1425 if (scsi_host_eh_past_deadline(shost)) { in scsi_eh_stu()
1483 static int scsi_eh_bus_device_reset(struct Scsi_Host *shost, in scsi_eh_bus_device_reset() argument
1491 shost_for_each_device(sdev, shost) { in scsi_eh_bus_device_reset()
1492 if (scsi_host_eh_past_deadline(shost)) { in scsi_eh_bus_device_reset()
1545 static int scsi_eh_target_reset(struct Scsi_Host *shost, in scsi_eh_target_reset() argument
1559 if (scsi_host_eh_past_deadline(shost)) { in scsi_eh_target_reset()
1564 shost_printk(KERN_INFO, shost, in scsi_eh_target_reset()
1574 shost_printk(KERN_INFO, shost, in scsi_eh_target_reset()
1580 shost_printk(KERN_INFO, shost, in scsi_eh_target_reset()
1607 static int scsi_eh_bus_reset(struct Scsi_Host *shost, in scsi_eh_bus_reset() argument
1623 for (channel = 0; channel <= shost->max_channel; channel++) { in scsi_eh_bus_reset()
1624 if (scsi_host_eh_past_deadline(shost)) { in scsi_eh_bus_reset()
1627 shost_printk(KERN_INFO, shost, in scsi_eh_bus_reset()
1648 shost_printk(KERN_INFO, shost, in scsi_eh_bus_reset()
1665 shost_printk(KERN_INFO, shost, in scsi_eh_bus_reset()
1679 static int scsi_eh_host_reset(struct Scsi_Host *shost, in scsi_eh_host_reset() argument
1692 shost_printk(KERN_INFO, shost, in scsi_eh_host_reset()
1705 shost_printk(KERN_INFO, shost, in scsi_eh_host_reset()
2021 static void scsi_restart_operations(struct Scsi_Host *shost) in scsi_restart_operations() argument
2031 shost_for_each_device(sdev, shost) { in scsi_restart_operations()
2044 shost_printk(KERN_INFO, shost, "waking up host to restart\n")); in scsi_restart_operations()
2046 spin_lock_irqsave(shost->host_lock, flags); in scsi_restart_operations()
2047 if (scsi_host_set_state(shost, SHOST_RUNNING)) in scsi_restart_operations()
2048 if (scsi_host_set_state(shost, SHOST_CANCEL)) in scsi_restart_operations()
2049 BUG_ON(scsi_host_set_state(shost, SHOST_DEL)); in scsi_restart_operations()
2050 spin_unlock_irqrestore(shost->host_lock, flags); in scsi_restart_operations()
2052 wake_up(&shost->host_wait); in scsi_restart_operations()
2060 scsi_run_host_queues(shost); in scsi_restart_operations()
2070 spin_lock_irqsave(shost->host_lock, flags); in scsi_restart_operations()
2071 if (shost->host_eh_scheduled) in scsi_restart_operations()
2072 if (scsi_host_set_state(shost, SHOST_RECOVERY)) in scsi_restart_operations()
2073 WARN_ON(scsi_host_set_state(shost, SHOST_CANCEL_RECOVERY)); in scsi_restart_operations()
2074 spin_unlock_irqrestore(shost->host_lock, flags); in scsi_restart_operations()
2083 void scsi_eh_ready_devs(struct Scsi_Host *shost, in scsi_eh_ready_devs() argument
2087 if (!scsi_eh_stu(shost, work_q, done_q)) in scsi_eh_ready_devs()
2088 if (!scsi_eh_bus_device_reset(shost, work_q, done_q)) in scsi_eh_ready_devs()
2089 if (!scsi_eh_target_reset(shost, work_q, done_q)) in scsi_eh_ready_devs()
2090 if (!scsi_eh_bus_reset(shost, work_q, done_q)) in scsi_eh_ready_devs()
2091 if (!scsi_eh_host_reset(shost, work_q, done_q)) in scsi_eh_ready_devs()
2155 static void scsi_unjam_host(struct Scsi_Host *shost) in scsi_unjam_host() argument
2161 spin_lock_irqsave(shost->host_lock, flags); in scsi_unjam_host()
2162 list_splice_init(&shost->eh_cmd_q, &eh_work_q); in scsi_unjam_host()
2163 spin_unlock_irqrestore(shost->host_lock, flags); in scsi_unjam_host()
2165 SCSI_LOG_ERROR_RECOVERY(1, scsi_eh_prt_fail_stats(shost, &eh_work_q)); in scsi_unjam_host()
2168 scsi_eh_ready_devs(shost, &eh_work_q, &eh_done_q); in scsi_unjam_host()
2170 spin_lock_irqsave(shost->host_lock, flags); in scsi_unjam_host()
2171 if (shost->eh_deadline != -1) in scsi_unjam_host()
2172 shost->last_reset = 0; in scsi_unjam_host()
2173 spin_unlock_irqrestore(shost->host_lock, flags); in scsi_unjam_host()
2187 struct Scsi_Host *shost = data; in scsi_error_handler() local
2206 if ((shost->host_failed == 0 && shost->host_eh_scheduled == 0) || in scsi_error_handler()
2207 shost->host_failed != scsi_host_busy(shost)) { in scsi_error_handler()
2209 shost_printk(KERN_INFO, shost, in scsi_error_handler()
2211 shost->host_no)); in scsi_error_handler()
2218 shost_printk(KERN_INFO, shost, in scsi_error_handler()
2220 shost->host_no, shost->host_eh_scheduled, in scsi_error_handler()
2221 shost->host_failed, in scsi_error_handler()
2222 scsi_host_busy(shost))); in scsi_error_handler()
2229 if (!shost->eh_noresume && scsi_autopm_get_host(shost) != 0) { in scsi_error_handler()
2231 shost_printk(KERN_ERR, shost, in scsi_error_handler()
2233 shost->host_no)); in scsi_error_handler()
2237 if (shost->transportt->eh_strategy_handler) in scsi_error_handler()
2238 shost->transportt->eh_strategy_handler(shost); in scsi_error_handler()
2240 scsi_unjam_host(shost); in scsi_error_handler()
2243 shost->host_failed = 0; in scsi_error_handler()
2252 scsi_restart_operations(shost); in scsi_error_handler()
2253 if (!shost->eh_noresume) in scsi_error_handler()
2254 scsi_autopm_put_host(shost); in scsi_error_handler()
2259 shost_printk(KERN_INFO, shost, in scsi_error_handler()
2261 shost->host_no)); in scsi_error_handler()
2262 shost->ehandler = NULL; in scsi_error_handler()
2287 void scsi_report_bus_reset(struct Scsi_Host *shost, int channel) in scsi_report_bus_reset() argument
2291 __shost_for_each_device(sdev, shost) { in scsi_report_bus_reset()
2320 void scsi_report_device_reset(struct Scsi_Host *shost, int channel, int target) in scsi_report_device_reset() argument
2324 __shost_for_each_device(sdev, shost) { in scsi_report_device_reset()
2346 struct Scsi_Host *shost = dev->host; in scsi_ioctl_reset() local
2358 if (scsi_autopm_get_host(shost) < 0) in scsi_ioctl_reset()
2363 shost->hostt->cmd_size, GFP_KERNEL); in scsi_ioctl_reset()
2380 spin_lock_irqsave(shost->host_lock, flags); in scsi_ioctl_reset()
2381 shost->tmf_in_progress = 1; in scsi_ioctl_reset()
2382 spin_unlock_irqrestore(shost->host_lock, flags); in scsi_ioctl_reset()
2415 spin_lock_irqsave(shost->host_lock, flags); in scsi_ioctl_reset()
2416 shost->tmf_in_progress = 0; in scsi_ioctl_reset()
2417 spin_unlock_irqrestore(shost->host_lock, flags); in scsi_ioctl_reset()
2424 shost_printk(KERN_INFO, shost, in scsi_ioctl_reset()
2427 wake_up(&shost->host_wait); in scsi_ioctl_reset()
2428 scsi_run_host_queues(shost); in scsi_ioctl_reset()
2433 scsi_autopm_put_host(shost); in scsi_ioctl_reset()