1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun * Linux Driver for Mylex DAC960/AcceleRAID/eXtremeRAID PCI RAID Controllers
4*4882a593Smuzhiyun *
5*4882a593Smuzhiyun * This driver supports the newer, SCSI-based firmware interface only.
6*4882a593Smuzhiyun *
7*4882a593Smuzhiyun * Copyright 2017 Hannes Reinecke, SUSE Linux GmbH <hare@suse.com>
8*4882a593Smuzhiyun *
9*4882a593Smuzhiyun * Based on the original DAC960 driver, which has
10*4882a593Smuzhiyun * Copyright 1998-2001 by Leonard N. Zubkoff <lnz@dandelion.com>
11*4882a593Smuzhiyun * Portions Copyright 2002 by Mylex (An IBM Business Unit)
12*4882a593Smuzhiyun */
13*4882a593Smuzhiyun
14*4882a593Smuzhiyun #include <linux/module.h>
15*4882a593Smuzhiyun #include <linux/types.h>
16*4882a593Smuzhiyun #include <linux/delay.h>
17*4882a593Smuzhiyun #include <linux/interrupt.h>
18*4882a593Smuzhiyun #include <linux/pci.h>
19*4882a593Smuzhiyun #include <linux/raid_class.h>
20*4882a593Smuzhiyun #include <asm/unaligned.h>
21*4882a593Smuzhiyun #include <scsi/scsi.h>
22*4882a593Smuzhiyun #include <scsi/scsi_host.h>
23*4882a593Smuzhiyun #include <scsi/scsi_device.h>
24*4882a593Smuzhiyun #include <scsi/scsi_cmnd.h>
25*4882a593Smuzhiyun #include <scsi/scsi_tcq.h>
26*4882a593Smuzhiyun #include "myrs.h"
27*4882a593Smuzhiyun
28*4882a593Smuzhiyun static struct raid_template *myrs_raid_template;
29*4882a593Smuzhiyun
30*4882a593Smuzhiyun static struct myrs_devstate_name_entry {
31*4882a593Smuzhiyun enum myrs_devstate state;
32*4882a593Smuzhiyun char *name;
33*4882a593Smuzhiyun } myrs_devstate_name_list[] = {
34*4882a593Smuzhiyun { MYRS_DEVICE_UNCONFIGURED, "Unconfigured" },
35*4882a593Smuzhiyun { MYRS_DEVICE_ONLINE, "Online" },
36*4882a593Smuzhiyun { MYRS_DEVICE_REBUILD, "Rebuild" },
37*4882a593Smuzhiyun { MYRS_DEVICE_MISSING, "Missing" },
38*4882a593Smuzhiyun { MYRS_DEVICE_SUSPECTED_CRITICAL, "SuspectedCritical" },
39*4882a593Smuzhiyun { MYRS_DEVICE_OFFLINE, "Offline" },
40*4882a593Smuzhiyun { MYRS_DEVICE_CRITICAL, "Critical" },
41*4882a593Smuzhiyun { MYRS_DEVICE_SUSPECTED_DEAD, "SuspectedDead" },
42*4882a593Smuzhiyun { MYRS_DEVICE_COMMANDED_OFFLINE, "CommandedOffline" },
43*4882a593Smuzhiyun { MYRS_DEVICE_STANDBY, "Standby" },
44*4882a593Smuzhiyun { MYRS_DEVICE_INVALID_STATE, "Invalid" },
45*4882a593Smuzhiyun };
46*4882a593Smuzhiyun
myrs_devstate_name(enum myrs_devstate state)47*4882a593Smuzhiyun static char *myrs_devstate_name(enum myrs_devstate state)
48*4882a593Smuzhiyun {
49*4882a593Smuzhiyun struct myrs_devstate_name_entry *entry = myrs_devstate_name_list;
50*4882a593Smuzhiyun int i;
51*4882a593Smuzhiyun
52*4882a593Smuzhiyun for (i = 0; i < ARRAY_SIZE(myrs_devstate_name_list); i++) {
53*4882a593Smuzhiyun if (entry[i].state == state)
54*4882a593Smuzhiyun return entry[i].name;
55*4882a593Smuzhiyun }
56*4882a593Smuzhiyun return NULL;
57*4882a593Smuzhiyun }
58*4882a593Smuzhiyun
59*4882a593Smuzhiyun static struct myrs_raid_level_name_entry {
60*4882a593Smuzhiyun enum myrs_raid_level level;
61*4882a593Smuzhiyun char *name;
62*4882a593Smuzhiyun } myrs_raid_level_name_list[] = {
63*4882a593Smuzhiyun { MYRS_RAID_LEVEL0, "RAID0" },
64*4882a593Smuzhiyun { MYRS_RAID_LEVEL1, "RAID1" },
65*4882a593Smuzhiyun { MYRS_RAID_LEVEL3, "RAID3 right asymmetric parity" },
66*4882a593Smuzhiyun { MYRS_RAID_LEVEL5, "RAID5 right asymmetric parity" },
67*4882a593Smuzhiyun { MYRS_RAID_LEVEL6, "RAID6" },
68*4882a593Smuzhiyun { MYRS_RAID_JBOD, "JBOD" },
69*4882a593Smuzhiyun { MYRS_RAID_NEWSPAN, "New Mylex SPAN" },
70*4882a593Smuzhiyun { MYRS_RAID_LEVEL3F, "RAID3 fixed parity" },
71*4882a593Smuzhiyun { MYRS_RAID_LEVEL3L, "RAID3 left symmetric parity" },
72*4882a593Smuzhiyun { MYRS_RAID_SPAN, "Mylex SPAN" },
73*4882a593Smuzhiyun { MYRS_RAID_LEVEL5L, "RAID5 left symmetric parity" },
74*4882a593Smuzhiyun { MYRS_RAID_LEVELE, "RAIDE (concatenation)" },
75*4882a593Smuzhiyun { MYRS_RAID_PHYSICAL, "Physical device" },
76*4882a593Smuzhiyun };
77*4882a593Smuzhiyun
myrs_raid_level_name(enum myrs_raid_level level)78*4882a593Smuzhiyun static char *myrs_raid_level_name(enum myrs_raid_level level)
79*4882a593Smuzhiyun {
80*4882a593Smuzhiyun struct myrs_raid_level_name_entry *entry = myrs_raid_level_name_list;
81*4882a593Smuzhiyun int i;
82*4882a593Smuzhiyun
83*4882a593Smuzhiyun for (i = 0; i < ARRAY_SIZE(myrs_raid_level_name_list); i++) {
84*4882a593Smuzhiyun if (entry[i].level == level)
85*4882a593Smuzhiyun return entry[i].name;
86*4882a593Smuzhiyun }
87*4882a593Smuzhiyun return NULL;
88*4882a593Smuzhiyun }
89*4882a593Smuzhiyun
90*4882a593Smuzhiyun /*
91*4882a593Smuzhiyun * myrs_reset_cmd - clears critical fields in struct myrs_cmdblk
92*4882a593Smuzhiyun */
myrs_reset_cmd(struct myrs_cmdblk * cmd_blk)93*4882a593Smuzhiyun static inline void myrs_reset_cmd(struct myrs_cmdblk *cmd_blk)
94*4882a593Smuzhiyun {
95*4882a593Smuzhiyun union myrs_cmd_mbox *mbox = &cmd_blk->mbox;
96*4882a593Smuzhiyun
97*4882a593Smuzhiyun memset(mbox, 0, sizeof(union myrs_cmd_mbox));
98*4882a593Smuzhiyun cmd_blk->status = 0;
99*4882a593Smuzhiyun }
100*4882a593Smuzhiyun
101*4882a593Smuzhiyun /*
102*4882a593Smuzhiyun * myrs_qcmd - queues Command for DAC960 V2 Series Controllers.
103*4882a593Smuzhiyun */
myrs_qcmd(struct myrs_hba * cs,struct myrs_cmdblk * cmd_blk)104*4882a593Smuzhiyun static void myrs_qcmd(struct myrs_hba *cs, struct myrs_cmdblk *cmd_blk)
105*4882a593Smuzhiyun {
106*4882a593Smuzhiyun void __iomem *base = cs->io_base;
107*4882a593Smuzhiyun union myrs_cmd_mbox *mbox = &cmd_blk->mbox;
108*4882a593Smuzhiyun union myrs_cmd_mbox *next_mbox = cs->next_cmd_mbox;
109*4882a593Smuzhiyun
110*4882a593Smuzhiyun cs->write_cmd_mbox(next_mbox, mbox);
111*4882a593Smuzhiyun
112*4882a593Smuzhiyun if (cs->prev_cmd_mbox1->words[0] == 0 ||
113*4882a593Smuzhiyun cs->prev_cmd_mbox2->words[0] == 0)
114*4882a593Smuzhiyun cs->get_cmd_mbox(base);
115*4882a593Smuzhiyun
116*4882a593Smuzhiyun cs->prev_cmd_mbox2 = cs->prev_cmd_mbox1;
117*4882a593Smuzhiyun cs->prev_cmd_mbox1 = next_mbox;
118*4882a593Smuzhiyun
119*4882a593Smuzhiyun if (++next_mbox > cs->last_cmd_mbox)
120*4882a593Smuzhiyun next_mbox = cs->first_cmd_mbox;
121*4882a593Smuzhiyun
122*4882a593Smuzhiyun cs->next_cmd_mbox = next_mbox;
123*4882a593Smuzhiyun }
124*4882a593Smuzhiyun
125*4882a593Smuzhiyun /*
126*4882a593Smuzhiyun * myrs_exec_cmd - executes V2 Command and waits for completion.
127*4882a593Smuzhiyun */
myrs_exec_cmd(struct myrs_hba * cs,struct myrs_cmdblk * cmd_blk)128*4882a593Smuzhiyun static void myrs_exec_cmd(struct myrs_hba *cs,
129*4882a593Smuzhiyun struct myrs_cmdblk *cmd_blk)
130*4882a593Smuzhiyun {
131*4882a593Smuzhiyun DECLARE_COMPLETION_ONSTACK(complete);
132*4882a593Smuzhiyun unsigned long flags;
133*4882a593Smuzhiyun
134*4882a593Smuzhiyun cmd_blk->complete = &complete;
135*4882a593Smuzhiyun spin_lock_irqsave(&cs->queue_lock, flags);
136*4882a593Smuzhiyun myrs_qcmd(cs, cmd_blk);
137*4882a593Smuzhiyun spin_unlock_irqrestore(&cs->queue_lock, flags);
138*4882a593Smuzhiyun
139*4882a593Smuzhiyun WARN_ON(in_interrupt());
140*4882a593Smuzhiyun wait_for_completion(&complete);
141*4882a593Smuzhiyun }
142*4882a593Smuzhiyun
143*4882a593Smuzhiyun /*
144*4882a593Smuzhiyun * myrs_report_progress - prints progress message
145*4882a593Smuzhiyun */
myrs_report_progress(struct myrs_hba * cs,unsigned short ldev_num,unsigned char * msg,unsigned long blocks,unsigned long size)146*4882a593Smuzhiyun static void myrs_report_progress(struct myrs_hba *cs, unsigned short ldev_num,
147*4882a593Smuzhiyun unsigned char *msg, unsigned long blocks,
148*4882a593Smuzhiyun unsigned long size)
149*4882a593Smuzhiyun {
150*4882a593Smuzhiyun shost_printk(KERN_INFO, cs->host,
151*4882a593Smuzhiyun "Logical Drive %d: %s in Progress: %d%% completed\n",
152*4882a593Smuzhiyun ldev_num, msg,
153*4882a593Smuzhiyun (100 * (int)(blocks >> 7)) / (int)(size >> 7));
154*4882a593Smuzhiyun }
155*4882a593Smuzhiyun
156*4882a593Smuzhiyun /*
157*4882a593Smuzhiyun * myrs_get_ctlr_info - executes a Controller Information IOCTL Command
158*4882a593Smuzhiyun */
myrs_get_ctlr_info(struct myrs_hba * cs)159*4882a593Smuzhiyun static unsigned char myrs_get_ctlr_info(struct myrs_hba *cs)
160*4882a593Smuzhiyun {
161*4882a593Smuzhiyun struct myrs_cmdblk *cmd_blk = &cs->dcmd_blk;
162*4882a593Smuzhiyun union myrs_cmd_mbox *mbox = &cmd_blk->mbox;
163*4882a593Smuzhiyun dma_addr_t ctlr_info_addr;
164*4882a593Smuzhiyun union myrs_sgl *sgl;
165*4882a593Smuzhiyun unsigned char status;
166*4882a593Smuzhiyun unsigned short ldev_present, ldev_critical, ldev_offline;
167*4882a593Smuzhiyun
168*4882a593Smuzhiyun ldev_present = cs->ctlr_info->ldev_present;
169*4882a593Smuzhiyun ldev_critical = cs->ctlr_info->ldev_critical;
170*4882a593Smuzhiyun ldev_offline = cs->ctlr_info->ldev_offline;
171*4882a593Smuzhiyun
172*4882a593Smuzhiyun ctlr_info_addr = dma_map_single(&cs->pdev->dev, cs->ctlr_info,
173*4882a593Smuzhiyun sizeof(struct myrs_ctlr_info),
174*4882a593Smuzhiyun DMA_FROM_DEVICE);
175*4882a593Smuzhiyun if (dma_mapping_error(&cs->pdev->dev, ctlr_info_addr))
176*4882a593Smuzhiyun return MYRS_STATUS_FAILED;
177*4882a593Smuzhiyun
178*4882a593Smuzhiyun mutex_lock(&cs->dcmd_mutex);
179*4882a593Smuzhiyun myrs_reset_cmd(cmd_blk);
180*4882a593Smuzhiyun mbox->ctlr_info.id = MYRS_DCMD_TAG;
181*4882a593Smuzhiyun mbox->ctlr_info.opcode = MYRS_CMD_OP_IOCTL;
182*4882a593Smuzhiyun mbox->ctlr_info.control.dma_ctrl_to_host = true;
183*4882a593Smuzhiyun mbox->ctlr_info.control.no_autosense = true;
184*4882a593Smuzhiyun mbox->ctlr_info.dma_size = sizeof(struct myrs_ctlr_info);
185*4882a593Smuzhiyun mbox->ctlr_info.ctlr_num = 0;
186*4882a593Smuzhiyun mbox->ctlr_info.ioctl_opcode = MYRS_IOCTL_GET_CTLR_INFO;
187*4882a593Smuzhiyun sgl = &mbox->ctlr_info.dma_addr;
188*4882a593Smuzhiyun sgl->sge[0].sge_addr = ctlr_info_addr;
189*4882a593Smuzhiyun sgl->sge[0].sge_count = mbox->ctlr_info.dma_size;
190*4882a593Smuzhiyun dev_dbg(&cs->host->shost_gendev, "Sending GetControllerInfo\n");
191*4882a593Smuzhiyun myrs_exec_cmd(cs, cmd_blk);
192*4882a593Smuzhiyun status = cmd_blk->status;
193*4882a593Smuzhiyun mutex_unlock(&cs->dcmd_mutex);
194*4882a593Smuzhiyun dma_unmap_single(&cs->pdev->dev, ctlr_info_addr,
195*4882a593Smuzhiyun sizeof(struct myrs_ctlr_info), DMA_FROM_DEVICE);
196*4882a593Smuzhiyun if (status == MYRS_STATUS_SUCCESS) {
197*4882a593Smuzhiyun if (cs->ctlr_info->bg_init_active +
198*4882a593Smuzhiyun cs->ctlr_info->ldev_init_active +
199*4882a593Smuzhiyun cs->ctlr_info->pdev_init_active +
200*4882a593Smuzhiyun cs->ctlr_info->cc_active +
201*4882a593Smuzhiyun cs->ctlr_info->rbld_active +
202*4882a593Smuzhiyun cs->ctlr_info->exp_active != 0)
203*4882a593Smuzhiyun cs->needs_update = true;
204*4882a593Smuzhiyun if (cs->ctlr_info->ldev_present != ldev_present ||
205*4882a593Smuzhiyun cs->ctlr_info->ldev_critical != ldev_critical ||
206*4882a593Smuzhiyun cs->ctlr_info->ldev_offline != ldev_offline)
207*4882a593Smuzhiyun shost_printk(KERN_INFO, cs->host,
208*4882a593Smuzhiyun "Logical drive count changes (%d/%d/%d)\n",
209*4882a593Smuzhiyun cs->ctlr_info->ldev_critical,
210*4882a593Smuzhiyun cs->ctlr_info->ldev_offline,
211*4882a593Smuzhiyun cs->ctlr_info->ldev_present);
212*4882a593Smuzhiyun }
213*4882a593Smuzhiyun
214*4882a593Smuzhiyun return status;
215*4882a593Smuzhiyun }
216*4882a593Smuzhiyun
217*4882a593Smuzhiyun /*
218*4882a593Smuzhiyun * myrs_get_ldev_info - executes a Logical Device Information IOCTL Command
219*4882a593Smuzhiyun */
myrs_get_ldev_info(struct myrs_hba * cs,unsigned short ldev_num,struct myrs_ldev_info * ldev_info)220*4882a593Smuzhiyun static unsigned char myrs_get_ldev_info(struct myrs_hba *cs,
221*4882a593Smuzhiyun unsigned short ldev_num, struct myrs_ldev_info *ldev_info)
222*4882a593Smuzhiyun {
223*4882a593Smuzhiyun struct myrs_cmdblk *cmd_blk = &cs->dcmd_blk;
224*4882a593Smuzhiyun union myrs_cmd_mbox *mbox = &cmd_blk->mbox;
225*4882a593Smuzhiyun dma_addr_t ldev_info_addr;
226*4882a593Smuzhiyun struct myrs_ldev_info ldev_info_orig;
227*4882a593Smuzhiyun union myrs_sgl *sgl;
228*4882a593Smuzhiyun unsigned char status;
229*4882a593Smuzhiyun
230*4882a593Smuzhiyun memcpy(&ldev_info_orig, ldev_info, sizeof(struct myrs_ldev_info));
231*4882a593Smuzhiyun ldev_info_addr = dma_map_single(&cs->pdev->dev, ldev_info,
232*4882a593Smuzhiyun sizeof(struct myrs_ldev_info),
233*4882a593Smuzhiyun DMA_FROM_DEVICE);
234*4882a593Smuzhiyun if (dma_mapping_error(&cs->pdev->dev, ldev_info_addr))
235*4882a593Smuzhiyun return MYRS_STATUS_FAILED;
236*4882a593Smuzhiyun
237*4882a593Smuzhiyun mutex_lock(&cs->dcmd_mutex);
238*4882a593Smuzhiyun myrs_reset_cmd(cmd_blk);
239*4882a593Smuzhiyun mbox->ldev_info.id = MYRS_DCMD_TAG;
240*4882a593Smuzhiyun mbox->ldev_info.opcode = MYRS_CMD_OP_IOCTL;
241*4882a593Smuzhiyun mbox->ldev_info.control.dma_ctrl_to_host = true;
242*4882a593Smuzhiyun mbox->ldev_info.control.no_autosense = true;
243*4882a593Smuzhiyun mbox->ldev_info.dma_size = sizeof(struct myrs_ldev_info);
244*4882a593Smuzhiyun mbox->ldev_info.ldev.ldev_num = ldev_num;
245*4882a593Smuzhiyun mbox->ldev_info.ioctl_opcode = MYRS_IOCTL_GET_LDEV_INFO_VALID;
246*4882a593Smuzhiyun sgl = &mbox->ldev_info.dma_addr;
247*4882a593Smuzhiyun sgl->sge[0].sge_addr = ldev_info_addr;
248*4882a593Smuzhiyun sgl->sge[0].sge_count = mbox->ldev_info.dma_size;
249*4882a593Smuzhiyun dev_dbg(&cs->host->shost_gendev,
250*4882a593Smuzhiyun "Sending GetLogicalDeviceInfoValid for ldev %d\n", ldev_num);
251*4882a593Smuzhiyun myrs_exec_cmd(cs, cmd_blk);
252*4882a593Smuzhiyun status = cmd_blk->status;
253*4882a593Smuzhiyun mutex_unlock(&cs->dcmd_mutex);
254*4882a593Smuzhiyun dma_unmap_single(&cs->pdev->dev, ldev_info_addr,
255*4882a593Smuzhiyun sizeof(struct myrs_ldev_info), DMA_FROM_DEVICE);
256*4882a593Smuzhiyun if (status == MYRS_STATUS_SUCCESS) {
257*4882a593Smuzhiyun unsigned short ldev_num = ldev_info->ldev_num;
258*4882a593Smuzhiyun struct myrs_ldev_info *new = ldev_info;
259*4882a593Smuzhiyun struct myrs_ldev_info *old = &ldev_info_orig;
260*4882a593Smuzhiyun unsigned long ldev_size = new->cfg_devsize;
261*4882a593Smuzhiyun
262*4882a593Smuzhiyun if (new->dev_state != old->dev_state) {
263*4882a593Smuzhiyun const char *name;
264*4882a593Smuzhiyun
265*4882a593Smuzhiyun name = myrs_devstate_name(new->dev_state);
266*4882a593Smuzhiyun shost_printk(KERN_INFO, cs->host,
267*4882a593Smuzhiyun "Logical Drive %d is now %s\n",
268*4882a593Smuzhiyun ldev_num, name ? name : "Invalid");
269*4882a593Smuzhiyun }
270*4882a593Smuzhiyun if ((new->soft_errs != old->soft_errs) ||
271*4882a593Smuzhiyun (new->cmds_failed != old->cmds_failed) ||
272*4882a593Smuzhiyun (new->deferred_write_errs != old->deferred_write_errs))
273*4882a593Smuzhiyun shost_printk(KERN_INFO, cs->host,
274*4882a593Smuzhiyun "Logical Drive %d Errors: Soft = %d, Failed = %d, Deferred Write = %d\n",
275*4882a593Smuzhiyun ldev_num, new->soft_errs,
276*4882a593Smuzhiyun new->cmds_failed,
277*4882a593Smuzhiyun new->deferred_write_errs);
278*4882a593Smuzhiyun if (new->bg_init_active)
279*4882a593Smuzhiyun myrs_report_progress(cs, ldev_num,
280*4882a593Smuzhiyun "Background Initialization",
281*4882a593Smuzhiyun new->bg_init_lba, ldev_size);
282*4882a593Smuzhiyun else if (new->fg_init_active)
283*4882a593Smuzhiyun myrs_report_progress(cs, ldev_num,
284*4882a593Smuzhiyun "Foreground Initialization",
285*4882a593Smuzhiyun new->fg_init_lba, ldev_size);
286*4882a593Smuzhiyun else if (new->migration_active)
287*4882a593Smuzhiyun myrs_report_progress(cs, ldev_num,
288*4882a593Smuzhiyun "Data Migration",
289*4882a593Smuzhiyun new->migration_lba, ldev_size);
290*4882a593Smuzhiyun else if (new->patrol_active)
291*4882a593Smuzhiyun myrs_report_progress(cs, ldev_num,
292*4882a593Smuzhiyun "Patrol Operation",
293*4882a593Smuzhiyun new->patrol_lba, ldev_size);
294*4882a593Smuzhiyun if (old->bg_init_active && !new->bg_init_active)
295*4882a593Smuzhiyun shost_printk(KERN_INFO, cs->host,
296*4882a593Smuzhiyun "Logical Drive %d: Background Initialization %s\n",
297*4882a593Smuzhiyun ldev_num,
298*4882a593Smuzhiyun (new->ldev_control.ldev_init_done ?
299*4882a593Smuzhiyun "Completed" : "Failed"));
300*4882a593Smuzhiyun }
301*4882a593Smuzhiyun return status;
302*4882a593Smuzhiyun }
303*4882a593Smuzhiyun
304*4882a593Smuzhiyun /*
305*4882a593Smuzhiyun * myrs_get_pdev_info - executes a "Read Physical Device Information" Command
306*4882a593Smuzhiyun */
myrs_get_pdev_info(struct myrs_hba * cs,unsigned char channel,unsigned char target,unsigned char lun,struct myrs_pdev_info * pdev_info)307*4882a593Smuzhiyun static unsigned char myrs_get_pdev_info(struct myrs_hba *cs,
308*4882a593Smuzhiyun unsigned char channel, unsigned char target, unsigned char lun,
309*4882a593Smuzhiyun struct myrs_pdev_info *pdev_info)
310*4882a593Smuzhiyun {
311*4882a593Smuzhiyun struct myrs_cmdblk *cmd_blk = &cs->dcmd_blk;
312*4882a593Smuzhiyun union myrs_cmd_mbox *mbox = &cmd_blk->mbox;
313*4882a593Smuzhiyun dma_addr_t pdev_info_addr;
314*4882a593Smuzhiyun union myrs_sgl *sgl;
315*4882a593Smuzhiyun unsigned char status;
316*4882a593Smuzhiyun
317*4882a593Smuzhiyun pdev_info_addr = dma_map_single(&cs->pdev->dev, pdev_info,
318*4882a593Smuzhiyun sizeof(struct myrs_pdev_info),
319*4882a593Smuzhiyun DMA_FROM_DEVICE);
320*4882a593Smuzhiyun if (dma_mapping_error(&cs->pdev->dev, pdev_info_addr))
321*4882a593Smuzhiyun return MYRS_STATUS_FAILED;
322*4882a593Smuzhiyun
323*4882a593Smuzhiyun mutex_lock(&cs->dcmd_mutex);
324*4882a593Smuzhiyun myrs_reset_cmd(cmd_blk);
325*4882a593Smuzhiyun mbox->pdev_info.opcode = MYRS_CMD_OP_IOCTL;
326*4882a593Smuzhiyun mbox->pdev_info.id = MYRS_DCMD_TAG;
327*4882a593Smuzhiyun mbox->pdev_info.control.dma_ctrl_to_host = true;
328*4882a593Smuzhiyun mbox->pdev_info.control.no_autosense = true;
329*4882a593Smuzhiyun mbox->pdev_info.dma_size = sizeof(struct myrs_pdev_info);
330*4882a593Smuzhiyun mbox->pdev_info.pdev.lun = lun;
331*4882a593Smuzhiyun mbox->pdev_info.pdev.target = target;
332*4882a593Smuzhiyun mbox->pdev_info.pdev.channel = channel;
333*4882a593Smuzhiyun mbox->pdev_info.ioctl_opcode = MYRS_IOCTL_GET_PDEV_INFO_VALID;
334*4882a593Smuzhiyun sgl = &mbox->pdev_info.dma_addr;
335*4882a593Smuzhiyun sgl->sge[0].sge_addr = pdev_info_addr;
336*4882a593Smuzhiyun sgl->sge[0].sge_count = mbox->pdev_info.dma_size;
337*4882a593Smuzhiyun dev_dbg(&cs->host->shost_gendev,
338*4882a593Smuzhiyun "Sending GetPhysicalDeviceInfoValid for pdev %d:%d:%d\n",
339*4882a593Smuzhiyun channel, target, lun);
340*4882a593Smuzhiyun myrs_exec_cmd(cs, cmd_blk);
341*4882a593Smuzhiyun status = cmd_blk->status;
342*4882a593Smuzhiyun mutex_unlock(&cs->dcmd_mutex);
343*4882a593Smuzhiyun dma_unmap_single(&cs->pdev->dev, pdev_info_addr,
344*4882a593Smuzhiyun sizeof(struct myrs_pdev_info), DMA_FROM_DEVICE);
345*4882a593Smuzhiyun return status;
346*4882a593Smuzhiyun }
347*4882a593Smuzhiyun
348*4882a593Smuzhiyun /*
349*4882a593Smuzhiyun * myrs_dev_op - executes a "Device Operation" Command
350*4882a593Smuzhiyun */
myrs_dev_op(struct myrs_hba * cs,enum myrs_ioctl_opcode opcode,enum myrs_opdev opdev)351*4882a593Smuzhiyun static unsigned char myrs_dev_op(struct myrs_hba *cs,
352*4882a593Smuzhiyun enum myrs_ioctl_opcode opcode, enum myrs_opdev opdev)
353*4882a593Smuzhiyun {
354*4882a593Smuzhiyun struct myrs_cmdblk *cmd_blk = &cs->dcmd_blk;
355*4882a593Smuzhiyun union myrs_cmd_mbox *mbox = &cmd_blk->mbox;
356*4882a593Smuzhiyun unsigned char status;
357*4882a593Smuzhiyun
358*4882a593Smuzhiyun mutex_lock(&cs->dcmd_mutex);
359*4882a593Smuzhiyun myrs_reset_cmd(cmd_blk);
360*4882a593Smuzhiyun mbox->dev_op.opcode = MYRS_CMD_OP_IOCTL;
361*4882a593Smuzhiyun mbox->dev_op.id = MYRS_DCMD_TAG;
362*4882a593Smuzhiyun mbox->dev_op.control.dma_ctrl_to_host = true;
363*4882a593Smuzhiyun mbox->dev_op.control.no_autosense = true;
364*4882a593Smuzhiyun mbox->dev_op.ioctl_opcode = opcode;
365*4882a593Smuzhiyun mbox->dev_op.opdev = opdev;
366*4882a593Smuzhiyun myrs_exec_cmd(cs, cmd_blk);
367*4882a593Smuzhiyun status = cmd_blk->status;
368*4882a593Smuzhiyun mutex_unlock(&cs->dcmd_mutex);
369*4882a593Smuzhiyun return status;
370*4882a593Smuzhiyun }
371*4882a593Smuzhiyun
372*4882a593Smuzhiyun /*
373*4882a593Smuzhiyun * myrs_translate_pdev - translates a Physical Device Channel and
374*4882a593Smuzhiyun * TargetID into a Logical Device.
375*4882a593Smuzhiyun */
myrs_translate_pdev(struct myrs_hba * cs,unsigned char channel,unsigned char target,unsigned char lun,struct myrs_devmap * devmap)376*4882a593Smuzhiyun static unsigned char myrs_translate_pdev(struct myrs_hba *cs,
377*4882a593Smuzhiyun unsigned char channel, unsigned char target, unsigned char lun,
378*4882a593Smuzhiyun struct myrs_devmap *devmap)
379*4882a593Smuzhiyun {
380*4882a593Smuzhiyun struct pci_dev *pdev = cs->pdev;
381*4882a593Smuzhiyun dma_addr_t devmap_addr;
382*4882a593Smuzhiyun struct myrs_cmdblk *cmd_blk;
383*4882a593Smuzhiyun union myrs_cmd_mbox *mbox;
384*4882a593Smuzhiyun union myrs_sgl *sgl;
385*4882a593Smuzhiyun unsigned char status;
386*4882a593Smuzhiyun
387*4882a593Smuzhiyun memset(devmap, 0x0, sizeof(struct myrs_devmap));
388*4882a593Smuzhiyun devmap_addr = dma_map_single(&pdev->dev, devmap,
389*4882a593Smuzhiyun sizeof(struct myrs_devmap),
390*4882a593Smuzhiyun DMA_FROM_DEVICE);
391*4882a593Smuzhiyun if (dma_mapping_error(&pdev->dev, devmap_addr))
392*4882a593Smuzhiyun return MYRS_STATUS_FAILED;
393*4882a593Smuzhiyun
394*4882a593Smuzhiyun mutex_lock(&cs->dcmd_mutex);
395*4882a593Smuzhiyun cmd_blk = &cs->dcmd_blk;
396*4882a593Smuzhiyun mbox = &cmd_blk->mbox;
397*4882a593Smuzhiyun mbox->pdev_info.opcode = MYRS_CMD_OP_IOCTL;
398*4882a593Smuzhiyun mbox->pdev_info.control.dma_ctrl_to_host = true;
399*4882a593Smuzhiyun mbox->pdev_info.control.no_autosense = true;
400*4882a593Smuzhiyun mbox->pdev_info.dma_size = sizeof(struct myrs_devmap);
401*4882a593Smuzhiyun mbox->pdev_info.pdev.target = target;
402*4882a593Smuzhiyun mbox->pdev_info.pdev.channel = channel;
403*4882a593Smuzhiyun mbox->pdev_info.pdev.lun = lun;
404*4882a593Smuzhiyun mbox->pdev_info.ioctl_opcode = MYRS_IOCTL_XLATE_PDEV_TO_LDEV;
405*4882a593Smuzhiyun sgl = &mbox->pdev_info.dma_addr;
406*4882a593Smuzhiyun sgl->sge[0].sge_addr = devmap_addr;
407*4882a593Smuzhiyun sgl->sge[0].sge_count = mbox->pdev_info.dma_size;
408*4882a593Smuzhiyun
409*4882a593Smuzhiyun myrs_exec_cmd(cs, cmd_blk);
410*4882a593Smuzhiyun status = cmd_blk->status;
411*4882a593Smuzhiyun mutex_unlock(&cs->dcmd_mutex);
412*4882a593Smuzhiyun dma_unmap_single(&pdev->dev, devmap_addr,
413*4882a593Smuzhiyun sizeof(struct myrs_devmap), DMA_FROM_DEVICE);
414*4882a593Smuzhiyun return status;
415*4882a593Smuzhiyun }
416*4882a593Smuzhiyun
417*4882a593Smuzhiyun /*
418*4882a593Smuzhiyun * myrs_get_event - executes a Get Event Command
419*4882a593Smuzhiyun */
myrs_get_event(struct myrs_hba * cs,unsigned int event_num,struct myrs_event * event_buf)420*4882a593Smuzhiyun static unsigned char myrs_get_event(struct myrs_hba *cs,
421*4882a593Smuzhiyun unsigned int event_num, struct myrs_event *event_buf)
422*4882a593Smuzhiyun {
423*4882a593Smuzhiyun struct pci_dev *pdev = cs->pdev;
424*4882a593Smuzhiyun dma_addr_t event_addr;
425*4882a593Smuzhiyun struct myrs_cmdblk *cmd_blk = &cs->mcmd_blk;
426*4882a593Smuzhiyun union myrs_cmd_mbox *mbox = &cmd_blk->mbox;
427*4882a593Smuzhiyun union myrs_sgl *sgl;
428*4882a593Smuzhiyun unsigned char status;
429*4882a593Smuzhiyun
430*4882a593Smuzhiyun event_addr = dma_map_single(&pdev->dev, event_buf,
431*4882a593Smuzhiyun sizeof(struct myrs_event), DMA_FROM_DEVICE);
432*4882a593Smuzhiyun if (dma_mapping_error(&pdev->dev, event_addr))
433*4882a593Smuzhiyun return MYRS_STATUS_FAILED;
434*4882a593Smuzhiyun
435*4882a593Smuzhiyun mbox->get_event.opcode = MYRS_CMD_OP_IOCTL;
436*4882a593Smuzhiyun mbox->get_event.dma_size = sizeof(struct myrs_event);
437*4882a593Smuzhiyun mbox->get_event.evnum_upper = event_num >> 16;
438*4882a593Smuzhiyun mbox->get_event.ctlr_num = 0;
439*4882a593Smuzhiyun mbox->get_event.ioctl_opcode = MYRS_IOCTL_GET_EVENT;
440*4882a593Smuzhiyun mbox->get_event.evnum_lower = event_num & 0xFFFF;
441*4882a593Smuzhiyun sgl = &mbox->get_event.dma_addr;
442*4882a593Smuzhiyun sgl->sge[0].sge_addr = event_addr;
443*4882a593Smuzhiyun sgl->sge[0].sge_count = mbox->get_event.dma_size;
444*4882a593Smuzhiyun myrs_exec_cmd(cs, cmd_blk);
445*4882a593Smuzhiyun status = cmd_blk->status;
446*4882a593Smuzhiyun dma_unmap_single(&pdev->dev, event_addr,
447*4882a593Smuzhiyun sizeof(struct myrs_event), DMA_FROM_DEVICE);
448*4882a593Smuzhiyun
449*4882a593Smuzhiyun return status;
450*4882a593Smuzhiyun }
451*4882a593Smuzhiyun
452*4882a593Smuzhiyun /*
453*4882a593Smuzhiyun * myrs_get_fwstatus - executes a Get Health Status Command
454*4882a593Smuzhiyun */
myrs_get_fwstatus(struct myrs_hba * cs)455*4882a593Smuzhiyun static unsigned char myrs_get_fwstatus(struct myrs_hba *cs)
456*4882a593Smuzhiyun {
457*4882a593Smuzhiyun struct myrs_cmdblk *cmd_blk = &cs->mcmd_blk;
458*4882a593Smuzhiyun union myrs_cmd_mbox *mbox = &cmd_blk->mbox;
459*4882a593Smuzhiyun union myrs_sgl *sgl;
460*4882a593Smuzhiyun unsigned char status = cmd_blk->status;
461*4882a593Smuzhiyun
462*4882a593Smuzhiyun myrs_reset_cmd(cmd_blk);
463*4882a593Smuzhiyun mbox->common.opcode = MYRS_CMD_OP_IOCTL;
464*4882a593Smuzhiyun mbox->common.id = MYRS_MCMD_TAG;
465*4882a593Smuzhiyun mbox->common.control.dma_ctrl_to_host = true;
466*4882a593Smuzhiyun mbox->common.control.no_autosense = true;
467*4882a593Smuzhiyun mbox->common.dma_size = sizeof(struct myrs_fwstat);
468*4882a593Smuzhiyun mbox->common.ioctl_opcode = MYRS_IOCTL_GET_HEALTH_STATUS;
469*4882a593Smuzhiyun sgl = &mbox->common.dma_addr;
470*4882a593Smuzhiyun sgl->sge[0].sge_addr = cs->fwstat_addr;
471*4882a593Smuzhiyun sgl->sge[0].sge_count = mbox->ctlr_info.dma_size;
472*4882a593Smuzhiyun dev_dbg(&cs->host->shost_gendev, "Sending GetHealthStatus\n");
473*4882a593Smuzhiyun myrs_exec_cmd(cs, cmd_blk);
474*4882a593Smuzhiyun status = cmd_blk->status;
475*4882a593Smuzhiyun
476*4882a593Smuzhiyun return status;
477*4882a593Smuzhiyun }
478*4882a593Smuzhiyun
479*4882a593Smuzhiyun /*
480*4882a593Smuzhiyun * myrs_enable_mmio_mbox - enables the Memory Mailbox Interface
481*4882a593Smuzhiyun */
myrs_enable_mmio_mbox(struct myrs_hba * cs,enable_mbox_t enable_mbox_fn)482*4882a593Smuzhiyun static bool myrs_enable_mmio_mbox(struct myrs_hba *cs,
483*4882a593Smuzhiyun enable_mbox_t enable_mbox_fn)
484*4882a593Smuzhiyun {
485*4882a593Smuzhiyun void __iomem *base = cs->io_base;
486*4882a593Smuzhiyun struct pci_dev *pdev = cs->pdev;
487*4882a593Smuzhiyun union myrs_cmd_mbox *cmd_mbox;
488*4882a593Smuzhiyun struct myrs_stat_mbox *stat_mbox;
489*4882a593Smuzhiyun union myrs_cmd_mbox *mbox;
490*4882a593Smuzhiyun dma_addr_t mbox_addr;
491*4882a593Smuzhiyun unsigned char status = MYRS_STATUS_FAILED;
492*4882a593Smuzhiyun
493*4882a593Smuzhiyun if (dma_set_mask(&pdev->dev, DMA_BIT_MASK(64)))
494*4882a593Smuzhiyun if (dma_set_mask(&pdev->dev, DMA_BIT_MASK(32))) {
495*4882a593Smuzhiyun dev_err(&pdev->dev, "DMA mask out of range\n");
496*4882a593Smuzhiyun return false;
497*4882a593Smuzhiyun }
498*4882a593Smuzhiyun
499*4882a593Smuzhiyun /* Temporary dma mapping, used only in the scope of this function */
500*4882a593Smuzhiyun mbox = dma_alloc_coherent(&pdev->dev, sizeof(union myrs_cmd_mbox),
501*4882a593Smuzhiyun &mbox_addr, GFP_KERNEL);
502*4882a593Smuzhiyun if (dma_mapping_error(&pdev->dev, mbox_addr))
503*4882a593Smuzhiyun return false;
504*4882a593Smuzhiyun
505*4882a593Smuzhiyun /* These are the base addresses for the command memory mailbox array */
506*4882a593Smuzhiyun cs->cmd_mbox_size = MYRS_MAX_CMD_MBOX * sizeof(union myrs_cmd_mbox);
507*4882a593Smuzhiyun cmd_mbox = dma_alloc_coherent(&pdev->dev, cs->cmd_mbox_size,
508*4882a593Smuzhiyun &cs->cmd_mbox_addr, GFP_KERNEL);
509*4882a593Smuzhiyun if (dma_mapping_error(&pdev->dev, cs->cmd_mbox_addr)) {
510*4882a593Smuzhiyun dev_err(&pdev->dev, "Failed to map command mailbox\n");
511*4882a593Smuzhiyun goto out_free;
512*4882a593Smuzhiyun }
513*4882a593Smuzhiyun cs->first_cmd_mbox = cmd_mbox;
514*4882a593Smuzhiyun cmd_mbox += MYRS_MAX_CMD_MBOX - 1;
515*4882a593Smuzhiyun cs->last_cmd_mbox = cmd_mbox;
516*4882a593Smuzhiyun cs->next_cmd_mbox = cs->first_cmd_mbox;
517*4882a593Smuzhiyun cs->prev_cmd_mbox1 = cs->last_cmd_mbox;
518*4882a593Smuzhiyun cs->prev_cmd_mbox2 = cs->last_cmd_mbox - 1;
519*4882a593Smuzhiyun
520*4882a593Smuzhiyun /* These are the base addresses for the status memory mailbox array */
521*4882a593Smuzhiyun cs->stat_mbox_size = MYRS_MAX_STAT_MBOX * sizeof(struct myrs_stat_mbox);
522*4882a593Smuzhiyun stat_mbox = dma_alloc_coherent(&pdev->dev, cs->stat_mbox_size,
523*4882a593Smuzhiyun &cs->stat_mbox_addr, GFP_KERNEL);
524*4882a593Smuzhiyun if (dma_mapping_error(&pdev->dev, cs->stat_mbox_addr)) {
525*4882a593Smuzhiyun dev_err(&pdev->dev, "Failed to map status mailbox\n");
526*4882a593Smuzhiyun goto out_free;
527*4882a593Smuzhiyun }
528*4882a593Smuzhiyun
529*4882a593Smuzhiyun cs->first_stat_mbox = stat_mbox;
530*4882a593Smuzhiyun stat_mbox += MYRS_MAX_STAT_MBOX - 1;
531*4882a593Smuzhiyun cs->last_stat_mbox = stat_mbox;
532*4882a593Smuzhiyun cs->next_stat_mbox = cs->first_stat_mbox;
533*4882a593Smuzhiyun
534*4882a593Smuzhiyun cs->fwstat_buf = dma_alloc_coherent(&pdev->dev,
535*4882a593Smuzhiyun sizeof(struct myrs_fwstat),
536*4882a593Smuzhiyun &cs->fwstat_addr, GFP_KERNEL);
537*4882a593Smuzhiyun if (dma_mapping_error(&pdev->dev, cs->fwstat_addr)) {
538*4882a593Smuzhiyun dev_err(&pdev->dev, "Failed to map firmware health buffer\n");
539*4882a593Smuzhiyun cs->fwstat_buf = NULL;
540*4882a593Smuzhiyun goto out_free;
541*4882a593Smuzhiyun }
542*4882a593Smuzhiyun cs->ctlr_info = kzalloc(sizeof(struct myrs_ctlr_info),
543*4882a593Smuzhiyun GFP_KERNEL | GFP_DMA);
544*4882a593Smuzhiyun if (!cs->ctlr_info)
545*4882a593Smuzhiyun goto out_free;
546*4882a593Smuzhiyun
547*4882a593Smuzhiyun cs->event_buf = kzalloc(sizeof(struct myrs_event),
548*4882a593Smuzhiyun GFP_KERNEL | GFP_DMA);
549*4882a593Smuzhiyun if (!cs->event_buf)
550*4882a593Smuzhiyun goto out_free;
551*4882a593Smuzhiyun
552*4882a593Smuzhiyun /* Enable the Memory Mailbox Interface. */
553*4882a593Smuzhiyun memset(mbox, 0, sizeof(union myrs_cmd_mbox));
554*4882a593Smuzhiyun mbox->set_mbox.id = 1;
555*4882a593Smuzhiyun mbox->set_mbox.opcode = MYRS_CMD_OP_IOCTL;
556*4882a593Smuzhiyun mbox->set_mbox.control.no_autosense = true;
557*4882a593Smuzhiyun mbox->set_mbox.first_cmd_mbox_size_kb =
558*4882a593Smuzhiyun (MYRS_MAX_CMD_MBOX * sizeof(union myrs_cmd_mbox)) >> 10;
559*4882a593Smuzhiyun mbox->set_mbox.first_stat_mbox_size_kb =
560*4882a593Smuzhiyun (MYRS_MAX_STAT_MBOX * sizeof(struct myrs_stat_mbox)) >> 10;
561*4882a593Smuzhiyun mbox->set_mbox.second_cmd_mbox_size_kb = 0;
562*4882a593Smuzhiyun mbox->set_mbox.second_stat_mbox_size_kb = 0;
563*4882a593Smuzhiyun mbox->set_mbox.sense_len = 0;
564*4882a593Smuzhiyun mbox->set_mbox.ioctl_opcode = MYRS_IOCTL_SET_MEM_MBOX;
565*4882a593Smuzhiyun mbox->set_mbox.fwstat_buf_size_kb = 1;
566*4882a593Smuzhiyun mbox->set_mbox.fwstat_buf_addr = cs->fwstat_addr;
567*4882a593Smuzhiyun mbox->set_mbox.first_cmd_mbox_addr = cs->cmd_mbox_addr;
568*4882a593Smuzhiyun mbox->set_mbox.first_stat_mbox_addr = cs->stat_mbox_addr;
569*4882a593Smuzhiyun status = enable_mbox_fn(base, mbox_addr);
570*4882a593Smuzhiyun
571*4882a593Smuzhiyun out_free:
572*4882a593Smuzhiyun dma_free_coherent(&pdev->dev, sizeof(union myrs_cmd_mbox),
573*4882a593Smuzhiyun mbox, mbox_addr);
574*4882a593Smuzhiyun if (status != MYRS_STATUS_SUCCESS)
575*4882a593Smuzhiyun dev_err(&pdev->dev, "Failed to enable mailbox, status %X\n",
576*4882a593Smuzhiyun status);
577*4882a593Smuzhiyun return (status == MYRS_STATUS_SUCCESS);
578*4882a593Smuzhiyun }
579*4882a593Smuzhiyun
580*4882a593Smuzhiyun /*
581*4882a593Smuzhiyun * myrs_get_config - reads the Configuration Information
582*4882a593Smuzhiyun */
myrs_get_config(struct myrs_hba * cs)583*4882a593Smuzhiyun static int myrs_get_config(struct myrs_hba *cs)
584*4882a593Smuzhiyun {
585*4882a593Smuzhiyun struct myrs_ctlr_info *info = cs->ctlr_info;
586*4882a593Smuzhiyun struct Scsi_Host *shost = cs->host;
587*4882a593Smuzhiyun unsigned char status;
588*4882a593Smuzhiyun unsigned char model[20];
589*4882a593Smuzhiyun unsigned char fw_version[12];
590*4882a593Smuzhiyun int i, model_len;
591*4882a593Smuzhiyun
592*4882a593Smuzhiyun /* Get data into dma-able area, then copy into permanent location */
593*4882a593Smuzhiyun mutex_lock(&cs->cinfo_mutex);
594*4882a593Smuzhiyun status = myrs_get_ctlr_info(cs);
595*4882a593Smuzhiyun mutex_unlock(&cs->cinfo_mutex);
596*4882a593Smuzhiyun if (status != MYRS_STATUS_SUCCESS) {
597*4882a593Smuzhiyun shost_printk(KERN_ERR, shost,
598*4882a593Smuzhiyun "Failed to get controller information\n");
599*4882a593Smuzhiyun return -ENODEV;
600*4882a593Smuzhiyun }
601*4882a593Smuzhiyun
602*4882a593Smuzhiyun /* Initialize the Controller Model Name and Full Model Name fields. */
603*4882a593Smuzhiyun model_len = sizeof(info->ctlr_name);
604*4882a593Smuzhiyun if (model_len > sizeof(model)-1)
605*4882a593Smuzhiyun model_len = sizeof(model)-1;
606*4882a593Smuzhiyun memcpy(model, info->ctlr_name, model_len);
607*4882a593Smuzhiyun model_len--;
608*4882a593Smuzhiyun while (model[model_len] == ' ' || model[model_len] == '\0')
609*4882a593Smuzhiyun model_len--;
610*4882a593Smuzhiyun model[++model_len] = '\0';
611*4882a593Smuzhiyun strcpy(cs->model_name, "DAC960 ");
612*4882a593Smuzhiyun strcat(cs->model_name, model);
613*4882a593Smuzhiyun /* Initialize the Controller Firmware Version field. */
614*4882a593Smuzhiyun sprintf(fw_version, "%d.%02d-%02d",
615*4882a593Smuzhiyun info->fw_major_version, info->fw_minor_version,
616*4882a593Smuzhiyun info->fw_turn_number);
617*4882a593Smuzhiyun if (info->fw_major_version == 6 &&
618*4882a593Smuzhiyun info->fw_minor_version == 0 &&
619*4882a593Smuzhiyun info->fw_turn_number < 1) {
620*4882a593Smuzhiyun shost_printk(KERN_WARNING, shost,
621*4882a593Smuzhiyun "FIRMWARE VERSION %s DOES NOT PROVIDE THE CONTROLLER\n"
622*4882a593Smuzhiyun "STATUS MONITORING FUNCTIONALITY NEEDED BY THIS DRIVER.\n"
623*4882a593Smuzhiyun "PLEASE UPGRADE TO VERSION 6.00-01 OR ABOVE.\n",
624*4882a593Smuzhiyun fw_version);
625*4882a593Smuzhiyun return -ENODEV;
626*4882a593Smuzhiyun }
627*4882a593Smuzhiyun /* Initialize the Controller Channels and Targets. */
628*4882a593Smuzhiyun shost->max_channel = info->physchan_present + info->virtchan_present;
629*4882a593Smuzhiyun shost->max_id = info->max_targets[0];
630*4882a593Smuzhiyun for (i = 1; i < 16; i++) {
631*4882a593Smuzhiyun if (!info->max_targets[i])
632*4882a593Smuzhiyun continue;
633*4882a593Smuzhiyun if (shost->max_id < info->max_targets[i])
634*4882a593Smuzhiyun shost->max_id = info->max_targets[i];
635*4882a593Smuzhiyun }
636*4882a593Smuzhiyun
637*4882a593Smuzhiyun /*
638*4882a593Smuzhiyun * Initialize the Controller Queue Depth, Driver Queue Depth,
639*4882a593Smuzhiyun * Logical Drive Count, Maximum Blocks per Command, Controller
640*4882a593Smuzhiyun * Scatter/Gather Limit, and Driver Scatter/Gather Limit.
641*4882a593Smuzhiyun * The Driver Queue Depth must be at most three less than
642*4882a593Smuzhiyun * the Controller Queue Depth; tag '1' is reserved for
643*4882a593Smuzhiyun * direct commands, and tag '2' for monitoring commands.
644*4882a593Smuzhiyun */
645*4882a593Smuzhiyun shost->can_queue = info->max_tcq - 3;
646*4882a593Smuzhiyun if (shost->can_queue > MYRS_MAX_CMD_MBOX - 3)
647*4882a593Smuzhiyun shost->can_queue = MYRS_MAX_CMD_MBOX - 3;
648*4882a593Smuzhiyun shost->max_sectors = info->max_transfer_size;
649*4882a593Smuzhiyun shost->sg_tablesize = info->max_sge;
650*4882a593Smuzhiyun if (shost->sg_tablesize > MYRS_SG_LIMIT)
651*4882a593Smuzhiyun shost->sg_tablesize = MYRS_SG_LIMIT;
652*4882a593Smuzhiyun
653*4882a593Smuzhiyun shost_printk(KERN_INFO, shost,
654*4882a593Smuzhiyun "Configuring %s PCI RAID Controller\n", model);
655*4882a593Smuzhiyun shost_printk(KERN_INFO, shost,
656*4882a593Smuzhiyun " Firmware Version: %s, Channels: %d, Memory Size: %dMB\n",
657*4882a593Smuzhiyun fw_version, info->physchan_present, info->mem_size_mb);
658*4882a593Smuzhiyun
659*4882a593Smuzhiyun shost_printk(KERN_INFO, shost,
660*4882a593Smuzhiyun " Controller Queue Depth: %d, Maximum Blocks per Command: %d\n",
661*4882a593Smuzhiyun shost->can_queue, shost->max_sectors);
662*4882a593Smuzhiyun
663*4882a593Smuzhiyun shost_printk(KERN_INFO, shost,
664*4882a593Smuzhiyun " Driver Queue Depth: %d, Scatter/Gather Limit: %d of %d Segments\n",
665*4882a593Smuzhiyun shost->can_queue, shost->sg_tablesize, MYRS_SG_LIMIT);
666*4882a593Smuzhiyun for (i = 0; i < info->physchan_max; i++) {
667*4882a593Smuzhiyun if (!info->max_targets[i])
668*4882a593Smuzhiyun continue;
669*4882a593Smuzhiyun shost_printk(KERN_INFO, shost,
670*4882a593Smuzhiyun " Device Channel %d: max %d devices\n",
671*4882a593Smuzhiyun i, info->max_targets[i]);
672*4882a593Smuzhiyun }
673*4882a593Smuzhiyun shost_printk(KERN_INFO, shost,
674*4882a593Smuzhiyun " Physical: %d/%d channels, %d disks, %d devices\n",
675*4882a593Smuzhiyun info->physchan_present, info->physchan_max,
676*4882a593Smuzhiyun info->pdisk_present, info->pdev_present);
677*4882a593Smuzhiyun
678*4882a593Smuzhiyun shost_printk(KERN_INFO, shost,
679*4882a593Smuzhiyun " Logical: %d/%d channels, %d disks\n",
680*4882a593Smuzhiyun info->virtchan_present, info->virtchan_max,
681*4882a593Smuzhiyun info->ldev_present);
682*4882a593Smuzhiyun return 0;
683*4882a593Smuzhiyun }
684*4882a593Smuzhiyun
685*4882a593Smuzhiyun /*
686*4882a593Smuzhiyun * myrs_log_event - prints a Controller Event message
687*4882a593Smuzhiyun */
688*4882a593Smuzhiyun static struct {
689*4882a593Smuzhiyun int ev_code;
690*4882a593Smuzhiyun unsigned char *ev_msg;
691*4882a593Smuzhiyun } myrs_ev_list[] = {
692*4882a593Smuzhiyun /* Physical Device Events (0x0000 - 0x007F) */
693*4882a593Smuzhiyun { 0x0001, "P Online" },
694*4882a593Smuzhiyun { 0x0002, "P Standby" },
695*4882a593Smuzhiyun { 0x0005, "P Automatic Rebuild Started" },
696*4882a593Smuzhiyun { 0x0006, "P Manual Rebuild Started" },
697*4882a593Smuzhiyun { 0x0007, "P Rebuild Completed" },
698*4882a593Smuzhiyun { 0x0008, "P Rebuild Cancelled" },
699*4882a593Smuzhiyun { 0x0009, "P Rebuild Failed for Unknown Reasons" },
700*4882a593Smuzhiyun { 0x000A, "P Rebuild Failed due to New Physical Device" },
701*4882a593Smuzhiyun { 0x000B, "P Rebuild Failed due to Logical Drive Failure" },
702*4882a593Smuzhiyun { 0x000C, "S Offline" },
703*4882a593Smuzhiyun { 0x000D, "P Found" },
704*4882a593Smuzhiyun { 0x000E, "P Removed" },
705*4882a593Smuzhiyun { 0x000F, "P Unconfigured" },
706*4882a593Smuzhiyun { 0x0010, "P Expand Capacity Started" },
707*4882a593Smuzhiyun { 0x0011, "P Expand Capacity Completed" },
708*4882a593Smuzhiyun { 0x0012, "P Expand Capacity Failed" },
709*4882a593Smuzhiyun { 0x0013, "P Command Timed Out" },
710*4882a593Smuzhiyun { 0x0014, "P Command Aborted" },
711*4882a593Smuzhiyun { 0x0015, "P Command Retried" },
712*4882a593Smuzhiyun { 0x0016, "P Parity Error" },
713*4882a593Smuzhiyun { 0x0017, "P Soft Error" },
714*4882a593Smuzhiyun { 0x0018, "P Miscellaneous Error" },
715*4882a593Smuzhiyun { 0x0019, "P Reset" },
716*4882a593Smuzhiyun { 0x001A, "P Active Spare Found" },
717*4882a593Smuzhiyun { 0x001B, "P Warm Spare Found" },
718*4882a593Smuzhiyun { 0x001C, "S Sense Data Received" },
719*4882a593Smuzhiyun { 0x001D, "P Initialization Started" },
720*4882a593Smuzhiyun { 0x001E, "P Initialization Completed" },
721*4882a593Smuzhiyun { 0x001F, "P Initialization Failed" },
722*4882a593Smuzhiyun { 0x0020, "P Initialization Cancelled" },
723*4882a593Smuzhiyun { 0x0021, "P Failed because Write Recovery Failed" },
724*4882a593Smuzhiyun { 0x0022, "P Failed because SCSI Bus Reset Failed" },
725*4882a593Smuzhiyun { 0x0023, "P Failed because of Double Check Condition" },
726*4882a593Smuzhiyun { 0x0024, "P Failed because Device Cannot Be Accessed" },
727*4882a593Smuzhiyun { 0x0025, "P Failed because of Gross Error on SCSI Processor" },
728*4882a593Smuzhiyun { 0x0026, "P Failed because of Bad Tag from Device" },
729*4882a593Smuzhiyun { 0x0027, "P Failed because of Command Timeout" },
730*4882a593Smuzhiyun { 0x0028, "P Failed because of System Reset" },
731*4882a593Smuzhiyun { 0x0029, "P Failed because of Busy Status or Parity Error" },
732*4882a593Smuzhiyun { 0x002A, "P Failed because Host Set Device to Failed State" },
733*4882a593Smuzhiyun { 0x002B, "P Failed because of Selection Timeout" },
734*4882a593Smuzhiyun { 0x002C, "P Failed because of SCSI Bus Phase Error" },
735*4882a593Smuzhiyun { 0x002D, "P Failed because Device Returned Unknown Status" },
736*4882a593Smuzhiyun { 0x002E, "P Failed because Device Not Ready" },
737*4882a593Smuzhiyun { 0x002F, "P Failed because Device Not Found at Startup" },
738*4882a593Smuzhiyun { 0x0030, "P Failed because COD Write Operation Failed" },
739*4882a593Smuzhiyun { 0x0031, "P Failed because BDT Write Operation Failed" },
740*4882a593Smuzhiyun { 0x0039, "P Missing at Startup" },
741*4882a593Smuzhiyun { 0x003A, "P Start Rebuild Failed due to Physical Drive Too Small" },
742*4882a593Smuzhiyun { 0x003C, "P Temporarily Offline Device Automatically Made Online" },
743*4882a593Smuzhiyun { 0x003D, "P Standby Rebuild Started" },
744*4882a593Smuzhiyun /* Logical Device Events (0x0080 - 0x00FF) */
745*4882a593Smuzhiyun { 0x0080, "M Consistency Check Started" },
746*4882a593Smuzhiyun { 0x0081, "M Consistency Check Completed" },
747*4882a593Smuzhiyun { 0x0082, "M Consistency Check Cancelled" },
748*4882a593Smuzhiyun { 0x0083, "M Consistency Check Completed With Errors" },
749*4882a593Smuzhiyun { 0x0084, "M Consistency Check Failed due to Logical Drive Failure" },
750*4882a593Smuzhiyun { 0x0085, "M Consistency Check Failed due to Physical Device Failure" },
751*4882a593Smuzhiyun { 0x0086, "L Offline" },
752*4882a593Smuzhiyun { 0x0087, "L Critical" },
753*4882a593Smuzhiyun { 0x0088, "L Online" },
754*4882a593Smuzhiyun { 0x0089, "M Automatic Rebuild Started" },
755*4882a593Smuzhiyun { 0x008A, "M Manual Rebuild Started" },
756*4882a593Smuzhiyun { 0x008B, "M Rebuild Completed" },
757*4882a593Smuzhiyun { 0x008C, "M Rebuild Cancelled" },
758*4882a593Smuzhiyun { 0x008D, "M Rebuild Failed for Unknown Reasons" },
759*4882a593Smuzhiyun { 0x008E, "M Rebuild Failed due to New Physical Device" },
760*4882a593Smuzhiyun { 0x008F, "M Rebuild Failed due to Logical Drive Failure" },
761*4882a593Smuzhiyun { 0x0090, "M Initialization Started" },
762*4882a593Smuzhiyun { 0x0091, "M Initialization Completed" },
763*4882a593Smuzhiyun { 0x0092, "M Initialization Cancelled" },
764*4882a593Smuzhiyun { 0x0093, "M Initialization Failed" },
765*4882a593Smuzhiyun { 0x0094, "L Found" },
766*4882a593Smuzhiyun { 0x0095, "L Deleted" },
767*4882a593Smuzhiyun { 0x0096, "M Expand Capacity Started" },
768*4882a593Smuzhiyun { 0x0097, "M Expand Capacity Completed" },
769*4882a593Smuzhiyun { 0x0098, "M Expand Capacity Failed" },
770*4882a593Smuzhiyun { 0x0099, "L Bad Block Found" },
771*4882a593Smuzhiyun { 0x009A, "L Size Changed" },
772*4882a593Smuzhiyun { 0x009B, "L Type Changed" },
773*4882a593Smuzhiyun { 0x009C, "L Bad Data Block Found" },
774*4882a593Smuzhiyun { 0x009E, "L Read of Data Block in BDT" },
775*4882a593Smuzhiyun { 0x009F, "L Write Back Data for Disk Block Lost" },
776*4882a593Smuzhiyun { 0x00A0, "L Temporarily Offline RAID-5/3 Drive Made Online" },
777*4882a593Smuzhiyun { 0x00A1, "L Temporarily Offline RAID-6/1/0/7 Drive Made Online" },
778*4882a593Smuzhiyun { 0x00A2, "L Standby Rebuild Started" },
779*4882a593Smuzhiyun /* Fault Management Events (0x0100 - 0x017F) */
780*4882a593Smuzhiyun { 0x0140, "E Fan %d Failed" },
781*4882a593Smuzhiyun { 0x0141, "E Fan %d OK" },
782*4882a593Smuzhiyun { 0x0142, "E Fan %d Not Present" },
783*4882a593Smuzhiyun { 0x0143, "E Power Supply %d Failed" },
784*4882a593Smuzhiyun { 0x0144, "E Power Supply %d OK" },
785*4882a593Smuzhiyun { 0x0145, "E Power Supply %d Not Present" },
786*4882a593Smuzhiyun { 0x0146, "E Temperature Sensor %d Temperature Exceeds Safe Limit" },
787*4882a593Smuzhiyun { 0x0147, "E Temperature Sensor %d Temperature Exceeds Working Limit" },
788*4882a593Smuzhiyun { 0x0148, "E Temperature Sensor %d Temperature Normal" },
789*4882a593Smuzhiyun { 0x0149, "E Temperature Sensor %d Not Present" },
790*4882a593Smuzhiyun { 0x014A, "E Enclosure Management Unit %d Access Critical" },
791*4882a593Smuzhiyun { 0x014B, "E Enclosure Management Unit %d Access OK" },
792*4882a593Smuzhiyun { 0x014C, "E Enclosure Management Unit %d Access Offline" },
793*4882a593Smuzhiyun /* Controller Events (0x0180 - 0x01FF) */
794*4882a593Smuzhiyun { 0x0181, "C Cache Write Back Error" },
795*4882a593Smuzhiyun { 0x0188, "C Battery Backup Unit Found" },
796*4882a593Smuzhiyun { 0x0189, "C Battery Backup Unit Charge Level Low" },
797*4882a593Smuzhiyun { 0x018A, "C Battery Backup Unit Charge Level OK" },
798*4882a593Smuzhiyun { 0x0193, "C Installation Aborted" },
799*4882a593Smuzhiyun { 0x0195, "C Battery Backup Unit Physically Removed" },
800*4882a593Smuzhiyun { 0x0196, "C Memory Error During Warm Boot" },
801*4882a593Smuzhiyun { 0x019E, "C Memory Soft ECC Error Corrected" },
802*4882a593Smuzhiyun { 0x019F, "C Memory Hard ECC Error Corrected" },
803*4882a593Smuzhiyun { 0x01A2, "C Battery Backup Unit Failed" },
804*4882a593Smuzhiyun { 0x01AB, "C Mirror Race Recovery Failed" },
805*4882a593Smuzhiyun { 0x01AC, "C Mirror Race on Critical Drive" },
806*4882a593Smuzhiyun /* Controller Internal Processor Events */
807*4882a593Smuzhiyun { 0x0380, "C Internal Controller Hung" },
808*4882a593Smuzhiyun { 0x0381, "C Internal Controller Firmware Breakpoint" },
809*4882a593Smuzhiyun { 0x0390, "C Internal Controller i960 Processor Specific Error" },
810*4882a593Smuzhiyun { 0x03A0, "C Internal Controller StrongARM Processor Specific Error" },
811*4882a593Smuzhiyun { 0, "" }
812*4882a593Smuzhiyun };
813*4882a593Smuzhiyun
myrs_log_event(struct myrs_hba * cs,struct myrs_event * ev)814*4882a593Smuzhiyun static void myrs_log_event(struct myrs_hba *cs, struct myrs_event *ev)
815*4882a593Smuzhiyun {
816*4882a593Smuzhiyun unsigned char msg_buf[MYRS_LINE_BUFFER_SIZE];
817*4882a593Smuzhiyun int ev_idx = 0, ev_code;
818*4882a593Smuzhiyun unsigned char ev_type, *ev_msg;
819*4882a593Smuzhiyun struct Scsi_Host *shost = cs->host;
820*4882a593Smuzhiyun struct scsi_device *sdev;
821*4882a593Smuzhiyun struct scsi_sense_hdr sshdr = {0};
822*4882a593Smuzhiyun unsigned char sense_info[4];
823*4882a593Smuzhiyun unsigned char cmd_specific[4];
824*4882a593Smuzhiyun
825*4882a593Smuzhiyun if (ev->ev_code == 0x1C) {
826*4882a593Smuzhiyun if (!scsi_normalize_sense(ev->sense_data, 40, &sshdr)) {
827*4882a593Smuzhiyun memset(&sshdr, 0x0, sizeof(sshdr));
828*4882a593Smuzhiyun memset(sense_info, 0x0, sizeof(sense_info));
829*4882a593Smuzhiyun memset(cmd_specific, 0x0, sizeof(cmd_specific));
830*4882a593Smuzhiyun } else {
831*4882a593Smuzhiyun memcpy(sense_info, &ev->sense_data[3], 4);
832*4882a593Smuzhiyun memcpy(cmd_specific, &ev->sense_data[7], 4);
833*4882a593Smuzhiyun }
834*4882a593Smuzhiyun }
835*4882a593Smuzhiyun if (sshdr.sense_key == VENDOR_SPECIFIC &&
836*4882a593Smuzhiyun (sshdr.asc == 0x80 || sshdr.asc == 0x81))
837*4882a593Smuzhiyun ev->ev_code = ((sshdr.asc - 0x80) << 8 | sshdr.ascq);
838*4882a593Smuzhiyun while (true) {
839*4882a593Smuzhiyun ev_code = myrs_ev_list[ev_idx].ev_code;
840*4882a593Smuzhiyun if (ev_code == ev->ev_code || ev_code == 0)
841*4882a593Smuzhiyun break;
842*4882a593Smuzhiyun ev_idx++;
843*4882a593Smuzhiyun }
844*4882a593Smuzhiyun ev_type = myrs_ev_list[ev_idx].ev_msg[0];
845*4882a593Smuzhiyun ev_msg = &myrs_ev_list[ev_idx].ev_msg[2];
846*4882a593Smuzhiyun if (ev_code == 0) {
847*4882a593Smuzhiyun shost_printk(KERN_WARNING, shost,
848*4882a593Smuzhiyun "Unknown Controller Event Code %04X\n",
849*4882a593Smuzhiyun ev->ev_code);
850*4882a593Smuzhiyun return;
851*4882a593Smuzhiyun }
852*4882a593Smuzhiyun switch (ev_type) {
853*4882a593Smuzhiyun case 'P':
854*4882a593Smuzhiyun sdev = scsi_device_lookup(shost, ev->channel,
855*4882a593Smuzhiyun ev->target, 0);
856*4882a593Smuzhiyun sdev_printk(KERN_INFO, sdev, "event %d: Physical Device %s\n",
857*4882a593Smuzhiyun ev->ev_seq, ev_msg);
858*4882a593Smuzhiyun if (sdev && sdev->hostdata &&
859*4882a593Smuzhiyun sdev->channel < cs->ctlr_info->physchan_present) {
860*4882a593Smuzhiyun struct myrs_pdev_info *pdev_info = sdev->hostdata;
861*4882a593Smuzhiyun
862*4882a593Smuzhiyun switch (ev->ev_code) {
863*4882a593Smuzhiyun case 0x0001:
864*4882a593Smuzhiyun case 0x0007:
865*4882a593Smuzhiyun pdev_info->dev_state = MYRS_DEVICE_ONLINE;
866*4882a593Smuzhiyun break;
867*4882a593Smuzhiyun case 0x0002:
868*4882a593Smuzhiyun pdev_info->dev_state = MYRS_DEVICE_STANDBY;
869*4882a593Smuzhiyun break;
870*4882a593Smuzhiyun case 0x000C:
871*4882a593Smuzhiyun pdev_info->dev_state = MYRS_DEVICE_OFFLINE;
872*4882a593Smuzhiyun break;
873*4882a593Smuzhiyun case 0x000E:
874*4882a593Smuzhiyun pdev_info->dev_state = MYRS_DEVICE_MISSING;
875*4882a593Smuzhiyun break;
876*4882a593Smuzhiyun case 0x000F:
877*4882a593Smuzhiyun pdev_info->dev_state = MYRS_DEVICE_UNCONFIGURED;
878*4882a593Smuzhiyun break;
879*4882a593Smuzhiyun }
880*4882a593Smuzhiyun }
881*4882a593Smuzhiyun break;
882*4882a593Smuzhiyun case 'L':
883*4882a593Smuzhiyun shost_printk(KERN_INFO, shost,
884*4882a593Smuzhiyun "event %d: Logical Drive %d %s\n",
885*4882a593Smuzhiyun ev->ev_seq, ev->lun, ev_msg);
886*4882a593Smuzhiyun cs->needs_update = true;
887*4882a593Smuzhiyun break;
888*4882a593Smuzhiyun case 'M':
889*4882a593Smuzhiyun shost_printk(KERN_INFO, shost,
890*4882a593Smuzhiyun "event %d: Logical Drive %d %s\n",
891*4882a593Smuzhiyun ev->ev_seq, ev->lun, ev_msg);
892*4882a593Smuzhiyun cs->needs_update = true;
893*4882a593Smuzhiyun break;
894*4882a593Smuzhiyun case 'S':
895*4882a593Smuzhiyun if (sshdr.sense_key == NO_SENSE ||
896*4882a593Smuzhiyun (sshdr.sense_key == NOT_READY &&
897*4882a593Smuzhiyun sshdr.asc == 0x04 && (sshdr.ascq == 0x01 ||
898*4882a593Smuzhiyun sshdr.ascq == 0x02)))
899*4882a593Smuzhiyun break;
900*4882a593Smuzhiyun shost_printk(KERN_INFO, shost,
901*4882a593Smuzhiyun "event %d: Physical Device %d:%d %s\n",
902*4882a593Smuzhiyun ev->ev_seq, ev->channel, ev->target, ev_msg);
903*4882a593Smuzhiyun shost_printk(KERN_INFO, shost,
904*4882a593Smuzhiyun "Physical Device %d:%d Sense Key = %X, ASC = %02X, ASCQ = %02X\n",
905*4882a593Smuzhiyun ev->channel, ev->target,
906*4882a593Smuzhiyun sshdr.sense_key, sshdr.asc, sshdr.ascq);
907*4882a593Smuzhiyun shost_printk(KERN_INFO, shost,
908*4882a593Smuzhiyun "Physical Device %d:%d Sense Information = %02X%02X%02X%02X %02X%02X%02X%02X\n",
909*4882a593Smuzhiyun ev->channel, ev->target,
910*4882a593Smuzhiyun sense_info[0], sense_info[1],
911*4882a593Smuzhiyun sense_info[2], sense_info[3],
912*4882a593Smuzhiyun cmd_specific[0], cmd_specific[1],
913*4882a593Smuzhiyun cmd_specific[2], cmd_specific[3]);
914*4882a593Smuzhiyun break;
915*4882a593Smuzhiyun case 'E':
916*4882a593Smuzhiyun if (cs->disable_enc_msg)
917*4882a593Smuzhiyun break;
918*4882a593Smuzhiyun sprintf(msg_buf, ev_msg, ev->lun);
919*4882a593Smuzhiyun shost_printk(KERN_INFO, shost, "event %d: Enclosure %d %s\n",
920*4882a593Smuzhiyun ev->ev_seq, ev->target, msg_buf);
921*4882a593Smuzhiyun break;
922*4882a593Smuzhiyun case 'C':
923*4882a593Smuzhiyun shost_printk(KERN_INFO, shost, "event %d: Controller %s\n",
924*4882a593Smuzhiyun ev->ev_seq, ev_msg);
925*4882a593Smuzhiyun break;
926*4882a593Smuzhiyun default:
927*4882a593Smuzhiyun shost_printk(KERN_INFO, shost,
928*4882a593Smuzhiyun "event %d: Unknown Event Code %04X\n",
929*4882a593Smuzhiyun ev->ev_seq, ev->ev_code);
930*4882a593Smuzhiyun break;
931*4882a593Smuzhiyun }
932*4882a593Smuzhiyun }
933*4882a593Smuzhiyun
934*4882a593Smuzhiyun /*
935*4882a593Smuzhiyun * SCSI sysfs interface functions
936*4882a593Smuzhiyun */
raid_state_show(struct device * dev,struct device_attribute * attr,char * buf)937*4882a593Smuzhiyun static ssize_t raid_state_show(struct device *dev,
938*4882a593Smuzhiyun struct device_attribute *attr, char *buf)
939*4882a593Smuzhiyun {
940*4882a593Smuzhiyun struct scsi_device *sdev = to_scsi_device(dev);
941*4882a593Smuzhiyun struct myrs_hba *cs = shost_priv(sdev->host);
942*4882a593Smuzhiyun int ret;
943*4882a593Smuzhiyun
944*4882a593Smuzhiyun if (!sdev->hostdata)
945*4882a593Smuzhiyun return snprintf(buf, 16, "Unknown\n");
946*4882a593Smuzhiyun
947*4882a593Smuzhiyun if (sdev->channel >= cs->ctlr_info->physchan_present) {
948*4882a593Smuzhiyun struct myrs_ldev_info *ldev_info = sdev->hostdata;
949*4882a593Smuzhiyun const char *name;
950*4882a593Smuzhiyun
951*4882a593Smuzhiyun name = myrs_devstate_name(ldev_info->dev_state);
952*4882a593Smuzhiyun if (name)
953*4882a593Smuzhiyun ret = snprintf(buf, 32, "%s\n", name);
954*4882a593Smuzhiyun else
955*4882a593Smuzhiyun ret = snprintf(buf, 32, "Invalid (%02X)\n",
956*4882a593Smuzhiyun ldev_info->dev_state);
957*4882a593Smuzhiyun } else {
958*4882a593Smuzhiyun struct myrs_pdev_info *pdev_info;
959*4882a593Smuzhiyun const char *name;
960*4882a593Smuzhiyun
961*4882a593Smuzhiyun pdev_info = sdev->hostdata;
962*4882a593Smuzhiyun name = myrs_devstate_name(pdev_info->dev_state);
963*4882a593Smuzhiyun if (name)
964*4882a593Smuzhiyun ret = snprintf(buf, 32, "%s\n", name);
965*4882a593Smuzhiyun else
966*4882a593Smuzhiyun ret = snprintf(buf, 32, "Invalid (%02X)\n",
967*4882a593Smuzhiyun pdev_info->dev_state);
968*4882a593Smuzhiyun }
969*4882a593Smuzhiyun return ret;
970*4882a593Smuzhiyun }
971*4882a593Smuzhiyun
raid_state_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)972*4882a593Smuzhiyun static ssize_t raid_state_store(struct device *dev,
973*4882a593Smuzhiyun struct device_attribute *attr, const char *buf, size_t count)
974*4882a593Smuzhiyun {
975*4882a593Smuzhiyun struct scsi_device *sdev = to_scsi_device(dev);
976*4882a593Smuzhiyun struct myrs_hba *cs = shost_priv(sdev->host);
977*4882a593Smuzhiyun struct myrs_cmdblk *cmd_blk;
978*4882a593Smuzhiyun union myrs_cmd_mbox *mbox;
979*4882a593Smuzhiyun enum myrs_devstate new_state;
980*4882a593Smuzhiyun unsigned short ldev_num;
981*4882a593Smuzhiyun unsigned char status;
982*4882a593Smuzhiyun
983*4882a593Smuzhiyun if (!strncmp(buf, "offline", 7) ||
984*4882a593Smuzhiyun !strncmp(buf, "kill", 4))
985*4882a593Smuzhiyun new_state = MYRS_DEVICE_OFFLINE;
986*4882a593Smuzhiyun else if (!strncmp(buf, "online", 6))
987*4882a593Smuzhiyun new_state = MYRS_DEVICE_ONLINE;
988*4882a593Smuzhiyun else if (!strncmp(buf, "standby", 7))
989*4882a593Smuzhiyun new_state = MYRS_DEVICE_STANDBY;
990*4882a593Smuzhiyun else
991*4882a593Smuzhiyun return -EINVAL;
992*4882a593Smuzhiyun
993*4882a593Smuzhiyun if (sdev->channel < cs->ctlr_info->physchan_present) {
994*4882a593Smuzhiyun struct myrs_pdev_info *pdev_info = sdev->hostdata;
995*4882a593Smuzhiyun struct myrs_devmap *pdev_devmap =
996*4882a593Smuzhiyun (struct myrs_devmap *)&pdev_info->rsvd13;
997*4882a593Smuzhiyun
998*4882a593Smuzhiyun if (pdev_info->dev_state == new_state) {
999*4882a593Smuzhiyun sdev_printk(KERN_INFO, sdev,
1000*4882a593Smuzhiyun "Device already in %s\n",
1001*4882a593Smuzhiyun myrs_devstate_name(new_state));
1002*4882a593Smuzhiyun return count;
1003*4882a593Smuzhiyun }
1004*4882a593Smuzhiyun status = myrs_translate_pdev(cs, sdev->channel, sdev->id,
1005*4882a593Smuzhiyun sdev->lun, pdev_devmap);
1006*4882a593Smuzhiyun if (status != MYRS_STATUS_SUCCESS)
1007*4882a593Smuzhiyun return -ENXIO;
1008*4882a593Smuzhiyun ldev_num = pdev_devmap->ldev_num;
1009*4882a593Smuzhiyun } else {
1010*4882a593Smuzhiyun struct myrs_ldev_info *ldev_info = sdev->hostdata;
1011*4882a593Smuzhiyun
1012*4882a593Smuzhiyun if (ldev_info->dev_state == new_state) {
1013*4882a593Smuzhiyun sdev_printk(KERN_INFO, sdev,
1014*4882a593Smuzhiyun "Device already in %s\n",
1015*4882a593Smuzhiyun myrs_devstate_name(new_state));
1016*4882a593Smuzhiyun return count;
1017*4882a593Smuzhiyun }
1018*4882a593Smuzhiyun ldev_num = ldev_info->ldev_num;
1019*4882a593Smuzhiyun }
1020*4882a593Smuzhiyun mutex_lock(&cs->dcmd_mutex);
1021*4882a593Smuzhiyun cmd_blk = &cs->dcmd_blk;
1022*4882a593Smuzhiyun myrs_reset_cmd(cmd_blk);
1023*4882a593Smuzhiyun mbox = &cmd_blk->mbox;
1024*4882a593Smuzhiyun mbox->common.opcode = MYRS_CMD_OP_IOCTL;
1025*4882a593Smuzhiyun mbox->common.id = MYRS_DCMD_TAG;
1026*4882a593Smuzhiyun mbox->common.control.dma_ctrl_to_host = true;
1027*4882a593Smuzhiyun mbox->common.control.no_autosense = true;
1028*4882a593Smuzhiyun mbox->set_devstate.ioctl_opcode = MYRS_IOCTL_SET_DEVICE_STATE;
1029*4882a593Smuzhiyun mbox->set_devstate.state = new_state;
1030*4882a593Smuzhiyun mbox->set_devstate.ldev.ldev_num = ldev_num;
1031*4882a593Smuzhiyun myrs_exec_cmd(cs, cmd_blk);
1032*4882a593Smuzhiyun status = cmd_blk->status;
1033*4882a593Smuzhiyun mutex_unlock(&cs->dcmd_mutex);
1034*4882a593Smuzhiyun if (status == MYRS_STATUS_SUCCESS) {
1035*4882a593Smuzhiyun if (sdev->channel < cs->ctlr_info->physchan_present) {
1036*4882a593Smuzhiyun struct myrs_pdev_info *pdev_info = sdev->hostdata;
1037*4882a593Smuzhiyun
1038*4882a593Smuzhiyun pdev_info->dev_state = new_state;
1039*4882a593Smuzhiyun } else {
1040*4882a593Smuzhiyun struct myrs_ldev_info *ldev_info = sdev->hostdata;
1041*4882a593Smuzhiyun
1042*4882a593Smuzhiyun ldev_info->dev_state = new_state;
1043*4882a593Smuzhiyun }
1044*4882a593Smuzhiyun sdev_printk(KERN_INFO, sdev,
1045*4882a593Smuzhiyun "Set device state to %s\n",
1046*4882a593Smuzhiyun myrs_devstate_name(new_state));
1047*4882a593Smuzhiyun return count;
1048*4882a593Smuzhiyun }
1049*4882a593Smuzhiyun sdev_printk(KERN_INFO, sdev,
1050*4882a593Smuzhiyun "Failed to set device state to %s, status 0x%02x\n",
1051*4882a593Smuzhiyun myrs_devstate_name(new_state), status);
1052*4882a593Smuzhiyun return -EINVAL;
1053*4882a593Smuzhiyun }
1054*4882a593Smuzhiyun static DEVICE_ATTR_RW(raid_state);
1055*4882a593Smuzhiyun
raid_level_show(struct device * dev,struct device_attribute * attr,char * buf)1056*4882a593Smuzhiyun static ssize_t raid_level_show(struct device *dev,
1057*4882a593Smuzhiyun struct device_attribute *attr, char *buf)
1058*4882a593Smuzhiyun {
1059*4882a593Smuzhiyun struct scsi_device *sdev = to_scsi_device(dev);
1060*4882a593Smuzhiyun struct myrs_hba *cs = shost_priv(sdev->host);
1061*4882a593Smuzhiyun const char *name = NULL;
1062*4882a593Smuzhiyun
1063*4882a593Smuzhiyun if (!sdev->hostdata)
1064*4882a593Smuzhiyun return snprintf(buf, 16, "Unknown\n");
1065*4882a593Smuzhiyun
1066*4882a593Smuzhiyun if (sdev->channel >= cs->ctlr_info->physchan_present) {
1067*4882a593Smuzhiyun struct myrs_ldev_info *ldev_info;
1068*4882a593Smuzhiyun
1069*4882a593Smuzhiyun ldev_info = sdev->hostdata;
1070*4882a593Smuzhiyun name = myrs_raid_level_name(ldev_info->raid_level);
1071*4882a593Smuzhiyun if (!name)
1072*4882a593Smuzhiyun return snprintf(buf, 32, "Invalid (%02X)\n",
1073*4882a593Smuzhiyun ldev_info->dev_state);
1074*4882a593Smuzhiyun
1075*4882a593Smuzhiyun } else
1076*4882a593Smuzhiyun name = myrs_raid_level_name(MYRS_RAID_PHYSICAL);
1077*4882a593Smuzhiyun
1078*4882a593Smuzhiyun return snprintf(buf, 32, "%s\n", name);
1079*4882a593Smuzhiyun }
1080*4882a593Smuzhiyun static DEVICE_ATTR_RO(raid_level);
1081*4882a593Smuzhiyun
rebuild_show(struct device * dev,struct device_attribute * attr,char * buf)1082*4882a593Smuzhiyun static ssize_t rebuild_show(struct device *dev,
1083*4882a593Smuzhiyun struct device_attribute *attr, char *buf)
1084*4882a593Smuzhiyun {
1085*4882a593Smuzhiyun struct scsi_device *sdev = to_scsi_device(dev);
1086*4882a593Smuzhiyun struct myrs_hba *cs = shost_priv(sdev->host);
1087*4882a593Smuzhiyun struct myrs_ldev_info *ldev_info;
1088*4882a593Smuzhiyun unsigned short ldev_num;
1089*4882a593Smuzhiyun unsigned char status;
1090*4882a593Smuzhiyun
1091*4882a593Smuzhiyun if (sdev->channel < cs->ctlr_info->physchan_present)
1092*4882a593Smuzhiyun return snprintf(buf, 32, "physical device - not rebuilding\n");
1093*4882a593Smuzhiyun
1094*4882a593Smuzhiyun ldev_info = sdev->hostdata;
1095*4882a593Smuzhiyun ldev_num = ldev_info->ldev_num;
1096*4882a593Smuzhiyun status = myrs_get_ldev_info(cs, ldev_num, ldev_info);
1097*4882a593Smuzhiyun if (status != MYRS_STATUS_SUCCESS) {
1098*4882a593Smuzhiyun sdev_printk(KERN_INFO, sdev,
1099*4882a593Smuzhiyun "Failed to get device information, status 0x%02x\n",
1100*4882a593Smuzhiyun status);
1101*4882a593Smuzhiyun return -EIO;
1102*4882a593Smuzhiyun }
1103*4882a593Smuzhiyun if (ldev_info->rbld_active) {
1104*4882a593Smuzhiyun return snprintf(buf, 32, "rebuilding block %zu of %zu\n",
1105*4882a593Smuzhiyun (size_t)ldev_info->rbld_lba,
1106*4882a593Smuzhiyun (size_t)ldev_info->cfg_devsize);
1107*4882a593Smuzhiyun } else
1108*4882a593Smuzhiyun return snprintf(buf, 32, "not rebuilding\n");
1109*4882a593Smuzhiyun }
1110*4882a593Smuzhiyun
rebuild_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)1111*4882a593Smuzhiyun static ssize_t rebuild_store(struct device *dev,
1112*4882a593Smuzhiyun struct device_attribute *attr, const char *buf, size_t count)
1113*4882a593Smuzhiyun {
1114*4882a593Smuzhiyun struct scsi_device *sdev = to_scsi_device(dev);
1115*4882a593Smuzhiyun struct myrs_hba *cs = shost_priv(sdev->host);
1116*4882a593Smuzhiyun struct myrs_ldev_info *ldev_info;
1117*4882a593Smuzhiyun struct myrs_cmdblk *cmd_blk;
1118*4882a593Smuzhiyun union myrs_cmd_mbox *mbox;
1119*4882a593Smuzhiyun unsigned short ldev_num;
1120*4882a593Smuzhiyun unsigned char status;
1121*4882a593Smuzhiyun int rebuild, ret;
1122*4882a593Smuzhiyun
1123*4882a593Smuzhiyun if (sdev->channel < cs->ctlr_info->physchan_present)
1124*4882a593Smuzhiyun return -EINVAL;
1125*4882a593Smuzhiyun
1126*4882a593Smuzhiyun ldev_info = sdev->hostdata;
1127*4882a593Smuzhiyun if (!ldev_info)
1128*4882a593Smuzhiyun return -ENXIO;
1129*4882a593Smuzhiyun ldev_num = ldev_info->ldev_num;
1130*4882a593Smuzhiyun
1131*4882a593Smuzhiyun ret = kstrtoint(buf, 0, &rebuild);
1132*4882a593Smuzhiyun if (ret)
1133*4882a593Smuzhiyun return ret;
1134*4882a593Smuzhiyun
1135*4882a593Smuzhiyun status = myrs_get_ldev_info(cs, ldev_num, ldev_info);
1136*4882a593Smuzhiyun if (status != MYRS_STATUS_SUCCESS) {
1137*4882a593Smuzhiyun sdev_printk(KERN_INFO, sdev,
1138*4882a593Smuzhiyun "Failed to get device information, status 0x%02x\n",
1139*4882a593Smuzhiyun status);
1140*4882a593Smuzhiyun return -EIO;
1141*4882a593Smuzhiyun }
1142*4882a593Smuzhiyun
1143*4882a593Smuzhiyun if (rebuild && ldev_info->rbld_active) {
1144*4882a593Smuzhiyun sdev_printk(KERN_INFO, sdev,
1145*4882a593Smuzhiyun "Rebuild Not Initiated; already in progress\n");
1146*4882a593Smuzhiyun return -EALREADY;
1147*4882a593Smuzhiyun }
1148*4882a593Smuzhiyun if (!rebuild && !ldev_info->rbld_active) {
1149*4882a593Smuzhiyun sdev_printk(KERN_INFO, sdev,
1150*4882a593Smuzhiyun "Rebuild Not Cancelled; no rebuild in progress\n");
1151*4882a593Smuzhiyun return count;
1152*4882a593Smuzhiyun }
1153*4882a593Smuzhiyun
1154*4882a593Smuzhiyun mutex_lock(&cs->dcmd_mutex);
1155*4882a593Smuzhiyun cmd_blk = &cs->dcmd_blk;
1156*4882a593Smuzhiyun myrs_reset_cmd(cmd_blk);
1157*4882a593Smuzhiyun mbox = &cmd_blk->mbox;
1158*4882a593Smuzhiyun mbox->common.opcode = MYRS_CMD_OP_IOCTL;
1159*4882a593Smuzhiyun mbox->common.id = MYRS_DCMD_TAG;
1160*4882a593Smuzhiyun mbox->common.control.dma_ctrl_to_host = true;
1161*4882a593Smuzhiyun mbox->common.control.no_autosense = true;
1162*4882a593Smuzhiyun if (rebuild) {
1163*4882a593Smuzhiyun mbox->ldev_info.ldev.ldev_num = ldev_num;
1164*4882a593Smuzhiyun mbox->ldev_info.ioctl_opcode = MYRS_IOCTL_RBLD_DEVICE_START;
1165*4882a593Smuzhiyun } else {
1166*4882a593Smuzhiyun mbox->ldev_info.ldev.ldev_num = ldev_num;
1167*4882a593Smuzhiyun mbox->ldev_info.ioctl_opcode = MYRS_IOCTL_RBLD_DEVICE_STOP;
1168*4882a593Smuzhiyun }
1169*4882a593Smuzhiyun myrs_exec_cmd(cs, cmd_blk);
1170*4882a593Smuzhiyun status = cmd_blk->status;
1171*4882a593Smuzhiyun mutex_unlock(&cs->dcmd_mutex);
1172*4882a593Smuzhiyun if (status) {
1173*4882a593Smuzhiyun sdev_printk(KERN_INFO, sdev,
1174*4882a593Smuzhiyun "Rebuild Not %s, status 0x%02x\n",
1175*4882a593Smuzhiyun rebuild ? "Initiated" : "Cancelled", status);
1176*4882a593Smuzhiyun ret = -EIO;
1177*4882a593Smuzhiyun } else {
1178*4882a593Smuzhiyun sdev_printk(KERN_INFO, sdev, "Rebuild %s\n",
1179*4882a593Smuzhiyun rebuild ? "Initiated" : "Cancelled");
1180*4882a593Smuzhiyun ret = count;
1181*4882a593Smuzhiyun }
1182*4882a593Smuzhiyun
1183*4882a593Smuzhiyun return ret;
1184*4882a593Smuzhiyun }
1185*4882a593Smuzhiyun static DEVICE_ATTR_RW(rebuild);
1186*4882a593Smuzhiyun
consistency_check_show(struct device * dev,struct device_attribute * attr,char * buf)1187*4882a593Smuzhiyun static ssize_t consistency_check_show(struct device *dev,
1188*4882a593Smuzhiyun struct device_attribute *attr, char *buf)
1189*4882a593Smuzhiyun {
1190*4882a593Smuzhiyun struct scsi_device *sdev = to_scsi_device(dev);
1191*4882a593Smuzhiyun struct myrs_hba *cs = shost_priv(sdev->host);
1192*4882a593Smuzhiyun struct myrs_ldev_info *ldev_info;
1193*4882a593Smuzhiyun unsigned short ldev_num;
1194*4882a593Smuzhiyun unsigned char status;
1195*4882a593Smuzhiyun
1196*4882a593Smuzhiyun if (sdev->channel < cs->ctlr_info->physchan_present)
1197*4882a593Smuzhiyun return snprintf(buf, 32, "physical device - not checking\n");
1198*4882a593Smuzhiyun
1199*4882a593Smuzhiyun ldev_info = sdev->hostdata;
1200*4882a593Smuzhiyun if (!ldev_info)
1201*4882a593Smuzhiyun return -ENXIO;
1202*4882a593Smuzhiyun ldev_num = ldev_info->ldev_num;
1203*4882a593Smuzhiyun status = myrs_get_ldev_info(cs, ldev_num, ldev_info);
1204*4882a593Smuzhiyun if (ldev_info->cc_active)
1205*4882a593Smuzhiyun return snprintf(buf, 32, "checking block %zu of %zu\n",
1206*4882a593Smuzhiyun (size_t)ldev_info->cc_lba,
1207*4882a593Smuzhiyun (size_t)ldev_info->cfg_devsize);
1208*4882a593Smuzhiyun else
1209*4882a593Smuzhiyun return snprintf(buf, 32, "not checking\n");
1210*4882a593Smuzhiyun }
1211*4882a593Smuzhiyun
consistency_check_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)1212*4882a593Smuzhiyun static ssize_t consistency_check_store(struct device *dev,
1213*4882a593Smuzhiyun struct device_attribute *attr, const char *buf, size_t count)
1214*4882a593Smuzhiyun {
1215*4882a593Smuzhiyun struct scsi_device *sdev = to_scsi_device(dev);
1216*4882a593Smuzhiyun struct myrs_hba *cs = shost_priv(sdev->host);
1217*4882a593Smuzhiyun struct myrs_ldev_info *ldev_info;
1218*4882a593Smuzhiyun struct myrs_cmdblk *cmd_blk;
1219*4882a593Smuzhiyun union myrs_cmd_mbox *mbox;
1220*4882a593Smuzhiyun unsigned short ldev_num;
1221*4882a593Smuzhiyun unsigned char status;
1222*4882a593Smuzhiyun int check, ret;
1223*4882a593Smuzhiyun
1224*4882a593Smuzhiyun if (sdev->channel < cs->ctlr_info->physchan_present)
1225*4882a593Smuzhiyun return -EINVAL;
1226*4882a593Smuzhiyun
1227*4882a593Smuzhiyun ldev_info = sdev->hostdata;
1228*4882a593Smuzhiyun if (!ldev_info)
1229*4882a593Smuzhiyun return -ENXIO;
1230*4882a593Smuzhiyun ldev_num = ldev_info->ldev_num;
1231*4882a593Smuzhiyun
1232*4882a593Smuzhiyun ret = kstrtoint(buf, 0, &check);
1233*4882a593Smuzhiyun if (ret)
1234*4882a593Smuzhiyun return ret;
1235*4882a593Smuzhiyun
1236*4882a593Smuzhiyun status = myrs_get_ldev_info(cs, ldev_num, ldev_info);
1237*4882a593Smuzhiyun if (status != MYRS_STATUS_SUCCESS) {
1238*4882a593Smuzhiyun sdev_printk(KERN_INFO, sdev,
1239*4882a593Smuzhiyun "Failed to get device information, status 0x%02x\n",
1240*4882a593Smuzhiyun status);
1241*4882a593Smuzhiyun return -EIO;
1242*4882a593Smuzhiyun }
1243*4882a593Smuzhiyun if (check && ldev_info->cc_active) {
1244*4882a593Smuzhiyun sdev_printk(KERN_INFO, sdev,
1245*4882a593Smuzhiyun "Consistency Check Not Initiated; "
1246*4882a593Smuzhiyun "already in progress\n");
1247*4882a593Smuzhiyun return -EALREADY;
1248*4882a593Smuzhiyun }
1249*4882a593Smuzhiyun if (!check && !ldev_info->cc_active) {
1250*4882a593Smuzhiyun sdev_printk(KERN_INFO, sdev,
1251*4882a593Smuzhiyun "Consistency Check Not Cancelled; "
1252*4882a593Smuzhiyun "check not in progress\n");
1253*4882a593Smuzhiyun return count;
1254*4882a593Smuzhiyun }
1255*4882a593Smuzhiyun
1256*4882a593Smuzhiyun mutex_lock(&cs->dcmd_mutex);
1257*4882a593Smuzhiyun cmd_blk = &cs->dcmd_blk;
1258*4882a593Smuzhiyun myrs_reset_cmd(cmd_blk);
1259*4882a593Smuzhiyun mbox = &cmd_blk->mbox;
1260*4882a593Smuzhiyun mbox->common.opcode = MYRS_CMD_OP_IOCTL;
1261*4882a593Smuzhiyun mbox->common.id = MYRS_DCMD_TAG;
1262*4882a593Smuzhiyun mbox->common.control.dma_ctrl_to_host = true;
1263*4882a593Smuzhiyun mbox->common.control.no_autosense = true;
1264*4882a593Smuzhiyun if (check) {
1265*4882a593Smuzhiyun mbox->cc.ldev.ldev_num = ldev_num;
1266*4882a593Smuzhiyun mbox->cc.ioctl_opcode = MYRS_IOCTL_CC_START;
1267*4882a593Smuzhiyun mbox->cc.restore_consistency = true;
1268*4882a593Smuzhiyun mbox->cc.initialized_area_only = false;
1269*4882a593Smuzhiyun } else {
1270*4882a593Smuzhiyun mbox->cc.ldev.ldev_num = ldev_num;
1271*4882a593Smuzhiyun mbox->cc.ioctl_opcode = MYRS_IOCTL_CC_STOP;
1272*4882a593Smuzhiyun }
1273*4882a593Smuzhiyun myrs_exec_cmd(cs, cmd_blk);
1274*4882a593Smuzhiyun status = cmd_blk->status;
1275*4882a593Smuzhiyun mutex_unlock(&cs->dcmd_mutex);
1276*4882a593Smuzhiyun if (status != MYRS_STATUS_SUCCESS) {
1277*4882a593Smuzhiyun sdev_printk(KERN_INFO, sdev,
1278*4882a593Smuzhiyun "Consistency Check Not %s, status 0x%02x\n",
1279*4882a593Smuzhiyun check ? "Initiated" : "Cancelled", status);
1280*4882a593Smuzhiyun ret = -EIO;
1281*4882a593Smuzhiyun } else {
1282*4882a593Smuzhiyun sdev_printk(KERN_INFO, sdev, "Consistency Check %s\n",
1283*4882a593Smuzhiyun check ? "Initiated" : "Cancelled");
1284*4882a593Smuzhiyun ret = count;
1285*4882a593Smuzhiyun }
1286*4882a593Smuzhiyun
1287*4882a593Smuzhiyun return ret;
1288*4882a593Smuzhiyun }
1289*4882a593Smuzhiyun static DEVICE_ATTR_RW(consistency_check);
1290*4882a593Smuzhiyun
1291*4882a593Smuzhiyun static struct device_attribute *myrs_sdev_attrs[] = {
1292*4882a593Smuzhiyun &dev_attr_consistency_check,
1293*4882a593Smuzhiyun &dev_attr_rebuild,
1294*4882a593Smuzhiyun &dev_attr_raid_state,
1295*4882a593Smuzhiyun &dev_attr_raid_level,
1296*4882a593Smuzhiyun NULL,
1297*4882a593Smuzhiyun };
1298*4882a593Smuzhiyun
serial_show(struct device * dev,struct device_attribute * attr,char * buf)1299*4882a593Smuzhiyun static ssize_t serial_show(struct device *dev,
1300*4882a593Smuzhiyun struct device_attribute *attr, char *buf)
1301*4882a593Smuzhiyun {
1302*4882a593Smuzhiyun struct Scsi_Host *shost = class_to_shost(dev);
1303*4882a593Smuzhiyun struct myrs_hba *cs = shost_priv(shost);
1304*4882a593Smuzhiyun char serial[17];
1305*4882a593Smuzhiyun
1306*4882a593Smuzhiyun memcpy(serial, cs->ctlr_info->serial_number, 16);
1307*4882a593Smuzhiyun serial[16] = '\0';
1308*4882a593Smuzhiyun return snprintf(buf, 16, "%s\n", serial);
1309*4882a593Smuzhiyun }
1310*4882a593Smuzhiyun static DEVICE_ATTR_RO(serial);
1311*4882a593Smuzhiyun
ctlr_num_show(struct device * dev,struct device_attribute * attr,char * buf)1312*4882a593Smuzhiyun static ssize_t ctlr_num_show(struct device *dev,
1313*4882a593Smuzhiyun struct device_attribute *attr, char *buf)
1314*4882a593Smuzhiyun {
1315*4882a593Smuzhiyun struct Scsi_Host *shost = class_to_shost(dev);
1316*4882a593Smuzhiyun struct myrs_hba *cs = shost_priv(shost);
1317*4882a593Smuzhiyun
1318*4882a593Smuzhiyun return snprintf(buf, 20, "%d\n", cs->host->host_no);
1319*4882a593Smuzhiyun }
1320*4882a593Smuzhiyun static DEVICE_ATTR_RO(ctlr_num);
1321*4882a593Smuzhiyun
1322*4882a593Smuzhiyun static struct myrs_cpu_type_tbl {
1323*4882a593Smuzhiyun enum myrs_cpu_type type;
1324*4882a593Smuzhiyun char *name;
1325*4882a593Smuzhiyun } myrs_cpu_type_names[] = {
1326*4882a593Smuzhiyun { MYRS_CPUTYPE_i960CA, "i960CA" },
1327*4882a593Smuzhiyun { MYRS_CPUTYPE_i960RD, "i960RD" },
1328*4882a593Smuzhiyun { MYRS_CPUTYPE_i960RN, "i960RN" },
1329*4882a593Smuzhiyun { MYRS_CPUTYPE_i960RP, "i960RP" },
1330*4882a593Smuzhiyun { MYRS_CPUTYPE_NorthBay, "NorthBay" },
1331*4882a593Smuzhiyun { MYRS_CPUTYPE_StrongArm, "StrongARM" },
1332*4882a593Smuzhiyun { MYRS_CPUTYPE_i960RM, "i960RM" },
1333*4882a593Smuzhiyun };
1334*4882a593Smuzhiyun
processor_show(struct device * dev,struct device_attribute * attr,char * buf)1335*4882a593Smuzhiyun static ssize_t processor_show(struct device *dev,
1336*4882a593Smuzhiyun struct device_attribute *attr, char *buf)
1337*4882a593Smuzhiyun {
1338*4882a593Smuzhiyun struct Scsi_Host *shost = class_to_shost(dev);
1339*4882a593Smuzhiyun struct myrs_hba *cs = shost_priv(shost);
1340*4882a593Smuzhiyun struct myrs_cpu_type_tbl *tbl;
1341*4882a593Smuzhiyun const char *first_processor = NULL;
1342*4882a593Smuzhiyun const char *second_processor = NULL;
1343*4882a593Smuzhiyun struct myrs_ctlr_info *info = cs->ctlr_info;
1344*4882a593Smuzhiyun ssize_t ret;
1345*4882a593Smuzhiyun int i;
1346*4882a593Smuzhiyun
1347*4882a593Smuzhiyun if (info->cpu[0].cpu_count) {
1348*4882a593Smuzhiyun tbl = myrs_cpu_type_names;
1349*4882a593Smuzhiyun for (i = 0; i < ARRAY_SIZE(myrs_cpu_type_names); i++) {
1350*4882a593Smuzhiyun if (tbl[i].type == info->cpu[0].cpu_type) {
1351*4882a593Smuzhiyun first_processor = tbl[i].name;
1352*4882a593Smuzhiyun break;
1353*4882a593Smuzhiyun }
1354*4882a593Smuzhiyun }
1355*4882a593Smuzhiyun }
1356*4882a593Smuzhiyun if (info->cpu[1].cpu_count) {
1357*4882a593Smuzhiyun tbl = myrs_cpu_type_names;
1358*4882a593Smuzhiyun for (i = 0; i < ARRAY_SIZE(myrs_cpu_type_names); i++) {
1359*4882a593Smuzhiyun if (tbl[i].type == info->cpu[1].cpu_type) {
1360*4882a593Smuzhiyun second_processor = tbl[i].name;
1361*4882a593Smuzhiyun break;
1362*4882a593Smuzhiyun }
1363*4882a593Smuzhiyun }
1364*4882a593Smuzhiyun }
1365*4882a593Smuzhiyun if (first_processor && second_processor)
1366*4882a593Smuzhiyun ret = snprintf(buf, 64, "1: %s (%s, %d cpus)\n"
1367*4882a593Smuzhiyun "2: %s (%s, %d cpus)\n",
1368*4882a593Smuzhiyun info->cpu[0].cpu_name,
1369*4882a593Smuzhiyun first_processor, info->cpu[0].cpu_count,
1370*4882a593Smuzhiyun info->cpu[1].cpu_name,
1371*4882a593Smuzhiyun second_processor, info->cpu[1].cpu_count);
1372*4882a593Smuzhiyun else if (first_processor && !second_processor)
1373*4882a593Smuzhiyun ret = snprintf(buf, 64, "1: %s (%s, %d cpus)\n2: absent\n",
1374*4882a593Smuzhiyun info->cpu[0].cpu_name,
1375*4882a593Smuzhiyun first_processor, info->cpu[0].cpu_count);
1376*4882a593Smuzhiyun else if (!first_processor && second_processor)
1377*4882a593Smuzhiyun ret = snprintf(buf, 64, "1: absent\n2: %s (%s, %d cpus)\n",
1378*4882a593Smuzhiyun info->cpu[1].cpu_name,
1379*4882a593Smuzhiyun second_processor, info->cpu[1].cpu_count);
1380*4882a593Smuzhiyun else
1381*4882a593Smuzhiyun ret = snprintf(buf, 64, "1: absent\n2: absent\n");
1382*4882a593Smuzhiyun
1383*4882a593Smuzhiyun return ret;
1384*4882a593Smuzhiyun }
1385*4882a593Smuzhiyun static DEVICE_ATTR_RO(processor);
1386*4882a593Smuzhiyun
model_show(struct device * dev,struct device_attribute * attr,char * buf)1387*4882a593Smuzhiyun static ssize_t model_show(struct device *dev,
1388*4882a593Smuzhiyun struct device_attribute *attr, char *buf)
1389*4882a593Smuzhiyun {
1390*4882a593Smuzhiyun struct Scsi_Host *shost = class_to_shost(dev);
1391*4882a593Smuzhiyun struct myrs_hba *cs = shost_priv(shost);
1392*4882a593Smuzhiyun
1393*4882a593Smuzhiyun return snprintf(buf, 28, "%s\n", cs->model_name);
1394*4882a593Smuzhiyun }
1395*4882a593Smuzhiyun static DEVICE_ATTR_RO(model);
1396*4882a593Smuzhiyun
ctlr_type_show(struct device * dev,struct device_attribute * attr,char * buf)1397*4882a593Smuzhiyun static ssize_t ctlr_type_show(struct device *dev,
1398*4882a593Smuzhiyun struct device_attribute *attr, char *buf)
1399*4882a593Smuzhiyun {
1400*4882a593Smuzhiyun struct Scsi_Host *shost = class_to_shost(dev);
1401*4882a593Smuzhiyun struct myrs_hba *cs = shost_priv(shost);
1402*4882a593Smuzhiyun
1403*4882a593Smuzhiyun return snprintf(buf, 4, "%d\n", cs->ctlr_info->ctlr_type);
1404*4882a593Smuzhiyun }
1405*4882a593Smuzhiyun static DEVICE_ATTR_RO(ctlr_type);
1406*4882a593Smuzhiyun
cache_size_show(struct device * dev,struct device_attribute * attr,char * buf)1407*4882a593Smuzhiyun static ssize_t cache_size_show(struct device *dev,
1408*4882a593Smuzhiyun struct device_attribute *attr, char *buf)
1409*4882a593Smuzhiyun {
1410*4882a593Smuzhiyun struct Scsi_Host *shost = class_to_shost(dev);
1411*4882a593Smuzhiyun struct myrs_hba *cs = shost_priv(shost);
1412*4882a593Smuzhiyun
1413*4882a593Smuzhiyun return snprintf(buf, 8, "%d MB\n", cs->ctlr_info->cache_size_mb);
1414*4882a593Smuzhiyun }
1415*4882a593Smuzhiyun static DEVICE_ATTR_RO(cache_size);
1416*4882a593Smuzhiyun
firmware_show(struct device * dev,struct device_attribute * attr,char * buf)1417*4882a593Smuzhiyun static ssize_t firmware_show(struct device *dev,
1418*4882a593Smuzhiyun struct device_attribute *attr, char *buf)
1419*4882a593Smuzhiyun {
1420*4882a593Smuzhiyun struct Scsi_Host *shost = class_to_shost(dev);
1421*4882a593Smuzhiyun struct myrs_hba *cs = shost_priv(shost);
1422*4882a593Smuzhiyun
1423*4882a593Smuzhiyun return snprintf(buf, 16, "%d.%02d-%02d\n",
1424*4882a593Smuzhiyun cs->ctlr_info->fw_major_version,
1425*4882a593Smuzhiyun cs->ctlr_info->fw_minor_version,
1426*4882a593Smuzhiyun cs->ctlr_info->fw_turn_number);
1427*4882a593Smuzhiyun }
1428*4882a593Smuzhiyun static DEVICE_ATTR_RO(firmware);
1429*4882a593Smuzhiyun
discovery_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)1430*4882a593Smuzhiyun static ssize_t discovery_store(struct device *dev,
1431*4882a593Smuzhiyun struct device_attribute *attr, const char *buf, size_t count)
1432*4882a593Smuzhiyun {
1433*4882a593Smuzhiyun struct Scsi_Host *shost = class_to_shost(dev);
1434*4882a593Smuzhiyun struct myrs_hba *cs = shost_priv(shost);
1435*4882a593Smuzhiyun struct myrs_cmdblk *cmd_blk;
1436*4882a593Smuzhiyun union myrs_cmd_mbox *mbox;
1437*4882a593Smuzhiyun unsigned char status;
1438*4882a593Smuzhiyun
1439*4882a593Smuzhiyun mutex_lock(&cs->dcmd_mutex);
1440*4882a593Smuzhiyun cmd_blk = &cs->dcmd_blk;
1441*4882a593Smuzhiyun myrs_reset_cmd(cmd_blk);
1442*4882a593Smuzhiyun mbox = &cmd_blk->mbox;
1443*4882a593Smuzhiyun mbox->common.opcode = MYRS_CMD_OP_IOCTL;
1444*4882a593Smuzhiyun mbox->common.id = MYRS_DCMD_TAG;
1445*4882a593Smuzhiyun mbox->common.control.dma_ctrl_to_host = true;
1446*4882a593Smuzhiyun mbox->common.control.no_autosense = true;
1447*4882a593Smuzhiyun mbox->common.ioctl_opcode = MYRS_IOCTL_START_DISCOVERY;
1448*4882a593Smuzhiyun myrs_exec_cmd(cs, cmd_blk);
1449*4882a593Smuzhiyun status = cmd_blk->status;
1450*4882a593Smuzhiyun mutex_unlock(&cs->dcmd_mutex);
1451*4882a593Smuzhiyun if (status != MYRS_STATUS_SUCCESS) {
1452*4882a593Smuzhiyun shost_printk(KERN_INFO, shost,
1453*4882a593Smuzhiyun "Discovery Not Initiated, status %02X\n",
1454*4882a593Smuzhiyun status);
1455*4882a593Smuzhiyun return -EINVAL;
1456*4882a593Smuzhiyun }
1457*4882a593Smuzhiyun shost_printk(KERN_INFO, shost, "Discovery Initiated\n");
1458*4882a593Smuzhiyun cs->next_evseq = 0;
1459*4882a593Smuzhiyun cs->needs_update = true;
1460*4882a593Smuzhiyun queue_delayed_work(cs->work_q, &cs->monitor_work, 1);
1461*4882a593Smuzhiyun flush_delayed_work(&cs->monitor_work);
1462*4882a593Smuzhiyun shost_printk(KERN_INFO, shost, "Discovery Completed\n");
1463*4882a593Smuzhiyun
1464*4882a593Smuzhiyun return count;
1465*4882a593Smuzhiyun }
1466*4882a593Smuzhiyun static DEVICE_ATTR_WO(discovery);
1467*4882a593Smuzhiyun
flush_cache_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)1468*4882a593Smuzhiyun static ssize_t flush_cache_store(struct device *dev,
1469*4882a593Smuzhiyun struct device_attribute *attr, const char *buf, size_t count)
1470*4882a593Smuzhiyun {
1471*4882a593Smuzhiyun struct Scsi_Host *shost = class_to_shost(dev);
1472*4882a593Smuzhiyun struct myrs_hba *cs = shost_priv(shost);
1473*4882a593Smuzhiyun unsigned char status;
1474*4882a593Smuzhiyun
1475*4882a593Smuzhiyun status = myrs_dev_op(cs, MYRS_IOCTL_FLUSH_DEVICE_DATA,
1476*4882a593Smuzhiyun MYRS_RAID_CONTROLLER);
1477*4882a593Smuzhiyun if (status == MYRS_STATUS_SUCCESS) {
1478*4882a593Smuzhiyun shost_printk(KERN_INFO, shost, "Cache Flush Completed\n");
1479*4882a593Smuzhiyun return count;
1480*4882a593Smuzhiyun }
1481*4882a593Smuzhiyun shost_printk(KERN_INFO, shost,
1482*4882a593Smuzhiyun "Cache Flush failed, status 0x%02x\n", status);
1483*4882a593Smuzhiyun return -EIO;
1484*4882a593Smuzhiyun }
1485*4882a593Smuzhiyun static DEVICE_ATTR_WO(flush_cache);
1486*4882a593Smuzhiyun
disable_enclosure_messages_show(struct device * dev,struct device_attribute * attr,char * buf)1487*4882a593Smuzhiyun static ssize_t disable_enclosure_messages_show(struct device *dev,
1488*4882a593Smuzhiyun struct device_attribute *attr, char *buf)
1489*4882a593Smuzhiyun {
1490*4882a593Smuzhiyun struct Scsi_Host *shost = class_to_shost(dev);
1491*4882a593Smuzhiyun struct myrs_hba *cs = shost_priv(shost);
1492*4882a593Smuzhiyun
1493*4882a593Smuzhiyun return snprintf(buf, 3, "%d\n", cs->disable_enc_msg);
1494*4882a593Smuzhiyun }
1495*4882a593Smuzhiyun
disable_enclosure_messages_store(struct device * dev,struct device_attribute * attr,const char * buf,size_t count)1496*4882a593Smuzhiyun static ssize_t disable_enclosure_messages_store(struct device *dev,
1497*4882a593Smuzhiyun struct device_attribute *attr, const char *buf, size_t count)
1498*4882a593Smuzhiyun {
1499*4882a593Smuzhiyun struct scsi_device *sdev = to_scsi_device(dev);
1500*4882a593Smuzhiyun struct myrs_hba *cs = shost_priv(sdev->host);
1501*4882a593Smuzhiyun int value, ret;
1502*4882a593Smuzhiyun
1503*4882a593Smuzhiyun ret = kstrtoint(buf, 0, &value);
1504*4882a593Smuzhiyun if (ret)
1505*4882a593Smuzhiyun return ret;
1506*4882a593Smuzhiyun
1507*4882a593Smuzhiyun if (value > 2)
1508*4882a593Smuzhiyun return -EINVAL;
1509*4882a593Smuzhiyun
1510*4882a593Smuzhiyun cs->disable_enc_msg = value;
1511*4882a593Smuzhiyun return count;
1512*4882a593Smuzhiyun }
1513*4882a593Smuzhiyun static DEVICE_ATTR_RW(disable_enclosure_messages);
1514*4882a593Smuzhiyun
1515*4882a593Smuzhiyun static struct device_attribute *myrs_shost_attrs[] = {
1516*4882a593Smuzhiyun &dev_attr_serial,
1517*4882a593Smuzhiyun &dev_attr_ctlr_num,
1518*4882a593Smuzhiyun &dev_attr_processor,
1519*4882a593Smuzhiyun &dev_attr_model,
1520*4882a593Smuzhiyun &dev_attr_ctlr_type,
1521*4882a593Smuzhiyun &dev_attr_cache_size,
1522*4882a593Smuzhiyun &dev_attr_firmware,
1523*4882a593Smuzhiyun &dev_attr_discovery,
1524*4882a593Smuzhiyun &dev_attr_flush_cache,
1525*4882a593Smuzhiyun &dev_attr_disable_enclosure_messages,
1526*4882a593Smuzhiyun NULL,
1527*4882a593Smuzhiyun };
1528*4882a593Smuzhiyun
1529*4882a593Smuzhiyun /*
1530*4882a593Smuzhiyun * SCSI midlayer interface
1531*4882a593Smuzhiyun */
myrs_host_reset(struct scsi_cmnd * scmd)1532*4882a593Smuzhiyun static int myrs_host_reset(struct scsi_cmnd *scmd)
1533*4882a593Smuzhiyun {
1534*4882a593Smuzhiyun struct Scsi_Host *shost = scmd->device->host;
1535*4882a593Smuzhiyun struct myrs_hba *cs = shost_priv(shost);
1536*4882a593Smuzhiyun
1537*4882a593Smuzhiyun cs->reset(cs->io_base);
1538*4882a593Smuzhiyun return SUCCESS;
1539*4882a593Smuzhiyun }
1540*4882a593Smuzhiyun
myrs_mode_sense(struct myrs_hba * cs,struct scsi_cmnd * scmd,struct myrs_ldev_info * ldev_info)1541*4882a593Smuzhiyun static void myrs_mode_sense(struct myrs_hba *cs, struct scsi_cmnd *scmd,
1542*4882a593Smuzhiyun struct myrs_ldev_info *ldev_info)
1543*4882a593Smuzhiyun {
1544*4882a593Smuzhiyun unsigned char modes[32], *mode_pg;
1545*4882a593Smuzhiyun bool dbd;
1546*4882a593Smuzhiyun size_t mode_len;
1547*4882a593Smuzhiyun
1548*4882a593Smuzhiyun dbd = (scmd->cmnd[1] & 0x08) == 0x08;
1549*4882a593Smuzhiyun if (dbd) {
1550*4882a593Smuzhiyun mode_len = 24;
1551*4882a593Smuzhiyun mode_pg = &modes[4];
1552*4882a593Smuzhiyun } else {
1553*4882a593Smuzhiyun mode_len = 32;
1554*4882a593Smuzhiyun mode_pg = &modes[12];
1555*4882a593Smuzhiyun }
1556*4882a593Smuzhiyun memset(modes, 0, sizeof(modes));
1557*4882a593Smuzhiyun modes[0] = mode_len - 1;
1558*4882a593Smuzhiyun modes[2] = 0x10; /* Enable FUA */
1559*4882a593Smuzhiyun if (ldev_info->ldev_control.wce == MYRS_LOGICALDEVICE_RO)
1560*4882a593Smuzhiyun modes[2] |= 0x80;
1561*4882a593Smuzhiyun if (!dbd) {
1562*4882a593Smuzhiyun unsigned char *block_desc = &modes[4];
1563*4882a593Smuzhiyun
1564*4882a593Smuzhiyun modes[3] = 8;
1565*4882a593Smuzhiyun put_unaligned_be32(ldev_info->cfg_devsize, &block_desc[0]);
1566*4882a593Smuzhiyun put_unaligned_be32(ldev_info->devsize_bytes, &block_desc[5]);
1567*4882a593Smuzhiyun }
1568*4882a593Smuzhiyun mode_pg[0] = 0x08;
1569*4882a593Smuzhiyun mode_pg[1] = 0x12;
1570*4882a593Smuzhiyun if (ldev_info->ldev_control.rce == MYRS_READCACHE_DISABLED)
1571*4882a593Smuzhiyun mode_pg[2] |= 0x01;
1572*4882a593Smuzhiyun if (ldev_info->ldev_control.wce == MYRS_WRITECACHE_ENABLED ||
1573*4882a593Smuzhiyun ldev_info->ldev_control.wce == MYRS_INTELLIGENT_WRITECACHE_ENABLED)
1574*4882a593Smuzhiyun mode_pg[2] |= 0x04;
1575*4882a593Smuzhiyun if (ldev_info->cacheline_size) {
1576*4882a593Smuzhiyun mode_pg[2] |= 0x08;
1577*4882a593Smuzhiyun put_unaligned_be16(1 << ldev_info->cacheline_size,
1578*4882a593Smuzhiyun &mode_pg[14]);
1579*4882a593Smuzhiyun }
1580*4882a593Smuzhiyun
1581*4882a593Smuzhiyun scsi_sg_copy_from_buffer(scmd, modes, mode_len);
1582*4882a593Smuzhiyun }
1583*4882a593Smuzhiyun
myrs_queuecommand(struct Scsi_Host * shost,struct scsi_cmnd * scmd)1584*4882a593Smuzhiyun static int myrs_queuecommand(struct Scsi_Host *shost,
1585*4882a593Smuzhiyun struct scsi_cmnd *scmd)
1586*4882a593Smuzhiyun {
1587*4882a593Smuzhiyun struct myrs_hba *cs = shost_priv(shost);
1588*4882a593Smuzhiyun struct myrs_cmdblk *cmd_blk = scsi_cmd_priv(scmd);
1589*4882a593Smuzhiyun union myrs_cmd_mbox *mbox = &cmd_blk->mbox;
1590*4882a593Smuzhiyun struct scsi_device *sdev = scmd->device;
1591*4882a593Smuzhiyun union myrs_sgl *hw_sge;
1592*4882a593Smuzhiyun dma_addr_t sense_addr;
1593*4882a593Smuzhiyun struct scatterlist *sgl;
1594*4882a593Smuzhiyun unsigned long flags, timeout;
1595*4882a593Smuzhiyun int nsge;
1596*4882a593Smuzhiyun
1597*4882a593Smuzhiyun if (!scmd->device->hostdata) {
1598*4882a593Smuzhiyun scmd->result = (DID_NO_CONNECT << 16);
1599*4882a593Smuzhiyun scmd->scsi_done(scmd);
1600*4882a593Smuzhiyun return 0;
1601*4882a593Smuzhiyun }
1602*4882a593Smuzhiyun
1603*4882a593Smuzhiyun switch (scmd->cmnd[0]) {
1604*4882a593Smuzhiyun case REPORT_LUNS:
1605*4882a593Smuzhiyun scsi_build_sense_buffer(0, scmd->sense_buffer, ILLEGAL_REQUEST,
1606*4882a593Smuzhiyun 0x20, 0x0);
1607*4882a593Smuzhiyun scmd->result = (DRIVER_SENSE << 24) | SAM_STAT_CHECK_CONDITION;
1608*4882a593Smuzhiyun scmd->scsi_done(scmd);
1609*4882a593Smuzhiyun return 0;
1610*4882a593Smuzhiyun case MODE_SENSE:
1611*4882a593Smuzhiyun if (scmd->device->channel >= cs->ctlr_info->physchan_present) {
1612*4882a593Smuzhiyun struct myrs_ldev_info *ldev_info = sdev->hostdata;
1613*4882a593Smuzhiyun
1614*4882a593Smuzhiyun if ((scmd->cmnd[2] & 0x3F) != 0x3F &&
1615*4882a593Smuzhiyun (scmd->cmnd[2] & 0x3F) != 0x08) {
1616*4882a593Smuzhiyun /* Illegal request, invalid field in CDB */
1617*4882a593Smuzhiyun scsi_build_sense_buffer(0, scmd->sense_buffer,
1618*4882a593Smuzhiyun ILLEGAL_REQUEST, 0x24, 0);
1619*4882a593Smuzhiyun scmd->result = (DRIVER_SENSE << 24) |
1620*4882a593Smuzhiyun SAM_STAT_CHECK_CONDITION;
1621*4882a593Smuzhiyun } else {
1622*4882a593Smuzhiyun myrs_mode_sense(cs, scmd, ldev_info);
1623*4882a593Smuzhiyun scmd->result = (DID_OK << 16);
1624*4882a593Smuzhiyun }
1625*4882a593Smuzhiyun scmd->scsi_done(scmd);
1626*4882a593Smuzhiyun return 0;
1627*4882a593Smuzhiyun }
1628*4882a593Smuzhiyun break;
1629*4882a593Smuzhiyun }
1630*4882a593Smuzhiyun
1631*4882a593Smuzhiyun myrs_reset_cmd(cmd_blk);
1632*4882a593Smuzhiyun cmd_blk->sense = dma_pool_alloc(cs->sense_pool, GFP_ATOMIC,
1633*4882a593Smuzhiyun &sense_addr);
1634*4882a593Smuzhiyun if (!cmd_blk->sense)
1635*4882a593Smuzhiyun return SCSI_MLQUEUE_HOST_BUSY;
1636*4882a593Smuzhiyun cmd_blk->sense_addr = sense_addr;
1637*4882a593Smuzhiyun
1638*4882a593Smuzhiyun timeout = scmd->request->timeout;
1639*4882a593Smuzhiyun if (scmd->cmd_len <= 10) {
1640*4882a593Smuzhiyun if (scmd->device->channel >= cs->ctlr_info->physchan_present) {
1641*4882a593Smuzhiyun struct myrs_ldev_info *ldev_info = sdev->hostdata;
1642*4882a593Smuzhiyun
1643*4882a593Smuzhiyun mbox->SCSI_10.opcode = MYRS_CMD_OP_SCSI_10;
1644*4882a593Smuzhiyun mbox->SCSI_10.pdev.lun = ldev_info->lun;
1645*4882a593Smuzhiyun mbox->SCSI_10.pdev.target = ldev_info->target;
1646*4882a593Smuzhiyun mbox->SCSI_10.pdev.channel = ldev_info->channel;
1647*4882a593Smuzhiyun mbox->SCSI_10.pdev.ctlr = 0;
1648*4882a593Smuzhiyun } else {
1649*4882a593Smuzhiyun mbox->SCSI_10.opcode = MYRS_CMD_OP_SCSI_10_PASSTHRU;
1650*4882a593Smuzhiyun mbox->SCSI_10.pdev.lun = sdev->lun;
1651*4882a593Smuzhiyun mbox->SCSI_10.pdev.target = sdev->id;
1652*4882a593Smuzhiyun mbox->SCSI_10.pdev.channel = sdev->channel;
1653*4882a593Smuzhiyun }
1654*4882a593Smuzhiyun mbox->SCSI_10.id = scmd->request->tag + 3;
1655*4882a593Smuzhiyun mbox->SCSI_10.control.dma_ctrl_to_host =
1656*4882a593Smuzhiyun (scmd->sc_data_direction == DMA_FROM_DEVICE);
1657*4882a593Smuzhiyun if (scmd->request->cmd_flags & REQ_FUA)
1658*4882a593Smuzhiyun mbox->SCSI_10.control.fua = true;
1659*4882a593Smuzhiyun mbox->SCSI_10.dma_size = scsi_bufflen(scmd);
1660*4882a593Smuzhiyun mbox->SCSI_10.sense_addr = cmd_blk->sense_addr;
1661*4882a593Smuzhiyun mbox->SCSI_10.sense_len = MYRS_SENSE_SIZE;
1662*4882a593Smuzhiyun mbox->SCSI_10.cdb_len = scmd->cmd_len;
1663*4882a593Smuzhiyun if (timeout > 60) {
1664*4882a593Smuzhiyun mbox->SCSI_10.tmo.tmo_scale = MYRS_TMO_SCALE_MINUTES;
1665*4882a593Smuzhiyun mbox->SCSI_10.tmo.tmo_val = timeout / 60;
1666*4882a593Smuzhiyun } else {
1667*4882a593Smuzhiyun mbox->SCSI_10.tmo.tmo_scale = MYRS_TMO_SCALE_SECONDS;
1668*4882a593Smuzhiyun mbox->SCSI_10.tmo.tmo_val = timeout;
1669*4882a593Smuzhiyun }
1670*4882a593Smuzhiyun memcpy(&mbox->SCSI_10.cdb, scmd->cmnd, scmd->cmd_len);
1671*4882a593Smuzhiyun hw_sge = &mbox->SCSI_10.dma_addr;
1672*4882a593Smuzhiyun cmd_blk->dcdb = NULL;
1673*4882a593Smuzhiyun } else {
1674*4882a593Smuzhiyun dma_addr_t dcdb_dma;
1675*4882a593Smuzhiyun
1676*4882a593Smuzhiyun cmd_blk->dcdb = dma_pool_alloc(cs->dcdb_pool, GFP_ATOMIC,
1677*4882a593Smuzhiyun &dcdb_dma);
1678*4882a593Smuzhiyun if (!cmd_blk->dcdb) {
1679*4882a593Smuzhiyun dma_pool_free(cs->sense_pool, cmd_blk->sense,
1680*4882a593Smuzhiyun cmd_blk->sense_addr);
1681*4882a593Smuzhiyun cmd_blk->sense = NULL;
1682*4882a593Smuzhiyun cmd_blk->sense_addr = 0;
1683*4882a593Smuzhiyun return SCSI_MLQUEUE_HOST_BUSY;
1684*4882a593Smuzhiyun }
1685*4882a593Smuzhiyun cmd_blk->dcdb_dma = dcdb_dma;
1686*4882a593Smuzhiyun if (scmd->device->channel >= cs->ctlr_info->physchan_present) {
1687*4882a593Smuzhiyun struct myrs_ldev_info *ldev_info = sdev->hostdata;
1688*4882a593Smuzhiyun
1689*4882a593Smuzhiyun mbox->SCSI_255.opcode = MYRS_CMD_OP_SCSI_256;
1690*4882a593Smuzhiyun mbox->SCSI_255.pdev.lun = ldev_info->lun;
1691*4882a593Smuzhiyun mbox->SCSI_255.pdev.target = ldev_info->target;
1692*4882a593Smuzhiyun mbox->SCSI_255.pdev.channel = ldev_info->channel;
1693*4882a593Smuzhiyun mbox->SCSI_255.pdev.ctlr = 0;
1694*4882a593Smuzhiyun } else {
1695*4882a593Smuzhiyun mbox->SCSI_255.opcode = MYRS_CMD_OP_SCSI_255_PASSTHRU;
1696*4882a593Smuzhiyun mbox->SCSI_255.pdev.lun = sdev->lun;
1697*4882a593Smuzhiyun mbox->SCSI_255.pdev.target = sdev->id;
1698*4882a593Smuzhiyun mbox->SCSI_255.pdev.channel = sdev->channel;
1699*4882a593Smuzhiyun }
1700*4882a593Smuzhiyun mbox->SCSI_255.id = scmd->request->tag + 3;
1701*4882a593Smuzhiyun mbox->SCSI_255.control.dma_ctrl_to_host =
1702*4882a593Smuzhiyun (scmd->sc_data_direction == DMA_FROM_DEVICE);
1703*4882a593Smuzhiyun if (scmd->request->cmd_flags & REQ_FUA)
1704*4882a593Smuzhiyun mbox->SCSI_255.control.fua = true;
1705*4882a593Smuzhiyun mbox->SCSI_255.dma_size = scsi_bufflen(scmd);
1706*4882a593Smuzhiyun mbox->SCSI_255.sense_addr = cmd_blk->sense_addr;
1707*4882a593Smuzhiyun mbox->SCSI_255.sense_len = MYRS_SENSE_SIZE;
1708*4882a593Smuzhiyun mbox->SCSI_255.cdb_len = scmd->cmd_len;
1709*4882a593Smuzhiyun mbox->SCSI_255.cdb_addr = cmd_blk->dcdb_dma;
1710*4882a593Smuzhiyun if (timeout > 60) {
1711*4882a593Smuzhiyun mbox->SCSI_255.tmo.tmo_scale = MYRS_TMO_SCALE_MINUTES;
1712*4882a593Smuzhiyun mbox->SCSI_255.tmo.tmo_val = timeout / 60;
1713*4882a593Smuzhiyun } else {
1714*4882a593Smuzhiyun mbox->SCSI_255.tmo.tmo_scale = MYRS_TMO_SCALE_SECONDS;
1715*4882a593Smuzhiyun mbox->SCSI_255.tmo.tmo_val = timeout;
1716*4882a593Smuzhiyun }
1717*4882a593Smuzhiyun memcpy(cmd_blk->dcdb, scmd->cmnd, scmd->cmd_len);
1718*4882a593Smuzhiyun hw_sge = &mbox->SCSI_255.dma_addr;
1719*4882a593Smuzhiyun }
1720*4882a593Smuzhiyun if (scmd->sc_data_direction == DMA_NONE)
1721*4882a593Smuzhiyun goto submit;
1722*4882a593Smuzhiyun nsge = scsi_dma_map(scmd);
1723*4882a593Smuzhiyun if (nsge == 1) {
1724*4882a593Smuzhiyun sgl = scsi_sglist(scmd);
1725*4882a593Smuzhiyun hw_sge->sge[0].sge_addr = (u64)sg_dma_address(sgl);
1726*4882a593Smuzhiyun hw_sge->sge[0].sge_count = (u64)sg_dma_len(sgl);
1727*4882a593Smuzhiyun } else {
1728*4882a593Smuzhiyun struct myrs_sge *hw_sgl;
1729*4882a593Smuzhiyun dma_addr_t hw_sgl_addr;
1730*4882a593Smuzhiyun int i;
1731*4882a593Smuzhiyun
1732*4882a593Smuzhiyun if (nsge > 2) {
1733*4882a593Smuzhiyun hw_sgl = dma_pool_alloc(cs->sg_pool, GFP_ATOMIC,
1734*4882a593Smuzhiyun &hw_sgl_addr);
1735*4882a593Smuzhiyun if (WARN_ON(!hw_sgl)) {
1736*4882a593Smuzhiyun if (cmd_blk->dcdb) {
1737*4882a593Smuzhiyun dma_pool_free(cs->dcdb_pool,
1738*4882a593Smuzhiyun cmd_blk->dcdb,
1739*4882a593Smuzhiyun cmd_blk->dcdb_dma);
1740*4882a593Smuzhiyun cmd_blk->dcdb = NULL;
1741*4882a593Smuzhiyun cmd_blk->dcdb_dma = 0;
1742*4882a593Smuzhiyun }
1743*4882a593Smuzhiyun dma_pool_free(cs->sense_pool,
1744*4882a593Smuzhiyun cmd_blk->sense,
1745*4882a593Smuzhiyun cmd_blk->sense_addr);
1746*4882a593Smuzhiyun cmd_blk->sense = NULL;
1747*4882a593Smuzhiyun cmd_blk->sense_addr = 0;
1748*4882a593Smuzhiyun return SCSI_MLQUEUE_HOST_BUSY;
1749*4882a593Smuzhiyun }
1750*4882a593Smuzhiyun cmd_blk->sgl = hw_sgl;
1751*4882a593Smuzhiyun cmd_blk->sgl_addr = hw_sgl_addr;
1752*4882a593Smuzhiyun if (scmd->cmd_len <= 10)
1753*4882a593Smuzhiyun mbox->SCSI_10.control.add_sge_mem = true;
1754*4882a593Smuzhiyun else
1755*4882a593Smuzhiyun mbox->SCSI_255.control.add_sge_mem = true;
1756*4882a593Smuzhiyun hw_sge->ext.sge0_len = nsge;
1757*4882a593Smuzhiyun hw_sge->ext.sge0_addr = cmd_blk->sgl_addr;
1758*4882a593Smuzhiyun } else
1759*4882a593Smuzhiyun hw_sgl = hw_sge->sge;
1760*4882a593Smuzhiyun
1761*4882a593Smuzhiyun scsi_for_each_sg(scmd, sgl, nsge, i) {
1762*4882a593Smuzhiyun if (WARN_ON(!hw_sgl)) {
1763*4882a593Smuzhiyun scsi_dma_unmap(scmd);
1764*4882a593Smuzhiyun scmd->result = (DID_ERROR << 16);
1765*4882a593Smuzhiyun scmd->scsi_done(scmd);
1766*4882a593Smuzhiyun return 0;
1767*4882a593Smuzhiyun }
1768*4882a593Smuzhiyun hw_sgl->sge_addr = (u64)sg_dma_address(sgl);
1769*4882a593Smuzhiyun hw_sgl->sge_count = (u64)sg_dma_len(sgl);
1770*4882a593Smuzhiyun hw_sgl++;
1771*4882a593Smuzhiyun }
1772*4882a593Smuzhiyun }
1773*4882a593Smuzhiyun submit:
1774*4882a593Smuzhiyun spin_lock_irqsave(&cs->queue_lock, flags);
1775*4882a593Smuzhiyun myrs_qcmd(cs, cmd_blk);
1776*4882a593Smuzhiyun spin_unlock_irqrestore(&cs->queue_lock, flags);
1777*4882a593Smuzhiyun
1778*4882a593Smuzhiyun return 0;
1779*4882a593Smuzhiyun }
1780*4882a593Smuzhiyun
myrs_translate_ldev(struct myrs_hba * cs,struct scsi_device * sdev)1781*4882a593Smuzhiyun static unsigned short myrs_translate_ldev(struct myrs_hba *cs,
1782*4882a593Smuzhiyun struct scsi_device *sdev)
1783*4882a593Smuzhiyun {
1784*4882a593Smuzhiyun unsigned short ldev_num;
1785*4882a593Smuzhiyun unsigned int chan_offset =
1786*4882a593Smuzhiyun sdev->channel - cs->ctlr_info->physchan_present;
1787*4882a593Smuzhiyun
1788*4882a593Smuzhiyun ldev_num = sdev->id + chan_offset * sdev->host->max_id;
1789*4882a593Smuzhiyun
1790*4882a593Smuzhiyun return ldev_num;
1791*4882a593Smuzhiyun }
1792*4882a593Smuzhiyun
myrs_slave_alloc(struct scsi_device * sdev)1793*4882a593Smuzhiyun static int myrs_slave_alloc(struct scsi_device *sdev)
1794*4882a593Smuzhiyun {
1795*4882a593Smuzhiyun struct myrs_hba *cs = shost_priv(sdev->host);
1796*4882a593Smuzhiyun unsigned char status;
1797*4882a593Smuzhiyun
1798*4882a593Smuzhiyun if (sdev->channel > sdev->host->max_channel)
1799*4882a593Smuzhiyun return 0;
1800*4882a593Smuzhiyun
1801*4882a593Smuzhiyun if (sdev->channel >= cs->ctlr_info->physchan_present) {
1802*4882a593Smuzhiyun struct myrs_ldev_info *ldev_info;
1803*4882a593Smuzhiyun unsigned short ldev_num;
1804*4882a593Smuzhiyun
1805*4882a593Smuzhiyun if (sdev->lun > 0)
1806*4882a593Smuzhiyun return -ENXIO;
1807*4882a593Smuzhiyun
1808*4882a593Smuzhiyun ldev_num = myrs_translate_ldev(cs, sdev);
1809*4882a593Smuzhiyun
1810*4882a593Smuzhiyun ldev_info = kzalloc(sizeof(*ldev_info), GFP_KERNEL|GFP_DMA);
1811*4882a593Smuzhiyun if (!ldev_info)
1812*4882a593Smuzhiyun return -ENOMEM;
1813*4882a593Smuzhiyun
1814*4882a593Smuzhiyun status = myrs_get_ldev_info(cs, ldev_num, ldev_info);
1815*4882a593Smuzhiyun if (status != MYRS_STATUS_SUCCESS) {
1816*4882a593Smuzhiyun sdev->hostdata = NULL;
1817*4882a593Smuzhiyun kfree(ldev_info);
1818*4882a593Smuzhiyun } else {
1819*4882a593Smuzhiyun enum raid_level level;
1820*4882a593Smuzhiyun
1821*4882a593Smuzhiyun dev_dbg(&sdev->sdev_gendev,
1822*4882a593Smuzhiyun "Logical device mapping %d:%d:%d -> %d\n",
1823*4882a593Smuzhiyun ldev_info->channel, ldev_info->target,
1824*4882a593Smuzhiyun ldev_info->lun, ldev_info->ldev_num);
1825*4882a593Smuzhiyun
1826*4882a593Smuzhiyun sdev->hostdata = ldev_info;
1827*4882a593Smuzhiyun switch (ldev_info->raid_level) {
1828*4882a593Smuzhiyun case MYRS_RAID_LEVEL0:
1829*4882a593Smuzhiyun level = RAID_LEVEL_LINEAR;
1830*4882a593Smuzhiyun break;
1831*4882a593Smuzhiyun case MYRS_RAID_LEVEL1:
1832*4882a593Smuzhiyun level = RAID_LEVEL_1;
1833*4882a593Smuzhiyun break;
1834*4882a593Smuzhiyun case MYRS_RAID_LEVEL3:
1835*4882a593Smuzhiyun case MYRS_RAID_LEVEL3F:
1836*4882a593Smuzhiyun case MYRS_RAID_LEVEL3L:
1837*4882a593Smuzhiyun level = RAID_LEVEL_3;
1838*4882a593Smuzhiyun break;
1839*4882a593Smuzhiyun case MYRS_RAID_LEVEL5:
1840*4882a593Smuzhiyun case MYRS_RAID_LEVEL5L:
1841*4882a593Smuzhiyun level = RAID_LEVEL_5;
1842*4882a593Smuzhiyun break;
1843*4882a593Smuzhiyun case MYRS_RAID_LEVEL6:
1844*4882a593Smuzhiyun level = RAID_LEVEL_6;
1845*4882a593Smuzhiyun break;
1846*4882a593Smuzhiyun case MYRS_RAID_LEVELE:
1847*4882a593Smuzhiyun case MYRS_RAID_NEWSPAN:
1848*4882a593Smuzhiyun case MYRS_RAID_SPAN:
1849*4882a593Smuzhiyun level = RAID_LEVEL_LINEAR;
1850*4882a593Smuzhiyun break;
1851*4882a593Smuzhiyun case MYRS_RAID_JBOD:
1852*4882a593Smuzhiyun level = RAID_LEVEL_JBOD;
1853*4882a593Smuzhiyun break;
1854*4882a593Smuzhiyun default:
1855*4882a593Smuzhiyun level = RAID_LEVEL_UNKNOWN;
1856*4882a593Smuzhiyun break;
1857*4882a593Smuzhiyun }
1858*4882a593Smuzhiyun raid_set_level(myrs_raid_template,
1859*4882a593Smuzhiyun &sdev->sdev_gendev, level);
1860*4882a593Smuzhiyun if (ldev_info->dev_state != MYRS_DEVICE_ONLINE) {
1861*4882a593Smuzhiyun const char *name;
1862*4882a593Smuzhiyun
1863*4882a593Smuzhiyun name = myrs_devstate_name(ldev_info->dev_state);
1864*4882a593Smuzhiyun sdev_printk(KERN_DEBUG, sdev,
1865*4882a593Smuzhiyun "logical device in state %s\n",
1866*4882a593Smuzhiyun name ? name : "Invalid");
1867*4882a593Smuzhiyun }
1868*4882a593Smuzhiyun }
1869*4882a593Smuzhiyun } else {
1870*4882a593Smuzhiyun struct myrs_pdev_info *pdev_info;
1871*4882a593Smuzhiyun
1872*4882a593Smuzhiyun pdev_info = kzalloc(sizeof(*pdev_info), GFP_KERNEL|GFP_DMA);
1873*4882a593Smuzhiyun if (!pdev_info)
1874*4882a593Smuzhiyun return -ENOMEM;
1875*4882a593Smuzhiyun
1876*4882a593Smuzhiyun status = myrs_get_pdev_info(cs, sdev->channel,
1877*4882a593Smuzhiyun sdev->id, sdev->lun,
1878*4882a593Smuzhiyun pdev_info);
1879*4882a593Smuzhiyun if (status != MYRS_STATUS_SUCCESS) {
1880*4882a593Smuzhiyun sdev->hostdata = NULL;
1881*4882a593Smuzhiyun kfree(pdev_info);
1882*4882a593Smuzhiyun return -ENXIO;
1883*4882a593Smuzhiyun }
1884*4882a593Smuzhiyun sdev->hostdata = pdev_info;
1885*4882a593Smuzhiyun }
1886*4882a593Smuzhiyun return 0;
1887*4882a593Smuzhiyun }
1888*4882a593Smuzhiyun
myrs_slave_configure(struct scsi_device * sdev)1889*4882a593Smuzhiyun static int myrs_slave_configure(struct scsi_device *sdev)
1890*4882a593Smuzhiyun {
1891*4882a593Smuzhiyun struct myrs_hba *cs = shost_priv(sdev->host);
1892*4882a593Smuzhiyun struct myrs_ldev_info *ldev_info;
1893*4882a593Smuzhiyun
1894*4882a593Smuzhiyun if (sdev->channel > sdev->host->max_channel)
1895*4882a593Smuzhiyun return -ENXIO;
1896*4882a593Smuzhiyun
1897*4882a593Smuzhiyun if (sdev->channel < cs->ctlr_info->physchan_present) {
1898*4882a593Smuzhiyun /* Skip HBA device */
1899*4882a593Smuzhiyun if (sdev->type == TYPE_RAID)
1900*4882a593Smuzhiyun return -ENXIO;
1901*4882a593Smuzhiyun sdev->no_uld_attach = 1;
1902*4882a593Smuzhiyun return 0;
1903*4882a593Smuzhiyun }
1904*4882a593Smuzhiyun if (sdev->lun != 0)
1905*4882a593Smuzhiyun return -ENXIO;
1906*4882a593Smuzhiyun
1907*4882a593Smuzhiyun ldev_info = sdev->hostdata;
1908*4882a593Smuzhiyun if (!ldev_info)
1909*4882a593Smuzhiyun return -ENXIO;
1910*4882a593Smuzhiyun if (ldev_info->ldev_control.wce == MYRS_WRITECACHE_ENABLED ||
1911*4882a593Smuzhiyun ldev_info->ldev_control.wce == MYRS_INTELLIGENT_WRITECACHE_ENABLED)
1912*4882a593Smuzhiyun sdev->wce_default_on = 1;
1913*4882a593Smuzhiyun sdev->tagged_supported = 1;
1914*4882a593Smuzhiyun return 0;
1915*4882a593Smuzhiyun }
1916*4882a593Smuzhiyun
myrs_slave_destroy(struct scsi_device * sdev)1917*4882a593Smuzhiyun static void myrs_slave_destroy(struct scsi_device *sdev)
1918*4882a593Smuzhiyun {
1919*4882a593Smuzhiyun kfree(sdev->hostdata);
1920*4882a593Smuzhiyun }
1921*4882a593Smuzhiyun
1922*4882a593Smuzhiyun static struct scsi_host_template myrs_template = {
1923*4882a593Smuzhiyun .module = THIS_MODULE,
1924*4882a593Smuzhiyun .name = "DAC960",
1925*4882a593Smuzhiyun .proc_name = "myrs",
1926*4882a593Smuzhiyun .queuecommand = myrs_queuecommand,
1927*4882a593Smuzhiyun .eh_host_reset_handler = myrs_host_reset,
1928*4882a593Smuzhiyun .slave_alloc = myrs_slave_alloc,
1929*4882a593Smuzhiyun .slave_configure = myrs_slave_configure,
1930*4882a593Smuzhiyun .slave_destroy = myrs_slave_destroy,
1931*4882a593Smuzhiyun .cmd_size = sizeof(struct myrs_cmdblk),
1932*4882a593Smuzhiyun .shost_attrs = myrs_shost_attrs,
1933*4882a593Smuzhiyun .sdev_attrs = myrs_sdev_attrs,
1934*4882a593Smuzhiyun .this_id = -1,
1935*4882a593Smuzhiyun };
1936*4882a593Smuzhiyun
myrs_alloc_host(struct pci_dev * pdev,const struct pci_device_id * entry)1937*4882a593Smuzhiyun static struct myrs_hba *myrs_alloc_host(struct pci_dev *pdev,
1938*4882a593Smuzhiyun const struct pci_device_id *entry)
1939*4882a593Smuzhiyun {
1940*4882a593Smuzhiyun struct Scsi_Host *shost;
1941*4882a593Smuzhiyun struct myrs_hba *cs;
1942*4882a593Smuzhiyun
1943*4882a593Smuzhiyun shost = scsi_host_alloc(&myrs_template, sizeof(struct myrs_hba));
1944*4882a593Smuzhiyun if (!shost)
1945*4882a593Smuzhiyun return NULL;
1946*4882a593Smuzhiyun
1947*4882a593Smuzhiyun shost->max_cmd_len = 16;
1948*4882a593Smuzhiyun shost->max_lun = 256;
1949*4882a593Smuzhiyun cs = shost_priv(shost);
1950*4882a593Smuzhiyun mutex_init(&cs->dcmd_mutex);
1951*4882a593Smuzhiyun mutex_init(&cs->cinfo_mutex);
1952*4882a593Smuzhiyun cs->host = shost;
1953*4882a593Smuzhiyun
1954*4882a593Smuzhiyun return cs;
1955*4882a593Smuzhiyun }
1956*4882a593Smuzhiyun
1957*4882a593Smuzhiyun /*
1958*4882a593Smuzhiyun * RAID template functions
1959*4882a593Smuzhiyun */
1960*4882a593Smuzhiyun
1961*4882a593Smuzhiyun /**
1962*4882a593Smuzhiyun * myrs_is_raid - return boolean indicating device is raid volume
1963*4882a593Smuzhiyun * @dev the device struct object
1964*4882a593Smuzhiyun */
1965*4882a593Smuzhiyun static int
myrs_is_raid(struct device * dev)1966*4882a593Smuzhiyun myrs_is_raid(struct device *dev)
1967*4882a593Smuzhiyun {
1968*4882a593Smuzhiyun struct scsi_device *sdev = to_scsi_device(dev);
1969*4882a593Smuzhiyun struct myrs_hba *cs = shost_priv(sdev->host);
1970*4882a593Smuzhiyun
1971*4882a593Smuzhiyun return (sdev->channel >= cs->ctlr_info->physchan_present) ? 1 : 0;
1972*4882a593Smuzhiyun }
1973*4882a593Smuzhiyun
1974*4882a593Smuzhiyun /**
1975*4882a593Smuzhiyun * myrs_get_resync - get raid volume resync percent complete
1976*4882a593Smuzhiyun * @dev the device struct object
1977*4882a593Smuzhiyun */
1978*4882a593Smuzhiyun static void
myrs_get_resync(struct device * dev)1979*4882a593Smuzhiyun myrs_get_resync(struct device *dev)
1980*4882a593Smuzhiyun {
1981*4882a593Smuzhiyun struct scsi_device *sdev = to_scsi_device(dev);
1982*4882a593Smuzhiyun struct myrs_hba *cs = shost_priv(sdev->host);
1983*4882a593Smuzhiyun struct myrs_ldev_info *ldev_info = sdev->hostdata;
1984*4882a593Smuzhiyun u64 percent_complete = 0;
1985*4882a593Smuzhiyun u8 status;
1986*4882a593Smuzhiyun
1987*4882a593Smuzhiyun if (sdev->channel < cs->ctlr_info->physchan_present || !ldev_info)
1988*4882a593Smuzhiyun return;
1989*4882a593Smuzhiyun if (ldev_info->rbld_active) {
1990*4882a593Smuzhiyun unsigned short ldev_num = ldev_info->ldev_num;
1991*4882a593Smuzhiyun
1992*4882a593Smuzhiyun status = myrs_get_ldev_info(cs, ldev_num, ldev_info);
1993*4882a593Smuzhiyun percent_complete = ldev_info->rbld_lba * 100;
1994*4882a593Smuzhiyun do_div(percent_complete, ldev_info->cfg_devsize);
1995*4882a593Smuzhiyun }
1996*4882a593Smuzhiyun raid_set_resync(myrs_raid_template, dev, percent_complete);
1997*4882a593Smuzhiyun }
1998*4882a593Smuzhiyun
1999*4882a593Smuzhiyun /**
2000*4882a593Smuzhiyun * myrs_get_state - get raid volume status
2001*4882a593Smuzhiyun * @dev the device struct object
2002*4882a593Smuzhiyun */
2003*4882a593Smuzhiyun static void
myrs_get_state(struct device * dev)2004*4882a593Smuzhiyun myrs_get_state(struct device *dev)
2005*4882a593Smuzhiyun {
2006*4882a593Smuzhiyun struct scsi_device *sdev = to_scsi_device(dev);
2007*4882a593Smuzhiyun struct myrs_hba *cs = shost_priv(sdev->host);
2008*4882a593Smuzhiyun struct myrs_ldev_info *ldev_info = sdev->hostdata;
2009*4882a593Smuzhiyun enum raid_state state = RAID_STATE_UNKNOWN;
2010*4882a593Smuzhiyun
2011*4882a593Smuzhiyun if (sdev->channel < cs->ctlr_info->physchan_present || !ldev_info)
2012*4882a593Smuzhiyun state = RAID_STATE_UNKNOWN;
2013*4882a593Smuzhiyun else {
2014*4882a593Smuzhiyun switch (ldev_info->dev_state) {
2015*4882a593Smuzhiyun case MYRS_DEVICE_ONLINE:
2016*4882a593Smuzhiyun state = RAID_STATE_ACTIVE;
2017*4882a593Smuzhiyun break;
2018*4882a593Smuzhiyun case MYRS_DEVICE_SUSPECTED_CRITICAL:
2019*4882a593Smuzhiyun case MYRS_DEVICE_CRITICAL:
2020*4882a593Smuzhiyun state = RAID_STATE_DEGRADED;
2021*4882a593Smuzhiyun break;
2022*4882a593Smuzhiyun case MYRS_DEVICE_REBUILD:
2023*4882a593Smuzhiyun state = RAID_STATE_RESYNCING;
2024*4882a593Smuzhiyun break;
2025*4882a593Smuzhiyun case MYRS_DEVICE_UNCONFIGURED:
2026*4882a593Smuzhiyun case MYRS_DEVICE_INVALID_STATE:
2027*4882a593Smuzhiyun state = RAID_STATE_UNKNOWN;
2028*4882a593Smuzhiyun break;
2029*4882a593Smuzhiyun default:
2030*4882a593Smuzhiyun state = RAID_STATE_OFFLINE;
2031*4882a593Smuzhiyun }
2032*4882a593Smuzhiyun }
2033*4882a593Smuzhiyun raid_set_state(myrs_raid_template, dev, state);
2034*4882a593Smuzhiyun }
2035*4882a593Smuzhiyun
2036*4882a593Smuzhiyun static struct raid_function_template myrs_raid_functions = {
2037*4882a593Smuzhiyun .cookie = &myrs_template,
2038*4882a593Smuzhiyun .is_raid = myrs_is_raid,
2039*4882a593Smuzhiyun .get_resync = myrs_get_resync,
2040*4882a593Smuzhiyun .get_state = myrs_get_state,
2041*4882a593Smuzhiyun };
2042*4882a593Smuzhiyun
2043*4882a593Smuzhiyun /*
2044*4882a593Smuzhiyun * PCI interface functions
2045*4882a593Smuzhiyun */
myrs_flush_cache(struct myrs_hba * cs)2046*4882a593Smuzhiyun static void myrs_flush_cache(struct myrs_hba *cs)
2047*4882a593Smuzhiyun {
2048*4882a593Smuzhiyun myrs_dev_op(cs, MYRS_IOCTL_FLUSH_DEVICE_DATA, MYRS_RAID_CONTROLLER);
2049*4882a593Smuzhiyun }
2050*4882a593Smuzhiyun
myrs_handle_scsi(struct myrs_hba * cs,struct myrs_cmdblk * cmd_blk,struct scsi_cmnd * scmd)2051*4882a593Smuzhiyun static void myrs_handle_scsi(struct myrs_hba *cs, struct myrs_cmdblk *cmd_blk,
2052*4882a593Smuzhiyun struct scsi_cmnd *scmd)
2053*4882a593Smuzhiyun {
2054*4882a593Smuzhiyun unsigned char status;
2055*4882a593Smuzhiyun
2056*4882a593Smuzhiyun if (!cmd_blk)
2057*4882a593Smuzhiyun return;
2058*4882a593Smuzhiyun
2059*4882a593Smuzhiyun scsi_dma_unmap(scmd);
2060*4882a593Smuzhiyun status = cmd_blk->status;
2061*4882a593Smuzhiyun if (cmd_blk->sense) {
2062*4882a593Smuzhiyun if (status == MYRS_STATUS_FAILED && cmd_blk->sense_len) {
2063*4882a593Smuzhiyun unsigned int sense_len = SCSI_SENSE_BUFFERSIZE;
2064*4882a593Smuzhiyun
2065*4882a593Smuzhiyun if (sense_len > cmd_blk->sense_len)
2066*4882a593Smuzhiyun sense_len = cmd_blk->sense_len;
2067*4882a593Smuzhiyun memcpy(scmd->sense_buffer, cmd_blk->sense, sense_len);
2068*4882a593Smuzhiyun }
2069*4882a593Smuzhiyun dma_pool_free(cs->sense_pool, cmd_blk->sense,
2070*4882a593Smuzhiyun cmd_blk->sense_addr);
2071*4882a593Smuzhiyun cmd_blk->sense = NULL;
2072*4882a593Smuzhiyun cmd_blk->sense_addr = 0;
2073*4882a593Smuzhiyun }
2074*4882a593Smuzhiyun if (cmd_blk->dcdb) {
2075*4882a593Smuzhiyun dma_pool_free(cs->dcdb_pool, cmd_blk->dcdb,
2076*4882a593Smuzhiyun cmd_blk->dcdb_dma);
2077*4882a593Smuzhiyun cmd_blk->dcdb = NULL;
2078*4882a593Smuzhiyun cmd_blk->dcdb_dma = 0;
2079*4882a593Smuzhiyun }
2080*4882a593Smuzhiyun if (cmd_blk->sgl) {
2081*4882a593Smuzhiyun dma_pool_free(cs->sg_pool, cmd_blk->sgl,
2082*4882a593Smuzhiyun cmd_blk->sgl_addr);
2083*4882a593Smuzhiyun cmd_blk->sgl = NULL;
2084*4882a593Smuzhiyun cmd_blk->sgl_addr = 0;
2085*4882a593Smuzhiyun }
2086*4882a593Smuzhiyun if (cmd_blk->residual)
2087*4882a593Smuzhiyun scsi_set_resid(scmd, cmd_blk->residual);
2088*4882a593Smuzhiyun if (status == MYRS_STATUS_DEVICE_NON_RESPONSIVE ||
2089*4882a593Smuzhiyun status == MYRS_STATUS_DEVICE_NON_RESPONSIVE2)
2090*4882a593Smuzhiyun scmd->result = (DID_BAD_TARGET << 16);
2091*4882a593Smuzhiyun else
2092*4882a593Smuzhiyun scmd->result = (DID_OK << 16) | status;
2093*4882a593Smuzhiyun scmd->scsi_done(scmd);
2094*4882a593Smuzhiyun }
2095*4882a593Smuzhiyun
myrs_handle_cmdblk(struct myrs_hba * cs,struct myrs_cmdblk * cmd_blk)2096*4882a593Smuzhiyun static void myrs_handle_cmdblk(struct myrs_hba *cs, struct myrs_cmdblk *cmd_blk)
2097*4882a593Smuzhiyun {
2098*4882a593Smuzhiyun if (!cmd_blk)
2099*4882a593Smuzhiyun return;
2100*4882a593Smuzhiyun
2101*4882a593Smuzhiyun if (cmd_blk->complete) {
2102*4882a593Smuzhiyun complete(cmd_blk->complete);
2103*4882a593Smuzhiyun cmd_blk->complete = NULL;
2104*4882a593Smuzhiyun }
2105*4882a593Smuzhiyun }
2106*4882a593Smuzhiyun
myrs_monitor(struct work_struct * work)2107*4882a593Smuzhiyun static void myrs_monitor(struct work_struct *work)
2108*4882a593Smuzhiyun {
2109*4882a593Smuzhiyun struct myrs_hba *cs = container_of(work, struct myrs_hba,
2110*4882a593Smuzhiyun monitor_work.work);
2111*4882a593Smuzhiyun struct Scsi_Host *shost = cs->host;
2112*4882a593Smuzhiyun struct myrs_ctlr_info *info = cs->ctlr_info;
2113*4882a593Smuzhiyun unsigned int epoch = cs->fwstat_buf->epoch;
2114*4882a593Smuzhiyun unsigned long interval = MYRS_PRIMARY_MONITOR_INTERVAL;
2115*4882a593Smuzhiyun unsigned char status;
2116*4882a593Smuzhiyun
2117*4882a593Smuzhiyun dev_dbg(&shost->shost_gendev, "monitor tick\n");
2118*4882a593Smuzhiyun
2119*4882a593Smuzhiyun status = myrs_get_fwstatus(cs);
2120*4882a593Smuzhiyun
2121*4882a593Smuzhiyun if (cs->needs_update) {
2122*4882a593Smuzhiyun cs->needs_update = false;
2123*4882a593Smuzhiyun mutex_lock(&cs->cinfo_mutex);
2124*4882a593Smuzhiyun status = myrs_get_ctlr_info(cs);
2125*4882a593Smuzhiyun mutex_unlock(&cs->cinfo_mutex);
2126*4882a593Smuzhiyun }
2127*4882a593Smuzhiyun if (cs->fwstat_buf->next_evseq - cs->next_evseq > 0) {
2128*4882a593Smuzhiyun status = myrs_get_event(cs, cs->next_evseq,
2129*4882a593Smuzhiyun cs->event_buf);
2130*4882a593Smuzhiyun if (status == MYRS_STATUS_SUCCESS) {
2131*4882a593Smuzhiyun myrs_log_event(cs, cs->event_buf);
2132*4882a593Smuzhiyun cs->next_evseq++;
2133*4882a593Smuzhiyun interval = 1;
2134*4882a593Smuzhiyun }
2135*4882a593Smuzhiyun }
2136*4882a593Smuzhiyun
2137*4882a593Smuzhiyun if (time_after(jiffies, cs->secondary_monitor_time
2138*4882a593Smuzhiyun + MYRS_SECONDARY_MONITOR_INTERVAL))
2139*4882a593Smuzhiyun cs->secondary_monitor_time = jiffies;
2140*4882a593Smuzhiyun
2141*4882a593Smuzhiyun if (info->bg_init_active +
2142*4882a593Smuzhiyun info->ldev_init_active +
2143*4882a593Smuzhiyun info->pdev_init_active +
2144*4882a593Smuzhiyun info->cc_active +
2145*4882a593Smuzhiyun info->rbld_active +
2146*4882a593Smuzhiyun info->exp_active != 0) {
2147*4882a593Smuzhiyun struct scsi_device *sdev;
2148*4882a593Smuzhiyun
2149*4882a593Smuzhiyun shost_for_each_device(sdev, shost) {
2150*4882a593Smuzhiyun struct myrs_ldev_info *ldev_info;
2151*4882a593Smuzhiyun int ldev_num;
2152*4882a593Smuzhiyun
2153*4882a593Smuzhiyun if (sdev->channel < info->physchan_present)
2154*4882a593Smuzhiyun continue;
2155*4882a593Smuzhiyun ldev_info = sdev->hostdata;
2156*4882a593Smuzhiyun if (!ldev_info)
2157*4882a593Smuzhiyun continue;
2158*4882a593Smuzhiyun ldev_num = ldev_info->ldev_num;
2159*4882a593Smuzhiyun myrs_get_ldev_info(cs, ldev_num, ldev_info);
2160*4882a593Smuzhiyun }
2161*4882a593Smuzhiyun cs->needs_update = true;
2162*4882a593Smuzhiyun }
2163*4882a593Smuzhiyun if (epoch == cs->epoch &&
2164*4882a593Smuzhiyun cs->fwstat_buf->next_evseq == cs->next_evseq &&
2165*4882a593Smuzhiyun (cs->needs_update == false ||
2166*4882a593Smuzhiyun time_before(jiffies, cs->primary_monitor_time
2167*4882a593Smuzhiyun + MYRS_PRIMARY_MONITOR_INTERVAL))) {
2168*4882a593Smuzhiyun interval = MYRS_SECONDARY_MONITOR_INTERVAL;
2169*4882a593Smuzhiyun }
2170*4882a593Smuzhiyun
2171*4882a593Smuzhiyun if (interval > 1)
2172*4882a593Smuzhiyun cs->primary_monitor_time = jiffies;
2173*4882a593Smuzhiyun queue_delayed_work(cs->work_q, &cs->monitor_work, interval);
2174*4882a593Smuzhiyun }
2175*4882a593Smuzhiyun
myrs_create_mempools(struct pci_dev * pdev,struct myrs_hba * cs)2176*4882a593Smuzhiyun static bool myrs_create_mempools(struct pci_dev *pdev, struct myrs_hba *cs)
2177*4882a593Smuzhiyun {
2178*4882a593Smuzhiyun struct Scsi_Host *shost = cs->host;
2179*4882a593Smuzhiyun size_t elem_size, elem_align;
2180*4882a593Smuzhiyun
2181*4882a593Smuzhiyun elem_align = sizeof(struct myrs_sge);
2182*4882a593Smuzhiyun elem_size = shost->sg_tablesize * elem_align;
2183*4882a593Smuzhiyun cs->sg_pool = dma_pool_create("myrs_sg", &pdev->dev,
2184*4882a593Smuzhiyun elem_size, elem_align, 0);
2185*4882a593Smuzhiyun if (cs->sg_pool == NULL) {
2186*4882a593Smuzhiyun shost_printk(KERN_ERR, shost,
2187*4882a593Smuzhiyun "Failed to allocate SG pool\n");
2188*4882a593Smuzhiyun return false;
2189*4882a593Smuzhiyun }
2190*4882a593Smuzhiyun
2191*4882a593Smuzhiyun cs->sense_pool = dma_pool_create("myrs_sense", &pdev->dev,
2192*4882a593Smuzhiyun MYRS_SENSE_SIZE, sizeof(int), 0);
2193*4882a593Smuzhiyun if (cs->sense_pool == NULL) {
2194*4882a593Smuzhiyun dma_pool_destroy(cs->sg_pool);
2195*4882a593Smuzhiyun cs->sg_pool = NULL;
2196*4882a593Smuzhiyun shost_printk(KERN_ERR, shost,
2197*4882a593Smuzhiyun "Failed to allocate sense data pool\n");
2198*4882a593Smuzhiyun return false;
2199*4882a593Smuzhiyun }
2200*4882a593Smuzhiyun
2201*4882a593Smuzhiyun cs->dcdb_pool = dma_pool_create("myrs_dcdb", &pdev->dev,
2202*4882a593Smuzhiyun MYRS_DCDB_SIZE,
2203*4882a593Smuzhiyun sizeof(unsigned char), 0);
2204*4882a593Smuzhiyun if (!cs->dcdb_pool) {
2205*4882a593Smuzhiyun dma_pool_destroy(cs->sg_pool);
2206*4882a593Smuzhiyun cs->sg_pool = NULL;
2207*4882a593Smuzhiyun dma_pool_destroy(cs->sense_pool);
2208*4882a593Smuzhiyun cs->sense_pool = NULL;
2209*4882a593Smuzhiyun shost_printk(KERN_ERR, shost,
2210*4882a593Smuzhiyun "Failed to allocate DCDB pool\n");
2211*4882a593Smuzhiyun return false;
2212*4882a593Smuzhiyun }
2213*4882a593Smuzhiyun
2214*4882a593Smuzhiyun snprintf(cs->work_q_name, sizeof(cs->work_q_name),
2215*4882a593Smuzhiyun "myrs_wq_%d", shost->host_no);
2216*4882a593Smuzhiyun cs->work_q = create_singlethread_workqueue(cs->work_q_name);
2217*4882a593Smuzhiyun if (!cs->work_q) {
2218*4882a593Smuzhiyun dma_pool_destroy(cs->dcdb_pool);
2219*4882a593Smuzhiyun cs->dcdb_pool = NULL;
2220*4882a593Smuzhiyun dma_pool_destroy(cs->sg_pool);
2221*4882a593Smuzhiyun cs->sg_pool = NULL;
2222*4882a593Smuzhiyun dma_pool_destroy(cs->sense_pool);
2223*4882a593Smuzhiyun cs->sense_pool = NULL;
2224*4882a593Smuzhiyun shost_printk(KERN_ERR, shost,
2225*4882a593Smuzhiyun "Failed to create workqueue\n");
2226*4882a593Smuzhiyun return false;
2227*4882a593Smuzhiyun }
2228*4882a593Smuzhiyun
2229*4882a593Smuzhiyun /* Initialize the Monitoring Timer. */
2230*4882a593Smuzhiyun INIT_DELAYED_WORK(&cs->monitor_work, myrs_monitor);
2231*4882a593Smuzhiyun queue_delayed_work(cs->work_q, &cs->monitor_work, 1);
2232*4882a593Smuzhiyun
2233*4882a593Smuzhiyun return true;
2234*4882a593Smuzhiyun }
2235*4882a593Smuzhiyun
myrs_destroy_mempools(struct myrs_hba * cs)2236*4882a593Smuzhiyun static void myrs_destroy_mempools(struct myrs_hba *cs)
2237*4882a593Smuzhiyun {
2238*4882a593Smuzhiyun cancel_delayed_work_sync(&cs->monitor_work);
2239*4882a593Smuzhiyun destroy_workqueue(cs->work_q);
2240*4882a593Smuzhiyun
2241*4882a593Smuzhiyun dma_pool_destroy(cs->sg_pool);
2242*4882a593Smuzhiyun dma_pool_destroy(cs->dcdb_pool);
2243*4882a593Smuzhiyun dma_pool_destroy(cs->sense_pool);
2244*4882a593Smuzhiyun }
2245*4882a593Smuzhiyun
myrs_unmap(struct myrs_hba * cs)2246*4882a593Smuzhiyun static void myrs_unmap(struct myrs_hba *cs)
2247*4882a593Smuzhiyun {
2248*4882a593Smuzhiyun kfree(cs->event_buf);
2249*4882a593Smuzhiyun kfree(cs->ctlr_info);
2250*4882a593Smuzhiyun if (cs->fwstat_buf) {
2251*4882a593Smuzhiyun dma_free_coherent(&cs->pdev->dev, sizeof(struct myrs_fwstat),
2252*4882a593Smuzhiyun cs->fwstat_buf, cs->fwstat_addr);
2253*4882a593Smuzhiyun cs->fwstat_buf = NULL;
2254*4882a593Smuzhiyun }
2255*4882a593Smuzhiyun if (cs->first_stat_mbox) {
2256*4882a593Smuzhiyun dma_free_coherent(&cs->pdev->dev, cs->stat_mbox_size,
2257*4882a593Smuzhiyun cs->first_stat_mbox, cs->stat_mbox_addr);
2258*4882a593Smuzhiyun cs->first_stat_mbox = NULL;
2259*4882a593Smuzhiyun }
2260*4882a593Smuzhiyun if (cs->first_cmd_mbox) {
2261*4882a593Smuzhiyun dma_free_coherent(&cs->pdev->dev, cs->cmd_mbox_size,
2262*4882a593Smuzhiyun cs->first_cmd_mbox, cs->cmd_mbox_addr);
2263*4882a593Smuzhiyun cs->first_cmd_mbox = NULL;
2264*4882a593Smuzhiyun }
2265*4882a593Smuzhiyun }
2266*4882a593Smuzhiyun
myrs_cleanup(struct myrs_hba * cs)2267*4882a593Smuzhiyun static void myrs_cleanup(struct myrs_hba *cs)
2268*4882a593Smuzhiyun {
2269*4882a593Smuzhiyun struct pci_dev *pdev = cs->pdev;
2270*4882a593Smuzhiyun
2271*4882a593Smuzhiyun /* Free the memory mailbox, status, and related structures */
2272*4882a593Smuzhiyun myrs_unmap(cs);
2273*4882a593Smuzhiyun
2274*4882a593Smuzhiyun if (cs->mmio_base) {
2275*4882a593Smuzhiyun if (cs->disable_intr)
2276*4882a593Smuzhiyun cs->disable_intr(cs);
2277*4882a593Smuzhiyun iounmap(cs->mmio_base);
2278*4882a593Smuzhiyun cs->mmio_base = NULL;
2279*4882a593Smuzhiyun }
2280*4882a593Smuzhiyun if (cs->irq)
2281*4882a593Smuzhiyun free_irq(cs->irq, cs);
2282*4882a593Smuzhiyun if (cs->io_addr)
2283*4882a593Smuzhiyun release_region(cs->io_addr, 0x80);
2284*4882a593Smuzhiyun pci_set_drvdata(pdev, NULL);
2285*4882a593Smuzhiyun pci_disable_device(pdev);
2286*4882a593Smuzhiyun scsi_host_put(cs->host);
2287*4882a593Smuzhiyun }
2288*4882a593Smuzhiyun
myrs_detect(struct pci_dev * pdev,const struct pci_device_id * entry)2289*4882a593Smuzhiyun static struct myrs_hba *myrs_detect(struct pci_dev *pdev,
2290*4882a593Smuzhiyun const struct pci_device_id *entry)
2291*4882a593Smuzhiyun {
2292*4882a593Smuzhiyun struct myrs_privdata *privdata =
2293*4882a593Smuzhiyun (struct myrs_privdata *)entry->driver_data;
2294*4882a593Smuzhiyun irq_handler_t irq_handler = privdata->irq_handler;
2295*4882a593Smuzhiyun unsigned int mmio_size = privdata->mmio_size;
2296*4882a593Smuzhiyun struct myrs_hba *cs = NULL;
2297*4882a593Smuzhiyun
2298*4882a593Smuzhiyun cs = myrs_alloc_host(pdev, entry);
2299*4882a593Smuzhiyun if (!cs) {
2300*4882a593Smuzhiyun dev_err(&pdev->dev, "Unable to allocate Controller\n");
2301*4882a593Smuzhiyun return NULL;
2302*4882a593Smuzhiyun }
2303*4882a593Smuzhiyun cs->pdev = pdev;
2304*4882a593Smuzhiyun
2305*4882a593Smuzhiyun if (pci_enable_device(pdev))
2306*4882a593Smuzhiyun goto Failure;
2307*4882a593Smuzhiyun
2308*4882a593Smuzhiyun cs->pci_addr = pci_resource_start(pdev, 0);
2309*4882a593Smuzhiyun
2310*4882a593Smuzhiyun pci_set_drvdata(pdev, cs);
2311*4882a593Smuzhiyun spin_lock_init(&cs->queue_lock);
2312*4882a593Smuzhiyun /* Map the Controller Register Window. */
2313*4882a593Smuzhiyun if (mmio_size < PAGE_SIZE)
2314*4882a593Smuzhiyun mmio_size = PAGE_SIZE;
2315*4882a593Smuzhiyun cs->mmio_base = ioremap(cs->pci_addr & PAGE_MASK, mmio_size);
2316*4882a593Smuzhiyun if (cs->mmio_base == NULL) {
2317*4882a593Smuzhiyun dev_err(&pdev->dev,
2318*4882a593Smuzhiyun "Unable to map Controller Register Window\n");
2319*4882a593Smuzhiyun goto Failure;
2320*4882a593Smuzhiyun }
2321*4882a593Smuzhiyun
2322*4882a593Smuzhiyun cs->io_base = cs->mmio_base + (cs->pci_addr & ~PAGE_MASK);
2323*4882a593Smuzhiyun if (privdata->hw_init(pdev, cs, cs->io_base))
2324*4882a593Smuzhiyun goto Failure;
2325*4882a593Smuzhiyun
2326*4882a593Smuzhiyun /* Acquire shared access to the IRQ Channel. */
2327*4882a593Smuzhiyun if (request_irq(pdev->irq, irq_handler, IRQF_SHARED, "myrs", cs) < 0) {
2328*4882a593Smuzhiyun dev_err(&pdev->dev,
2329*4882a593Smuzhiyun "Unable to acquire IRQ Channel %d\n", pdev->irq);
2330*4882a593Smuzhiyun goto Failure;
2331*4882a593Smuzhiyun }
2332*4882a593Smuzhiyun cs->irq = pdev->irq;
2333*4882a593Smuzhiyun return cs;
2334*4882a593Smuzhiyun
2335*4882a593Smuzhiyun Failure:
2336*4882a593Smuzhiyun dev_err(&pdev->dev,
2337*4882a593Smuzhiyun "Failed to initialize Controller\n");
2338*4882a593Smuzhiyun myrs_cleanup(cs);
2339*4882a593Smuzhiyun return NULL;
2340*4882a593Smuzhiyun }
2341*4882a593Smuzhiyun
2342*4882a593Smuzhiyun /*
2343*4882a593Smuzhiyun * myrs_err_status reports Controller BIOS Messages passed through
2344*4882a593Smuzhiyun * the Error Status Register when the driver performs the BIOS handshaking.
2345*4882a593Smuzhiyun * It returns true for fatal errors and false otherwise.
2346*4882a593Smuzhiyun */
2347*4882a593Smuzhiyun
myrs_err_status(struct myrs_hba * cs,unsigned char status,unsigned char parm0,unsigned char parm1)2348*4882a593Smuzhiyun static bool myrs_err_status(struct myrs_hba *cs, unsigned char status,
2349*4882a593Smuzhiyun unsigned char parm0, unsigned char parm1)
2350*4882a593Smuzhiyun {
2351*4882a593Smuzhiyun struct pci_dev *pdev = cs->pdev;
2352*4882a593Smuzhiyun
2353*4882a593Smuzhiyun switch (status) {
2354*4882a593Smuzhiyun case 0x00:
2355*4882a593Smuzhiyun dev_info(&pdev->dev,
2356*4882a593Smuzhiyun "Physical Device %d:%d Not Responding\n",
2357*4882a593Smuzhiyun parm1, parm0);
2358*4882a593Smuzhiyun break;
2359*4882a593Smuzhiyun case 0x08:
2360*4882a593Smuzhiyun dev_notice(&pdev->dev, "Spinning Up Drives\n");
2361*4882a593Smuzhiyun break;
2362*4882a593Smuzhiyun case 0x30:
2363*4882a593Smuzhiyun dev_notice(&pdev->dev, "Configuration Checksum Error\n");
2364*4882a593Smuzhiyun break;
2365*4882a593Smuzhiyun case 0x60:
2366*4882a593Smuzhiyun dev_notice(&pdev->dev, "Mirror Race Recovery Failed\n");
2367*4882a593Smuzhiyun break;
2368*4882a593Smuzhiyun case 0x70:
2369*4882a593Smuzhiyun dev_notice(&pdev->dev, "Mirror Race Recovery In Progress\n");
2370*4882a593Smuzhiyun break;
2371*4882a593Smuzhiyun case 0x90:
2372*4882a593Smuzhiyun dev_notice(&pdev->dev, "Physical Device %d:%d COD Mismatch\n",
2373*4882a593Smuzhiyun parm1, parm0);
2374*4882a593Smuzhiyun break;
2375*4882a593Smuzhiyun case 0xA0:
2376*4882a593Smuzhiyun dev_notice(&pdev->dev, "Logical Drive Installation Aborted\n");
2377*4882a593Smuzhiyun break;
2378*4882a593Smuzhiyun case 0xB0:
2379*4882a593Smuzhiyun dev_notice(&pdev->dev, "Mirror Race On A Critical Logical Drive\n");
2380*4882a593Smuzhiyun break;
2381*4882a593Smuzhiyun case 0xD0:
2382*4882a593Smuzhiyun dev_notice(&pdev->dev, "New Controller Configuration Found\n");
2383*4882a593Smuzhiyun break;
2384*4882a593Smuzhiyun case 0xF0:
2385*4882a593Smuzhiyun dev_err(&pdev->dev, "Fatal Memory Parity Error\n");
2386*4882a593Smuzhiyun return true;
2387*4882a593Smuzhiyun default:
2388*4882a593Smuzhiyun dev_err(&pdev->dev, "Unknown Initialization Error %02X\n",
2389*4882a593Smuzhiyun status);
2390*4882a593Smuzhiyun return true;
2391*4882a593Smuzhiyun }
2392*4882a593Smuzhiyun return false;
2393*4882a593Smuzhiyun }
2394*4882a593Smuzhiyun
2395*4882a593Smuzhiyun /*
2396*4882a593Smuzhiyun * Hardware-specific functions
2397*4882a593Smuzhiyun */
2398*4882a593Smuzhiyun
2399*4882a593Smuzhiyun /*
2400*4882a593Smuzhiyun * DAC960 GEM Series Controllers.
2401*4882a593Smuzhiyun */
2402*4882a593Smuzhiyun
DAC960_GEM_hw_mbox_new_cmd(void __iomem * base)2403*4882a593Smuzhiyun static inline void DAC960_GEM_hw_mbox_new_cmd(void __iomem *base)
2404*4882a593Smuzhiyun {
2405*4882a593Smuzhiyun __le32 val = cpu_to_le32(DAC960_GEM_IDB_HWMBOX_NEW_CMD << 24);
2406*4882a593Smuzhiyun
2407*4882a593Smuzhiyun writel(val, base + DAC960_GEM_IDB_READ_OFFSET);
2408*4882a593Smuzhiyun }
2409*4882a593Smuzhiyun
DAC960_GEM_ack_hw_mbox_status(void __iomem * base)2410*4882a593Smuzhiyun static inline void DAC960_GEM_ack_hw_mbox_status(void __iomem *base)
2411*4882a593Smuzhiyun {
2412*4882a593Smuzhiyun __le32 val = cpu_to_le32(DAC960_GEM_IDB_HWMBOX_ACK_STS << 24);
2413*4882a593Smuzhiyun
2414*4882a593Smuzhiyun writel(val, base + DAC960_GEM_IDB_CLEAR_OFFSET);
2415*4882a593Smuzhiyun }
2416*4882a593Smuzhiyun
DAC960_GEM_gen_intr(void __iomem * base)2417*4882a593Smuzhiyun static inline void DAC960_GEM_gen_intr(void __iomem *base)
2418*4882a593Smuzhiyun {
2419*4882a593Smuzhiyun __le32 val = cpu_to_le32(DAC960_GEM_IDB_GEN_IRQ << 24);
2420*4882a593Smuzhiyun
2421*4882a593Smuzhiyun writel(val, base + DAC960_GEM_IDB_READ_OFFSET);
2422*4882a593Smuzhiyun }
2423*4882a593Smuzhiyun
DAC960_GEM_reset_ctrl(void __iomem * base)2424*4882a593Smuzhiyun static inline void DAC960_GEM_reset_ctrl(void __iomem *base)
2425*4882a593Smuzhiyun {
2426*4882a593Smuzhiyun __le32 val = cpu_to_le32(DAC960_GEM_IDB_CTRL_RESET << 24);
2427*4882a593Smuzhiyun
2428*4882a593Smuzhiyun writel(val, base + DAC960_GEM_IDB_READ_OFFSET);
2429*4882a593Smuzhiyun }
2430*4882a593Smuzhiyun
DAC960_GEM_mem_mbox_new_cmd(void __iomem * base)2431*4882a593Smuzhiyun static inline void DAC960_GEM_mem_mbox_new_cmd(void __iomem *base)
2432*4882a593Smuzhiyun {
2433*4882a593Smuzhiyun __le32 val = cpu_to_le32(DAC960_GEM_IDB_HWMBOX_NEW_CMD << 24);
2434*4882a593Smuzhiyun
2435*4882a593Smuzhiyun writel(val, base + DAC960_GEM_IDB_READ_OFFSET);
2436*4882a593Smuzhiyun }
2437*4882a593Smuzhiyun
DAC960_GEM_hw_mbox_is_full(void __iomem * base)2438*4882a593Smuzhiyun static inline bool DAC960_GEM_hw_mbox_is_full(void __iomem *base)
2439*4882a593Smuzhiyun {
2440*4882a593Smuzhiyun __le32 val;
2441*4882a593Smuzhiyun
2442*4882a593Smuzhiyun val = readl(base + DAC960_GEM_IDB_READ_OFFSET);
2443*4882a593Smuzhiyun return (le32_to_cpu(val) >> 24) & DAC960_GEM_IDB_HWMBOX_FULL;
2444*4882a593Smuzhiyun }
2445*4882a593Smuzhiyun
DAC960_GEM_init_in_progress(void __iomem * base)2446*4882a593Smuzhiyun static inline bool DAC960_GEM_init_in_progress(void __iomem *base)
2447*4882a593Smuzhiyun {
2448*4882a593Smuzhiyun __le32 val;
2449*4882a593Smuzhiyun
2450*4882a593Smuzhiyun val = readl(base + DAC960_GEM_IDB_READ_OFFSET);
2451*4882a593Smuzhiyun return (le32_to_cpu(val) >> 24) & DAC960_GEM_IDB_INIT_IN_PROGRESS;
2452*4882a593Smuzhiyun }
2453*4882a593Smuzhiyun
DAC960_GEM_ack_hw_mbox_intr(void __iomem * base)2454*4882a593Smuzhiyun static inline void DAC960_GEM_ack_hw_mbox_intr(void __iomem *base)
2455*4882a593Smuzhiyun {
2456*4882a593Smuzhiyun __le32 val = cpu_to_le32(DAC960_GEM_ODB_HWMBOX_ACK_IRQ << 24);
2457*4882a593Smuzhiyun
2458*4882a593Smuzhiyun writel(val, base + DAC960_GEM_ODB_CLEAR_OFFSET);
2459*4882a593Smuzhiyun }
2460*4882a593Smuzhiyun
DAC960_GEM_ack_mem_mbox_intr(void __iomem * base)2461*4882a593Smuzhiyun static inline void DAC960_GEM_ack_mem_mbox_intr(void __iomem *base)
2462*4882a593Smuzhiyun {
2463*4882a593Smuzhiyun __le32 val = cpu_to_le32(DAC960_GEM_ODB_MMBOX_ACK_IRQ << 24);
2464*4882a593Smuzhiyun
2465*4882a593Smuzhiyun writel(val, base + DAC960_GEM_ODB_CLEAR_OFFSET);
2466*4882a593Smuzhiyun }
2467*4882a593Smuzhiyun
DAC960_GEM_ack_intr(void __iomem * base)2468*4882a593Smuzhiyun static inline void DAC960_GEM_ack_intr(void __iomem *base)
2469*4882a593Smuzhiyun {
2470*4882a593Smuzhiyun __le32 val = cpu_to_le32((DAC960_GEM_ODB_HWMBOX_ACK_IRQ |
2471*4882a593Smuzhiyun DAC960_GEM_ODB_MMBOX_ACK_IRQ) << 24);
2472*4882a593Smuzhiyun
2473*4882a593Smuzhiyun writel(val, base + DAC960_GEM_ODB_CLEAR_OFFSET);
2474*4882a593Smuzhiyun }
2475*4882a593Smuzhiyun
DAC960_GEM_hw_mbox_status_available(void __iomem * base)2476*4882a593Smuzhiyun static inline bool DAC960_GEM_hw_mbox_status_available(void __iomem *base)
2477*4882a593Smuzhiyun {
2478*4882a593Smuzhiyun __le32 val;
2479*4882a593Smuzhiyun
2480*4882a593Smuzhiyun val = readl(base + DAC960_GEM_ODB_READ_OFFSET);
2481*4882a593Smuzhiyun return (le32_to_cpu(val) >> 24) & DAC960_GEM_ODB_HWMBOX_STS_AVAIL;
2482*4882a593Smuzhiyun }
2483*4882a593Smuzhiyun
DAC960_GEM_mem_mbox_status_available(void __iomem * base)2484*4882a593Smuzhiyun static inline bool DAC960_GEM_mem_mbox_status_available(void __iomem *base)
2485*4882a593Smuzhiyun {
2486*4882a593Smuzhiyun __le32 val;
2487*4882a593Smuzhiyun
2488*4882a593Smuzhiyun val = readl(base + DAC960_GEM_ODB_READ_OFFSET);
2489*4882a593Smuzhiyun return (le32_to_cpu(val) >> 24) & DAC960_GEM_ODB_MMBOX_STS_AVAIL;
2490*4882a593Smuzhiyun }
2491*4882a593Smuzhiyun
DAC960_GEM_enable_intr(void __iomem * base)2492*4882a593Smuzhiyun static inline void DAC960_GEM_enable_intr(void __iomem *base)
2493*4882a593Smuzhiyun {
2494*4882a593Smuzhiyun __le32 val = cpu_to_le32((DAC960_GEM_IRQMASK_HWMBOX_IRQ |
2495*4882a593Smuzhiyun DAC960_GEM_IRQMASK_MMBOX_IRQ) << 24);
2496*4882a593Smuzhiyun writel(val, base + DAC960_GEM_IRQMASK_CLEAR_OFFSET);
2497*4882a593Smuzhiyun }
2498*4882a593Smuzhiyun
DAC960_GEM_disable_intr(void __iomem * base)2499*4882a593Smuzhiyun static inline void DAC960_GEM_disable_intr(void __iomem *base)
2500*4882a593Smuzhiyun {
2501*4882a593Smuzhiyun __le32 val = 0;
2502*4882a593Smuzhiyun
2503*4882a593Smuzhiyun writel(val, base + DAC960_GEM_IRQMASK_READ_OFFSET);
2504*4882a593Smuzhiyun }
2505*4882a593Smuzhiyun
DAC960_GEM_intr_enabled(void __iomem * base)2506*4882a593Smuzhiyun static inline bool DAC960_GEM_intr_enabled(void __iomem *base)
2507*4882a593Smuzhiyun {
2508*4882a593Smuzhiyun __le32 val;
2509*4882a593Smuzhiyun
2510*4882a593Smuzhiyun val = readl(base + DAC960_GEM_IRQMASK_READ_OFFSET);
2511*4882a593Smuzhiyun return !((le32_to_cpu(val) >> 24) &
2512*4882a593Smuzhiyun (DAC960_GEM_IRQMASK_HWMBOX_IRQ |
2513*4882a593Smuzhiyun DAC960_GEM_IRQMASK_MMBOX_IRQ));
2514*4882a593Smuzhiyun }
2515*4882a593Smuzhiyun
DAC960_GEM_write_cmd_mbox(union myrs_cmd_mbox * mem_mbox,union myrs_cmd_mbox * mbox)2516*4882a593Smuzhiyun static inline void DAC960_GEM_write_cmd_mbox(union myrs_cmd_mbox *mem_mbox,
2517*4882a593Smuzhiyun union myrs_cmd_mbox *mbox)
2518*4882a593Smuzhiyun {
2519*4882a593Smuzhiyun memcpy(&mem_mbox->words[1], &mbox->words[1],
2520*4882a593Smuzhiyun sizeof(union myrs_cmd_mbox) - sizeof(unsigned int));
2521*4882a593Smuzhiyun /* Barrier to avoid reordering */
2522*4882a593Smuzhiyun wmb();
2523*4882a593Smuzhiyun mem_mbox->words[0] = mbox->words[0];
2524*4882a593Smuzhiyun /* Barrier to force PCI access */
2525*4882a593Smuzhiyun mb();
2526*4882a593Smuzhiyun }
2527*4882a593Smuzhiyun
DAC960_GEM_write_hw_mbox(void __iomem * base,dma_addr_t cmd_mbox_addr)2528*4882a593Smuzhiyun static inline void DAC960_GEM_write_hw_mbox(void __iomem *base,
2529*4882a593Smuzhiyun dma_addr_t cmd_mbox_addr)
2530*4882a593Smuzhiyun {
2531*4882a593Smuzhiyun dma_addr_writeql(cmd_mbox_addr, base + DAC960_GEM_CMDMBX_OFFSET);
2532*4882a593Smuzhiyun }
2533*4882a593Smuzhiyun
DAC960_GEM_read_cmd_ident(void __iomem * base)2534*4882a593Smuzhiyun static inline unsigned short DAC960_GEM_read_cmd_ident(void __iomem *base)
2535*4882a593Smuzhiyun {
2536*4882a593Smuzhiyun return readw(base + DAC960_GEM_CMDSTS_OFFSET);
2537*4882a593Smuzhiyun }
2538*4882a593Smuzhiyun
DAC960_GEM_read_cmd_status(void __iomem * base)2539*4882a593Smuzhiyun static inline unsigned char DAC960_GEM_read_cmd_status(void __iomem *base)
2540*4882a593Smuzhiyun {
2541*4882a593Smuzhiyun return readw(base + DAC960_GEM_CMDSTS_OFFSET + 2);
2542*4882a593Smuzhiyun }
2543*4882a593Smuzhiyun
2544*4882a593Smuzhiyun static inline bool
DAC960_GEM_read_error_status(void __iomem * base,unsigned char * error,unsigned char * param0,unsigned char * param1)2545*4882a593Smuzhiyun DAC960_GEM_read_error_status(void __iomem *base, unsigned char *error,
2546*4882a593Smuzhiyun unsigned char *param0, unsigned char *param1)
2547*4882a593Smuzhiyun {
2548*4882a593Smuzhiyun __le32 val;
2549*4882a593Smuzhiyun
2550*4882a593Smuzhiyun val = readl(base + DAC960_GEM_ERRSTS_READ_OFFSET);
2551*4882a593Smuzhiyun if (!((le32_to_cpu(val) >> 24) & DAC960_GEM_ERRSTS_PENDING))
2552*4882a593Smuzhiyun return false;
2553*4882a593Smuzhiyun *error = val & ~(DAC960_GEM_ERRSTS_PENDING << 24);
2554*4882a593Smuzhiyun *param0 = readb(base + DAC960_GEM_CMDMBX_OFFSET + 0);
2555*4882a593Smuzhiyun *param1 = readb(base + DAC960_GEM_CMDMBX_OFFSET + 1);
2556*4882a593Smuzhiyun writel(0x03000000, base + DAC960_GEM_ERRSTS_CLEAR_OFFSET);
2557*4882a593Smuzhiyun return true;
2558*4882a593Smuzhiyun }
2559*4882a593Smuzhiyun
2560*4882a593Smuzhiyun static inline unsigned char
DAC960_GEM_mbox_init(void __iomem * base,dma_addr_t mbox_addr)2561*4882a593Smuzhiyun DAC960_GEM_mbox_init(void __iomem *base, dma_addr_t mbox_addr)
2562*4882a593Smuzhiyun {
2563*4882a593Smuzhiyun unsigned char status;
2564*4882a593Smuzhiyun
2565*4882a593Smuzhiyun while (DAC960_GEM_hw_mbox_is_full(base))
2566*4882a593Smuzhiyun udelay(1);
2567*4882a593Smuzhiyun DAC960_GEM_write_hw_mbox(base, mbox_addr);
2568*4882a593Smuzhiyun DAC960_GEM_hw_mbox_new_cmd(base);
2569*4882a593Smuzhiyun while (!DAC960_GEM_hw_mbox_status_available(base))
2570*4882a593Smuzhiyun udelay(1);
2571*4882a593Smuzhiyun status = DAC960_GEM_read_cmd_status(base);
2572*4882a593Smuzhiyun DAC960_GEM_ack_hw_mbox_intr(base);
2573*4882a593Smuzhiyun DAC960_GEM_ack_hw_mbox_status(base);
2574*4882a593Smuzhiyun
2575*4882a593Smuzhiyun return status;
2576*4882a593Smuzhiyun }
2577*4882a593Smuzhiyun
DAC960_GEM_hw_init(struct pci_dev * pdev,struct myrs_hba * cs,void __iomem * base)2578*4882a593Smuzhiyun static int DAC960_GEM_hw_init(struct pci_dev *pdev,
2579*4882a593Smuzhiyun struct myrs_hba *cs, void __iomem *base)
2580*4882a593Smuzhiyun {
2581*4882a593Smuzhiyun int timeout = 0;
2582*4882a593Smuzhiyun unsigned char status, parm0, parm1;
2583*4882a593Smuzhiyun
2584*4882a593Smuzhiyun DAC960_GEM_disable_intr(base);
2585*4882a593Smuzhiyun DAC960_GEM_ack_hw_mbox_status(base);
2586*4882a593Smuzhiyun udelay(1000);
2587*4882a593Smuzhiyun while (DAC960_GEM_init_in_progress(base) &&
2588*4882a593Smuzhiyun timeout < MYRS_MAILBOX_TIMEOUT) {
2589*4882a593Smuzhiyun if (DAC960_GEM_read_error_status(base, &status,
2590*4882a593Smuzhiyun &parm0, &parm1) &&
2591*4882a593Smuzhiyun myrs_err_status(cs, status, parm0, parm1))
2592*4882a593Smuzhiyun return -EIO;
2593*4882a593Smuzhiyun udelay(10);
2594*4882a593Smuzhiyun timeout++;
2595*4882a593Smuzhiyun }
2596*4882a593Smuzhiyun if (timeout == MYRS_MAILBOX_TIMEOUT) {
2597*4882a593Smuzhiyun dev_err(&pdev->dev,
2598*4882a593Smuzhiyun "Timeout waiting for Controller Initialisation\n");
2599*4882a593Smuzhiyun return -ETIMEDOUT;
2600*4882a593Smuzhiyun }
2601*4882a593Smuzhiyun if (!myrs_enable_mmio_mbox(cs, DAC960_GEM_mbox_init)) {
2602*4882a593Smuzhiyun dev_err(&pdev->dev,
2603*4882a593Smuzhiyun "Unable to Enable Memory Mailbox Interface\n");
2604*4882a593Smuzhiyun DAC960_GEM_reset_ctrl(base);
2605*4882a593Smuzhiyun return -EAGAIN;
2606*4882a593Smuzhiyun }
2607*4882a593Smuzhiyun DAC960_GEM_enable_intr(base);
2608*4882a593Smuzhiyun cs->write_cmd_mbox = DAC960_GEM_write_cmd_mbox;
2609*4882a593Smuzhiyun cs->get_cmd_mbox = DAC960_GEM_mem_mbox_new_cmd;
2610*4882a593Smuzhiyun cs->disable_intr = DAC960_GEM_disable_intr;
2611*4882a593Smuzhiyun cs->reset = DAC960_GEM_reset_ctrl;
2612*4882a593Smuzhiyun return 0;
2613*4882a593Smuzhiyun }
2614*4882a593Smuzhiyun
DAC960_GEM_intr_handler(int irq,void * arg)2615*4882a593Smuzhiyun static irqreturn_t DAC960_GEM_intr_handler(int irq, void *arg)
2616*4882a593Smuzhiyun {
2617*4882a593Smuzhiyun struct myrs_hba *cs = arg;
2618*4882a593Smuzhiyun void __iomem *base = cs->io_base;
2619*4882a593Smuzhiyun struct myrs_stat_mbox *next_stat_mbox;
2620*4882a593Smuzhiyun unsigned long flags;
2621*4882a593Smuzhiyun
2622*4882a593Smuzhiyun spin_lock_irqsave(&cs->queue_lock, flags);
2623*4882a593Smuzhiyun DAC960_GEM_ack_intr(base);
2624*4882a593Smuzhiyun next_stat_mbox = cs->next_stat_mbox;
2625*4882a593Smuzhiyun while (next_stat_mbox->id > 0) {
2626*4882a593Smuzhiyun unsigned short id = next_stat_mbox->id;
2627*4882a593Smuzhiyun struct scsi_cmnd *scmd = NULL;
2628*4882a593Smuzhiyun struct myrs_cmdblk *cmd_blk = NULL;
2629*4882a593Smuzhiyun
2630*4882a593Smuzhiyun if (id == MYRS_DCMD_TAG)
2631*4882a593Smuzhiyun cmd_blk = &cs->dcmd_blk;
2632*4882a593Smuzhiyun else if (id == MYRS_MCMD_TAG)
2633*4882a593Smuzhiyun cmd_blk = &cs->mcmd_blk;
2634*4882a593Smuzhiyun else {
2635*4882a593Smuzhiyun scmd = scsi_host_find_tag(cs->host, id - 3);
2636*4882a593Smuzhiyun if (scmd)
2637*4882a593Smuzhiyun cmd_blk = scsi_cmd_priv(scmd);
2638*4882a593Smuzhiyun }
2639*4882a593Smuzhiyun if (cmd_blk) {
2640*4882a593Smuzhiyun cmd_blk->status = next_stat_mbox->status;
2641*4882a593Smuzhiyun cmd_blk->sense_len = next_stat_mbox->sense_len;
2642*4882a593Smuzhiyun cmd_blk->residual = next_stat_mbox->residual;
2643*4882a593Smuzhiyun } else
2644*4882a593Smuzhiyun dev_err(&cs->pdev->dev,
2645*4882a593Smuzhiyun "Unhandled command completion %d\n", id);
2646*4882a593Smuzhiyun
2647*4882a593Smuzhiyun memset(next_stat_mbox, 0, sizeof(struct myrs_stat_mbox));
2648*4882a593Smuzhiyun if (++next_stat_mbox > cs->last_stat_mbox)
2649*4882a593Smuzhiyun next_stat_mbox = cs->first_stat_mbox;
2650*4882a593Smuzhiyun
2651*4882a593Smuzhiyun if (cmd_blk) {
2652*4882a593Smuzhiyun if (id < 3)
2653*4882a593Smuzhiyun myrs_handle_cmdblk(cs, cmd_blk);
2654*4882a593Smuzhiyun else
2655*4882a593Smuzhiyun myrs_handle_scsi(cs, cmd_blk, scmd);
2656*4882a593Smuzhiyun }
2657*4882a593Smuzhiyun }
2658*4882a593Smuzhiyun cs->next_stat_mbox = next_stat_mbox;
2659*4882a593Smuzhiyun spin_unlock_irqrestore(&cs->queue_lock, flags);
2660*4882a593Smuzhiyun return IRQ_HANDLED;
2661*4882a593Smuzhiyun }
2662*4882a593Smuzhiyun
2663*4882a593Smuzhiyun struct myrs_privdata DAC960_GEM_privdata = {
2664*4882a593Smuzhiyun .hw_init = DAC960_GEM_hw_init,
2665*4882a593Smuzhiyun .irq_handler = DAC960_GEM_intr_handler,
2666*4882a593Smuzhiyun .mmio_size = DAC960_GEM_mmio_size,
2667*4882a593Smuzhiyun };
2668*4882a593Smuzhiyun
2669*4882a593Smuzhiyun /*
2670*4882a593Smuzhiyun * DAC960 BA Series Controllers.
2671*4882a593Smuzhiyun */
2672*4882a593Smuzhiyun
DAC960_BA_hw_mbox_new_cmd(void __iomem * base)2673*4882a593Smuzhiyun static inline void DAC960_BA_hw_mbox_new_cmd(void __iomem *base)
2674*4882a593Smuzhiyun {
2675*4882a593Smuzhiyun writeb(DAC960_BA_IDB_HWMBOX_NEW_CMD, base + DAC960_BA_IDB_OFFSET);
2676*4882a593Smuzhiyun }
2677*4882a593Smuzhiyun
DAC960_BA_ack_hw_mbox_status(void __iomem * base)2678*4882a593Smuzhiyun static inline void DAC960_BA_ack_hw_mbox_status(void __iomem *base)
2679*4882a593Smuzhiyun {
2680*4882a593Smuzhiyun writeb(DAC960_BA_IDB_HWMBOX_ACK_STS, base + DAC960_BA_IDB_OFFSET);
2681*4882a593Smuzhiyun }
2682*4882a593Smuzhiyun
DAC960_BA_gen_intr(void __iomem * base)2683*4882a593Smuzhiyun static inline void DAC960_BA_gen_intr(void __iomem *base)
2684*4882a593Smuzhiyun {
2685*4882a593Smuzhiyun writeb(DAC960_BA_IDB_GEN_IRQ, base + DAC960_BA_IDB_OFFSET);
2686*4882a593Smuzhiyun }
2687*4882a593Smuzhiyun
DAC960_BA_reset_ctrl(void __iomem * base)2688*4882a593Smuzhiyun static inline void DAC960_BA_reset_ctrl(void __iomem *base)
2689*4882a593Smuzhiyun {
2690*4882a593Smuzhiyun writeb(DAC960_BA_IDB_CTRL_RESET, base + DAC960_BA_IDB_OFFSET);
2691*4882a593Smuzhiyun }
2692*4882a593Smuzhiyun
DAC960_BA_mem_mbox_new_cmd(void __iomem * base)2693*4882a593Smuzhiyun static inline void DAC960_BA_mem_mbox_new_cmd(void __iomem *base)
2694*4882a593Smuzhiyun {
2695*4882a593Smuzhiyun writeb(DAC960_BA_IDB_MMBOX_NEW_CMD, base + DAC960_BA_IDB_OFFSET);
2696*4882a593Smuzhiyun }
2697*4882a593Smuzhiyun
DAC960_BA_hw_mbox_is_full(void __iomem * base)2698*4882a593Smuzhiyun static inline bool DAC960_BA_hw_mbox_is_full(void __iomem *base)
2699*4882a593Smuzhiyun {
2700*4882a593Smuzhiyun u8 val;
2701*4882a593Smuzhiyun
2702*4882a593Smuzhiyun val = readb(base + DAC960_BA_IDB_OFFSET);
2703*4882a593Smuzhiyun return !(val & DAC960_BA_IDB_HWMBOX_EMPTY);
2704*4882a593Smuzhiyun }
2705*4882a593Smuzhiyun
DAC960_BA_init_in_progress(void __iomem * base)2706*4882a593Smuzhiyun static inline bool DAC960_BA_init_in_progress(void __iomem *base)
2707*4882a593Smuzhiyun {
2708*4882a593Smuzhiyun u8 val;
2709*4882a593Smuzhiyun
2710*4882a593Smuzhiyun val = readb(base + DAC960_BA_IDB_OFFSET);
2711*4882a593Smuzhiyun return !(val & DAC960_BA_IDB_INIT_DONE);
2712*4882a593Smuzhiyun }
2713*4882a593Smuzhiyun
DAC960_BA_ack_hw_mbox_intr(void __iomem * base)2714*4882a593Smuzhiyun static inline void DAC960_BA_ack_hw_mbox_intr(void __iomem *base)
2715*4882a593Smuzhiyun {
2716*4882a593Smuzhiyun writeb(DAC960_BA_ODB_HWMBOX_ACK_IRQ, base + DAC960_BA_ODB_OFFSET);
2717*4882a593Smuzhiyun }
2718*4882a593Smuzhiyun
DAC960_BA_ack_mem_mbox_intr(void __iomem * base)2719*4882a593Smuzhiyun static inline void DAC960_BA_ack_mem_mbox_intr(void __iomem *base)
2720*4882a593Smuzhiyun {
2721*4882a593Smuzhiyun writeb(DAC960_BA_ODB_MMBOX_ACK_IRQ, base + DAC960_BA_ODB_OFFSET);
2722*4882a593Smuzhiyun }
2723*4882a593Smuzhiyun
DAC960_BA_ack_intr(void __iomem * base)2724*4882a593Smuzhiyun static inline void DAC960_BA_ack_intr(void __iomem *base)
2725*4882a593Smuzhiyun {
2726*4882a593Smuzhiyun writeb(DAC960_BA_ODB_HWMBOX_ACK_IRQ | DAC960_BA_ODB_MMBOX_ACK_IRQ,
2727*4882a593Smuzhiyun base + DAC960_BA_ODB_OFFSET);
2728*4882a593Smuzhiyun }
2729*4882a593Smuzhiyun
DAC960_BA_hw_mbox_status_available(void __iomem * base)2730*4882a593Smuzhiyun static inline bool DAC960_BA_hw_mbox_status_available(void __iomem *base)
2731*4882a593Smuzhiyun {
2732*4882a593Smuzhiyun u8 val;
2733*4882a593Smuzhiyun
2734*4882a593Smuzhiyun val = readb(base + DAC960_BA_ODB_OFFSET);
2735*4882a593Smuzhiyun return val & DAC960_BA_ODB_HWMBOX_STS_AVAIL;
2736*4882a593Smuzhiyun }
2737*4882a593Smuzhiyun
DAC960_BA_mem_mbox_status_available(void __iomem * base)2738*4882a593Smuzhiyun static inline bool DAC960_BA_mem_mbox_status_available(void __iomem *base)
2739*4882a593Smuzhiyun {
2740*4882a593Smuzhiyun u8 val;
2741*4882a593Smuzhiyun
2742*4882a593Smuzhiyun val = readb(base + DAC960_BA_ODB_OFFSET);
2743*4882a593Smuzhiyun return val & DAC960_BA_ODB_MMBOX_STS_AVAIL;
2744*4882a593Smuzhiyun }
2745*4882a593Smuzhiyun
DAC960_BA_enable_intr(void __iomem * base)2746*4882a593Smuzhiyun static inline void DAC960_BA_enable_intr(void __iomem *base)
2747*4882a593Smuzhiyun {
2748*4882a593Smuzhiyun writeb(~DAC960_BA_IRQMASK_DISABLE_IRQ, base + DAC960_BA_IRQMASK_OFFSET);
2749*4882a593Smuzhiyun }
2750*4882a593Smuzhiyun
DAC960_BA_disable_intr(void __iomem * base)2751*4882a593Smuzhiyun static inline void DAC960_BA_disable_intr(void __iomem *base)
2752*4882a593Smuzhiyun {
2753*4882a593Smuzhiyun writeb(0xFF, base + DAC960_BA_IRQMASK_OFFSET);
2754*4882a593Smuzhiyun }
2755*4882a593Smuzhiyun
DAC960_BA_intr_enabled(void __iomem * base)2756*4882a593Smuzhiyun static inline bool DAC960_BA_intr_enabled(void __iomem *base)
2757*4882a593Smuzhiyun {
2758*4882a593Smuzhiyun u8 val;
2759*4882a593Smuzhiyun
2760*4882a593Smuzhiyun val = readb(base + DAC960_BA_IRQMASK_OFFSET);
2761*4882a593Smuzhiyun return !(val & DAC960_BA_IRQMASK_DISABLE_IRQ);
2762*4882a593Smuzhiyun }
2763*4882a593Smuzhiyun
DAC960_BA_write_cmd_mbox(union myrs_cmd_mbox * mem_mbox,union myrs_cmd_mbox * mbox)2764*4882a593Smuzhiyun static inline void DAC960_BA_write_cmd_mbox(union myrs_cmd_mbox *mem_mbox,
2765*4882a593Smuzhiyun union myrs_cmd_mbox *mbox)
2766*4882a593Smuzhiyun {
2767*4882a593Smuzhiyun memcpy(&mem_mbox->words[1], &mbox->words[1],
2768*4882a593Smuzhiyun sizeof(union myrs_cmd_mbox) - sizeof(unsigned int));
2769*4882a593Smuzhiyun /* Barrier to avoid reordering */
2770*4882a593Smuzhiyun wmb();
2771*4882a593Smuzhiyun mem_mbox->words[0] = mbox->words[0];
2772*4882a593Smuzhiyun /* Barrier to force PCI access */
2773*4882a593Smuzhiyun mb();
2774*4882a593Smuzhiyun }
2775*4882a593Smuzhiyun
2776*4882a593Smuzhiyun
DAC960_BA_write_hw_mbox(void __iomem * base,dma_addr_t cmd_mbox_addr)2777*4882a593Smuzhiyun static inline void DAC960_BA_write_hw_mbox(void __iomem *base,
2778*4882a593Smuzhiyun dma_addr_t cmd_mbox_addr)
2779*4882a593Smuzhiyun {
2780*4882a593Smuzhiyun dma_addr_writeql(cmd_mbox_addr, base + DAC960_BA_CMDMBX_OFFSET);
2781*4882a593Smuzhiyun }
2782*4882a593Smuzhiyun
DAC960_BA_read_cmd_ident(void __iomem * base)2783*4882a593Smuzhiyun static inline unsigned short DAC960_BA_read_cmd_ident(void __iomem *base)
2784*4882a593Smuzhiyun {
2785*4882a593Smuzhiyun return readw(base + DAC960_BA_CMDSTS_OFFSET);
2786*4882a593Smuzhiyun }
2787*4882a593Smuzhiyun
DAC960_BA_read_cmd_status(void __iomem * base)2788*4882a593Smuzhiyun static inline unsigned char DAC960_BA_read_cmd_status(void __iomem *base)
2789*4882a593Smuzhiyun {
2790*4882a593Smuzhiyun return readw(base + DAC960_BA_CMDSTS_OFFSET + 2);
2791*4882a593Smuzhiyun }
2792*4882a593Smuzhiyun
2793*4882a593Smuzhiyun static inline bool
DAC960_BA_read_error_status(void __iomem * base,unsigned char * error,unsigned char * param0,unsigned char * param1)2794*4882a593Smuzhiyun DAC960_BA_read_error_status(void __iomem *base, unsigned char *error,
2795*4882a593Smuzhiyun unsigned char *param0, unsigned char *param1)
2796*4882a593Smuzhiyun {
2797*4882a593Smuzhiyun u8 val;
2798*4882a593Smuzhiyun
2799*4882a593Smuzhiyun val = readb(base + DAC960_BA_ERRSTS_OFFSET);
2800*4882a593Smuzhiyun if (!(val & DAC960_BA_ERRSTS_PENDING))
2801*4882a593Smuzhiyun return false;
2802*4882a593Smuzhiyun val &= ~DAC960_BA_ERRSTS_PENDING;
2803*4882a593Smuzhiyun *error = val;
2804*4882a593Smuzhiyun *param0 = readb(base + DAC960_BA_CMDMBX_OFFSET + 0);
2805*4882a593Smuzhiyun *param1 = readb(base + DAC960_BA_CMDMBX_OFFSET + 1);
2806*4882a593Smuzhiyun writeb(0xFF, base + DAC960_BA_ERRSTS_OFFSET);
2807*4882a593Smuzhiyun return true;
2808*4882a593Smuzhiyun }
2809*4882a593Smuzhiyun
2810*4882a593Smuzhiyun static inline unsigned char
DAC960_BA_mbox_init(void __iomem * base,dma_addr_t mbox_addr)2811*4882a593Smuzhiyun DAC960_BA_mbox_init(void __iomem *base, dma_addr_t mbox_addr)
2812*4882a593Smuzhiyun {
2813*4882a593Smuzhiyun unsigned char status;
2814*4882a593Smuzhiyun
2815*4882a593Smuzhiyun while (DAC960_BA_hw_mbox_is_full(base))
2816*4882a593Smuzhiyun udelay(1);
2817*4882a593Smuzhiyun DAC960_BA_write_hw_mbox(base, mbox_addr);
2818*4882a593Smuzhiyun DAC960_BA_hw_mbox_new_cmd(base);
2819*4882a593Smuzhiyun while (!DAC960_BA_hw_mbox_status_available(base))
2820*4882a593Smuzhiyun udelay(1);
2821*4882a593Smuzhiyun status = DAC960_BA_read_cmd_status(base);
2822*4882a593Smuzhiyun DAC960_BA_ack_hw_mbox_intr(base);
2823*4882a593Smuzhiyun DAC960_BA_ack_hw_mbox_status(base);
2824*4882a593Smuzhiyun
2825*4882a593Smuzhiyun return status;
2826*4882a593Smuzhiyun }
2827*4882a593Smuzhiyun
DAC960_BA_hw_init(struct pci_dev * pdev,struct myrs_hba * cs,void __iomem * base)2828*4882a593Smuzhiyun static int DAC960_BA_hw_init(struct pci_dev *pdev,
2829*4882a593Smuzhiyun struct myrs_hba *cs, void __iomem *base)
2830*4882a593Smuzhiyun {
2831*4882a593Smuzhiyun int timeout = 0;
2832*4882a593Smuzhiyun unsigned char status, parm0, parm1;
2833*4882a593Smuzhiyun
2834*4882a593Smuzhiyun DAC960_BA_disable_intr(base);
2835*4882a593Smuzhiyun DAC960_BA_ack_hw_mbox_status(base);
2836*4882a593Smuzhiyun udelay(1000);
2837*4882a593Smuzhiyun while (DAC960_BA_init_in_progress(base) &&
2838*4882a593Smuzhiyun timeout < MYRS_MAILBOX_TIMEOUT) {
2839*4882a593Smuzhiyun if (DAC960_BA_read_error_status(base, &status,
2840*4882a593Smuzhiyun &parm0, &parm1) &&
2841*4882a593Smuzhiyun myrs_err_status(cs, status, parm0, parm1))
2842*4882a593Smuzhiyun return -EIO;
2843*4882a593Smuzhiyun udelay(10);
2844*4882a593Smuzhiyun timeout++;
2845*4882a593Smuzhiyun }
2846*4882a593Smuzhiyun if (timeout == MYRS_MAILBOX_TIMEOUT) {
2847*4882a593Smuzhiyun dev_err(&pdev->dev,
2848*4882a593Smuzhiyun "Timeout waiting for Controller Initialisation\n");
2849*4882a593Smuzhiyun return -ETIMEDOUT;
2850*4882a593Smuzhiyun }
2851*4882a593Smuzhiyun if (!myrs_enable_mmio_mbox(cs, DAC960_BA_mbox_init)) {
2852*4882a593Smuzhiyun dev_err(&pdev->dev,
2853*4882a593Smuzhiyun "Unable to Enable Memory Mailbox Interface\n");
2854*4882a593Smuzhiyun DAC960_BA_reset_ctrl(base);
2855*4882a593Smuzhiyun return -EAGAIN;
2856*4882a593Smuzhiyun }
2857*4882a593Smuzhiyun DAC960_BA_enable_intr(base);
2858*4882a593Smuzhiyun cs->write_cmd_mbox = DAC960_BA_write_cmd_mbox;
2859*4882a593Smuzhiyun cs->get_cmd_mbox = DAC960_BA_mem_mbox_new_cmd;
2860*4882a593Smuzhiyun cs->disable_intr = DAC960_BA_disable_intr;
2861*4882a593Smuzhiyun cs->reset = DAC960_BA_reset_ctrl;
2862*4882a593Smuzhiyun return 0;
2863*4882a593Smuzhiyun }
2864*4882a593Smuzhiyun
DAC960_BA_intr_handler(int irq,void * arg)2865*4882a593Smuzhiyun static irqreturn_t DAC960_BA_intr_handler(int irq, void *arg)
2866*4882a593Smuzhiyun {
2867*4882a593Smuzhiyun struct myrs_hba *cs = arg;
2868*4882a593Smuzhiyun void __iomem *base = cs->io_base;
2869*4882a593Smuzhiyun struct myrs_stat_mbox *next_stat_mbox;
2870*4882a593Smuzhiyun unsigned long flags;
2871*4882a593Smuzhiyun
2872*4882a593Smuzhiyun spin_lock_irqsave(&cs->queue_lock, flags);
2873*4882a593Smuzhiyun DAC960_BA_ack_intr(base);
2874*4882a593Smuzhiyun next_stat_mbox = cs->next_stat_mbox;
2875*4882a593Smuzhiyun while (next_stat_mbox->id > 0) {
2876*4882a593Smuzhiyun unsigned short id = next_stat_mbox->id;
2877*4882a593Smuzhiyun struct scsi_cmnd *scmd = NULL;
2878*4882a593Smuzhiyun struct myrs_cmdblk *cmd_blk = NULL;
2879*4882a593Smuzhiyun
2880*4882a593Smuzhiyun if (id == MYRS_DCMD_TAG)
2881*4882a593Smuzhiyun cmd_blk = &cs->dcmd_blk;
2882*4882a593Smuzhiyun else if (id == MYRS_MCMD_TAG)
2883*4882a593Smuzhiyun cmd_blk = &cs->mcmd_blk;
2884*4882a593Smuzhiyun else {
2885*4882a593Smuzhiyun scmd = scsi_host_find_tag(cs->host, id - 3);
2886*4882a593Smuzhiyun if (scmd)
2887*4882a593Smuzhiyun cmd_blk = scsi_cmd_priv(scmd);
2888*4882a593Smuzhiyun }
2889*4882a593Smuzhiyun if (cmd_blk) {
2890*4882a593Smuzhiyun cmd_blk->status = next_stat_mbox->status;
2891*4882a593Smuzhiyun cmd_blk->sense_len = next_stat_mbox->sense_len;
2892*4882a593Smuzhiyun cmd_blk->residual = next_stat_mbox->residual;
2893*4882a593Smuzhiyun } else
2894*4882a593Smuzhiyun dev_err(&cs->pdev->dev,
2895*4882a593Smuzhiyun "Unhandled command completion %d\n", id);
2896*4882a593Smuzhiyun
2897*4882a593Smuzhiyun memset(next_stat_mbox, 0, sizeof(struct myrs_stat_mbox));
2898*4882a593Smuzhiyun if (++next_stat_mbox > cs->last_stat_mbox)
2899*4882a593Smuzhiyun next_stat_mbox = cs->first_stat_mbox;
2900*4882a593Smuzhiyun
2901*4882a593Smuzhiyun if (cmd_blk) {
2902*4882a593Smuzhiyun if (id < 3)
2903*4882a593Smuzhiyun myrs_handle_cmdblk(cs, cmd_blk);
2904*4882a593Smuzhiyun else
2905*4882a593Smuzhiyun myrs_handle_scsi(cs, cmd_blk, scmd);
2906*4882a593Smuzhiyun }
2907*4882a593Smuzhiyun }
2908*4882a593Smuzhiyun cs->next_stat_mbox = next_stat_mbox;
2909*4882a593Smuzhiyun spin_unlock_irqrestore(&cs->queue_lock, flags);
2910*4882a593Smuzhiyun return IRQ_HANDLED;
2911*4882a593Smuzhiyun }
2912*4882a593Smuzhiyun
2913*4882a593Smuzhiyun struct myrs_privdata DAC960_BA_privdata = {
2914*4882a593Smuzhiyun .hw_init = DAC960_BA_hw_init,
2915*4882a593Smuzhiyun .irq_handler = DAC960_BA_intr_handler,
2916*4882a593Smuzhiyun .mmio_size = DAC960_BA_mmio_size,
2917*4882a593Smuzhiyun };
2918*4882a593Smuzhiyun
2919*4882a593Smuzhiyun /*
2920*4882a593Smuzhiyun * DAC960 LP Series Controllers.
2921*4882a593Smuzhiyun */
2922*4882a593Smuzhiyun
DAC960_LP_hw_mbox_new_cmd(void __iomem * base)2923*4882a593Smuzhiyun static inline void DAC960_LP_hw_mbox_new_cmd(void __iomem *base)
2924*4882a593Smuzhiyun {
2925*4882a593Smuzhiyun writeb(DAC960_LP_IDB_HWMBOX_NEW_CMD, base + DAC960_LP_IDB_OFFSET);
2926*4882a593Smuzhiyun }
2927*4882a593Smuzhiyun
DAC960_LP_ack_hw_mbox_status(void __iomem * base)2928*4882a593Smuzhiyun static inline void DAC960_LP_ack_hw_mbox_status(void __iomem *base)
2929*4882a593Smuzhiyun {
2930*4882a593Smuzhiyun writeb(DAC960_LP_IDB_HWMBOX_ACK_STS, base + DAC960_LP_IDB_OFFSET);
2931*4882a593Smuzhiyun }
2932*4882a593Smuzhiyun
DAC960_LP_gen_intr(void __iomem * base)2933*4882a593Smuzhiyun static inline void DAC960_LP_gen_intr(void __iomem *base)
2934*4882a593Smuzhiyun {
2935*4882a593Smuzhiyun writeb(DAC960_LP_IDB_GEN_IRQ, base + DAC960_LP_IDB_OFFSET);
2936*4882a593Smuzhiyun }
2937*4882a593Smuzhiyun
DAC960_LP_reset_ctrl(void __iomem * base)2938*4882a593Smuzhiyun static inline void DAC960_LP_reset_ctrl(void __iomem *base)
2939*4882a593Smuzhiyun {
2940*4882a593Smuzhiyun writeb(DAC960_LP_IDB_CTRL_RESET, base + DAC960_LP_IDB_OFFSET);
2941*4882a593Smuzhiyun }
2942*4882a593Smuzhiyun
DAC960_LP_mem_mbox_new_cmd(void __iomem * base)2943*4882a593Smuzhiyun static inline void DAC960_LP_mem_mbox_new_cmd(void __iomem *base)
2944*4882a593Smuzhiyun {
2945*4882a593Smuzhiyun writeb(DAC960_LP_IDB_MMBOX_NEW_CMD, base + DAC960_LP_IDB_OFFSET);
2946*4882a593Smuzhiyun }
2947*4882a593Smuzhiyun
DAC960_LP_hw_mbox_is_full(void __iomem * base)2948*4882a593Smuzhiyun static inline bool DAC960_LP_hw_mbox_is_full(void __iomem *base)
2949*4882a593Smuzhiyun {
2950*4882a593Smuzhiyun u8 val;
2951*4882a593Smuzhiyun
2952*4882a593Smuzhiyun val = readb(base + DAC960_LP_IDB_OFFSET);
2953*4882a593Smuzhiyun return val & DAC960_LP_IDB_HWMBOX_FULL;
2954*4882a593Smuzhiyun }
2955*4882a593Smuzhiyun
DAC960_LP_init_in_progress(void __iomem * base)2956*4882a593Smuzhiyun static inline bool DAC960_LP_init_in_progress(void __iomem *base)
2957*4882a593Smuzhiyun {
2958*4882a593Smuzhiyun u8 val;
2959*4882a593Smuzhiyun
2960*4882a593Smuzhiyun val = readb(base + DAC960_LP_IDB_OFFSET);
2961*4882a593Smuzhiyun return val & DAC960_LP_IDB_INIT_IN_PROGRESS;
2962*4882a593Smuzhiyun }
2963*4882a593Smuzhiyun
DAC960_LP_ack_hw_mbox_intr(void __iomem * base)2964*4882a593Smuzhiyun static inline void DAC960_LP_ack_hw_mbox_intr(void __iomem *base)
2965*4882a593Smuzhiyun {
2966*4882a593Smuzhiyun writeb(DAC960_LP_ODB_HWMBOX_ACK_IRQ, base + DAC960_LP_ODB_OFFSET);
2967*4882a593Smuzhiyun }
2968*4882a593Smuzhiyun
DAC960_LP_ack_mem_mbox_intr(void __iomem * base)2969*4882a593Smuzhiyun static inline void DAC960_LP_ack_mem_mbox_intr(void __iomem *base)
2970*4882a593Smuzhiyun {
2971*4882a593Smuzhiyun writeb(DAC960_LP_ODB_MMBOX_ACK_IRQ, base + DAC960_LP_ODB_OFFSET);
2972*4882a593Smuzhiyun }
2973*4882a593Smuzhiyun
DAC960_LP_ack_intr(void __iomem * base)2974*4882a593Smuzhiyun static inline void DAC960_LP_ack_intr(void __iomem *base)
2975*4882a593Smuzhiyun {
2976*4882a593Smuzhiyun writeb(DAC960_LP_ODB_HWMBOX_ACK_IRQ | DAC960_LP_ODB_MMBOX_ACK_IRQ,
2977*4882a593Smuzhiyun base + DAC960_LP_ODB_OFFSET);
2978*4882a593Smuzhiyun }
2979*4882a593Smuzhiyun
DAC960_LP_hw_mbox_status_available(void __iomem * base)2980*4882a593Smuzhiyun static inline bool DAC960_LP_hw_mbox_status_available(void __iomem *base)
2981*4882a593Smuzhiyun {
2982*4882a593Smuzhiyun u8 val;
2983*4882a593Smuzhiyun
2984*4882a593Smuzhiyun val = readb(base + DAC960_LP_ODB_OFFSET);
2985*4882a593Smuzhiyun return val & DAC960_LP_ODB_HWMBOX_STS_AVAIL;
2986*4882a593Smuzhiyun }
2987*4882a593Smuzhiyun
DAC960_LP_mem_mbox_status_available(void __iomem * base)2988*4882a593Smuzhiyun static inline bool DAC960_LP_mem_mbox_status_available(void __iomem *base)
2989*4882a593Smuzhiyun {
2990*4882a593Smuzhiyun u8 val;
2991*4882a593Smuzhiyun
2992*4882a593Smuzhiyun val = readb(base + DAC960_LP_ODB_OFFSET);
2993*4882a593Smuzhiyun return val & DAC960_LP_ODB_MMBOX_STS_AVAIL;
2994*4882a593Smuzhiyun }
2995*4882a593Smuzhiyun
DAC960_LP_enable_intr(void __iomem * base)2996*4882a593Smuzhiyun static inline void DAC960_LP_enable_intr(void __iomem *base)
2997*4882a593Smuzhiyun {
2998*4882a593Smuzhiyun writeb(~DAC960_LP_IRQMASK_DISABLE_IRQ, base + DAC960_LP_IRQMASK_OFFSET);
2999*4882a593Smuzhiyun }
3000*4882a593Smuzhiyun
DAC960_LP_disable_intr(void __iomem * base)3001*4882a593Smuzhiyun static inline void DAC960_LP_disable_intr(void __iomem *base)
3002*4882a593Smuzhiyun {
3003*4882a593Smuzhiyun writeb(0xFF, base + DAC960_LP_IRQMASK_OFFSET);
3004*4882a593Smuzhiyun }
3005*4882a593Smuzhiyun
DAC960_LP_intr_enabled(void __iomem * base)3006*4882a593Smuzhiyun static inline bool DAC960_LP_intr_enabled(void __iomem *base)
3007*4882a593Smuzhiyun {
3008*4882a593Smuzhiyun u8 val;
3009*4882a593Smuzhiyun
3010*4882a593Smuzhiyun val = readb(base + DAC960_LP_IRQMASK_OFFSET);
3011*4882a593Smuzhiyun return !(val & DAC960_LP_IRQMASK_DISABLE_IRQ);
3012*4882a593Smuzhiyun }
3013*4882a593Smuzhiyun
DAC960_LP_write_cmd_mbox(union myrs_cmd_mbox * mem_mbox,union myrs_cmd_mbox * mbox)3014*4882a593Smuzhiyun static inline void DAC960_LP_write_cmd_mbox(union myrs_cmd_mbox *mem_mbox,
3015*4882a593Smuzhiyun union myrs_cmd_mbox *mbox)
3016*4882a593Smuzhiyun {
3017*4882a593Smuzhiyun memcpy(&mem_mbox->words[1], &mbox->words[1],
3018*4882a593Smuzhiyun sizeof(union myrs_cmd_mbox) - sizeof(unsigned int));
3019*4882a593Smuzhiyun /* Barrier to avoid reordering */
3020*4882a593Smuzhiyun wmb();
3021*4882a593Smuzhiyun mem_mbox->words[0] = mbox->words[0];
3022*4882a593Smuzhiyun /* Barrier to force PCI access */
3023*4882a593Smuzhiyun mb();
3024*4882a593Smuzhiyun }
3025*4882a593Smuzhiyun
DAC960_LP_write_hw_mbox(void __iomem * base,dma_addr_t cmd_mbox_addr)3026*4882a593Smuzhiyun static inline void DAC960_LP_write_hw_mbox(void __iomem *base,
3027*4882a593Smuzhiyun dma_addr_t cmd_mbox_addr)
3028*4882a593Smuzhiyun {
3029*4882a593Smuzhiyun dma_addr_writeql(cmd_mbox_addr, base + DAC960_LP_CMDMBX_OFFSET);
3030*4882a593Smuzhiyun }
3031*4882a593Smuzhiyun
DAC960_LP_read_cmd_ident(void __iomem * base)3032*4882a593Smuzhiyun static inline unsigned short DAC960_LP_read_cmd_ident(void __iomem *base)
3033*4882a593Smuzhiyun {
3034*4882a593Smuzhiyun return readw(base + DAC960_LP_CMDSTS_OFFSET);
3035*4882a593Smuzhiyun }
3036*4882a593Smuzhiyun
DAC960_LP_read_cmd_status(void __iomem * base)3037*4882a593Smuzhiyun static inline unsigned char DAC960_LP_read_cmd_status(void __iomem *base)
3038*4882a593Smuzhiyun {
3039*4882a593Smuzhiyun return readw(base + DAC960_LP_CMDSTS_OFFSET + 2);
3040*4882a593Smuzhiyun }
3041*4882a593Smuzhiyun
3042*4882a593Smuzhiyun static inline bool
DAC960_LP_read_error_status(void __iomem * base,unsigned char * error,unsigned char * param0,unsigned char * param1)3043*4882a593Smuzhiyun DAC960_LP_read_error_status(void __iomem *base, unsigned char *error,
3044*4882a593Smuzhiyun unsigned char *param0, unsigned char *param1)
3045*4882a593Smuzhiyun {
3046*4882a593Smuzhiyun u8 val;
3047*4882a593Smuzhiyun
3048*4882a593Smuzhiyun val = readb(base + DAC960_LP_ERRSTS_OFFSET);
3049*4882a593Smuzhiyun if (!(val & DAC960_LP_ERRSTS_PENDING))
3050*4882a593Smuzhiyun return false;
3051*4882a593Smuzhiyun val &= ~DAC960_LP_ERRSTS_PENDING;
3052*4882a593Smuzhiyun *error = val;
3053*4882a593Smuzhiyun *param0 = readb(base + DAC960_LP_CMDMBX_OFFSET + 0);
3054*4882a593Smuzhiyun *param1 = readb(base + DAC960_LP_CMDMBX_OFFSET + 1);
3055*4882a593Smuzhiyun writeb(0xFF, base + DAC960_LP_ERRSTS_OFFSET);
3056*4882a593Smuzhiyun return true;
3057*4882a593Smuzhiyun }
3058*4882a593Smuzhiyun
3059*4882a593Smuzhiyun static inline unsigned char
DAC960_LP_mbox_init(void __iomem * base,dma_addr_t mbox_addr)3060*4882a593Smuzhiyun DAC960_LP_mbox_init(void __iomem *base, dma_addr_t mbox_addr)
3061*4882a593Smuzhiyun {
3062*4882a593Smuzhiyun unsigned char status;
3063*4882a593Smuzhiyun
3064*4882a593Smuzhiyun while (DAC960_LP_hw_mbox_is_full(base))
3065*4882a593Smuzhiyun udelay(1);
3066*4882a593Smuzhiyun DAC960_LP_write_hw_mbox(base, mbox_addr);
3067*4882a593Smuzhiyun DAC960_LP_hw_mbox_new_cmd(base);
3068*4882a593Smuzhiyun while (!DAC960_LP_hw_mbox_status_available(base))
3069*4882a593Smuzhiyun udelay(1);
3070*4882a593Smuzhiyun status = DAC960_LP_read_cmd_status(base);
3071*4882a593Smuzhiyun DAC960_LP_ack_hw_mbox_intr(base);
3072*4882a593Smuzhiyun DAC960_LP_ack_hw_mbox_status(base);
3073*4882a593Smuzhiyun
3074*4882a593Smuzhiyun return status;
3075*4882a593Smuzhiyun }
3076*4882a593Smuzhiyun
DAC960_LP_hw_init(struct pci_dev * pdev,struct myrs_hba * cs,void __iomem * base)3077*4882a593Smuzhiyun static int DAC960_LP_hw_init(struct pci_dev *pdev,
3078*4882a593Smuzhiyun struct myrs_hba *cs, void __iomem *base)
3079*4882a593Smuzhiyun {
3080*4882a593Smuzhiyun int timeout = 0;
3081*4882a593Smuzhiyun unsigned char status, parm0, parm1;
3082*4882a593Smuzhiyun
3083*4882a593Smuzhiyun DAC960_LP_disable_intr(base);
3084*4882a593Smuzhiyun DAC960_LP_ack_hw_mbox_status(base);
3085*4882a593Smuzhiyun udelay(1000);
3086*4882a593Smuzhiyun while (DAC960_LP_init_in_progress(base) &&
3087*4882a593Smuzhiyun timeout < MYRS_MAILBOX_TIMEOUT) {
3088*4882a593Smuzhiyun if (DAC960_LP_read_error_status(base, &status,
3089*4882a593Smuzhiyun &parm0, &parm1) &&
3090*4882a593Smuzhiyun myrs_err_status(cs, status, parm0, parm1))
3091*4882a593Smuzhiyun return -EIO;
3092*4882a593Smuzhiyun udelay(10);
3093*4882a593Smuzhiyun timeout++;
3094*4882a593Smuzhiyun }
3095*4882a593Smuzhiyun if (timeout == MYRS_MAILBOX_TIMEOUT) {
3096*4882a593Smuzhiyun dev_err(&pdev->dev,
3097*4882a593Smuzhiyun "Timeout waiting for Controller Initialisation\n");
3098*4882a593Smuzhiyun return -ETIMEDOUT;
3099*4882a593Smuzhiyun }
3100*4882a593Smuzhiyun if (!myrs_enable_mmio_mbox(cs, DAC960_LP_mbox_init)) {
3101*4882a593Smuzhiyun dev_err(&pdev->dev,
3102*4882a593Smuzhiyun "Unable to Enable Memory Mailbox Interface\n");
3103*4882a593Smuzhiyun DAC960_LP_reset_ctrl(base);
3104*4882a593Smuzhiyun return -ENODEV;
3105*4882a593Smuzhiyun }
3106*4882a593Smuzhiyun DAC960_LP_enable_intr(base);
3107*4882a593Smuzhiyun cs->write_cmd_mbox = DAC960_LP_write_cmd_mbox;
3108*4882a593Smuzhiyun cs->get_cmd_mbox = DAC960_LP_mem_mbox_new_cmd;
3109*4882a593Smuzhiyun cs->disable_intr = DAC960_LP_disable_intr;
3110*4882a593Smuzhiyun cs->reset = DAC960_LP_reset_ctrl;
3111*4882a593Smuzhiyun
3112*4882a593Smuzhiyun return 0;
3113*4882a593Smuzhiyun }
3114*4882a593Smuzhiyun
DAC960_LP_intr_handler(int irq,void * arg)3115*4882a593Smuzhiyun static irqreturn_t DAC960_LP_intr_handler(int irq, void *arg)
3116*4882a593Smuzhiyun {
3117*4882a593Smuzhiyun struct myrs_hba *cs = arg;
3118*4882a593Smuzhiyun void __iomem *base = cs->io_base;
3119*4882a593Smuzhiyun struct myrs_stat_mbox *next_stat_mbox;
3120*4882a593Smuzhiyun unsigned long flags;
3121*4882a593Smuzhiyun
3122*4882a593Smuzhiyun spin_lock_irqsave(&cs->queue_lock, flags);
3123*4882a593Smuzhiyun DAC960_LP_ack_intr(base);
3124*4882a593Smuzhiyun next_stat_mbox = cs->next_stat_mbox;
3125*4882a593Smuzhiyun while (next_stat_mbox->id > 0) {
3126*4882a593Smuzhiyun unsigned short id = next_stat_mbox->id;
3127*4882a593Smuzhiyun struct scsi_cmnd *scmd = NULL;
3128*4882a593Smuzhiyun struct myrs_cmdblk *cmd_blk = NULL;
3129*4882a593Smuzhiyun
3130*4882a593Smuzhiyun if (id == MYRS_DCMD_TAG)
3131*4882a593Smuzhiyun cmd_blk = &cs->dcmd_blk;
3132*4882a593Smuzhiyun else if (id == MYRS_MCMD_TAG)
3133*4882a593Smuzhiyun cmd_blk = &cs->mcmd_blk;
3134*4882a593Smuzhiyun else {
3135*4882a593Smuzhiyun scmd = scsi_host_find_tag(cs->host, id - 3);
3136*4882a593Smuzhiyun if (scmd)
3137*4882a593Smuzhiyun cmd_blk = scsi_cmd_priv(scmd);
3138*4882a593Smuzhiyun }
3139*4882a593Smuzhiyun if (cmd_blk) {
3140*4882a593Smuzhiyun cmd_blk->status = next_stat_mbox->status;
3141*4882a593Smuzhiyun cmd_blk->sense_len = next_stat_mbox->sense_len;
3142*4882a593Smuzhiyun cmd_blk->residual = next_stat_mbox->residual;
3143*4882a593Smuzhiyun } else
3144*4882a593Smuzhiyun dev_err(&cs->pdev->dev,
3145*4882a593Smuzhiyun "Unhandled command completion %d\n", id);
3146*4882a593Smuzhiyun
3147*4882a593Smuzhiyun memset(next_stat_mbox, 0, sizeof(struct myrs_stat_mbox));
3148*4882a593Smuzhiyun if (++next_stat_mbox > cs->last_stat_mbox)
3149*4882a593Smuzhiyun next_stat_mbox = cs->first_stat_mbox;
3150*4882a593Smuzhiyun
3151*4882a593Smuzhiyun if (cmd_blk) {
3152*4882a593Smuzhiyun if (id < 3)
3153*4882a593Smuzhiyun myrs_handle_cmdblk(cs, cmd_blk);
3154*4882a593Smuzhiyun else
3155*4882a593Smuzhiyun myrs_handle_scsi(cs, cmd_blk, scmd);
3156*4882a593Smuzhiyun }
3157*4882a593Smuzhiyun }
3158*4882a593Smuzhiyun cs->next_stat_mbox = next_stat_mbox;
3159*4882a593Smuzhiyun spin_unlock_irqrestore(&cs->queue_lock, flags);
3160*4882a593Smuzhiyun return IRQ_HANDLED;
3161*4882a593Smuzhiyun }
3162*4882a593Smuzhiyun
3163*4882a593Smuzhiyun struct myrs_privdata DAC960_LP_privdata = {
3164*4882a593Smuzhiyun .hw_init = DAC960_LP_hw_init,
3165*4882a593Smuzhiyun .irq_handler = DAC960_LP_intr_handler,
3166*4882a593Smuzhiyun .mmio_size = DAC960_LP_mmio_size,
3167*4882a593Smuzhiyun };
3168*4882a593Smuzhiyun
3169*4882a593Smuzhiyun /*
3170*4882a593Smuzhiyun * Module functions
3171*4882a593Smuzhiyun */
3172*4882a593Smuzhiyun static int
myrs_probe(struct pci_dev * dev,const struct pci_device_id * entry)3173*4882a593Smuzhiyun myrs_probe(struct pci_dev *dev, const struct pci_device_id *entry)
3174*4882a593Smuzhiyun {
3175*4882a593Smuzhiyun struct myrs_hba *cs;
3176*4882a593Smuzhiyun int ret;
3177*4882a593Smuzhiyun
3178*4882a593Smuzhiyun cs = myrs_detect(dev, entry);
3179*4882a593Smuzhiyun if (!cs)
3180*4882a593Smuzhiyun return -ENODEV;
3181*4882a593Smuzhiyun
3182*4882a593Smuzhiyun ret = myrs_get_config(cs);
3183*4882a593Smuzhiyun if (ret < 0) {
3184*4882a593Smuzhiyun myrs_cleanup(cs);
3185*4882a593Smuzhiyun return ret;
3186*4882a593Smuzhiyun }
3187*4882a593Smuzhiyun
3188*4882a593Smuzhiyun if (!myrs_create_mempools(dev, cs)) {
3189*4882a593Smuzhiyun ret = -ENOMEM;
3190*4882a593Smuzhiyun goto failed;
3191*4882a593Smuzhiyun }
3192*4882a593Smuzhiyun
3193*4882a593Smuzhiyun ret = scsi_add_host(cs->host, &dev->dev);
3194*4882a593Smuzhiyun if (ret) {
3195*4882a593Smuzhiyun dev_err(&dev->dev, "scsi_add_host failed with %d\n", ret);
3196*4882a593Smuzhiyun myrs_destroy_mempools(cs);
3197*4882a593Smuzhiyun goto failed;
3198*4882a593Smuzhiyun }
3199*4882a593Smuzhiyun scsi_scan_host(cs->host);
3200*4882a593Smuzhiyun return 0;
3201*4882a593Smuzhiyun failed:
3202*4882a593Smuzhiyun myrs_cleanup(cs);
3203*4882a593Smuzhiyun return ret;
3204*4882a593Smuzhiyun }
3205*4882a593Smuzhiyun
3206*4882a593Smuzhiyun
myrs_remove(struct pci_dev * pdev)3207*4882a593Smuzhiyun static void myrs_remove(struct pci_dev *pdev)
3208*4882a593Smuzhiyun {
3209*4882a593Smuzhiyun struct myrs_hba *cs = pci_get_drvdata(pdev);
3210*4882a593Smuzhiyun
3211*4882a593Smuzhiyun if (cs == NULL)
3212*4882a593Smuzhiyun return;
3213*4882a593Smuzhiyun
3214*4882a593Smuzhiyun shost_printk(KERN_NOTICE, cs->host, "Flushing Cache...");
3215*4882a593Smuzhiyun myrs_flush_cache(cs);
3216*4882a593Smuzhiyun myrs_destroy_mempools(cs);
3217*4882a593Smuzhiyun myrs_cleanup(cs);
3218*4882a593Smuzhiyun }
3219*4882a593Smuzhiyun
3220*4882a593Smuzhiyun
3221*4882a593Smuzhiyun static const struct pci_device_id myrs_id_table[] = {
3222*4882a593Smuzhiyun {
3223*4882a593Smuzhiyun PCI_DEVICE_SUB(PCI_VENDOR_ID_MYLEX,
3224*4882a593Smuzhiyun PCI_DEVICE_ID_MYLEX_DAC960_GEM,
3225*4882a593Smuzhiyun PCI_VENDOR_ID_MYLEX, PCI_ANY_ID),
3226*4882a593Smuzhiyun .driver_data = (unsigned long) &DAC960_GEM_privdata,
3227*4882a593Smuzhiyun },
3228*4882a593Smuzhiyun {
3229*4882a593Smuzhiyun PCI_DEVICE_DATA(MYLEX, DAC960_BA, &DAC960_BA_privdata),
3230*4882a593Smuzhiyun },
3231*4882a593Smuzhiyun {
3232*4882a593Smuzhiyun PCI_DEVICE_DATA(MYLEX, DAC960_LP, &DAC960_LP_privdata),
3233*4882a593Smuzhiyun },
3234*4882a593Smuzhiyun {0, },
3235*4882a593Smuzhiyun };
3236*4882a593Smuzhiyun
3237*4882a593Smuzhiyun MODULE_DEVICE_TABLE(pci, myrs_id_table);
3238*4882a593Smuzhiyun
3239*4882a593Smuzhiyun static struct pci_driver myrs_pci_driver = {
3240*4882a593Smuzhiyun .name = "myrs",
3241*4882a593Smuzhiyun .id_table = myrs_id_table,
3242*4882a593Smuzhiyun .probe = myrs_probe,
3243*4882a593Smuzhiyun .remove = myrs_remove,
3244*4882a593Smuzhiyun };
3245*4882a593Smuzhiyun
myrs_init_module(void)3246*4882a593Smuzhiyun static int __init myrs_init_module(void)
3247*4882a593Smuzhiyun {
3248*4882a593Smuzhiyun int ret;
3249*4882a593Smuzhiyun
3250*4882a593Smuzhiyun myrs_raid_template = raid_class_attach(&myrs_raid_functions);
3251*4882a593Smuzhiyun if (!myrs_raid_template)
3252*4882a593Smuzhiyun return -ENODEV;
3253*4882a593Smuzhiyun
3254*4882a593Smuzhiyun ret = pci_register_driver(&myrs_pci_driver);
3255*4882a593Smuzhiyun if (ret)
3256*4882a593Smuzhiyun raid_class_release(myrs_raid_template);
3257*4882a593Smuzhiyun
3258*4882a593Smuzhiyun return ret;
3259*4882a593Smuzhiyun }
3260*4882a593Smuzhiyun
myrs_cleanup_module(void)3261*4882a593Smuzhiyun static void __exit myrs_cleanup_module(void)
3262*4882a593Smuzhiyun {
3263*4882a593Smuzhiyun pci_unregister_driver(&myrs_pci_driver);
3264*4882a593Smuzhiyun raid_class_release(myrs_raid_template);
3265*4882a593Smuzhiyun }
3266*4882a593Smuzhiyun
3267*4882a593Smuzhiyun module_init(myrs_init_module);
3268*4882a593Smuzhiyun module_exit(myrs_cleanup_module);
3269*4882a593Smuzhiyun
3270*4882a593Smuzhiyun MODULE_DESCRIPTION("Mylex DAC960/AcceleRAID/eXtremeRAID driver (SCSI Interface)");
3271*4882a593Smuzhiyun MODULE_AUTHOR("Hannes Reinecke <hare@suse.com>");
3272*4882a593Smuzhiyun MODULE_LICENSE("GPL");
3273