Pull sgi-drivers-makefile-cleanup into release branch
[GitHub/mt8127/android_kernel_alcatel_ttab.git] / drivers / scsi / aic7xxx / aic79xx_osm.c
1 /*
2 * Adaptec AIC79xx device driver for Linux.
3 *
4 * $Id: //depot/aic7xxx/linux/drivers/scsi/aic7xxx/aic79xx_osm.c#171 $
5 *
6 * --------------------------------------------------------------------------
7 * Copyright (c) 1994-2000 Justin T. Gibbs.
8 * Copyright (c) 1997-1999 Doug Ledford
9 * Copyright (c) 2000-2003 Adaptec Inc.
10 * All rights reserved.
11 *
12 * Redistribution and use in source and binary forms, with or without
13 * modification, are permitted provided that the following conditions
14 * are met:
15 * 1. Redistributions of source code must retain the above copyright
16 * notice, this list of conditions, and the following disclaimer,
17 * without modification.
18 * 2. Redistributions in binary form must reproduce at minimum a disclaimer
19 * substantially similar to the "NO WARRANTY" disclaimer below
20 * ("Disclaimer") and any redistribution must be conditioned upon
21 * including a substantially similar Disclaimer requirement for further
22 * binary redistribution.
23 * 3. Neither the names of the above-listed copyright holders nor the names
24 * of any contributors may be used to endorse or promote products derived
25 * from this software without specific prior written permission.
26 *
27 * Alternatively, this software may be distributed under the terms of the
28 * GNU General Public License ("GPL") version 2 as published by the Free
29 * Software Foundation.
30 *
31 * NO WARRANTY
32 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
33 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
34 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTIBILITY AND FITNESS FOR
35 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
36 * HOLDERS OR CONTRIBUTORS BE LIABLE FOR SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
37 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
38 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
39 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
40 * STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING
41 * IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
42 * POSSIBILITY OF SUCH DAMAGES.
43 */
44
45 #include "aic79xx_osm.h"
46 #include "aic79xx_inline.h"
47 #include <scsi/scsicam.h>
48
49 static struct scsi_transport_template *ahd_linux_transport_template = NULL;
50
51 #include <linux/init.h> /* __setup */
52 #include <linux/mm.h> /* For fetching system memory size */
53 #include <linux/blkdev.h> /* For block_size() */
54 #include <linux/delay.h> /* For ssleep/msleep */
55 #include <linux/device.h>
56
57 /*
58 * Bucket size for counting good commands in between bad ones.
59 */
60 #define AHD_LINUX_ERR_THRESH 1000
61
62 /*
63 * Set this to the delay in seconds after SCSI bus reset.
64 * Note, we honor this only for the initial bus reset.
65 * The scsi error recovery code performs its own bus settle
66 * delay handling for error recovery actions.
67 */
68 #ifdef CONFIG_AIC79XX_RESET_DELAY_MS
69 #define AIC79XX_RESET_DELAY CONFIG_AIC79XX_RESET_DELAY_MS
70 #else
71 #define AIC79XX_RESET_DELAY 5000
72 #endif
73
74 /*
75 * To change the default number of tagged transactions allowed per-device,
76 * add a line to the lilo.conf file like:
77 * append="aic79xx=verbose,tag_info:{{32,32,32,32},{32,32,32,32}}"
78 * which will result in the first four devices on the first two
79 * controllers being set to a tagged queue depth of 32.
80 *
81 * The tag_commands is an array of 16 to allow for wide and twin adapters.
82 * Twin adapters will use indexes 0-7 for channel 0, and indexes 8-15
83 * for channel 1.
84 */
85 typedef struct {
86 uint16_t tag_commands[16]; /* Allow for wide/twin adapters. */
87 } adapter_tag_info_t;
88
89 /*
90 * Modify this as you see fit for your system.
91 *
92 * 0 tagged queuing disabled
93 * 1 <= n <= 253 n == max tags ever dispatched.
94 *
95 * The driver will throttle the number of commands dispatched to a
96 * device if it returns queue full. For devices with a fixed maximum
97 * queue depth, the driver will eventually determine this depth and
98 * lock it in (a console message is printed to indicate that a lock
99 * has occurred). On some devices, queue full is returned for a temporary
100 * resource shortage. These devices will return queue full at varying
101 * depths. The driver will throttle back when the queue fulls occur and
102 * attempt to slowly increase the depth over time as the device recovers
103 * from the resource shortage.
104 *
105 * In this example, the first line will disable tagged queueing for all
106 * the devices on the first probed aic79xx adapter.
107 *
108 * The second line enables tagged queueing with 4 commands/LUN for IDs
109 * (0, 2-11, 13-15), disables tagged queueing for ID 12, and tells the
110 * driver to attempt to use up to 64 tags for ID 1.
111 *
112 * The third line is the same as the first line.
113 *
114 * The fourth line disables tagged queueing for devices 0 and 3. It
115 * enables tagged queueing for the other IDs, with 16 commands/LUN
116 * for IDs 1 and 4, 127 commands/LUN for ID 8, and 4 commands/LUN for
117 * IDs 2, 5-7, and 9-15.
118 */
119
120 /*
121 * NOTE: The below structure is for reference only, the actual structure
122 * to modify in order to change things is just below this comment block.
123 adapter_tag_info_t aic79xx_tag_info[] =
124 {
125 {{0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0}},
126 {{4, 64, 4, 4, 4, 4, 4, 4, 4, 4, 4, 4, 0, 4, 4, 4}},
127 {{0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0}},
128 {{0, 16, 4, 0, 16, 4, 4, 4, 127, 4, 4, 4, 4, 4, 4, 4}}
129 };
130 */
131
132 #ifdef CONFIG_AIC79XX_CMDS_PER_DEVICE
133 #define AIC79XX_CMDS_PER_DEVICE CONFIG_AIC79XX_CMDS_PER_DEVICE
134 #else
135 #define AIC79XX_CMDS_PER_DEVICE AHD_MAX_QUEUE
136 #endif
137
138 #define AIC79XX_CONFIGED_TAG_COMMANDS { \
139 AIC79XX_CMDS_PER_DEVICE, AIC79XX_CMDS_PER_DEVICE, \
140 AIC79XX_CMDS_PER_DEVICE, AIC79XX_CMDS_PER_DEVICE, \
141 AIC79XX_CMDS_PER_DEVICE, AIC79XX_CMDS_PER_DEVICE, \
142 AIC79XX_CMDS_PER_DEVICE, AIC79XX_CMDS_PER_DEVICE, \
143 AIC79XX_CMDS_PER_DEVICE, AIC79XX_CMDS_PER_DEVICE, \
144 AIC79XX_CMDS_PER_DEVICE, AIC79XX_CMDS_PER_DEVICE, \
145 AIC79XX_CMDS_PER_DEVICE, AIC79XX_CMDS_PER_DEVICE, \
146 AIC79XX_CMDS_PER_DEVICE, AIC79XX_CMDS_PER_DEVICE \
147 }
148
149 /*
150 * By default, use the number of commands specified by
151 * the users kernel configuration.
152 */
153 static adapter_tag_info_t aic79xx_tag_info[] =
154 {
155 {AIC79XX_CONFIGED_TAG_COMMANDS},
156 {AIC79XX_CONFIGED_TAG_COMMANDS},
157 {AIC79XX_CONFIGED_TAG_COMMANDS},
158 {AIC79XX_CONFIGED_TAG_COMMANDS},
159 {AIC79XX_CONFIGED_TAG_COMMANDS},
160 {AIC79XX_CONFIGED_TAG_COMMANDS},
161 {AIC79XX_CONFIGED_TAG_COMMANDS},
162 {AIC79XX_CONFIGED_TAG_COMMANDS},
163 {AIC79XX_CONFIGED_TAG_COMMANDS},
164 {AIC79XX_CONFIGED_TAG_COMMANDS},
165 {AIC79XX_CONFIGED_TAG_COMMANDS},
166 {AIC79XX_CONFIGED_TAG_COMMANDS},
167 {AIC79XX_CONFIGED_TAG_COMMANDS},
168 {AIC79XX_CONFIGED_TAG_COMMANDS},
169 {AIC79XX_CONFIGED_TAG_COMMANDS},
170 {AIC79XX_CONFIGED_TAG_COMMANDS}
171 };
172
173 /*
174 * The I/O cell on the chip is very configurable in respect to its analog
175 * characteristics. Set the defaults here; they can be overriden with
176 * the proper insmod parameters.
177 */
178 struct ahd_linux_iocell_opts
179 {
180 uint8_t precomp;
181 uint8_t slewrate;
182 uint8_t amplitude;
183 };
184 #define AIC79XX_DEFAULT_PRECOMP 0xFF
185 #define AIC79XX_DEFAULT_SLEWRATE 0xFF
186 #define AIC79XX_DEFAULT_AMPLITUDE 0xFF
187 #define AIC79XX_DEFAULT_IOOPTS \
188 { \
189 AIC79XX_DEFAULT_PRECOMP, \
190 AIC79XX_DEFAULT_SLEWRATE, \
191 AIC79XX_DEFAULT_AMPLITUDE \
192 }
193 #define AIC79XX_PRECOMP_INDEX 0
194 #define AIC79XX_SLEWRATE_INDEX 1
195 #define AIC79XX_AMPLITUDE_INDEX 2
196 static struct ahd_linux_iocell_opts aic79xx_iocell_info[] =
197 {
198 AIC79XX_DEFAULT_IOOPTS,
199 AIC79XX_DEFAULT_IOOPTS,
200 AIC79XX_DEFAULT_IOOPTS,
201 AIC79XX_DEFAULT_IOOPTS,
202 AIC79XX_DEFAULT_IOOPTS,
203 AIC79XX_DEFAULT_IOOPTS,
204 AIC79XX_DEFAULT_IOOPTS,
205 AIC79XX_DEFAULT_IOOPTS,
206 AIC79XX_DEFAULT_IOOPTS,
207 AIC79XX_DEFAULT_IOOPTS,
208 AIC79XX_DEFAULT_IOOPTS,
209 AIC79XX_DEFAULT_IOOPTS,
210 AIC79XX_DEFAULT_IOOPTS,
211 AIC79XX_DEFAULT_IOOPTS,
212 AIC79XX_DEFAULT_IOOPTS,
213 AIC79XX_DEFAULT_IOOPTS
214 };
215
216 /*
217 * There should be a specific return value for this in scsi.h, but
218 * it seems that most drivers ignore it.
219 */
220 #define DID_UNDERFLOW DID_ERROR
221
222 void
223 ahd_print_path(struct ahd_softc *ahd, struct scb *scb)
224 {
225 printk("(scsi%d:%c:%d:%d): ",
226 ahd->platform_data->host->host_no,
227 scb != NULL ? SCB_GET_CHANNEL(ahd, scb) : 'X',
228 scb != NULL ? SCB_GET_TARGET(ahd, scb) : -1,
229 scb != NULL ? SCB_GET_LUN(scb) : -1);
230 }
231
232 /*
233 * XXX - these options apply unilaterally to _all_ adapters
234 * cards in the system. This should be fixed. Exceptions to this
235 * rule are noted in the comments.
236 */
237
238 /*
239 * Skip the scsi bus reset. Non 0 make us skip the reset at startup. This
240 * has no effect on any later resets that might occur due to things like
241 * SCSI bus timeouts.
242 */
243 static uint32_t aic79xx_no_reset;
244
245 /*
246 * Certain PCI motherboards will scan PCI devices from highest to lowest,
247 * others scan from lowest to highest, and they tend to do all kinds of
248 * strange things when they come into contact with PCI bridge chips. The
249 * net result of all this is that the PCI card that is actually used to boot
250 * the machine is very hard to detect. Most motherboards go from lowest
251 * PCI slot number to highest, and the first SCSI controller found is the
252 * one you boot from. The only exceptions to this are when a controller
253 * has its BIOS disabled. So, we by default sort all of our SCSI controllers
254 * from lowest PCI slot number to highest PCI slot number. We also force
255 * all controllers with their BIOS disabled to the end of the list. This
256 * works on *almost* all computers. Where it doesn't work, we have this
257 * option. Setting this option to non-0 will reverse the order of the sort
258 * to highest first, then lowest, but will still leave cards with their BIOS
259 * disabled at the very end. That should fix everyone up unless there are
260 * really strange cirumstances.
261 */
262 static uint32_t aic79xx_reverse_scan;
263
264 /*
265 * Should we force EXTENDED translation on a controller.
266 * 0 == Use whatever is in the SEEPROM or default to off
267 * 1 == Use whatever is in the SEEPROM or default to on
268 */
269 static uint32_t aic79xx_extended;
270
271 /*
272 * PCI bus parity checking of the Adaptec controllers. This is somewhat
273 * dubious at best. To my knowledge, this option has never actually
274 * solved a PCI parity problem, but on certain machines with broken PCI
275 * chipset configurations, it can generate tons of false error messages.
276 * It's included in the driver for completeness.
277 * 0 = Shut off PCI parity check
278 * non-0 = Enable PCI parity check
279 *
280 * NOTE: you can't actually pass -1 on the lilo prompt. So, to set this
281 * variable to -1 you would actually want to simply pass the variable
282 * name without a number. That will invert the 0 which will result in
283 * -1.
284 */
285 static uint32_t aic79xx_pci_parity = ~0;
286
287 /*
288 * There are lots of broken chipsets in the world. Some of them will
289 * violate the PCI spec when we issue byte sized memory writes to our
290 * controller. I/O mapped register access, if allowed by the given
291 * platform, will work in almost all cases.
292 */
293 uint32_t aic79xx_allow_memio = ~0;
294
295 /*
296 * So that we can set how long each device is given as a selection timeout.
297 * The table of values goes like this:
298 * 0 - 256ms
299 * 1 - 128ms
300 * 2 - 64ms
301 * 3 - 32ms
302 * We default to 256ms because some older devices need a longer time
303 * to respond to initial selection.
304 */
305 static uint32_t aic79xx_seltime;
306
307 /*
308 * Certain devices do not perform any aging on commands. Should the
309 * device be saturated by commands in one portion of the disk, it is
310 * possible for transactions on far away sectors to never be serviced.
311 * To handle these devices, we can periodically send an ordered tag to
312 * force all outstanding transactions to be serviced prior to a new
313 * transaction.
314 */
315 uint32_t aic79xx_periodic_otag;
316
317 /* Some storage boxes are using an LSI chip which has a bug making it
318 * impossible to use aic79xx Rev B chip in 320 speeds. The following
319 * storage boxes have been reported to be buggy:
320 * EonStor 3U 16-Bay: U16U-G3A3
321 * EonStor 2U 12-Bay: U12U-G3A3
322 * SentinelRAID: 2500F R5 / R6
323 * SentinelRAID: 2500F R1
324 * SentinelRAID: 2500F/1500F
325 * SentinelRAID: 150F
326 *
327 * To get around this LSI bug, you can set your board to 160 mode
328 * or you can enable the SLOWCRC bit.
329 */
330 uint32_t aic79xx_slowcrc;
331
332 /*
333 * Module information and settable options.
334 */
335 static char *aic79xx = NULL;
336
337 MODULE_AUTHOR("Maintainer: Justin T. Gibbs <gibbs@scsiguy.com>");
338 MODULE_DESCRIPTION("Adaptec Aic790X U320 SCSI Host Bus Adapter driver");
339 MODULE_LICENSE("Dual BSD/GPL");
340 MODULE_VERSION(AIC79XX_DRIVER_VERSION);
341 module_param(aic79xx, charp, 0444);
342 MODULE_PARM_DESC(aic79xx,
343 "period delimited, options string.\n"
344 " verbose Enable verbose/diagnostic logging\n"
345 " allow_memio Allow device registers to be memory mapped\n"
346 " debug Bitmask of debug values to enable\n"
347 " no_reset Supress initial bus resets\n"
348 " extended Enable extended geometry on all controllers\n"
349 " periodic_otag Send an ordered tagged transaction\n"
350 " periodically to prevent tag starvation.\n"
351 " This may be required by some older disk\n"
352 " or drives/RAID arrays.\n"
353 " reverse_scan Sort PCI devices highest Bus/Slot to lowest\n"
354 " tag_info:<tag_str> Set per-target tag depth\n"
355 " global_tag_depth:<int> Global tag depth for all targets on all buses\n"
356 " slewrate:<slewrate_list>Set the signal slew rate (0-15).\n"
357 " precomp:<pcomp_list> Set the signal precompensation (0-7).\n"
358 " amplitude:<int> Set the signal amplitude (0-7).\n"
359 " seltime:<int> Selection Timeout:\n"
360 " (0/256ms,1/128ms,2/64ms,3/32ms)\n"
361 " slowcrc Turn on the SLOWCRC bit (Rev B only)\n"
362 "\n"
363 " Sample /etc/modprobe.conf line:\n"
364 " Enable verbose logging\n"
365 " Set tag depth on Controller 2/Target 2 to 10 tags\n"
366 " Shorten the selection timeout to 128ms\n"
367 "\n"
368 " options aic79xx 'aic79xx=verbose.tag_info:{{}.{}.{..10}}.seltime:1'\n"
369 "\n");
370
371 static void ahd_linux_handle_scsi_status(struct ahd_softc *,
372 struct scsi_device *,
373 struct scb *);
374 static void ahd_linux_queue_cmd_complete(struct ahd_softc *ahd,
375 struct scsi_cmnd *cmd);
376 static void ahd_linux_sem_timeout(u_long arg);
377 static int ahd_linux_queue_recovery_cmd(struct scsi_cmnd *cmd, scb_flag flag);
378 static void ahd_linux_initialize_scsi_bus(struct ahd_softc *ahd);
379 static u_int ahd_linux_user_tagdepth(struct ahd_softc *ahd,
380 struct ahd_devinfo *devinfo);
381 static void ahd_linux_device_queue_depth(struct scsi_device *);
382 static int ahd_linux_run_command(struct ahd_softc*,
383 struct ahd_linux_device *,
384 struct scsi_cmnd *);
385 static void ahd_linux_setup_tag_info_global(char *p);
386 static int aic79xx_setup(char *c);
387
388 static int ahd_linux_unit;
389
390
391 /****************************** Inlines ***************************************/
392 static __inline void ahd_linux_unmap_scb(struct ahd_softc*, struct scb*);
393
394 static __inline void
395 ahd_linux_unmap_scb(struct ahd_softc *ahd, struct scb *scb)
396 {
397 struct scsi_cmnd *cmd;
398 int direction;
399
400 cmd = scb->io_ctx;
401 direction = cmd->sc_data_direction;
402 ahd_sync_sglist(ahd, scb, BUS_DMASYNC_POSTWRITE);
403 if (cmd->use_sg != 0) {
404 struct scatterlist *sg;
405
406 sg = (struct scatterlist *)cmd->request_buffer;
407 pci_unmap_sg(ahd->dev_softc, sg, cmd->use_sg, direction);
408 } else if (cmd->request_bufflen != 0) {
409 pci_unmap_single(ahd->dev_softc,
410 scb->platform_data->buf_busaddr,
411 cmd->request_bufflen, direction);
412 }
413 }
414
415 /******************************** Macros **************************************/
416 #define BUILD_SCSIID(ahd, cmd) \
417 (((scmd_id(cmd) << TID_SHIFT) & TID) | (ahd)->our_id)
418
419 /*
420 * Return a string describing the driver.
421 */
422 static const char *
423 ahd_linux_info(struct Scsi_Host *host)
424 {
425 static char buffer[512];
426 char ahd_info[256];
427 char *bp;
428 struct ahd_softc *ahd;
429
430 bp = &buffer[0];
431 ahd = *(struct ahd_softc **)host->hostdata;
432 memset(bp, 0, sizeof(buffer));
433 strcpy(bp, "Adaptec AIC79XX PCI-X SCSI HBA DRIVER, Rev ");
434 strcat(bp, AIC79XX_DRIVER_VERSION);
435 strcat(bp, "\n");
436 strcat(bp, " <");
437 strcat(bp, ahd->description);
438 strcat(bp, ">\n");
439 strcat(bp, " ");
440 ahd_controller_info(ahd, ahd_info);
441 strcat(bp, ahd_info);
442 strcat(bp, "\n");
443
444 return (bp);
445 }
446
447 /*
448 * Queue an SCB to the controller.
449 */
450 static int
451 ahd_linux_queue(struct scsi_cmnd * cmd, void (*scsi_done) (struct scsi_cmnd *))
452 {
453 struct ahd_softc *ahd;
454 struct ahd_linux_device *dev = scsi_transport_device_data(cmd->device);
455 int rtn = SCSI_MLQUEUE_HOST_BUSY;
456 unsigned long flags;
457
458 ahd = *(struct ahd_softc **)cmd->device->host->hostdata;
459
460 ahd_lock(ahd, &flags);
461 if (ahd->platform_data->qfrozen == 0) {
462 cmd->scsi_done = scsi_done;
463 cmd->result = CAM_REQ_INPROG << 16;
464 rtn = ahd_linux_run_command(ahd, dev, cmd);
465
466 }
467 ahd_unlock(ahd, &flags);
468 return rtn;
469 }
470
471 static inline struct scsi_target **
472 ahd_linux_target_in_softc(struct scsi_target *starget)
473 {
474 struct ahd_softc *ahd =
475 *((struct ahd_softc **)dev_to_shost(&starget->dev)->hostdata);
476 unsigned int target_offset;
477
478 target_offset = starget->id;
479 if (starget->channel != 0)
480 target_offset += 8;
481
482 return &ahd->platform_data->starget[target_offset];
483 }
484
485 static int
486 ahd_linux_target_alloc(struct scsi_target *starget)
487 {
488 struct ahd_softc *ahd =
489 *((struct ahd_softc **)dev_to_shost(&starget->dev)->hostdata);
490 unsigned long flags;
491 struct scsi_target **ahd_targp = ahd_linux_target_in_softc(starget);
492 struct ahd_linux_target *targ = scsi_transport_target_data(starget);
493 struct ahd_devinfo devinfo;
494 struct ahd_initiator_tinfo *tinfo;
495 struct ahd_tmode_tstate *tstate;
496 char channel = starget->channel + 'A';
497
498 ahd_lock(ahd, &flags);
499
500 BUG_ON(*ahd_targp != NULL);
501
502 *ahd_targp = starget;
503 memset(targ, 0, sizeof(*targ));
504
505 tinfo = ahd_fetch_transinfo(ahd, channel, ahd->our_id,
506 starget->id, &tstate);
507 ahd_compile_devinfo(&devinfo, ahd->our_id, starget->id,
508 CAM_LUN_WILDCARD, channel,
509 ROLE_INITIATOR);
510 spi_min_period(starget) = AHD_SYNCRATE_MAX; /* We can do U320 */
511 if ((ahd->bugs & AHD_PACED_NEGTABLE_BUG) != 0)
512 spi_max_offset(starget) = MAX_OFFSET_PACED_BUG;
513 else
514 spi_max_offset(starget) = MAX_OFFSET_PACED;
515 spi_max_width(starget) = ahd->features & AHD_WIDE;
516
517 ahd_set_syncrate(ahd, &devinfo, 0, 0, 0,
518 AHD_TRANS_GOAL, /*paused*/FALSE);
519 ahd_set_width(ahd, &devinfo, MSG_EXT_WDTR_BUS_8_BIT,
520 AHD_TRANS_GOAL, /*paused*/FALSE);
521 ahd_unlock(ahd, &flags);
522
523 return 0;
524 }
525
526 static void
527 ahd_linux_target_destroy(struct scsi_target *starget)
528 {
529 struct scsi_target **ahd_targp = ahd_linux_target_in_softc(starget);
530
531 *ahd_targp = NULL;
532 }
533
534 static int
535 ahd_linux_slave_alloc(struct scsi_device *sdev)
536 {
537 struct ahd_softc *ahd =
538 *((struct ahd_softc **)sdev->host->hostdata);
539 struct scsi_target *starget = sdev->sdev_target;
540 struct ahd_linux_target *targ = scsi_transport_target_data(starget);
541 struct ahd_linux_device *dev;
542
543 if (bootverbose)
544 printf("%s: Slave Alloc %d\n", ahd_name(ahd), sdev->id);
545
546 BUG_ON(targ->sdev[sdev->lun] != NULL);
547
548 dev = scsi_transport_device_data(sdev);
549 memset(dev, 0, sizeof(*dev));
550
551 /*
552 * We start out life using untagged
553 * transactions of which we allow one.
554 */
555 dev->openings = 1;
556
557 /*
558 * Set maxtags to 0. This will be changed if we
559 * later determine that we are dealing with
560 * a tagged queuing capable device.
561 */
562 dev->maxtags = 0;
563
564 targ->sdev[sdev->lun] = sdev;
565
566 return (0);
567 }
568
569 static int
570 ahd_linux_slave_configure(struct scsi_device *sdev)
571 {
572 struct ahd_softc *ahd;
573
574 ahd = *((struct ahd_softc **)sdev->host->hostdata);
575 if (bootverbose)
576 sdev_printk(KERN_INFO, sdev, "Slave Configure\n");
577
578 ahd_linux_device_queue_depth(sdev);
579
580 /* Initial Domain Validation */
581 if (!spi_initial_dv(sdev->sdev_target))
582 spi_dv_device(sdev);
583
584 return 0;
585 }
586
587 static void
588 ahd_linux_slave_destroy(struct scsi_device *sdev)
589 {
590 struct ahd_softc *ahd;
591 struct ahd_linux_device *dev = scsi_transport_device_data(sdev);
592 struct ahd_linux_target *targ = scsi_transport_target_data(sdev->sdev_target);
593
594 ahd = *((struct ahd_softc **)sdev->host->hostdata);
595 if (bootverbose)
596 printf("%s: Slave Destroy %d\n", ahd_name(ahd), sdev->id);
597
598 BUG_ON(dev->active);
599
600 targ->sdev[sdev->lun] = NULL;
601
602 }
603
604 #if defined(__i386__)
605 /*
606 * Return the disk geometry for the given SCSI device.
607 */
608 static int
609 ahd_linux_biosparam(struct scsi_device *sdev, struct block_device *bdev,
610 sector_t capacity, int geom[])
611 {
612 uint8_t *bh;
613 int heads;
614 int sectors;
615 int cylinders;
616 int ret;
617 int extended;
618 struct ahd_softc *ahd;
619
620 ahd = *((struct ahd_softc **)sdev->host->hostdata);
621
622 bh = scsi_bios_ptable(bdev);
623 if (bh) {
624 ret = scsi_partsize(bh, capacity,
625 &geom[2], &geom[0], &geom[1]);
626 kfree(bh);
627 if (ret != -1)
628 return (ret);
629 }
630 heads = 64;
631 sectors = 32;
632 cylinders = aic_sector_div(capacity, heads, sectors);
633
634 if (aic79xx_extended != 0)
635 extended = 1;
636 else
637 extended = (ahd->flags & AHD_EXTENDED_TRANS_A) != 0;
638 if (extended && cylinders >= 1024) {
639 heads = 255;
640 sectors = 63;
641 cylinders = aic_sector_div(capacity, heads, sectors);
642 }
643 geom[0] = heads;
644 geom[1] = sectors;
645 geom[2] = cylinders;
646 return (0);
647 }
648 #endif
649
650 /*
651 * Abort the current SCSI command(s).
652 */
653 static int
654 ahd_linux_abort(struct scsi_cmnd *cmd)
655 {
656 int error;
657
658 error = ahd_linux_queue_recovery_cmd(cmd, SCB_ABORT);
659 if (error != 0)
660 printf("aic79xx_abort returns 0x%x\n", error);
661 return error;
662 }
663
664 /*
665 * Attempt to send a target reset message to the device that timed out.
666 */
667 static int
668 ahd_linux_dev_reset(struct scsi_cmnd *cmd)
669 {
670 int error;
671
672 error = ahd_linux_queue_recovery_cmd(cmd, SCB_DEVICE_RESET);
673 if (error != 0)
674 printf("aic79xx_dev_reset returns 0x%x\n", error);
675 return error;
676 }
677
678 /*
679 * Reset the SCSI bus.
680 */
681 static int
682 ahd_linux_bus_reset(struct scsi_cmnd *cmd)
683 {
684 struct ahd_softc *ahd;
685 u_long s;
686 int found;
687
688 ahd = *(struct ahd_softc **)cmd->device->host->hostdata;
689 #ifdef AHD_DEBUG
690 if ((ahd_debug & AHD_SHOW_RECOVERY) != 0)
691 printf("%s: Bus reset called for cmd %p\n",
692 ahd_name(ahd), cmd);
693 #endif
694 ahd_lock(ahd, &s);
695 found = ahd_reset_channel(ahd, scmd_channel(cmd) + 'A',
696 /*initiate reset*/TRUE);
697 ahd_unlock(ahd, &s);
698
699 if (bootverbose)
700 printf("%s: SCSI bus reset delivered. "
701 "%d SCBs aborted.\n", ahd_name(ahd), found);
702
703 return (SUCCESS);
704 }
705
706 struct scsi_host_template aic79xx_driver_template = {
707 .module = THIS_MODULE,
708 .name = "aic79xx",
709 .proc_name = "aic79xx",
710 .proc_info = ahd_linux_proc_info,
711 .info = ahd_linux_info,
712 .queuecommand = ahd_linux_queue,
713 .eh_abort_handler = ahd_linux_abort,
714 .eh_device_reset_handler = ahd_linux_dev_reset,
715 .eh_bus_reset_handler = ahd_linux_bus_reset,
716 #if defined(__i386__)
717 .bios_param = ahd_linux_biosparam,
718 #endif
719 .can_queue = AHD_MAX_QUEUE,
720 .this_id = -1,
721 .cmd_per_lun = 2,
722 .use_clustering = ENABLE_CLUSTERING,
723 .slave_alloc = ahd_linux_slave_alloc,
724 .slave_configure = ahd_linux_slave_configure,
725 .slave_destroy = ahd_linux_slave_destroy,
726 .target_alloc = ahd_linux_target_alloc,
727 .target_destroy = ahd_linux_target_destroy,
728 };
729
730 /******************************** Bus DMA *************************************/
731 int
732 ahd_dma_tag_create(struct ahd_softc *ahd, bus_dma_tag_t parent,
733 bus_size_t alignment, bus_size_t boundary,
734 dma_addr_t lowaddr, dma_addr_t highaddr,
735 bus_dma_filter_t *filter, void *filterarg,
736 bus_size_t maxsize, int nsegments,
737 bus_size_t maxsegsz, int flags, bus_dma_tag_t *ret_tag)
738 {
739 bus_dma_tag_t dmat;
740
741 dmat = malloc(sizeof(*dmat), M_DEVBUF, M_NOWAIT);
742 if (dmat == NULL)
743 return (ENOMEM);
744
745 /*
746 * Linux is very simplistic about DMA memory. For now don't
747 * maintain all specification information. Once Linux supplies
748 * better facilities for doing these operations, or the
749 * needs of this particular driver change, we might need to do
750 * more here.
751 */
752 dmat->alignment = alignment;
753 dmat->boundary = boundary;
754 dmat->maxsize = maxsize;
755 *ret_tag = dmat;
756 return (0);
757 }
758
759 void
760 ahd_dma_tag_destroy(struct ahd_softc *ahd, bus_dma_tag_t dmat)
761 {
762 free(dmat, M_DEVBUF);
763 }
764
765 int
766 ahd_dmamem_alloc(struct ahd_softc *ahd, bus_dma_tag_t dmat, void** vaddr,
767 int flags, bus_dmamap_t *mapp)
768 {
769 *vaddr = pci_alloc_consistent(ahd->dev_softc,
770 dmat->maxsize, mapp);
771 if (*vaddr == NULL)
772 return (ENOMEM);
773 return(0);
774 }
775
776 void
777 ahd_dmamem_free(struct ahd_softc *ahd, bus_dma_tag_t dmat,
778 void* vaddr, bus_dmamap_t map)
779 {
780 pci_free_consistent(ahd->dev_softc, dmat->maxsize,
781 vaddr, map);
782 }
783
784 int
785 ahd_dmamap_load(struct ahd_softc *ahd, bus_dma_tag_t dmat, bus_dmamap_t map,
786 void *buf, bus_size_t buflen, bus_dmamap_callback_t *cb,
787 void *cb_arg, int flags)
788 {
789 /*
790 * Assume for now that this will only be used during
791 * initialization and not for per-transaction buffer mapping.
792 */
793 bus_dma_segment_t stack_sg;
794
795 stack_sg.ds_addr = map;
796 stack_sg.ds_len = dmat->maxsize;
797 cb(cb_arg, &stack_sg, /*nseg*/1, /*error*/0);
798 return (0);
799 }
800
801 void
802 ahd_dmamap_destroy(struct ahd_softc *ahd, bus_dma_tag_t dmat, bus_dmamap_t map)
803 {
804 }
805
806 int
807 ahd_dmamap_unload(struct ahd_softc *ahd, bus_dma_tag_t dmat, bus_dmamap_t map)
808 {
809 /* Nothing to do */
810 return (0);
811 }
812
813 /********************* Platform Dependent Functions ***************************/
814 /*
815 * Compare "left hand" softc with "right hand" softc, returning:
816 * < 0 - lahd has a lower priority than rahd
817 * 0 - Softcs are equal
818 * > 0 - lahd has a higher priority than rahd
819 */
820 int
821 ahd_softc_comp(struct ahd_softc *lahd, struct ahd_softc *rahd)
822 {
823 int value;
824
825 /*
826 * Under Linux, cards are ordered as follows:
827 * 1) PCI devices that are marked as the boot controller.
828 * 2) PCI devices with BIOS enabled sorted by bus/slot/func.
829 * 3) All remaining PCI devices sorted by bus/slot/func.
830 */
831 #if 0
832 value = (lahd->flags & AHD_BOOT_CHANNEL)
833 - (rahd->flags & AHD_BOOT_CHANNEL);
834 if (value != 0)
835 /* Controllers set for boot have a *higher* priority */
836 return (value);
837 #endif
838
839 value = (lahd->flags & AHD_BIOS_ENABLED)
840 - (rahd->flags & AHD_BIOS_ENABLED);
841 if (value != 0)
842 /* Controllers with BIOS enabled have a *higher* priority */
843 return (value);
844
845 /* Still equal. Sort by bus/slot/func. */
846 if (aic79xx_reverse_scan != 0)
847 value = ahd_get_pci_bus(lahd->dev_softc)
848 - ahd_get_pci_bus(rahd->dev_softc);
849 else
850 value = ahd_get_pci_bus(rahd->dev_softc)
851 - ahd_get_pci_bus(lahd->dev_softc);
852 if (value != 0)
853 return (value);
854 if (aic79xx_reverse_scan != 0)
855 value = ahd_get_pci_slot(lahd->dev_softc)
856 - ahd_get_pci_slot(rahd->dev_softc);
857 else
858 value = ahd_get_pci_slot(rahd->dev_softc)
859 - ahd_get_pci_slot(lahd->dev_softc);
860 if (value != 0)
861 return (value);
862
863 value = rahd->channel - lahd->channel;
864 return (value);
865 }
866
867 static void
868 ahd_linux_setup_iocell_info(u_long index, int instance, int targ, int32_t value)
869 {
870
871 if ((instance >= 0)
872 && (instance < NUM_ELEMENTS(aic79xx_iocell_info))) {
873 uint8_t *iocell_info;
874
875 iocell_info = (uint8_t*)&aic79xx_iocell_info[instance];
876 iocell_info[index] = value & 0xFFFF;
877 if (bootverbose)
878 printf("iocell[%d:%ld] = %d\n", instance, index, value);
879 }
880 }
881
882 static void
883 ahd_linux_setup_tag_info_global(char *p)
884 {
885 int tags, i, j;
886
887 tags = simple_strtoul(p + 1, NULL, 0) & 0xff;
888 printf("Setting Global Tags= %d\n", tags);
889
890 for (i = 0; i < NUM_ELEMENTS(aic79xx_tag_info); i++) {
891 for (j = 0; j < AHD_NUM_TARGETS; j++) {
892 aic79xx_tag_info[i].tag_commands[j] = tags;
893 }
894 }
895 }
896
897 static void
898 ahd_linux_setup_tag_info(u_long arg, int instance, int targ, int32_t value)
899 {
900
901 if ((instance >= 0) && (targ >= 0)
902 && (instance < NUM_ELEMENTS(aic79xx_tag_info))
903 && (targ < AHD_NUM_TARGETS)) {
904 aic79xx_tag_info[instance].tag_commands[targ] = value & 0x1FF;
905 if (bootverbose)
906 printf("tag_info[%d:%d] = %d\n", instance, targ, value);
907 }
908 }
909
910 static char *
911 ahd_parse_brace_option(char *opt_name, char *opt_arg, char *end, int depth,
912 void (*callback)(u_long, int, int, int32_t),
913 u_long callback_arg)
914 {
915 char *tok_end;
916 char *tok_end2;
917 int i;
918 int instance;
919 int targ;
920 int done;
921 char tok_list[] = {'.', ',', '{', '}', '\0'};
922
923 /* All options use a ':' name/arg separator */
924 if (*opt_arg != ':')
925 return (opt_arg);
926 opt_arg++;
927 instance = -1;
928 targ = -1;
929 done = FALSE;
930 /*
931 * Restore separator that may be in
932 * the middle of our option argument.
933 */
934 tok_end = strchr(opt_arg, '\0');
935 if (tok_end < end)
936 *tok_end = ',';
937 while (!done) {
938 switch (*opt_arg) {
939 case '{':
940 if (instance == -1) {
941 instance = 0;
942 } else {
943 if (depth > 1) {
944 if (targ == -1)
945 targ = 0;
946 } else {
947 printf("Malformed Option %s\n",
948 opt_name);
949 done = TRUE;
950 }
951 }
952 opt_arg++;
953 break;
954 case '}':
955 if (targ != -1)
956 targ = -1;
957 else if (instance != -1)
958 instance = -1;
959 opt_arg++;
960 break;
961 case ',':
962 case '.':
963 if (instance == -1)
964 done = TRUE;
965 else if (targ >= 0)
966 targ++;
967 else if (instance >= 0)
968 instance++;
969 opt_arg++;
970 break;
971 case '\0':
972 done = TRUE;
973 break;
974 default:
975 tok_end = end;
976 for (i = 0; tok_list[i]; i++) {
977 tok_end2 = strchr(opt_arg, tok_list[i]);
978 if ((tok_end2) && (tok_end2 < tok_end))
979 tok_end = tok_end2;
980 }
981 callback(callback_arg, instance, targ,
982 simple_strtol(opt_arg, NULL, 0));
983 opt_arg = tok_end;
984 break;
985 }
986 }
987 return (opt_arg);
988 }
989
990 /*
991 * Handle Linux boot parameters. This routine allows for assigning a value
992 * to a parameter with a ':' between the parameter and the value.
993 * ie. aic79xx=stpwlev:1,extended
994 */
995 static int
996 aic79xx_setup(char *s)
997 {
998 int i, n;
999 char *p;
1000 char *end;
1001
1002 static struct {
1003 const char *name;
1004 uint32_t *flag;
1005 } options[] = {
1006 { "extended", &aic79xx_extended },
1007 { "no_reset", &aic79xx_no_reset },
1008 { "verbose", &aic79xx_verbose },
1009 { "allow_memio", &aic79xx_allow_memio},
1010 #ifdef AHD_DEBUG
1011 { "debug", &ahd_debug },
1012 #endif
1013 { "reverse_scan", &aic79xx_reverse_scan },
1014 { "periodic_otag", &aic79xx_periodic_otag },
1015 { "pci_parity", &aic79xx_pci_parity },
1016 { "seltime", &aic79xx_seltime },
1017 { "tag_info", NULL },
1018 { "global_tag_depth", NULL},
1019 { "slewrate", NULL },
1020 { "precomp", NULL },
1021 { "amplitude", NULL },
1022 { "slowcrc", &aic79xx_slowcrc },
1023 };
1024
1025 end = strchr(s, '\0');
1026
1027 /*
1028 * XXX ia64 gcc isn't smart enough to know that NUM_ELEMENTS
1029 * will never be 0 in this case.
1030 */
1031 n = 0;
1032
1033 while ((p = strsep(&s, ",.")) != NULL) {
1034 if (*p == '\0')
1035 continue;
1036 for (i = 0; i < NUM_ELEMENTS(options); i++) {
1037
1038 n = strlen(options[i].name);
1039 if (strncmp(options[i].name, p, n) == 0)
1040 break;
1041 }
1042 if (i == NUM_ELEMENTS(options))
1043 continue;
1044
1045 if (strncmp(p, "global_tag_depth", n) == 0) {
1046 ahd_linux_setup_tag_info_global(p + n);
1047 } else if (strncmp(p, "tag_info", n) == 0) {
1048 s = ahd_parse_brace_option("tag_info", p + n, end,
1049 2, ahd_linux_setup_tag_info, 0);
1050 } else if (strncmp(p, "slewrate", n) == 0) {
1051 s = ahd_parse_brace_option("slewrate",
1052 p + n, end, 1, ahd_linux_setup_iocell_info,
1053 AIC79XX_SLEWRATE_INDEX);
1054 } else if (strncmp(p, "precomp", n) == 0) {
1055 s = ahd_parse_brace_option("precomp",
1056 p + n, end, 1, ahd_linux_setup_iocell_info,
1057 AIC79XX_PRECOMP_INDEX);
1058 } else if (strncmp(p, "amplitude", n) == 0) {
1059 s = ahd_parse_brace_option("amplitude",
1060 p + n, end, 1, ahd_linux_setup_iocell_info,
1061 AIC79XX_AMPLITUDE_INDEX);
1062 } else if (p[n] == ':') {
1063 *(options[i].flag) = simple_strtoul(p + n + 1, NULL, 0);
1064 } else if (!strncmp(p, "verbose", n)) {
1065 *(options[i].flag) = 1;
1066 } else {
1067 *(options[i].flag) ^= 0xFFFFFFFF;
1068 }
1069 }
1070 return 1;
1071 }
1072
1073 __setup("aic79xx=", aic79xx_setup);
1074
1075 uint32_t aic79xx_verbose;
1076
1077 int
1078 ahd_linux_register_host(struct ahd_softc *ahd, struct scsi_host_template *template)
1079 {
1080 char buf[80];
1081 struct Scsi_Host *host;
1082 char *new_name;
1083 u_long s;
1084 int retval;
1085
1086 template->name = ahd->description;
1087 host = scsi_host_alloc(template, sizeof(struct ahd_softc *));
1088 if (host == NULL)
1089 return (ENOMEM);
1090
1091 *((struct ahd_softc **)host->hostdata) = ahd;
1092 ahd->platform_data->host = host;
1093 host->can_queue = AHD_MAX_QUEUE;
1094 host->cmd_per_lun = 2;
1095 host->sg_tablesize = AHD_NSEG;
1096 host->this_id = ahd->our_id;
1097 host->irq = ahd->platform_data->irq;
1098 host->max_id = (ahd->features & AHD_WIDE) ? 16 : 8;
1099 host->max_lun = AHD_NUM_LUNS;
1100 host->max_channel = 0;
1101 host->sg_tablesize = AHD_NSEG;
1102 ahd_lock(ahd, &s);
1103 ahd_set_unit(ahd, ahd_linux_unit++);
1104 ahd_unlock(ahd, &s);
1105 sprintf(buf, "scsi%d", host->host_no);
1106 new_name = malloc(strlen(buf) + 1, M_DEVBUF, M_NOWAIT);
1107 if (new_name != NULL) {
1108 strcpy(new_name, buf);
1109 ahd_set_name(ahd, new_name);
1110 }
1111 host->unique_id = ahd->unit;
1112 ahd_linux_initialize_scsi_bus(ahd);
1113 ahd_intr_enable(ahd, TRUE);
1114
1115 host->transportt = ahd_linux_transport_template;
1116
1117 retval = scsi_add_host(host, &ahd->dev_softc->dev);
1118 if (retval) {
1119 printk(KERN_WARNING "aic79xx: scsi_add_host failed\n");
1120 scsi_host_put(host);
1121 return retval;
1122 }
1123
1124 scsi_scan_host(host);
1125 return 0;
1126 }
1127
1128 uint64_t
1129 ahd_linux_get_memsize(void)
1130 {
1131 struct sysinfo si;
1132
1133 si_meminfo(&si);
1134 return ((uint64_t)si.totalram << PAGE_SHIFT);
1135 }
1136
1137 /*
1138 * Place the SCSI bus into a known state by either resetting it,
1139 * or forcing transfer negotiations on the next command to any
1140 * target.
1141 */
1142 static void
1143 ahd_linux_initialize_scsi_bus(struct ahd_softc *ahd)
1144 {
1145 u_int target_id;
1146 u_int numtarg;
1147 unsigned long s;
1148
1149 target_id = 0;
1150 numtarg = 0;
1151
1152 if (aic79xx_no_reset != 0)
1153 ahd->flags &= ~AHD_RESET_BUS_A;
1154
1155 if ((ahd->flags & AHD_RESET_BUS_A) != 0)
1156 ahd_reset_channel(ahd, 'A', /*initiate_reset*/TRUE);
1157 else
1158 numtarg = (ahd->features & AHD_WIDE) ? 16 : 8;
1159
1160 ahd_lock(ahd, &s);
1161
1162 /*
1163 * Force negotiation to async for all targets that
1164 * will not see an initial bus reset.
1165 */
1166 for (; target_id < numtarg; target_id++) {
1167 struct ahd_devinfo devinfo;
1168 struct ahd_initiator_tinfo *tinfo;
1169 struct ahd_tmode_tstate *tstate;
1170
1171 tinfo = ahd_fetch_transinfo(ahd, 'A', ahd->our_id,
1172 target_id, &tstate);
1173 ahd_compile_devinfo(&devinfo, ahd->our_id, target_id,
1174 CAM_LUN_WILDCARD, 'A', ROLE_INITIATOR);
1175 ahd_update_neg_request(ahd, &devinfo, tstate,
1176 tinfo, AHD_NEG_ALWAYS);
1177 }
1178 ahd_unlock(ahd, &s);
1179 /* Give the bus some time to recover */
1180 if ((ahd->flags & AHD_RESET_BUS_A) != 0) {
1181 ahd_freeze_simq(ahd);
1182 msleep(AIC79XX_RESET_DELAY);
1183 ahd_release_simq(ahd);
1184 }
1185 }
1186
1187 int
1188 ahd_platform_alloc(struct ahd_softc *ahd, void *platform_arg)
1189 {
1190 ahd->platform_data =
1191 malloc(sizeof(struct ahd_platform_data), M_DEVBUF, M_NOWAIT);
1192 if (ahd->platform_data == NULL)
1193 return (ENOMEM);
1194 memset(ahd->platform_data, 0, sizeof(struct ahd_platform_data));
1195 ahd->platform_data->irq = AHD_LINUX_NOIRQ;
1196 ahd_lockinit(ahd);
1197 init_MUTEX_LOCKED(&ahd->platform_data->eh_sem);
1198 ahd->seltime = (aic79xx_seltime & 0x3) << 4;
1199 return (0);
1200 }
1201
1202 void
1203 ahd_platform_free(struct ahd_softc *ahd)
1204 {
1205 struct scsi_target *starget;
1206 int i, j;
1207
1208 if (ahd->platform_data != NULL) {
1209 /* destroy all of the device and target objects */
1210 for (i = 0; i < AHD_NUM_TARGETS; i++) {
1211 starget = ahd->platform_data->starget[i];
1212 if (starget != NULL) {
1213 for (j = 0; j < AHD_NUM_LUNS; j++) {
1214 struct ahd_linux_target *targ =
1215 scsi_transport_target_data(starget);
1216 if (targ->sdev[j] == NULL)
1217 continue;
1218 targ->sdev[j] = NULL;
1219 }
1220 ahd->platform_data->starget[i] = NULL;
1221 }
1222 }
1223
1224 if (ahd->platform_data->irq != AHD_LINUX_NOIRQ)
1225 free_irq(ahd->platform_data->irq, ahd);
1226 if (ahd->tags[0] == BUS_SPACE_PIO
1227 && ahd->bshs[0].ioport != 0)
1228 release_region(ahd->bshs[0].ioport, 256);
1229 if (ahd->tags[1] == BUS_SPACE_PIO
1230 && ahd->bshs[1].ioport != 0)
1231 release_region(ahd->bshs[1].ioport, 256);
1232 if (ahd->tags[0] == BUS_SPACE_MEMIO
1233 && ahd->bshs[0].maddr != NULL) {
1234 iounmap(ahd->bshs[0].maddr);
1235 release_mem_region(ahd->platform_data->mem_busaddr,
1236 0x1000);
1237 }
1238 if (ahd->platform_data->host)
1239 scsi_host_put(ahd->platform_data->host);
1240
1241 free(ahd->platform_data, M_DEVBUF);
1242 }
1243 }
1244
1245 void
1246 ahd_platform_init(struct ahd_softc *ahd)
1247 {
1248 /*
1249 * Lookup and commit any modified IO Cell options.
1250 */
1251 if (ahd->unit < NUM_ELEMENTS(aic79xx_iocell_info)) {
1252 struct ahd_linux_iocell_opts *iocell_opts;
1253
1254 iocell_opts = &aic79xx_iocell_info[ahd->unit];
1255 if (iocell_opts->precomp != AIC79XX_DEFAULT_PRECOMP)
1256 AHD_SET_PRECOMP(ahd, iocell_opts->precomp);
1257 if (iocell_opts->slewrate != AIC79XX_DEFAULT_SLEWRATE)
1258 AHD_SET_SLEWRATE(ahd, iocell_opts->slewrate);
1259 if (iocell_opts->amplitude != AIC79XX_DEFAULT_AMPLITUDE)
1260 AHD_SET_AMPLITUDE(ahd, iocell_opts->amplitude);
1261 }
1262
1263 }
1264
1265 void
1266 ahd_platform_freeze_devq(struct ahd_softc *ahd, struct scb *scb)
1267 {
1268 ahd_platform_abort_scbs(ahd, SCB_GET_TARGET(ahd, scb),
1269 SCB_GET_CHANNEL(ahd, scb),
1270 SCB_GET_LUN(scb), SCB_LIST_NULL,
1271 ROLE_UNKNOWN, CAM_REQUEUE_REQ);
1272 }
1273
1274 void
1275 ahd_platform_set_tags(struct ahd_softc *ahd, struct ahd_devinfo *devinfo,
1276 ahd_queue_alg alg)
1277 {
1278 struct scsi_target *starget;
1279 struct ahd_linux_target *targ;
1280 struct ahd_linux_device *dev;
1281 struct scsi_device *sdev;
1282 int was_queuing;
1283 int now_queuing;
1284
1285 starget = ahd->platform_data->starget[devinfo->target];
1286 targ = scsi_transport_target_data(starget);
1287 BUG_ON(targ == NULL);
1288 sdev = targ->sdev[devinfo->lun];
1289 if (sdev == NULL)
1290 return;
1291
1292 dev = scsi_transport_device_data(sdev);
1293
1294 if (dev == NULL)
1295 return;
1296 was_queuing = dev->flags & (AHD_DEV_Q_BASIC|AHD_DEV_Q_TAGGED);
1297 switch (alg) {
1298 default:
1299 case AHD_QUEUE_NONE:
1300 now_queuing = 0;
1301 break;
1302 case AHD_QUEUE_BASIC:
1303 now_queuing = AHD_DEV_Q_BASIC;
1304 break;
1305 case AHD_QUEUE_TAGGED:
1306 now_queuing = AHD_DEV_Q_TAGGED;
1307 break;
1308 }
1309 if ((dev->flags & AHD_DEV_FREEZE_TIL_EMPTY) == 0
1310 && (was_queuing != now_queuing)
1311 && (dev->active != 0)) {
1312 dev->flags |= AHD_DEV_FREEZE_TIL_EMPTY;
1313 dev->qfrozen++;
1314 }
1315
1316 dev->flags &= ~(AHD_DEV_Q_BASIC|AHD_DEV_Q_TAGGED|AHD_DEV_PERIODIC_OTAG);
1317 if (now_queuing) {
1318 u_int usertags;
1319
1320 usertags = ahd_linux_user_tagdepth(ahd, devinfo);
1321 if (!was_queuing) {
1322 /*
1323 * Start out agressively and allow our
1324 * dynamic queue depth algorithm to take
1325 * care of the rest.
1326 */
1327 dev->maxtags = usertags;
1328 dev->openings = dev->maxtags - dev->active;
1329 }
1330 if (dev->maxtags == 0) {
1331 /*
1332 * Queueing is disabled by the user.
1333 */
1334 dev->openings = 1;
1335 } else if (alg == AHD_QUEUE_TAGGED) {
1336 dev->flags |= AHD_DEV_Q_TAGGED;
1337 if (aic79xx_periodic_otag != 0)
1338 dev->flags |= AHD_DEV_PERIODIC_OTAG;
1339 } else
1340 dev->flags |= AHD_DEV_Q_BASIC;
1341 } else {
1342 /* We can only have one opening. */
1343 dev->maxtags = 0;
1344 dev->openings = 1 - dev->active;
1345 }
1346
1347 switch ((dev->flags & (AHD_DEV_Q_BASIC|AHD_DEV_Q_TAGGED))) {
1348 case AHD_DEV_Q_BASIC:
1349 scsi_adjust_queue_depth(sdev,
1350 MSG_SIMPLE_TASK,
1351 dev->openings + dev->active);
1352 break;
1353 case AHD_DEV_Q_TAGGED:
1354 scsi_adjust_queue_depth(sdev,
1355 MSG_ORDERED_TASK,
1356 dev->openings + dev->active);
1357 break;
1358 default:
1359 /*
1360 * We allow the OS to queue 2 untagged transactions to
1361 * us at any time even though we can only execute them
1362 * serially on the controller/device. This should
1363 * remove some latency.
1364 */
1365 scsi_adjust_queue_depth(sdev,
1366 /*NON-TAGGED*/0,
1367 /*queue depth*/2);
1368 break;
1369 }
1370 }
1371
1372 int
1373 ahd_platform_abort_scbs(struct ahd_softc *ahd, int target, char channel,
1374 int lun, u_int tag, role_t role, uint32_t status)
1375 {
1376 return 0;
1377 }
1378
1379 static u_int
1380 ahd_linux_user_tagdepth(struct ahd_softc *ahd, struct ahd_devinfo *devinfo)
1381 {
1382 static int warned_user;
1383 u_int tags;
1384
1385 tags = 0;
1386 if ((ahd->user_discenable & devinfo->target_mask) != 0) {
1387 if (ahd->unit >= NUM_ELEMENTS(aic79xx_tag_info)) {
1388
1389 if (warned_user == 0) {
1390 printf(KERN_WARNING
1391 "aic79xx: WARNING: Insufficient tag_info instances\n"
1392 "aic79xx: for installed controllers. Using defaults\n"
1393 "aic79xx: Please update the aic79xx_tag_info array in\n"
1394 "aic79xx: the aic79xx_osm.c source file.\n");
1395 warned_user++;
1396 }
1397 tags = AHD_MAX_QUEUE;
1398 } else {
1399 adapter_tag_info_t *tag_info;
1400
1401 tag_info = &aic79xx_tag_info[ahd->unit];
1402 tags = tag_info->tag_commands[devinfo->target_offset];
1403 if (tags > AHD_MAX_QUEUE)
1404 tags = AHD_MAX_QUEUE;
1405 }
1406 }
1407 return (tags);
1408 }
1409
1410 /*
1411 * Determines the queue depth for a given device.
1412 */
1413 static void
1414 ahd_linux_device_queue_depth(struct scsi_device *sdev)
1415 {
1416 struct ahd_devinfo devinfo;
1417 u_int tags;
1418 struct ahd_softc *ahd = *((struct ahd_softc **)sdev->host->hostdata);
1419
1420 ahd_compile_devinfo(&devinfo,
1421 ahd->our_id,
1422 sdev->sdev_target->id, sdev->lun,
1423 sdev->sdev_target->channel == 0 ? 'A' : 'B',
1424 ROLE_INITIATOR);
1425 tags = ahd_linux_user_tagdepth(ahd, &devinfo);
1426 if (tags != 0 && sdev->tagged_supported != 0) {
1427
1428 ahd_set_tags(ahd, &devinfo, AHD_QUEUE_TAGGED);
1429 ahd_print_devinfo(ahd, &devinfo);
1430 printf("Tagged Queuing enabled. Depth %d\n", tags);
1431 } else {
1432 ahd_set_tags(ahd, &devinfo, AHD_QUEUE_NONE);
1433 }
1434 }
1435
1436 static int
1437 ahd_linux_run_command(struct ahd_softc *ahd, struct ahd_linux_device *dev,
1438 struct scsi_cmnd *cmd)
1439 {
1440 struct scb *scb;
1441 struct hardware_scb *hscb;
1442 struct ahd_initiator_tinfo *tinfo;
1443 struct ahd_tmode_tstate *tstate;
1444 u_int col_idx;
1445 uint16_t mask;
1446
1447 /*
1448 * Get an scb to use.
1449 */
1450 tinfo = ahd_fetch_transinfo(ahd, 'A', ahd->our_id,
1451 cmd->device->id, &tstate);
1452 if ((dev->flags & (AHD_DEV_Q_TAGGED|AHD_DEV_Q_BASIC)) == 0
1453 || (tinfo->curr.ppr_options & MSG_EXT_PPR_IU_REQ) != 0) {
1454 col_idx = AHD_NEVER_COL_IDX;
1455 } else {
1456 col_idx = AHD_BUILD_COL_IDX(cmd->device->id,
1457 cmd->device->lun);
1458 }
1459 if ((scb = ahd_get_scb(ahd, col_idx)) == NULL) {
1460 ahd->flags |= AHD_RESOURCE_SHORTAGE;
1461 return SCSI_MLQUEUE_HOST_BUSY;
1462 }
1463
1464 scb->io_ctx = cmd;
1465 scb->platform_data->dev = dev;
1466 hscb = scb->hscb;
1467 cmd->host_scribble = (char *)scb;
1468
1469 /*
1470 * Fill out basics of the HSCB.
1471 */
1472 hscb->control = 0;
1473 hscb->scsiid = BUILD_SCSIID(ahd, cmd);
1474 hscb->lun = cmd->device->lun;
1475 scb->hscb->task_management = 0;
1476 mask = SCB_GET_TARGET_MASK(ahd, scb);
1477
1478 if ((ahd->user_discenable & mask) != 0)
1479 hscb->control |= DISCENB;
1480
1481 if ((tinfo->curr.ppr_options & MSG_EXT_PPR_IU_REQ) != 0)
1482 scb->flags |= SCB_PACKETIZED;
1483
1484 if ((tstate->auto_negotiate & mask) != 0) {
1485 scb->flags |= SCB_AUTO_NEGOTIATE;
1486 scb->hscb->control |= MK_MESSAGE;
1487 } else if (cmd->cmnd[0] == INQUIRY
1488 && (tinfo->curr.offset != 0
1489 || tinfo->curr.width != MSG_EXT_WDTR_BUS_8_BIT
1490 || tinfo->curr.ppr_options != 0)
1491 && (tinfo->curr.ppr_options & MSG_EXT_PPR_IU_REQ)==0) {
1492 /*
1493 * The SCSI spec requires inquiry
1494 * commands to complete without
1495 * reporting unit attention conditions.
1496 * Because of this, an inquiry command
1497 * that occurs just after a device is
1498 * reset will result in a data phase
1499 * with mismatched negotiated rates.
1500 * The core already forces a renegotiation
1501 * for reset events that are visible to
1502 * our controller or that we initiate,
1503 * but a third party device reset or a
1504 * hot-plug insertion can still cause this
1505 * issue. Therefore, we force a re-negotiation
1506 * for every inquiry command unless we
1507 * are async.
1508 */
1509 scb->flags |= SCB_NEGOTIATE;
1510 scb->hscb->control |= MK_MESSAGE;
1511 }
1512
1513 if ((dev->flags & (AHD_DEV_Q_TAGGED|AHD_DEV_Q_BASIC)) != 0) {
1514 int msg_bytes;
1515 uint8_t tag_msgs[2];
1516
1517 msg_bytes = scsi_populate_tag_msg(cmd, tag_msgs);
1518 if (msg_bytes && tag_msgs[0] != MSG_SIMPLE_TASK) {
1519 hscb->control |= tag_msgs[0];
1520 if (tag_msgs[0] == MSG_ORDERED_TASK)
1521 dev->commands_since_idle_or_otag = 0;
1522 } else
1523 if (dev->commands_since_idle_or_otag == AHD_OTAG_THRESH
1524 && (dev->flags & AHD_DEV_Q_TAGGED) != 0) {
1525 hscb->control |= MSG_ORDERED_TASK;
1526 dev->commands_since_idle_or_otag = 0;
1527 } else {
1528 hscb->control |= MSG_SIMPLE_TASK;
1529 }
1530 }
1531
1532 hscb->cdb_len = cmd->cmd_len;
1533 memcpy(hscb->shared_data.idata.cdb, cmd->cmnd, hscb->cdb_len);
1534
1535 scb->platform_data->xfer_len = 0;
1536 ahd_set_residual(scb, 0);
1537 ahd_set_sense_residual(scb, 0);
1538 scb->sg_count = 0;
1539 if (cmd->use_sg != 0) {
1540 void *sg;
1541 struct scatterlist *cur_seg;
1542 u_int nseg;
1543 int dir;
1544
1545 cur_seg = (struct scatterlist *)cmd->request_buffer;
1546 dir = cmd->sc_data_direction;
1547 nseg = pci_map_sg(ahd->dev_softc, cur_seg,
1548 cmd->use_sg, dir);
1549 scb->platform_data->xfer_len = 0;
1550 for (sg = scb->sg_list; nseg > 0; nseg--, cur_seg++) {
1551 dma_addr_t addr;
1552 bus_size_t len;
1553
1554 addr = sg_dma_address(cur_seg);
1555 len = sg_dma_len(cur_seg);
1556 scb->platform_data->xfer_len += len;
1557 sg = ahd_sg_setup(ahd, scb, sg, addr, len,
1558 /*last*/nseg == 1);
1559 }
1560 } else if (cmd->request_bufflen != 0) {
1561 void *sg;
1562 dma_addr_t addr;
1563 int dir;
1564
1565 sg = scb->sg_list;
1566 dir = cmd->sc_data_direction;
1567 addr = pci_map_single(ahd->dev_softc,
1568 cmd->request_buffer,
1569 cmd->request_bufflen, dir);
1570 scb->platform_data->xfer_len = cmd->request_bufflen;
1571 scb->platform_data->buf_busaddr = addr;
1572 sg = ahd_sg_setup(ahd, scb, sg, addr,
1573 cmd->request_bufflen, /*last*/TRUE);
1574 }
1575
1576 LIST_INSERT_HEAD(&ahd->pending_scbs, scb, pending_links);
1577 dev->openings--;
1578 dev->active++;
1579 dev->commands_issued++;
1580
1581 if ((dev->flags & AHD_DEV_PERIODIC_OTAG) != 0)
1582 dev->commands_since_idle_or_otag++;
1583 scb->flags |= SCB_ACTIVE;
1584 ahd_queue_scb(ahd, scb);
1585
1586 return 0;
1587 }
1588
1589 /*
1590 * SCSI controller interrupt handler.
1591 */
1592 irqreturn_t
1593 ahd_linux_isr(int irq, void *dev_id, struct pt_regs * regs)
1594 {
1595 struct ahd_softc *ahd;
1596 u_long flags;
1597 int ours;
1598
1599 ahd = (struct ahd_softc *) dev_id;
1600 ahd_lock(ahd, &flags);
1601 ours = ahd_intr(ahd);
1602 ahd_unlock(ahd, &flags);
1603 return IRQ_RETVAL(ours);
1604 }
1605
1606 void
1607 ahd_platform_flushwork(struct ahd_softc *ahd)
1608 {
1609
1610 }
1611
1612 void
1613 ahd_send_async(struct ahd_softc *ahd, char channel,
1614 u_int target, u_int lun, ac_code code, void *arg)
1615 {
1616 switch (code) {
1617 case AC_TRANSFER_NEG:
1618 {
1619 char buf[80];
1620 struct scsi_target *starget;
1621 struct ahd_linux_target *targ;
1622 struct info_str info;
1623 struct ahd_initiator_tinfo *tinfo;
1624 struct ahd_tmode_tstate *tstate;
1625 unsigned int target_ppr_options;
1626
1627 BUG_ON(target == CAM_TARGET_WILDCARD);
1628
1629 info.buffer = buf;
1630 info.length = sizeof(buf);
1631 info.offset = 0;
1632 info.pos = 0;
1633 tinfo = ahd_fetch_transinfo(ahd, channel, ahd->our_id,
1634 target, &tstate);
1635
1636 /*
1637 * Don't bother reporting results while
1638 * negotiations are still pending.
1639 */
1640 if (tinfo->curr.period != tinfo->goal.period
1641 || tinfo->curr.width != tinfo->goal.width
1642 || tinfo->curr.offset != tinfo->goal.offset
1643 || tinfo->curr.ppr_options != tinfo->goal.ppr_options)
1644 if (bootverbose == 0)
1645 break;
1646
1647 /*
1648 * Don't bother reporting results that
1649 * are identical to those last reported.
1650 */
1651 starget = ahd->platform_data->starget[target];
1652 if (starget == NULL)
1653 break;
1654 targ = scsi_transport_target_data(starget);
1655
1656 target_ppr_options =
1657 (spi_dt(starget) ? MSG_EXT_PPR_DT_REQ : 0)
1658 + (spi_qas(starget) ? MSG_EXT_PPR_QAS_REQ : 0)
1659 + (spi_iu(starget) ? MSG_EXT_PPR_IU_REQ : 0)
1660 + (spi_rd_strm(starget) ? MSG_EXT_PPR_RD_STRM : 0)
1661 + (spi_pcomp_en(starget) ? MSG_EXT_PPR_PCOMP_EN : 0)
1662 + (spi_rti(starget) ? MSG_EXT_PPR_RTI : 0)
1663 + (spi_wr_flow(starget) ? MSG_EXT_PPR_WR_FLOW : 0)
1664 + (spi_hold_mcs(starget) ? MSG_EXT_PPR_HOLD_MCS : 0);
1665
1666 if (tinfo->curr.period == spi_period(starget)
1667 && tinfo->curr.width == spi_width(starget)
1668 && tinfo->curr.offset == spi_offset(starget)
1669 && tinfo->curr.ppr_options == target_ppr_options)
1670 if (bootverbose == 0)
1671 break;
1672
1673 spi_period(starget) = tinfo->curr.period;
1674 spi_width(starget) = tinfo->curr.width;
1675 spi_offset(starget) = tinfo->curr.offset;
1676 spi_dt(starget) = tinfo->curr.ppr_options & MSG_EXT_PPR_DT_REQ ? 1 : 0;
1677 spi_qas(starget) = tinfo->curr.ppr_options & MSG_EXT_PPR_QAS_REQ ? 1 : 0;
1678 spi_iu(starget) = tinfo->curr.ppr_options & MSG_EXT_PPR_IU_REQ ? 1 : 0;
1679 spi_rd_strm(starget) = tinfo->curr.ppr_options & MSG_EXT_PPR_RD_STRM ? 1 : 0;
1680 spi_pcomp_en(starget) = tinfo->curr.ppr_options & MSG_EXT_PPR_PCOMP_EN ? 1 : 0;
1681 spi_rti(starget) = tinfo->curr.ppr_options & MSG_EXT_PPR_RTI ? 1 : 0;
1682 spi_wr_flow(starget) = tinfo->curr.ppr_options & MSG_EXT_PPR_WR_FLOW ? 1 : 0;
1683 spi_hold_mcs(starget) = tinfo->curr.ppr_options & MSG_EXT_PPR_HOLD_MCS ? 1 : 0;
1684 spi_display_xfer_agreement(starget);
1685 break;
1686 }
1687 case AC_SENT_BDR:
1688 {
1689 WARN_ON(lun != CAM_LUN_WILDCARD);
1690 scsi_report_device_reset(ahd->platform_data->host,
1691 channel - 'A', target);
1692 break;
1693 }
1694 case AC_BUS_RESET:
1695 if (ahd->platform_data->host != NULL) {
1696 scsi_report_bus_reset(ahd->platform_data->host,
1697 channel - 'A');
1698 }
1699 break;
1700 default:
1701 panic("ahd_send_async: Unexpected async event");
1702 }
1703 }
1704
1705 /*
1706 * Calls the higher level scsi done function and frees the scb.
1707 */
1708 void
1709 ahd_done(struct ahd_softc *ahd, struct scb *scb)
1710 {
1711 struct scsi_cmnd *cmd;
1712 struct ahd_linux_device *dev;
1713
1714 if ((scb->flags & SCB_ACTIVE) == 0) {
1715 printf("SCB %d done'd twice\n", SCB_GET_TAG(scb));
1716 ahd_dump_card_state(ahd);
1717 panic("Stopping for safety");
1718 }
1719 LIST_REMOVE(scb, pending_links);
1720 cmd = scb->io_ctx;
1721 dev = scb->platform_data->dev;
1722 dev->active--;
1723 dev->openings++;
1724 if ((cmd->result & (CAM_DEV_QFRZN << 16)) != 0) {
1725 cmd->result &= ~(CAM_DEV_QFRZN << 16);
1726 dev->qfrozen--;
1727 }
1728 ahd_linux_unmap_scb(ahd, scb);
1729
1730 /*
1731 * Guard against stale sense data.
1732 * The Linux mid-layer assumes that sense
1733 * was retrieved anytime the first byte of
1734 * the sense buffer looks "sane".
1735 */
1736 cmd->sense_buffer[0] = 0;
1737 if (ahd_get_transaction_status(scb) == CAM_REQ_INPROG) {
1738 uint32_t amount_xferred;
1739
1740 amount_xferred =
1741 ahd_get_transfer_length(scb) - ahd_get_residual(scb);
1742 if ((scb->flags & SCB_TRANSMISSION_ERROR) != 0) {
1743 #ifdef AHD_DEBUG
1744 if ((ahd_debug & AHD_SHOW_MISC) != 0) {
1745 ahd_print_path(ahd, scb);
1746 printf("Set CAM_UNCOR_PARITY\n");
1747 }
1748 #endif
1749 ahd_set_transaction_status(scb, CAM_UNCOR_PARITY);
1750 #ifdef AHD_REPORT_UNDERFLOWS
1751 /*
1752 * This code is disabled by default as some
1753 * clients of the SCSI system do not properly
1754 * initialize the underflow parameter. This
1755 * results in spurious termination of commands
1756 * that complete as expected (e.g. underflow is
1757 * allowed as command can return variable amounts
1758 * of data.
1759 */
1760 } else if (amount_xferred < scb->io_ctx->underflow) {
1761 u_int i;
1762
1763 ahd_print_path(ahd, scb);
1764 printf("CDB:");
1765 for (i = 0; i < scb->io_ctx->cmd_len; i++)
1766 printf(" 0x%x", scb->io_ctx->cmnd[i]);
1767 printf("\n");
1768 ahd_print_path(ahd, scb);
1769 printf("Saw underflow (%ld of %ld bytes). "
1770 "Treated as error\n",
1771 ahd_get_residual(scb),
1772 ahd_get_transfer_length(scb));
1773 ahd_set_transaction_status(scb, CAM_DATA_RUN_ERR);
1774 #endif
1775 } else {
1776 ahd_set_transaction_status(scb, CAM_REQ_CMP);
1777 }
1778 } else if (ahd_get_transaction_status(scb) == CAM_SCSI_STATUS_ERROR) {
1779 ahd_linux_handle_scsi_status(ahd, cmd->device, scb);
1780 }
1781
1782 if (dev->openings == 1
1783 && ahd_get_transaction_status(scb) == CAM_REQ_CMP
1784 && ahd_get_scsi_status(scb) != SCSI_STATUS_QUEUE_FULL)
1785 dev->tag_success_count++;
1786 /*
1787 * Some devices deal with temporary internal resource
1788 * shortages by returning queue full. When the queue
1789 * full occurrs, we throttle back. Slowly try to get
1790 * back to our previous queue depth.
1791 */
1792 if ((dev->openings + dev->active) < dev->maxtags
1793 && dev->tag_success_count > AHD_TAG_SUCCESS_INTERVAL) {
1794 dev->tag_success_count = 0;
1795 dev->openings++;
1796 }
1797
1798 if (dev->active == 0)
1799 dev->commands_since_idle_or_otag = 0;
1800
1801 if ((scb->flags & SCB_RECOVERY_SCB) != 0) {
1802 printf("Recovery SCB completes\n");
1803 if (ahd_get_transaction_status(scb) == CAM_BDR_SENT
1804 || ahd_get_transaction_status(scb) == CAM_REQ_ABORTED)
1805 ahd_set_transaction_status(scb, CAM_CMD_TIMEOUT);
1806 if ((ahd->platform_data->flags & AHD_SCB_UP_EH_SEM) != 0) {
1807 ahd->platform_data->flags &= ~AHD_SCB_UP_EH_SEM;
1808 up(&ahd->platform_data->eh_sem);
1809 }
1810 }
1811
1812 ahd_free_scb(ahd, scb);
1813 ahd_linux_queue_cmd_complete(ahd, cmd);
1814 }
1815
1816 static void
1817 ahd_linux_handle_scsi_status(struct ahd_softc *ahd,
1818 struct scsi_device *sdev, struct scb *scb)
1819 {
1820 struct ahd_devinfo devinfo;
1821 struct ahd_linux_device *dev = scsi_transport_device_data(sdev);
1822
1823 ahd_compile_devinfo(&devinfo,
1824 ahd->our_id,
1825 sdev->sdev_target->id, sdev->lun,
1826 sdev->sdev_target->channel == 0 ? 'A' : 'B',
1827 ROLE_INITIATOR);
1828
1829 /*
1830 * We don't currently trust the mid-layer to
1831 * properly deal with queue full or busy. So,
1832 * when one occurs, we tell the mid-layer to
1833 * unconditionally requeue the command to us
1834 * so that we can retry it ourselves. We also
1835 * implement our own throttling mechanism so
1836 * we don't clobber the device with too many
1837 * commands.
1838 */
1839 switch (ahd_get_scsi_status(scb)) {
1840 default:
1841 break;
1842 case SCSI_STATUS_CHECK_COND:
1843 case SCSI_STATUS_CMD_TERMINATED:
1844 {
1845 struct scsi_cmnd *cmd;
1846
1847 /*
1848 * Copy sense information to the OS's cmd
1849 * structure if it is available.
1850 */
1851 cmd = scb->io_ctx;
1852 if ((scb->flags & (SCB_SENSE|SCB_PKT_SENSE)) != 0) {
1853 struct scsi_status_iu_header *siu;
1854 u_int sense_size;
1855 u_int sense_offset;
1856
1857 if (scb->flags & SCB_SENSE) {
1858 sense_size = MIN(sizeof(struct scsi_sense_data)
1859 - ahd_get_sense_residual(scb),
1860 sizeof(cmd->sense_buffer));
1861 sense_offset = 0;
1862 } else {
1863 /*
1864 * Copy only the sense data into the provided
1865 * buffer.
1866 */
1867 siu = (struct scsi_status_iu_header *)
1868 scb->sense_data;
1869 sense_size = MIN(scsi_4btoul(siu->sense_length),
1870 sizeof(cmd->sense_buffer));
1871 sense_offset = SIU_SENSE_OFFSET(siu);
1872 }
1873
1874 memset(cmd->sense_buffer, 0, sizeof(cmd->sense_buffer));
1875 memcpy(cmd->sense_buffer,
1876 ahd_get_sense_buf(ahd, scb)
1877 + sense_offset, sense_size);
1878 cmd->result |= (DRIVER_SENSE << 24);
1879
1880 #ifdef AHD_DEBUG
1881 if (ahd_debug & AHD_SHOW_SENSE) {
1882 int i;
1883
1884 printf("Copied %d bytes of sense data at %d:",
1885 sense_size, sense_offset);
1886 for (i = 0; i < sense_size; i++) {
1887 if ((i & 0xF) == 0)
1888 printf("\n");
1889 printf("0x%x ", cmd->sense_buffer[i]);
1890 }
1891 printf("\n");
1892 }
1893 #endif
1894 }
1895 break;
1896 }
1897 case SCSI_STATUS_QUEUE_FULL:
1898 /*
1899 * By the time the core driver has returned this
1900 * command, all other commands that were queued
1901 * to us but not the device have been returned.
1902 * This ensures that dev->active is equal to
1903 * the number of commands actually queued to
1904 * the device.
1905 */
1906 dev->tag_success_count = 0;
1907 if (dev->active != 0) {
1908 /*
1909 * Drop our opening count to the number
1910 * of commands currently outstanding.
1911 */
1912 dev->openings = 0;
1913 #ifdef AHD_DEBUG
1914 if ((ahd_debug & AHD_SHOW_QFULL) != 0) {
1915 ahd_print_path(ahd, scb);
1916 printf("Dropping tag count to %d\n",
1917 dev->active);
1918 }
1919 #endif
1920 if (dev->active == dev->tags_on_last_queuefull) {
1921
1922 dev->last_queuefull_same_count++;
1923 /*
1924 * If we repeatedly see a queue full
1925 * at the same queue depth, this
1926 * device has a fixed number of tag
1927 * slots. Lock in this tag depth
1928 * so we stop seeing queue fulls from
1929 * this device.
1930 */
1931 if (dev->last_queuefull_same_count
1932 == AHD_LOCK_TAGS_COUNT) {
1933 dev->maxtags = dev->active;
1934 ahd_print_path(ahd, scb);
1935 printf("Locking max tag count at %d\n",
1936 dev->active);
1937 }
1938 } else {
1939 dev->tags_on_last_queuefull = dev->active;
1940 dev->last_queuefull_same_count = 0;
1941 }
1942 ahd_set_transaction_status(scb, CAM_REQUEUE_REQ);
1943 ahd_set_scsi_status(scb, SCSI_STATUS_OK);
1944 ahd_platform_set_tags(ahd, &devinfo,
1945 (dev->flags & AHD_DEV_Q_BASIC)
1946 ? AHD_QUEUE_BASIC : AHD_QUEUE_TAGGED);
1947 break;
1948 }
1949 /*
1950 * Drop down to a single opening, and treat this
1951 * as if the target returned BUSY SCSI status.
1952 */
1953 dev->openings = 1;
1954 ahd_platform_set_tags(ahd, &devinfo,
1955 (dev->flags & AHD_DEV_Q_BASIC)
1956 ? AHD_QUEUE_BASIC : AHD_QUEUE_TAGGED);
1957 ahd_set_scsi_status(scb, SCSI_STATUS_BUSY);
1958 }
1959 }
1960
1961 static void
1962 ahd_linux_queue_cmd_complete(struct ahd_softc *ahd, struct scsi_cmnd *cmd)
1963 {
1964 /*
1965 * Map CAM error codes into Linux Error codes. We
1966 * avoid the conversion so that the DV code has the
1967 * full error information available when making
1968 * state change decisions.
1969 */
1970 {
1971 uint32_t status;
1972 u_int new_status;
1973
1974 status = ahd_cmd_get_transaction_status(cmd);
1975 switch (status) {
1976 case CAM_REQ_INPROG:
1977 case CAM_REQ_CMP:
1978 case CAM_SCSI_STATUS_ERROR:
1979 new_status = DID_OK;
1980 break;
1981 case CAM_REQ_ABORTED:
1982 new_status = DID_ABORT;
1983 break;
1984 case CAM_BUSY:
1985 new_status = DID_BUS_BUSY;
1986 break;
1987 case CAM_REQ_INVALID:
1988 case CAM_PATH_INVALID:
1989 new_status = DID_BAD_TARGET;
1990 break;
1991 case CAM_SEL_TIMEOUT:
1992 new_status = DID_NO_CONNECT;
1993 break;
1994 case CAM_SCSI_BUS_RESET:
1995 case CAM_BDR_SENT:
1996 new_status = DID_RESET;
1997 break;
1998 case CAM_UNCOR_PARITY:
1999 new_status = DID_PARITY;
2000 break;
2001 case CAM_CMD_TIMEOUT:
2002 new_status = DID_TIME_OUT;
2003 break;
2004 case CAM_UA_ABORT:
2005 case CAM_REQ_CMP_ERR:
2006 case CAM_AUTOSENSE_FAIL:
2007 case CAM_NO_HBA:
2008 case CAM_DATA_RUN_ERR:
2009 case CAM_UNEXP_BUSFREE:
2010 case CAM_SEQUENCE_FAIL:
2011 case CAM_CCB_LEN_ERR:
2012 case CAM_PROVIDE_FAIL:
2013 case CAM_REQ_TERMIO:
2014 case CAM_UNREC_HBA_ERROR:
2015 case CAM_REQ_TOO_BIG:
2016 new_status = DID_ERROR;
2017 break;
2018 case CAM_REQUEUE_REQ:
2019 new_status = DID_REQUEUE;
2020 break;
2021 default:
2022 /* We should never get here */
2023 new_status = DID_ERROR;
2024 break;
2025 }
2026
2027 ahd_cmd_set_transaction_status(cmd, new_status);
2028 }
2029
2030 cmd->scsi_done(cmd);
2031 }
2032
2033 static void
2034 ahd_linux_sem_timeout(u_long arg)
2035 {
2036 struct ahd_softc *ahd;
2037 u_long s;
2038
2039 ahd = (struct ahd_softc *)arg;
2040
2041 ahd_lock(ahd, &s);
2042 if ((ahd->platform_data->flags & AHD_SCB_UP_EH_SEM) != 0) {
2043 ahd->platform_data->flags &= ~AHD_SCB_UP_EH_SEM;
2044 up(&ahd->platform_data->eh_sem);
2045 }
2046 ahd_unlock(ahd, &s);
2047 }
2048
2049 void
2050 ahd_freeze_simq(struct ahd_softc *ahd)
2051 {
2052 unsigned long s;
2053
2054 ahd_lock(ahd, &s);
2055 ahd->platform_data->qfrozen++;
2056 if (ahd->platform_data->qfrozen == 1) {
2057 scsi_block_requests(ahd->platform_data->host);
2058 ahd_platform_abort_scbs(ahd, CAM_TARGET_WILDCARD, ALL_CHANNELS,
2059 CAM_LUN_WILDCARD, SCB_LIST_NULL,
2060 ROLE_INITIATOR, CAM_REQUEUE_REQ);
2061 }
2062 ahd_unlock(ahd, &s);
2063 }
2064
2065 void
2066 ahd_release_simq(struct ahd_softc *ahd)
2067 {
2068 u_long s;
2069 int unblock_reqs;
2070
2071 unblock_reqs = 0;
2072 ahd_lock(ahd, &s);
2073 if (ahd->platform_data->qfrozen > 0)
2074 ahd->platform_data->qfrozen--;
2075 if (ahd->platform_data->qfrozen == 0) {
2076 unblock_reqs = 1;
2077 }
2078 ahd_unlock(ahd, &s);
2079 /*
2080 * There is still a race here. The mid-layer
2081 * should keep its own freeze count and use
2082 * a bottom half handler to run the queues
2083 * so we can unblock with our own lock held.
2084 */
2085 if (unblock_reqs)
2086 scsi_unblock_requests(ahd->platform_data->host);
2087 }
2088
2089 static int
2090 ahd_linux_queue_recovery_cmd(struct scsi_cmnd *cmd, scb_flag flag)
2091 {
2092 struct ahd_softc *ahd;
2093 struct ahd_linux_device *dev;
2094 struct scb *pending_scb;
2095 u_int saved_scbptr;
2096 u_int active_scbptr;
2097 u_int last_phase;
2098 u_int saved_scsiid;
2099 u_int cdb_byte;
2100 int retval;
2101 int was_paused;
2102 int paused;
2103 int wait;
2104 int disconnected;
2105 int found;
2106 ahd_mode_state saved_modes;
2107 unsigned long flags;
2108
2109 pending_scb = NULL;
2110 paused = FALSE;
2111 wait = FALSE;
2112 ahd = *(struct ahd_softc **)cmd->device->host->hostdata;
2113
2114 scmd_printk(KERN_INFO, cmd,
2115 "Attempting to queue a%s message:",
2116 flag == SCB_ABORT ? "n ABORT" : " TARGET RESET");
2117
2118 printf("CDB:");
2119 for (cdb_byte = 0; cdb_byte < cmd->cmd_len; cdb_byte++)
2120 printf(" 0x%x", cmd->cmnd[cdb_byte]);
2121 printf("\n");
2122
2123 ahd_lock(ahd, &flags);
2124
2125 /*
2126 * First determine if we currently own this command.
2127 * Start by searching the device queue. If not found
2128 * there, check the pending_scb list. If not found
2129 * at all, and the system wanted us to just abort the
2130 * command, return success.
2131 */
2132 dev = scsi_transport_device_data(cmd->device);
2133
2134 if (dev == NULL) {
2135 /*
2136 * No target device for this command exists,
2137 * so we must not still own the command.
2138 */
2139 scmd_printk(KERN_INFO, cmd, "Is not an active device\n");
2140 retval = SUCCESS;
2141 goto no_cmd;
2142 }
2143
2144 /*
2145 * See if we can find a matching cmd in the pending list.
2146 */
2147 LIST_FOREACH(pending_scb, &ahd->pending_scbs, pending_links) {
2148 if (pending_scb->io_ctx == cmd)
2149 break;
2150 }
2151
2152 if (pending_scb == NULL && flag == SCB_DEVICE_RESET) {
2153
2154 /* Any SCB for this device will do for a target reset */
2155 LIST_FOREACH(pending_scb, &ahd->pending_scbs, pending_links) {
2156 if (ahd_match_scb(ahd, pending_scb,
2157 scmd_id(cmd),
2158 scmd_channel(cmd) + 'A',
2159 CAM_LUN_WILDCARD,
2160 SCB_LIST_NULL, ROLE_INITIATOR))
2161 break;
2162 }
2163 }
2164
2165 if (pending_scb == NULL) {
2166 scmd_printk(KERN_INFO, cmd, "Command not found\n");
2167 goto no_cmd;
2168 }
2169
2170 if ((pending_scb->flags & SCB_RECOVERY_SCB) != 0) {
2171 /*
2172 * We can't queue two recovery actions using the same SCB
2173 */
2174 retval = FAILED;
2175 goto done;
2176 }
2177
2178 /*
2179 * Ensure that the card doesn't do anything
2180 * behind our back. Also make sure that we
2181 * didn't "just" miss an interrupt that would
2182 * affect this cmd.
2183 */
2184 was_paused = ahd_is_paused(ahd);
2185 ahd_pause_and_flushwork(ahd);
2186 paused = TRUE;
2187
2188 if ((pending_scb->flags & SCB_ACTIVE) == 0) {
2189 scmd_printk(KERN_INFO, cmd, "Command already completed\n");
2190 goto no_cmd;
2191 }
2192
2193 printf("%s: At time of recovery, card was %spaused\n",
2194 ahd_name(ahd), was_paused ? "" : "not ");
2195 ahd_dump_card_state(ahd);
2196
2197 disconnected = TRUE;
2198 if (flag == SCB_ABORT) {
2199 if (ahd_search_qinfifo(ahd, cmd->device->id,
2200 cmd->device->channel + 'A',
2201 cmd->device->lun,
2202 pending_scb->hscb->tag,
2203 ROLE_INITIATOR, CAM_REQ_ABORTED,
2204 SEARCH_COMPLETE) > 0) {
2205 printf("%s:%d:%d:%d: Cmd aborted from QINFIFO\n",
2206 ahd_name(ahd), cmd->device->channel,
2207 cmd->device->id, cmd->device->lun);
2208 retval = SUCCESS;
2209 goto done;
2210 }
2211 } else if (ahd_search_qinfifo(ahd, cmd->device->id,
2212 cmd->device->channel + 'A',
2213 cmd->device->lun, pending_scb->hscb->tag,
2214 ROLE_INITIATOR, /*status*/0,
2215 SEARCH_COUNT) > 0) {
2216 disconnected = FALSE;
2217 }
2218
2219 saved_modes = ahd_save_modes(ahd);
2220 ahd_set_modes(ahd, AHD_MODE_SCSI, AHD_MODE_SCSI);
2221 last_phase = ahd_inb(ahd, LASTPHASE);
2222 saved_scbptr = ahd_get_scbptr(ahd);
2223 active_scbptr = saved_scbptr;
2224 if (disconnected && ((last_phase != P_BUSFREE) ||
2225 (ahd_inb(ahd, SEQ_FLAGS) & NOT_IDENTIFIED) == 0)) {
2226 struct scb *bus_scb;
2227
2228 bus_scb = ahd_lookup_scb(ahd, active_scbptr);
2229 if (bus_scb == pending_scb)
2230 disconnected = FALSE;
2231 else if (flag != SCB_ABORT
2232 && ahd_inb(ahd, SAVED_SCSIID) == pending_scb->hscb->scsiid
2233 && ahd_inb(ahd, SAVED_LUN) == SCB_GET_LUN(pending_scb))
2234 disconnected = FALSE;
2235 }
2236
2237 /*
2238 * At this point, pending_scb is the scb associated with the
2239 * passed in command. That command is currently active on the
2240 * bus or is in the disconnected state.
2241 */
2242 saved_scsiid = ahd_inb(ahd, SAVED_SCSIID);
2243 if (SCB_GET_TAG(pending_scb) == active_scbptr
2244 || (flag == SCB_DEVICE_RESET
2245 && SCSIID_TARGET(ahd, saved_scsiid) == scmd_id(cmd))) {
2246
2247 /*
2248 * We're active on the bus, so assert ATN
2249 * and hope that the target responds.
2250 */
2251 pending_scb = ahd_lookup_scb(ahd, active_scbptr);
2252 pending_scb->flags |= SCB_RECOVERY_SCB|SCB_DEVICE_RESET;
2253 ahd_outb(ahd, MSG_OUT, HOST_MSG);
2254 ahd_outb(ahd, SCSISIGO, last_phase|ATNO);
2255 scmd_printk(KERN_INFO, cmd, "BDR message in message buffer\n");
2256 wait = TRUE;
2257 } else if (last_phase != P_BUSFREE
2258 && ahd_inb(ahd, SCSIPHASE) == 0) {
2259 /*
2260 * SCB is not identified, there
2261 * is no pending REQ, and the sequencer
2262 * has not seen a busfree. Looks like
2263 * a stuck connection waiting to
2264 * go busfree. Reset the bus.
2265 */
2266 found = ahd_reset_channel(ahd, cmd->device->channel + 'A',
2267 /*Initiate Reset*/TRUE);
2268 printf("%s: Issued Channel %c Bus Reset. "
2269 "%d SCBs aborted\n", ahd_name(ahd),
2270 cmd->device->channel + 'A', found);
2271 } else if (disconnected) {
2272
2273 /*
2274 * Actually re-queue this SCB in an attempt
2275 * to select the device before it reconnects.
2276 */
2277 pending_scb->flags |= SCB_RECOVERY_SCB|flag;
2278 ahd_set_scbptr(ahd, SCB_GET_TAG(pending_scb));
2279 pending_scb->hscb->cdb_len = 0;
2280 pending_scb->hscb->task_attribute = 0;
2281 pending_scb->hscb->task_management = SIU_TASKMGMT_ABORT_TASK;
2282
2283 if ((pending_scb->flags & SCB_PACKETIZED) != 0) {
2284 /*
2285 * Mark the SCB has having an outstanding
2286 * task management function. Should the command
2287 * complete normally before the task management
2288 * function can be sent, the host will be notified
2289 * to abort our requeued SCB.
2290 */
2291 ahd_outb(ahd, SCB_TASK_MANAGEMENT,
2292 pending_scb->hscb->task_management);
2293 } else {
2294 /*
2295 * If non-packetized, set the MK_MESSAGE control
2296 * bit indicating that we desire to send a message.
2297 * We also set the disconnected flag since there is
2298 * no guarantee that our SCB control byte matches
2299 * the version on the card. We don't want the
2300 * sequencer to abort the command thinking an
2301 * unsolicited reselection occurred.
2302 */
2303 pending_scb->hscb->control |= MK_MESSAGE|DISCONNECTED;
2304
2305 /*
2306 * The sequencer will never re-reference the
2307 * in-core SCB. To make sure we are notified
2308 * during reslection, set the MK_MESSAGE flag in
2309 * the card's copy of the SCB.
2310 */
2311 ahd_outb(ahd, SCB_CONTROL,
2312 ahd_inb(ahd, SCB_CONTROL)|MK_MESSAGE);
2313 }
2314
2315 /*
2316 * Clear out any entries in the QINFIFO first
2317 * so we are the next SCB for this target
2318 * to run.
2319 */
2320 ahd_search_qinfifo(ahd, cmd->device->id,
2321 cmd->device->channel + 'A', cmd->device->lun,
2322 SCB_LIST_NULL, ROLE_INITIATOR,
2323 CAM_REQUEUE_REQ, SEARCH_COMPLETE);
2324 ahd_qinfifo_requeue_tail(ahd, pending_scb);
2325 ahd_set_scbptr(ahd, saved_scbptr);
2326 ahd_print_path(ahd, pending_scb);
2327 printf("Device is disconnected, re-queuing SCB\n");
2328 wait = TRUE;
2329 } else {
2330 scmd_printk(KERN_INFO, cmd, "Unable to deliver message\n");
2331 retval = FAILED;
2332 goto done;
2333 }
2334
2335 no_cmd:
2336 /*
2337 * Our assumption is that if we don't have the command, no
2338 * recovery action was required, so we return success. Again,
2339 * the semantics of the mid-layer recovery engine are not
2340 * well defined, so this may change in time.
2341 */
2342 retval = SUCCESS;
2343 done:
2344 if (paused)
2345 ahd_unpause(ahd);
2346 if (wait) {
2347 struct timer_list timer;
2348 int ret;
2349
2350 ahd->platform_data->flags |= AHD_SCB_UP_EH_SEM;
2351 ahd_unlock(ahd, &flags);
2352
2353 init_timer(&timer);
2354 timer.data = (u_long)ahd;
2355 timer.expires = jiffies + (5 * HZ);
2356 timer.function = ahd_linux_sem_timeout;
2357 add_timer(&timer);
2358 printf("%s: Recovery code sleeping\n", ahd_name(ahd));
2359 down(&ahd->platform_data->eh_sem);
2360 printf("%s: Recovery code awake\n", ahd_name(ahd));
2361 ret = del_timer_sync(&timer);
2362 if (ret == 0) {
2363 printf("%s: Timer Expired (active %d)\n",
2364 ahd_name(ahd), dev->active);
2365 retval = FAILED;
2366 }
2367 } else
2368 ahd_unlock(ahd, &flags);
2369
2370 return (retval);
2371 }
2372
2373 static void ahd_linux_set_width(struct scsi_target *starget, int width)
2374 {
2375 struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
2376 struct ahd_softc *ahd = *((struct ahd_softc **)shost->hostdata);
2377 struct ahd_devinfo devinfo;
2378 unsigned long flags;
2379
2380 ahd_compile_devinfo(&devinfo, shost->this_id, starget->id, 0,
2381 starget->channel + 'A', ROLE_INITIATOR);
2382 ahd_lock(ahd, &flags);
2383 ahd_set_width(ahd, &devinfo, width, AHD_TRANS_GOAL, FALSE);
2384 ahd_unlock(ahd, &flags);
2385 }
2386
2387 static void ahd_linux_set_period(struct scsi_target *starget, int period)
2388 {
2389 struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
2390 struct ahd_softc *ahd = *((struct ahd_softc **)shost->hostdata);
2391 struct ahd_tmode_tstate *tstate;
2392 struct ahd_initiator_tinfo *tinfo
2393 = ahd_fetch_transinfo(ahd,
2394 starget->channel + 'A',
2395 shost->this_id, starget->id, &tstate);
2396 struct ahd_devinfo devinfo;
2397 unsigned int ppr_options = tinfo->goal.ppr_options;
2398 unsigned int dt;
2399 unsigned long flags;
2400 unsigned long offset = tinfo->goal.offset;
2401
2402 #ifdef AHD_DEBUG
2403 if ((ahd_debug & AHD_SHOW_DV) != 0)
2404 printf("%s: set period to %d\n", ahd_name(ahd), period);
2405 #endif
2406 if (offset == 0)
2407 offset = MAX_OFFSET;
2408
2409 if (period < 8)
2410 period = 8;
2411 if (period < 10) {
2412 ppr_options |= MSG_EXT_PPR_DT_REQ;
2413 if (period == 8)
2414 ppr_options |= MSG_EXT_PPR_IU_REQ;
2415 }
2416
2417 dt = ppr_options & MSG_EXT_PPR_DT_REQ;
2418
2419 ahd_compile_devinfo(&devinfo, shost->this_id, starget->id, 0,
2420 starget->channel + 'A', ROLE_INITIATOR);
2421
2422 /* all PPR requests apart from QAS require wide transfers */
2423 if (ppr_options & ~MSG_EXT_PPR_QAS_REQ) {
2424 if (spi_width(starget) == 0)
2425 ppr_options &= MSG_EXT_PPR_QAS_REQ;
2426 }
2427
2428 ahd_find_syncrate(ahd, &period, &ppr_options,
2429 dt ? AHD_SYNCRATE_MAX : AHD_SYNCRATE_ULTRA2);
2430
2431 ahd_lock(ahd, &flags);
2432 ahd_set_syncrate(ahd, &devinfo, period, offset,
2433 ppr_options, AHD_TRANS_GOAL, FALSE);
2434 ahd_unlock(ahd, &flags);
2435 }
2436
2437 static void ahd_linux_set_offset(struct scsi_target *starget, int offset)
2438 {
2439 struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
2440 struct ahd_softc *ahd = *((struct ahd_softc **)shost->hostdata);
2441 struct ahd_tmode_tstate *tstate;
2442 struct ahd_initiator_tinfo *tinfo
2443 = ahd_fetch_transinfo(ahd,
2444 starget->channel + 'A',
2445 shost->this_id, starget->id, &tstate);
2446 struct ahd_devinfo devinfo;
2447 unsigned int ppr_options = 0;
2448 unsigned int period = 0;
2449 unsigned int dt = ppr_options & MSG_EXT_PPR_DT_REQ;
2450 unsigned long flags;
2451
2452 #ifdef AHD_DEBUG
2453 if ((ahd_debug & AHD_SHOW_DV) != 0)
2454 printf("%s: set offset to %d\n", ahd_name(ahd), offset);
2455 #endif
2456
2457 ahd_compile_devinfo(&devinfo, shost->this_id, starget->id, 0,
2458 starget->channel + 'A', ROLE_INITIATOR);
2459 if (offset != 0) {
2460 period = tinfo->goal.period;
2461 ppr_options = tinfo->goal.ppr_options;
2462 ahd_find_syncrate(ahd, &period, &ppr_options,
2463 dt ? AHD_SYNCRATE_MAX : AHD_SYNCRATE_ULTRA2);
2464 }
2465
2466 ahd_lock(ahd, &flags);
2467 ahd_set_syncrate(ahd, &devinfo, period, offset, ppr_options,
2468 AHD_TRANS_GOAL, FALSE);
2469 ahd_unlock(ahd, &flags);
2470 }
2471
2472 static void ahd_linux_set_dt(struct scsi_target *starget, int dt)
2473 {
2474 struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
2475 struct ahd_softc *ahd = *((struct ahd_softc **)shost->hostdata);
2476 struct ahd_tmode_tstate *tstate;
2477 struct ahd_initiator_tinfo *tinfo
2478 = ahd_fetch_transinfo(ahd,
2479 starget->channel + 'A',
2480 shost->this_id, starget->id, &tstate);
2481 struct ahd_devinfo devinfo;
2482 unsigned int ppr_options = tinfo->goal.ppr_options
2483 & ~MSG_EXT_PPR_DT_REQ;
2484 unsigned int period = tinfo->goal.period;
2485 unsigned int width = tinfo->goal.width;
2486 unsigned long flags;
2487
2488 #ifdef AHD_DEBUG
2489 if ((ahd_debug & AHD_SHOW_DV) != 0)
2490 printf("%s: %s DT\n", ahd_name(ahd),
2491 dt ? "enabling" : "disabling");
2492 #endif
2493 if (dt) {
2494 ppr_options |= MSG_EXT_PPR_DT_REQ;
2495 if (!width)
2496 ahd_linux_set_width(starget, 1);
2497 } else {
2498 if (period <= 9)
2499 period = 10; /* If resetting DT, period must be >= 25ns */
2500 /* IU is invalid without DT set */
2501 ppr_options &= ~MSG_EXT_PPR_IU_REQ;
2502 }
2503 ahd_compile_devinfo(&devinfo, shost->this_id, starget->id, 0,
2504 starget->channel + 'A', ROLE_INITIATOR);
2505 ahd_find_syncrate(ahd, &period, &ppr_options,
2506 dt ? AHD_SYNCRATE_MAX : AHD_SYNCRATE_ULTRA2);
2507
2508 ahd_lock(ahd, &flags);
2509 ahd_set_syncrate(ahd, &devinfo, period, tinfo->goal.offset,
2510 ppr_options, AHD_TRANS_GOAL, FALSE);
2511 ahd_unlock(ahd, &flags);
2512 }
2513
2514 static void ahd_linux_set_qas(struct scsi_target *starget, int qas)
2515 {
2516 struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
2517 struct ahd_softc *ahd = *((struct ahd_softc **)shost->hostdata);
2518 struct ahd_tmode_tstate *tstate;
2519 struct ahd_initiator_tinfo *tinfo
2520 = ahd_fetch_transinfo(ahd,
2521 starget->channel + 'A',
2522 shost->this_id, starget->id, &tstate);
2523 struct ahd_devinfo devinfo;
2524 unsigned int ppr_options = tinfo->goal.ppr_options
2525 & ~MSG_EXT_PPR_QAS_REQ;
2526 unsigned int period = tinfo->goal.period;
2527 unsigned int dt;
2528 unsigned long flags;
2529
2530 #ifdef AHD_DEBUG
2531 if ((ahd_debug & AHD_SHOW_DV) != 0)
2532 printf("%s: %s QAS\n", ahd_name(ahd),
2533 qas ? "enabling" : "disabling");
2534 #endif
2535
2536 if (qas) {
2537 ppr_options |= MSG_EXT_PPR_QAS_REQ;
2538 }
2539
2540 dt = ppr_options & MSG_EXT_PPR_DT_REQ;
2541
2542 ahd_compile_devinfo(&devinfo, shost->this_id, starget->id, 0,
2543 starget->channel + 'A', ROLE_INITIATOR);
2544 ahd_find_syncrate(ahd, &period, &ppr_options,
2545 dt ? AHD_SYNCRATE_MAX : AHD_SYNCRATE_ULTRA2);
2546
2547 ahd_lock(ahd, &flags);
2548 ahd_set_syncrate(ahd, &devinfo, period, tinfo->goal.offset,
2549 ppr_options, AHD_TRANS_GOAL, FALSE);
2550 ahd_unlock(ahd, &flags);
2551 }
2552
2553 static void ahd_linux_set_iu(struct scsi_target *starget, int iu)
2554 {
2555 struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
2556 struct ahd_softc *ahd = *((struct ahd_softc **)shost->hostdata);
2557 struct ahd_tmode_tstate *tstate;
2558 struct ahd_initiator_tinfo *tinfo
2559 = ahd_fetch_transinfo(ahd,
2560 starget->channel + 'A',
2561 shost->this_id, starget->id, &tstate);
2562 struct ahd_devinfo devinfo;
2563 unsigned int ppr_options = tinfo->goal.ppr_options
2564 & ~MSG_EXT_PPR_IU_REQ;
2565 unsigned int period = tinfo->goal.period;
2566 unsigned int dt;
2567 unsigned long flags;
2568
2569 #ifdef AHD_DEBUG
2570 if ((ahd_debug & AHD_SHOW_DV) != 0)
2571 printf("%s: %s IU\n", ahd_name(ahd),
2572 iu ? "enabling" : "disabling");
2573 #endif
2574
2575 if (iu) {
2576 ppr_options |= MSG_EXT_PPR_IU_REQ;
2577 ppr_options |= MSG_EXT_PPR_DT_REQ; /* IU requires DT */
2578 }
2579
2580 dt = ppr_options & MSG_EXT_PPR_DT_REQ;
2581
2582 ahd_compile_devinfo(&devinfo, shost->this_id, starget->id, 0,
2583 starget->channel + 'A', ROLE_INITIATOR);
2584 ahd_find_syncrate(ahd, &period, &ppr_options,
2585 dt ? AHD_SYNCRATE_MAX : AHD_SYNCRATE_ULTRA2);
2586
2587 ahd_lock(ahd, &flags);
2588 ahd_set_syncrate(ahd, &devinfo, period, tinfo->goal.offset,
2589 ppr_options, AHD_TRANS_GOAL, FALSE);
2590 ahd_unlock(ahd, &flags);
2591 }
2592
2593 static void ahd_linux_set_rd_strm(struct scsi_target *starget, int rdstrm)
2594 {
2595 struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
2596 struct ahd_softc *ahd = *((struct ahd_softc **)shost->hostdata);
2597 struct ahd_tmode_tstate *tstate;
2598 struct ahd_initiator_tinfo *tinfo
2599 = ahd_fetch_transinfo(ahd,
2600 starget->channel + 'A',
2601 shost->this_id, starget->id, &tstate);
2602 struct ahd_devinfo devinfo;
2603 unsigned int ppr_options = tinfo->goal.ppr_options
2604 & ~MSG_EXT_PPR_RD_STRM;
2605 unsigned int period = tinfo->goal.period;
2606 unsigned int dt = ppr_options & MSG_EXT_PPR_DT_REQ;
2607 unsigned long flags;
2608
2609 #ifdef AHD_DEBUG
2610 if ((ahd_debug & AHD_SHOW_DV) != 0)
2611 printf("%s: %s Read Streaming\n", ahd_name(ahd),
2612 rdstrm ? "enabling" : "disabling");
2613 #endif
2614
2615 if (rdstrm)
2616 ppr_options |= MSG_EXT_PPR_RD_STRM;
2617
2618 ahd_compile_devinfo(&devinfo, shost->this_id, starget->id, 0,
2619 starget->channel + 'A', ROLE_INITIATOR);
2620 ahd_find_syncrate(ahd, &period, &ppr_options,
2621 dt ? AHD_SYNCRATE_MAX : AHD_SYNCRATE_ULTRA2);
2622
2623 ahd_lock(ahd, &flags);
2624 ahd_set_syncrate(ahd, &devinfo, period, tinfo->goal.offset,
2625 ppr_options, AHD_TRANS_GOAL, FALSE);
2626 ahd_unlock(ahd, &flags);
2627 }
2628
2629 static void ahd_linux_set_wr_flow(struct scsi_target *starget, int wrflow)
2630 {
2631 struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
2632 struct ahd_softc *ahd = *((struct ahd_softc **)shost->hostdata);
2633 struct ahd_tmode_tstate *tstate;
2634 struct ahd_initiator_tinfo *tinfo
2635 = ahd_fetch_transinfo(ahd,
2636 starget->channel + 'A',
2637 shost->this_id, starget->id, &tstate);
2638 struct ahd_devinfo devinfo;
2639 unsigned int ppr_options = tinfo->goal.ppr_options
2640 & ~MSG_EXT_PPR_WR_FLOW;
2641 unsigned int period = tinfo->goal.period;
2642 unsigned int dt = ppr_options & MSG_EXT_PPR_DT_REQ;
2643 unsigned long flags;
2644
2645 #ifdef AHD_DEBUG
2646 if ((ahd_debug & AHD_SHOW_DV) != 0)
2647 printf("%s: %s Write Flow Control\n", ahd_name(ahd),
2648 wrflow ? "enabling" : "disabling");
2649 #endif
2650
2651 if (wrflow)
2652 ppr_options |= MSG_EXT_PPR_WR_FLOW;
2653
2654 ahd_compile_devinfo(&devinfo, shost->this_id, starget->id, 0,
2655 starget->channel + 'A', ROLE_INITIATOR);
2656 ahd_find_syncrate(ahd, &period, &ppr_options,
2657 dt ? AHD_SYNCRATE_MAX : AHD_SYNCRATE_ULTRA2);
2658
2659 ahd_lock(ahd, &flags);
2660 ahd_set_syncrate(ahd, &devinfo, period, tinfo->goal.offset,
2661 ppr_options, AHD_TRANS_GOAL, FALSE);
2662 ahd_unlock(ahd, &flags);
2663 }
2664
2665 static void ahd_linux_set_rti(struct scsi_target *starget, int rti)
2666 {
2667 struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
2668 struct ahd_softc *ahd = *((struct ahd_softc **)shost->hostdata);
2669 struct ahd_tmode_tstate *tstate;
2670 struct ahd_initiator_tinfo *tinfo
2671 = ahd_fetch_transinfo(ahd,
2672 starget->channel + 'A',
2673 shost->this_id, starget->id, &tstate);
2674 struct ahd_devinfo devinfo;
2675 unsigned int ppr_options = tinfo->goal.ppr_options
2676 & ~MSG_EXT_PPR_RTI;
2677 unsigned int period = tinfo->goal.period;
2678 unsigned int dt = ppr_options & MSG_EXT_PPR_DT_REQ;
2679 unsigned long flags;
2680
2681 if ((ahd->features & AHD_RTI) == 0) {
2682 #ifdef AHD_DEBUG
2683 if ((ahd_debug & AHD_SHOW_DV) != 0)
2684 printf("%s: RTI not available\n", ahd_name(ahd));
2685 #endif
2686 return;
2687 }
2688
2689 #ifdef AHD_DEBUG
2690 if ((ahd_debug & AHD_SHOW_DV) != 0)
2691 printf("%s: %s RTI\n", ahd_name(ahd),
2692 rti ? "enabling" : "disabling");
2693 #endif
2694
2695 if (rti)
2696 ppr_options |= MSG_EXT_PPR_RTI;
2697
2698 ahd_compile_devinfo(&devinfo, shost->this_id, starget->id, 0,
2699 starget->channel + 'A', ROLE_INITIATOR);
2700 ahd_find_syncrate(ahd, &period, &ppr_options,
2701 dt ? AHD_SYNCRATE_MAX : AHD_SYNCRATE_ULTRA2);
2702
2703 ahd_lock(ahd, &flags);
2704 ahd_set_syncrate(ahd, &devinfo, period, tinfo->goal.offset,
2705 ppr_options, AHD_TRANS_GOAL, FALSE);
2706 ahd_unlock(ahd, &flags);
2707 }
2708
2709 static void ahd_linux_set_pcomp_en(struct scsi_target *starget, int pcomp)
2710 {
2711 struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
2712 struct ahd_softc *ahd = *((struct ahd_softc **)shost->hostdata);
2713 struct ahd_tmode_tstate *tstate;
2714 struct ahd_initiator_tinfo *tinfo
2715 = ahd_fetch_transinfo(ahd,
2716 starget->channel + 'A',
2717 shost->this_id, starget->id, &tstate);
2718 struct ahd_devinfo devinfo;
2719 unsigned int ppr_options = tinfo->goal.ppr_options
2720 & ~MSG_EXT_PPR_PCOMP_EN;
2721 unsigned int period = tinfo->goal.period;
2722 unsigned int dt = ppr_options & MSG_EXT_PPR_DT_REQ;
2723 unsigned long flags;
2724
2725 #ifdef AHD_DEBUG
2726 if ((ahd_debug & AHD_SHOW_DV) != 0)
2727 printf("%s: %s Precompensation\n", ahd_name(ahd),
2728 pcomp ? "Enable" : "Disable");
2729 #endif
2730
2731 if (pcomp)
2732 ppr_options |= MSG_EXT_PPR_PCOMP_EN;
2733
2734 ahd_compile_devinfo(&devinfo, shost->this_id, starget->id, 0,
2735 starget->channel + 'A', ROLE_INITIATOR);
2736 ahd_find_syncrate(ahd, &period, &ppr_options,
2737 dt ? AHD_SYNCRATE_MAX : AHD_SYNCRATE_ULTRA2);
2738
2739 ahd_lock(ahd, &flags);
2740 ahd_set_syncrate(ahd, &devinfo, period, tinfo->goal.offset,
2741 ppr_options, AHD_TRANS_GOAL, FALSE);
2742 ahd_unlock(ahd, &flags);
2743 }
2744
2745 static void ahd_linux_set_hold_mcs(struct scsi_target *starget, int hold)
2746 {
2747 struct Scsi_Host *shost = dev_to_shost(starget->dev.parent);
2748 struct ahd_softc *ahd = *((struct ahd_softc **)shost->hostdata);
2749 struct ahd_tmode_tstate *tstate;
2750 struct ahd_initiator_tinfo *tinfo
2751 = ahd_fetch_transinfo(ahd,
2752 starget->channel + 'A',
2753 shost->this_id, starget->id, &tstate);
2754 struct ahd_devinfo devinfo;
2755 unsigned int ppr_options = tinfo->goal.ppr_options
2756 & ~MSG_EXT_PPR_HOLD_MCS;
2757 unsigned int period = tinfo->goal.period;
2758 unsigned int dt = ppr_options & MSG_EXT_PPR_DT_REQ;
2759 unsigned long flags;
2760
2761 if (hold)
2762 ppr_options |= MSG_EXT_PPR_HOLD_MCS;
2763
2764 ahd_compile_devinfo(&devinfo, shost->this_id, starget->id, 0,
2765 starget->channel + 'A', ROLE_INITIATOR);
2766 ahd_find_syncrate(ahd, &period, &ppr_options,
2767 dt ? AHD_SYNCRATE_MAX : AHD_SYNCRATE_ULTRA2);
2768
2769 ahd_lock(ahd, &flags);
2770 ahd_set_syncrate(ahd, &devinfo, period, tinfo->goal.offset,
2771 ppr_options, AHD_TRANS_GOAL, FALSE);
2772 ahd_unlock(ahd, &flags);
2773 }
2774
2775
2776
2777 static struct spi_function_template ahd_linux_transport_functions = {
2778 .set_offset = ahd_linux_set_offset,
2779 .show_offset = 1,
2780 .set_period = ahd_linux_set_period,
2781 .show_period = 1,
2782 .set_width = ahd_linux_set_width,
2783 .show_width = 1,
2784 .set_dt = ahd_linux_set_dt,
2785 .show_dt = 1,
2786 .set_iu = ahd_linux_set_iu,
2787 .show_iu = 1,
2788 .set_qas = ahd_linux_set_qas,
2789 .show_qas = 1,
2790 .set_rd_strm = ahd_linux_set_rd_strm,
2791 .show_rd_strm = 1,
2792 .set_wr_flow = ahd_linux_set_wr_flow,
2793 .show_wr_flow = 1,
2794 .set_rti = ahd_linux_set_rti,
2795 .show_rti = 1,
2796 .set_pcomp_en = ahd_linux_set_pcomp_en,
2797 .show_pcomp_en = 1,
2798 .set_hold_mcs = ahd_linux_set_hold_mcs,
2799 .show_hold_mcs = 1,
2800 };
2801
2802 static int __init
2803 ahd_linux_init(void)
2804 {
2805 int error = 0;
2806
2807 /*
2808 * If we've been passed any parameters, process them now.
2809 */
2810 if (aic79xx)
2811 aic79xx_setup(aic79xx);
2812
2813 ahd_linux_transport_template =
2814 spi_attach_transport(&ahd_linux_transport_functions);
2815 if (!ahd_linux_transport_template)
2816 return -ENODEV;
2817
2818 scsi_transport_reserve_target(ahd_linux_transport_template,
2819 sizeof(struct ahd_linux_target));
2820 scsi_transport_reserve_device(ahd_linux_transport_template,
2821 sizeof(struct ahd_linux_device));
2822
2823 error = ahd_linux_pci_init();
2824 if (error)
2825 spi_release_transport(ahd_linux_transport_template);
2826 return error;
2827 }
2828
2829 static void __exit
2830 ahd_linux_exit(void)
2831 {
2832 ahd_linux_pci_exit();
2833 spi_release_transport(ahd_linux_transport_template);
2834 }
2835
2836 module_init(ahd_linux_init);
2837 module_exit(ahd_linux_exit);