]> bbs.cooldavid.org Git - net-next-2.6.git/blame - drivers/scsi/53c700.c
[SCSI] hptiop: don't use cmnd->bufflen
[net-next-2.6.git] / drivers / scsi / 53c700.c
CommitLineData
1da177e4
LT
1/* -*- mode: c; c-basic-offset: 8 -*- */
2
3/* NCR (or Symbios) 53c700 and 53c700-66 Driver
4 *
5 * Copyright (C) 2001 by James.Bottomley@HansenPartnership.com
6**-----------------------------------------------------------------------------
7**
8** This program is free software; you can redistribute it and/or modify
9** it under the terms of the GNU General Public License as published by
10** the Free Software Foundation; either version 2 of the License, or
11** (at your option) any later version.
12**
13** This program is distributed in the hope that it will be useful,
14** but WITHOUT ANY WARRANTY; without even the implied warranty of
15** MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
16** GNU General Public License for more details.
17**
18** You should have received a copy of the GNU General Public License
19** along with this program; if not, write to the Free Software
20** Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
21**
22**-----------------------------------------------------------------------------
23 */
24
25/* Notes:
26 *
27 * This driver is designed exclusively for these chips (virtually the
28 * earliest of the scripts engine chips). They need their own drivers
29 * because they are missing so many of the scripts and snazzy register
30 * features of their elder brothers (the 710, 720 and 770).
31 *
32 * The 700 is the lowliest of the line, it can only do async SCSI.
33 * The 700-66 can at least do synchronous SCSI up to 10MHz.
34 *
35 * The 700 chip has no host bus interface logic of its own. However,
36 * it is usually mapped to a location with well defined register
37 * offsets. Therefore, if you can determine the base address and the
38 * irq your board incorporating this chip uses, you can probably use
39 * this driver to run it (although you'll probably have to write a
40 * minimal wrapper for the purpose---see the NCR_D700 driver for
41 * details about how to do this).
42 *
43 *
44 * TODO List:
45 *
46 * 1. Better statistics in the proc fs
47 *
48 * 2. Implement message queue (queues SCSI messages like commands) and make
49 * the abort and device reset functions use them.
50 * */
51
52/* CHANGELOG
53 *
54 * Version 2.8
55 *
56 * Fixed bad bug affecting tag starvation processing (previously the
57 * driver would hang the system if too many tags starved. Also fixed
58 * bad bug having to do with 10 byte command processing and REQUEST
59 * SENSE (the command would loop forever getting a transfer length
60 * mismatch in the CMD phase).
61 *
62 * Version 2.7
63 *
64 * Fixed scripts problem which caused certain devices (notably CDRWs)
65 * to hang on initial INQUIRY. Updated NCR_700_readl/writel to use
66 * __raw_readl/writel for parisc compatibility (Thomas
67 * Bogendoerfer). Added missing SCp->request_bufflen initialisation
68 * for sense requests (Ryan Bradetich).
69 *
70 * Version 2.6
71 *
72 * Following test of the 64 bit parisc kernel by Richard Hirst,
73 * several problems have now been corrected. Also adds support for
74 * consistent memory allocation.
75 *
76 * Version 2.5
77 *
78 * More Compatibility changes for 710 (now actually works). Enhanced
79 * support for odd clock speeds which constrain SDTR negotiations.
80 * correct cacheline separation for scsi messages and status for
81 * incoherent architectures. Use of the pci mapping functions on
82 * buffers to begin support for 64 bit drivers.
83 *
84 * Version 2.4
85 *
86 * Added support for the 53c710 chip (in 53c700 emulation mode only---no
87 * special 53c710 instructions or registers are used).
88 *
89 * Version 2.3
90 *
91 * More endianness/cache coherency changes.
92 *
93 * Better bad device handling (handles devices lying about tag
94 * queueing support and devices which fail to provide sense data on
95 * contingent allegiance conditions)
96 *
97 * Many thanks to Richard Hirst <rhirst@linuxcare.com> for patiently
98 * debugging this driver on the parisc architecture and suggesting
99 * many improvements and bug fixes.
100 *
101 * Thanks also go to Linuxcare Inc. for providing several PARISC
102 * machines for me to debug the driver on.
103 *
104 * Version 2.2
105 *
106 * Made the driver mem or io mapped; added endian invariance; added
107 * dma cache flushing operations for architectures which need it;
108 * added support for more varied clocking speeds.
109 *
110 * Version 2.1
111 *
112 * Initial modularisation from the D700. See NCR_D700.c for the rest of
113 * the changelog.
114 * */
115#define NCR_700_VERSION "2.8"
116
117#include <linux/config.h>
118#include <linux/kernel.h>
119#include <linux/types.h>
120#include <linux/string.h>
121#include <linux/ioport.h>
122#include <linux/delay.h>
123#include <linux/spinlock.h>
124#include <linux/completion.h>
125#include <linux/sched.h>
126#include <linux/init.h>
127#include <linux/proc_fs.h>
128#include <linux/blkdev.h>
129#include <linux/module.h>
130#include <linux/interrupt.h>
017560fc 131#include <linux/device.h>
1da177e4
LT
132#include <asm/dma.h>
133#include <asm/system.h>
134#include <asm/io.h>
135#include <asm/pgtable.h>
136#include <asm/byteorder.h>
137
138#include <scsi/scsi.h>
139#include <scsi/scsi_cmnd.h>
140#include <scsi/scsi_dbg.h>
141#include <scsi/scsi_eh.h>
142#include <scsi/scsi_host.h>
143#include <scsi/scsi_tcq.h>
144#include <scsi/scsi_transport.h>
145#include <scsi/scsi_transport_spi.h>
146
147#include "53c700.h"
148
149/* NOTE: For 64 bit drivers there are points in the code where we use
150 * a non dereferenceable pointer to point to a structure in dma-able
151 * memory (which is 32 bits) so that we can use all of the structure
152 * operations but take the address at the end. This macro allows us
153 * to truncate the 64 bit pointer down to 32 bits without the compiler
154 * complaining */
155#define to32bit(x) ((__u32)((unsigned long)(x)))
156
157#ifdef NCR_700_DEBUG
158#define STATIC
159#else
160#define STATIC static
161#endif
162
163MODULE_AUTHOR("James Bottomley");
164MODULE_DESCRIPTION("53c700 and 53c700-66 Driver");
165MODULE_LICENSE("GPL");
166
167/* This is the script */
168#include "53c700_d.h"
169
170
171STATIC int NCR_700_queuecommand(struct scsi_cmnd *, void (*done)(struct scsi_cmnd *));
172STATIC int NCR_700_abort(struct scsi_cmnd * SCpnt);
173STATIC int NCR_700_bus_reset(struct scsi_cmnd * SCpnt);
1da177e4
LT
174STATIC int NCR_700_host_reset(struct scsi_cmnd * SCpnt);
175STATIC void NCR_700_chip_setup(struct Scsi_Host *host);
176STATIC void NCR_700_chip_reset(struct Scsi_Host *host);
177STATIC int NCR_700_slave_configure(struct scsi_device *SDpnt);
178STATIC void NCR_700_slave_destroy(struct scsi_device *SDpnt);
179static int NCR_700_change_queue_depth(struct scsi_device *SDpnt, int depth);
180static int NCR_700_change_queue_type(struct scsi_device *SDpnt, int depth);
181
182STATIC struct device_attribute *NCR_700_dev_attrs[];
183
184STATIC struct scsi_transport_template *NCR_700_transport_template = NULL;
185
186static char *NCR_700_phase[] = {
187 "",
188 "after selection",
189 "before command phase",
190 "after command phase",
191 "after status phase",
192 "after data in phase",
193 "after data out phase",
194 "during data phase",
195};
196
197static char *NCR_700_condition[] = {
198 "",
199 "NOT MSG_OUT",
200 "UNEXPECTED PHASE",
201 "NOT MSG_IN",
202 "UNEXPECTED MSG",
203 "MSG_IN",
204 "SDTR_MSG RECEIVED",
205 "REJECT_MSG RECEIVED",
206 "DISCONNECT_MSG RECEIVED",
207 "MSG_OUT",
208 "DATA_IN",
209
210};
211
212static char *NCR_700_fatal_messages[] = {
213 "unexpected message after reselection",
214 "still MSG_OUT after message injection",
215 "not MSG_IN after selection",
216 "Illegal message length received",
217};
218
219static char *NCR_700_SBCL_bits[] = {
220 "IO ",
221 "CD ",
222 "MSG ",
223 "ATN ",
224 "SEL ",
225 "BSY ",
226 "ACK ",
227 "REQ ",
228};
229
230static char *NCR_700_SBCL_to_phase[] = {
231 "DATA_OUT",
232 "DATA_IN",
233 "CMD_OUT",
234 "STATE",
235 "ILLEGAL PHASE",
236 "ILLEGAL PHASE",
237 "MSG OUT",
238 "MSG IN",
239};
240
1da177e4
LT
241/* This translates the SDTR message offset and period to a value
242 * which can be loaded into the SXFER_REG.
243 *
244 * NOTE: According to SCSI-2, the true transfer period (in ns) is
245 * actually four times this period value */
246static inline __u8
247NCR_700_offset_period_to_sxfer(struct NCR_700_Host_Parameters *hostdata,
248 __u8 offset, __u8 period)
249{
250 int XFERP;
251
252 __u8 min_xferp = (hostdata->chip710
253 ? NCR_710_MIN_XFERP : NCR_700_MIN_XFERP);
254 __u8 max_offset = (hostdata->chip710
255 ? NCR_710_MAX_OFFSET : NCR_700_MAX_OFFSET);
256
257 if(offset == 0)
258 return 0;
259
260 if(period < hostdata->min_period) {
6ea3c0b2 261 printk(KERN_WARNING "53c700: Period %dns is less than this chip's minimum, setting to %d\n", period*4, NCR_700_MIN_PERIOD*4);
1da177e4
LT
262 period = hostdata->min_period;
263 }
264 XFERP = (period*4 * hostdata->sync_clock)/1000 - 4;
265 if(offset > max_offset) {
266 printk(KERN_WARNING "53c700: Offset %d exceeds chip maximum, setting to %d\n",
267 offset, max_offset);
268 offset = max_offset;
269 }
270 if(XFERP < min_xferp) {
271 printk(KERN_WARNING "53c700: XFERP %d is less than minium, setting to %d\n",
272 XFERP, min_xferp);
273 XFERP = min_xferp;
274 }
275 return (offset & 0x0f) | (XFERP & 0x07)<<4;
276}
277
278static inline __u8
279NCR_700_get_SXFER(struct scsi_device *SDp)
280{
281 struct NCR_700_Host_Parameters *hostdata =
282 (struct NCR_700_Host_Parameters *)SDp->host->hostdata[0];
283
284 return NCR_700_offset_period_to_sxfer(hostdata,
285 spi_offset(SDp->sdev_target),
286 spi_period(SDp->sdev_target));
287}
288
289struct Scsi_Host *
290NCR_700_detect(struct scsi_host_template *tpnt,
291 struct NCR_700_Host_Parameters *hostdata, struct device *dev)
292{
293 dma_addr_t pScript, pSlots;
294 __u8 *memory;
295 __u32 *script;
296 struct Scsi_Host *host;
297 static int banner = 0;
298 int j;
299
300 if(tpnt->sdev_attrs == NULL)
301 tpnt->sdev_attrs = NCR_700_dev_attrs;
302
303 memory = dma_alloc_noncoherent(hostdata->dev, TOTAL_MEM_SIZE,
304 &pScript, GFP_KERNEL);
305 if(memory == NULL) {
306 printk(KERN_ERR "53c700: Failed to allocate memory for driver, detatching\n");
307 return NULL;
308 }
309
310 script = (__u32 *)memory;
311 hostdata->msgin = memory + MSGIN_OFFSET;
312 hostdata->msgout = memory + MSGOUT_OFFSET;
313 hostdata->status = memory + STATUS_OFFSET;
314 /* all of these offsets are L1_CACHE_BYTES separated. It is fatal
315 * if this isn't sufficient separation to avoid dma flushing issues */
316 BUG_ON(!dma_is_consistent(pScript) && L1_CACHE_BYTES < dma_get_cache_alignment());
317 hostdata->slots = (struct NCR_700_command_slot *)(memory + SLOTS_OFFSET);
318 hostdata->dev = dev;
6391a113 319
1da177e4
LT
320 pSlots = pScript + SLOTS_OFFSET;
321
322 /* Fill in the missing routines from the host template */
323 tpnt->queuecommand = NCR_700_queuecommand;
324 tpnt->eh_abort_handler = NCR_700_abort;
1da177e4
LT
325 tpnt->eh_bus_reset_handler = NCR_700_bus_reset;
326 tpnt->eh_host_reset_handler = NCR_700_host_reset;
327 tpnt->can_queue = NCR_700_COMMAND_SLOTS_PER_HOST;
328 tpnt->sg_tablesize = NCR_700_SG_SEGMENTS;
329 tpnt->cmd_per_lun = NCR_700_CMD_PER_LUN;
330 tpnt->use_clustering = ENABLE_CLUSTERING;
331 tpnt->slave_configure = NCR_700_slave_configure;
332 tpnt->slave_destroy = NCR_700_slave_destroy;
333 tpnt->change_queue_depth = NCR_700_change_queue_depth;
334 tpnt->change_queue_type = NCR_700_change_queue_type;
6391a113 335
1da177e4
LT
336 if(tpnt->name == NULL)
337 tpnt->name = "53c700";
338 if(tpnt->proc_name == NULL)
339 tpnt->proc_name = "53c700";
1da177e4
LT
340
341 host = scsi_host_alloc(tpnt, 4);
342 if (!host)
343 return NULL;
344 memset(hostdata->slots, 0, sizeof(struct NCR_700_command_slot)
345 * NCR_700_COMMAND_SLOTS_PER_HOST);
6391a113 346 for (j = 0; j < NCR_700_COMMAND_SLOTS_PER_HOST; j++) {
1da177e4
LT
347 dma_addr_t offset = (dma_addr_t)((unsigned long)&hostdata->slots[j].SG[0]
348 - (unsigned long)&hostdata->slots[0].SG[0]);
349 hostdata->slots[j].pSG = (struct NCR_700_SG_List *)((unsigned long)(pSlots + offset));
350 if(j == 0)
351 hostdata->free_list = &hostdata->slots[j];
352 else
353 hostdata->slots[j-1].ITL_forw = &hostdata->slots[j];
354 hostdata->slots[j].state = NCR_700_SLOT_FREE;
355 }
356
6391a113 357 for (j = 0; j < ARRAY_SIZE(SCRIPT); j++)
1da177e4 358 script[j] = bS_to_host(SCRIPT[j]);
1da177e4
LT
359
360 /* adjust all labels to be bus physical */
6391a113 361 for (j = 0; j < PATCHES; j++)
1da177e4 362 script[LABELPATCHES[j]] = bS_to_host(pScript + SCRIPT[LABELPATCHES[j]]);
1da177e4
LT
363 /* now patch up fixed addresses. */
364 script_patch_32(script, MessageLocation,
365 pScript + MSGOUT_OFFSET);
366 script_patch_32(script, StatusAddress,
367 pScript + STATUS_OFFSET);
368 script_patch_32(script, ReceiveMsgAddress,
369 pScript + MSGIN_OFFSET);
370
371 hostdata->script = script;
372 hostdata->pScript = pScript;
373 dma_sync_single_for_device(hostdata->dev, pScript, sizeof(SCRIPT), DMA_TO_DEVICE);
374 hostdata->state = NCR_700_HOST_FREE;
375 hostdata->cmd = NULL;
2b89dad0 376 host->max_id = 8;
1da177e4
LT
377 host->max_lun = NCR_700_MAX_LUNS;
378 BUG_ON(NCR_700_transport_template == NULL);
379 host->transportt = NCR_700_transport_template;
56fece20 380 host->unique_id = (unsigned long)hostdata->base;
1da177e4
LT
381 hostdata->eh_complete = NULL;
382 host->hostdata[0] = (unsigned long)hostdata;
383 /* kick the chip */
384 NCR_700_writeb(0xff, host, CTEST9_REG);
6391a113 385 if (hostdata->chip710)
1da177e4
LT
386 hostdata->rev = (NCR_700_readb(host, CTEST8_REG)>>4) & 0x0f;
387 else
388 hostdata->rev = (NCR_700_readb(host, CTEST7_REG)>>4) & 0x0f;
389 hostdata->fast = (NCR_700_readb(host, CTEST9_REG) == 0);
6391a113 390 if (banner == 0) {
1da177e4
LT
391 printk(KERN_NOTICE "53c700: Version " NCR_700_VERSION " By James.Bottomley@HansenPartnership.com\n");
392 banner = 1;
393 }
394 printk(KERN_NOTICE "scsi%d: %s rev %d %s\n", host->host_no,
6391a113 395 hostdata->chip710 ? "53c710" :
1da177e4
LT
396 (hostdata->fast ? "53c700-66" : "53c700"),
397 hostdata->rev, hostdata->differential ?
398 "(Differential)" : "");
399 /* reset the chip */
400 NCR_700_chip_reset(host);
401
402 if (scsi_add_host(host, dev)) {
403 dev_printk(KERN_ERR, dev, "53c700: scsi_add_host failed\n");
404 scsi_host_put(host);
405 return NULL;
406 }
407
408 spi_signalling(host) = hostdata->differential ? SPI_SIGNAL_HVD :
409 SPI_SIGNAL_SE;
410
411 return host;
412}
413
414int
415NCR_700_release(struct Scsi_Host *host)
416{
417 struct NCR_700_Host_Parameters *hostdata =
418 (struct NCR_700_Host_Parameters *)host->hostdata[0];
419
420 dma_free_noncoherent(hostdata->dev, TOTAL_MEM_SIZE,
421 hostdata->script, hostdata->pScript);
422 return 1;
423}
424
425static inline __u8
426NCR_700_identify(int can_disconnect, __u8 lun)
427{
428 return IDENTIFY_BASE |
429 ((can_disconnect) ? 0x40 : 0) |
430 (lun & NCR_700_LUN_MASK);
431}
432
433/*
434 * Function : static int data_residual (Scsi_Host *host)
435 *
436 * Purpose : return residual data count of what's in the chip. If you
437 * really want to know what this function is doing, it's almost a
438 * direct transcription of the algorithm described in the 53c710
439 * guide, except that the DBC and DFIFO registers are only 6 bits
440 * wide on a 53c700.
441 *
442 * Inputs : host - SCSI host */
443static inline int
444NCR_700_data_residual (struct Scsi_Host *host) {
445 struct NCR_700_Host_Parameters *hostdata =
446 (struct NCR_700_Host_Parameters *)host->hostdata[0];
447 int count, synchronous = 0;
448 unsigned int ddir;
449
450 if(hostdata->chip710) {
451 count = ((NCR_700_readb(host, DFIFO_REG) & 0x7f) -
452 (NCR_700_readl(host, DBC_REG) & 0x7f)) & 0x7f;
453 } else {
454 count = ((NCR_700_readb(host, DFIFO_REG) & 0x3f) -
455 (NCR_700_readl(host, DBC_REG) & 0x3f)) & 0x3f;
456 }
457
458 if(hostdata->fast)
459 synchronous = NCR_700_readb(host, SXFER_REG) & 0x0f;
460
461 /* get the data direction */
462 ddir = NCR_700_readb(host, CTEST0_REG) & 0x01;
463
464 if (ddir) {
465 /* Receive */
466 if (synchronous)
467 count += (NCR_700_readb(host, SSTAT2_REG) & 0xf0) >> 4;
468 else
469 if (NCR_700_readb(host, SSTAT1_REG) & SIDL_REG_FULL)
470 ++count;
471 } else {
472 /* Send */
473 __u8 sstat = NCR_700_readb(host, SSTAT1_REG);
474 if (sstat & SODL_REG_FULL)
475 ++count;
476 if (synchronous && (sstat & SODR_REG_FULL))
477 ++count;
478 }
479#ifdef NCR_700_DEBUG
480 if(count)
481 printk("RESIDUAL IS %d (ddir %d)\n", count, ddir);
482#endif
483 return count;
484}
485
486/* print out the SCSI wires and corresponding phase from the SBCL register
487 * in the chip */
488static inline char *
489sbcl_to_string(__u8 sbcl)
490{
491 int i;
492 static char ret[256];
493
494 ret[0]='\0';
495 for(i=0; i<8; i++) {
496 if((1<<i) & sbcl)
497 strcat(ret, NCR_700_SBCL_bits[i]);
498 }
499 strcat(ret, NCR_700_SBCL_to_phase[sbcl & 0x07]);
500 return ret;
501}
502
503static inline __u8
504bitmap_to_number(__u8 bitmap)
505{
506 __u8 i;
507
508 for(i=0; i<8 && !(bitmap &(1<<i)); i++)
509 ;
510 return i;
511}
512
513/* Pull a slot off the free list */
514STATIC struct NCR_700_command_slot *
515find_empty_slot(struct NCR_700_Host_Parameters *hostdata)
516{
517 struct NCR_700_command_slot *slot = hostdata->free_list;
518
519 if(slot == NULL) {
520 /* sanity check */
521 if(hostdata->command_slot_count != NCR_700_COMMAND_SLOTS_PER_HOST)
522 printk(KERN_ERR "SLOTS FULL, but count is %d, should be %d\n", hostdata->command_slot_count, NCR_700_COMMAND_SLOTS_PER_HOST);
523 return NULL;
524 }
525
526 if(slot->state != NCR_700_SLOT_FREE)
527 /* should panic! */
528 printk(KERN_ERR "BUSY SLOT ON FREE LIST!!!\n");
529
530
531 hostdata->free_list = slot->ITL_forw;
532 slot->ITL_forw = NULL;
533
534
535 /* NOTE: set the state to busy here, not queued, since this
536 * indicates the slot is in use and cannot be run by the IRQ
537 * finish routine. If we cannot queue the command when it
538 * is properly build, we then change to NCR_700_SLOT_QUEUED */
539 slot->state = NCR_700_SLOT_BUSY;
540 hostdata->command_slot_count++;
541
542 return slot;
543}
544
545STATIC void
546free_slot(struct NCR_700_command_slot *slot,
547 struct NCR_700_Host_Parameters *hostdata)
548{
549 if((slot->state & NCR_700_SLOT_MASK) != NCR_700_SLOT_MAGIC) {
550 printk(KERN_ERR "53c700: SLOT %p is not MAGIC!!!\n", slot);
551 }
552 if(slot->state == NCR_700_SLOT_FREE) {
553 printk(KERN_ERR "53c700: SLOT %p is FREE!!!\n", slot);
554 }
555
556 slot->resume_offset = 0;
557 slot->cmnd = NULL;
558 slot->state = NCR_700_SLOT_FREE;
559 slot->ITL_forw = hostdata->free_list;
560 hostdata->free_list = slot;
561 hostdata->command_slot_count--;
562}
563
564
565/* This routine really does very little. The command is indexed on
566 the ITL and (if tagged) the ITLQ lists in _queuecommand */
567STATIC void
568save_for_reselection(struct NCR_700_Host_Parameters *hostdata,
569 struct scsi_cmnd *SCp, __u32 dsp)
570{
571 /* Its just possible that this gets executed twice */
572 if(SCp != NULL) {
573 struct NCR_700_command_slot *slot =
574 (struct NCR_700_command_slot *)SCp->host_scribble;
575
576 slot->resume_offset = dsp;
577 }
578 hostdata->state = NCR_700_HOST_FREE;
579 hostdata->cmd = NULL;
580}
581
582STATIC inline void
583NCR_700_unmap(struct NCR_700_Host_Parameters *hostdata, struct scsi_cmnd *SCp,
584 struct NCR_700_command_slot *slot)
585{
586 if(SCp->sc_data_direction != DMA_NONE &&
587 SCp->sc_data_direction != DMA_BIDIRECTIONAL) {
588 if(SCp->use_sg) {
589 dma_unmap_sg(hostdata->dev, SCp->buffer,
590 SCp->use_sg, SCp->sc_data_direction);
591 } else {
592 dma_unmap_single(hostdata->dev, slot->dma_handle,
593 SCp->request_bufflen,
594 SCp->sc_data_direction);
595 }
596 }
597}
598
599STATIC inline void
600NCR_700_scsi_done(struct NCR_700_Host_Parameters *hostdata,
601 struct scsi_cmnd *SCp, int result)
602{
603 hostdata->state = NCR_700_HOST_FREE;
604 hostdata->cmd = NULL;
605
606 if(SCp != NULL) {
607 struct NCR_700_command_slot *slot =
608 (struct NCR_700_command_slot *)SCp->host_scribble;
609
610 NCR_700_unmap(hostdata, SCp, slot);
611 dma_unmap_single(hostdata->dev, slot->pCmd,
612 sizeof(SCp->cmnd), DMA_TO_DEVICE);
613 if(SCp->cmnd[0] == REQUEST_SENSE && SCp->cmnd[6] == NCR_700_INTERNAL_SENSE_MAGIC) {
614#ifdef NCR_700_DEBUG
615 printk(" ORIGINAL CMD %p RETURNED %d, new return is %d sense is\n",
616 SCp, SCp->cmnd[7], result);
617 scsi_print_sense("53c700", SCp);
618
619#endif
620 /* restore the old result if the request sense was
621 * successful */
622 if(result == 0)
623 result = SCp->cmnd[7];
624 /* now restore the original command */
625 memcpy((void *) SCp->cmnd, (void *) SCp->data_cmnd,
626 sizeof(SCp->data_cmnd));
627 SCp->request_buffer = SCp->buffer;
628 SCp->request_bufflen = SCp->bufflen;
629 SCp->use_sg = SCp->old_use_sg;
630 SCp->cmd_len = SCp->old_cmd_len;
631 SCp->sc_data_direction = SCp->sc_old_data_direction;
632 SCp->underflow = SCp->old_underflow;
633
634 }
635 free_slot(slot, hostdata);
636#ifdef NCR_700_DEBUG
637 if(NCR_700_get_depth(SCp->device) == 0 ||
638 NCR_700_get_depth(SCp->device) > SCp->device->queue_depth)
639 printk(KERN_ERR "Invalid depth in NCR_700_scsi_done(): %d\n",
640 NCR_700_get_depth(SCp->device));
641#endif /* NCR_700_DEBUG */
642 NCR_700_set_depth(SCp->device, NCR_700_get_depth(SCp->device) - 1);
643
644 SCp->host_scribble = NULL;
645 SCp->result = result;
646 SCp->scsi_done(SCp);
647 } else {
648 printk(KERN_ERR "53c700: SCSI DONE HAS NULL SCp\n");
649 }
650}
651
652
653STATIC void
654NCR_700_internal_bus_reset(struct Scsi_Host *host)
655{
656 /* Bus reset */
657 NCR_700_writeb(ASSERT_RST, host, SCNTL1_REG);
658 udelay(50);
659 NCR_700_writeb(0, host, SCNTL1_REG);
660
661}
662
663STATIC void
664NCR_700_chip_setup(struct Scsi_Host *host)
665{
666 struct NCR_700_Host_Parameters *hostdata =
667 (struct NCR_700_Host_Parameters *)host->hostdata[0];
668 __u32 dcntl_extra = 0;
669 __u8 min_period;
670 __u8 min_xferp = (hostdata->chip710 ? NCR_710_MIN_XFERP : NCR_700_MIN_XFERP);
671
672 if(hostdata->chip710) {
673 __u8 burst_disable = hostdata->burst_disable
674 ? BURST_DISABLE : 0;
675 dcntl_extra = COMPAT_700_MODE;
676
677 NCR_700_writeb(dcntl_extra, host, DCNTL_REG);
678 NCR_700_writeb(BURST_LENGTH_8 | hostdata->dmode_extra,
679 host, DMODE_710_REG);
680 NCR_700_writeb(burst_disable | (hostdata->differential ?
681 DIFF : 0), host, CTEST7_REG);
682 NCR_700_writeb(BTB_TIMER_DISABLE, host, CTEST0_REG);
683 NCR_700_writeb(FULL_ARBITRATION | ENABLE_PARITY | PARITY
684 | AUTO_ATN, host, SCNTL0_REG);
685 } else {
686 NCR_700_writeb(BURST_LENGTH_8 | hostdata->dmode_extra,
687 host, DMODE_700_REG);
688 NCR_700_writeb(hostdata->differential ?
689 DIFF : 0, host, CTEST7_REG);
690 if(hostdata->fast) {
691 /* this is for 700-66, does nothing on 700 */
692 NCR_700_writeb(LAST_DIS_ENBL | ENABLE_ACTIVE_NEGATION
693 | GENERATE_RECEIVE_PARITY, host,
694 CTEST8_REG);
695 } else {
696 NCR_700_writeb(FULL_ARBITRATION | ENABLE_PARITY
697 | PARITY | AUTO_ATN, host, SCNTL0_REG);
698 }
699 }
700
701 NCR_700_writeb(1 << host->this_id, host, SCID_REG);
702 NCR_700_writeb(0, host, SBCL_REG);
703 NCR_700_writeb(ASYNC_OPERATION, host, SXFER_REG);
704
705 NCR_700_writeb(PHASE_MM_INT | SEL_TIMEOUT_INT | GROSS_ERR_INT | UX_DISC_INT
706 | RST_INT | PAR_ERR_INT | SELECT_INT, host, SIEN_REG);
707
708 NCR_700_writeb(ABORT_INT | INT_INST_INT | ILGL_INST_INT, host, DIEN_REG);
709 NCR_700_writeb(ENABLE_SELECT, host, SCNTL1_REG);
710 if(hostdata->clock > 75) {
711 printk(KERN_ERR "53c700: Clock speed %dMHz is too high: 75Mhz is the maximum this chip can be driven at\n", hostdata->clock);
712 /* do the best we can, but the async clock will be out
713 * of spec: sync divider 2, async divider 3 */
714 DEBUG(("53c700: sync 2 async 3\n"));
715 NCR_700_writeb(SYNC_DIV_2_0, host, SBCL_REG);
716 NCR_700_writeb(ASYNC_DIV_3_0 | dcntl_extra, host, DCNTL_REG);
717 hostdata->sync_clock = hostdata->clock/2;
718 } else if(hostdata->clock > 50 && hostdata->clock <= 75) {
719 /* sync divider 1.5, async divider 3 */
720 DEBUG(("53c700: sync 1.5 async 3\n"));
721 NCR_700_writeb(SYNC_DIV_1_5, host, SBCL_REG);
722 NCR_700_writeb(ASYNC_DIV_3_0 | dcntl_extra, host, DCNTL_REG);
723 hostdata->sync_clock = hostdata->clock*2;
724 hostdata->sync_clock /= 3;
725
726 } else if(hostdata->clock > 37 && hostdata->clock <= 50) {
727 /* sync divider 1, async divider 2 */
728 DEBUG(("53c700: sync 1 async 2\n"));
729 NCR_700_writeb(SYNC_DIV_1_0, host, SBCL_REG);
730 NCR_700_writeb(ASYNC_DIV_2_0 | dcntl_extra, host, DCNTL_REG);
731 hostdata->sync_clock = hostdata->clock;
732 } else if(hostdata->clock > 25 && hostdata->clock <=37) {
733 /* sync divider 1, async divider 1.5 */
734 DEBUG(("53c700: sync 1 async 1.5\n"));
735 NCR_700_writeb(SYNC_DIV_1_0, host, SBCL_REG);
736 NCR_700_writeb(ASYNC_DIV_1_5 | dcntl_extra, host, DCNTL_REG);
737 hostdata->sync_clock = hostdata->clock;
738 } else {
739 DEBUG(("53c700: sync 1 async 1\n"));
740 NCR_700_writeb(SYNC_DIV_1_0, host, SBCL_REG);
741 NCR_700_writeb(ASYNC_DIV_1_0 | dcntl_extra, host, DCNTL_REG);
742 /* sync divider 1, async divider 1 */
743 hostdata->sync_clock = hostdata->clock;
744 }
745 /* Calculate the actual minimum period that can be supported
746 * by our synchronous clock speed. See the 710 manual for
747 * exact details of this calculation which is based on a
748 * setting of the SXFER register */
749 min_period = 1000*(4+min_xferp)/(4*hostdata->sync_clock);
750 hostdata->min_period = NCR_700_MIN_PERIOD;
751 if(min_period > NCR_700_MIN_PERIOD)
752 hostdata->min_period = min_period;
753}
754
755STATIC void
756NCR_700_chip_reset(struct Scsi_Host *host)
757{
758 struct NCR_700_Host_Parameters *hostdata =
759 (struct NCR_700_Host_Parameters *)host->hostdata[0];
760 if(hostdata->chip710) {
761 NCR_700_writeb(SOFTWARE_RESET_710, host, ISTAT_REG);
762 udelay(100);
763
764 NCR_700_writeb(0, host, ISTAT_REG);
765 } else {
766 NCR_700_writeb(SOFTWARE_RESET, host, DCNTL_REG);
767 udelay(100);
768
769 NCR_700_writeb(0, host, DCNTL_REG);
770 }
771
772 mdelay(1000);
773
774 NCR_700_chip_setup(host);
775}
776
777/* The heart of the message processing engine is that the instruction
778 * immediately after the INT is the normal case (and so must be CLEAR
779 * ACK). If we want to do something else, we call that routine in
780 * scripts and set temp to be the normal case + 8 (skipping the CLEAR
781 * ACK) so that the routine returns correctly to resume its activity
782 * */
783STATIC __u32
784process_extended_message(struct Scsi_Host *host,
785 struct NCR_700_Host_Parameters *hostdata,
786 struct scsi_cmnd *SCp, __u32 dsp, __u32 dsps)
787{
788 __u32 resume_offset = dsp, temp = dsp + 8;
789 __u8 pun = 0xff, lun = 0xff;
790
791 if(SCp != NULL) {
792 pun = SCp->device->id;
793 lun = SCp->device->lun;
794 }
795
796 switch(hostdata->msgin[2]) {
797 case A_SDTR_MSG:
798 if(SCp != NULL && NCR_700_is_flag_set(SCp->device, NCR_700_DEV_BEGIN_SYNC_NEGOTIATION)) {
799 struct scsi_target *starget = SCp->device->sdev_target;
800 __u8 period = hostdata->msgin[3];
801 __u8 offset = hostdata->msgin[4];
802
803 if(offset == 0 || period == 0) {
804 offset = 0;
805 period = 0;
806 }
807
808 spi_offset(starget) = offset;
809 spi_period(starget) = period;
810
811 if(NCR_700_is_flag_set(SCp->device, NCR_700_DEV_PRINT_SYNC_NEGOTIATION)) {
812 spi_display_xfer_agreement(starget);
813 NCR_700_clear_flag(SCp->device, NCR_700_DEV_PRINT_SYNC_NEGOTIATION);
814 }
815
816 NCR_700_set_flag(SCp->device, NCR_700_DEV_NEGOTIATED_SYNC);
817 NCR_700_clear_flag(SCp->device, NCR_700_DEV_BEGIN_SYNC_NEGOTIATION);
818
819 NCR_700_writeb(NCR_700_get_SXFER(SCp->device),
820 host, SXFER_REG);
821
822 } else {
823 /* SDTR message out of the blue, reject it */
017560fc
JG
824 shost_printk(KERN_WARNING, host,
825 "Unexpected SDTR msg\n");
1da177e4
LT
826 hostdata->msgout[0] = A_REJECT_MSG;
827 dma_cache_sync(hostdata->msgout, 1, DMA_TO_DEVICE);
828 script_patch_16(hostdata->script, MessageCount, 1);
829 /* SendMsgOut returns, so set up the return
830 * address */
831 resume_offset = hostdata->pScript + Ent_SendMessageWithATN;
832 }
833 break;
834
835 case A_WDTR_MSG:
836 printk(KERN_INFO "scsi%d: (%d:%d), Unsolicited WDTR after CMD, Rejecting\n",
837 host->host_no, pun, lun);
838 hostdata->msgout[0] = A_REJECT_MSG;
839 dma_cache_sync(hostdata->msgout, 1, DMA_TO_DEVICE);
840 script_patch_16(hostdata->script, MessageCount, 1);
841 resume_offset = hostdata->pScript + Ent_SendMessageWithATN;
842
843 break;
844
845 default:
846 printk(KERN_INFO "scsi%d (%d:%d): Unexpected message %s: ",
847 host->host_no, pun, lun,
848 NCR_700_phase[(dsps & 0xf00) >> 8]);
1abfd370 849 spi_print_msg(hostdata->msgin);
1da177e4
LT
850 printk("\n");
851 /* just reject it */
852 hostdata->msgout[0] = A_REJECT_MSG;
853 dma_cache_sync(hostdata->msgout, 1, DMA_TO_DEVICE);
854 script_patch_16(hostdata->script, MessageCount, 1);
855 /* SendMsgOut returns, so set up the return
856 * address */
857 resume_offset = hostdata->pScript + Ent_SendMessageWithATN;
858 }
859 NCR_700_writel(temp, host, TEMP_REG);
860 return resume_offset;
861}
862
863STATIC __u32
864process_message(struct Scsi_Host *host, struct NCR_700_Host_Parameters *hostdata,
865 struct scsi_cmnd *SCp, __u32 dsp, __u32 dsps)
866{
867 /* work out where to return to */
868 __u32 temp = dsp + 8, resume_offset = dsp;
869 __u8 pun = 0xff, lun = 0xff;
870
871 if(SCp != NULL) {
872 pun = SCp->device->id;
873 lun = SCp->device->lun;
874 }
875
876#ifdef NCR_700_DEBUG
877 printk("scsi%d (%d:%d): message %s: ", host->host_no, pun, lun,
878 NCR_700_phase[(dsps & 0xf00) >> 8]);
1abfd370 879 spi_print_msg(hostdata->msgin);
1da177e4
LT
880 printk("\n");
881#endif
882
883 switch(hostdata->msgin[0]) {
884
885 case A_EXTENDED_MSG:
886 resume_offset = process_extended_message(host, hostdata, SCp,
887 dsp, dsps);
888 break;
889
890 case A_REJECT_MSG:
891 if(SCp != NULL && NCR_700_is_flag_set(SCp->device, NCR_700_DEV_BEGIN_SYNC_NEGOTIATION)) {
892 /* Rejected our sync negotiation attempt */
893 spi_period(SCp->device->sdev_target) =
894 spi_offset(SCp->device->sdev_target) = 0;
895 NCR_700_set_flag(SCp->device, NCR_700_DEV_NEGOTIATED_SYNC);
896 NCR_700_clear_flag(SCp->device, NCR_700_DEV_BEGIN_SYNC_NEGOTIATION);
897 } else if(SCp != NULL && NCR_700_get_tag_neg_state(SCp->device) == NCR_700_DURING_TAG_NEGOTIATION) {
898 /* rejected our first simple tag message */
017560fc
JG
899 scmd_printk(KERN_WARNING, SCp,
900 "Rejected first tag queue attempt, turning off tag queueing\n");
1da177e4
LT
901 /* we're done negotiating */
902 NCR_700_set_tag_neg_state(SCp->device, NCR_700_FINISHED_TAG_NEGOTIATION);
017560fc 903 hostdata->tag_negotiated &= ~(1<<scmd_id(SCp));
1da177e4
LT
904 SCp->device->tagged_supported = 0;
905 scsi_deactivate_tcq(SCp->device, host->cmd_per_lun);
906 } else {
017560fc
JG
907 shost_printk(KERN_WARNING, host,
908 "(%d:%d) Unexpected REJECT Message %s\n",
909 pun, lun,
1da177e4
LT
910 NCR_700_phase[(dsps & 0xf00) >> 8]);
911 /* however, just ignore it */
912 }
913 break;
914
915 case A_PARITY_ERROR_MSG:
916 printk(KERN_ERR "scsi%d (%d:%d) Parity Error!\n", host->host_no,
917 pun, lun);
918 NCR_700_internal_bus_reset(host);
919 break;
920 case A_SIMPLE_TAG_MSG:
921 printk(KERN_INFO "scsi%d (%d:%d) SIMPLE TAG %d %s\n", host->host_no,
922 pun, lun, hostdata->msgin[1],
923 NCR_700_phase[(dsps & 0xf00) >> 8]);
924 /* just ignore it */
925 break;
926 default:
927 printk(KERN_INFO "scsi%d (%d:%d): Unexpected message %s: ",
928 host->host_no, pun, lun,
929 NCR_700_phase[(dsps & 0xf00) >> 8]);
930
1abfd370 931 spi_print_msg(hostdata->msgin);
1da177e4
LT
932 printk("\n");
933 /* just reject it */
934 hostdata->msgout[0] = A_REJECT_MSG;
935 dma_cache_sync(hostdata->msgout, 1, DMA_TO_DEVICE);
936 script_patch_16(hostdata->script, MessageCount, 1);
937 /* SendMsgOut returns, so set up the return
938 * address */
939 resume_offset = hostdata->pScript + Ent_SendMessageWithATN;
940
941 break;
942 }
943 NCR_700_writel(temp, host, TEMP_REG);
944 /* set us up to receive another message */
945 dma_cache_sync(hostdata->msgin, MSG_ARRAY_SIZE, DMA_FROM_DEVICE);
946 return resume_offset;
947}
948
949STATIC __u32
950process_script_interrupt(__u32 dsps, __u32 dsp, struct scsi_cmnd *SCp,
951 struct Scsi_Host *host,
952 struct NCR_700_Host_Parameters *hostdata)
953{
954 __u32 resume_offset = 0;
955 __u8 pun = 0xff, lun=0xff;
956
957 if(SCp != NULL) {
958 pun = SCp->device->id;
959 lun = SCp->device->lun;
960 }
961
962 if(dsps == A_GOOD_STATUS_AFTER_STATUS) {
963 DEBUG((" COMMAND COMPLETE, status=%02x\n",
964 hostdata->status[0]));
965 /* OK, if TCQ still under negotiation, we now know it works */
966 if (NCR_700_get_tag_neg_state(SCp->device) == NCR_700_DURING_TAG_NEGOTIATION)
967 NCR_700_set_tag_neg_state(SCp->device,
968 NCR_700_FINISHED_TAG_NEGOTIATION);
969
970 /* check for contingent allegiance contitions */
971 if(status_byte(hostdata->status[0]) == CHECK_CONDITION ||
972 status_byte(hostdata->status[0]) == COMMAND_TERMINATED) {
973 struct NCR_700_command_slot *slot =
974 (struct NCR_700_command_slot *)SCp->host_scribble;
975 if(SCp->cmnd[0] == REQUEST_SENSE) {
976 /* OOPS: bad device, returning another
977 * contingent allegiance condition */
017560fc
JG
978 scmd_printk(KERN_ERR, SCp,
979 "broken device is looping in contingent allegiance: ignoring\n");
1da177e4
LT
980 NCR_700_scsi_done(hostdata, SCp, hostdata->status[0]);
981 } else {
982#ifdef NCR_DEBUG
983 scsi_print_command(SCp);
984 printk(" cmd %p has status %d, requesting sense\n",
985 SCp, hostdata->status[0]);
986#endif
987 /* we can destroy the command here
988 * because the contingent allegiance
989 * condition will cause a retry which
990 * will re-copy the command from the
991 * saved data_cmnd. We also unmap any
992 * data associated with the command
993 * here */
994 NCR_700_unmap(hostdata, SCp, slot);
995
996 SCp->cmnd[0] = REQUEST_SENSE;
997 SCp->cmnd[1] = (SCp->device->lun & 0x7) << 5;
998 SCp->cmnd[2] = 0;
999 SCp->cmnd[3] = 0;
1000 SCp->cmnd[4] = sizeof(SCp->sense_buffer);
1001 SCp->cmnd[5] = 0;
1002 SCp->cmd_len = 6;
1003 /* Here's a quiet hack: the
1004 * REQUEST_SENSE command is six bytes,
1005 * so store a flag indicating that
1006 * this was an internal sense request
1007 * and the original status at the end
1008 * of the command */
1009 SCp->cmnd[6] = NCR_700_INTERNAL_SENSE_MAGIC;
1010 SCp->cmnd[7] = hostdata->status[0];
1011 SCp->use_sg = 0;
1012 SCp->sc_data_direction = DMA_FROM_DEVICE;
1013 dma_sync_single_for_device(hostdata->dev, slot->pCmd,
1014 SCp->cmd_len, DMA_TO_DEVICE);
1015 SCp->request_bufflen = sizeof(SCp->sense_buffer);
1016 slot->dma_handle = dma_map_single(hostdata->dev, SCp->sense_buffer, sizeof(SCp->sense_buffer), DMA_FROM_DEVICE);
1017 slot->SG[0].ins = bS_to_host(SCRIPT_MOVE_DATA_IN | sizeof(SCp->sense_buffer));
1018 slot->SG[0].pAddr = bS_to_host(slot->dma_handle);
1019 slot->SG[1].ins = bS_to_host(SCRIPT_RETURN);
1020 slot->SG[1].pAddr = 0;
1021 slot->resume_offset = hostdata->pScript;
1022 dma_cache_sync(slot->SG, sizeof(slot->SG[0])*2, DMA_TO_DEVICE);
1023 dma_cache_sync(SCp->sense_buffer, sizeof(SCp->sense_buffer), DMA_FROM_DEVICE);
1024
1025 /* queue the command for reissue */
1026 slot->state = NCR_700_SLOT_QUEUED;
1027 hostdata->state = NCR_700_HOST_FREE;
1028 hostdata->cmd = NULL;
1029 }
1030 } else {
1031 // Currently rely on the mid layer evaluation
1032 // of the tag queuing capability
1033 //
1034 //if(status_byte(hostdata->status[0]) == GOOD &&
1035 // SCp->cmnd[0] == INQUIRY && SCp->use_sg == 0) {
1036 // /* Piggy back the tag queueing support
1037 // * on this command */
1038 // dma_sync_single_for_cpu(hostdata->dev,
1039 // slot->dma_handle,
1040 // SCp->request_bufflen,
1041 // DMA_FROM_DEVICE);
1042 // if(((char *)SCp->request_buffer)[7] & 0x02) {
017560fc
JG
1043 // scmd_printk(KERN_INFO, SCp,
1044 // "Enabling Tag Command Queuing\n");
1045 // hostdata->tag_negotiated |= (1<<scmd_id(SCp));
1da177e4
LT
1046 // NCR_700_set_flag(SCp->device, NCR_700_DEV_BEGIN_TAG_QUEUEING);
1047 // } else {
1048 // NCR_700_clear_flag(SCp->device, NCR_700_DEV_BEGIN_TAG_QUEUEING);
017560fc 1049 // hostdata->tag_negotiated &= ~(1<<scmd_id(SCp));
1da177e4
LT
1050 // }
1051 //}
1052 NCR_700_scsi_done(hostdata, SCp, hostdata->status[0]);
1053 }
1054 } else if((dsps & 0xfffff0f0) == A_UNEXPECTED_PHASE) {
1055 __u8 i = (dsps & 0xf00) >> 8;
1056
017560fc 1057 scmd_printk(KERN_ERR, SCp, "UNEXPECTED PHASE %s (%s)\n",
1da177e4
LT
1058 NCR_700_phase[i],
1059 sbcl_to_string(NCR_700_readb(host, SBCL_REG)));
017560fc
JG
1060 scmd_printk(KERN_ERR, SCp, " len = %d, cmd =",
1061 SCp->cmd_len);
1da177e4
LT
1062 scsi_print_command(SCp);
1063
1064 NCR_700_internal_bus_reset(host);
1065 } else if((dsps & 0xfffff000) == A_FATAL) {
1066 int i = (dsps & 0xfff);
1067
1068 printk(KERN_ERR "scsi%d: (%d:%d) FATAL ERROR: %s\n",
1069 host->host_no, pun, lun, NCR_700_fatal_messages[i]);
1070 if(dsps == A_FATAL_ILLEGAL_MSG_LENGTH) {
1071 printk(KERN_ERR " msg begins %02x %02x\n",
1072 hostdata->msgin[0], hostdata->msgin[1]);
1073 }
1074 NCR_700_internal_bus_reset(host);
1075 } else if((dsps & 0xfffff0f0) == A_DISCONNECT) {
1076#ifdef NCR_700_DEBUG
1077 __u8 i = (dsps & 0xf00) >> 8;
1078
1079 printk("scsi%d: (%d:%d), DISCONNECTED (%d) %s\n",
1080 host->host_no, pun, lun,
1081 i, NCR_700_phase[i]);
1082#endif
1083 save_for_reselection(hostdata, SCp, dsp);
1084
1085 } else if(dsps == A_RESELECTION_IDENTIFIED) {
1086 __u8 lun;
1087 struct NCR_700_command_slot *slot;
1088 __u8 reselection_id = hostdata->reselection_id;
1089 struct scsi_device *SDp;
1090
1091 lun = hostdata->msgin[0] & 0x1f;
1092
1093 hostdata->reselection_id = 0xff;
1094 DEBUG(("scsi%d: (%d:%d) RESELECTED!\n",
1095 host->host_no, reselection_id, lun));
1096 /* clear the reselection indicator */
1097 SDp = __scsi_device_lookup(host, 0, reselection_id, lun);
1098 if(unlikely(SDp == NULL)) {
1099 printk(KERN_ERR "scsi%d: (%d:%d) HAS NO device\n",
1100 host->host_no, reselection_id, lun);
1101 BUG();
1102 }
1103 if(hostdata->msgin[1] == A_SIMPLE_TAG_MSG) {
1104 struct scsi_cmnd *SCp = scsi_find_tag(SDp, hostdata->msgin[2]);
1105 if(unlikely(SCp == NULL)) {
1106 printk(KERN_ERR "scsi%d: (%d:%d) no saved request for tag %d\n",
1107 host->host_no, reselection_id, lun, hostdata->msgin[2]);
1108 BUG();
1109 }
1110
1111 slot = (struct NCR_700_command_slot *)SCp->host_scribble;
017560fc
JG
1112 DDEBUG(KERN_DEBUG, SDp,
1113 "reselection is tag %d, slot %p(%d)\n",
1114 hostdata->msgin[2], slot, slot->tag);
1da177e4
LT
1115 } else {
1116 struct scsi_cmnd *SCp = scsi_find_tag(SDp, SCSI_NO_TAG);
1117 if(unlikely(SCp == NULL)) {
017560fc
JG
1118 sdev_printk(KERN_ERR, SDp,
1119 "no saved request for untagged cmd\n");
1da177e4
LT
1120 BUG();
1121 }
1122 slot = (struct NCR_700_command_slot *)SCp->host_scribble;
1123 }
1124
1125 if(slot == NULL) {
1126 printk(KERN_ERR "scsi%d: (%d:%d) RESELECTED but no saved command (MSG = %02x %02x %02x)!!\n",
1127 host->host_no, reselection_id, lun,
1128 hostdata->msgin[0], hostdata->msgin[1],
1129 hostdata->msgin[2]);
1130 } else {
1131 if(hostdata->state != NCR_700_HOST_BUSY)
1132 printk(KERN_ERR "scsi%d: FATAL, host not busy during valid reselection!\n",
1133 host->host_no);
1134 resume_offset = slot->resume_offset;
1135 hostdata->cmd = slot->cmnd;
1136
1137 /* re-patch for this command */
1138 script_patch_32_abs(hostdata->script, CommandAddress,
1139 slot->pCmd);
1140 script_patch_16(hostdata->script,
1141 CommandCount, slot->cmnd->cmd_len);
1142 script_patch_32_abs(hostdata->script, SGScriptStartAddress,
1143 to32bit(&slot->pSG[0].ins));
1144
1145 /* Note: setting SXFER only works if we're
1146 * still in the MESSAGE phase, so it is vital
1147 * that ACK is still asserted when we process
1148 * the reselection message. The resume offset
1149 * should therefore always clear ACK */
1150 NCR_700_writeb(NCR_700_get_SXFER(hostdata->cmd->device),
1151 host, SXFER_REG);
1152 dma_cache_sync(hostdata->msgin,
1153 MSG_ARRAY_SIZE, DMA_FROM_DEVICE);
1154 dma_cache_sync(hostdata->msgout,
1155 MSG_ARRAY_SIZE, DMA_TO_DEVICE);
1156 /* I'm just being paranoid here, the command should
1157 * already have been flushed from the cache */
1158 dma_cache_sync(slot->cmnd->cmnd,
1159 slot->cmnd->cmd_len, DMA_TO_DEVICE);
1160
1161
1162
1163 }
1164 } else if(dsps == A_RESELECTED_DURING_SELECTION) {
1165
1166 /* This section is full of debugging code because I've
1167 * never managed to reach it. I think what happens is
1168 * that, because the 700 runs with selection
1169 * interrupts enabled the whole time that we take a
1170 * selection interrupt before we manage to get to the
1171 * reselected script interrupt */
1172
1173 __u8 reselection_id = NCR_700_readb(host, SFBR_REG);
1174 struct NCR_700_command_slot *slot;
1175
1176 /* Take out our own ID */
1177 reselection_id &= ~(1<<host->this_id);
1178
1179 /* I've never seen this happen, so keep this as a printk rather
1180 * than a debug */
1181 printk(KERN_INFO "scsi%d: (%d:%d) RESELECTION DURING SELECTION, dsp=%08x[%04x] state=%d, count=%d\n",
1182 host->host_no, reselection_id, lun, dsp, dsp - hostdata->pScript, hostdata->state, hostdata->command_slot_count);
1183
1184 {
1185 /* FIXME: DEBUGGING CODE */
1186 __u32 SG = (__u32)bS_to_cpu(hostdata->script[A_SGScriptStartAddress_used[0]]);
1187 int i;
1188
1189 for(i=0; i< NCR_700_COMMAND_SLOTS_PER_HOST; i++) {
1190 if(SG >= to32bit(&hostdata->slots[i].pSG[0])
1191 && SG <= to32bit(&hostdata->slots[i].pSG[NCR_700_SG_SEGMENTS]))
1192 break;
1193 }
1194 printk(KERN_INFO "IDENTIFIED SG segment as being %08x in slot %p, cmd %p, slot->resume_offset=%08x\n", SG, &hostdata->slots[i], hostdata->slots[i].cmnd, hostdata->slots[i].resume_offset);
1195 SCp = hostdata->slots[i].cmnd;
1196 }
1197
1198 if(SCp != NULL) {
1199 slot = (struct NCR_700_command_slot *)SCp->host_scribble;
1200 /* change slot from busy to queued to redo command */
1201 slot->state = NCR_700_SLOT_QUEUED;
1202 }
1203 hostdata->cmd = NULL;
1204
1205 if(reselection_id == 0) {
1206 if(hostdata->reselection_id == 0xff) {
1207 printk(KERN_ERR "scsi%d: Invalid reselection during selection!!\n", host->host_no);
1208 return 0;
1209 } else {
1210 printk(KERN_ERR "scsi%d: script reselected and we took a selection interrupt\n",
1211 host->host_no);
1212 reselection_id = hostdata->reselection_id;
1213 }
1214 } else {
1215
1216 /* convert to real ID */
1217 reselection_id = bitmap_to_number(reselection_id);
1218 }
1219 hostdata->reselection_id = reselection_id;
1220 /* just in case we have a stale simple tag message, clear it */
1221 hostdata->msgin[1] = 0;
1222 dma_cache_sync(hostdata->msgin,
1223 MSG_ARRAY_SIZE, DMA_BIDIRECTIONAL);
1224 if(hostdata->tag_negotiated & (1<<reselection_id)) {
1225 resume_offset = hostdata->pScript + Ent_GetReselectionWithTag;
1226 } else {
1227 resume_offset = hostdata->pScript + Ent_GetReselectionData;
1228 }
1229 } else if(dsps == A_COMPLETED_SELECTION_AS_TARGET) {
1230 /* we've just disconnected from the bus, do nothing since
1231 * a return here will re-run the queued command slot
1232 * that may have been interrupted by the initial selection */
1233 DEBUG((" SELECTION COMPLETED\n"));
1234 } else if((dsps & 0xfffff0f0) == A_MSG_IN) {
1235 resume_offset = process_message(host, hostdata, SCp,
1236 dsp, dsps);
1237 } else if((dsps & 0xfffff000) == 0) {
1238 __u8 i = (dsps & 0xf0) >> 4, j = (dsps & 0xf00) >> 8;
1239 printk(KERN_ERR "scsi%d: (%d:%d), unhandled script condition %s %s at %04x\n",
1240 host->host_no, pun, lun, NCR_700_condition[i],
1241 NCR_700_phase[j], dsp - hostdata->pScript);
1242 if(SCp != NULL) {
1243 scsi_print_command(SCp);
1244
1245 if(SCp->use_sg) {
1246 for(i = 0; i < SCp->use_sg + 1; i++) {
1247 printk(KERN_INFO " SG[%d].length = %d, move_insn=%08x, addr %08x\n", i, ((struct scatterlist *)SCp->buffer)[i].length, ((struct NCR_700_command_slot *)SCp->host_scribble)->SG[i].ins, ((struct NCR_700_command_slot *)SCp->host_scribble)->SG[i].pAddr);
1248 }
1249 }
1250 }
1251 NCR_700_internal_bus_reset(host);
1252 } else if((dsps & 0xfffff000) == A_DEBUG_INTERRUPT) {
1253 printk(KERN_NOTICE "scsi%d (%d:%d) DEBUG INTERRUPT %d AT %08x[%04x], continuing\n",
1254 host->host_no, pun, lun, dsps & 0xfff, dsp, dsp - hostdata->pScript);
1255 resume_offset = dsp;
1256 } else {
1257 printk(KERN_ERR "scsi%d: (%d:%d), unidentified script interrupt 0x%x at %04x\n",
1258 host->host_no, pun, lun, dsps, dsp - hostdata->pScript);
1259 NCR_700_internal_bus_reset(host);
1260 }
1261 return resume_offset;
1262}
1263
1264/* We run the 53c700 with selection interrupts always enabled. This
1265 * means that the chip may be selected as soon as the bus frees. On a
1266 * busy bus, this can be before the scripts engine finishes its
1267 * processing. Therefore, part of the selection processing has to be
1268 * to find out what the scripts engine is doing and complete the
1269 * function if necessary (i.e. process the pending disconnect or save
1270 * the interrupted initial selection */
1271STATIC inline __u32
1272process_selection(struct Scsi_Host *host, __u32 dsp)
1273{
1274 __u8 id = 0; /* Squash compiler warning */
1275 int count = 0;
1276 __u32 resume_offset = 0;
1277 struct NCR_700_Host_Parameters *hostdata =
1278 (struct NCR_700_Host_Parameters *)host->hostdata[0];
1279 struct scsi_cmnd *SCp = hostdata->cmd;
1280 __u8 sbcl;
1281
1282 for(count = 0; count < 5; count++) {
1283 id = NCR_700_readb(host, hostdata->chip710 ?
1284 CTEST9_REG : SFBR_REG);
1285
1286 /* Take out our own ID */
1287 id &= ~(1<<host->this_id);
1288 if(id != 0)
1289 break;
1290 udelay(5);
1291 }
1292 sbcl = NCR_700_readb(host, SBCL_REG);
1293 if((sbcl & SBCL_IO) == 0) {
1294 /* mark as having been selected rather than reselected */
1295 id = 0xff;
1296 } else {
1297 /* convert to real ID */
1298 hostdata->reselection_id = id = bitmap_to_number(id);
1299 DEBUG(("scsi%d: Reselected by %d\n",
1300 host->host_no, id));
1301 }
1302 if(hostdata->state == NCR_700_HOST_BUSY && SCp != NULL) {
1303 struct NCR_700_command_slot *slot =
1304 (struct NCR_700_command_slot *)SCp->host_scribble;
1305 DEBUG((" ID %d WARNING: RESELECTION OF BUSY HOST, saving cmd %p, slot %p, addr %x [%04x], resume %x!\n", id, hostdata->cmd, slot, dsp, dsp - hostdata->pScript, resume_offset));
1306
1307 switch(dsp - hostdata->pScript) {
1308 case Ent_Disconnect1:
1309 case Ent_Disconnect2:
1310 save_for_reselection(hostdata, SCp, Ent_Disconnect2 + hostdata->pScript);
1311 break;
1312 case Ent_Disconnect3:
1313 case Ent_Disconnect4:
1314 save_for_reselection(hostdata, SCp, Ent_Disconnect4 + hostdata->pScript);
1315 break;
1316 case Ent_Disconnect5:
1317 case Ent_Disconnect6:
1318 save_for_reselection(hostdata, SCp, Ent_Disconnect6 + hostdata->pScript);
1319 break;
1320 case Ent_Disconnect7:
1321 case Ent_Disconnect8:
1322 save_for_reselection(hostdata, SCp, Ent_Disconnect8 + hostdata->pScript);
1323 break;
1324 case Ent_Finish1:
1325 case Ent_Finish2:
1326 process_script_interrupt(A_GOOD_STATUS_AFTER_STATUS, dsp, SCp, host, hostdata);
1327 break;
1328
1329 default:
1330 slot->state = NCR_700_SLOT_QUEUED;
1331 break;
1332 }
1333 }
1334 hostdata->state = NCR_700_HOST_BUSY;
1335 hostdata->cmd = NULL;
1336 /* clear any stale simple tag message */
1337 hostdata->msgin[1] = 0;
1338 dma_cache_sync(hostdata->msgin, MSG_ARRAY_SIZE,
1339 DMA_BIDIRECTIONAL);
1340
1341 if(id == 0xff) {
1342 /* Selected as target, Ignore */
1343 resume_offset = hostdata->pScript + Ent_SelectedAsTarget;
1344 } else if(hostdata->tag_negotiated & (1<<id)) {
1345 resume_offset = hostdata->pScript + Ent_GetReselectionWithTag;
1346 } else {
1347 resume_offset = hostdata->pScript + Ent_GetReselectionData;
1348 }
1349 return resume_offset;
1350}
1351
1352static inline void
1353NCR_700_clear_fifo(struct Scsi_Host *host) {
1354 const struct NCR_700_Host_Parameters *hostdata
1355 = (struct NCR_700_Host_Parameters *)host->hostdata[0];
1356 if(hostdata->chip710) {
1357 NCR_700_writeb(CLR_FIFO_710, host, CTEST8_REG);
1358 } else {
1359 NCR_700_writeb(CLR_FIFO, host, DFIFO_REG);
1360 }
1361}
1362
1363static inline void
1364NCR_700_flush_fifo(struct Scsi_Host *host) {
1365 const struct NCR_700_Host_Parameters *hostdata
1366 = (struct NCR_700_Host_Parameters *)host->hostdata[0];
1367 if(hostdata->chip710) {
1368 NCR_700_writeb(FLUSH_DMA_FIFO_710, host, CTEST8_REG);
1369 udelay(10);
1370 NCR_700_writeb(0, host, CTEST8_REG);
1371 } else {
1372 NCR_700_writeb(FLUSH_DMA_FIFO, host, DFIFO_REG);
1373 udelay(10);
1374 NCR_700_writeb(0, host, DFIFO_REG);
1375 }
1376}
1377
1378
1379/* The queue lock with interrupts disabled must be held on entry to
1380 * this function */
1381STATIC int
1382NCR_700_start_command(struct scsi_cmnd *SCp)
1383{
1384 struct NCR_700_command_slot *slot =
1385 (struct NCR_700_command_slot *)SCp->host_scribble;
1386 struct NCR_700_Host_Parameters *hostdata =
1387 (struct NCR_700_Host_Parameters *)SCp->device->host->hostdata[0];
1388 __u16 count = 1; /* for IDENTIFY message */
1389
1390 if(hostdata->state != NCR_700_HOST_FREE) {
1391 /* keep this inside the lock to close the race window where
1392 * the running command finishes on another CPU while we don't
1393 * change the state to queued on this one */
1394 slot->state = NCR_700_SLOT_QUEUED;
1395
1396 DEBUG(("scsi%d: host busy, queueing command %p, slot %p\n",
1397 SCp->device->host->host_no, slot->cmnd, slot));
1398 return 0;
1399 }
1400 hostdata->state = NCR_700_HOST_BUSY;
1401 hostdata->cmd = SCp;
1402 slot->state = NCR_700_SLOT_BUSY;
1403 /* keep interrupts disabled until we have the command correctly
1404 * set up so we cannot take a selection interrupt */
1405
1406 hostdata->msgout[0] = NCR_700_identify(SCp->cmnd[0] != REQUEST_SENSE,
1407 SCp->device->lun);
1408 /* for INQUIRY or REQUEST_SENSE commands, we cannot be sure
1409 * if the negotiated transfer parameters still hold, so
1410 * always renegotiate them */
1411 if(SCp->cmnd[0] == INQUIRY || SCp->cmnd[0] == REQUEST_SENSE) {
1412 NCR_700_clear_flag(SCp->device, NCR_700_DEV_NEGOTIATED_SYNC);
1413 }
1414
1415 /* REQUEST_SENSE is asking for contingent I_T_L(_Q) status.
1416 * If a contingent allegiance condition exists, the device
1417 * will refuse all tags, so send the request sense as untagged
1418 * */
422c0d61 1419 if((hostdata->tag_negotiated & (1<<scmd_id(SCp)))
1da177e4
LT
1420 && (slot->tag != SCSI_NO_TAG && SCp->cmnd[0] != REQUEST_SENSE)) {
1421 count += scsi_populate_tag_msg(SCp, &hostdata->msgout[count]);
1422 }
1423
1424 if(hostdata->fast &&
1425 NCR_700_is_flag_clear(SCp->device, NCR_700_DEV_NEGOTIATED_SYNC)) {
6ea3c0b2
MW
1426 count += spi_populate_sync_msg(&hostdata->msgout[count],
1427 spi_period(SCp->device->sdev_target),
1428 spi_offset(SCp->device->sdev_target));
1da177e4
LT
1429 NCR_700_set_flag(SCp->device, NCR_700_DEV_BEGIN_SYNC_NEGOTIATION);
1430 }
1431
1432 script_patch_16(hostdata->script, MessageCount, count);
1433
1434
1435 script_patch_ID(hostdata->script,
422c0d61 1436 Device_ID, 1<<scmd_id(SCp));
1da177e4
LT
1437
1438 script_patch_32_abs(hostdata->script, CommandAddress,
1439 slot->pCmd);
1440 script_patch_16(hostdata->script, CommandCount, SCp->cmd_len);
1441 /* finally plumb the beginning of the SG list into the script
1442 * */
1443 script_patch_32_abs(hostdata->script, SGScriptStartAddress,
1444 to32bit(&slot->pSG[0].ins));
1445 NCR_700_clear_fifo(SCp->device->host);
1446
1447 if(slot->resume_offset == 0)
1448 slot->resume_offset = hostdata->pScript;
1449 /* now perform all the writebacks and invalidates */
1450 dma_cache_sync(hostdata->msgout, count, DMA_TO_DEVICE);
1451 dma_cache_sync(hostdata->msgin, MSG_ARRAY_SIZE,
1452 DMA_FROM_DEVICE);
1453 dma_cache_sync(SCp->cmnd, SCp->cmd_len, DMA_TO_DEVICE);
1454 dma_cache_sync(hostdata->status, 1, DMA_FROM_DEVICE);
1455
1456 /* set the synchronous period/offset */
1457 NCR_700_writeb(NCR_700_get_SXFER(SCp->device),
1458 SCp->device->host, SXFER_REG);
1459 NCR_700_writel(slot->temp, SCp->device->host, TEMP_REG);
1460 NCR_700_writel(slot->resume_offset, SCp->device->host, DSP_REG);
1461
1462 return 1;
1463}
1464
1465irqreturn_t
1466NCR_700_intr(int irq, void *dev_id, struct pt_regs *regs)
1467{
1468 struct Scsi_Host *host = (struct Scsi_Host *)dev_id;
1469 struct NCR_700_Host_Parameters *hostdata =
1470 (struct NCR_700_Host_Parameters *)host->hostdata[0];
1471 __u8 istat;
1472 __u32 resume_offset = 0;
1473 __u8 pun = 0xff, lun = 0xff;
1474 unsigned long flags;
1475 int handled = 0;
1476
1477 /* Use the host lock to serialise acess to the 53c700
1478 * hardware. Note: In future, we may need to take the queue
1479 * lock to enter the done routines. When that happens, we
1480 * need to ensure that for this driver, the host lock and the
1481 * queue lock point to the same thing. */
1482 spin_lock_irqsave(host->host_lock, flags);
1483 if((istat = NCR_700_readb(host, ISTAT_REG))
1484 & (SCSI_INT_PENDING | DMA_INT_PENDING)) {
1485 __u32 dsps;
1486 __u8 sstat0 = 0, dstat = 0;
1487 __u32 dsp;
1488 struct scsi_cmnd *SCp = hostdata->cmd;
1489 enum NCR_700_Host_State state;
1490
1491 handled = 1;
1492 state = hostdata->state;
1493 SCp = hostdata->cmd;
1494
1495 if(istat & SCSI_INT_PENDING) {
1496 udelay(10);
1497
1498 sstat0 = NCR_700_readb(host, SSTAT0_REG);
1499 }
1500
1501 if(istat & DMA_INT_PENDING) {
1502 udelay(10);
1503
1504 dstat = NCR_700_readb(host, DSTAT_REG);
1505 }
1506
1507 dsps = NCR_700_readl(host, DSPS_REG);
1508 dsp = NCR_700_readl(host, DSP_REG);
1509
1510 DEBUG(("scsi%d: istat %02x sstat0 %02x dstat %02x dsp %04x[%08x] dsps 0x%x\n",
1511 host->host_no, istat, sstat0, dstat,
1512 (dsp - (__u32)(hostdata->pScript))/4,
1513 dsp, dsps));
1514
1515 if(SCp != NULL) {
1516 pun = SCp->device->id;
1517 lun = SCp->device->lun;
1518 }
1519
1520 if(sstat0 & SCSI_RESET_DETECTED) {
1521 struct scsi_device *SDp;
1522 int i;
1523
1524 hostdata->state = NCR_700_HOST_BUSY;
1525
1526 printk(KERN_ERR "scsi%d: Bus Reset detected, executing command %p, slot %p, dsp %08x[%04x]\n",
1527 host->host_no, SCp, SCp == NULL ? NULL : SCp->host_scribble, dsp, dsp - hostdata->pScript);
1528
1529 scsi_report_bus_reset(host, 0);
1530
1531 /* clear all the negotiated parameters */
1532 __shost_for_each_device(SDp, host)
1533 SDp->hostdata = NULL;
1534
1535 /* clear all the slots and their pending commands */
1536 for(i = 0; i < NCR_700_COMMAND_SLOTS_PER_HOST; i++) {
1537 struct scsi_cmnd *SCp;
1538 struct NCR_700_command_slot *slot =
1539 &hostdata->slots[i];
1540
1541 if(slot->state == NCR_700_SLOT_FREE)
1542 continue;
1543
1544 SCp = slot->cmnd;
1545 printk(KERN_ERR " failing command because of reset, slot %p, cmnd %p\n",
1546 slot, SCp);
1547 free_slot(slot, hostdata);
1548 SCp->host_scribble = NULL;
1549 NCR_700_set_depth(SCp->device, 0);
1550 /* NOTE: deadlock potential here: we
1551 * rely on mid-layer guarantees that
1552 * scsi_done won't try to issue the
1553 * command again otherwise we'll
1554 * deadlock on the
1555 * hostdata->state_lock */
1556 SCp->result = DID_RESET << 16;
1557 SCp->scsi_done(SCp);
1558 }
1559 mdelay(25);
1560 NCR_700_chip_setup(host);
1561
1562 hostdata->state = NCR_700_HOST_FREE;
1563 hostdata->cmd = NULL;
1564 /* signal back if this was an eh induced reset */
1565 if(hostdata->eh_complete != NULL)
1566 complete(hostdata->eh_complete);
1567 goto out_unlock;
1568 } else if(sstat0 & SELECTION_TIMEOUT) {
1569 DEBUG(("scsi%d: (%d:%d) selection timeout\n",
1570 host->host_no, pun, lun));
1571 NCR_700_scsi_done(hostdata, SCp, DID_NO_CONNECT<<16);
1572 } else if(sstat0 & PHASE_MISMATCH) {
1573 struct NCR_700_command_slot *slot = (SCp == NULL) ? NULL :
1574 (struct NCR_700_command_slot *)SCp->host_scribble;
1575
1576 if(dsp == Ent_SendMessage + 8 + hostdata->pScript) {
1577 /* It wants to reply to some part of
1578 * our message */
1579#ifdef NCR_700_DEBUG
1580 __u32 temp = NCR_700_readl(host, TEMP_REG);
1581 int count = (hostdata->script[Ent_SendMessage/4] & 0xffffff) - ((NCR_700_readl(host, DBC_REG) & 0xffffff) + NCR_700_data_residual(host));
1582 printk("scsi%d (%d:%d) PHASE MISMATCH IN SEND MESSAGE %d remain, return %p[%04x], phase %s\n", host->host_no, pun, lun, count, (void *)temp, temp - hostdata->pScript, sbcl_to_string(NCR_700_readb(host, SBCL_REG)));
1583#endif
1584 resume_offset = hostdata->pScript + Ent_SendMessagePhaseMismatch;
1585 } else if(dsp >= to32bit(&slot->pSG[0].ins) &&
1586 dsp <= to32bit(&slot->pSG[NCR_700_SG_SEGMENTS].ins)) {
1587 int data_transfer = NCR_700_readl(host, DBC_REG) & 0xffffff;
1588 int SGcount = (dsp - to32bit(&slot->pSG[0].ins))/sizeof(struct NCR_700_SG_List);
1589 int residual = NCR_700_data_residual(host);
1590 int i;
1591#ifdef NCR_700_DEBUG
1592 __u32 naddr = NCR_700_readl(host, DNAD_REG);
1593
1594 printk("scsi%d: (%d:%d) Expected phase mismatch in slot->SG[%d], transferred 0x%x\n",
1595 host->host_no, pun, lun,
1596 SGcount, data_transfer);
1597 scsi_print_command(SCp);
1598 if(residual) {
1599 printk("scsi%d: (%d:%d) Expected phase mismatch in slot->SG[%d], transferred 0x%x, residual %d\n",
1600 host->host_no, pun, lun,
1601 SGcount, data_transfer, residual);
1602 }
1603#endif
1604 data_transfer += residual;
1605
1606 if(data_transfer != 0) {
1607 int count;
1608 __u32 pAddr;
1609
1610 SGcount--;
1611
1612 count = (bS_to_cpu(slot->SG[SGcount].ins) & 0x00ffffff);
1613 DEBUG(("DATA TRANSFER MISMATCH, count = %d, transferred %d\n", count, count-data_transfer));
1614 slot->SG[SGcount].ins &= bS_to_host(0xff000000);
1615 slot->SG[SGcount].ins |= bS_to_host(data_transfer);
1616 pAddr = bS_to_cpu(slot->SG[SGcount].pAddr);
1617 pAddr += (count - data_transfer);
1618#ifdef NCR_700_DEBUG
1619 if(pAddr != naddr) {
1620 printk("scsi%d (%d:%d) transfer mismatch pAddr=%lx, naddr=%lx, data_transfer=%d, residual=%d\n", host->host_no, pun, lun, (unsigned long)pAddr, (unsigned long)naddr, data_transfer, residual);
1621 }
1622#endif
1623 slot->SG[SGcount].pAddr = bS_to_host(pAddr);
1624 }
1625 /* set the executed moves to nops */
1626 for(i=0; i<SGcount; i++) {
1627 slot->SG[i].ins = bS_to_host(SCRIPT_NOP);
1628 slot->SG[i].pAddr = 0;
1629 }
1630 dma_cache_sync(slot->SG, sizeof(slot->SG), DMA_TO_DEVICE);
1631 /* and pretend we disconnected after
1632 * the command phase */
1633 resume_offset = hostdata->pScript + Ent_MsgInDuringData;
1634 /* make sure all the data is flushed */
1635 NCR_700_flush_fifo(host);
1636 } else {
1637 __u8 sbcl = NCR_700_readb(host, SBCL_REG);
1638 printk(KERN_ERR "scsi%d: (%d:%d) phase mismatch at %04x, phase %s\n",
1639 host->host_no, pun, lun, dsp - hostdata->pScript, sbcl_to_string(sbcl));
1640 NCR_700_internal_bus_reset(host);
1641 }
1642
1643 } else if(sstat0 & SCSI_GROSS_ERROR) {
1644 printk(KERN_ERR "scsi%d: (%d:%d) GROSS ERROR\n",
1645 host->host_no, pun, lun);
1646 NCR_700_scsi_done(hostdata, SCp, DID_ERROR<<16);
1647 } else if(sstat0 & PARITY_ERROR) {
1648 printk(KERN_ERR "scsi%d: (%d:%d) PARITY ERROR\n",
1649 host->host_no, pun, lun);
1650 NCR_700_scsi_done(hostdata, SCp, DID_ERROR<<16);
1651 } else if(dstat & SCRIPT_INT_RECEIVED) {
1652 DEBUG(("scsi%d: (%d:%d) ====>SCRIPT INTERRUPT<====\n",
1653 host->host_no, pun, lun));
1654 resume_offset = process_script_interrupt(dsps, dsp, SCp, host, hostdata);
1655 } else if(dstat & (ILGL_INST_DETECTED)) {
1656 printk(KERN_ERR "scsi%d: (%d:%d) Illegal Instruction detected at 0x%08x[0x%x]!!!\n"
1657 " Please email James.Bottomley@HansenPartnership.com with the details\n",
1658 host->host_no, pun, lun,
1659 dsp, dsp - hostdata->pScript);
1660 NCR_700_scsi_done(hostdata, SCp, DID_ERROR<<16);
1661 } else if(dstat & (WATCH_DOG_INTERRUPT|ABORTED)) {
1662 printk(KERN_ERR "scsi%d: (%d:%d) serious DMA problem, dstat=%02x\n",
1663 host->host_no, pun, lun, dstat);
1664 NCR_700_scsi_done(hostdata, SCp, DID_ERROR<<16);
1665 }
1666
1667
1668 /* NOTE: selection interrupt processing MUST occur
1669 * after script interrupt processing to correctly cope
1670 * with the case where we process a disconnect and
1671 * then get reselected before we process the
1672 * disconnection */
1673 if(sstat0 & SELECTED) {
1674 /* FIXME: It currently takes at least FOUR
1675 * interrupts to complete a command that
1676 * disconnects: one for the disconnect, one
1677 * for the reselection, one to get the
1678 * reselection data and one to complete the
1679 * command. If we guess the reselected
1680 * command here and prepare it, we only need
1681 * to get a reselection data interrupt if we
1682 * guessed wrongly. Since the interrupt
1683 * overhead is much greater than the command
1684 * setup, this would be an efficient
1685 * optimisation particularly as we probably
1686 * only have one outstanding command on a
1687 * target most of the time */
1688
1689 resume_offset = process_selection(host, dsp);
1690
1691 }
1692
1693 }
1694
1695 if(resume_offset) {
1696 if(hostdata->state != NCR_700_HOST_BUSY) {
1697 printk(KERN_ERR "scsi%d: Driver error: resume at 0x%08x [0x%04x] with non busy host!\n",
1698 host->host_no, resume_offset, resume_offset - hostdata->pScript);
1699 hostdata->state = NCR_700_HOST_BUSY;
1700 }
1701
1702 DEBUG(("Attempting to resume at %x\n", resume_offset));
1703 NCR_700_clear_fifo(host);
1704 NCR_700_writel(resume_offset, host, DSP_REG);
1705 }
1706 /* There is probably a technical no-no about this: If we're a
1707 * shared interrupt and we got this interrupt because the
1708 * other device needs servicing not us, we're still going to
1709 * check our queued commands here---of course, there shouldn't
1710 * be any outstanding.... */
1711 if(hostdata->state == NCR_700_HOST_FREE) {
1712 int i;
1713
1714 for(i = 0; i < NCR_700_COMMAND_SLOTS_PER_HOST; i++) {
1715 /* fairness: always run the queue from the last
1716 * position we left off */
1717 int j = (i + hostdata->saved_slot_position)
1718 % NCR_700_COMMAND_SLOTS_PER_HOST;
1719
1720 if(hostdata->slots[j].state != NCR_700_SLOT_QUEUED)
1721 continue;
1722 if(NCR_700_start_command(hostdata->slots[j].cmnd)) {
1723 DEBUG(("scsi%d: Issuing saved command slot %p, cmd %p\t\n",
1724 host->host_no, &hostdata->slots[j],
1725 hostdata->slots[j].cmnd));
1726 hostdata->saved_slot_position = j + 1;
1727 }
1728
1729 break;
1730 }
1731 }
1732 out_unlock:
1733 spin_unlock_irqrestore(host->host_lock, flags);
1734 return IRQ_RETVAL(handled);
1735}
1736
1737STATIC int
1738NCR_700_queuecommand(struct scsi_cmnd *SCp, void (*done)(struct scsi_cmnd *))
1739{
1740 struct NCR_700_Host_Parameters *hostdata =
1741 (struct NCR_700_Host_Parameters *)SCp->device->host->hostdata[0];
1742 __u32 move_ins;
1743 enum dma_data_direction direction;
1744 struct NCR_700_command_slot *slot;
1745
1746 if(hostdata->command_slot_count >= NCR_700_COMMAND_SLOTS_PER_HOST) {
1747 /* We're over our allocation, this should never happen
1748 * since we report the max allocation to the mid layer */
1749 printk(KERN_WARNING "scsi%d: Command depth has gone over queue depth\n", SCp->device->host->host_no);
1750 return 1;
1751 }
1752 /* check for untagged commands. We cannot have any outstanding
1753 * commands if we accept them. Commands could be untagged because:
1754 *
1755 * - The tag negotiated bitmap is clear
1756 * - The blk layer sent and untagged command
1757 */
1758 if(NCR_700_get_depth(SCp->device) != 0
017560fc 1759 && (!(hostdata->tag_negotiated & (1<<scmd_id(SCp)))
1da177e4 1760 || !blk_rq_tagged(SCp->request))) {
017560fc
JG
1761 CDEBUG(KERN_ERR, SCp, "has non zero depth %d\n",
1762 NCR_700_get_depth(SCp->device));
1da177e4
LT
1763 return SCSI_MLQUEUE_DEVICE_BUSY;
1764 }
1765 if(NCR_700_get_depth(SCp->device) >= SCp->device->queue_depth) {
017560fc
JG
1766 CDEBUG(KERN_ERR, SCp, "has max tag depth %d\n",
1767 NCR_700_get_depth(SCp->device));
1da177e4
LT
1768 return SCSI_MLQUEUE_DEVICE_BUSY;
1769 }
1770 NCR_700_set_depth(SCp->device, NCR_700_get_depth(SCp->device) + 1);
1771
1772 /* begin the command here */
1773 /* no need to check for NULL, test for command_slot_count above
1774 * ensures a slot is free */
1775 slot = find_empty_slot(hostdata);
1776
1777 slot->cmnd = SCp;
1778
1779 SCp->scsi_done = done;
1780 SCp->host_scribble = (unsigned char *)slot;
1781 SCp->SCp.ptr = NULL;
1782 SCp->SCp.buffer = NULL;
1783
1784#ifdef NCR_700_DEBUG
1785 printk("53c700: scsi%d, command ", SCp->device->host->host_no);
1786 scsi_print_command(SCp);
1787#endif
1788 if(blk_rq_tagged(SCp->request)
017560fc 1789 && (hostdata->tag_negotiated &(1<<scmd_id(SCp))) == 0
1da177e4 1790 && NCR_700_get_tag_neg_state(SCp->device) == NCR_700_START_TAG_NEGOTIATION) {
017560fc
JG
1791 scmd_printk(KERN_ERR, SCp, "Enabling Tag Command Queuing\n");
1792 hostdata->tag_negotiated |= (1<<scmd_id(SCp));
1da177e4
LT
1793 NCR_700_set_tag_neg_state(SCp->device, NCR_700_DURING_TAG_NEGOTIATION);
1794 }
1795
1796 /* here we may have to process an untagged command. The gate
1797 * above ensures that this will be the only one outstanding,
1798 * so clear the tag negotiated bit.
1799 *
1800 * FIXME: This will royally screw up on multiple LUN devices
1801 * */
1802 if(!blk_rq_tagged(SCp->request)
017560fc
JG
1803 && (hostdata->tag_negotiated &(1<<scmd_id(SCp)))) {
1804 scmd_printk(KERN_INFO, SCp, "Disabling Tag Command Queuing\n");
1805 hostdata->tag_negotiated &= ~(1<<scmd_id(SCp));
1da177e4
LT
1806 }
1807
017560fc 1808 if((hostdata->tag_negotiated &(1<<scmd_id(SCp)))
1da177e4
LT
1809 && scsi_get_tag_type(SCp->device)) {
1810 slot->tag = SCp->request->tag;
017560fc
JG
1811 CDEBUG(KERN_DEBUG, SCp, "sending out tag %d, slot %p\n",
1812 slot->tag, slot);
1da177e4
LT
1813 } else {
1814 slot->tag = SCSI_NO_TAG;
1815 /* must populate current_cmnd for scsi_find_tag to work */
1816 SCp->device->current_cmnd = SCp;
1817 }
1818 /* sanity check: some of the commands generated by the mid-layer
1819 * have an eccentric idea of their sc_data_direction */
1820 if(!SCp->use_sg && !SCp->request_bufflen
1821 && SCp->sc_data_direction != DMA_NONE) {
1822#ifdef NCR_700_DEBUG
1823 printk("53c700: Command");
1824 scsi_print_command(SCp);
1825 printk("Has wrong data direction %d\n", SCp->sc_data_direction);
1826#endif
1827 SCp->sc_data_direction = DMA_NONE;
1828 }
1829
1830 switch (SCp->cmnd[0]) {
1831 case REQUEST_SENSE:
1832 /* clear the internal sense magic */
1833 SCp->cmnd[6] = 0;
1834 /* fall through */
1835 default:
1836 /* OK, get it from the command */
1837 switch(SCp->sc_data_direction) {
1838 case DMA_BIDIRECTIONAL:
1839 default:
1840 printk(KERN_ERR "53c700: Unknown command for data direction ");
1841 scsi_print_command(SCp);
1842
1843 move_ins = 0;
1844 break;
1845 case DMA_NONE:
1846 move_ins = 0;
1847 break;
1848 case DMA_FROM_DEVICE:
1849 move_ins = SCRIPT_MOVE_DATA_IN;
1850 break;
1851 case DMA_TO_DEVICE:
1852 move_ins = SCRIPT_MOVE_DATA_OUT;
1853 break;
1854 }
1855 }
1856
1857 /* now build the scatter gather list */
1858 direction = SCp->sc_data_direction;
1859 if(move_ins != 0) {
1860 int i;
1861 int sg_count;
1862 dma_addr_t vPtr = 0;
1863 __u32 count = 0;
1864
1865 if(SCp->use_sg) {
1866 sg_count = dma_map_sg(hostdata->dev, SCp->buffer,
1867 SCp->use_sg, direction);
1868 } else {
1869 vPtr = dma_map_single(hostdata->dev,
1870 SCp->request_buffer,
1871 SCp->request_bufflen,
1872 direction);
1873 count = SCp->request_bufflen;
1874 slot->dma_handle = vPtr;
1875 sg_count = 1;
1876 }
1877
1878
1879 for(i = 0; i < sg_count; i++) {
1880
1881 if(SCp->use_sg) {
1882 struct scatterlist *sg = SCp->buffer;
1883
1884 vPtr = sg_dma_address(&sg[i]);
1885 count = sg_dma_len(&sg[i]);
1886 }
1887
1888 slot->SG[i].ins = bS_to_host(move_ins | count);
1889 DEBUG((" scatter block %d: move %d[%08x] from 0x%lx\n",
1890 i, count, slot->SG[i].ins, (unsigned long)vPtr));
1891 slot->SG[i].pAddr = bS_to_host(vPtr);
1892 }
1893 slot->SG[i].ins = bS_to_host(SCRIPT_RETURN);
1894 slot->SG[i].pAddr = 0;
1895 dma_cache_sync(slot->SG, sizeof(slot->SG), DMA_TO_DEVICE);
1896 DEBUG((" SETTING %08lx to %x\n",
1897 (&slot->pSG[i].ins),
1898 slot->SG[i].ins));
1899 }
1900 slot->resume_offset = 0;
1901 slot->pCmd = dma_map_single(hostdata->dev, SCp->cmnd,
1902 sizeof(SCp->cmnd), DMA_TO_DEVICE);
1903 NCR_700_start_command(SCp);
1904 return 0;
1905}
1906
1907STATIC int
1908NCR_700_abort(struct scsi_cmnd * SCp)
1909{
1910 struct NCR_700_command_slot *slot;
1911
017560fc
JG
1912 scmd_printk(KERN_INFO, SCp,
1913 "New error handler wants to abort command\n\t");
1da177e4
LT
1914 scsi_print_command(SCp);
1915
1916 slot = (struct NCR_700_command_slot *)SCp->host_scribble;
1917
1918 if(slot == NULL)
1919 /* no outstanding command to abort */
1920 return SUCCESS;
1921 if(SCp->cmnd[0] == TEST_UNIT_READY) {
1922 /* FIXME: This is because of a problem in the new
1923 * error handler. When it is in error recovery, it
1924 * will send a TUR to a device it thinks may still be
1925 * showing a problem. If the TUR isn't responded to,
1926 * it will abort it and mark the device off line.
1927 * Unfortunately, it does no other error recovery, so
1928 * this would leave us with an outstanding command
1929 * occupying a slot. Rather than allow this to
1930 * happen, we issue a bus reset to force all
1931 * outstanding commands to terminate here. */
1932 NCR_700_internal_bus_reset(SCp->device->host);
1933 /* still drop through and return failed */
1934 }
1935 return FAILED;
1936
1937}
1938
1939STATIC int
1940NCR_700_bus_reset(struct scsi_cmnd * SCp)
1941{
1942 DECLARE_COMPLETION(complete);
1943 struct NCR_700_Host_Parameters *hostdata =
1944 (struct NCR_700_Host_Parameters *)SCp->device->host->hostdata[0];
1945
017560fc
JG
1946 scmd_printk(KERN_INFO, SCp,
1947 "New error handler wants BUS reset, cmd %p\n\t", SCp);
1da177e4 1948 scsi_print_command(SCp);
68b3aa7c 1949
1da177e4
LT
1950 /* In theory, eh_complete should always be null because the
1951 * eh is single threaded, but just in case we're handling a
1952 * reset via sg or something */
68b3aa7c
JG
1953 spin_lock_irq(SCp->device->host->host_lock);
1954 while (hostdata->eh_complete != NULL) {
1da177e4
LT
1955 spin_unlock_irq(SCp->device->host->host_lock);
1956 msleep_interruptible(100);
1957 spin_lock_irq(SCp->device->host->host_lock);
1958 }
68b3aa7c 1959
1da177e4
LT
1960 hostdata->eh_complete = &complete;
1961 NCR_700_internal_bus_reset(SCp->device->host);
68b3aa7c 1962
1da177e4
LT
1963 spin_unlock_irq(SCp->device->host->host_lock);
1964 wait_for_completion(&complete);
1965 spin_lock_irq(SCp->device->host->host_lock);
68b3aa7c 1966
1da177e4
LT
1967 hostdata->eh_complete = NULL;
1968 /* Revalidate the transport parameters of the failing device */
1969 if(hostdata->fast)
1970 spi_schedule_dv_device(SCp->device);
68b3aa7c
JG
1971
1972 spin_unlock_irq(SCp->device->host->host_lock);
1da177e4
LT
1973 return SUCCESS;
1974}
1975
1da177e4
LT
1976STATIC int
1977NCR_700_host_reset(struct scsi_cmnd * SCp)
1978{
017560fc 1979 scmd_printk(KERN_INFO, SCp, "New error handler wants HOST reset\n\t");
1da177e4
LT
1980 scsi_print_command(SCp);
1981
df0ae249
JG
1982 spin_lock_irq(SCp->device->host->host_lock);
1983
1da177e4
LT
1984 NCR_700_internal_bus_reset(SCp->device->host);
1985 NCR_700_chip_reset(SCp->device->host);
df0ae249
JG
1986
1987 spin_unlock_irq(SCp->device->host->host_lock);
1988
1da177e4
LT
1989 return SUCCESS;
1990}
1991
1992STATIC void
1993NCR_700_set_period(struct scsi_target *STp, int period)
1994{
1995 struct Scsi_Host *SHp = dev_to_shost(STp->dev.parent);
1996 struct NCR_700_Host_Parameters *hostdata =
1997 (struct NCR_700_Host_Parameters *)SHp->hostdata[0];
1998
1999 if(!hostdata->fast)
2000 return;
2001
2002 if(period < hostdata->min_period)
2003 period = hostdata->min_period;
2004
2005 spi_period(STp) = period;
2006 spi_flags(STp) &= ~(NCR_700_DEV_NEGOTIATED_SYNC |
2007 NCR_700_DEV_BEGIN_SYNC_NEGOTIATION);
2008 spi_flags(STp) |= NCR_700_DEV_PRINT_SYNC_NEGOTIATION;
2009}
2010
2011STATIC void
2012NCR_700_set_offset(struct scsi_target *STp, int offset)
2013{
2014 struct Scsi_Host *SHp = dev_to_shost(STp->dev.parent);
2015 struct NCR_700_Host_Parameters *hostdata =
2016 (struct NCR_700_Host_Parameters *)SHp->hostdata[0];
2017 int max_offset = hostdata->chip710
2018 ? NCR_710_MAX_OFFSET : NCR_700_MAX_OFFSET;
2019
2020 if(!hostdata->fast)
2021 return;
2022
2023 if(offset > max_offset)
2024 offset = max_offset;
2025
2026 /* if we're currently async, make sure the period is reasonable */
2027 if(spi_offset(STp) == 0 && (spi_period(STp) < hostdata->min_period ||
2028 spi_period(STp) > 0xff))
2029 spi_period(STp) = hostdata->min_period;
2030
2031 spi_offset(STp) = offset;
2032 spi_flags(STp) &= ~(NCR_700_DEV_NEGOTIATED_SYNC |
2033 NCR_700_DEV_BEGIN_SYNC_NEGOTIATION);
2034 spi_flags(STp) |= NCR_700_DEV_PRINT_SYNC_NEGOTIATION;
2035}
2036
2037
2038
2039STATIC int
2040NCR_700_slave_configure(struct scsi_device *SDp)
2041{
2042 struct NCR_700_Host_Parameters *hostdata =
2043 (struct NCR_700_Host_Parameters *)SDp->host->hostdata[0];
2044
2045 /* to do here: allocate memory; build a queue_full list */
2046 if(SDp->tagged_supported) {
2047 scsi_set_tag_type(SDp, MSG_ORDERED_TAG);
2048 scsi_activate_tcq(SDp, NCR_700_DEFAULT_TAGS);
2049 NCR_700_set_tag_neg_state(SDp, NCR_700_START_TAG_NEGOTIATION);
2050 } else {
2051 /* initialise to default depth */
2052 scsi_adjust_queue_depth(SDp, 0, SDp->host->cmd_per_lun);
2053 }
2054 if(hostdata->fast) {
2055 /* Find the correct offset and period via domain validation */
2056 if (!spi_initial_dv(SDp->sdev_target))
2057 spi_dv_device(SDp);
2058 } else {
2059 spi_offset(SDp->sdev_target) = 0;
2060 spi_period(SDp->sdev_target) = 0;
2061 }
2062 return 0;
2063}
2064
2065STATIC void
2066NCR_700_slave_destroy(struct scsi_device *SDp)
2067{
2068 /* to do here: deallocate memory */
2069}
2070
2071static int
2072NCR_700_change_queue_depth(struct scsi_device *SDp, int depth)
2073{
2074 if (depth > NCR_700_MAX_TAGS)
2075 depth = NCR_700_MAX_TAGS;
2076
2077 scsi_adjust_queue_depth(SDp, scsi_get_tag_type(SDp), depth);
2078 return depth;
2079}
2080
2081static int NCR_700_change_queue_type(struct scsi_device *SDp, int tag_type)
2082{
2083 int change_tag = ((tag_type ==0 && scsi_get_tag_type(SDp) != 0)
2084 || (tag_type != 0 && scsi_get_tag_type(SDp) == 0));
2085 struct NCR_700_Host_Parameters *hostdata =
2086 (struct NCR_700_Host_Parameters *)SDp->host->hostdata[0];
2087
2088 scsi_set_tag_type(SDp, tag_type);
2089
2090 /* We have a global (per target) flag to track whether TCQ is
2091 * enabled, so we'll be turning it off for the entire target here.
2092 * our tag algorithm will fail if we mix tagged and untagged commands,
2093 * so quiesce the device before doing this */
2094 if (change_tag)
2095 scsi_target_quiesce(SDp->sdev_target);
2096
2097 if (!tag_type) {
2098 /* shift back to the default unqueued number of commands
2099 * (the user can still raise this) */
2100 scsi_deactivate_tcq(SDp, SDp->host->cmd_per_lun);
422c0d61 2101 hostdata->tag_negotiated &= ~(1 << sdev_id(SDp));
1da177e4
LT
2102 } else {
2103 /* Here, we cleared the negotiation flag above, so this
2104 * will force the driver to renegotiate */
2105 scsi_activate_tcq(SDp, SDp->queue_depth);
2106 if (change_tag)
2107 NCR_700_set_tag_neg_state(SDp, NCR_700_START_TAG_NEGOTIATION);
2108 }
2109 if (change_tag)
2110 scsi_target_resume(SDp->sdev_target);
2111
2112 return tag_type;
2113}
2114
2115static ssize_t
10523b3b 2116NCR_700_show_active_tags(struct device *dev, struct device_attribute *attr, char *buf)
1da177e4
LT
2117{
2118 struct scsi_device *SDp = to_scsi_device(dev);
2119
2120 return snprintf(buf, 20, "%d\n", NCR_700_get_depth(SDp));
2121}
2122
2123static struct device_attribute NCR_700_active_tags_attr = {
2124 .attr = {
2125 .name = "active_tags",
2126 .mode = S_IRUGO,
2127 },
2128 .show = NCR_700_show_active_tags,
2129};
2130
2131STATIC struct device_attribute *NCR_700_dev_attrs[] = {
2132 &NCR_700_active_tags_attr,
2133 NULL,
2134};
2135
2136EXPORT_SYMBOL(NCR_700_detect);
2137EXPORT_SYMBOL(NCR_700_release);
2138EXPORT_SYMBOL(NCR_700_intr);
2139
2140static struct spi_function_template NCR_700_transport_functions = {
2141 .set_period = NCR_700_set_period,
2142 .show_period = 1,
2143 .set_offset = NCR_700_set_offset,
2144 .show_offset = 1,
2145};
2146
2147static int __init NCR_700_init(void)
2148{
2149 NCR_700_transport_template = spi_attach_transport(&NCR_700_transport_functions);
2150 if(!NCR_700_transport_template)
2151 return -ENODEV;
2152 return 0;
2153}
2154
2155static void __exit NCR_700_exit(void)
2156{
2157 spi_release_transport(NCR_700_transport_template);
2158}
2159
2160module_init(NCR_700_init);
2161module_exit(NCR_700_exit);
2162