dma-mapping: add the device argument to dma_mapping_error()
[GitHub/mt8127/android_kernel_alcatel_ttab.git] / drivers / scsi / ibmvscsi / ibmvstgt.c
CommitLineData
0e5d030b
FT
1/*
2 * IBM eServer i/pSeries Virtual SCSI Target Driver
3 * Copyright (C) 2003-2005 Dave Boutcher (boutcher@us.ibm.com) IBM Corp.
4 * Santiago Leon (santil@us.ibm.com) IBM Corp.
5 * Linda Xie (lxie@us.ibm.com) IBM Corp.
6 *
7 * Copyright (C) 2005-2006 FUJITA Tomonori <tomof@acm.org>
8 *
9 * This program is free software; you can redistribute it and/or modify
10 * it under the terms of the GNU General Public License as published by
11 * the Free Software Foundation; either version 2 of the License, or
12 * (at your option) any later version.
13 *
14 * This program is distributed in the hope that it will be useful,
15 * but WITHOUT ANY WARRANTY; without even the implied warranty of
16 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
17 * GNU General Public License for more details.
18 *
19 * You should have received a copy of the GNU General Public License
20 * along with this program; if not, write to the Free Software
21 * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307
22 * USA
23 */
24#include <linux/interrupt.h>
25#include <linux/module.h>
26#include <scsi/scsi.h>
27#include <scsi/scsi_host.h>
17b0bcfa 28#include <scsi/scsi_transport_srp.h>
0e5d030b
FT
29#include <scsi/scsi_tgt.h>
30#include <scsi/libsrp.h>
31#include <asm/hvcall.h>
32#include <asm/iommu.h>
33#include <asm/prom.h>
34#include <asm/vio.h>
35
36#include "ibmvscsi.h"
37
38#define INITIAL_SRP_LIMIT 16
ad1331a7 39#define DEFAULT_MAX_SECTORS 256
0e5d030b
FT
40
41#define TGT_NAME "ibmvstgt"
42
43/*
44 * Hypervisor calls.
45 */
46#define h_copy_rdma(l, sa, sb, da, db) \
47 plpar_hcall_norets(H_COPY_RDMA, l, sa, sb, da, db)
48#define h_send_crq(ua, l, h) \
49 plpar_hcall_norets(H_SEND_CRQ, ua, l, h)
50#define h_reg_crq(ua, tok, sz)\
51 plpar_hcall_norets(H_REG_CRQ, ua, tok, sz);
52#define h_free_crq(ua) \
53 plpar_hcall_norets(H_FREE_CRQ, ua);
54
55/* tmp - will replace with SCSI logging stuff */
56#define eprintk(fmt, args...) \
57do { \
58 printk("%s(%d) " fmt, __FUNCTION__, __LINE__, ##args); \
59} while (0)
60/* #define dprintk eprintk */
61#define dprintk(fmt, args...)
62
63struct vio_port {
64 struct vio_dev *dma_dev;
65
66 struct crq_queue crq_queue;
67 struct work_struct crq_work;
68
69 unsigned long liobn;
70 unsigned long riobn;
2fdb611d 71 struct srp_target *target;
17b0bcfa
FT
72
73 struct srp_rport *rport;
0e5d030b
FT
74};
75
76static struct workqueue_struct *vtgtd;
17b0bcfa 77static struct scsi_transport_template *ibmvstgt_transport_template;
0e5d030b
FT
78
79/*
80 * These are fixed for the system and come from the Open Firmware device tree.
81 * We just store them here to save getting them every time.
82 */
83static char system_id[64] = "";
84static char partition_name[97] = "UNKNOWN";
85static unsigned int partition_number = -1;
86
87static struct vio_port *target_to_port(struct srp_target *target)
88{
89 return (struct vio_port *) target->ldata;
90}
91
92static inline union viosrp_iu *vio_iu(struct iu_entry *iue)
93{
94 return (union viosrp_iu *) (iue->sbuf->buf);
95}
96
97static int send_iu(struct iu_entry *iue, uint64_t length, uint8_t format)
98{
99 struct srp_target *target = iue->target;
100 struct vio_port *vport = target_to_port(target);
101 long rc, rc1;
102 union {
103 struct viosrp_crq cooked;
104 uint64_t raw[2];
105 } crq;
106
107 /* First copy the SRP */
108 rc = h_copy_rdma(length, vport->liobn, iue->sbuf->dma,
109 vport->riobn, iue->remote_token);
110
111 if (rc)
112 eprintk("Error %ld transferring data\n", rc);
113
114 crq.cooked.valid = 0x80;
115 crq.cooked.format = format;
116 crq.cooked.reserved = 0x00;
117 crq.cooked.timeout = 0x00;
118 crq.cooked.IU_length = length;
119 crq.cooked.IU_data_ptr = vio_iu(iue)->srp.rsp.tag;
120
121 if (rc == 0)
122 crq.cooked.status = 0x99; /* Just needs to be non-zero */
123 else
124 crq.cooked.status = 0x00;
125
126 rc1 = h_send_crq(vport->dma_dev->unit_address, crq.raw[0], crq.raw[1]);
127
128 if (rc1) {
129 eprintk("%ld sending response\n", rc1);
130 return rc1;
131 }
132
133 return rc;
134}
135
136#define SRP_RSP_SENSE_DATA_LEN 18
137
138static int send_rsp(struct iu_entry *iue, struct scsi_cmnd *sc,
139 unsigned char status, unsigned char asc)
140{
141 union viosrp_iu *iu = vio_iu(iue);
142 uint64_t tag = iu->srp.rsp.tag;
143
144 /* If the linked bit is on and status is good */
145 if (test_bit(V_LINKED, &iue->flags) && (status == NO_SENSE))
146 status = 0x10;
147
148 memset(iu, 0, sizeof(struct srp_rsp));
149 iu->srp.rsp.opcode = SRP_RSP;
150 iu->srp.rsp.req_lim_delta = 1;
151 iu->srp.rsp.tag = tag;
152
153 if (test_bit(V_DIOVER, &iue->flags))
154 iu->srp.rsp.flags |= SRP_RSP_FLAG_DIOVER;
155
156 iu->srp.rsp.data_in_res_cnt = 0;
157 iu->srp.rsp.data_out_res_cnt = 0;
158
159 iu->srp.rsp.flags &= ~SRP_RSP_FLAG_RSPVALID;
160
161 iu->srp.rsp.resp_data_len = 0;
162 iu->srp.rsp.status = status;
163 if (status) {
164 uint8_t *sense = iu->srp.rsp.data;
165
166 if (sc) {
167 iu->srp.rsp.flags |= SRP_RSP_FLAG_SNSVALID;
168 iu->srp.rsp.sense_data_len = SCSI_SENSE_BUFFERSIZE;
169 memcpy(sense, sc->sense_buffer, SCSI_SENSE_BUFFERSIZE);
170 } else {
171 iu->srp.rsp.status = SAM_STAT_CHECK_CONDITION;
172 iu->srp.rsp.flags |= SRP_RSP_FLAG_SNSVALID;
173 iu->srp.rsp.sense_data_len = SRP_RSP_SENSE_DATA_LEN;
174
175 /* Valid bit and 'current errors' */
176 sense[0] = (0x1 << 7 | 0x70);
177 /* Sense key */
178 sense[2] = status;
179 /* Additional sense length */
180 sense[7] = 0xa; /* 10 bytes */
181 /* Additional sense code */
182 sense[12] = asc;
183 }
184 }
185
186 send_iu(iue, sizeof(iu->srp.rsp) + SRP_RSP_SENSE_DATA_LEN,
187 VIOSRP_SRP_FORMAT);
188
189 return 0;
190}
191
192static void handle_cmd_queue(struct srp_target *target)
193{
194 struct Scsi_Host *shost = target->shost;
17b0bcfa 195 struct srp_rport *rport = target_to_port(target)->rport;
0e5d030b
FT
196 struct iu_entry *iue;
197 struct srp_cmd *cmd;
198 unsigned long flags;
199 int err;
200
201retry:
202 spin_lock_irqsave(&target->lock, flags);
203
204 list_for_each_entry(iue, &target->cmd_queue, ilist) {
205 if (!test_and_set_bit(V_FLYING, &iue->flags)) {
206 spin_unlock_irqrestore(&target->lock, flags);
207 cmd = iue->sbuf->buf;
17b0bcfa
FT
208 err = srp_cmd_queue(shost, cmd, iue,
209 (unsigned long)rport, 0);
0e5d030b
FT
210 if (err) {
211 eprintk("cannot queue cmd %p %d\n", cmd, err);
212 srp_iu_put(iue);
213 }
214 goto retry;
215 }
216 }
217
218 spin_unlock_irqrestore(&target->lock, flags);
219}
220
221static int ibmvstgt_rdma(struct scsi_cmnd *sc, struct scatterlist *sg, int nsg,
222 struct srp_direct_buf *md, int nmd,
223 enum dma_data_direction dir, unsigned int rest)
224{
225 struct iu_entry *iue = (struct iu_entry *) sc->SCp.ptr;
226 struct srp_target *target = iue->target;
227 struct vio_port *vport = target_to_port(target);
228 dma_addr_t token;
229 long err;
230 unsigned int done = 0;
231 int i, sidx, soff;
232
233 sidx = soff = 0;
234 token = sg_dma_address(sg + sidx);
235
236 for (i = 0; i < nmd && rest; i++) {
237 unsigned int mdone, mlen;
238
239 mlen = min(rest, md[i].len);
240 for (mdone = 0; mlen;) {
241 int slen = min(sg_dma_len(sg + sidx) - soff, mlen);
242
243 if (dir == DMA_TO_DEVICE)
244 err = h_copy_rdma(slen,
245 vport->riobn,
246 md[i].va + mdone,
247 vport->liobn,
248 token + soff);
249 else
250 err = h_copy_rdma(slen,
251 vport->liobn,
252 token + soff,
253 vport->riobn,
254 md[i].va + mdone);
255
256 if (err != H_SUCCESS) {
ad1331a7
BB
257 eprintk("rdma error %d %d %ld\n", dir, slen, err);
258 return -EIO;
0e5d030b
FT
259 }
260
261 mlen -= slen;
262 mdone += slen;
263 soff += slen;
264 done += slen;
265
266 if (soff == sg_dma_len(sg + sidx)) {
267 sidx++;
268 soff = 0;
269 token = sg_dma_address(sg + sidx);
270
271 if (sidx > nsg) {
272 eprintk("out of sg %p %d %d\n",
273 iue, sidx, nsg);
ad1331a7 274 return -EIO;
0e5d030b
FT
275 }
276 }
277 };
278
279 rest -= mlen;
280 }
0e5d030b
FT
281 return 0;
282}
283
0e5d030b
FT
284static int ibmvstgt_cmd_done(struct scsi_cmnd *sc,
285 void (*done)(struct scsi_cmnd *))
286{
287 unsigned long flags;
288 struct iu_entry *iue = (struct iu_entry *) sc->SCp.ptr;
289 struct srp_target *target = iue->target;
ad1331a7 290 int err = 0;
0e5d030b 291
bc7e380a 292 dprintk("%p %p %x %u\n", iue, target, vio_iu(iue)->srp.cmd.cdb[0],
21f1e91d 293 scsi_sg_count(sc));
bc7e380a 294
f10ab66f 295 if (scsi_sg_count(sc))
ad1331a7 296 err = srp_transfer_data(sc, &vio_iu(iue)->srp.cmd, ibmvstgt_rdma, 1, 1);
0e5d030b
FT
297
298 spin_lock_irqsave(&target->lock, flags);
299 list_del(&iue->ilist);
300 spin_unlock_irqrestore(&target->lock, flags);
301
ad1331a7 302 if (err|| sc->result != SAM_STAT_GOOD) {
0e5d030b
FT
303 eprintk("operation failed %p %d %x\n",
304 iue, sc->result, vio_iu(iue)->srp.cmd.cdb[0]);
305 send_rsp(iue, sc, HARDWARE_ERROR, 0x00);
306 } else
307 send_rsp(iue, sc, NO_SENSE, 0x00);
308
309 done(sc);
310 srp_iu_put(iue);
311 return 0;
312}
313
314int send_adapter_info(struct iu_entry *iue,
315 dma_addr_t remote_buffer, uint16_t length)
316{
317 struct srp_target *target = iue->target;
318 struct vio_port *vport = target_to_port(target);
319 struct Scsi_Host *shost = target->shost;
320 dma_addr_t data_token;
321 struct mad_adapter_info_data *info;
322 int err;
323
324 info = dma_alloc_coherent(target->dev, sizeof(*info), &data_token,
325 GFP_KERNEL);
326 if (!info) {
327 eprintk("bad dma_alloc_coherent %p\n", target);
328 return 1;
329 }
330
331 /* Get remote info */
332 err = h_copy_rdma(sizeof(*info), vport->riobn, remote_buffer,
333 vport->liobn, data_token);
334 if (err == H_SUCCESS) {
335 dprintk("Client connect: %s (%d)\n",
336 info->partition_name, info->partition_number);
337 }
338
339 memset(info, 0, sizeof(*info));
340
341 strcpy(info->srp_version, "16.a");
342 strncpy(info->partition_name, partition_name,
343 sizeof(info->partition_name));
344 info->partition_number = partition_number;
345 info->mad_version = 1;
346 info->os_type = 2;
347 info->port_max_txu[0] = shost->hostt->max_sectors << 9;
348
349 /* Send our info to remote */
350 err = h_copy_rdma(sizeof(*info), vport->liobn, data_token,
351 vport->riobn, remote_buffer);
352
353 dma_free_coherent(target->dev, sizeof(*info), info, data_token);
354
355 if (err != H_SUCCESS) {
356 eprintk("Error sending adapter info %d\n", err);
357 return 1;
358 }
359
360 return 0;
361}
362
363static void process_login(struct iu_entry *iue)
364{
365 union viosrp_iu *iu = vio_iu(iue);
366 struct srp_login_rsp *rsp = &iu->srp.login_rsp;
367 uint64_t tag = iu->srp.rsp.tag;
17b0bcfa
FT
368 struct Scsi_Host *shost = iue->target->shost;
369 struct srp_target *target = host_to_srp_target(shost);
370 struct vio_port *vport = target_to_port(target);
371 struct srp_rport_identifiers ids;
372
373 memset(&ids, 0, sizeof(ids));
374 sprintf(ids.port_id, "%x", vport->dma_dev->unit_address);
375 ids.roles = SRP_RPORT_ROLE_INITIATOR;
376 if (!vport->rport)
377 vport->rport = srp_rport_add(shost, &ids);
0e5d030b
FT
378
379 /* TODO handle case that requested size is wrong and
380 * buffer format is wrong
381 */
382 memset(iu, 0, sizeof(struct srp_login_rsp));
383 rsp->opcode = SRP_LOGIN_RSP;
384 rsp->req_lim_delta = INITIAL_SRP_LIMIT;
385 rsp->tag = tag;
386 rsp->max_it_iu_len = sizeof(union srp_iu);
387 rsp->max_ti_iu_len = sizeof(union srp_iu);
388 /* direct and indirect */
389 rsp->buf_fmt = SRP_BUF_FORMAT_DIRECT | SRP_BUF_FORMAT_INDIRECT;
390
391 send_iu(iue, sizeof(*rsp), VIOSRP_SRP_FORMAT);
392}
393
394static inline void queue_cmd(struct iu_entry *iue)
395{
396 struct srp_target *target = iue->target;
397 unsigned long flags;
398
399 spin_lock_irqsave(&target->lock, flags);
400 list_add_tail(&iue->ilist, &target->cmd_queue);
401 spin_unlock_irqrestore(&target->lock, flags);
402}
403
404static int process_tsk_mgmt(struct iu_entry *iue)
405{
406 union viosrp_iu *iu = vio_iu(iue);
407 int fn;
408
409 dprintk("%p %u\n", iue, iu->srp.tsk_mgmt.tsk_mgmt_func);
410
411 switch (iu->srp.tsk_mgmt.tsk_mgmt_func) {
412 case SRP_TSK_ABORT_TASK:
413 fn = ABORT_TASK;
414 break;
415 case SRP_TSK_ABORT_TASK_SET:
416 fn = ABORT_TASK_SET;
417 break;
418 case SRP_TSK_CLEAR_TASK_SET:
419 fn = CLEAR_TASK_SET;
420 break;
421 case SRP_TSK_LUN_RESET:
422 fn = LOGICAL_UNIT_RESET;
423 break;
424 case SRP_TSK_CLEAR_ACA:
425 fn = CLEAR_ACA;
426 break;
427 default:
428 fn = 0;
429 }
430 if (fn)
17b0bcfa
FT
431 scsi_tgt_tsk_mgmt_request(iue->target->shost,
432 (unsigned long)iue->target->shost,
433 fn,
0e5d030b
FT
434 iu->srp.tsk_mgmt.task_tag,
435 (struct scsi_lun *) &iu->srp.tsk_mgmt.lun,
436 iue);
437 else
438 send_rsp(iue, NULL, ILLEGAL_REQUEST, 0x20);
439
440 return !fn;
441}
442
443static int process_mad_iu(struct iu_entry *iue)
444{
445 union viosrp_iu *iu = vio_iu(iue);
446 struct viosrp_adapter_info *info;
447 struct viosrp_host_config *conf;
448
449 switch (iu->mad.empty_iu.common.type) {
450 case VIOSRP_EMPTY_IU_TYPE:
451 eprintk("%s\n", "Unsupported EMPTY MAD IU");
452 break;
453 case VIOSRP_ERROR_LOG_TYPE:
454 eprintk("%s\n", "Unsupported ERROR LOG MAD IU");
455 iu->mad.error_log.common.status = 1;
456 send_iu(iue, sizeof(iu->mad.error_log), VIOSRP_MAD_FORMAT);
457 break;
458 case VIOSRP_ADAPTER_INFO_TYPE:
459 info = &iu->mad.adapter_info;
460 info->common.status = send_adapter_info(iue, info->buffer,
461 info->common.length);
462 send_iu(iue, sizeof(*info), VIOSRP_MAD_FORMAT);
463 break;
464 case VIOSRP_HOST_CONFIG_TYPE:
465 conf = &iu->mad.host_config;
466 conf->common.status = 1;
467 send_iu(iue, sizeof(*conf), VIOSRP_MAD_FORMAT);
468 break;
469 default:
470 eprintk("Unknown type %u\n", iu->srp.rsp.opcode);
471 }
472
473 return 1;
474}
475
476static int process_srp_iu(struct iu_entry *iue)
477{
478 union viosrp_iu *iu = vio_iu(iue);
479 int done = 1;
480 u8 opcode = iu->srp.rsp.opcode;
481
482 switch (opcode) {
483 case SRP_LOGIN_REQ:
484 process_login(iue);
485 break;
486 case SRP_TSK_MGMT:
487 done = process_tsk_mgmt(iue);
488 break;
489 case SRP_CMD:
490 queue_cmd(iue);
491 done = 0;
492 break;
493 case SRP_LOGIN_RSP:
494 case SRP_I_LOGOUT:
495 case SRP_T_LOGOUT:
496 case SRP_RSP:
497 case SRP_CRED_REQ:
498 case SRP_CRED_RSP:
499 case SRP_AER_REQ:
500 case SRP_AER_RSP:
501 eprintk("Unsupported type %u\n", opcode);
502 break;
503 default:
504 eprintk("Unknown type %u\n", opcode);
505 }
506
507 return done;
508}
509
510static void process_iu(struct viosrp_crq *crq, struct srp_target *target)
511{
512 struct vio_port *vport = target_to_port(target);
513 struct iu_entry *iue;
ad1331a7
BB
514 long err;
515 int done = 1;
0e5d030b
FT
516
517 iue = srp_iu_get(target);
518 if (!iue) {
519 eprintk("Error getting IU from pool, %p\n", target);
520 return;
521 }
522
523 iue->remote_token = crq->IU_data_ptr;
524
525 err = h_copy_rdma(crq->IU_length, vport->riobn,
526 iue->remote_token, vport->liobn, iue->sbuf->dma);
527
528 if (err != H_SUCCESS) {
529 eprintk("%ld transferring data error %p\n", err, iue);
0e5d030b
FT
530 goto out;
531 }
532
533 if (crq->format == VIOSRP_MAD_FORMAT)
534 done = process_mad_iu(iue);
535 else
536 done = process_srp_iu(iue);
537out:
538 if (done)
539 srp_iu_put(iue);
540}
541
15aafa2f 542static irqreturn_t ibmvstgt_interrupt(int dummy, void *data)
0e5d030b 543{
15aafa2f 544 struct srp_target *target = data;
0e5d030b
FT
545 struct vio_port *vport = target_to_port(target);
546
547 vio_disable_interrupts(vport->dma_dev);
548 queue_work(vtgtd, &vport->crq_work);
549
550 return IRQ_HANDLED;
551}
552
553static int crq_queue_create(struct crq_queue *queue, struct srp_target *target)
554{
555 int err;
556 struct vio_port *vport = target_to_port(target);
557
558 queue->msgs = (struct viosrp_crq *) get_zeroed_page(GFP_KERNEL);
559 if (!queue->msgs)
560 goto malloc_failed;
561 queue->size = PAGE_SIZE / sizeof(*queue->msgs);
562
563 queue->msg_token = dma_map_single(target->dev, queue->msgs,
564 queue->size * sizeof(*queue->msgs),
565 DMA_BIDIRECTIONAL);
566
8d8bb39b 567 if (dma_mapping_error(target->dev, queue->msg_token))
0e5d030b
FT
568 goto map_failed;
569
570 err = h_reg_crq(vport->dma_dev->unit_address, queue->msg_token,
571 PAGE_SIZE);
572
573 /* If the adapter was left active for some reason (like kexec)
574 * try freeing and re-registering
575 */
576 if (err == H_RESOURCE) {
577 do {
578 err = h_free_crq(vport->dma_dev->unit_address);
579 } while (err == H_BUSY || H_IS_LONG_BUSY(err));
580
581 err = h_reg_crq(vport->dma_dev->unit_address, queue->msg_token,
582 PAGE_SIZE);
583 }
584
585 if (err != H_SUCCESS && err != 2) {
586 eprintk("Error 0x%x opening virtual adapter\n", err);
587 goto reg_crq_failed;
588 }
589
590 err = request_irq(vport->dma_dev->irq, &ibmvstgt_interrupt,
38515e90 591 IRQF_DISABLED, "ibmvstgt", target);
0e5d030b
FT
592 if (err)
593 goto req_irq_failed;
594
595 vio_enable_interrupts(vport->dma_dev);
596
597 h_send_crq(vport->dma_dev->unit_address, 0xC001000000000000, 0);
598
599 queue->cur = 0;
600 spin_lock_init(&queue->lock);
601
602 return 0;
603
604req_irq_failed:
605 do {
606 err = h_free_crq(vport->dma_dev->unit_address);
607 } while (err == H_BUSY || H_IS_LONG_BUSY(err));
608
609reg_crq_failed:
610 dma_unmap_single(target->dev, queue->msg_token,
611 queue->size * sizeof(*queue->msgs), DMA_BIDIRECTIONAL);
612map_failed:
613 free_page((unsigned long) queue->msgs);
614
615malloc_failed:
616 return -ENOMEM;
617}
618
619static void crq_queue_destroy(struct srp_target *target)
620{
621 struct vio_port *vport = target_to_port(target);
622 struct crq_queue *queue = &vport->crq_queue;
623 int err;
624
625 free_irq(vport->dma_dev->irq, target);
626 do {
627 err = h_free_crq(vport->dma_dev->unit_address);
628 } while (err == H_BUSY || H_IS_LONG_BUSY(err));
629
630 dma_unmap_single(target->dev, queue->msg_token,
631 queue->size * sizeof(*queue->msgs), DMA_BIDIRECTIONAL);
632
633 free_page((unsigned long) queue->msgs);
634}
635
636static void process_crq(struct viosrp_crq *crq, struct srp_target *target)
637{
638 struct vio_port *vport = target_to_port(target);
639 dprintk("%x %x\n", crq->valid, crq->format);
640
641 switch (crq->valid) {
642 case 0xC0:
643 /* initialization */
644 switch (crq->format) {
645 case 0x01:
646 h_send_crq(vport->dma_dev->unit_address,
647 0xC002000000000000, 0);
648 break;
649 case 0x02:
650 break;
651 default:
652 eprintk("Unknown format %u\n", crq->format);
653 }
654 break;
655 case 0xFF:
656 /* transport event */
657 break;
658 case 0x80:
659 /* real payload */
660 switch (crq->format) {
661 case VIOSRP_SRP_FORMAT:
662 case VIOSRP_MAD_FORMAT:
663 process_iu(crq, target);
664 break;
665 case VIOSRP_OS400_FORMAT:
666 case VIOSRP_AIX_FORMAT:
667 case VIOSRP_LINUX_FORMAT:
668 case VIOSRP_INLINE_FORMAT:
669 eprintk("Unsupported format %u\n", crq->format);
670 break;
671 default:
672 eprintk("Unknown format %u\n", crq->format);
673 }
674 break;
675 default:
676 eprintk("unknown message type 0x%02x!?\n", crq->valid);
677 }
678}
679
680static inline struct viosrp_crq *next_crq(struct crq_queue *queue)
681{
682 struct viosrp_crq *crq;
683 unsigned long flags;
684
685 spin_lock_irqsave(&queue->lock, flags);
686 crq = &queue->msgs[queue->cur];
687 if (crq->valid & 0x80) {
688 if (++queue->cur == queue->size)
689 queue->cur = 0;
690 } else
691 crq = NULL;
692 spin_unlock_irqrestore(&queue->lock, flags);
693
694 return crq;
695}
696
2fdb611d 697static void handle_crq(struct work_struct *work)
0e5d030b 698{
2fdb611d
AV
699 struct vio_port *vport = container_of(work, struct vio_port, crq_work);
700 struct srp_target *target = vport->target;
0e5d030b
FT
701 struct viosrp_crq *crq;
702 int done = 0;
703
704 while (!done) {
705 while ((crq = next_crq(&vport->crq_queue)) != NULL) {
706 process_crq(crq, target);
707 crq->valid = 0x00;
708 }
709
710 vio_enable_interrupts(vport->dma_dev);
711
712 crq = next_crq(&vport->crq_queue);
713 if (crq) {
714 vio_disable_interrupts(vport->dma_dev);
715 process_crq(crq, target);
716 crq->valid = 0x00;
717 } else
718 done = 1;
719 }
720
721 handle_cmd_queue(target);
722}
723
724
725static int ibmvstgt_eh_abort_handler(struct scsi_cmnd *sc)
726{
727 unsigned long flags;
728 struct iu_entry *iue = (struct iu_entry *) sc->SCp.ptr;
729 struct srp_target *target = iue->target;
730
731 dprintk("%p %p %x\n", iue, target, vio_iu(iue)->srp.cmd.cdb[0]);
732
733 spin_lock_irqsave(&target->lock, flags);
734 list_del(&iue->ilist);
735 spin_unlock_irqrestore(&target->lock, flags);
736
737 srp_iu_put(iue);
738
739 return 0;
740}
741
17b0bcfa
FT
742static int ibmvstgt_tsk_mgmt_response(struct Scsi_Host *shost,
743 u64 itn_id, u64 mid, int result)
0e5d030b
FT
744{
745 struct iu_entry *iue = (struct iu_entry *) ((void *) mid);
746 union viosrp_iu *iu = vio_iu(iue);
747 unsigned char status, asc;
748
749 eprintk("%p %d\n", iue, result);
750 status = NO_SENSE;
751 asc = 0;
752
753 switch (iu->srp.tsk_mgmt.tsk_mgmt_func) {
754 case SRP_TSK_ABORT_TASK:
755 asc = 0x14;
756 if (result)
757 status = ABORTED_COMMAND;
758 break;
759 default:
760 break;
761 }
762
763 send_rsp(iue, NULL, status, asc);
764 srp_iu_put(iue);
765
766 return 0;
767}
768
17b0bcfa
FT
769static int ibmvstgt_it_nexus_response(struct Scsi_Host *shost, u64 itn_id,
770 int result)
771{
772 struct srp_target *target = host_to_srp_target(shost);
773 struct vio_port *vport = target_to_port(target);
774
775 if (result) {
776 eprintk("%p %d\n", shost, result);
777 srp_rport_del(vport->rport);
778 vport->rport = NULL;
779 }
780 return 0;
781}
782
ee959b00
TJ
783static ssize_t system_id_show(struct device *dev,
784 struct device_attribute *attr, char *buf)
0e5d030b
FT
785{
786 return snprintf(buf, PAGE_SIZE, "%s\n", system_id);
787}
788
ee959b00
TJ
789static ssize_t partition_number_show(struct device *dev,
790 struct device_attribute *attr, char *buf)
0e5d030b
FT
791{
792 return snprintf(buf, PAGE_SIZE, "%x\n", partition_number);
793}
794
ee959b00
TJ
795static ssize_t unit_address_show(struct device *dev,
796 struct device_attribute *attr, char *buf)
0e5d030b 797{
ee959b00 798 struct Scsi_Host *shost = class_to_shost(dev);
0e5d030b
FT
799 struct srp_target *target = host_to_srp_target(shost);
800 struct vio_port *vport = target_to_port(target);
801 return snprintf(buf, PAGE_SIZE, "%x\n", vport->dma_dev->unit_address);
802}
803
ee959b00
TJ
804static DEVICE_ATTR(system_id, S_IRUGO, system_id_show, NULL);
805static DEVICE_ATTR(partition_number, S_IRUGO, partition_number_show, NULL);
806static DEVICE_ATTR(unit_address, S_IRUGO, unit_address_show, NULL);
0e5d030b 807
ee959b00
TJ
808static struct device_attribute *ibmvstgt_attrs[] = {
809 &dev_attr_system_id,
810 &dev_attr_partition_number,
811 &dev_attr_unit_address,
0e5d030b
FT
812 NULL,
813};
814
815static struct scsi_host_template ibmvstgt_sht = {
816 .name = TGT_NAME,
817 .module = THIS_MODULE,
818 .can_queue = INITIAL_SRP_LIMIT,
819 .sg_tablesize = SG_ALL,
820 .use_clustering = DISABLE_CLUSTERING,
821 .max_sectors = DEFAULT_MAX_SECTORS,
822 .transfer_response = ibmvstgt_cmd_done,
0e5d030b 823 .eh_abort_handler = ibmvstgt_eh_abort_handler,
0e5d030b
FT
824 .shost_attrs = ibmvstgt_attrs,
825 .proc_name = TGT_NAME,
72e39ea7 826 .supported_mode = MODE_TARGET,
0e5d030b
FT
827};
828
829static int ibmvstgt_probe(struct vio_dev *dev, const struct vio_device_id *id)
830{
831 struct Scsi_Host *shost;
832 struct srp_target *target;
833 struct vio_port *vport;
834 unsigned int *dma, dma_size;
835 int err = -ENOMEM;
836
837 vport = kzalloc(sizeof(struct vio_port), GFP_KERNEL);
838 if (!vport)
839 return err;
840 shost = scsi_host_alloc(&ibmvstgt_sht, sizeof(struct srp_target));
841 if (!shost)
842 goto free_vport;
17b0bcfa 843 shost->transportt = ibmvstgt_transport_template;
0e5d030b
FT
844
845 target = host_to_srp_target(shost);
846 target->shost = shost;
847 vport->dma_dev = dev;
848 target->ldata = vport;
2fdb611d 849 vport->target = target;
0e5d030b
FT
850 err = srp_target_alloc(target, &dev->dev, INITIAL_SRP_LIMIT,
851 SRP_MAX_IU_LEN);
852 if (err)
853 goto put_host;
854
855 dma = (unsigned int *) vio_get_attribute(dev, "ibm,my-dma-window",
856 &dma_size);
857 if (!dma || dma_size != 40) {
858 eprintk("Couldn't get window property %d\n", dma_size);
859 err = -EIO;
860 goto free_srp_target;
861 }
862 vport->liobn = dma[0];
863 vport->riobn = dma[5];
864
2fdb611d 865 INIT_WORK(&vport->crq_work, handle_crq);
0e5d030b
FT
866
867 err = crq_queue_create(&vport->crq_queue, target);
868 if (err)
869 goto free_srp_target;
870
871 err = scsi_add_host(shost, target->dev);
872 if (err)
873 goto destroy_queue;
0e5d030b 874
b560665c
FT
875 err = scsi_tgt_alloc_queue(shost);
876 if (err)
877 goto destroy_queue;
878
17b0bcfa 879 return 0;
0e5d030b
FT
880destroy_queue:
881 crq_queue_destroy(target);
882free_srp_target:
883 srp_target_free(target);
884put_host:
885 scsi_host_put(shost);
886free_vport:
887 kfree(vport);
888 return err;
889}
890
891static int ibmvstgt_remove(struct vio_dev *dev)
892{
893 struct srp_target *target = (struct srp_target *) dev->dev.driver_data;
894 struct Scsi_Host *shost = target->shost;
895 struct vio_port *vport = target->ldata;
896
897 crq_queue_destroy(target);
17b0bcfa 898 srp_remove_host(shost);
0e5d030b
FT
899 scsi_remove_host(shost);
900 scsi_tgt_free_queue(shost);
901 srp_target_free(target);
902 kfree(vport);
903 scsi_host_put(shost);
904 return 0;
905}
906
907static struct vio_device_id ibmvstgt_device_table[] __devinitdata = {
908 {"v-scsi-host", "IBM,v-scsi-host"},
909 {"",""}
910};
911
912MODULE_DEVICE_TABLE(vio, ibmvstgt_device_table);
913
914static struct vio_driver ibmvstgt_driver = {
915 .id_table = ibmvstgt_device_table,
916 .probe = ibmvstgt_probe,
917 .remove = ibmvstgt_remove,
918 .driver = {
919 .name = "ibmvscsis",
920 .owner = THIS_MODULE,
921 }
922};
923
924static int get_system_info(void)
925{
926 struct device_node *rootdn;
927 const char *id, *model, *name;
a7edd0e6 928 const unsigned int *num;
0e5d030b 929
8c8dc322 930 rootdn = of_find_node_by_path("/");
0e5d030b
FT
931 if (!rootdn)
932 return -ENOENT;
933
40cd3a45
SR
934 model = of_get_property(rootdn, "model", NULL);
935 id = of_get_property(rootdn, "system-id", NULL);
0e5d030b
FT
936 if (model && id)
937 snprintf(system_id, sizeof(system_id), "%s-%s", model, id);
938
40cd3a45 939 name = of_get_property(rootdn, "ibm,partition-name", NULL);
0e5d030b
FT
940 if (name)
941 strncpy(partition_name, name, sizeof(partition_name));
942
40cd3a45 943 num = of_get_property(rootdn, "ibm,partition-no", NULL);
0e5d030b
FT
944 if (num)
945 partition_number = *num;
946
8c8dc322 947 of_node_put(rootdn);
0e5d030b
FT
948 return 0;
949}
950
17b0bcfa 951static struct srp_function_template ibmvstgt_transport_functions = {
e9906fe8 952 .tsk_mgmt_response = ibmvstgt_tsk_mgmt_response,
17b0bcfa
FT
953 .it_nexus_response = ibmvstgt_it_nexus_response,
954};
955
0e5d030b
FT
956static int ibmvstgt_init(void)
957{
958 int err = -ENOMEM;
959
960 printk("IBM eServer i/pSeries Virtual SCSI Target Driver\n");
961
17b0bcfa
FT
962 ibmvstgt_transport_template =
963 srp_attach_transport(&ibmvstgt_transport_functions);
964 if (!ibmvstgt_transport_template)
965 return err;
966
0e5d030b
FT
967 vtgtd = create_workqueue("ibmvtgtd");
968 if (!vtgtd)
17b0bcfa 969 goto release_transport;
0e5d030b
FT
970
971 err = get_system_info();
972 if (err)
973 goto destroy_wq;
974
975 err = vio_register_driver(&ibmvstgt_driver);
976 if (err)
977 goto destroy_wq;
978
979 return 0;
0e5d030b
FT
980destroy_wq:
981 destroy_workqueue(vtgtd);
17b0bcfa
FT
982release_transport:
983 srp_release_transport(ibmvstgt_transport_template);
0e5d030b
FT
984 return err;
985}
986
987static void ibmvstgt_exit(void)
988{
989 printk("Unregister IBM virtual SCSI driver\n");
990
991 destroy_workqueue(vtgtd);
992 vio_unregister_driver(&ibmvstgt_driver);
17b0bcfa 993 srp_release_transport(ibmvstgt_transport_template);
0e5d030b
FT
994}
995
996MODULE_DESCRIPTION("IBM Virtual SCSI Target");
997MODULE_AUTHOR("Santiago Leon");
998MODULE_LICENSE("GPL");
999
1000module_init(ibmvstgt_init);
1001module_exit(ibmvstgt_exit);