Merge tag 'kvm-s390-master-4.12-2' of git://git.kernel.org/pub/scm/linux/kernel/git...
[GitHub/LineageOS/android_kernel_motorola_exynos9610.git] / drivers / usb / gadget / function / f_mass_storage.c
1 /*
2 * f_mass_storage.c -- Mass Storage USB Composite Function
3 *
4 * Copyright (C) 2003-2008 Alan Stern
5 * Copyright (C) 2009 Samsung Electronics
6 * Author: Michal Nazarewicz <mina86@mina86.com>
7 * All rights reserved.
8 *
9 * Redistribution and use in source and binary forms, with or without
10 * modification, are permitted provided that the following conditions
11 * are met:
12 * 1. Redistributions of source code must retain the above copyright
13 * notice, this list of conditions, and the following disclaimer,
14 * without modification.
15 * 2. Redistributions in binary form must reproduce the above copyright
16 * notice, this list of conditions and the following disclaimer in the
17 * documentation and/or other materials provided with the distribution.
18 * 3. The names of the above-listed copyright holders may not be used
19 * to endorse or promote products derived from this software without
20 * specific prior written permission.
21 *
22 * ALTERNATIVELY, this software may be distributed under the terms of the
23 * GNU General Public License ("GPL") as published by the Free Software
24 * Foundation, either version 2 of that License or (at your option) any
25 * later version.
26 *
27 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS
28 * IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO,
29 * THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
30 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR
31 * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
32 * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
33 * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
34 * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF
35 * LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
36 * NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
37 * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
38 */
39
40 /*
41 * The Mass Storage Function acts as a USB Mass Storage device,
42 * appearing to the host as a disk drive or as a CD-ROM drive. In
43 * addition to providing an example of a genuinely useful composite
44 * function for a USB device, it also illustrates a technique of
45 * double-buffering for increased throughput.
46 *
47 * For more information about MSF and in particular its module
48 * parameters and sysfs interface read the
49 * <Documentation/usb/mass-storage.txt> file.
50 */
51
52 /*
53 * MSF is configured by specifying a fsg_config structure. It has the
54 * following fields:
55 *
56 * nluns Number of LUNs function have (anywhere from 1
57 * to FSG_MAX_LUNS).
58 * luns An array of LUN configuration values. This
59 * should be filled for each LUN that
60 * function will include (ie. for "nluns"
61 * LUNs). Each element of the array has
62 * the following fields:
63 * ->filename The path to the backing file for the LUN.
64 * Required if LUN is not marked as
65 * removable.
66 * ->ro Flag specifying access to the LUN shall be
67 * read-only. This is implied if CD-ROM
68 * emulation is enabled as well as when
69 * it was impossible to open "filename"
70 * in R/W mode.
71 * ->removable Flag specifying that LUN shall be indicated as
72 * being removable.
73 * ->cdrom Flag specifying that LUN shall be reported as
74 * being a CD-ROM.
75 * ->nofua Flag specifying that FUA flag in SCSI WRITE(10,12)
76 * commands for this LUN shall be ignored.
77 *
78 * vendor_name
79 * product_name
80 * release Information used as a reply to INQUIRY
81 * request. To use default set to NULL,
82 * NULL, 0xffff respectively. The first
83 * field should be 8 and the second 16
84 * characters or less.
85 *
86 * can_stall Set to permit function to halt bulk endpoints.
87 * Disabled on some USB devices known not
88 * to work correctly. You should set it
89 * to true.
90 *
91 * If "removable" is not set for a LUN then a backing file must be
92 * specified. If it is set, then NULL filename means the LUN's medium
93 * is not loaded (an empty string as "filename" in the fsg_config
94 * structure causes error). The CD-ROM emulation includes a single
95 * data track and no audio tracks; hence there need be only one
96 * backing file per LUN.
97 *
98 * This function is heavily based on "File-backed Storage Gadget" by
99 * Alan Stern which in turn is heavily based on "Gadget Zero" by David
100 * Brownell. The driver's SCSI command interface was based on the
101 * "Information technology - Small Computer System Interface - 2"
102 * document from X3T9.2 Project 375D, Revision 10L, 7-SEP-93,
103 * available at <http://www.t10.org/ftp/t10/drafts/s2/s2-r10l.pdf>.
104 * The single exception is opcode 0x23 (READ FORMAT CAPACITIES), which
105 * was based on the "Universal Serial Bus Mass Storage Class UFI
106 * Command Specification" document, Revision 1.0, December 14, 1998,
107 * available at
108 * <http://www.usb.org/developers/devclass_docs/usbmass-ufi10.pdf>.
109 */
110
111 /*
112 * Driver Design
113 *
114 * The MSF is fairly straightforward. There is a main kernel
115 * thread that handles most of the work. Interrupt routines field
116 * callbacks from the controller driver: bulk- and interrupt-request
117 * completion notifications, endpoint-0 events, and disconnect events.
118 * Completion events are passed to the main thread by wakeup calls. Many
119 * ep0 requests are handled at interrupt time, but SetInterface,
120 * SetConfiguration, and device reset requests are forwarded to the
121 * thread in the form of "exceptions" using SIGUSR1 signals (since they
122 * should interrupt any ongoing file I/O operations).
123 *
124 * The thread's main routine implements the standard command/data/status
125 * parts of a SCSI interaction. It and its subroutines are full of tests
126 * for pending signals/exceptions -- all this polling is necessary since
127 * the kernel has no setjmp/longjmp equivalents. (Maybe this is an
128 * indication that the driver really wants to be running in userspace.)
129 * An important point is that so long as the thread is alive it keeps an
130 * open reference to the backing file. This will prevent unmounting
131 * the backing file's underlying filesystem and could cause problems
132 * during system shutdown, for example. To prevent such problems, the
133 * thread catches INT, TERM, and KILL signals and converts them into
134 * an EXIT exception.
135 *
136 * In normal operation the main thread is started during the gadget's
137 * fsg_bind() callback and stopped during fsg_unbind(). But it can
138 * also exit when it receives a signal, and there's no point leaving
139 * the gadget running when the thread is dead. As of this moment, MSF
140 * provides no way to deregister the gadget when thread dies -- maybe
141 * a callback functions is needed.
142 *
143 * To provide maximum throughput, the driver uses a circular pipeline of
144 * buffer heads (struct fsg_buffhd). In principle the pipeline can be
145 * arbitrarily long; in practice the benefits don't justify having more
146 * than 2 stages (i.e., double buffering). But it helps to think of the
147 * pipeline as being a long one. Each buffer head contains a bulk-in and
148 * a bulk-out request pointer (since the buffer can be used for both
149 * output and input -- directions always are given from the host's
150 * point of view) as well as a pointer to the buffer and various state
151 * variables.
152 *
153 * Use of the pipeline follows a simple protocol. There is a variable
154 * (fsg->next_buffhd_to_fill) that points to the next buffer head to use.
155 * At any time that buffer head may still be in use from an earlier
156 * request, so each buffer head has a state variable indicating whether
157 * it is EMPTY, FULL, or BUSY. Typical use involves waiting for the
158 * buffer head to be EMPTY, filling the buffer either by file I/O or by
159 * USB I/O (during which the buffer head is BUSY), and marking the buffer
160 * head FULL when the I/O is complete. Then the buffer will be emptied
161 * (again possibly by USB I/O, during which it is marked BUSY) and
162 * finally marked EMPTY again (possibly by a completion routine).
163 *
164 * A module parameter tells the driver to avoid stalling the bulk
165 * endpoints wherever the transport specification allows. This is
166 * necessary for some UDCs like the SuperH, which cannot reliably clear a
167 * halt on a bulk endpoint. However, under certain circumstances the
168 * Bulk-only specification requires a stall. In such cases the driver
169 * will halt the endpoint and set a flag indicating that it should clear
170 * the halt in software during the next device reset. Hopefully this
171 * will permit everything to work correctly. Furthermore, although the
172 * specification allows the bulk-out endpoint to halt when the host sends
173 * too much data, implementing this would cause an unavoidable race.
174 * The driver will always use the "no-stall" approach for OUT transfers.
175 *
176 * One subtle point concerns sending status-stage responses for ep0
177 * requests. Some of these requests, such as device reset, can involve
178 * interrupting an ongoing file I/O operation, which might take an
179 * arbitrarily long time. During that delay the host might give up on
180 * the original ep0 request and issue a new one. When that happens the
181 * driver should not notify the host about completion of the original
182 * request, as the host will no longer be waiting for it. So the driver
183 * assigns to each ep0 request a unique tag, and it keeps track of the
184 * tag value of the request associated with a long-running exception
185 * (device-reset, interface-change, or configuration-change). When the
186 * exception handler is finished, the status-stage response is submitted
187 * only if the current ep0 request tag is equal to the exception request
188 * tag. Thus only the most recently received ep0 request will get a
189 * status-stage response.
190 *
191 * Warning: This driver source file is too long. It ought to be split up
192 * into a header file plus about 3 separate .c files, to handle the details
193 * of the Gadget, USB Mass Storage, and SCSI protocols.
194 */
195
196
197 /* #define VERBOSE_DEBUG */
198 /* #define DUMP_MSGS */
199
200 #include <linux/blkdev.h>
201 #include <linux/completion.h>
202 #include <linux/dcache.h>
203 #include <linux/delay.h>
204 #include <linux/device.h>
205 #include <linux/fcntl.h>
206 #include <linux/file.h>
207 #include <linux/fs.h>
208 #include <linux/kref.h>
209 #include <linux/kthread.h>
210 #include <linux/sched/signal.h>
211 #include <linux/limits.h>
212 #include <linux/rwsem.h>
213 #include <linux/slab.h>
214 #include <linux/spinlock.h>
215 #include <linux/string.h>
216 #include <linux/freezer.h>
217 #include <linux/module.h>
218 #include <linux/uaccess.h>
219
220 #include <linux/usb/ch9.h>
221 #include <linux/usb/gadget.h>
222 #include <linux/usb/composite.h>
223
224 #include "configfs.h"
225
226
227 /*------------------------------------------------------------------------*/
228
229 #define FSG_DRIVER_DESC "Mass Storage Function"
230 #define FSG_DRIVER_VERSION "2009/09/11"
231
232 static const char fsg_string_interface[] = "Mass Storage";
233
234 #include "storage_common.h"
235 #include "f_mass_storage.h"
236
237 /* Static strings, in UTF-8 (for simplicity we use only ASCII characters) */
238 static struct usb_string fsg_strings[] = {
239 {FSG_STRING_INTERFACE, fsg_string_interface},
240 {}
241 };
242
243 static struct usb_gadget_strings fsg_stringtab = {
244 .language = 0x0409, /* en-us */
245 .strings = fsg_strings,
246 };
247
248 static struct usb_gadget_strings *fsg_strings_array[] = {
249 &fsg_stringtab,
250 NULL,
251 };
252
253 /*-------------------------------------------------------------------------*/
254
255 struct fsg_dev;
256 struct fsg_common;
257
258 /* Data shared by all the FSG instances. */
259 struct fsg_common {
260 struct usb_gadget *gadget;
261 struct usb_composite_dev *cdev;
262 struct fsg_dev *fsg, *new_fsg;
263 wait_queue_head_t fsg_wait;
264
265 /* filesem protects: backing files in use */
266 struct rw_semaphore filesem;
267
268 /* lock protects: state, all the req_busy's */
269 spinlock_t lock;
270
271 struct usb_ep *ep0; /* Copy of gadget->ep0 */
272 struct usb_request *ep0req; /* Copy of cdev->req */
273 unsigned int ep0_req_tag;
274
275 struct fsg_buffhd *next_buffhd_to_fill;
276 struct fsg_buffhd *next_buffhd_to_drain;
277 struct fsg_buffhd *buffhds;
278 unsigned int fsg_num_buffers;
279
280 int cmnd_size;
281 u8 cmnd[MAX_COMMAND_SIZE];
282
283 unsigned int lun;
284 struct fsg_lun *luns[FSG_MAX_LUNS];
285 struct fsg_lun *curlun;
286
287 unsigned int bulk_out_maxpacket;
288 enum fsg_state state; /* For exception handling */
289 unsigned int exception_req_tag;
290
291 enum data_direction data_dir;
292 u32 data_size;
293 u32 data_size_from_cmnd;
294 u32 tag;
295 u32 residue;
296 u32 usb_amount_left;
297
298 unsigned int can_stall:1;
299 unsigned int free_storage_on_release:1;
300 unsigned int phase_error:1;
301 unsigned int short_packet_received:1;
302 unsigned int bad_lun_okay:1;
303 unsigned int running:1;
304 unsigned int sysfs:1;
305
306 int thread_wakeup_needed;
307 struct completion thread_notifier;
308 struct task_struct *thread_task;
309
310 /* Callback functions. */
311 const struct fsg_operations *ops;
312 /* Gadget's private data. */
313 void *private_data;
314
315 char inquiry_string[INQUIRY_STRING_LEN];
316
317 struct kref ref;
318 };
319
320 struct fsg_dev {
321 struct usb_function function;
322 struct usb_gadget *gadget; /* Copy of cdev->gadget */
323 struct fsg_common *common;
324
325 u16 interface_number;
326
327 unsigned int bulk_in_enabled:1;
328 unsigned int bulk_out_enabled:1;
329
330 unsigned long atomic_bitflags;
331 #define IGNORE_BULK_OUT 0
332
333 struct usb_ep *bulk_in;
334 struct usb_ep *bulk_out;
335 };
336
337 static inline int __fsg_is_set(struct fsg_common *common,
338 const char *func, unsigned line)
339 {
340 if (common->fsg)
341 return 1;
342 ERROR(common, "common->fsg is NULL in %s at %u\n", func, line);
343 WARN_ON(1);
344 return 0;
345 }
346
347 #define fsg_is_set(common) likely(__fsg_is_set(common, __func__, __LINE__))
348
349 static inline struct fsg_dev *fsg_from_func(struct usb_function *f)
350 {
351 return container_of(f, struct fsg_dev, function);
352 }
353
354 typedef void (*fsg_routine_t)(struct fsg_dev *);
355
356 static int exception_in_progress(struct fsg_common *common)
357 {
358 return common->state > FSG_STATE_IDLE;
359 }
360
361 /* Make bulk-out requests be divisible by the maxpacket size */
362 static void set_bulk_out_req_length(struct fsg_common *common,
363 struct fsg_buffhd *bh, unsigned int length)
364 {
365 unsigned int rem;
366
367 bh->bulk_out_intended_length = length;
368 rem = length % common->bulk_out_maxpacket;
369 if (rem > 0)
370 length += common->bulk_out_maxpacket - rem;
371 bh->outreq->length = length;
372 }
373
374
375 /*-------------------------------------------------------------------------*/
376
377 static int fsg_set_halt(struct fsg_dev *fsg, struct usb_ep *ep)
378 {
379 const char *name;
380
381 if (ep == fsg->bulk_in)
382 name = "bulk-in";
383 else if (ep == fsg->bulk_out)
384 name = "bulk-out";
385 else
386 name = ep->name;
387 DBG(fsg, "%s set halt\n", name);
388 return usb_ep_set_halt(ep);
389 }
390
391
392 /*-------------------------------------------------------------------------*/
393
394 /* These routines may be called in process context or in_irq */
395
396 /* Caller must hold fsg->lock */
397 static void wakeup_thread(struct fsg_common *common)
398 {
399 /*
400 * Ensure the reading of thread_wakeup_needed
401 * and the writing of bh->state are completed
402 */
403 smp_mb();
404 /* Tell the main thread that something has happened */
405 common->thread_wakeup_needed = 1;
406 if (common->thread_task)
407 wake_up_process(common->thread_task);
408 }
409
410 static void raise_exception(struct fsg_common *common, enum fsg_state new_state)
411 {
412 unsigned long flags;
413
414 /*
415 * Do nothing if a higher-priority exception is already in progress.
416 * If a lower-or-equal priority exception is in progress, preempt it
417 * and notify the main thread by sending it a signal.
418 */
419 spin_lock_irqsave(&common->lock, flags);
420 if (common->state <= new_state) {
421 common->exception_req_tag = common->ep0_req_tag;
422 common->state = new_state;
423 if (common->thread_task)
424 send_sig_info(SIGUSR1, SEND_SIG_FORCED,
425 common->thread_task);
426 }
427 spin_unlock_irqrestore(&common->lock, flags);
428 }
429
430
431 /*-------------------------------------------------------------------------*/
432
433 static int ep0_queue(struct fsg_common *common)
434 {
435 int rc;
436
437 rc = usb_ep_queue(common->ep0, common->ep0req, GFP_ATOMIC);
438 common->ep0->driver_data = common;
439 if (rc != 0 && rc != -ESHUTDOWN) {
440 /* We can't do much more than wait for a reset */
441 WARNING(common, "error in submission: %s --> %d\n",
442 common->ep0->name, rc);
443 }
444 return rc;
445 }
446
447
448 /*-------------------------------------------------------------------------*/
449
450 /* Completion handlers. These always run in_irq. */
451
452 static void bulk_in_complete(struct usb_ep *ep, struct usb_request *req)
453 {
454 struct fsg_common *common = ep->driver_data;
455 struct fsg_buffhd *bh = req->context;
456
457 if (req->status || req->actual != req->length)
458 DBG(common, "%s --> %d, %u/%u\n", __func__,
459 req->status, req->actual, req->length);
460 if (req->status == -ECONNRESET) /* Request was cancelled */
461 usb_ep_fifo_flush(ep);
462
463 /* Hold the lock while we update the request and buffer states */
464 smp_wmb();
465 spin_lock(&common->lock);
466 bh->inreq_busy = 0;
467 bh->state = BUF_STATE_EMPTY;
468 wakeup_thread(common);
469 spin_unlock(&common->lock);
470 }
471
472 static void bulk_out_complete(struct usb_ep *ep, struct usb_request *req)
473 {
474 struct fsg_common *common = ep->driver_data;
475 struct fsg_buffhd *bh = req->context;
476
477 dump_msg(common, "bulk-out", req->buf, req->actual);
478 if (req->status || req->actual != bh->bulk_out_intended_length)
479 DBG(common, "%s --> %d, %u/%u\n", __func__,
480 req->status, req->actual, bh->bulk_out_intended_length);
481 if (req->status == -ECONNRESET) /* Request was cancelled */
482 usb_ep_fifo_flush(ep);
483
484 /* Hold the lock while we update the request and buffer states */
485 smp_wmb();
486 spin_lock(&common->lock);
487 bh->outreq_busy = 0;
488 bh->state = BUF_STATE_FULL;
489 wakeup_thread(common);
490 spin_unlock(&common->lock);
491 }
492
493 static int _fsg_common_get_max_lun(struct fsg_common *common)
494 {
495 int i = ARRAY_SIZE(common->luns) - 1;
496
497 while (i >= 0 && !common->luns[i])
498 --i;
499
500 return i;
501 }
502
503 static int fsg_setup(struct usb_function *f,
504 const struct usb_ctrlrequest *ctrl)
505 {
506 struct fsg_dev *fsg = fsg_from_func(f);
507 struct usb_request *req = fsg->common->ep0req;
508 u16 w_index = le16_to_cpu(ctrl->wIndex);
509 u16 w_value = le16_to_cpu(ctrl->wValue);
510 u16 w_length = le16_to_cpu(ctrl->wLength);
511
512 if (!fsg_is_set(fsg->common))
513 return -EOPNOTSUPP;
514
515 ++fsg->common->ep0_req_tag; /* Record arrival of a new request */
516 req->context = NULL;
517 req->length = 0;
518 dump_msg(fsg, "ep0-setup", (u8 *) ctrl, sizeof(*ctrl));
519
520 switch (ctrl->bRequest) {
521
522 case US_BULK_RESET_REQUEST:
523 if (ctrl->bRequestType !=
524 (USB_DIR_OUT | USB_TYPE_CLASS | USB_RECIP_INTERFACE))
525 break;
526 if (w_index != fsg->interface_number || w_value != 0 ||
527 w_length != 0)
528 return -EDOM;
529
530 /*
531 * Raise an exception to stop the current operation
532 * and reinitialize our state.
533 */
534 DBG(fsg, "bulk reset request\n");
535 raise_exception(fsg->common, FSG_STATE_RESET);
536 return USB_GADGET_DELAYED_STATUS;
537
538 case US_BULK_GET_MAX_LUN:
539 if (ctrl->bRequestType !=
540 (USB_DIR_IN | USB_TYPE_CLASS | USB_RECIP_INTERFACE))
541 break;
542 if (w_index != fsg->interface_number || w_value != 0 ||
543 w_length != 1)
544 return -EDOM;
545 VDBG(fsg, "get max LUN\n");
546 *(u8 *)req->buf = _fsg_common_get_max_lun(fsg->common);
547
548 /* Respond with data/status */
549 req->length = min((u16)1, w_length);
550 return ep0_queue(fsg->common);
551 }
552
553 VDBG(fsg,
554 "unknown class-specific control req %02x.%02x v%04x i%04x l%u\n",
555 ctrl->bRequestType, ctrl->bRequest,
556 le16_to_cpu(ctrl->wValue), w_index, w_length);
557 return -EOPNOTSUPP;
558 }
559
560
561 /*-------------------------------------------------------------------------*/
562
563 /* All the following routines run in process context */
564
565 /* Use this for bulk or interrupt transfers, not ep0 */
566 static void start_transfer(struct fsg_dev *fsg, struct usb_ep *ep,
567 struct usb_request *req, int *pbusy,
568 enum fsg_buffer_state *state)
569 {
570 int rc;
571
572 if (ep == fsg->bulk_in)
573 dump_msg(fsg, "bulk-in", req->buf, req->length);
574
575 spin_lock_irq(&fsg->common->lock);
576 *pbusy = 1;
577 *state = BUF_STATE_BUSY;
578 spin_unlock_irq(&fsg->common->lock);
579
580 rc = usb_ep_queue(ep, req, GFP_KERNEL);
581 if (rc == 0)
582 return; /* All good, we're done */
583
584 *pbusy = 0;
585 *state = BUF_STATE_EMPTY;
586
587 /* We can't do much more than wait for a reset */
588
589 /*
590 * Note: currently the net2280 driver fails zero-length
591 * submissions if DMA is enabled.
592 */
593 if (rc != -ESHUTDOWN && !(rc == -EOPNOTSUPP && req->length == 0))
594 WARNING(fsg, "error in submission: %s --> %d\n", ep->name, rc);
595 }
596
597 static bool start_in_transfer(struct fsg_common *common, struct fsg_buffhd *bh)
598 {
599 if (!fsg_is_set(common))
600 return false;
601 start_transfer(common->fsg, common->fsg->bulk_in,
602 bh->inreq, &bh->inreq_busy, &bh->state);
603 return true;
604 }
605
606 static bool start_out_transfer(struct fsg_common *common, struct fsg_buffhd *bh)
607 {
608 if (!fsg_is_set(common))
609 return false;
610 start_transfer(common->fsg, common->fsg->bulk_out,
611 bh->outreq, &bh->outreq_busy, &bh->state);
612 return true;
613 }
614
615 static int sleep_thread(struct fsg_common *common, bool can_freeze)
616 {
617 int rc = 0;
618
619 /* Wait until a signal arrives or we are woken up */
620 for (;;) {
621 if (can_freeze)
622 try_to_freeze();
623 set_current_state(TASK_INTERRUPTIBLE);
624 if (signal_pending(current)) {
625 rc = -EINTR;
626 break;
627 }
628 if (common->thread_wakeup_needed)
629 break;
630 schedule();
631 }
632 __set_current_state(TASK_RUNNING);
633 common->thread_wakeup_needed = 0;
634
635 /*
636 * Ensure the writing of thread_wakeup_needed
637 * and the reading of bh->state are completed
638 */
639 smp_mb();
640 return rc;
641 }
642
643
644 /*-------------------------------------------------------------------------*/
645
646 static int do_read(struct fsg_common *common)
647 {
648 struct fsg_lun *curlun = common->curlun;
649 u32 lba;
650 struct fsg_buffhd *bh;
651 int rc;
652 u32 amount_left;
653 loff_t file_offset, file_offset_tmp;
654 unsigned int amount;
655 ssize_t nread;
656
657 /*
658 * Get the starting Logical Block Address and check that it's
659 * not too big.
660 */
661 if (common->cmnd[0] == READ_6)
662 lba = get_unaligned_be24(&common->cmnd[1]);
663 else {
664 lba = get_unaligned_be32(&common->cmnd[2]);
665
666 /*
667 * We allow DPO (Disable Page Out = don't save data in the
668 * cache) and FUA (Force Unit Access = don't read from the
669 * cache), but we don't implement them.
670 */
671 if ((common->cmnd[1] & ~0x18) != 0) {
672 curlun->sense_data = SS_INVALID_FIELD_IN_CDB;
673 return -EINVAL;
674 }
675 }
676 if (lba >= curlun->num_sectors) {
677 curlun->sense_data = SS_LOGICAL_BLOCK_ADDRESS_OUT_OF_RANGE;
678 return -EINVAL;
679 }
680 file_offset = ((loff_t) lba) << curlun->blkbits;
681
682 /* Carry out the file reads */
683 amount_left = common->data_size_from_cmnd;
684 if (unlikely(amount_left == 0))
685 return -EIO; /* No default reply */
686
687 for (;;) {
688 /*
689 * Figure out how much we need to read:
690 * Try to read the remaining amount.
691 * But don't read more than the buffer size.
692 * And don't try to read past the end of the file.
693 */
694 amount = min(amount_left, FSG_BUFLEN);
695 amount = min((loff_t)amount,
696 curlun->file_length - file_offset);
697
698 /* Wait for the next buffer to become available */
699 bh = common->next_buffhd_to_fill;
700 while (bh->state != BUF_STATE_EMPTY) {
701 rc = sleep_thread(common, false);
702 if (rc)
703 return rc;
704 }
705
706 /*
707 * If we were asked to read past the end of file,
708 * end with an empty buffer.
709 */
710 if (amount == 0) {
711 curlun->sense_data =
712 SS_LOGICAL_BLOCK_ADDRESS_OUT_OF_RANGE;
713 curlun->sense_data_info =
714 file_offset >> curlun->blkbits;
715 curlun->info_valid = 1;
716 bh->inreq->length = 0;
717 bh->state = BUF_STATE_FULL;
718 break;
719 }
720
721 /* Perform the read */
722 file_offset_tmp = file_offset;
723 nread = vfs_read(curlun->filp,
724 (char __user *)bh->buf,
725 amount, &file_offset_tmp);
726 VLDBG(curlun, "file read %u @ %llu -> %d\n", amount,
727 (unsigned long long)file_offset, (int)nread);
728 if (signal_pending(current))
729 return -EINTR;
730
731 if (nread < 0) {
732 LDBG(curlun, "error in file read: %d\n", (int)nread);
733 nread = 0;
734 } else if (nread < amount) {
735 LDBG(curlun, "partial file read: %d/%u\n",
736 (int)nread, amount);
737 nread = round_down(nread, curlun->blksize);
738 }
739 file_offset += nread;
740 amount_left -= nread;
741 common->residue -= nread;
742
743 /*
744 * Except at the end of the transfer, nread will be
745 * equal to the buffer size, which is divisible by the
746 * bulk-in maxpacket size.
747 */
748 bh->inreq->length = nread;
749 bh->state = BUF_STATE_FULL;
750
751 /* If an error occurred, report it and its position */
752 if (nread < amount) {
753 curlun->sense_data = SS_UNRECOVERED_READ_ERROR;
754 curlun->sense_data_info =
755 file_offset >> curlun->blkbits;
756 curlun->info_valid = 1;
757 break;
758 }
759
760 if (amount_left == 0)
761 break; /* No more left to read */
762
763 /* Send this buffer and go read some more */
764 bh->inreq->zero = 0;
765 if (!start_in_transfer(common, bh))
766 /* Don't know what to do if common->fsg is NULL */
767 return -EIO;
768 common->next_buffhd_to_fill = bh->next;
769 }
770
771 return -EIO; /* No default reply */
772 }
773
774
775 /*-------------------------------------------------------------------------*/
776
777 static int do_write(struct fsg_common *common)
778 {
779 struct fsg_lun *curlun = common->curlun;
780 u32 lba;
781 struct fsg_buffhd *bh;
782 int get_some_more;
783 u32 amount_left_to_req, amount_left_to_write;
784 loff_t usb_offset, file_offset, file_offset_tmp;
785 unsigned int amount;
786 ssize_t nwritten;
787 int rc;
788
789 if (curlun->ro) {
790 curlun->sense_data = SS_WRITE_PROTECTED;
791 return -EINVAL;
792 }
793 spin_lock(&curlun->filp->f_lock);
794 curlun->filp->f_flags &= ~O_SYNC; /* Default is not to wait */
795 spin_unlock(&curlun->filp->f_lock);
796
797 /*
798 * Get the starting Logical Block Address and check that it's
799 * not too big
800 */
801 if (common->cmnd[0] == WRITE_6)
802 lba = get_unaligned_be24(&common->cmnd[1]);
803 else {
804 lba = get_unaligned_be32(&common->cmnd[2]);
805
806 /*
807 * We allow DPO (Disable Page Out = don't save data in the
808 * cache) and FUA (Force Unit Access = write directly to the
809 * medium). We don't implement DPO; we implement FUA by
810 * performing synchronous output.
811 */
812 if (common->cmnd[1] & ~0x18) {
813 curlun->sense_data = SS_INVALID_FIELD_IN_CDB;
814 return -EINVAL;
815 }
816 if (!curlun->nofua && (common->cmnd[1] & 0x08)) { /* FUA */
817 spin_lock(&curlun->filp->f_lock);
818 curlun->filp->f_flags |= O_SYNC;
819 spin_unlock(&curlun->filp->f_lock);
820 }
821 }
822 if (lba >= curlun->num_sectors) {
823 curlun->sense_data = SS_LOGICAL_BLOCK_ADDRESS_OUT_OF_RANGE;
824 return -EINVAL;
825 }
826
827 /* Carry out the file writes */
828 get_some_more = 1;
829 file_offset = usb_offset = ((loff_t) lba) << curlun->blkbits;
830 amount_left_to_req = common->data_size_from_cmnd;
831 amount_left_to_write = common->data_size_from_cmnd;
832
833 while (amount_left_to_write > 0) {
834
835 /* Queue a request for more data from the host */
836 bh = common->next_buffhd_to_fill;
837 if (bh->state == BUF_STATE_EMPTY && get_some_more) {
838
839 /*
840 * Figure out how much we want to get:
841 * Try to get the remaining amount,
842 * but not more than the buffer size.
843 */
844 amount = min(amount_left_to_req, FSG_BUFLEN);
845
846 /* Beyond the end of the backing file? */
847 if (usb_offset >= curlun->file_length) {
848 get_some_more = 0;
849 curlun->sense_data =
850 SS_LOGICAL_BLOCK_ADDRESS_OUT_OF_RANGE;
851 curlun->sense_data_info =
852 usb_offset >> curlun->blkbits;
853 curlun->info_valid = 1;
854 continue;
855 }
856
857 /* Get the next buffer */
858 usb_offset += amount;
859 common->usb_amount_left -= amount;
860 amount_left_to_req -= amount;
861 if (amount_left_to_req == 0)
862 get_some_more = 0;
863
864 /*
865 * Except at the end of the transfer, amount will be
866 * equal to the buffer size, which is divisible by
867 * the bulk-out maxpacket size.
868 */
869 set_bulk_out_req_length(common, bh, amount);
870 if (!start_out_transfer(common, bh))
871 /* Dunno what to do if common->fsg is NULL */
872 return -EIO;
873 common->next_buffhd_to_fill = bh->next;
874 continue;
875 }
876
877 /* Write the received data to the backing file */
878 bh = common->next_buffhd_to_drain;
879 if (bh->state == BUF_STATE_EMPTY && !get_some_more)
880 break; /* We stopped early */
881 if (bh->state == BUF_STATE_FULL) {
882 smp_rmb();
883 common->next_buffhd_to_drain = bh->next;
884 bh->state = BUF_STATE_EMPTY;
885
886 /* Did something go wrong with the transfer? */
887 if (bh->outreq->status != 0) {
888 curlun->sense_data = SS_COMMUNICATION_FAILURE;
889 curlun->sense_data_info =
890 file_offset >> curlun->blkbits;
891 curlun->info_valid = 1;
892 break;
893 }
894
895 amount = bh->outreq->actual;
896 if (curlun->file_length - file_offset < amount) {
897 LERROR(curlun,
898 "write %u @ %llu beyond end %llu\n",
899 amount, (unsigned long long)file_offset,
900 (unsigned long long)curlun->file_length);
901 amount = curlun->file_length - file_offset;
902 }
903
904 /* Don't accept excess data. The spec doesn't say
905 * what to do in this case. We'll ignore the error.
906 */
907 amount = min(amount, bh->bulk_out_intended_length);
908
909 /* Don't write a partial block */
910 amount = round_down(amount, curlun->blksize);
911 if (amount == 0)
912 goto empty_write;
913
914 /* Perform the write */
915 file_offset_tmp = file_offset;
916 nwritten = vfs_write(curlun->filp,
917 (char __user *)bh->buf,
918 amount, &file_offset_tmp);
919 VLDBG(curlun, "file write %u @ %llu -> %d\n", amount,
920 (unsigned long long)file_offset, (int)nwritten);
921 if (signal_pending(current))
922 return -EINTR; /* Interrupted! */
923
924 if (nwritten < 0) {
925 LDBG(curlun, "error in file write: %d\n",
926 (int)nwritten);
927 nwritten = 0;
928 } else if (nwritten < amount) {
929 LDBG(curlun, "partial file write: %d/%u\n",
930 (int)nwritten, amount);
931 nwritten = round_down(nwritten, curlun->blksize);
932 }
933 file_offset += nwritten;
934 amount_left_to_write -= nwritten;
935 common->residue -= nwritten;
936
937 /* If an error occurred, report it and its position */
938 if (nwritten < amount) {
939 curlun->sense_data = SS_WRITE_ERROR;
940 curlun->sense_data_info =
941 file_offset >> curlun->blkbits;
942 curlun->info_valid = 1;
943 break;
944 }
945
946 empty_write:
947 /* Did the host decide to stop early? */
948 if (bh->outreq->actual < bh->bulk_out_intended_length) {
949 common->short_packet_received = 1;
950 break;
951 }
952 continue;
953 }
954
955 /* Wait for something to happen */
956 rc = sleep_thread(common, false);
957 if (rc)
958 return rc;
959 }
960
961 return -EIO; /* No default reply */
962 }
963
964
965 /*-------------------------------------------------------------------------*/
966
967 static int do_synchronize_cache(struct fsg_common *common)
968 {
969 struct fsg_lun *curlun = common->curlun;
970 int rc;
971
972 /* We ignore the requested LBA and write out all file's
973 * dirty data buffers. */
974 rc = fsg_lun_fsync_sub(curlun);
975 if (rc)
976 curlun->sense_data = SS_WRITE_ERROR;
977 return 0;
978 }
979
980
981 /*-------------------------------------------------------------------------*/
982
983 static void invalidate_sub(struct fsg_lun *curlun)
984 {
985 struct file *filp = curlun->filp;
986 struct inode *inode = file_inode(filp);
987 unsigned long rc;
988
989 rc = invalidate_mapping_pages(inode->i_mapping, 0, -1);
990 VLDBG(curlun, "invalidate_mapping_pages -> %ld\n", rc);
991 }
992
993 static int do_verify(struct fsg_common *common)
994 {
995 struct fsg_lun *curlun = common->curlun;
996 u32 lba;
997 u32 verification_length;
998 struct fsg_buffhd *bh = common->next_buffhd_to_fill;
999 loff_t file_offset, file_offset_tmp;
1000 u32 amount_left;
1001 unsigned int amount;
1002 ssize_t nread;
1003
1004 /*
1005 * Get the starting Logical Block Address and check that it's
1006 * not too big.
1007 */
1008 lba = get_unaligned_be32(&common->cmnd[2]);
1009 if (lba >= curlun->num_sectors) {
1010 curlun->sense_data = SS_LOGICAL_BLOCK_ADDRESS_OUT_OF_RANGE;
1011 return -EINVAL;
1012 }
1013
1014 /*
1015 * We allow DPO (Disable Page Out = don't save data in the
1016 * cache) but we don't implement it.
1017 */
1018 if (common->cmnd[1] & ~0x10) {
1019 curlun->sense_data = SS_INVALID_FIELD_IN_CDB;
1020 return -EINVAL;
1021 }
1022
1023 verification_length = get_unaligned_be16(&common->cmnd[7]);
1024 if (unlikely(verification_length == 0))
1025 return -EIO; /* No default reply */
1026
1027 /* Prepare to carry out the file verify */
1028 amount_left = verification_length << curlun->blkbits;
1029 file_offset = ((loff_t) lba) << curlun->blkbits;
1030
1031 /* Write out all the dirty buffers before invalidating them */
1032 fsg_lun_fsync_sub(curlun);
1033 if (signal_pending(current))
1034 return -EINTR;
1035
1036 invalidate_sub(curlun);
1037 if (signal_pending(current))
1038 return -EINTR;
1039
1040 /* Just try to read the requested blocks */
1041 while (amount_left > 0) {
1042 /*
1043 * Figure out how much we need to read:
1044 * Try to read the remaining amount, but not more than
1045 * the buffer size.
1046 * And don't try to read past the end of the file.
1047 */
1048 amount = min(amount_left, FSG_BUFLEN);
1049 amount = min((loff_t)amount,
1050 curlun->file_length - file_offset);
1051 if (amount == 0) {
1052 curlun->sense_data =
1053 SS_LOGICAL_BLOCK_ADDRESS_OUT_OF_RANGE;
1054 curlun->sense_data_info =
1055 file_offset >> curlun->blkbits;
1056 curlun->info_valid = 1;
1057 break;
1058 }
1059
1060 /* Perform the read */
1061 file_offset_tmp = file_offset;
1062 nread = vfs_read(curlun->filp,
1063 (char __user *) bh->buf,
1064 amount, &file_offset_tmp);
1065 VLDBG(curlun, "file read %u @ %llu -> %d\n", amount,
1066 (unsigned long long) file_offset,
1067 (int) nread);
1068 if (signal_pending(current))
1069 return -EINTR;
1070
1071 if (nread < 0) {
1072 LDBG(curlun, "error in file verify: %d\n", (int)nread);
1073 nread = 0;
1074 } else if (nread < amount) {
1075 LDBG(curlun, "partial file verify: %d/%u\n",
1076 (int)nread, amount);
1077 nread = round_down(nread, curlun->blksize);
1078 }
1079 if (nread == 0) {
1080 curlun->sense_data = SS_UNRECOVERED_READ_ERROR;
1081 curlun->sense_data_info =
1082 file_offset >> curlun->blkbits;
1083 curlun->info_valid = 1;
1084 break;
1085 }
1086 file_offset += nread;
1087 amount_left -= nread;
1088 }
1089 return 0;
1090 }
1091
1092
1093 /*-------------------------------------------------------------------------*/
1094
1095 static int do_inquiry(struct fsg_common *common, struct fsg_buffhd *bh)
1096 {
1097 struct fsg_lun *curlun = common->curlun;
1098 u8 *buf = (u8 *) bh->buf;
1099
1100 if (!curlun) { /* Unsupported LUNs are okay */
1101 common->bad_lun_okay = 1;
1102 memset(buf, 0, 36);
1103 buf[0] = TYPE_NO_LUN; /* Unsupported, no device-type */
1104 buf[4] = 31; /* Additional length */
1105 return 36;
1106 }
1107
1108 buf[0] = curlun->cdrom ? TYPE_ROM : TYPE_DISK;
1109 buf[1] = curlun->removable ? 0x80 : 0;
1110 buf[2] = 2; /* ANSI SCSI level 2 */
1111 buf[3] = 2; /* SCSI-2 INQUIRY data format */
1112 buf[4] = 31; /* Additional length */
1113 buf[5] = 0; /* No special options */
1114 buf[6] = 0;
1115 buf[7] = 0;
1116 if (curlun->inquiry_string[0])
1117 memcpy(buf + 8, curlun->inquiry_string,
1118 sizeof(curlun->inquiry_string));
1119 else
1120 memcpy(buf + 8, common->inquiry_string,
1121 sizeof(common->inquiry_string));
1122 return 36;
1123 }
1124
1125 static int do_request_sense(struct fsg_common *common, struct fsg_buffhd *bh)
1126 {
1127 struct fsg_lun *curlun = common->curlun;
1128 u8 *buf = (u8 *) bh->buf;
1129 u32 sd, sdinfo;
1130 int valid;
1131
1132 /*
1133 * From the SCSI-2 spec., section 7.9 (Unit attention condition):
1134 *
1135 * If a REQUEST SENSE command is received from an initiator
1136 * with a pending unit attention condition (before the target
1137 * generates the contingent allegiance condition), then the
1138 * target shall either:
1139 * a) report any pending sense data and preserve the unit
1140 * attention condition on the logical unit, or,
1141 * b) report the unit attention condition, may discard any
1142 * pending sense data, and clear the unit attention
1143 * condition on the logical unit for that initiator.
1144 *
1145 * FSG normally uses option a); enable this code to use option b).
1146 */
1147 #if 0
1148 if (curlun && curlun->unit_attention_data != SS_NO_SENSE) {
1149 curlun->sense_data = curlun->unit_attention_data;
1150 curlun->unit_attention_data = SS_NO_SENSE;
1151 }
1152 #endif
1153
1154 if (!curlun) { /* Unsupported LUNs are okay */
1155 common->bad_lun_okay = 1;
1156 sd = SS_LOGICAL_UNIT_NOT_SUPPORTED;
1157 sdinfo = 0;
1158 valid = 0;
1159 } else {
1160 sd = curlun->sense_data;
1161 sdinfo = curlun->sense_data_info;
1162 valid = curlun->info_valid << 7;
1163 curlun->sense_data = SS_NO_SENSE;
1164 curlun->sense_data_info = 0;
1165 curlun->info_valid = 0;
1166 }
1167
1168 memset(buf, 0, 18);
1169 buf[0] = valid | 0x70; /* Valid, current error */
1170 buf[2] = SK(sd);
1171 put_unaligned_be32(sdinfo, &buf[3]); /* Sense information */
1172 buf[7] = 18 - 8; /* Additional sense length */
1173 buf[12] = ASC(sd);
1174 buf[13] = ASCQ(sd);
1175 return 18;
1176 }
1177
1178 static int do_read_capacity(struct fsg_common *common, struct fsg_buffhd *bh)
1179 {
1180 struct fsg_lun *curlun = common->curlun;
1181 u32 lba = get_unaligned_be32(&common->cmnd[2]);
1182 int pmi = common->cmnd[8];
1183 u8 *buf = (u8 *)bh->buf;
1184
1185 /* Check the PMI and LBA fields */
1186 if (pmi > 1 || (pmi == 0 && lba != 0)) {
1187 curlun->sense_data = SS_INVALID_FIELD_IN_CDB;
1188 return -EINVAL;
1189 }
1190
1191 put_unaligned_be32(curlun->num_sectors - 1, &buf[0]);
1192 /* Max logical block */
1193 put_unaligned_be32(curlun->blksize, &buf[4]);/* Block length */
1194 return 8;
1195 }
1196
1197 static int do_read_header(struct fsg_common *common, struct fsg_buffhd *bh)
1198 {
1199 struct fsg_lun *curlun = common->curlun;
1200 int msf = common->cmnd[1] & 0x02;
1201 u32 lba = get_unaligned_be32(&common->cmnd[2]);
1202 u8 *buf = (u8 *)bh->buf;
1203
1204 if (common->cmnd[1] & ~0x02) { /* Mask away MSF */
1205 curlun->sense_data = SS_INVALID_FIELD_IN_CDB;
1206 return -EINVAL;
1207 }
1208 if (lba >= curlun->num_sectors) {
1209 curlun->sense_data = SS_LOGICAL_BLOCK_ADDRESS_OUT_OF_RANGE;
1210 return -EINVAL;
1211 }
1212
1213 memset(buf, 0, 8);
1214 buf[0] = 0x01; /* 2048 bytes of user data, rest is EC */
1215 store_cdrom_address(&buf[4], msf, lba);
1216 return 8;
1217 }
1218
1219 static int do_read_toc(struct fsg_common *common, struct fsg_buffhd *bh)
1220 {
1221 struct fsg_lun *curlun = common->curlun;
1222 int msf = common->cmnd[1] & 0x02;
1223 int start_track = common->cmnd[6];
1224 u8 *buf = (u8 *)bh->buf;
1225
1226 if ((common->cmnd[1] & ~0x02) != 0 || /* Mask away MSF */
1227 start_track > 1) {
1228 curlun->sense_data = SS_INVALID_FIELD_IN_CDB;
1229 return -EINVAL;
1230 }
1231
1232 memset(buf, 0, 20);
1233 buf[1] = (20-2); /* TOC data length */
1234 buf[2] = 1; /* First track number */
1235 buf[3] = 1; /* Last track number */
1236 buf[5] = 0x16; /* Data track, copying allowed */
1237 buf[6] = 0x01; /* Only track is number 1 */
1238 store_cdrom_address(&buf[8], msf, 0);
1239
1240 buf[13] = 0x16; /* Lead-out track is data */
1241 buf[14] = 0xAA; /* Lead-out track number */
1242 store_cdrom_address(&buf[16], msf, curlun->num_sectors);
1243 return 20;
1244 }
1245
1246 static int do_mode_sense(struct fsg_common *common, struct fsg_buffhd *bh)
1247 {
1248 struct fsg_lun *curlun = common->curlun;
1249 int mscmnd = common->cmnd[0];
1250 u8 *buf = (u8 *) bh->buf;
1251 u8 *buf0 = buf;
1252 int pc, page_code;
1253 int changeable_values, all_pages;
1254 int valid_page = 0;
1255 int len, limit;
1256
1257 if ((common->cmnd[1] & ~0x08) != 0) { /* Mask away DBD */
1258 curlun->sense_data = SS_INVALID_FIELD_IN_CDB;
1259 return -EINVAL;
1260 }
1261 pc = common->cmnd[2] >> 6;
1262 page_code = common->cmnd[2] & 0x3f;
1263 if (pc == 3) {
1264 curlun->sense_data = SS_SAVING_PARAMETERS_NOT_SUPPORTED;
1265 return -EINVAL;
1266 }
1267 changeable_values = (pc == 1);
1268 all_pages = (page_code == 0x3f);
1269
1270 /*
1271 * Write the mode parameter header. Fixed values are: default
1272 * medium type, no cache control (DPOFUA), and no block descriptors.
1273 * The only variable value is the WriteProtect bit. We will fill in
1274 * the mode data length later.
1275 */
1276 memset(buf, 0, 8);
1277 if (mscmnd == MODE_SENSE) {
1278 buf[2] = (curlun->ro ? 0x80 : 0x00); /* WP, DPOFUA */
1279 buf += 4;
1280 limit = 255;
1281 } else { /* MODE_SENSE_10 */
1282 buf[3] = (curlun->ro ? 0x80 : 0x00); /* WP, DPOFUA */
1283 buf += 8;
1284 limit = 65535; /* Should really be FSG_BUFLEN */
1285 }
1286
1287 /* No block descriptors */
1288
1289 /*
1290 * The mode pages, in numerical order. The only page we support
1291 * is the Caching page.
1292 */
1293 if (page_code == 0x08 || all_pages) {
1294 valid_page = 1;
1295 buf[0] = 0x08; /* Page code */
1296 buf[1] = 10; /* Page length */
1297 memset(buf+2, 0, 10); /* None of the fields are changeable */
1298
1299 if (!changeable_values) {
1300 buf[2] = 0x04; /* Write cache enable, */
1301 /* Read cache not disabled */
1302 /* No cache retention priorities */
1303 put_unaligned_be16(0xffff, &buf[4]);
1304 /* Don't disable prefetch */
1305 /* Minimum prefetch = 0 */
1306 put_unaligned_be16(0xffff, &buf[8]);
1307 /* Maximum prefetch */
1308 put_unaligned_be16(0xffff, &buf[10]);
1309 /* Maximum prefetch ceiling */
1310 }
1311 buf += 12;
1312 }
1313
1314 /*
1315 * Check that a valid page was requested and the mode data length
1316 * isn't too long.
1317 */
1318 len = buf - buf0;
1319 if (!valid_page || len > limit) {
1320 curlun->sense_data = SS_INVALID_FIELD_IN_CDB;
1321 return -EINVAL;
1322 }
1323
1324 /* Store the mode data length */
1325 if (mscmnd == MODE_SENSE)
1326 buf0[0] = len - 1;
1327 else
1328 put_unaligned_be16(len - 2, buf0);
1329 return len;
1330 }
1331
1332 static int do_start_stop(struct fsg_common *common)
1333 {
1334 struct fsg_lun *curlun = common->curlun;
1335 int loej, start;
1336
1337 if (!curlun) {
1338 return -EINVAL;
1339 } else if (!curlun->removable) {
1340 curlun->sense_data = SS_INVALID_COMMAND;
1341 return -EINVAL;
1342 } else if ((common->cmnd[1] & ~0x01) != 0 || /* Mask away Immed */
1343 (common->cmnd[4] & ~0x03) != 0) { /* Mask LoEj, Start */
1344 curlun->sense_data = SS_INVALID_FIELD_IN_CDB;
1345 return -EINVAL;
1346 }
1347
1348 loej = common->cmnd[4] & 0x02;
1349 start = common->cmnd[4] & 0x01;
1350
1351 /*
1352 * Our emulation doesn't support mounting; the medium is
1353 * available for use as soon as it is loaded.
1354 */
1355 if (start) {
1356 if (!fsg_lun_is_open(curlun)) {
1357 curlun->sense_data = SS_MEDIUM_NOT_PRESENT;
1358 return -EINVAL;
1359 }
1360 return 0;
1361 }
1362
1363 /* Are we allowed to unload the media? */
1364 if (curlun->prevent_medium_removal) {
1365 LDBG(curlun, "unload attempt prevented\n");
1366 curlun->sense_data = SS_MEDIUM_REMOVAL_PREVENTED;
1367 return -EINVAL;
1368 }
1369
1370 if (!loej)
1371 return 0;
1372
1373 up_read(&common->filesem);
1374 down_write(&common->filesem);
1375 fsg_lun_close(curlun);
1376 up_write(&common->filesem);
1377 down_read(&common->filesem);
1378
1379 return 0;
1380 }
1381
1382 static int do_prevent_allow(struct fsg_common *common)
1383 {
1384 struct fsg_lun *curlun = common->curlun;
1385 int prevent;
1386
1387 if (!common->curlun) {
1388 return -EINVAL;
1389 } else if (!common->curlun->removable) {
1390 common->curlun->sense_data = SS_INVALID_COMMAND;
1391 return -EINVAL;
1392 }
1393
1394 prevent = common->cmnd[4] & 0x01;
1395 if ((common->cmnd[4] & ~0x01) != 0) { /* Mask away Prevent */
1396 curlun->sense_data = SS_INVALID_FIELD_IN_CDB;
1397 return -EINVAL;
1398 }
1399
1400 if (curlun->prevent_medium_removal && !prevent)
1401 fsg_lun_fsync_sub(curlun);
1402 curlun->prevent_medium_removal = prevent;
1403 return 0;
1404 }
1405
1406 static int do_read_format_capacities(struct fsg_common *common,
1407 struct fsg_buffhd *bh)
1408 {
1409 struct fsg_lun *curlun = common->curlun;
1410 u8 *buf = (u8 *) bh->buf;
1411
1412 buf[0] = buf[1] = buf[2] = 0;
1413 buf[3] = 8; /* Only the Current/Maximum Capacity Descriptor */
1414 buf += 4;
1415
1416 put_unaligned_be32(curlun->num_sectors, &buf[0]);
1417 /* Number of blocks */
1418 put_unaligned_be32(curlun->blksize, &buf[4]);/* Block length */
1419 buf[4] = 0x02; /* Current capacity */
1420 return 12;
1421 }
1422
1423 static int do_mode_select(struct fsg_common *common, struct fsg_buffhd *bh)
1424 {
1425 struct fsg_lun *curlun = common->curlun;
1426
1427 /* We don't support MODE SELECT */
1428 if (curlun)
1429 curlun->sense_data = SS_INVALID_COMMAND;
1430 return -EINVAL;
1431 }
1432
1433
1434 /*-------------------------------------------------------------------------*/
1435
1436 static int halt_bulk_in_endpoint(struct fsg_dev *fsg)
1437 {
1438 int rc;
1439
1440 rc = fsg_set_halt(fsg, fsg->bulk_in);
1441 if (rc == -EAGAIN)
1442 VDBG(fsg, "delayed bulk-in endpoint halt\n");
1443 while (rc != 0) {
1444 if (rc != -EAGAIN) {
1445 WARNING(fsg, "usb_ep_set_halt -> %d\n", rc);
1446 rc = 0;
1447 break;
1448 }
1449
1450 /* Wait for a short time and then try again */
1451 if (msleep_interruptible(100) != 0)
1452 return -EINTR;
1453 rc = usb_ep_set_halt(fsg->bulk_in);
1454 }
1455 return rc;
1456 }
1457
1458 static int wedge_bulk_in_endpoint(struct fsg_dev *fsg)
1459 {
1460 int rc;
1461
1462 DBG(fsg, "bulk-in set wedge\n");
1463 rc = usb_ep_set_wedge(fsg->bulk_in);
1464 if (rc == -EAGAIN)
1465 VDBG(fsg, "delayed bulk-in endpoint wedge\n");
1466 while (rc != 0) {
1467 if (rc != -EAGAIN) {
1468 WARNING(fsg, "usb_ep_set_wedge -> %d\n", rc);
1469 rc = 0;
1470 break;
1471 }
1472
1473 /* Wait for a short time and then try again */
1474 if (msleep_interruptible(100) != 0)
1475 return -EINTR;
1476 rc = usb_ep_set_wedge(fsg->bulk_in);
1477 }
1478 return rc;
1479 }
1480
1481 static int throw_away_data(struct fsg_common *common)
1482 {
1483 struct fsg_buffhd *bh;
1484 u32 amount;
1485 int rc;
1486
1487 for (bh = common->next_buffhd_to_drain;
1488 bh->state != BUF_STATE_EMPTY || common->usb_amount_left > 0;
1489 bh = common->next_buffhd_to_drain) {
1490
1491 /* Throw away the data in a filled buffer */
1492 if (bh->state == BUF_STATE_FULL) {
1493 smp_rmb();
1494 bh->state = BUF_STATE_EMPTY;
1495 common->next_buffhd_to_drain = bh->next;
1496
1497 /* A short packet or an error ends everything */
1498 if (bh->outreq->actual < bh->bulk_out_intended_length ||
1499 bh->outreq->status != 0) {
1500 raise_exception(common,
1501 FSG_STATE_ABORT_BULK_OUT);
1502 return -EINTR;
1503 }
1504 continue;
1505 }
1506
1507 /* Try to submit another request if we need one */
1508 bh = common->next_buffhd_to_fill;
1509 if (bh->state == BUF_STATE_EMPTY
1510 && common->usb_amount_left > 0) {
1511 amount = min(common->usb_amount_left, FSG_BUFLEN);
1512
1513 /*
1514 * Except at the end of the transfer, amount will be
1515 * equal to the buffer size, which is divisible by
1516 * the bulk-out maxpacket size.
1517 */
1518 set_bulk_out_req_length(common, bh, amount);
1519 if (!start_out_transfer(common, bh))
1520 /* Dunno what to do if common->fsg is NULL */
1521 return -EIO;
1522 common->next_buffhd_to_fill = bh->next;
1523 common->usb_amount_left -= amount;
1524 continue;
1525 }
1526
1527 /* Otherwise wait for something to happen */
1528 rc = sleep_thread(common, true);
1529 if (rc)
1530 return rc;
1531 }
1532 return 0;
1533 }
1534
1535 static int finish_reply(struct fsg_common *common)
1536 {
1537 struct fsg_buffhd *bh = common->next_buffhd_to_fill;
1538 int rc = 0;
1539
1540 switch (common->data_dir) {
1541 case DATA_DIR_NONE:
1542 break; /* Nothing to send */
1543
1544 /*
1545 * If we don't know whether the host wants to read or write,
1546 * this must be CB or CBI with an unknown command. We mustn't
1547 * try to send or receive any data. So stall both bulk pipes
1548 * if we can and wait for a reset.
1549 */
1550 case DATA_DIR_UNKNOWN:
1551 if (!common->can_stall) {
1552 /* Nothing */
1553 } else if (fsg_is_set(common)) {
1554 fsg_set_halt(common->fsg, common->fsg->bulk_out);
1555 rc = halt_bulk_in_endpoint(common->fsg);
1556 } else {
1557 /* Don't know what to do if common->fsg is NULL */
1558 rc = -EIO;
1559 }
1560 break;
1561
1562 /* All but the last buffer of data must have already been sent */
1563 case DATA_DIR_TO_HOST:
1564 if (common->data_size == 0) {
1565 /* Nothing to send */
1566
1567 /* Don't know what to do if common->fsg is NULL */
1568 } else if (!fsg_is_set(common)) {
1569 rc = -EIO;
1570
1571 /* If there's no residue, simply send the last buffer */
1572 } else if (common->residue == 0) {
1573 bh->inreq->zero = 0;
1574 if (!start_in_transfer(common, bh))
1575 return -EIO;
1576 common->next_buffhd_to_fill = bh->next;
1577
1578 /*
1579 * For Bulk-only, mark the end of the data with a short
1580 * packet. If we are allowed to stall, halt the bulk-in
1581 * endpoint. (Note: This violates the Bulk-Only Transport
1582 * specification, which requires us to pad the data if we
1583 * don't halt the endpoint. Presumably nobody will mind.)
1584 */
1585 } else {
1586 bh->inreq->zero = 1;
1587 if (!start_in_transfer(common, bh))
1588 rc = -EIO;
1589 common->next_buffhd_to_fill = bh->next;
1590 if (common->can_stall)
1591 rc = halt_bulk_in_endpoint(common->fsg);
1592 }
1593 break;
1594
1595 /*
1596 * We have processed all we want from the data the host has sent.
1597 * There may still be outstanding bulk-out requests.
1598 */
1599 case DATA_DIR_FROM_HOST:
1600 if (common->residue == 0) {
1601 /* Nothing to receive */
1602
1603 /* Did the host stop sending unexpectedly early? */
1604 } else if (common->short_packet_received) {
1605 raise_exception(common, FSG_STATE_ABORT_BULK_OUT);
1606 rc = -EINTR;
1607
1608 /*
1609 * We haven't processed all the incoming data. Even though
1610 * we may be allowed to stall, doing so would cause a race.
1611 * The controller may already have ACK'ed all the remaining
1612 * bulk-out packets, in which case the host wouldn't see a
1613 * STALL. Not realizing the endpoint was halted, it wouldn't
1614 * clear the halt -- leading to problems later on.
1615 */
1616 #if 0
1617 } else if (common->can_stall) {
1618 if (fsg_is_set(common))
1619 fsg_set_halt(common->fsg,
1620 common->fsg->bulk_out);
1621 raise_exception(common, FSG_STATE_ABORT_BULK_OUT);
1622 rc = -EINTR;
1623 #endif
1624
1625 /*
1626 * We can't stall. Read in the excess data and throw it
1627 * all away.
1628 */
1629 } else {
1630 rc = throw_away_data(common);
1631 }
1632 break;
1633 }
1634 return rc;
1635 }
1636
1637 static int send_status(struct fsg_common *common)
1638 {
1639 struct fsg_lun *curlun = common->curlun;
1640 struct fsg_buffhd *bh;
1641 struct bulk_cs_wrap *csw;
1642 int rc;
1643 u8 status = US_BULK_STAT_OK;
1644 u32 sd, sdinfo = 0;
1645
1646 /* Wait for the next buffer to become available */
1647 bh = common->next_buffhd_to_fill;
1648 while (bh->state != BUF_STATE_EMPTY) {
1649 rc = sleep_thread(common, true);
1650 if (rc)
1651 return rc;
1652 }
1653
1654 if (curlun) {
1655 sd = curlun->sense_data;
1656 sdinfo = curlun->sense_data_info;
1657 } else if (common->bad_lun_okay)
1658 sd = SS_NO_SENSE;
1659 else
1660 sd = SS_LOGICAL_UNIT_NOT_SUPPORTED;
1661
1662 if (common->phase_error) {
1663 DBG(common, "sending phase-error status\n");
1664 status = US_BULK_STAT_PHASE;
1665 sd = SS_INVALID_COMMAND;
1666 } else if (sd != SS_NO_SENSE) {
1667 DBG(common, "sending command-failure status\n");
1668 status = US_BULK_STAT_FAIL;
1669 VDBG(common, " sense data: SK x%02x, ASC x%02x, ASCQ x%02x;"
1670 " info x%x\n",
1671 SK(sd), ASC(sd), ASCQ(sd), sdinfo);
1672 }
1673
1674 /* Store and send the Bulk-only CSW */
1675 csw = (void *)bh->buf;
1676
1677 csw->Signature = cpu_to_le32(US_BULK_CS_SIGN);
1678 csw->Tag = common->tag;
1679 csw->Residue = cpu_to_le32(common->residue);
1680 csw->Status = status;
1681
1682 bh->inreq->length = US_BULK_CS_WRAP_LEN;
1683 bh->inreq->zero = 0;
1684 if (!start_in_transfer(common, bh))
1685 /* Don't know what to do if common->fsg is NULL */
1686 return -EIO;
1687
1688 common->next_buffhd_to_fill = bh->next;
1689 return 0;
1690 }
1691
1692
1693 /*-------------------------------------------------------------------------*/
1694
1695 /*
1696 * Check whether the command is properly formed and whether its data size
1697 * and direction agree with the values we already have.
1698 */
1699 static int check_command(struct fsg_common *common, int cmnd_size,
1700 enum data_direction data_dir, unsigned int mask,
1701 int needs_medium, const char *name)
1702 {
1703 int i;
1704 unsigned int lun = common->cmnd[1] >> 5;
1705 static const char dirletter[4] = {'u', 'o', 'i', 'n'};
1706 char hdlen[20];
1707 struct fsg_lun *curlun;
1708
1709 hdlen[0] = 0;
1710 if (common->data_dir != DATA_DIR_UNKNOWN)
1711 sprintf(hdlen, ", H%c=%u", dirletter[(int) common->data_dir],
1712 common->data_size);
1713 VDBG(common, "SCSI command: %s; Dc=%d, D%c=%u; Hc=%d%s\n",
1714 name, cmnd_size, dirletter[(int) data_dir],
1715 common->data_size_from_cmnd, common->cmnd_size, hdlen);
1716
1717 /*
1718 * We can't reply at all until we know the correct data direction
1719 * and size.
1720 */
1721 if (common->data_size_from_cmnd == 0)
1722 data_dir = DATA_DIR_NONE;
1723 if (common->data_size < common->data_size_from_cmnd) {
1724 /*
1725 * Host data size < Device data size is a phase error.
1726 * Carry out the command, but only transfer as much as
1727 * we are allowed.
1728 */
1729 common->data_size_from_cmnd = common->data_size;
1730 common->phase_error = 1;
1731 }
1732 common->residue = common->data_size;
1733 common->usb_amount_left = common->data_size;
1734
1735 /* Conflicting data directions is a phase error */
1736 if (common->data_dir != data_dir && common->data_size_from_cmnd > 0) {
1737 common->phase_error = 1;
1738 return -EINVAL;
1739 }
1740
1741 /* Verify the length of the command itself */
1742 if (cmnd_size != common->cmnd_size) {
1743
1744 /*
1745 * Special case workaround: There are plenty of buggy SCSI
1746 * implementations. Many have issues with cbw->Length
1747 * field passing a wrong command size. For those cases we
1748 * always try to work around the problem by using the length
1749 * sent by the host side provided it is at least as large
1750 * as the correct command length.
1751 * Examples of such cases would be MS-Windows, which issues
1752 * REQUEST SENSE with cbw->Length == 12 where it should
1753 * be 6, and xbox360 issuing INQUIRY, TEST UNIT READY and
1754 * REQUEST SENSE with cbw->Length == 10 where it should
1755 * be 6 as well.
1756 */
1757 if (cmnd_size <= common->cmnd_size) {
1758 DBG(common, "%s is buggy! Expected length %d "
1759 "but we got %d\n", name,
1760 cmnd_size, common->cmnd_size);
1761 cmnd_size = common->cmnd_size;
1762 } else {
1763 common->phase_error = 1;
1764 return -EINVAL;
1765 }
1766 }
1767
1768 /* Check that the LUN values are consistent */
1769 if (common->lun != lun)
1770 DBG(common, "using LUN %u from CBW, not LUN %u from CDB\n",
1771 common->lun, lun);
1772
1773 /* Check the LUN */
1774 curlun = common->curlun;
1775 if (curlun) {
1776 if (common->cmnd[0] != REQUEST_SENSE) {
1777 curlun->sense_data = SS_NO_SENSE;
1778 curlun->sense_data_info = 0;
1779 curlun->info_valid = 0;
1780 }
1781 } else {
1782 common->bad_lun_okay = 0;
1783
1784 /*
1785 * INQUIRY and REQUEST SENSE commands are explicitly allowed
1786 * to use unsupported LUNs; all others may not.
1787 */
1788 if (common->cmnd[0] != INQUIRY &&
1789 common->cmnd[0] != REQUEST_SENSE) {
1790 DBG(common, "unsupported LUN %u\n", common->lun);
1791 return -EINVAL;
1792 }
1793 }
1794
1795 /*
1796 * If a unit attention condition exists, only INQUIRY and
1797 * REQUEST SENSE commands are allowed; anything else must fail.
1798 */
1799 if (curlun && curlun->unit_attention_data != SS_NO_SENSE &&
1800 common->cmnd[0] != INQUIRY &&
1801 common->cmnd[0] != REQUEST_SENSE) {
1802 curlun->sense_data = curlun->unit_attention_data;
1803 curlun->unit_attention_data = SS_NO_SENSE;
1804 return -EINVAL;
1805 }
1806
1807 /* Check that only command bytes listed in the mask are non-zero */
1808 common->cmnd[1] &= 0x1f; /* Mask away the LUN */
1809 for (i = 1; i < cmnd_size; ++i) {
1810 if (common->cmnd[i] && !(mask & (1 << i))) {
1811 if (curlun)
1812 curlun->sense_data = SS_INVALID_FIELD_IN_CDB;
1813 return -EINVAL;
1814 }
1815 }
1816
1817 /* If the medium isn't mounted and the command needs to access
1818 * it, return an error. */
1819 if (curlun && !fsg_lun_is_open(curlun) && needs_medium) {
1820 curlun->sense_data = SS_MEDIUM_NOT_PRESENT;
1821 return -EINVAL;
1822 }
1823
1824 return 0;
1825 }
1826
1827 /* wrapper of check_command for data size in blocks handling */
1828 static int check_command_size_in_blocks(struct fsg_common *common,
1829 int cmnd_size, enum data_direction data_dir,
1830 unsigned int mask, int needs_medium, const char *name)
1831 {
1832 if (common->curlun)
1833 common->data_size_from_cmnd <<= common->curlun->blkbits;
1834 return check_command(common, cmnd_size, data_dir,
1835 mask, needs_medium, name);
1836 }
1837
1838 static int do_scsi_command(struct fsg_common *common)
1839 {
1840 struct fsg_buffhd *bh;
1841 int rc;
1842 int reply = -EINVAL;
1843 int i;
1844 static char unknown[16];
1845
1846 dump_cdb(common);
1847
1848 /* Wait for the next buffer to become available for data or status */
1849 bh = common->next_buffhd_to_fill;
1850 common->next_buffhd_to_drain = bh;
1851 while (bh->state != BUF_STATE_EMPTY) {
1852 rc = sleep_thread(common, true);
1853 if (rc)
1854 return rc;
1855 }
1856 common->phase_error = 0;
1857 common->short_packet_received = 0;
1858
1859 down_read(&common->filesem); /* We're using the backing file */
1860 switch (common->cmnd[0]) {
1861
1862 case INQUIRY:
1863 common->data_size_from_cmnd = common->cmnd[4];
1864 reply = check_command(common, 6, DATA_DIR_TO_HOST,
1865 (1<<4), 0,
1866 "INQUIRY");
1867 if (reply == 0)
1868 reply = do_inquiry(common, bh);
1869 break;
1870
1871 case MODE_SELECT:
1872 common->data_size_from_cmnd = common->cmnd[4];
1873 reply = check_command(common, 6, DATA_DIR_FROM_HOST,
1874 (1<<1) | (1<<4), 0,
1875 "MODE SELECT(6)");
1876 if (reply == 0)
1877 reply = do_mode_select(common, bh);
1878 break;
1879
1880 case MODE_SELECT_10:
1881 common->data_size_from_cmnd =
1882 get_unaligned_be16(&common->cmnd[7]);
1883 reply = check_command(common, 10, DATA_DIR_FROM_HOST,
1884 (1<<1) | (3<<7), 0,
1885 "MODE SELECT(10)");
1886 if (reply == 0)
1887 reply = do_mode_select(common, bh);
1888 break;
1889
1890 case MODE_SENSE:
1891 common->data_size_from_cmnd = common->cmnd[4];
1892 reply = check_command(common, 6, DATA_DIR_TO_HOST,
1893 (1<<1) | (1<<2) | (1<<4), 0,
1894 "MODE SENSE(6)");
1895 if (reply == 0)
1896 reply = do_mode_sense(common, bh);
1897 break;
1898
1899 case MODE_SENSE_10:
1900 common->data_size_from_cmnd =
1901 get_unaligned_be16(&common->cmnd[7]);
1902 reply = check_command(common, 10, DATA_DIR_TO_HOST,
1903 (1<<1) | (1<<2) | (3<<7), 0,
1904 "MODE SENSE(10)");
1905 if (reply == 0)
1906 reply = do_mode_sense(common, bh);
1907 break;
1908
1909 case ALLOW_MEDIUM_REMOVAL:
1910 common->data_size_from_cmnd = 0;
1911 reply = check_command(common, 6, DATA_DIR_NONE,
1912 (1<<4), 0,
1913 "PREVENT-ALLOW MEDIUM REMOVAL");
1914 if (reply == 0)
1915 reply = do_prevent_allow(common);
1916 break;
1917
1918 case READ_6:
1919 i = common->cmnd[4];
1920 common->data_size_from_cmnd = (i == 0) ? 256 : i;
1921 reply = check_command_size_in_blocks(common, 6,
1922 DATA_DIR_TO_HOST,
1923 (7<<1) | (1<<4), 1,
1924 "READ(6)");
1925 if (reply == 0)
1926 reply = do_read(common);
1927 break;
1928
1929 case READ_10:
1930 common->data_size_from_cmnd =
1931 get_unaligned_be16(&common->cmnd[7]);
1932 reply = check_command_size_in_blocks(common, 10,
1933 DATA_DIR_TO_HOST,
1934 (1<<1) | (0xf<<2) | (3<<7), 1,
1935 "READ(10)");
1936 if (reply == 0)
1937 reply = do_read(common);
1938 break;
1939
1940 case READ_12:
1941 common->data_size_from_cmnd =
1942 get_unaligned_be32(&common->cmnd[6]);
1943 reply = check_command_size_in_blocks(common, 12,
1944 DATA_DIR_TO_HOST,
1945 (1<<1) | (0xf<<2) | (0xf<<6), 1,
1946 "READ(12)");
1947 if (reply == 0)
1948 reply = do_read(common);
1949 break;
1950
1951 case READ_CAPACITY:
1952 common->data_size_from_cmnd = 8;
1953 reply = check_command(common, 10, DATA_DIR_TO_HOST,
1954 (0xf<<2) | (1<<8), 1,
1955 "READ CAPACITY");
1956 if (reply == 0)
1957 reply = do_read_capacity(common, bh);
1958 break;
1959
1960 case READ_HEADER:
1961 if (!common->curlun || !common->curlun->cdrom)
1962 goto unknown_cmnd;
1963 common->data_size_from_cmnd =
1964 get_unaligned_be16(&common->cmnd[7]);
1965 reply = check_command(common, 10, DATA_DIR_TO_HOST,
1966 (3<<7) | (0x1f<<1), 1,
1967 "READ HEADER");
1968 if (reply == 0)
1969 reply = do_read_header(common, bh);
1970 break;
1971
1972 case READ_TOC:
1973 if (!common->curlun || !common->curlun->cdrom)
1974 goto unknown_cmnd;
1975 common->data_size_from_cmnd =
1976 get_unaligned_be16(&common->cmnd[7]);
1977 reply = check_command(common, 10, DATA_DIR_TO_HOST,
1978 (7<<6) | (1<<1), 1,
1979 "READ TOC");
1980 if (reply == 0)
1981 reply = do_read_toc(common, bh);
1982 break;
1983
1984 case READ_FORMAT_CAPACITIES:
1985 common->data_size_from_cmnd =
1986 get_unaligned_be16(&common->cmnd[7]);
1987 reply = check_command(common, 10, DATA_DIR_TO_HOST,
1988 (3<<7), 1,
1989 "READ FORMAT CAPACITIES");
1990 if (reply == 0)
1991 reply = do_read_format_capacities(common, bh);
1992 break;
1993
1994 case REQUEST_SENSE:
1995 common->data_size_from_cmnd = common->cmnd[4];
1996 reply = check_command(common, 6, DATA_DIR_TO_HOST,
1997 (1<<4), 0,
1998 "REQUEST SENSE");
1999 if (reply == 0)
2000 reply = do_request_sense(common, bh);
2001 break;
2002
2003 case START_STOP:
2004 common->data_size_from_cmnd = 0;
2005 reply = check_command(common, 6, DATA_DIR_NONE,
2006 (1<<1) | (1<<4), 0,
2007 "START-STOP UNIT");
2008 if (reply == 0)
2009 reply = do_start_stop(common);
2010 break;
2011
2012 case SYNCHRONIZE_CACHE:
2013 common->data_size_from_cmnd = 0;
2014 reply = check_command(common, 10, DATA_DIR_NONE,
2015 (0xf<<2) | (3<<7), 1,
2016 "SYNCHRONIZE CACHE");
2017 if (reply == 0)
2018 reply = do_synchronize_cache(common);
2019 break;
2020
2021 case TEST_UNIT_READY:
2022 common->data_size_from_cmnd = 0;
2023 reply = check_command(common, 6, DATA_DIR_NONE,
2024 0, 1,
2025 "TEST UNIT READY");
2026 break;
2027
2028 /*
2029 * Although optional, this command is used by MS-Windows. We
2030 * support a minimal version: BytChk must be 0.
2031 */
2032 case VERIFY:
2033 common->data_size_from_cmnd = 0;
2034 reply = check_command(common, 10, DATA_DIR_NONE,
2035 (1<<1) | (0xf<<2) | (3<<7), 1,
2036 "VERIFY");
2037 if (reply == 0)
2038 reply = do_verify(common);
2039 break;
2040
2041 case WRITE_6:
2042 i = common->cmnd[4];
2043 common->data_size_from_cmnd = (i == 0) ? 256 : i;
2044 reply = check_command_size_in_blocks(common, 6,
2045 DATA_DIR_FROM_HOST,
2046 (7<<1) | (1<<4), 1,
2047 "WRITE(6)");
2048 if (reply == 0)
2049 reply = do_write(common);
2050 break;
2051
2052 case WRITE_10:
2053 common->data_size_from_cmnd =
2054 get_unaligned_be16(&common->cmnd[7]);
2055 reply = check_command_size_in_blocks(common, 10,
2056 DATA_DIR_FROM_HOST,
2057 (1<<1) | (0xf<<2) | (3<<7), 1,
2058 "WRITE(10)");
2059 if (reply == 0)
2060 reply = do_write(common);
2061 break;
2062
2063 case WRITE_12:
2064 common->data_size_from_cmnd =
2065 get_unaligned_be32(&common->cmnd[6]);
2066 reply = check_command_size_in_blocks(common, 12,
2067 DATA_DIR_FROM_HOST,
2068 (1<<1) | (0xf<<2) | (0xf<<6), 1,
2069 "WRITE(12)");
2070 if (reply == 0)
2071 reply = do_write(common);
2072 break;
2073
2074 /*
2075 * Some mandatory commands that we recognize but don't implement.
2076 * They don't mean much in this setting. It's left as an exercise
2077 * for anyone interested to implement RESERVE and RELEASE in terms
2078 * of Posix locks.
2079 */
2080 case FORMAT_UNIT:
2081 case RELEASE:
2082 case RESERVE:
2083 case SEND_DIAGNOSTIC:
2084 /* Fall through */
2085
2086 default:
2087 unknown_cmnd:
2088 common->data_size_from_cmnd = 0;
2089 sprintf(unknown, "Unknown x%02x", common->cmnd[0]);
2090 reply = check_command(common, common->cmnd_size,
2091 DATA_DIR_UNKNOWN, ~0, 0, unknown);
2092 if (reply == 0) {
2093 common->curlun->sense_data = SS_INVALID_COMMAND;
2094 reply = -EINVAL;
2095 }
2096 break;
2097 }
2098 up_read(&common->filesem);
2099
2100 if (reply == -EINTR || signal_pending(current))
2101 return -EINTR;
2102
2103 /* Set up the single reply buffer for finish_reply() */
2104 if (reply == -EINVAL)
2105 reply = 0; /* Error reply length */
2106 if (reply >= 0 && common->data_dir == DATA_DIR_TO_HOST) {
2107 reply = min((u32)reply, common->data_size_from_cmnd);
2108 bh->inreq->length = reply;
2109 bh->state = BUF_STATE_FULL;
2110 common->residue -= reply;
2111 } /* Otherwise it's already set */
2112
2113 return 0;
2114 }
2115
2116
2117 /*-------------------------------------------------------------------------*/
2118
2119 static int received_cbw(struct fsg_dev *fsg, struct fsg_buffhd *bh)
2120 {
2121 struct usb_request *req = bh->outreq;
2122 struct bulk_cb_wrap *cbw = req->buf;
2123 struct fsg_common *common = fsg->common;
2124
2125 /* Was this a real packet? Should it be ignored? */
2126 if (req->status || test_bit(IGNORE_BULK_OUT, &fsg->atomic_bitflags))
2127 return -EINVAL;
2128
2129 /* Is the CBW valid? */
2130 if (req->actual != US_BULK_CB_WRAP_LEN ||
2131 cbw->Signature != cpu_to_le32(
2132 US_BULK_CB_SIGN)) {
2133 DBG(fsg, "invalid CBW: len %u sig 0x%x\n",
2134 req->actual,
2135 le32_to_cpu(cbw->Signature));
2136
2137 /*
2138 * The Bulk-only spec says we MUST stall the IN endpoint
2139 * (6.6.1), so it's unavoidable. It also says we must
2140 * retain this state until the next reset, but there's
2141 * no way to tell the controller driver it should ignore
2142 * Clear-Feature(HALT) requests.
2143 *
2144 * We aren't required to halt the OUT endpoint; instead
2145 * we can simply accept and discard any data received
2146 * until the next reset.
2147 */
2148 wedge_bulk_in_endpoint(fsg);
2149 set_bit(IGNORE_BULK_OUT, &fsg->atomic_bitflags);
2150 return -EINVAL;
2151 }
2152
2153 /* Is the CBW meaningful? */
2154 if (cbw->Lun >= ARRAY_SIZE(common->luns) ||
2155 cbw->Flags & ~US_BULK_FLAG_IN || cbw->Length <= 0 ||
2156 cbw->Length > MAX_COMMAND_SIZE) {
2157 DBG(fsg, "non-meaningful CBW: lun = %u, flags = 0x%x, "
2158 "cmdlen %u\n",
2159 cbw->Lun, cbw->Flags, cbw->Length);
2160
2161 /*
2162 * We can do anything we want here, so let's stall the
2163 * bulk pipes if we are allowed to.
2164 */
2165 if (common->can_stall) {
2166 fsg_set_halt(fsg, fsg->bulk_out);
2167 halt_bulk_in_endpoint(fsg);
2168 }
2169 return -EINVAL;
2170 }
2171
2172 /* Save the command for later */
2173 common->cmnd_size = cbw->Length;
2174 memcpy(common->cmnd, cbw->CDB, common->cmnd_size);
2175 if (cbw->Flags & US_BULK_FLAG_IN)
2176 common->data_dir = DATA_DIR_TO_HOST;
2177 else
2178 common->data_dir = DATA_DIR_FROM_HOST;
2179 common->data_size = le32_to_cpu(cbw->DataTransferLength);
2180 if (common->data_size == 0)
2181 common->data_dir = DATA_DIR_NONE;
2182 common->lun = cbw->Lun;
2183 if (common->lun < ARRAY_SIZE(common->luns))
2184 common->curlun = common->luns[common->lun];
2185 else
2186 common->curlun = NULL;
2187 common->tag = cbw->Tag;
2188 return 0;
2189 }
2190
2191 static int get_next_command(struct fsg_common *common)
2192 {
2193 struct fsg_buffhd *bh;
2194 int rc = 0;
2195
2196 /* Wait for the next buffer to become available */
2197 bh = common->next_buffhd_to_fill;
2198 while (bh->state != BUF_STATE_EMPTY) {
2199 rc = sleep_thread(common, true);
2200 if (rc)
2201 return rc;
2202 }
2203
2204 /* Queue a request to read a Bulk-only CBW */
2205 set_bulk_out_req_length(common, bh, US_BULK_CB_WRAP_LEN);
2206 if (!start_out_transfer(common, bh))
2207 /* Don't know what to do if common->fsg is NULL */
2208 return -EIO;
2209
2210 /*
2211 * We will drain the buffer in software, which means we
2212 * can reuse it for the next filling. No need to advance
2213 * next_buffhd_to_fill.
2214 */
2215
2216 /* Wait for the CBW to arrive */
2217 while (bh->state != BUF_STATE_FULL) {
2218 rc = sleep_thread(common, true);
2219 if (rc)
2220 return rc;
2221 }
2222 smp_rmb();
2223 rc = fsg_is_set(common) ? received_cbw(common->fsg, bh) : -EIO;
2224 bh->state = BUF_STATE_EMPTY;
2225
2226 return rc;
2227 }
2228
2229
2230 /*-------------------------------------------------------------------------*/
2231
2232 static int alloc_request(struct fsg_common *common, struct usb_ep *ep,
2233 struct usb_request **preq)
2234 {
2235 *preq = usb_ep_alloc_request(ep, GFP_ATOMIC);
2236 if (*preq)
2237 return 0;
2238 ERROR(common, "can't allocate request for %s\n", ep->name);
2239 return -ENOMEM;
2240 }
2241
2242 /* Reset interface setting and re-init endpoint state (toggle etc). */
2243 static int do_set_interface(struct fsg_common *common, struct fsg_dev *new_fsg)
2244 {
2245 struct fsg_dev *fsg;
2246 int i, rc = 0;
2247
2248 if (common->running)
2249 DBG(common, "reset interface\n");
2250
2251 reset:
2252 /* Deallocate the requests */
2253 if (common->fsg) {
2254 fsg = common->fsg;
2255
2256 for (i = 0; i < common->fsg_num_buffers; ++i) {
2257 struct fsg_buffhd *bh = &common->buffhds[i];
2258
2259 if (bh->inreq) {
2260 usb_ep_free_request(fsg->bulk_in, bh->inreq);
2261 bh->inreq = NULL;
2262 }
2263 if (bh->outreq) {
2264 usb_ep_free_request(fsg->bulk_out, bh->outreq);
2265 bh->outreq = NULL;
2266 }
2267 }
2268
2269 /* Disable the endpoints */
2270 if (fsg->bulk_in_enabled) {
2271 usb_ep_disable(fsg->bulk_in);
2272 fsg->bulk_in_enabled = 0;
2273 }
2274 if (fsg->bulk_out_enabled) {
2275 usb_ep_disable(fsg->bulk_out);
2276 fsg->bulk_out_enabled = 0;
2277 }
2278
2279 common->fsg = NULL;
2280 wake_up(&common->fsg_wait);
2281 }
2282
2283 common->running = 0;
2284 if (!new_fsg || rc)
2285 return rc;
2286
2287 common->fsg = new_fsg;
2288 fsg = common->fsg;
2289
2290 /* Enable the endpoints */
2291 rc = config_ep_by_speed(common->gadget, &(fsg->function), fsg->bulk_in);
2292 if (rc)
2293 goto reset;
2294 rc = usb_ep_enable(fsg->bulk_in);
2295 if (rc)
2296 goto reset;
2297 fsg->bulk_in->driver_data = common;
2298 fsg->bulk_in_enabled = 1;
2299
2300 rc = config_ep_by_speed(common->gadget, &(fsg->function),
2301 fsg->bulk_out);
2302 if (rc)
2303 goto reset;
2304 rc = usb_ep_enable(fsg->bulk_out);
2305 if (rc)
2306 goto reset;
2307 fsg->bulk_out->driver_data = common;
2308 fsg->bulk_out_enabled = 1;
2309 common->bulk_out_maxpacket = usb_endpoint_maxp(fsg->bulk_out->desc);
2310 clear_bit(IGNORE_BULK_OUT, &fsg->atomic_bitflags);
2311
2312 /* Allocate the requests */
2313 for (i = 0; i < common->fsg_num_buffers; ++i) {
2314 struct fsg_buffhd *bh = &common->buffhds[i];
2315
2316 rc = alloc_request(common, fsg->bulk_in, &bh->inreq);
2317 if (rc)
2318 goto reset;
2319 rc = alloc_request(common, fsg->bulk_out, &bh->outreq);
2320 if (rc)
2321 goto reset;
2322 bh->inreq->buf = bh->outreq->buf = bh->buf;
2323 bh->inreq->context = bh->outreq->context = bh;
2324 bh->inreq->complete = bulk_in_complete;
2325 bh->outreq->complete = bulk_out_complete;
2326 }
2327
2328 common->running = 1;
2329 for (i = 0; i < ARRAY_SIZE(common->luns); ++i)
2330 if (common->luns[i])
2331 common->luns[i]->unit_attention_data =
2332 SS_RESET_OCCURRED;
2333 return rc;
2334 }
2335
2336
2337 /****************************** ALT CONFIGS ******************************/
2338
2339 static int fsg_set_alt(struct usb_function *f, unsigned intf, unsigned alt)
2340 {
2341 struct fsg_dev *fsg = fsg_from_func(f);
2342 fsg->common->new_fsg = fsg;
2343 raise_exception(fsg->common, FSG_STATE_CONFIG_CHANGE);
2344 return USB_GADGET_DELAYED_STATUS;
2345 }
2346
2347 static void fsg_disable(struct usb_function *f)
2348 {
2349 struct fsg_dev *fsg = fsg_from_func(f);
2350 fsg->common->new_fsg = NULL;
2351 raise_exception(fsg->common, FSG_STATE_CONFIG_CHANGE);
2352 }
2353
2354
2355 /*-------------------------------------------------------------------------*/
2356
2357 static void handle_exception(struct fsg_common *common)
2358 {
2359 int i;
2360 struct fsg_buffhd *bh;
2361 enum fsg_state old_state;
2362 struct fsg_lun *curlun;
2363 unsigned int exception_req_tag;
2364
2365 /*
2366 * Clear the existing signals. Anything but SIGUSR1 is converted
2367 * into a high-priority EXIT exception.
2368 */
2369 for (;;) {
2370 int sig = kernel_dequeue_signal(NULL);
2371 if (!sig)
2372 break;
2373 if (sig != SIGUSR1) {
2374 if (common->state < FSG_STATE_EXIT)
2375 DBG(common, "Main thread exiting on signal\n");
2376 raise_exception(common, FSG_STATE_EXIT);
2377 }
2378 }
2379
2380 /* Cancel all the pending transfers */
2381 if (likely(common->fsg)) {
2382 for (i = 0; i < common->fsg_num_buffers; ++i) {
2383 bh = &common->buffhds[i];
2384 if (bh->inreq_busy)
2385 usb_ep_dequeue(common->fsg->bulk_in, bh->inreq);
2386 if (bh->outreq_busy)
2387 usb_ep_dequeue(common->fsg->bulk_out,
2388 bh->outreq);
2389 }
2390
2391 /* Wait until everything is idle */
2392 for (;;) {
2393 int num_active = 0;
2394 for (i = 0; i < common->fsg_num_buffers; ++i) {
2395 bh = &common->buffhds[i];
2396 num_active += bh->inreq_busy + bh->outreq_busy;
2397 }
2398 if (num_active == 0)
2399 break;
2400 if (sleep_thread(common, true))
2401 return;
2402 }
2403
2404 /* Clear out the controller's fifos */
2405 if (common->fsg->bulk_in_enabled)
2406 usb_ep_fifo_flush(common->fsg->bulk_in);
2407 if (common->fsg->bulk_out_enabled)
2408 usb_ep_fifo_flush(common->fsg->bulk_out);
2409 }
2410
2411 /*
2412 * Reset the I/O buffer states and pointers, the SCSI
2413 * state, and the exception. Then invoke the handler.
2414 */
2415 spin_lock_irq(&common->lock);
2416
2417 for (i = 0; i < common->fsg_num_buffers; ++i) {
2418 bh = &common->buffhds[i];
2419 bh->state = BUF_STATE_EMPTY;
2420 }
2421 common->next_buffhd_to_fill = &common->buffhds[0];
2422 common->next_buffhd_to_drain = &common->buffhds[0];
2423 exception_req_tag = common->exception_req_tag;
2424 old_state = common->state;
2425
2426 if (old_state == FSG_STATE_ABORT_BULK_OUT)
2427 common->state = FSG_STATE_STATUS_PHASE;
2428 else {
2429 for (i = 0; i < ARRAY_SIZE(common->luns); ++i) {
2430 curlun = common->luns[i];
2431 if (!curlun)
2432 continue;
2433 curlun->prevent_medium_removal = 0;
2434 curlun->sense_data = SS_NO_SENSE;
2435 curlun->unit_attention_data = SS_NO_SENSE;
2436 curlun->sense_data_info = 0;
2437 curlun->info_valid = 0;
2438 }
2439 common->state = FSG_STATE_IDLE;
2440 }
2441 spin_unlock_irq(&common->lock);
2442
2443 /* Carry out any extra actions required for the exception */
2444 switch (old_state) {
2445 case FSG_STATE_ABORT_BULK_OUT:
2446 send_status(common);
2447 spin_lock_irq(&common->lock);
2448 if (common->state == FSG_STATE_STATUS_PHASE)
2449 common->state = FSG_STATE_IDLE;
2450 spin_unlock_irq(&common->lock);
2451 break;
2452
2453 case FSG_STATE_RESET:
2454 /*
2455 * In case we were forced against our will to halt a
2456 * bulk endpoint, clear the halt now. (The SuperH UDC
2457 * requires this.)
2458 */
2459 if (!fsg_is_set(common))
2460 break;
2461 if (test_and_clear_bit(IGNORE_BULK_OUT,
2462 &common->fsg->atomic_bitflags))
2463 usb_ep_clear_halt(common->fsg->bulk_in);
2464
2465 if (common->ep0_req_tag == exception_req_tag)
2466 ep0_queue(common); /* Complete the status stage */
2467
2468 /*
2469 * Technically this should go here, but it would only be
2470 * a waste of time. Ditto for the INTERFACE_CHANGE and
2471 * CONFIG_CHANGE cases.
2472 */
2473 /* for (i = 0; i < common->ARRAY_SIZE(common->luns); ++i) */
2474 /* if (common->luns[i]) */
2475 /* common->luns[i]->unit_attention_data = */
2476 /* SS_RESET_OCCURRED; */
2477 break;
2478
2479 case FSG_STATE_CONFIG_CHANGE:
2480 do_set_interface(common, common->new_fsg);
2481 if (common->new_fsg)
2482 usb_composite_setup_continue(common->cdev);
2483 break;
2484
2485 case FSG_STATE_EXIT:
2486 case FSG_STATE_TERMINATED:
2487 do_set_interface(common, NULL); /* Free resources */
2488 spin_lock_irq(&common->lock);
2489 common->state = FSG_STATE_TERMINATED; /* Stop the thread */
2490 spin_unlock_irq(&common->lock);
2491 break;
2492
2493 case FSG_STATE_INTERFACE_CHANGE:
2494 case FSG_STATE_DISCONNECT:
2495 case FSG_STATE_COMMAND_PHASE:
2496 case FSG_STATE_DATA_PHASE:
2497 case FSG_STATE_STATUS_PHASE:
2498 case FSG_STATE_IDLE:
2499 break;
2500 }
2501 }
2502
2503
2504 /*-------------------------------------------------------------------------*/
2505
2506 static int fsg_main_thread(void *common_)
2507 {
2508 struct fsg_common *common = common_;
2509
2510 /*
2511 * Allow the thread to be killed by a signal, but set the signal mask
2512 * to block everything but INT, TERM, KILL, and USR1.
2513 */
2514 allow_signal(SIGINT);
2515 allow_signal(SIGTERM);
2516 allow_signal(SIGKILL);
2517 allow_signal(SIGUSR1);
2518
2519 /* Allow the thread to be frozen */
2520 set_freezable();
2521
2522 /*
2523 * Arrange for userspace references to be interpreted as kernel
2524 * pointers. That way we can pass a kernel pointer to a routine
2525 * that expects a __user pointer and it will work okay.
2526 */
2527 set_fs(get_ds());
2528
2529 /* The main loop */
2530 while (common->state != FSG_STATE_TERMINATED) {
2531 if (exception_in_progress(common) || signal_pending(current)) {
2532 handle_exception(common);
2533 continue;
2534 }
2535
2536 if (!common->running) {
2537 sleep_thread(common, true);
2538 continue;
2539 }
2540
2541 if (get_next_command(common))
2542 continue;
2543
2544 spin_lock_irq(&common->lock);
2545 if (!exception_in_progress(common))
2546 common->state = FSG_STATE_DATA_PHASE;
2547 spin_unlock_irq(&common->lock);
2548
2549 if (do_scsi_command(common) || finish_reply(common))
2550 continue;
2551
2552 spin_lock_irq(&common->lock);
2553 if (!exception_in_progress(common))
2554 common->state = FSG_STATE_STATUS_PHASE;
2555 spin_unlock_irq(&common->lock);
2556
2557 if (send_status(common))
2558 continue;
2559
2560 spin_lock_irq(&common->lock);
2561 if (!exception_in_progress(common))
2562 common->state = FSG_STATE_IDLE;
2563 spin_unlock_irq(&common->lock);
2564 }
2565
2566 spin_lock_irq(&common->lock);
2567 common->thread_task = NULL;
2568 spin_unlock_irq(&common->lock);
2569
2570 if (!common->ops || !common->ops->thread_exits
2571 || common->ops->thread_exits(common) < 0) {
2572 int i;
2573
2574 down_write(&common->filesem);
2575 for (i = 0; i < ARRAY_SIZE(common->luns); --i) {
2576 struct fsg_lun *curlun = common->luns[i];
2577 if (!curlun || !fsg_lun_is_open(curlun))
2578 continue;
2579
2580 fsg_lun_close(curlun);
2581 curlun->unit_attention_data = SS_MEDIUM_NOT_PRESENT;
2582 }
2583 up_write(&common->filesem);
2584 }
2585
2586 /* Let fsg_unbind() know the thread has exited */
2587 complete_and_exit(&common->thread_notifier, 0);
2588 }
2589
2590
2591 /*************************** DEVICE ATTRIBUTES ***************************/
2592
2593 static ssize_t ro_show(struct device *dev, struct device_attribute *attr, char *buf)
2594 {
2595 struct fsg_lun *curlun = fsg_lun_from_dev(dev);
2596
2597 return fsg_show_ro(curlun, buf);
2598 }
2599
2600 static ssize_t nofua_show(struct device *dev, struct device_attribute *attr,
2601 char *buf)
2602 {
2603 struct fsg_lun *curlun = fsg_lun_from_dev(dev);
2604
2605 return fsg_show_nofua(curlun, buf);
2606 }
2607
2608 static ssize_t file_show(struct device *dev, struct device_attribute *attr,
2609 char *buf)
2610 {
2611 struct fsg_lun *curlun = fsg_lun_from_dev(dev);
2612 struct rw_semaphore *filesem = dev_get_drvdata(dev);
2613
2614 return fsg_show_file(curlun, filesem, buf);
2615 }
2616
2617 static ssize_t ro_store(struct device *dev, struct device_attribute *attr,
2618 const char *buf, size_t count)
2619 {
2620 struct fsg_lun *curlun = fsg_lun_from_dev(dev);
2621 struct rw_semaphore *filesem = dev_get_drvdata(dev);
2622
2623 return fsg_store_ro(curlun, filesem, buf, count);
2624 }
2625
2626 static ssize_t nofua_store(struct device *dev, struct device_attribute *attr,
2627 const char *buf, size_t count)
2628 {
2629 struct fsg_lun *curlun = fsg_lun_from_dev(dev);
2630
2631 return fsg_store_nofua(curlun, buf, count);
2632 }
2633
2634 static ssize_t file_store(struct device *dev, struct device_attribute *attr,
2635 const char *buf, size_t count)
2636 {
2637 struct fsg_lun *curlun = fsg_lun_from_dev(dev);
2638 struct rw_semaphore *filesem = dev_get_drvdata(dev);
2639
2640 return fsg_store_file(curlun, filesem, buf, count);
2641 }
2642
2643 static DEVICE_ATTR_RW(nofua);
2644 /* mode wil be set in fsg_lun_attr_is_visible() */
2645 static DEVICE_ATTR(ro, 0, ro_show, ro_store);
2646 static DEVICE_ATTR(file, 0, file_show, file_store);
2647
2648 /****************************** FSG COMMON ******************************/
2649
2650 static void fsg_common_release(struct kref *ref);
2651
2652 static void fsg_lun_release(struct device *dev)
2653 {
2654 /* Nothing needs to be done */
2655 }
2656
2657 void fsg_common_get(struct fsg_common *common)
2658 {
2659 kref_get(&common->ref);
2660 }
2661 EXPORT_SYMBOL_GPL(fsg_common_get);
2662
2663 void fsg_common_put(struct fsg_common *common)
2664 {
2665 kref_put(&common->ref, fsg_common_release);
2666 }
2667 EXPORT_SYMBOL_GPL(fsg_common_put);
2668
2669 static struct fsg_common *fsg_common_setup(struct fsg_common *common)
2670 {
2671 if (!common) {
2672 common = kzalloc(sizeof(*common), GFP_KERNEL);
2673 if (!common)
2674 return ERR_PTR(-ENOMEM);
2675 common->free_storage_on_release = 1;
2676 } else {
2677 common->free_storage_on_release = 0;
2678 }
2679 init_rwsem(&common->filesem);
2680 spin_lock_init(&common->lock);
2681 kref_init(&common->ref);
2682 init_completion(&common->thread_notifier);
2683 init_waitqueue_head(&common->fsg_wait);
2684 common->state = FSG_STATE_TERMINATED;
2685 memset(common->luns, 0, sizeof(common->luns));
2686
2687 return common;
2688 }
2689
2690 void fsg_common_set_sysfs(struct fsg_common *common, bool sysfs)
2691 {
2692 common->sysfs = sysfs;
2693 }
2694 EXPORT_SYMBOL_GPL(fsg_common_set_sysfs);
2695
2696 static void _fsg_common_free_buffers(struct fsg_buffhd *buffhds, unsigned n)
2697 {
2698 if (buffhds) {
2699 struct fsg_buffhd *bh = buffhds;
2700 while (n--) {
2701 kfree(bh->buf);
2702 ++bh;
2703 }
2704 kfree(buffhds);
2705 }
2706 }
2707
2708 int fsg_common_set_num_buffers(struct fsg_common *common, unsigned int n)
2709 {
2710 struct fsg_buffhd *bh, *buffhds;
2711 int i;
2712
2713 buffhds = kcalloc(n, sizeof(*buffhds), GFP_KERNEL);
2714 if (!buffhds)
2715 return -ENOMEM;
2716
2717 /* Data buffers cyclic list */
2718 bh = buffhds;
2719 i = n;
2720 goto buffhds_first_it;
2721 do {
2722 bh->next = bh + 1;
2723 ++bh;
2724 buffhds_first_it:
2725 bh->buf = kmalloc(FSG_BUFLEN, GFP_KERNEL);
2726 if (unlikely(!bh->buf))
2727 goto error_release;
2728 } while (--i);
2729 bh->next = buffhds;
2730
2731 _fsg_common_free_buffers(common->buffhds, common->fsg_num_buffers);
2732 common->fsg_num_buffers = n;
2733 common->buffhds = buffhds;
2734
2735 return 0;
2736
2737 error_release:
2738 /*
2739 * "buf"s pointed to by heads after n - i are NULL
2740 * so releasing them won't hurt
2741 */
2742 _fsg_common_free_buffers(buffhds, n);
2743
2744 return -ENOMEM;
2745 }
2746 EXPORT_SYMBOL_GPL(fsg_common_set_num_buffers);
2747
2748 void fsg_common_remove_lun(struct fsg_lun *lun)
2749 {
2750 if (device_is_registered(&lun->dev))
2751 device_unregister(&lun->dev);
2752 fsg_lun_close(lun);
2753 kfree(lun);
2754 }
2755 EXPORT_SYMBOL_GPL(fsg_common_remove_lun);
2756
2757 static void _fsg_common_remove_luns(struct fsg_common *common, int n)
2758 {
2759 int i;
2760
2761 for (i = 0; i < n; ++i)
2762 if (common->luns[i]) {
2763 fsg_common_remove_lun(common->luns[i]);
2764 common->luns[i] = NULL;
2765 }
2766 }
2767
2768 void fsg_common_remove_luns(struct fsg_common *common)
2769 {
2770 _fsg_common_remove_luns(common, ARRAY_SIZE(common->luns));
2771 }
2772 EXPORT_SYMBOL_GPL(fsg_common_remove_luns);
2773
2774 void fsg_common_set_ops(struct fsg_common *common,
2775 const struct fsg_operations *ops)
2776 {
2777 common->ops = ops;
2778 }
2779 EXPORT_SYMBOL_GPL(fsg_common_set_ops);
2780
2781 void fsg_common_free_buffers(struct fsg_common *common)
2782 {
2783 _fsg_common_free_buffers(common->buffhds, common->fsg_num_buffers);
2784 common->buffhds = NULL;
2785 }
2786 EXPORT_SYMBOL_GPL(fsg_common_free_buffers);
2787
2788 int fsg_common_set_cdev(struct fsg_common *common,
2789 struct usb_composite_dev *cdev, bool can_stall)
2790 {
2791 struct usb_string *us;
2792
2793 common->gadget = cdev->gadget;
2794 common->ep0 = cdev->gadget->ep0;
2795 common->ep0req = cdev->req;
2796 common->cdev = cdev;
2797
2798 us = usb_gstrings_attach(cdev, fsg_strings_array,
2799 ARRAY_SIZE(fsg_strings));
2800 if (IS_ERR(us))
2801 return PTR_ERR(us);
2802
2803 fsg_intf_desc.iInterface = us[FSG_STRING_INTERFACE].id;
2804
2805 /*
2806 * Some peripheral controllers are known not to be able to
2807 * halt bulk endpoints correctly. If one of them is present,
2808 * disable stalls.
2809 */
2810 common->can_stall = can_stall &&
2811 gadget_is_stall_supported(common->gadget);
2812
2813 return 0;
2814 }
2815 EXPORT_SYMBOL_GPL(fsg_common_set_cdev);
2816
2817 static struct attribute *fsg_lun_dev_attrs[] = {
2818 &dev_attr_ro.attr,
2819 &dev_attr_file.attr,
2820 &dev_attr_nofua.attr,
2821 NULL
2822 };
2823
2824 static umode_t fsg_lun_dev_is_visible(struct kobject *kobj,
2825 struct attribute *attr, int idx)
2826 {
2827 struct device *dev = kobj_to_dev(kobj);
2828 struct fsg_lun *lun = fsg_lun_from_dev(dev);
2829
2830 if (attr == &dev_attr_ro.attr)
2831 return lun->cdrom ? S_IRUGO : (S_IWUSR | S_IRUGO);
2832 if (attr == &dev_attr_file.attr)
2833 return lun->removable ? (S_IWUSR | S_IRUGO) : S_IRUGO;
2834 return attr->mode;
2835 }
2836
2837 static const struct attribute_group fsg_lun_dev_group = {
2838 .attrs = fsg_lun_dev_attrs,
2839 .is_visible = fsg_lun_dev_is_visible,
2840 };
2841
2842 static const struct attribute_group *fsg_lun_dev_groups[] = {
2843 &fsg_lun_dev_group,
2844 NULL
2845 };
2846
2847 int fsg_common_create_lun(struct fsg_common *common, struct fsg_lun_config *cfg,
2848 unsigned int id, const char *name,
2849 const char **name_pfx)
2850 {
2851 struct fsg_lun *lun;
2852 char *pathbuf, *p;
2853 int rc = -ENOMEM;
2854
2855 if (id >= ARRAY_SIZE(common->luns))
2856 return -ENODEV;
2857
2858 if (common->luns[id])
2859 return -EBUSY;
2860
2861 if (!cfg->filename && !cfg->removable) {
2862 pr_err("no file given for LUN%d\n", id);
2863 return -EINVAL;
2864 }
2865
2866 lun = kzalloc(sizeof(*lun), GFP_KERNEL);
2867 if (!lun)
2868 return -ENOMEM;
2869
2870 lun->name_pfx = name_pfx;
2871
2872 lun->cdrom = !!cfg->cdrom;
2873 lun->ro = cfg->cdrom || cfg->ro;
2874 lun->initially_ro = lun->ro;
2875 lun->removable = !!cfg->removable;
2876
2877 if (!common->sysfs) {
2878 /* we DON'T own the name!*/
2879 lun->name = name;
2880 } else {
2881 lun->dev.release = fsg_lun_release;
2882 lun->dev.parent = &common->gadget->dev;
2883 lun->dev.groups = fsg_lun_dev_groups;
2884 dev_set_drvdata(&lun->dev, &common->filesem);
2885 dev_set_name(&lun->dev, "%s", name);
2886 lun->name = dev_name(&lun->dev);
2887
2888 rc = device_register(&lun->dev);
2889 if (rc) {
2890 pr_info("failed to register LUN%d: %d\n", id, rc);
2891 put_device(&lun->dev);
2892 goto error_sysfs;
2893 }
2894 }
2895
2896 common->luns[id] = lun;
2897
2898 if (cfg->filename) {
2899 rc = fsg_lun_open(lun, cfg->filename);
2900 if (rc)
2901 goto error_lun;
2902 }
2903
2904 pathbuf = kmalloc(PATH_MAX, GFP_KERNEL);
2905 p = "(no medium)";
2906 if (fsg_lun_is_open(lun)) {
2907 p = "(error)";
2908 if (pathbuf) {
2909 p = file_path(lun->filp, pathbuf, PATH_MAX);
2910 if (IS_ERR(p))
2911 p = "(error)";
2912 }
2913 }
2914 pr_info("LUN: %s%s%sfile: %s\n",
2915 lun->removable ? "removable " : "",
2916 lun->ro ? "read only " : "",
2917 lun->cdrom ? "CD-ROM " : "",
2918 p);
2919 kfree(pathbuf);
2920
2921 return 0;
2922
2923 error_lun:
2924 if (device_is_registered(&lun->dev))
2925 device_unregister(&lun->dev);
2926 fsg_lun_close(lun);
2927 common->luns[id] = NULL;
2928 error_sysfs:
2929 kfree(lun);
2930 return rc;
2931 }
2932 EXPORT_SYMBOL_GPL(fsg_common_create_lun);
2933
2934 int fsg_common_create_luns(struct fsg_common *common, struct fsg_config *cfg)
2935 {
2936 char buf[8]; /* enough for 100000000 different numbers, decimal */
2937 int i, rc;
2938
2939 fsg_common_remove_luns(common);
2940
2941 for (i = 0; i < cfg->nluns; ++i) {
2942 snprintf(buf, sizeof(buf), "lun%d", i);
2943 rc = fsg_common_create_lun(common, &cfg->luns[i], i, buf, NULL);
2944 if (rc)
2945 goto fail;
2946 }
2947
2948 pr_info("Number of LUNs=%d\n", cfg->nluns);
2949
2950 return 0;
2951
2952 fail:
2953 _fsg_common_remove_luns(common, i);
2954 return rc;
2955 }
2956 EXPORT_SYMBOL_GPL(fsg_common_create_luns);
2957
2958 void fsg_common_set_inquiry_string(struct fsg_common *common, const char *vn,
2959 const char *pn)
2960 {
2961 int i;
2962
2963 /* Prepare inquiryString */
2964 i = get_default_bcdDevice();
2965 snprintf(common->inquiry_string, sizeof(common->inquiry_string),
2966 "%-8s%-16s%04x", vn ?: "Linux",
2967 /* Assume product name dependent on the first LUN */
2968 pn ?: ((*common->luns)->cdrom
2969 ? "File-CD Gadget"
2970 : "File-Stor Gadget"),
2971 i);
2972 }
2973 EXPORT_SYMBOL_GPL(fsg_common_set_inquiry_string);
2974
2975 static void fsg_common_release(struct kref *ref)
2976 {
2977 struct fsg_common *common = container_of(ref, struct fsg_common, ref);
2978 int i;
2979
2980 /* If the thread isn't already dead, tell it to exit now */
2981 if (common->state != FSG_STATE_TERMINATED) {
2982 raise_exception(common, FSG_STATE_EXIT);
2983 wait_for_completion(&common->thread_notifier);
2984 common->thread_task = NULL;
2985 }
2986
2987 for (i = 0; i < ARRAY_SIZE(common->luns); ++i) {
2988 struct fsg_lun *lun = common->luns[i];
2989 if (!lun)
2990 continue;
2991 fsg_lun_close(lun);
2992 if (device_is_registered(&lun->dev))
2993 device_unregister(&lun->dev);
2994 kfree(lun);
2995 }
2996
2997 _fsg_common_free_buffers(common->buffhds, common->fsg_num_buffers);
2998 if (common->free_storage_on_release)
2999 kfree(common);
3000 }
3001
3002
3003 /*-------------------------------------------------------------------------*/
3004
3005 static int fsg_bind(struct usb_configuration *c, struct usb_function *f)
3006 {
3007 struct fsg_dev *fsg = fsg_from_func(f);
3008 struct fsg_common *common = fsg->common;
3009 struct usb_gadget *gadget = c->cdev->gadget;
3010 int i;
3011 struct usb_ep *ep;
3012 unsigned max_burst;
3013 int ret;
3014 struct fsg_opts *opts;
3015
3016 /* Don't allow to bind if we don't have at least one LUN */
3017 ret = _fsg_common_get_max_lun(common);
3018 if (ret < 0) {
3019 pr_err("There should be at least one LUN.\n");
3020 return -EINVAL;
3021 }
3022
3023 opts = fsg_opts_from_func_inst(f->fi);
3024 if (!opts->no_configfs) {
3025 ret = fsg_common_set_cdev(fsg->common, c->cdev,
3026 fsg->common->can_stall);
3027 if (ret)
3028 return ret;
3029 fsg_common_set_inquiry_string(fsg->common, NULL, NULL);
3030 }
3031
3032 if (!common->thread_task) {
3033 common->state = FSG_STATE_IDLE;
3034 common->thread_task =
3035 kthread_create(fsg_main_thread, common, "file-storage");
3036 if (IS_ERR(common->thread_task)) {
3037 int ret = PTR_ERR(common->thread_task);
3038 common->thread_task = NULL;
3039 common->state = FSG_STATE_TERMINATED;
3040 return ret;
3041 }
3042 DBG(common, "I/O thread pid: %d\n",
3043 task_pid_nr(common->thread_task));
3044 wake_up_process(common->thread_task);
3045 }
3046
3047 fsg->gadget = gadget;
3048
3049 /* New interface */
3050 i = usb_interface_id(c, f);
3051 if (i < 0)
3052 goto fail;
3053 fsg_intf_desc.bInterfaceNumber = i;
3054 fsg->interface_number = i;
3055
3056 /* Find all the endpoints we will use */
3057 ep = usb_ep_autoconfig(gadget, &fsg_fs_bulk_in_desc);
3058 if (!ep)
3059 goto autoconf_fail;
3060 fsg->bulk_in = ep;
3061
3062 ep = usb_ep_autoconfig(gadget, &fsg_fs_bulk_out_desc);
3063 if (!ep)
3064 goto autoconf_fail;
3065 fsg->bulk_out = ep;
3066
3067 /* Assume endpoint addresses are the same for both speeds */
3068 fsg_hs_bulk_in_desc.bEndpointAddress =
3069 fsg_fs_bulk_in_desc.bEndpointAddress;
3070 fsg_hs_bulk_out_desc.bEndpointAddress =
3071 fsg_fs_bulk_out_desc.bEndpointAddress;
3072
3073 /* Calculate bMaxBurst, we know packet size is 1024 */
3074 max_burst = min_t(unsigned, FSG_BUFLEN / 1024, 15);
3075
3076 fsg_ss_bulk_in_desc.bEndpointAddress =
3077 fsg_fs_bulk_in_desc.bEndpointAddress;
3078 fsg_ss_bulk_in_comp_desc.bMaxBurst = max_burst;
3079
3080 fsg_ss_bulk_out_desc.bEndpointAddress =
3081 fsg_fs_bulk_out_desc.bEndpointAddress;
3082 fsg_ss_bulk_out_comp_desc.bMaxBurst = max_burst;
3083
3084 ret = usb_assign_descriptors(f, fsg_fs_function, fsg_hs_function,
3085 fsg_ss_function, fsg_ss_function);
3086 if (ret)
3087 goto autoconf_fail;
3088
3089 return 0;
3090
3091 autoconf_fail:
3092 ERROR(fsg, "unable to autoconfigure all endpoints\n");
3093 i = -ENOTSUPP;
3094 fail:
3095 /* terminate the thread */
3096 if (fsg->common->state != FSG_STATE_TERMINATED) {
3097 raise_exception(fsg->common, FSG_STATE_EXIT);
3098 wait_for_completion(&fsg->common->thread_notifier);
3099 }
3100 return i;
3101 }
3102
3103 /****************************** ALLOCATE FUNCTION *************************/
3104
3105 static void fsg_unbind(struct usb_configuration *c, struct usb_function *f)
3106 {
3107 struct fsg_dev *fsg = fsg_from_func(f);
3108 struct fsg_common *common = fsg->common;
3109
3110 DBG(fsg, "unbind\n");
3111 if (fsg->common->fsg == fsg) {
3112 fsg->common->new_fsg = NULL;
3113 raise_exception(fsg->common, FSG_STATE_CONFIG_CHANGE);
3114 /* FIXME: make interruptible or killable somehow? */
3115 wait_event(common->fsg_wait, common->fsg != fsg);
3116 }
3117
3118 usb_free_all_descriptors(&fsg->function);
3119 }
3120
3121 static inline struct fsg_lun_opts *to_fsg_lun_opts(struct config_item *item)
3122 {
3123 return container_of(to_config_group(item), struct fsg_lun_opts, group);
3124 }
3125
3126 static inline struct fsg_opts *to_fsg_opts(struct config_item *item)
3127 {
3128 return container_of(to_config_group(item), struct fsg_opts,
3129 func_inst.group);
3130 }
3131
3132 static void fsg_lun_attr_release(struct config_item *item)
3133 {
3134 struct fsg_lun_opts *lun_opts;
3135
3136 lun_opts = to_fsg_lun_opts(item);
3137 kfree(lun_opts);
3138 }
3139
3140 static struct configfs_item_operations fsg_lun_item_ops = {
3141 .release = fsg_lun_attr_release,
3142 };
3143
3144 static ssize_t fsg_lun_opts_file_show(struct config_item *item, char *page)
3145 {
3146 struct fsg_lun_opts *opts = to_fsg_lun_opts(item);
3147 struct fsg_opts *fsg_opts = to_fsg_opts(opts->group.cg_item.ci_parent);
3148
3149 return fsg_show_file(opts->lun, &fsg_opts->common->filesem, page);
3150 }
3151
3152 static ssize_t fsg_lun_opts_file_store(struct config_item *item,
3153 const char *page, size_t len)
3154 {
3155 struct fsg_lun_opts *opts = to_fsg_lun_opts(item);
3156 struct fsg_opts *fsg_opts = to_fsg_opts(opts->group.cg_item.ci_parent);
3157
3158 return fsg_store_file(opts->lun, &fsg_opts->common->filesem, page, len);
3159 }
3160
3161 CONFIGFS_ATTR(fsg_lun_opts_, file);
3162
3163 static ssize_t fsg_lun_opts_ro_show(struct config_item *item, char *page)
3164 {
3165 return fsg_show_ro(to_fsg_lun_opts(item)->lun, page);
3166 }
3167
3168 static ssize_t fsg_lun_opts_ro_store(struct config_item *item,
3169 const char *page, size_t len)
3170 {
3171 struct fsg_lun_opts *opts = to_fsg_lun_opts(item);
3172 struct fsg_opts *fsg_opts = to_fsg_opts(opts->group.cg_item.ci_parent);
3173
3174 return fsg_store_ro(opts->lun, &fsg_opts->common->filesem, page, len);
3175 }
3176
3177 CONFIGFS_ATTR(fsg_lun_opts_, ro);
3178
3179 static ssize_t fsg_lun_opts_removable_show(struct config_item *item,
3180 char *page)
3181 {
3182 return fsg_show_removable(to_fsg_lun_opts(item)->lun, page);
3183 }
3184
3185 static ssize_t fsg_lun_opts_removable_store(struct config_item *item,
3186 const char *page, size_t len)
3187 {
3188 return fsg_store_removable(to_fsg_lun_opts(item)->lun, page, len);
3189 }
3190
3191 CONFIGFS_ATTR(fsg_lun_opts_, removable);
3192
3193 static ssize_t fsg_lun_opts_cdrom_show(struct config_item *item, char *page)
3194 {
3195 return fsg_show_cdrom(to_fsg_lun_opts(item)->lun, page);
3196 }
3197
3198 static ssize_t fsg_lun_opts_cdrom_store(struct config_item *item,
3199 const char *page, size_t len)
3200 {
3201 struct fsg_lun_opts *opts = to_fsg_lun_opts(item);
3202 struct fsg_opts *fsg_opts = to_fsg_opts(opts->group.cg_item.ci_parent);
3203
3204 return fsg_store_cdrom(opts->lun, &fsg_opts->common->filesem, page,
3205 len);
3206 }
3207
3208 CONFIGFS_ATTR(fsg_lun_opts_, cdrom);
3209
3210 static ssize_t fsg_lun_opts_nofua_show(struct config_item *item, char *page)
3211 {
3212 return fsg_show_nofua(to_fsg_lun_opts(item)->lun, page);
3213 }
3214
3215 static ssize_t fsg_lun_opts_nofua_store(struct config_item *item,
3216 const char *page, size_t len)
3217 {
3218 return fsg_store_nofua(to_fsg_lun_opts(item)->lun, page, len);
3219 }
3220
3221 CONFIGFS_ATTR(fsg_lun_opts_, nofua);
3222
3223 static ssize_t fsg_lun_opts_inquiry_string_show(struct config_item *item,
3224 char *page)
3225 {
3226 return fsg_show_inquiry_string(to_fsg_lun_opts(item)->lun, page);
3227 }
3228
3229 static ssize_t fsg_lun_opts_inquiry_string_store(struct config_item *item,
3230 const char *page, size_t len)
3231 {
3232 return fsg_store_inquiry_string(to_fsg_lun_opts(item)->lun, page, len);
3233 }
3234
3235 CONFIGFS_ATTR(fsg_lun_opts_, inquiry_string);
3236
3237 static struct configfs_attribute *fsg_lun_attrs[] = {
3238 &fsg_lun_opts_attr_file,
3239 &fsg_lun_opts_attr_ro,
3240 &fsg_lun_opts_attr_removable,
3241 &fsg_lun_opts_attr_cdrom,
3242 &fsg_lun_opts_attr_nofua,
3243 &fsg_lun_opts_attr_inquiry_string,
3244 NULL,
3245 };
3246
3247 static struct config_item_type fsg_lun_type = {
3248 .ct_item_ops = &fsg_lun_item_ops,
3249 .ct_attrs = fsg_lun_attrs,
3250 .ct_owner = THIS_MODULE,
3251 };
3252
3253 static struct config_group *fsg_lun_make(struct config_group *group,
3254 const char *name)
3255 {
3256 struct fsg_lun_opts *opts;
3257 struct fsg_opts *fsg_opts;
3258 struct fsg_lun_config config;
3259 char *num_str;
3260 u8 num;
3261 int ret;
3262
3263 num_str = strchr(name, '.');
3264 if (!num_str) {
3265 pr_err("Unable to locate . in LUN.NUMBER\n");
3266 return ERR_PTR(-EINVAL);
3267 }
3268 num_str++;
3269
3270 ret = kstrtou8(num_str, 0, &num);
3271 if (ret)
3272 return ERR_PTR(ret);
3273
3274 fsg_opts = to_fsg_opts(&group->cg_item);
3275 if (num >= FSG_MAX_LUNS)
3276 return ERR_PTR(-ERANGE);
3277
3278 mutex_lock(&fsg_opts->lock);
3279 if (fsg_opts->refcnt || fsg_opts->common->luns[num]) {
3280 ret = -EBUSY;
3281 goto out;
3282 }
3283
3284 opts = kzalloc(sizeof(*opts), GFP_KERNEL);
3285 if (!opts) {
3286 ret = -ENOMEM;
3287 goto out;
3288 }
3289
3290 memset(&config, 0, sizeof(config));
3291 config.removable = true;
3292
3293 ret = fsg_common_create_lun(fsg_opts->common, &config, num, name,
3294 (const char **)&group->cg_item.ci_name);
3295 if (ret) {
3296 kfree(opts);
3297 goto out;
3298 }
3299 opts->lun = fsg_opts->common->luns[num];
3300 opts->lun_id = num;
3301 mutex_unlock(&fsg_opts->lock);
3302
3303 config_group_init_type_name(&opts->group, name, &fsg_lun_type);
3304
3305 return &opts->group;
3306 out:
3307 mutex_unlock(&fsg_opts->lock);
3308 return ERR_PTR(ret);
3309 }
3310
3311 static void fsg_lun_drop(struct config_group *group, struct config_item *item)
3312 {
3313 struct fsg_lun_opts *lun_opts;
3314 struct fsg_opts *fsg_opts;
3315
3316 lun_opts = to_fsg_lun_opts(item);
3317 fsg_opts = to_fsg_opts(&group->cg_item);
3318
3319 mutex_lock(&fsg_opts->lock);
3320 if (fsg_opts->refcnt) {
3321 struct config_item *gadget;
3322
3323 gadget = group->cg_item.ci_parent->ci_parent;
3324 unregister_gadget_item(gadget);
3325 }
3326
3327 fsg_common_remove_lun(lun_opts->lun);
3328 fsg_opts->common->luns[lun_opts->lun_id] = NULL;
3329 lun_opts->lun_id = 0;
3330 mutex_unlock(&fsg_opts->lock);
3331
3332 config_item_put(item);
3333 }
3334
3335 static void fsg_attr_release(struct config_item *item)
3336 {
3337 struct fsg_opts *opts = to_fsg_opts(item);
3338
3339 usb_put_function_instance(&opts->func_inst);
3340 }
3341
3342 static struct configfs_item_operations fsg_item_ops = {
3343 .release = fsg_attr_release,
3344 };
3345
3346 static ssize_t fsg_opts_stall_show(struct config_item *item, char *page)
3347 {
3348 struct fsg_opts *opts = to_fsg_opts(item);
3349 int result;
3350
3351 mutex_lock(&opts->lock);
3352 result = sprintf(page, "%d", opts->common->can_stall);
3353 mutex_unlock(&opts->lock);
3354
3355 return result;
3356 }
3357
3358 static ssize_t fsg_opts_stall_store(struct config_item *item, const char *page,
3359 size_t len)
3360 {
3361 struct fsg_opts *opts = to_fsg_opts(item);
3362 int ret;
3363 bool stall;
3364
3365 mutex_lock(&opts->lock);
3366
3367 if (opts->refcnt) {
3368 mutex_unlock(&opts->lock);
3369 return -EBUSY;
3370 }
3371
3372 ret = strtobool(page, &stall);
3373 if (!ret) {
3374 opts->common->can_stall = stall;
3375 ret = len;
3376 }
3377
3378 mutex_unlock(&opts->lock);
3379
3380 return ret;
3381 }
3382
3383 CONFIGFS_ATTR(fsg_opts_, stall);
3384
3385 #ifdef CONFIG_USB_GADGET_DEBUG_FILES
3386 static ssize_t fsg_opts_num_buffers_show(struct config_item *item, char *page)
3387 {
3388 struct fsg_opts *opts = to_fsg_opts(item);
3389 int result;
3390
3391 mutex_lock(&opts->lock);
3392 result = sprintf(page, "%d", opts->common->fsg_num_buffers);
3393 mutex_unlock(&opts->lock);
3394
3395 return result;
3396 }
3397
3398 static ssize_t fsg_opts_num_buffers_store(struct config_item *item,
3399 const char *page, size_t len)
3400 {
3401 struct fsg_opts *opts = to_fsg_opts(item);
3402 int ret;
3403 u8 num;
3404
3405 mutex_lock(&opts->lock);
3406 if (opts->refcnt) {
3407 ret = -EBUSY;
3408 goto end;
3409 }
3410 ret = kstrtou8(page, 0, &num);
3411 if (ret)
3412 goto end;
3413
3414 fsg_common_set_num_buffers(opts->common, num);
3415 ret = len;
3416
3417 end:
3418 mutex_unlock(&opts->lock);
3419 return ret;
3420 }
3421
3422 CONFIGFS_ATTR(fsg_opts_, num_buffers);
3423 #endif
3424
3425 static struct configfs_attribute *fsg_attrs[] = {
3426 &fsg_opts_attr_stall,
3427 #ifdef CONFIG_USB_GADGET_DEBUG_FILES
3428 &fsg_opts_attr_num_buffers,
3429 #endif
3430 NULL,
3431 };
3432
3433 static struct configfs_group_operations fsg_group_ops = {
3434 .make_group = fsg_lun_make,
3435 .drop_item = fsg_lun_drop,
3436 };
3437
3438 static struct config_item_type fsg_func_type = {
3439 .ct_item_ops = &fsg_item_ops,
3440 .ct_group_ops = &fsg_group_ops,
3441 .ct_attrs = fsg_attrs,
3442 .ct_owner = THIS_MODULE,
3443 };
3444
3445 static void fsg_free_inst(struct usb_function_instance *fi)
3446 {
3447 struct fsg_opts *opts;
3448
3449 opts = fsg_opts_from_func_inst(fi);
3450 fsg_common_put(opts->common);
3451 kfree(opts);
3452 }
3453
3454 static struct usb_function_instance *fsg_alloc_inst(void)
3455 {
3456 struct fsg_opts *opts;
3457 struct fsg_lun_config config;
3458 int rc;
3459
3460 opts = kzalloc(sizeof(*opts), GFP_KERNEL);
3461 if (!opts)
3462 return ERR_PTR(-ENOMEM);
3463 mutex_init(&opts->lock);
3464 opts->func_inst.free_func_inst = fsg_free_inst;
3465 opts->common = fsg_common_setup(opts->common);
3466 if (IS_ERR(opts->common)) {
3467 rc = PTR_ERR(opts->common);
3468 goto release_opts;
3469 }
3470
3471 rc = fsg_common_set_num_buffers(opts->common,
3472 CONFIG_USB_GADGET_STORAGE_NUM_BUFFERS);
3473 if (rc)
3474 goto release_opts;
3475
3476 pr_info(FSG_DRIVER_DESC ", version: " FSG_DRIVER_VERSION "\n");
3477
3478 memset(&config, 0, sizeof(config));
3479 config.removable = true;
3480 rc = fsg_common_create_lun(opts->common, &config, 0, "lun.0",
3481 (const char **)&opts->func_inst.group.cg_item.ci_name);
3482 if (rc)
3483 goto release_buffers;
3484
3485 opts->lun0.lun = opts->common->luns[0];
3486 opts->lun0.lun_id = 0;
3487
3488 config_group_init_type_name(&opts->func_inst.group, "", &fsg_func_type);
3489
3490 config_group_init_type_name(&opts->lun0.group, "lun.0", &fsg_lun_type);
3491 configfs_add_default_group(&opts->lun0.group, &opts->func_inst.group);
3492
3493 return &opts->func_inst;
3494
3495 release_buffers:
3496 fsg_common_free_buffers(opts->common);
3497 release_opts:
3498 kfree(opts);
3499 return ERR_PTR(rc);
3500 }
3501
3502 static void fsg_free(struct usb_function *f)
3503 {
3504 struct fsg_dev *fsg;
3505 struct fsg_opts *opts;
3506
3507 fsg = container_of(f, struct fsg_dev, function);
3508 opts = container_of(f->fi, struct fsg_opts, func_inst);
3509
3510 mutex_lock(&opts->lock);
3511 opts->refcnt--;
3512 mutex_unlock(&opts->lock);
3513
3514 kfree(fsg);
3515 }
3516
3517 static struct usb_function *fsg_alloc(struct usb_function_instance *fi)
3518 {
3519 struct fsg_opts *opts = fsg_opts_from_func_inst(fi);
3520 struct fsg_common *common = opts->common;
3521 struct fsg_dev *fsg;
3522
3523 fsg = kzalloc(sizeof(*fsg), GFP_KERNEL);
3524 if (unlikely(!fsg))
3525 return ERR_PTR(-ENOMEM);
3526
3527 mutex_lock(&opts->lock);
3528 opts->refcnt++;
3529 mutex_unlock(&opts->lock);
3530
3531 fsg->function.name = FSG_DRIVER_DESC;
3532 fsg->function.bind = fsg_bind;
3533 fsg->function.unbind = fsg_unbind;
3534 fsg->function.setup = fsg_setup;
3535 fsg->function.set_alt = fsg_set_alt;
3536 fsg->function.disable = fsg_disable;
3537 fsg->function.free_func = fsg_free;
3538
3539 fsg->common = common;
3540
3541 return &fsg->function;
3542 }
3543
3544 DECLARE_USB_FUNCTION_INIT(mass_storage, fsg_alloc_inst, fsg_alloc);
3545 MODULE_LICENSE("GPL");
3546 MODULE_AUTHOR("Michal Nazarewicz");
3547
3548 /************************* Module parameters *************************/
3549
3550
3551 void fsg_config_from_params(struct fsg_config *cfg,
3552 const struct fsg_module_parameters *params,
3553 unsigned int fsg_num_buffers)
3554 {
3555 struct fsg_lun_config *lun;
3556 unsigned i;
3557
3558 /* Configure LUNs */
3559 cfg->nluns =
3560 min(params->luns ?: (params->file_count ?: 1u),
3561 (unsigned)FSG_MAX_LUNS);
3562 for (i = 0, lun = cfg->luns; i < cfg->nluns; ++i, ++lun) {
3563 lun->ro = !!params->ro[i];
3564 lun->cdrom = !!params->cdrom[i];
3565 lun->removable = !!params->removable[i];
3566 lun->filename =
3567 params->file_count > i && params->file[i][0]
3568 ? params->file[i]
3569 : NULL;
3570 }
3571
3572 /* Let MSF use defaults */
3573 cfg->vendor_name = NULL;
3574 cfg->product_name = NULL;
3575
3576 cfg->ops = NULL;
3577 cfg->private_data = NULL;
3578
3579 /* Finalise */
3580 cfg->can_stall = params->stall;
3581 cfg->fsg_num_buffers = fsg_num_buffers;
3582 }
3583 EXPORT_SYMBOL_GPL(fsg_config_from_params);