]>
Commit | Line | Data |
---|---|---|
f82bd046 HJ |
1 | /* |
2 | * | |
3 | * Copyright (c) 2009, Microsoft Corporation. | |
4 | * | |
5 | * This program is free software; you can redistribute it and/or modify it | |
6 | * under the terms and conditions of the GNU General Public License, | |
7 | * version 2, as published by the Free Software Foundation. | |
8 | * | |
9 | * This program is distributed in the hope it will be useful, but WITHOUT | |
10 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or | |
11 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for | |
12 | * more details. | |
13 | * | |
14 | * You should have received a copy of the GNU General Public License along with | |
15 | * this program; if not, write to the Free Software Foundation, Inc., 59 Temple | |
16 | * Place - Suite 330, Boston, MA 02111-1307 USA. | |
17 | * | |
18 | * Authors: | |
19 | * Hank Janssen <hjanssen@microsoft.com> | |
20 | * | |
21 | */ | |
22 | ||
f82bd046 HJ |
23 | #include <linux/init.h> |
24 | #include <linux/module.h> | |
25 | #include <linux/device.h> | |
26 | #include <linux/blkdev.h> | |
27 | #include <linux/major.h> | |
28 | #include <linux/delay.h> | |
29 | #include <linux/hdreg.h> | |
30 | ||
31 | #include <scsi/scsi.h> | |
32 | #include <scsi/scsi_cmnd.h> | |
33 | #include <scsi/scsi_eh.h> | |
34 | #include <scsi/scsi_dbg.h> | |
35 | ||
4983b39a | 36 | #include "osd.h" |
0fce4c2f | 37 | #include "include/logging.h" |
870cde80 | 38 | #include "vmbus.h" |
f82bd046 | 39 | |
0fce4c2f | 40 | #include "include/StorVscApi.h" |
f82bd046 | 41 | |
454f18a9 BP |
42 | |
43 | /* #defines */ | |
44 | ||
f82bd046 HJ |
45 | #define BLKVSC_MINORS 64 |
46 | ||
454f18a9 BP |
47 | |
48 | /* Data types */ | |
49 | ||
f82bd046 HJ |
50 | enum blkvsc_device_type { |
51 | UNKNOWN_DEV_TYPE, | |
52 | HARDDISK_TYPE, | |
53 | DVD_TYPE, | |
54 | }; | |
55 | ||
454f18a9 BP |
56 | /* |
57 | * This request ties the struct request and struct | |
0b3f6834 | 58 | * blkvsc_request/hv_storvsc_request together A struct request may be |
454f18a9 BP |
59 | * represented by 1 or more struct blkvsc_request |
60 | */ | |
f82bd046 HJ |
61 | struct blkvsc_request_group { |
62 | int outstanding; | |
63 | int status; | |
64 | ||
454f18a9 | 65 | struct list_head blkvsc_req_list; /* list of blkvsc_requests */ |
f82bd046 HJ |
66 | }; |
67 | ||
68 | ||
69 | struct blkvsc_request { | |
454f18a9 | 70 | struct list_head req_entry; /* blkvsc_request_group.blkvsc_req_list */ |
f82bd046 | 71 | |
454f18a9 | 72 | struct list_head pend_entry; /* block_device_context.pending_list */ |
f82bd046 | 73 | |
454f18a9 | 74 | struct request *req; /* This may be null if we generate a request internally */ |
f82bd046 | 75 | struct block_device_context *dev; |
454f18a9 | 76 | struct blkvsc_request_group *group; /* The group this request is part of. Maybe null */ |
f82bd046 HJ |
77 | |
78 | wait_queue_head_t wevent; | |
79 | int cond; | |
80 | ||
81 | int write; | |
82 | sector_t sector_start; | |
83 | unsigned long sector_count; | |
84 | ||
85 | unsigned char sense_buffer[SCSI_SENSE_BUFFERSIZE]; | |
86 | unsigned char cmd_len; | |
87 | unsigned char cmnd[MAX_COMMAND_SIZE]; | |
88 | ||
0b3f6834 | 89 | struct hv_storvsc_request request; |
454f18a9 BP |
90 | /* !!!DO NOT ADD ANYTHING BELOW HERE!!! Otherwise, memory can overlap, because - */ |
91 | /* The extension buffer falls right here and is pointed to by request.Extension; */ | |
f82bd046 HJ |
92 | }; |
93 | ||
454f18a9 | 94 | /* Per device structure */ |
f82bd046 | 95 | struct block_device_context { |
454f18a9 | 96 | struct device_context *device_ctx; /* point back to our device context */ |
f82bd046 HJ |
97 | struct kmem_cache *request_pool; |
98 | spinlock_t lock; | |
99 | struct gendisk *gd; | |
100 | enum blkvsc_device_type device_type; | |
101 | struct list_head pending_list; | |
102 | ||
103 | unsigned char device_id[64]; | |
104 | unsigned int device_id_len; | |
105 | int num_outstanding_reqs; | |
106 | int shutting_down; | |
107 | int media_not_present; | |
108 | unsigned int sector_size; | |
109 | sector_t capacity; | |
110 | unsigned int port; | |
111 | unsigned char path; | |
112 | unsigned char target; | |
113 | int users; | |
114 | }; | |
115 | ||
454f18a9 | 116 | /* Per driver */ |
f82bd046 | 117 | struct blkvsc_driver_context { |
454f18a9 | 118 | /* !! These must be the first 2 fields !! */ |
f82bd046 | 119 | struct driver_context drv_ctx; |
9f0c7d2c | 120 | struct storvsc_driver_object drv_obj; |
f82bd046 HJ |
121 | }; |
122 | ||
454f18a9 | 123 | /* Static decl */ |
f82bd046 HJ |
124 | static int blkvsc_probe(struct device *dev); |
125 | static int blkvsc_remove(struct device *device); | |
126 | static void blkvsc_shutdown(struct device *device); | |
127 | ||
39635f7d | 128 | static int blkvsc_open(struct block_device *bdev, fmode_t mode); |
77d2d9da | 129 | static int blkvsc_release(struct gendisk *disk, fmode_t mode); |
f82bd046 HJ |
130 | static int blkvsc_media_changed(struct gendisk *gd); |
131 | static int blkvsc_revalidate_disk(struct gendisk *gd); | |
132 | static int blkvsc_getgeo(struct block_device *bd, struct hd_geometry *hg); | |
dfe8b2d9 BP |
133 | static int blkvsc_ioctl(struct block_device *bd, fmode_t mode, |
134 | unsigned cmd, unsigned long argument); | |
f82bd046 | 135 | static void blkvsc_request(struct request_queue *queue); |
0b3f6834 | 136 | static void blkvsc_request_completion(struct hv_storvsc_request *request); |
f82bd046 | 137 | static int blkvsc_do_request(struct block_device_context *blkdev, struct request *req); |
0b3f6834 | 138 | static int blkvsc_submit_request(struct blkvsc_request *blkvsc_req, void (*request_completion)(struct hv_storvsc_request*) ); |
f82bd046 | 139 | static void blkvsc_init_rw(struct blkvsc_request *blkvsc_req); |
0b3f6834 | 140 | static void blkvsc_cmd_completion(struct hv_storvsc_request *request); |
f82bd046 HJ |
141 | static int blkvsc_do_inquiry(struct block_device_context *blkdev); |
142 | static int blkvsc_do_read_capacity(struct block_device_context *blkdev); | |
143 | static int blkvsc_do_read_capacity16(struct block_device_context *blkdev); | |
144 | static int blkvsc_do_flush(struct block_device_context *blkdev); | |
145 | static int blkvsc_cancel_pending_reqs(struct block_device_context *blkdev); | |
146 | static int blkvsc_do_pending_reqs(struct block_device_context *blkdev); | |
147 | ||
148 | ||
149 | static int blkvsc_ringbuffer_size = BLKVSC_RING_BUFFER_SIZE; | |
150 | ||
454f18a9 | 151 | /* The one and only one */ |
f82bd046 HJ |
152 | static struct blkvsc_driver_context g_blkvsc_drv; |
153 | ||
154 | ||
155 | static struct block_device_operations block_ops = | |
156 | { | |
157 | .owner = THIS_MODULE, | |
158 | .open = blkvsc_open, | |
159 | .release = blkvsc_release, | |
160 | .media_changed = blkvsc_media_changed, | |
161 | .revalidate_disk = blkvsc_revalidate_disk, | |
162 | .getgeo = blkvsc_getgeo, | |
163 | .ioctl = blkvsc_ioctl, | |
164 | }; | |
165 | ||
166 | /*++ | |
167 | ||
168 | Name: blkvsc_drv_init() | |
169 | ||
170 | Desc: BlkVsc driver initialization. | |
171 | ||
172 | --*/ | |
bd1de709 | 173 | static int blkvsc_drv_init(PFN_DRIVERINITIALIZE pfn_drv_init) |
f82bd046 HJ |
174 | { |
175 | int ret=0; | |
9f0c7d2c | 176 | struct storvsc_driver_object *storvsc_drv_obj = &g_blkvsc_drv.drv_obj; |
f82bd046 HJ |
177 | struct driver_context *drv_ctx=&g_blkvsc_drv.drv_ctx; |
178 | ||
179 | DPRINT_ENTER(BLKVSC_DRV); | |
180 | ||
181 | vmbus_get_interface(&storvsc_drv_obj->Base.VmbusChannelInterface); | |
182 | ||
183 | storvsc_drv_obj->RingBufferSize = blkvsc_ringbuffer_size; | |
184 | ||
454f18a9 | 185 | /* Callback to client driver to complete the initialization */ |
f82bd046 HJ |
186 | pfn_drv_init(&storvsc_drv_obj->Base); |
187 | ||
188 | drv_ctx->driver.name = storvsc_drv_obj->Base.name; | |
caf26a31 | 189 | memcpy(&drv_ctx->class_id, &storvsc_drv_obj->Base.deviceType, sizeof(struct hv_guid)); |
f82bd046 | 190 | |
f82bd046 HJ |
191 | drv_ctx->probe = blkvsc_probe; |
192 | drv_ctx->remove = blkvsc_remove; | |
193 | drv_ctx->shutdown = blkvsc_shutdown; | |
f82bd046 | 194 | |
454f18a9 | 195 | /* The driver belongs to vmbus */ |
5d48a1c2 | 196 | ret = vmbus_child_driver_register(drv_ctx); |
f82bd046 HJ |
197 | |
198 | DPRINT_EXIT(BLKVSC_DRV); | |
199 | ||
200 | return ret; | |
201 | } | |
202 | ||
203 | ||
204 | static int blkvsc_drv_exit_cb(struct device *dev, void *data) | |
205 | { | |
206 | struct device **curr = (struct device **)data; | |
207 | *curr = dev; | |
454f18a9 | 208 | return 1; /* stop iterating */ |
f82bd046 HJ |
209 | } |
210 | ||
211 | /*++ | |
212 | ||
213 | Name: blkvsc_drv_exit() | |
214 | ||
215 | Desc: | |
216 | ||
217 | --*/ | |
bd1de709 | 218 | static void blkvsc_drv_exit(void) |
f82bd046 | 219 | { |
9f0c7d2c | 220 | struct storvsc_driver_object *storvsc_drv_obj = &g_blkvsc_drv.drv_obj; |
f82bd046 | 221 | struct driver_context *drv_ctx=&g_blkvsc_drv.drv_ctx; |
f82bd046 | 222 | struct device *current_dev=NULL; |
2295ba2e | 223 | int ret; |
f82bd046 | 224 | |
f82bd046 HJ |
225 | DPRINT_ENTER(BLKVSC_DRV); |
226 | ||
227 | while (1) | |
228 | { | |
229 | current_dev = NULL; | |
230 | ||
454f18a9 | 231 | /* Get the device */ |
2295ba2e BP |
232 | ret = driver_for_each_device(&drv_ctx->driver, NULL, |
233 | (void *) ¤t_dev, | |
234 | blkvsc_drv_exit_cb); | |
235 | ||
236 | if (ret) | |
237 | DPRINT_WARN(BLKVSC_DRV, | |
238 | "driver_for_each_device returned %d", ret); | |
239 | ||
f82bd046 HJ |
240 | |
241 | if (current_dev == NULL) | |
242 | break; | |
243 | ||
454f18a9 | 244 | /* Initiate removal from the top-down */ |
f82bd046 HJ |
245 | device_unregister(current_dev); |
246 | } | |
247 | ||
248 | if (storvsc_drv_obj->Base.OnCleanup) | |
249 | storvsc_drv_obj->Base.OnCleanup(&storvsc_drv_obj->Base); | |
250 | ||
251 | vmbus_child_driver_unregister(drv_ctx); | |
252 | ||
253 | DPRINT_EXIT(BLKVSC_DRV); | |
254 | ||
255 | return; | |
256 | } | |
257 | ||
258 | /*++ | |
259 | ||
260 | Name: blkvsc_probe() | |
261 | ||
262 | Desc: Add a new device for this driver | |
263 | ||
264 | --*/ | |
265 | static int blkvsc_probe(struct device *device) | |
266 | { | |
267 | int ret=0; | |
268 | ||
269 | struct driver_context *driver_ctx = driver_to_driver_context(device->driver); | |
270 | struct blkvsc_driver_context *blkvsc_drv_ctx = (struct blkvsc_driver_context*)driver_ctx; | |
9f0c7d2c | 271 | struct storvsc_driver_object *storvsc_drv_obj = &blkvsc_drv_ctx->drv_obj; |
f82bd046 HJ |
272 | |
273 | struct device_context *device_ctx = device_to_device_context(device); | |
3d3b5518 | 274 | struct hv_device *device_obj = &device_ctx->device_obj; |
f82bd046 HJ |
275 | |
276 | struct block_device_context *blkdev=NULL; | |
9f0c7d2c | 277 | struct storvsc_device_info device_info; |
f82bd046 HJ |
278 | int major=0; |
279 | int devnum=0; | |
280 | ||
281 | static int ide0_registered=0; | |
282 | static int ide1_registered=0; | |
283 | ||
284 | DPRINT_ENTER(BLKVSC_DRV); | |
285 | ||
286 | DPRINT_DBG(BLKVSC_DRV, "blkvsc_probe - enter"); | |
287 | ||
288 | if (!storvsc_drv_obj->Base.OnDeviceAdd) | |
289 | { | |
290 | DPRINT_ERR(BLKVSC_DRV, "OnDeviceAdd() not set"); | |
291 | ||
292 | ret = -1; | |
293 | goto Cleanup; | |
294 | } | |
295 | ||
296 | blkdev = kzalloc(sizeof(struct block_device_context), GFP_KERNEL); | |
297 | if (!blkdev) | |
298 | { | |
299 | ret = -ENOMEM; | |
300 | goto Cleanup; | |
301 | } | |
302 | ||
303 | INIT_LIST_HEAD(&blkdev->pending_list); | |
304 | ||
454f18a9 | 305 | /* Initialize what we can here */ |
f82bd046 HJ |
306 | spin_lock_init(&blkdev->lock); |
307 | ||
308 | ASSERT(sizeof(struct blkvsc_request_group) <= sizeof(struct blkvsc_request)); | |
309 | ||
454f18a9 BP |
310 | blkdev->request_pool = kmem_cache_create(dev_name(&device_ctx->device), |
311 | sizeof(struct blkvsc_request) + storvsc_drv_obj->RequestExtSize, 0, | |
312 | SLAB_HWCACHE_ALIGN, NULL); | |
f82bd046 HJ |
313 | if (!blkdev->request_pool) |
314 | { | |
315 | ret = -ENOMEM; | |
316 | goto Cleanup; | |
317 | } | |
318 | ||
319 | ||
454f18a9 | 320 | /* Call to the vsc driver to add the device */ |
f82bd046 HJ |
321 | ret = storvsc_drv_obj->Base.OnDeviceAdd(device_obj, &device_info); |
322 | if (ret != 0) | |
323 | { | |
324 | DPRINT_ERR(BLKVSC_DRV, "unable to add blkvsc device"); | |
325 | goto Cleanup; | |
326 | } | |
327 | ||
328 | blkdev->device_ctx = device_ctx; | |
454f18a9 BP |
329 | blkdev->target = device_info.TargetId; /* this identified the device 0 or 1 */ |
330 | blkdev->path = device_info.PathId; /* this identified the ide ctrl 0 or 1 */ | |
f82bd046 | 331 | |
b57a68dc | 332 | dev_set_drvdata(device, blkdev); |
f82bd046 | 333 | |
454f18a9 | 334 | /* Calculate the major and device num */ |
f82bd046 HJ |
335 | if (blkdev->path == 0) |
336 | { | |
337 | major = IDE0_MAJOR; | |
454f18a9 | 338 | devnum = blkdev->path + blkdev->target; /* 0 or 1 */ |
f82bd046 HJ |
339 | |
340 | if (!ide0_registered) | |
341 | { | |
342 | ret = register_blkdev(major, "ide"); | |
343 | if (ret != 0) | |
344 | { | |
345 | DPRINT_ERR(BLKVSC_DRV, "register_blkdev() failed! ret %d", ret); | |
346 | goto Remove; | |
347 | } | |
348 | ||
349 | ide0_registered = 1; | |
350 | } | |
351 | } | |
352 | else if (blkdev->path == 1) | |
353 | { | |
354 | major = IDE1_MAJOR; | |
454f18a9 | 355 | devnum = blkdev->path + blkdev->target + 1; /* 2 or 3 */ |
f82bd046 HJ |
356 | |
357 | if (!ide1_registered) | |
358 | { | |
359 | ret = register_blkdev(major, "ide"); | |
360 | if (ret != 0) | |
361 | { | |
362 | DPRINT_ERR(BLKVSC_DRV, "register_blkdev() failed! ret %d", ret); | |
363 | goto Remove; | |
364 | } | |
365 | ||
366 | ide1_registered = 1; | |
367 | } | |
368 | ||
369 | } | |
370 | else | |
371 | { | |
372 | DPRINT_ERR(BLKVSC_DRV, "invalid pathid"); | |
373 | ret = -1; | |
374 | goto Cleanup; | |
375 | } | |
376 | ||
377 | DPRINT_INFO(BLKVSC_DRV, "blkvsc registered for major %d!!", major); | |
378 | ||
379 | blkdev->gd = alloc_disk(BLKVSC_MINORS); | |
380 | if (!blkdev->gd) | |
381 | { | |
382 | DPRINT_ERR(BLKVSC_DRV, "register_blkdev() failed! ret %d", ret); | |
383 | ret = -1; | |
384 | goto Cleanup; | |
385 | } | |
386 | ||
387 | blkdev->gd->queue = blk_init_queue(blkvsc_request, &blkdev->lock); | |
388 | ||
389 | blk_queue_max_segment_size(blkdev->gd->queue, PAGE_SIZE); | |
390 | blk_queue_max_phys_segments(blkdev->gd->queue, MAX_MULTIPAGE_BUFFER_COUNT); | |
391 | blk_queue_max_hw_segments(blkdev->gd->queue, MAX_MULTIPAGE_BUFFER_COUNT); | |
392 | blk_queue_segment_boundary(blkdev->gd->queue, PAGE_SIZE-1); | |
393 | blk_queue_bounce_limit(blkdev->gd->queue, BLK_BOUNCE_ANY); | |
394 | blk_queue_dma_alignment(blkdev->gd->queue, 511); | |
395 | ||
396 | blkdev->gd->major = major; | |
397 | if (devnum == 1 || devnum == 3) | |
398 | blkdev->gd->first_minor = BLKVSC_MINORS; | |
399 | else | |
400 | blkdev->gd->first_minor = 0; | |
401 | blkdev->gd->fops = &block_ops; | |
402 | blkdev->gd->private_data = blkdev; | |
403 | sprintf(blkdev->gd->disk_name, "hd%c", 'a'+ devnum); | |
404 | ||
405 | blkvsc_do_inquiry(blkdev); | |
406 | if (blkdev->device_type == DVD_TYPE) | |
407 | { | |
408 | set_disk_ro(blkdev->gd, 1); | |
409 | blkdev->gd->flags |= GENHD_FL_REMOVABLE; | |
410 | blkvsc_do_read_capacity(blkdev); | |
411 | } | |
412 | else | |
413 | { | |
414 | blkvsc_do_read_capacity16(blkdev); | |
415 | } | |
416 | ||
417 | set_capacity(blkdev->gd, blkdev->capacity * (blkdev->sector_size/512)); | |
0fce4c2f | 418 | blk_queue_logical_block_size(blkdev->gd->queue, blkdev->sector_size); |
454f18a9 | 419 | /* go! */ |
f82bd046 HJ |
420 | add_disk(blkdev->gd); |
421 | ||
627c156d | 422 | DPRINT_INFO(BLKVSC_DRV, "%s added!! capacity %lu sector_size %d", blkdev->gd->disk_name, (unsigned long) blkdev->capacity, blkdev->sector_size); |
f82bd046 HJ |
423 | |
424 | return ret; | |
425 | ||
426 | Remove: | |
427 | storvsc_drv_obj->Base.OnDeviceRemove(device_obj); | |
428 | ||
429 | Cleanup: | |
430 | if (blkdev) | |
431 | { | |
432 | if (blkdev->request_pool) | |
433 | { | |
434 | kmem_cache_destroy(blkdev->request_pool); | |
435 | blkdev->request_pool = NULL; | |
436 | } | |
437 | kfree(blkdev); | |
438 | blkdev = NULL; | |
439 | } | |
440 | ||
441 | DPRINT_EXIT(BLKVSC_DRV); | |
442 | ||
443 | return ret; | |
444 | } | |
445 | ||
446 | static void blkvsc_shutdown(struct device *device) | |
447 | { | |
b57a68dc | 448 | struct block_device_context *blkdev = dev_get_drvdata(device); |
f82bd046 HJ |
449 | unsigned long flags; |
450 | ||
451 | if (!blkdev) | |
452 | return; | |
453 | ||
454 | DPRINT_DBG(BLKVSC_DRV, "blkvsc_shutdown - users %d disk %s\n", blkdev->users, blkdev->gd->disk_name); | |
455 | ||
456 | spin_lock_irqsave(&blkdev->lock, flags); | |
457 | ||
458 | blkdev->shutting_down = 1; | |
459 | ||
460 | blk_stop_queue(blkdev->gd->queue); | |
461 | ||
462 | spin_unlock_irqrestore(&blkdev->lock, flags); | |
463 | ||
464 | while (blkdev->num_outstanding_reqs) | |
465 | { | |
466 | DPRINT_INFO(STORVSC, "waiting for %d requests to complete...", blkdev->num_outstanding_reqs); | |
467 | ||
468 | udelay(100); | |
469 | } | |
470 | ||
471 | blkvsc_do_flush(blkdev); | |
472 | ||
473 | spin_lock_irqsave(&blkdev->lock, flags); | |
474 | ||
475 | blkvsc_cancel_pending_reqs(blkdev); | |
476 | ||
477 | spin_unlock_irqrestore(&blkdev->lock, flags); | |
478 | } | |
479 | ||
480 | static int blkvsc_do_flush(struct block_device_context *blkdev) | |
481 | { | |
482 | struct blkvsc_request *blkvsc_req=NULL; | |
483 | ||
484 | DPRINT_DBG(BLKVSC_DRV, "blkvsc_do_flush()\n"); | |
485 | ||
486 | if (blkdev->device_type != HARDDISK_TYPE) | |
487 | return 0; | |
488 | ||
489 | blkvsc_req = kmem_cache_alloc(blkdev->request_pool, GFP_KERNEL); | |
490 | if (!blkvsc_req) | |
491 | { | |
492 | return -ENOMEM; | |
493 | } | |
494 | ||
495 | memset(blkvsc_req, 0, sizeof(struct blkvsc_request)); | |
496 | init_waitqueue_head(&blkvsc_req->wevent); | |
497 | blkvsc_req->dev = blkdev; | |
498 | blkvsc_req->req = NULL; | |
499 | blkvsc_req->write = 0; | |
500 | ||
501 | blkvsc_req->request.DataBuffer.PfnArray[0] = 0; | |
502 | blkvsc_req->request.DataBuffer.Offset = 0; | |
503 | blkvsc_req->request.DataBuffer.Length = 0; | |
504 | ||
505 | blkvsc_req->cmnd[0] = SYNCHRONIZE_CACHE; | |
506 | blkvsc_req->cmd_len = 10; | |
507 | ||
454f18a9 | 508 | /* Set this here since the completion routine may be invoked and completed before we return */ |
f82bd046 HJ |
509 | blkvsc_req->cond =0; |
510 | blkvsc_submit_request(blkvsc_req, blkvsc_cmd_completion); | |
511 | ||
512 | wait_event_interruptible(blkvsc_req->wevent, blkvsc_req->cond); | |
513 | ||
514 | kmem_cache_free(blkvsc_req->dev->request_pool, blkvsc_req); | |
515 | ||
516 | return 0; | |
517 | } | |
518 | ||
454f18a9 | 519 | /* Do a scsi INQUIRY cmd here to get the device type (ie disk or dvd) */ |
f82bd046 HJ |
520 | static int blkvsc_do_inquiry(struct block_device_context *blkdev) |
521 | { | |
522 | struct blkvsc_request *blkvsc_req=NULL; | |
523 | struct page *page_buf; | |
524 | unsigned char *buf; | |
525 | unsigned char device_type; | |
526 | ||
527 | DPRINT_DBG(BLKVSC_DRV, "blkvsc_do_inquiry()\n"); | |
528 | ||
529 | blkvsc_req = kmem_cache_alloc(blkdev->request_pool, GFP_KERNEL); | |
530 | if (!blkvsc_req) | |
531 | { | |
532 | return -ENOMEM; | |
533 | } | |
534 | ||
535 | memset(blkvsc_req, 0, sizeof(struct blkvsc_request)); | |
536 | page_buf = alloc_page(GFP_KERNEL); | |
537 | if (!page_buf) | |
538 | { | |
539 | kmem_cache_free(blkvsc_req->dev->request_pool, blkvsc_req); | |
540 | return -ENOMEM; | |
541 | } | |
542 | ||
543 | init_waitqueue_head(&blkvsc_req->wevent); | |
544 | blkvsc_req->dev = blkdev; | |
545 | blkvsc_req->req = NULL; | |
546 | blkvsc_req->write = 0; | |
547 | ||
548 | blkvsc_req->request.DataBuffer.PfnArray[0] = page_to_pfn(page_buf); | |
549 | blkvsc_req->request.DataBuffer.Offset = 0; | |
550 | blkvsc_req->request.DataBuffer.Length = 64; | |
551 | ||
552 | blkvsc_req->cmnd[0] = INQUIRY; | |
454f18a9 BP |
553 | blkvsc_req->cmnd[1] = 0x1; /* Get product data */ |
554 | blkvsc_req->cmnd[2] = 0x83; /* mode page 83 */ | |
f82bd046 HJ |
555 | blkvsc_req->cmnd[4] = 64; |
556 | blkvsc_req->cmd_len = 6; | |
557 | ||
454f18a9 | 558 | /* Set this here since the completion routine may be invoked and completed before we return */ |
f82bd046 HJ |
559 | blkvsc_req->cond =0; |
560 | ||
561 | blkvsc_submit_request(blkvsc_req, blkvsc_cmd_completion); | |
562 | ||
563 | DPRINT_DBG(BLKVSC_DRV, "waiting %p to complete - cond %d\n", blkvsc_req, blkvsc_req->cond); | |
564 | ||
565 | wait_event_interruptible(blkvsc_req->wevent, blkvsc_req->cond); | |
566 | ||
567 | buf = kmap(page_buf); | |
568 | ||
04f50c4d | 569 | /* print_hex_dump_bytes("", DUMP_PREFIX_NONE, buf, 64); */ |
454f18a9 | 570 | /* be to le */ |
f82bd046 HJ |
571 | device_type = buf[0] & 0x1F; |
572 | ||
573 | if (device_type == 0x0) | |
574 | { | |
575 | blkdev->device_type = HARDDISK_TYPE; | |
576 | } | |
577 | else if (device_type == 0x5) | |
578 | { | |
579 | blkdev->device_type = DVD_TYPE; | |
580 | } | |
581 | else | |
582 | { | |
454f18a9 | 583 | /* TODO: this is currently unsupported device type */ |
f82bd046 HJ |
584 | blkdev->device_type = UNKNOWN_DEV_TYPE; |
585 | } | |
586 | ||
587 | DPRINT_DBG(BLKVSC_DRV, "device type %d \n", device_type); | |
588 | ||
589 | blkdev->device_id_len = buf[7]; | |
590 | if (blkdev->device_id_len > 64) | |
591 | blkdev->device_id_len = 64; | |
592 | ||
593 | memcpy(blkdev->device_id, &buf[8], blkdev->device_id_len); | |
04f50c4d | 594 | /* printk_hex_dump_bytes("", DUMP_PREFIX_NONE, blkdev->device_id, |
454f18a9 | 595 | * blkdev->device_id_len); */ |
f82bd046 HJ |
596 | |
597 | kunmap(page_buf); | |
598 | ||
599 | __free_page(page_buf); | |
600 | ||
601 | kmem_cache_free(blkvsc_req->dev->request_pool, blkvsc_req); | |
602 | ||
603 | return 0; | |
604 | } | |
605 | ||
454f18a9 | 606 | /* Do a scsi READ_CAPACITY cmd here to get the size of the disk */ |
f82bd046 HJ |
607 | static int blkvsc_do_read_capacity(struct block_device_context *blkdev) |
608 | { | |
609 | struct blkvsc_request *blkvsc_req=NULL; | |
610 | struct page *page_buf; | |
611 | unsigned char *buf; | |
612 | struct scsi_sense_hdr sense_hdr; | |
613 | ||
614 | DPRINT_DBG(BLKVSC_DRV, "blkvsc_do_read_capacity()\n"); | |
615 | ||
616 | blkdev->sector_size = 0; | |
617 | blkdev->capacity = 0; | |
454f18a9 | 618 | blkdev->media_not_present = 0; /* assume a disk is present */ |
f82bd046 HJ |
619 | |
620 | blkvsc_req = kmem_cache_alloc(blkdev->request_pool, GFP_KERNEL); | |
621 | if (!blkvsc_req) | |
622 | { | |
623 | return -ENOMEM; | |
624 | } | |
625 | ||
626 | memset(blkvsc_req, 0, sizeof(struct blkvsc_request)); | |
627 | page_buf = alloc_page(GFP_KERNEL); | |
628 | if (!page_buf) | |
629 | { | |
630 | kmem_cache_free(blkvsc_req->dev->request_pool, blkvsc_req); | |
631 | return -ENOMEM; | |
632 | } | |
633 | ||
634 | init_waitqueue_head(&blkvsc_req->wevent); | |
635 | blkvsc_req->dev = blkdev; | |
636 | blkvsc_req->req = NULL; | |
637 | blkvsc_req->write = 0; | |
638 | ||
639 | blkvsc_req->request.DataBuffer.PfnArray[0] = page_to_pfn(page_buf); | |
640 | blkvsc_req->request.DataBuffer.Offset = 0; | |
641 | blkvsc_req->request.DataBuffer.Length = 8; | |
642 | ||
643 | blkvsc_req->cmnd[0] = READ_CAPACITY; | |
644 | blkvsc_req->cmd_len = 16; | |
645 | ||
454f18a9 BP |
646 | /* |
647 | * Set this here since the completion routine may be invoked | |
648 | * and completed before we return | |
649 | */ | |
f82bd046 HJ |
650 | blkvsc_req->cond =0; |
651 | ||
652 | blkvsc_submit_request(blkvsc_req, blkvsc_cmd_completion); | |
653 | ||
654 | DPRINT_DBG(BLKVSC_DRV, "waiting %p to complete - cond %d\n", blkvsc_req, blkvsc_req->cond); | |
655 | ||
656 | wait_event_interruptible(blkvsc_req->wevent, blkvsc_req->cond); | |
657 | ||
454f18a9 | 658 | /* check error */ |
f82bd046 HJ |
659 | if (blkvsc_req->request.Status) |
660 | { | |
661 | scsi_normalize_sense(blkvsc_req->sense_buffer, SCSI_SENSE_BUFFERSIZE, &sense_hdr); | |
662 | ||
454f18a9 | 663 | if (sense_hdr.asc == 0x3A) /* Medium not present */ |
f82bd046 HJ |
664 | { |
665 | blkdev->media_not_present = 1; | |
666 | } | |
667 | ||
668 | return 0; | |
669 | } | |
670 | buf = kmap(page_buf); | |
671 | ||
454f18a9 | 672 | /* be to le */ |
f82bd046 HJ |
673 | blkdev->capacity = ((buf[0] << 24) | (buf[1] << 16) | (buf[2] << 8) | buf[3]) + 1; |
674 | blkdev->sector_size = (buf[4] << 24) | (buf[5] << 16) | (buf[6] << 8) | buf[7]; | |
675 | ||
676 | kunmap(page_buf); | |
677 | ||
678 | __free_page(page_buf); | |
679 | ||
680 | kmem_cache_free(blkvsc_req->dev->request_pool, blkvsc_req); | |
681 | ||
682 | return 0; | |
683 | } | |
684 | ||
685 | ||
686 | static int blkvsc_do_read_capacity16(struct block_device_context *blkdev) | |
687 | { | |
688 | struct blkvsc_request *blkvsc_req=NULL; | |
689 | struct page *page_buf; | |
690 | unsigned char *buf; | |
691 | struct scsi_sense_hdr sense_hdr; | |
692 | ||
693 | DPRINT_DBG(BLKVSC_DRV, "blkvsc_do_read_capacity16()\n"); | |
694 | ||
695 | blkdev->sector_size = 0; | |
696 | blkdev->capacity = 0; | |
454f18a9 | 697 | blkdev->media_not_present = 0; /* assume a disk is present */ |
f82bd046 HJ |
698 | |
699 | blkvsc_req = kmem_cache_alloc(blkdev->request_pool, GFP_KERNEL); | |
700 | if (!blkvsc_req) | |
701 | { | |
702 | return -ENOMEM; | |
703 | } | |
704 | ||
705 | memset(blkvsc_req, 0, sizeof(struct blkvsc_request)); | |
706 | page_buf = alloc_page(GFP_KERNEL); | |
707 | if (!page_buf) | |
708 | { | |
709 | kmem_cache_free(blkvsc_req->dev->request_pool, blkvsc_req); | |
710 | return -ENOMEM; | |
711 | } | |
712 | ||
713 | init_waitqueue_head(&blkvsc_req->wevent); | |
714 | blkvsc_req->dev = blkdev; | |
715 | blkvsc_req->req = NULL; | |
716 | blkvsc_req->write = 0; | |
717 | ||
718 | blkvsc_req->request.DataBuffer.PfnArray[0] = page_to_pfn(page_buf); | |
719 | blkvsc_req->request.DataBuffer.Offset = 0; | |
720 | blkvsc_req->request.DataBuffer.Length = 12; | |
721 | ||
454f18a9 | 722 | blkvsc_req->cmnd[0] = 0x9E; /* READ_CAPACITY16; */ |
f82bd046 HJ |
723 | blkvsc_req->cmd_len = 16; |
724 | ||
454f18a9 BP |
725 | /* |
726 | * Set this here since the completion routine may be invoked | |
727 | * and completed before we return | |
728 | */ | |
f82bd046 HJ |
729 | blkvsc_req->cond =0; |
730 | ||
731 | blkvsc_submit_request(blkvsc_req, blkvsc_cmd_completion); | |
732 | ||
733 | DPRINT_DBG(BLKVSC_DRV, "waiting %p to complete - cond %d\n", blkvsc_req, blkvsc_req->cond); | |
734 | ||
735 | wait_event_interruptible(blkvsc_req->wevent, blkvsc_req->cond); | |
736 | ||
454f18a9 | 737 | /* check error */ |
f82bd046 HJ |
738 | if (blkvsc_req->request.Status) |
739 | { | |
740 | scsi_normalize_sense(blkvsc_req->sense_buffer, SCSI_SENSE_BUFFERSIZE, &sense_hdr); | |
741 | ||
454f18a9 | 742 | if (sense_hdr.asc == 0x3A) /* Medium not present */ |
f82bd046 HJ |
743 | { |
744 | blkdev->media_not_present = 1; | |
745 | } | |
746 | ||
747 | return 0; | |
748 | } | |
749 | buf = kmap(page_buf); | |
750 | ||
454f18a9 | 751 | /* be to le */ |
f82bd046 HJ |
752 | blkdev->capacity = be64_to_cpu(*(unsigned long long*) &buf[0]) + 1; |
753 | blkdev->sector_size = be32_to_cpu(*(unsigned int*)&buf[8]); | |
754 | ||
454f18a9 BP |
755 | /* blkdev->capacity = ((buf[0] << 24) | (buf[1] << 16) | (buf[2] << 8) | buf[3]) + 1; */ |
756 | /* blkdev->sector_size = (buf[4] << 24) | (buf[5] << 16) | (buf[6] << 8) | buf[7]; */ | |
f82bd046 HJ |
757 | |
758 | kunmap(page_buf); | |
759 | ||
760 | __free_page(page_buf); | |
761 | ||
762 | kmem_cache_free(blkvsc_req->dev->request_pool, blkvsc_req); | |
763 | ||
764 | return 0; | |
765 | } | |
766 | ||
767 | /*++ | |
768 | ||
769 | Name: blkvsc_remove() | |
770 | ||
771 | Desc: Callback when our device is removed | |
772 | ||
773 | --*/ | |
774 | static int blkvsc_remove(struct device *device) | |
775 | { | |
776 | int ret=0; | |
777 | ||
778 | struct driver_context *driver_ctx = driver_to_driver_context(device->driver); | |
779 | struct blkvsc_driver_context *blkvsc_drv_ctx = (struct blkvsc_driver_context*)driver_ctx; | |
9f0c7d2c | 780 | struct storvsc_driver_object *storvsc_drv_obj = &blkvsc_drv_ctx->drv_obj; |
f82bd046 HJ |
781 | |
782 | struct device_context *device_ctx = device_to_device_context(device); | |
3d3b5518 | 783 | struct hv_device *device_obj = &device_ctx->device_obj; |
b57a68dc | 784 | struct block_device_context *blkdev = dev_get_drvdata(device); |
f82bd046 HJ |
785 | unsigned long flags; |
786 | ||
787 | DPRINT_ENTER(BLKVSC_DRV); | |
788 | ||
789 | DPRINT_DBG(BLKVSC_DRV, "blkvsc_remove()\n"); | |
790 | ||
791 | if (!storvsc_drv_obj->Base.OnDeviceRemove) | |
792 | { | |
793 | DPRINT_EXIT(BLKVSC_DRV); | |
794 | return -1; | |
795 | } | |
796 | ||
454f18a9 | 797 | /* Call to the vsc driver to let it know that the device is being removed */ |
f82bd046 HJ |
798 | ret = storvsc_drv_obj->Base.OnDeviceRemove(device_obj); |
799 | if (ret != 0) | |
800 | { | |
454f18a9 | 801 | /* TODO: */ |
f82bd046 HJ |
802 | DPRINT_ERR(BLKVSC_DRV, "unable to remove blkvsc device (ret %d)", ret); |
803 | } | |
804 | ||
454f18a9 | 805 | /* Get to a known state */ |
f82bd046 HJ |
806 | spin_lock_irqsave(&blkdev->lock, flags); |
807 | ||
808 | blkdev->shutting_down = 1; | |
809 | ||
810 | blk_stop_queue(blkdev->gd->queue); | |
811 | ||
812 | spin_unlock_irqrestore(&blkdev->lock, flags); | |
813 | ||
814 | while (blkdev->num_outstanding_reqs) | |
815 | { | |
816 | DPRINT_INFO(STORVSC, "waiting for %d requests to complete...", blkdev->num_outstanding_reqs); | |
817 | ||
818 | udelay(100); | |
819 | } | |
820 | ||
821 | blkvsc_do_flush(blkdev); | |
822 | ||
823 | spin_lock_irqsave(&blkdev->lock, flags); | |
824 | ||
825 | blkvsc_cancel_pending_reqs(blkdev); | |
826 | ||
827 | spin_unlock_irqrestore(&blkdev->lock, flags); | |
828 | ||
829 | blk_cleanup_queue(blkdev->gd->queue); | |
830 | ||
831 | del_gendisk(blkdev->gd); | |
832 | ||
833 | kmem_cache_destroy(blkdev->request_pool); | |
834 | ||
835 | kfree(blkdev); | |
836 | ||
837 | DPRINT_EXIT(BLKVSC_DRV); | |
838 | ||
839 | return ret; | |
840 | } | |
841 | ||
842 | static void blkvsc_init_rw(struct blkvsc_request *blkvsc_req) | |
843 | { | |
844 | ASSERT(blkvsc_req->req); | |
845 | ASSERT(blkvsc_req->sector_count <= (MAX_MULTIPAGE_BUFFER_COUNT*8)); | |
846 | ||
847 | blkvsc_req->cmd_len = 16; | |
848 | ||
849 | if (blkvsc_req->sector_start > 0xffffffff) | |
850 | { | |
851 | if (rq_data_dir(blkvsc_req->req)) | |
852 | { | |
853 | blkvsc_req->write = 1; | |
854 | blkvsc_req->cmnd[0] = WRITE_16; | |
855 | } | |
856 | else | |
857 | { | |
858 | blkvsc_req->write = 0; | |
859 | blkvsc_req->cmnd[0] = READ_16; | |
860 | } | |
861 | ||
862 | blkvsc_req->cmnd[1] |= blk_fua_rq(blkvsc_req->req) ? 0x8 : 0; | |
863 | ||
864 | *(unsigned long long*)&blkvsc_req->cmnd[2] = cpu_to_be64(blkvsc_req->sector_start); | |
865 | *(unsigned int*)&blkvsc_req->cmnd[10] = cpu_to_be32(blkvsc_req->sector_count); | |
866 | } | |
867 | else if ((blkvsc_req->sector_count > 0xff) || (blkvsc_req->sector_start > 0x1fffff)) | |
868 | { | |
869 | if (rq_data_dir(blkvsc_req->req)) | |
870 | { | |
871 | blkvsc_req->write = 1; | |
872 | blkvsc_req->cmnd[0] = WRITE_10; | |
873 | } | |
874 | else | |
875 | { | |
876 | blkvsc_req->write = 0; | |
877 | blkvsc_req->cmnd[0] = READ_10; | |
878 | } | |
879 | ||
880 | blkvsc_req->cmnd[1] |= blk_fua_rq(blkvsc_req->req) ? 0x8 : 0; | |
881 | ||
882 | *(unsigned int *)&blkvsc_req->cmnd[2] = cpu_to_be32(blkvsc_req->sector_start); | |
883 | *(unsigned short*)&blkvsc_req->cmnd[7] = cpu_to_be16(blkvsc_req->sector_count); | |
884 | } | |
885 | else | |
886 | { | |
887 | if (rq_data_dir(blkvsc_req->req)) | |
888 | { | |
889 | blkvsc_req->write = 1; | |
890 | blkvsc_req->cmnd[0] = WRITE_6; | |
891 | } | |
892 | else | |
893 | { | |
894 | blkvsc_req->write = 0; | |
895 | blkvsc_req->cmnd[0] = READ_6; | |
896 | } | |
897 | ||
898 | *(unsigned int *)&blkvsc_req->cmnd[1] = cpu_to_be32(blkvsc_req->sector_start) >> 8; | |
899 | blkvsc_req->cmnd[1] &= 0x1f; | |
900 | blkvsc_req->cmnd[4] = (unsigned char) blkvsc_req->sector_count; | |
901 | } | |
902 | } | |
903 | ||
0b3f6834 | 904 | static int blkvsc_submit_request(struct blkvsc_request *blkvsc_req, void (*request_completion)(struct hv_storvsc_request*) ) |
f82bd046 HJ |
905 | { |
906 | struct block_device_context *blkdev = blkvsc_req->dev; | |
907 | struct device_context *device_ctx=blkdev->device_ctx; | |
908 | struct driver_context *driver_ctx = driver_to_driver_context(device_ctx->device.driver); | |
909 | struct blkvsc_driver_context *blkvsc_drv_ctx = (struct blkvsc_driver_context*)driver_ctx; | |
9f0c7d2c | 910 | struct storvsc_driver_object *storvsc_drv_obj = &blkvsc_drv_ctx->drv_obj; |
f82bd046 HJ |
911 | int ret =0; |
912 | ||
0b3f6834 | 913 | struct hv_storvsc_request *storvsc_req; |
f82bd046 | 914 | |
627c156d | 915 | DPRINT_DBG(BLKVSC_DRV, "blkvsc_submit_request() - req %p type %s start_sector %lu count %ld offset %d len %d\n", |
f82bd046 HJ |
916 | blkvsc_req, |
917 | (blkvsc_req->write)?"WRITE":"READ", | |
627c156d | 918 | (unsigned long) blkvsc_req->sector_start, |
f82bd046 HJ |
919 | blkvsc_req->sector_count, |
920 | blkvsc_req->request.DataBuffer.Offset, | |
921 | blkvsc_req->request.DataBuffer.Length); | |
922 | ||
923 | /*for (i=0; i < (blkvsc_req->request.DataBuffer.Length >> 12); i++) | |
924 | { | |
925 | DPRINT_DBG(BLKVSC_DRV, "blkvsc_submit_request() - req %p pfn[%d] %llx\n", | |
926 | blkvsc_req, | |
927 | i, | |
928 | blkvsc_req->request.DataBuffer.PfnArray[i]); | |
929 | }*/ | |
930 | ||
931 | storvsc_req = &blkvsc_req->request; | |
932 | storvsc_req->Extension = (void*)((unsigned long)blkvsc_req + sizeof(struct blkvsc_request)); | |
933 | ||
934 | storvsc_req->Type = blkvsc_req->write? WRITE_TYPE : READ_TYPE; | |
935 | ||
936 | storvsc_req->OnIOCompletion = request_completion; | |
937 | storvsc_req->Context = blkvsc_req; | |
938 | ||
939 | storvsc_req->Host = blkdev->port; | |
940 | storvsc_req->Bus = blkdev->path; | |
941 | storvsc_req->TargetId = blkdev->target; | |
454f18a9 | 942 | storvsc_req->LunId = 0; /* this is not really used at all */ |
f82bd046 HJ |
943 | |
944 | storvsc_req->CdbLen = blkvsc_req->cmd_len; | |
945 | storvsc_req->Cdb = blkvsc_req->cmnd; | |
946 | ||
947 | storvsc_req->SenseBuffer = blkvsc_req->sense_buffer; | |
948 | storvsc_req->SenseBufferSize = SCSI_SENSE_BUFFERSIZE; | |
949 | ||
950 | ret = storvsc_drv_obj->OnIORequest(&blkdev->device_ctx->device_obj, &blkvsc_req->request); | |
951 | if (ret == 0) | |
952 | { | |
953 | blkdev->num_outstanding_reqs++; | |
954 | } | |
955 | ||
956 | return ret; | |
957 | } | |
958 | ||
454f18a9 BP |
959 | |
960 | /* | |
961 | * We break the request into 1 or more blkvsc_requests and submit | |
962 | * them. If we cant submit them all, we put them on the | |
963 | * pending_list. The blkvsc_request() will work on the pending_list. | |
964 | */ | |
965 | ||
f82bd046 HJ |
966 | static int blkvsc_do_request(struct block_device_context *blkdev, struct request *req) |
967 | { | |
968 | struct bio *bio=NULL; | |
969 | struct bio_vec *bvec=NULL; | |
970 | struct bio_vec *prev_bvec=NULL; | |
971 | ||
972 | struct blkvsc_request *blkvsc_req=NULL; | |
973 | struct blkvsc_request *tmp; | |
974 | int databuf_idx=0; | |
975 | int seg_idx=0; | |
976 | ||
977 | sector_t start_sector; | |
978 | unsigned long num_sectors = 0; | |
979 | int ret=0; | |
980 | int pending=0; | |
981 | struct blkvsc_request_group *group=NULL; | |
982 | ||
627c156d | 983 | DPRINT_DBG(BLKVSC_DRV, "blkdev %p req %p sect %lu \n", blkdev, req, (unsigned long) blk_rq_pos(req)); |
f82bd046 | 984 | |
454f18a9 | 985 | /* Create a group to tie req to list of blkvsc_reqs */ |
f82bd046 HJ |
986 | group = (struct blkvsc_request_group*)kmem_cache_alloc(blkdev->request_pool, GFP_ATOMIC); |
987 | if (!group) | |
988 | { | |
989 | return -ENOMEM; | |
990 | } | |
991 | ||
992 | INIT_LIST_HEAD(&group->blkvsc_req_list); | |
993 | group->outstanding = group->status = 0; | |
994 | ||
0fce4c2f | 995 | start_sector = blk_rq_pos(req); |
f82bd046 | 996 | |
454f18a9 | 997 | /* foreach bio in the request */ |
f82bd046 HJ |
998 | if (req->bio) |
999 | for (bio = req->bio; bio; bio = bio->bi_next) | |
1000 | { | |
454f18a9 | 1001 | /* Map this bio into an existing or new storvsc request */ |
f82bd046 HJ |
1002 | bio_for_each_segment (bvec, bio, seg_idx) |
1003 | { | |
1004 | DPRINT_DBG(BLKVSC_DRV, "bio_for_each_segment() - req %p bio %p bvec %p seg_idx %d databuf_idx %d\n", | |
1005 | req, bio, bvec, seg_idx, databuf_idx); | |
1006 | ||
454f18a9 BP |
1007 | /* Get a new storvsc request */ |
1008 | if ( (!blkvsc_req) || /* 1st-time */ | |
f82bd046 | 1009 | (databuf_idx >= MAX_MULTIPAGE_BUFFER_COUNT) || |
454f18a9 BP |
1010 | (bvec->bv_offset != 0) || /* hole at the begin of page */ |
1011 | (prev_bvec && (prev_bvec->bv_len != PAGE_SIZE)) ) /* hold at the end of page */ | |
f82bd046 | 1012 | { |
454f18a9 | 1013 | /* submit the prev one */ |
f82bd046 HJ |
1014 | if (blkvsc_req) |
1015 | { | |
1016 | blkvsc_req->sector_start = start_sector; | |
1017 | sector_div(blkvsc_req->sector_start, (blkdev->sector_size >> 9)); | |
1018 | ||
1019 | blkvsc_req->sector_count = num_sectors / (blkdev->sector_size >> 9); | |
1020 | ||
1021 | blkvsc_init_rw(blkvsc_req); | |
1022 | } | |
1023 | ||
454f18a9 | 1024 | /* Create new blkvsc_req to represent the current bvec */ |
f82bd046 HJ |
1025 | blkvsc_req = kmem_cache_alloc(blkdev->request_pool, GFP_ATOMIC); |
1026 | if (!blkvsc_req) | |
1027 | { | |
454f18a9 | 1028 | /* free up everything */ |
f82bd046 HJ |
1029 | list_for_each_entry_safe(blkvsc_req, tmp, &group->blkvsc_req_list, req_entry) |
1030 | { | |
1031 | list_del(&blkvsc_req->req_entry); | |
1032 | kmem_cache_free(blkdev->request_pool, blkvsc_req); | |
1033 | } | |
1034 | ||
1035 | kmem_cache_free(blkdev->request_pool, group); | |
1036 | return -ENOMEM; | |
1037 | } | |
1038 | ||
1039 | memset(blkvsc_req, 0, sizeof(struct blkvsc_request)); | |
1040 | ||
1041 | blkvsc_req->dev = blkdev; | |
1042 | blkvsc_req->req = req; | |
1043 | blkvsc_req->request.DataBuffer.Offset = bvec->bv_offset; | |
1044 | blkvsc_req->request.DataBuffer.Length = 0; | |
1045 | ||
454f18a9 | 1046 | /* Add to the group */ |
f82bd046 HJ |
1047 | blkvsc_req->group = group; |
1048 | blkvsc_req->group->outstanding++; | |
1049 | list_add_tail(&blkvsc_req->req_entry, &blkvsc_req->group->blkvsc_req_list); | |
1050 | ||
1051 | start_sector += num_sectors; | |
1052 | num_sectors = 0; | |
1053 | databuf_idx = 0; | |
1054 | } | |
1055 | ||
454f18a9 | 1056 | /* Add the curr bvec/segment to the curr blkvsc_req */ |
f82bd046 HJ |
1057 | blkvsc_req->request.DataBuffer.PfnArray[databuf_idx] = page_to_pfn(bvec->bv_page); |
1058 | blkvsc_req->request.DataBuffer.Length += bvec->bv_len; | |
1059 | ||
1060 | prev_bvec = bvec; | |
1061 | ||
1062 | databuf_idx++; | |
1063 | num_sectors += bvec->bv_len >> 9; | |
1064 | ||
454f18a9 | 1065 | } /* bio_for_each_segment */ |
f82bd046 | 1066 | |
454f18a9 | 1067 | } /* rq_for_each_bio */ |
f82bd046 | 1068 | |
454f18a9 | 1069 | /* Handle the last one */ |
f82bd046 HJ |
1070 | if (blkvsc_req) |
1071 | { | |
1072 | DPRINT_DBG(BLKVSC_DRV, "blkdev %p req %p group %p count %d\n", blkdev, req, blkvsc_req->group, blkvsc_req->group->outstanding); | |
1073 | ||
1074 | blkvsc_req->sector_start = start_sector; | |
1075 | sector_div(blkvsc_req->sector_start, (blkdev->sector_size >> 9)); | |
1076 | ||
1077 | blkvsc_req->sector_count = num_sectors / (blkdev->sector_size >> 9); | |
1078 | ||
1079 | blkvsc_init_rw(blkvsc_req); | |
1080 | } | |
1081 | ||
1082 | list_for_each_entry(blkvsc_req, &group->blkvsc_req_list, req_entry) | |
1083 | { | |
1084 | if (pending) | |
1085 | { | |
627c156d | 1086 | DPRINT_DBG(BLKVSC_DRV, "adding blkvsc_req to pending_list - blkvsc_req %p start_sect %lu sect_count %ld (%lu %ld)\n", |
bafd2c2a | 1087 | blkvsc_req, (unsigned long)blkvsc_req->sector_start, blkvsc_req->sector_count, (unsigned long) start_sector, (unsigned long) num_sectors); |
f82bd046 HJ |
1088 | |
1089 | list_add_tail(&blkvsc_req->pend_entry, &blkdev->pending_list); | |
1090 | } | |
1091 | else | |
1092 | { | |
1093 | ret = blkvsc_submit_request(blkvsc_req, blkvsc_request_completion); | |
1094 | if (ret == -1) | |
1095 | { | |
1096 | pending = 1; | |
1097 | list_add_tail(&blkvsc_req->pend_entry, &blkdev->pending_list); | |
1098 | } | |
1099 | ||
627c156d BP |
1100 | DPRINT_DBG(BLKVSC_DRV, "submitted blkvsc_req %p start_sect %lu sect_count %ld (%lu %ld) ret %d\n", |
1101 | blkvsc_req, (unsigned long) blkvsc_req->sector_start, blkvsc_req->sector_count, (unsigned long) start_sector, num_sectors, ret); | |
f82bd046 HJ |
1102 | } |
1103 | } | |
1104 | ||
1105 | return pending; | |
1106 | } | |
1107 | ||
0b3f6834 | 1108 | static void blkvsc_cmd_completion(struct hv_storvsc_request *request) |
f82bd046 HJ |
1109 | { |
1110 | struct blkvsc_request *blkvsc_req=(struct blkvsc_request*)request->Context; | |
1111 | struct block_device_context *blkdev = (struct block_device_context*)blkvsc_req->dev; | |
1112 | ||
1113 | struct scsi_sense_hdr sense_hdr; | |
1114 | ||
1115 | DPRINT_DBG(BLKVSC_DRV, "blkvsc_cmd_completion() - req %p\n", blkvsc_req); | |
1116 | ||
1117 | blkdev->num_outstanding_reqs--; | |
1118 | ||
1119 | if (blkvsc_req->request.Status) | |
1120 | { | |
1121 | if (scsi_normalize_sense(blkvsc_req->sense_buffer, SCSI_SENSE_BUFFERSIZE, &sense_hdr)) | |
1122 | { | |
1123 | scsi_print_sense_hdr("blkvsc", &sense_hdr); | |
1124 | } | |
1125 | } | |
1126 | ||
1127 | blkvsc_req->cond =1; | |
1128 | wake_up_interruptible(&blkvsc_req->wevent); | |
1129 | } | |
1130 | ||
0b3f6834 | 1131 | static void blkvsc_request_completion(struct hv_storvsc_request *request) |
f82bd046 HJ |
1132 | { |
1133 | struct blkvsc_request *blkvsc_req=(struct blkvsc_request*)request->Context; | |
1134 | struct block_device_context *blkdev = (struct block_device_context*)blkvsc_req->dev; | |
1135 | unsigned long flags; | |
1136 | struct blkvsc_request *comp_req, *tmp; | |
1137 | ||
1138 | ASSERT(blkvsc_req->group); | |
1139 | ||
627c156d | 1140 | DPRINT_DBG(BLKVSC_DRV, "blkdev %p blkvsc_req %p group %p type %s sect_start %lu sect_count %ld len %d group outstd %d total outstd %d\n", |
f82bd046 HJ |
1141 | blkdev, |
1142 | blkvsc_req, | |
1143 | blkvsc_req->group, | |
1144 | (blkvsc_req->write)?"WRITE":"READ", | |
627c156d | 1145 | (unsigned long) blkvsc_req->sector_start, |
f82bd046 HJ |
1146 | blkvsc_req->sector_count, |
1147 | blkvsc_req->request.DataBuffer.Length, | |
1148 | blkvsc_req->group->outstanding, | |
1149 | blkdev->num_outstanding_reqs); | |
1150 | ||
1151 | spin_lock_irqsave(&blkdev->lock, flags); | |
1152 | ||
1153 | blkdev->num_outstanding_reqs--; | |
1154 | blkvsc_req->group->outstanding--; | |
1155 | ||
454f18a9 BP |
1156 | /* |
1157 | * Only start processing when all the blkvsc_reqs are | |
1158 | * completed. This guarantees no out-of-order blkvsc_req | |
1159 | * completion when calling end_that_request_first() | |
1160 | */ | |
f82bd046 HJ |
1161 | if (blkvsc_req->group->outstanding == 0) |
1162 | { | |
1163 | list_for_each_entry_safe(comp_req, tmp, &blkvsc_req->group->blkvsc_req_list, req_entry) | |
1164 | { | |
627c156d | 1165 | DPRINT_DBG(BLKVSC_DRV, "completing blkvsc_req %p sect_start %lu sect_count %ld \n", |
f82bd046 | 1166 | comp_req, |
627c156d | 1167 | (unsigned long) comp_req->sector_start, |
f82bd046 HJ |
1168 | comp_req->sector_count); |
1169 | ||
1170 | list_del(&comp_req->req_entry); | |
1171 | ||
f82bd046 HJ |
1172 | if (!__blk_end_request( |
1173 | comp_req->req, | |
1174 | (!comp_req->request.Status ? 0: -EIO), | |
1175 | comp_req->sector_count * blkdev->sector_size)) | |
1176 | { | |
454f18a9 | 1177 | /* All the sectors have been xferred ie the request is done */ |
f82bd046 HJ |
1178 | DPRINT_DBG(BLKVSC_DRV, "req %p COMPLETED\n", comp_req->req); |
1179 | kmem_cache_free(blkdev->request_pool, comp_req->group); | |
1180 | } | |
f82bd046 HJ |
1181 | |
1182 | kmem_cache_free(blkdev->request_pool, comp_req); | |
1183 | } | |
1184 | ||
1185 | if (!blkdev->shutting_down) | |
1186 | { | |
1187 | blkvsc_do_pending_reqs(blkdev); | |
1188 | blk_start_queue(blkdev->gd->queue); | |
1189 | blkvsc_request(blkdev->gd->queue); | |
1190 | } | |
1191 | } | |
1192 | ||
1193 | spin_unlock_irqrestore(&blkdev->lock, flags); | |
1194 | } | |
1195 | ||
1196 | static int blkvsc_cancel_pending_reqs(struct block_device_context *blkdev) | |
1197 | { | |
1198 | struct blkvsc_request *pend_req, *tmp; | |
1199 | struct blkvsc_request *comp_req, *tmp2; | |
1200 | ||
1201 | int ret=0; | |
1202 | ||
1203 | DPRINT_DBG(BLKVSC_DRV, "blkvsc_cancel_pending_reqs()"); | |
1204 | ||
454f18a9 | 1205 | /* Flush the pending list first */ |
f82bd046 HJ |
1206 | list_for_each_entry_safe(pend_req, tmp, &blkdev->pending_list, pend_entry) |
1207 | { | |
454f18a9 BP |
1208 | /* |
1209 | * The pend_req could be part of a partially completed | |
1210 | * request. If so, complete those req first until we | |
1211 | * hit the pend_req | |
1212 | */ | |
f82bd046 HJ |
1213 | list_for_each_entry_safe(comp_req, tmp2, &pend_req->group->blkvsc_req_list, req_entry) |
1214 | { | |
627c156d | 1215 | DPRINT_DBG(BLKVSC_DRV, "completing blkvsc_req %p sect_start %lu sect_count %ld \n", |
f82bd046 | 1216 | comp_req, |
627c156d | 1217 | (unsigned long) comp_req->sector_start, |
f82bd046 HJ |
1218 | comp_req->sector_count); |
1219 | ||
1220 | if (comp_req == pend_req) | |
1221 | break; | |
1222 | ||
1223 | list_del(&comp_req->req_entry); | |
1224 | ||
1225 | if (comp_req->req) | |
1226 | { | |
f82bd046 HJ |
1227 | ret = __blk_end_request( |
1228 | comp_req->req, | |
1229 | (!comp_req->request.Status ? 0 : -EIO), | |
1230 | comp_req->sector_count * blkdev->sector_size); | |
f82bd046 HJ |
1231 | ASSERT(ret != 0); |
1232 | } | |
1233 | ||
1234 | kmem_cache_free(blkdev->request_pool, comp_req); | |
1235 | } | |
1236 | ||
1237 | DPRINT_DBG(BLKVSC_DRV, "cancelling pending request - %p\n", pend_req); | |
1238 | ||
1239 | list_del(&pend_req->pend_entry); | |
1240 | ||
1241 | list_del(&pend_req->req_entry); | |
1242 | ||
1243 | if (comp_req->req) | |
1244 | { | |
f82bd046 HJ |
1245 | if (!__blk_end_request( |
1246 | pend_req->req, | |
1247 | -EIO, | |
1248 | pend_req->sector_count * blkdev->sector_size)) | |
1249 | { | |
454f18a9 | 1250 | /* All the sectors have been xferred ie the request is done */ |
f82bd046 HJ |
1251 | DPRINT_DBG(BLKVSC_DRV, "blkvsc_cancel_pending_reqs() - req %p COMPLETED\n", pend_req->req); |
1252 | kmem_cache_free(blkdev->request_pool, pend_req->group); | |
1253 | } | |
f82bd046 HJ |
1254 | } |
1255 | ||
1256 | kmem_cache_free(blkdev->request_pool, pend_req); | |
1257 | } | |
1258 | ||
1259 | return ret; | |
1260 | } | |
1261 | ||
1262 | static int blkvsc_do_pending_reqs(struct block_device_context *blkdev) | |
1263 | { | |
1264 | struct blkvsc_request *pend_req, *tmp; | |
1265 | int ret=0; | |
1266 | ||
454f18a9 | 1267 | /* Flush the pending list first */ |
f82bd046 HJ |
1268 | list_for_each_entry_safe(pend_req, tmp, &blkdev->pending_list, pend_entry) |
1269 | { | |
1270 | DPRINT_DBG(BLKVSC_DRV, "working off pending_list - %p\n", pend_req); | |
1271 | ||
1272 | ret = blkvsc_submit_request(pend_req, blkvsc_request_completion); | |
1273 | if (ret != 0) | |
1274 | { | |
1275 | break; | |
1276 | } | |
1277 | else | |
1278 | { | |
1279 | list_del(&pend_req->pend_entry); | |
1280 | } | |
1281 | } | |
1282 | ||
1283 | return ret; | |
1284 | } | |
1285 | ||
1286 | static void blkvsc_request(struct request_queue *queue) | |
1287 | { | |
1288 | struct block_device_context *blkdev = NULL; | |
1289 | struct request *req; | |
1290 | int ret=0; | |
1291 | ||
1292 | DPRINT_DBG(BLKVSC_DRV, "- enter \n"); | |
0fce4c2f | 1293 | while ((req = blk_peek_request(queue)) != NULL) |
f82bd046 HJ |
1294 | { |
1295 | DPRINT_DBG(BLKVSC_DRV, "- req %p\n", req); | |
1296 | ||
1297 | blkdev = req->rq_disk->private_data; | |
1298 | if (blkdev->shutting_down || !blk_fs_request(req) || blkdev->media_not_present) { | |
0fce4c2f | 1299 | __blk_end_request_cur(req, 0); |
f82bd046 HJ |
1300 | continue; |
1301 | } | |
1302 | ||
1303 | ret = blkvsc_do_pending_reqs(blkdev); | |
1304 | ||
1305 | if (ret != 0) | |
1306 | { | |
1307 | DPRINT_DBG(BLKVSC_DRV, "- stop queue - pending_list not empty\n"); | |
1308 | blk_stop_queue(queue); | |
1309 | break; | |
1310 | } | |
1311 | ||
0fce4c2f | 1312 | blk_start_request(req); |
f82bd046 HJ |
1313 | |
1314 | ret = blkvsc_do_request(blkdev, req); | |
1315 | if (ret > 0) | |
1316 | { | |
1317 | DPRINT_DBG(BLKVSC_DRV, "- stop queue - no room\n"); | |
1318 | blk_stop_queue(queue); | |
1319 | break; | |
1320 | } | |
1321 | else if (ret < 0) | |
1322 | { | |
1323 | DPRINT_DBG(BLKVSC_DRV, "- stop queue - no mem\n"); | |
1324 | blk_requeue_request(queue, req); | |
1325 | blk_stop_queue(queue); | |
1326 | break; | |
1327 | } | |
1328 | } | |
1329 | } | |
1330 | ||
39635f7d | 1331 | static int blkvsc_open(struct block_device *bdev, fmode_t mode) |
f82bd046 | 1332 | { |
39635f7d | 1333 | struct block_device_context *blkdev = bdev->bd_disk->private_data; |
f82bd046 HJ |
1334 | |
1335 | DPRINT_DBG(BLKVSC_DRV, "- users %d disk %s\n", blkdev->users, blkdev->gd->disk_name); | |
1336 | ||
1337 | spin_lock(&blkdev->lock); | |
1338 | ||
1339 | if (!blkdev->users && blkdev->device_type == DVD_TYPE) | |
1340 | { | |
1341 | spin_unlock(&blkdev->lock); | |
39635f7d | 1342 | check_disk_change(bdev); |
f82bd046 HJ |
1343 | spin_lock(&blkdev->lock); |
1344 | } | |
1345 | ||
1346 | blkdev->users++; | |
1347 | ||
1348 | spin_unlock(&blkdev->lock); | |
1349 | return 0; | |
1350 | } | |
1351 | ||
77d2d9da | 1352 | static int blkvsc_release(struct gendisk *disk, fmode_t mode) |
f82bd046 | 1353 | { |
77d2d9da | 1354 | struct block_device_context *blkdev = disk->private_data; |
f82bd046 HJ |
1355 | |
1356 | DPRINT_DBG(BLKVSC_DRV, "- users %d disk %s\n", blkdev->users, blkdev->gd->disk_name); | |
1357 | ||
1358 | spin_lock(&blkdev->lock); | |
1359 | if (blkdev->users == 1) | |
1360 | { | |
1361 | spin_unlock(&blkdev->lock); | |
1362 | blkvsc_do_flush(blkdev); | |
1363 | spin_lock(&blkdev->lock); | |
1364 | } | |
1365 | ||
1366 | blkdev->users--; | |
1367 | ||
1368 | spin_unlock(&blkdev->lock); | |
1369 | return 0; | |
1370 | } | |
1371 | ||
1372 | static int blkvsc_media_changed(struct gendisk *gd) | |
1373 | { | |
1374 | DPRINT_DBG(BLKVSC_DRV, "- enter\n"); | |
1375 | ||
1376 | return 1; | |
1377 | } | |
1378 | ||
1379 | static int blkvsc_revalidate_disk(struct gendisk *gd) | |
1380 | { | |
1381 | struct block_device_context *blkdev = gd->private_data; | |
1382 | ||
1383 | DPRINT_DBG(BLKVSC_DRV, "- enter\n"); | |
1384 | ||
1385 | if (blkdev->device_type == DVD_TYPE) | |
1386 | { | |
1387 | blkvsc_do_read_capacity(blkdev); | |
1388 | set_capacity(blkdev->gd, blkdev->capacity * (blkdev->sector_size/512)); | |
0fce4c2f | 1389 | blk_queue_logical_block_size(gd->queue, blkdev->sector_size); |
f82bd046 HJ |
1390 | } |
1391 | return 0; | |
1392 | } | |
1393 | ||
bd1de709 | 1394 | static int blkvsc_getgeo(struct block_device *bd, struct hd_geometry *hg) |
f82bd046 HJ |
1395 | { |
1396 | sector_t total_sectors = get_capacity(bd->bd_disk); | |
1397 | sector_t cylinder_times_heads=0; | |
1398 | sector_t temp=0; | |
1399 | ||
1400 | int sectors_per_track=0; | |
1401 | int heads=0; | |
1402 | int cylinders=0; | |
1403 | int rem=0; | |
1404 | ||
1405 | if (total_sectors > (65535 * 16 * 255)) { | |
454f18a9 | 1406 | total_sectors = (65535 * 16 * 255); |
f82bd046 HJ |
1407 | } |
1408 | ||
1409 | if (total_sectors >= (65535 * 16 * 63)) { | |
454f18a9 BP |
1410 | sectors_per_track = 255; |
1411 | heads = 16; | |
f82bd046 HJ |
1412 | |
1413 | cylinder_times_heads = total_sectors; | |
454f18a9 | 1414 | rem = sector_div(cylinder_times_heads, sectors_per_track); /* sector_div stores the quotient in cylinder_times_heads */ |
f82bd046 HJ |
1415 | } |
1416 | else | |
1417 | { | |
454f18a9 | 1418 | sectors_per_track = 17; |
f82bd046 HJ |
1419 | |
1420 | cylinder_times_heads = total_sectors; | |
454f18a9 | 1421 | rem = sector_div(cylinder_times_heads, sectors_per_track); /* sector_div stores the quotient in cylinder_times_heads */ |
f82bd046 HJ |
1422 | |
1423 | temp = cylinder_times_heads + 1023; | |
454f18a9 | 1424 | rem = sector_div(temp, 1024); /* sector_div stores the quotient in temp */ |
f82bd046 HJ |
1425 | |
1426 | heads = temp; | |
1427 | ||
454f18a9 BP |
1428 | if (heads < 4) { |
1429 | heads = 4; | |
1430 | } | |
f82bd046 | 1431 | |
454f18a9 BP |
1432 | if (cylinder_times_heads >= (heads * 1024) || (heads > 16)) { |
1433 | sectors_per_track = 31; | |
1434 | heads = 16; | |
f82bd046 HJ |
1435 | |
1436 | cylinder_times_heads = total_sectors; | |
454f18a9 BP |
1437 | rem = sector_div(cylinder_times_heads, sectors_per_track); /* sector_div stores the quotient in cylinder_times_heads */ |
1438 | } | |
f82bd046 | 1439 | |
454f18a9 BP |
1440 | if (cylinder_times_heads >= (heads * 1024)) { |
1441 | sectors_per_track = 63; | |
1442 | heads = 16; | |
f82bd046 HJ |
1443 | |
1444 | cylinder_times_heads = total_sectors; | |
454f18a9 BP |
1445 | rem = sector_div(cylinder_times_heads, sectors_per_track); /* sector_div stores the quotient in cylinder_times_heads */ |
1446 | } | |
f82bd046 HJ |
1447 | } |
1448 | ||
1449 | temp = cylinder_times_heads; | |
454f18a9 | 1450 | rem = sector_div(temp, heads); /* sector_div stores the quotient in temp */ |
f82bd046 HJ |
1451 | cylinders = temp; |
1452 | ||
1453 | hg->heads = heads; | |
1454 | hg->sectors = sectors_per_track; | |
1455 | hg->cylinders = cylinders; | |
1456 | ||
1457 | DPRINT_INFO(BLKVSC_DRV, "CHS (%d, %d, %d)", cylinders, heads, sectors_per_track); | |
1458 | ||
1459 | return 0; | |
1460 | } | |
1461 | ||
dfe8b2d9 BP |
1462 | static int blkvsc_ioctl(struct block_device *bd, fmode_t mode, |
1463 | unsigned cmd, unsigned long argument) | |
f82bd046 | 1464 | { |
b5788529 | 1465 | /* struct block_device_context *blkdev = bd->bd_disk->private_data; */ |
f82bd046 HJ |
1466 | int ret=0; |
1467 | ||
1468 | switch (cmd) | |
1469 | { | |
454f18a9 BP |
1470 | /* TODO: I think there is certain format for HDIO_GET_IDENTITY rather than just */ |
1471 | /* a GUID. Commented it out for now. */ | |
f82bd046 HJ |
1472 | /*case HDIO_GET_IDENTITY: |
1473 | DPRINT_INFO(BLKVSC_DRV, "HDIO_GET_IDENTITY\n"); | |
1474 | ||
1475 | if (copy_to_user((void __user *)arg, blkdev->device_id, blkdev->device_id_len)) | |
1476 | { | |
1477 | ret = -EFAULT; | |
1478 | } | |
1479 | ||
1480 | break;*/ | |
1481 | default: | |
1482 | ret = -EINVAL; | |
1483 | break; | |
1484 | } | |
1485 | ||
1486 | return ret; | |
1487 | } | |
1488 | ||
1489 | ||
1490 | MODULE_LICENSE("GPL"); | |
1491 | ||
1492 | static int __init blkvsc_init(void) | |
1493 | { | |
1494 | int ret; | |
1495 | ||
454f18a9 | 1496 | ASSERT(sizeof(sector_t) == 8); /* Make sure CONFIG_LBD is set */ |
f82bd046 HJ |
1497 | |
1498 | DPRINT_ENTER(BLKVSC_DRV); | |
1499 | ||
1500 | DPRINT_INFO(BLKVSC_DRV, "Blkvsc initializing...."); | |
1501 | ||
1502 | ret = blkvsc_drv_init(BlkVscInitialize); | |
1503 | ||
1504 | DPRINT_EXIT(BLKVSC_DRV); | |
1505 | ||
1506 | return ret; | |
1507 | } | |
1508 | ||
1509 | static void __exit blkvsc_exit(void) | |
1510 | { | |
1511 | DPRINT_ENTER(BLKVSC_DRV); | |
1512 | ||
1513 | blkvsc_drv_exit(); | |
1514 | ||
1515 | DPRINT_ENTER(BLKVSC_DRV); | |
1516 | } | |
1517 | ||
1518 | module_param(blkvsc_ringbuffer_size, int, S_IRUGO); | |
1519 | ||
1520 | module_init(blkvsc_init); | |
1521 | module_exit(blkvsc_exit); | |
1522 | ||
454f18a9 | 1523 | /* eof */ |