9fd097b149
In-kernel disk event polling doesn't matter for legacy/fringe drivers and may lead to infinite event loop if ->check_events() implementation generates events on level condition instead of edge. Now that block layer supports suppressing exporting unlisted events, simply leaving disk->events cleared allows these drivers to keep the internal revalidation behavior intact while avoiding weird interactions with userland event handler. Signed-off-by: Tejun Heo <tj@kernel.org> Cc: Kay Sievers <kay.sievers@vrfy.org> Signed-off-by: Jens Axboe <jaxboe@fusionio.com>
444 lines
11 KiB
C
444 lines
11 KiB
C
/*
|
|
* drivers/s390/char/tape_block.c
|
|
* block device frontend for tape device driver
|
|
*
|
|
* S390 and zSeries version
|
|
* Copyright (C) 2001,2003 IBM Deutschland Entwicklung GmbH, IBM Corporation
|
|
* Author(s): Carsten Otte <cotte@de.ibm.com>
|
|
* Tuan Ngo-Anh <ngoanh@de.ibm.com>
|
|
* Martin Schwidefsky <schwidefsky@de.ibm.com>
|
|
* Stefan Bader <shbader@de.ibm.com>
|
|
*/
|
|
|
|
#define KMSG_COMPONENT "tape"
|
|
#define pr_fmt(fmt) KMSG_COMPONENT ": " fmt
|
|
|
|
#include <linux/fs.h>
|
|
#include <linux/module.h>
|
|
#include <linux/blkdev.h>
|
|
#include <linux/mutex.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/buffer_head.h>
|
|
#include <linux/kernel.h>
|
|
|
|
#include <asm/debug.h>
|
|
|
|
#define TAPE_DBF_AREA tape_core_dbf
|
|
|
|
#include "tape.h"
|
|
|
|
#define TAPEBLOCK_MAX_SEC 100
|
|
#define TAPEBLOCK_MIN_REQUEUE 3
|
|
|
|
/*
|
|
* 2003/11/25 Stefan Bader <shbader@de.ibm.com>
|
|
*
|
|
* In 2.5/2.6 the block device request function is very likely to be called
|
|
* with disabled interrupts (e.g. generic_unplug_device). So the driver can't
|
|
* just call any function that tries to allocate CCW requests from that con-
|
|
* text since it might sleep. There are two choices to work around this:
|
|
* a) do not allocate with kmalloc but use its own memory pool
|
|
* b) take requests from the queue outside that context, knowing that
|
|
* allocation might sleep
|
|
*/
|
|
|
|
/*
|
|
* file operation structure for tape block frontend
|
|
*/
|
|
static DEFINE_MUTEX(tape_block_mutex);
|
|
static int tapeblock_open(struct block_device *, fmode_t);
|
|
static int tapeblock_release(struct gendisk *, fmode_t);
|
|
static unsigned int tapeblock_check_events(struct gendisk *, unsigned int);
|
|
static int tapeblock_revalidate_disk(struct gendisk *);
|
|
|
|
static const struct block_device_operations tapeblock_fops = {
|
|
.owner = THIS_MODULE,
|
|
.open = tapeblock_open,
|
|
.release = tapeblock_release,
|
|
.check_events = tapeblock_check_events,
|
|
.revalidate_disk = tapeblock_revalidate_disk,
|
|
};
|
|
|
|
static int tapeblock_major = 0;
|
|
|
|
static void
|
|
tapeblock_trigger_requeue(struct tape_device *device)
|
|
{
|
|
/* Protect against rescheduling. */
|
|
if (atomic_cmpxchg(&device->blk_data.requeue_scheduled, 0, 1) != 0)
|
|
return;
|
|
schedule_work(&device->blk_data.requeue_task);
|
|
}
|
|
|
|
/*
|
|
* Post finished request.
|
|
*/
|
|
static void
|
|
__tapeblock_end_request(struct tape_request *ccw_req, void *data)
|
|
{
|
|
struct tape_device *device;
|
|
struct request *req;
|
|
|
|
DBF_LH(6, "__tapeblock_end_request()\n");
|
|
|
|
device = ccw_req->device;
|
|
req = (struct request *) data;
|
|
blk_end_request_all(req, (ccw_req->rc == 0) ? 0 : -EIO);
|
|
if (ccw_req->rc == 0)
|
|
/* Update position. */
|
|
device->blk_data.block_position =
|
|
(blk_rq_pos(req) + blk_rq_sectors(req)) >> TAPEBLOCK_HSEC_S2B;
|
|
else
|
|
/* We lost the position information due to an error. */
|
|
device->blk_data.block_position = -1;
|
|
device->discipline->free_bread(ccw_req);
|
|
if (!list_empty(&device->req_queue) ||
|
|
blk_peek_request(device->blk_data.request_queue))
|
|
tapeblock_trigger_requeue(device);
|
|
}
|
|
|
|
/*
|
|
* Feed the tape device CCW queue with requests supplied in a list.
|
|
*/
|
|
static int
|
|
tapeblock_start_request(struct tape_device *device, struct request *req)
|
|
{
|
|
struct tape_request * ccw_req;
|
|
int rc;
|
|
|
|
DBF_LH(6, "tapeblock_start_request(%p, %p)\n", device, req);
|
|
|
|
ccw_req = device->discipline->bread(device, req);
|
|
if (IS_ERR(ccw_req)) {
|
|
DBF_EVENT(1, "TBLOCK: bread failed\n");
|
|
blk_end_request_all(req, -EIO);
|
|
return PTR_ERR(ccw_req);
|
|
}
|
|
ccw_req->callback = __tapeblock_end_request;
|
|
ccw_req->callback_data = (void *) req;
|
|
ccw_req->retries = TAPEBLOCK_RETRIES;
|
|
|
|
rc = tape_do_io_async(device, ccw_req);
|
|
if (rc) {
|
|
/*
|
|
* Start/enqueueing failed. No retries in
|
|
* this case.
|
|
*/
|
|
blk_end_request_all(req, -EIO);
|
|
device->discipline->free_bread(ccw_req);
|
|
}
|
|
|
|
return rc;
|
|
}
|
|
|
|
/*
|
|
* Move requests from the block device request queue to the tape device ccw
|
|
* queue.
|
|
*/
|
|
static void
|
|
tapeblock_requeue(struct work_struct *work) {
|
|
struct tape_blk_data * blkdat;
|
|
struct tape_device * device;
|
|
struct request_queue * queue;
|
|
int nr_queued;
|
|
struct request * req;
|
|
struct list_head * l;
|
|
int rc;
|
|
|
|
blkdat = container_of(work, struct tape_blk_data, requeue_task);
|
|
device = blkdat->device;
|
|
if (!device)
|
|
return;
|
|
|
|
spin_lock_irq(get_ccwdev_lock(device->cdev));
|
|
queue = device->blk_data.request_queue;
|
|
|
|
/* Count number of requests on ccw queue. */
|
|
nr_queued = 0;
|
|
list_for_each(l, &device->req_queue)
|
|
nr_queued++;
|
|
spin_unlock(get_ccwdev_lock(device->cdev));
|
|
|
|
spin_lock_irq(&device->blk_data.request_queue_lock);
|
|
while (
|
|
blk_peek_request(queue) &&
|
|
nr_queued < TAPEBLOCK_MIN_REQUEUE
|
|
) {
|
|
req = blk_fetch_request(queue);
|
|
if (rq_data_dir(req) == WRITE) {
|
|
DBF_EVENT(1, "TBLOCK: Rejecting write request\n");
|
|
spin_unlock_irq(&device->blk_data.request_queue_lock);
|
|
blk_end_request_all(req, -EIO);
|
|
spin_lock_irq(&device->blk_data.request_queue_lock);
|
|
continue;
|
|
}
|
|
nr_queued++;
|
|
spin_unlock_irq(&device->blk_data.request_queue_lock);
|
|
rc = tapeblock_start_request(device, req);
|
|
spin_lock_irq(&device->blk_data.request_queue_lock);
|
|
}
|
|
spin_unlock_irq(&device->blk_data.request_queue_lock);
|
|
atomic_set(&device->blk_data.requeue_scheduled, 0);
|
|
}
|
|
|
|
/*
|
|
* Tape request queue function. Called from ll_rw_blk.c
|
|
*/
|
|
static void
|
|
tapeblock_request_fn(struct request_queue *queue)
|
|
{
|
|
struct tape_device *device;
|
|
|
|
device = (struct tape_device *) queue->queuedata;
|
|
DBF_LH(6, "tapeblock_request_fn(device=%p)\n", device);
|
|
BUG_ON(device == NULL);
|
|
tapeblock_trigger_requeue(device);
|
|
}
|
|
|
|
/*
|
|
* This function is called for every new tapedevice
|
|
*/
|
|
int
|
|
tapeblock_setup_device(struct tape_device * device)
|
|
{
|
|
struct tape_blk_data * blkdat;
|
|
struct gendisk * disk;
|
|
int rc;
|
|
|
|
blkdat = &device->blk_data;
|
|
blkdat->device = device;
|
|
spin_lock_init(&blkdat->request_queue_lock);
|
|
atomic_set(&blkdat->requeue_scheduled, 0);
|
|
|
|
blkdat->request_queue = blk_init_queue(
|
|
tapeblock_request_fn,
|
|
&blkdat->request_queue_lock
|
|
);
|
|
if (!blkdat->request_queue)
|
|
return -ENOMEM;
|
|
|
|
rc = elevator_change(blkdat->request_queue, "noop");
|
|
if (rc)
|
|
goto cleanup_queue;
|
|
|
|
blk_queue_logical_block_size(blkdat->request_queue, TAPEBLOCK_HSEC_SIZE);
|
|
blk_queue_max_hw_sectors(blkdat->request_queue, TAPEBLOCK_MAX_SEC);
|
|
blk_queue_max_segments(blkdat->request_queue, -1L);
|
|
blk_queue_max_segment_size(blkdat->request_queue, -1L);
|
|
blk_queue_segment_boundary(blkdat->request_queue, -1L);
|
|
|
|
disk = alloc_disk(1);
|
|
if (!disk) {
|
|
rc = -ENOMEM;
|
|
goto cleanup_queue;
|
|
}
|
|
|
|
disk->major = tapeblock_major;
|
|
disk->first_minor = device->first_minor;
|
|
disk->fops = &tapeblock_fops;
|
|
disk->private_data = tape_get_device(device);
|
|
disk->queue = blkdat->request_queue;
|
|
set_capacity(disk, 0);
|
|
sprintf(disk->disk_name, "btibm%d",
|
|
device->first_minor / TAPE_MINORS_PER_DEV);
|
|
|
|
blkdat->disk = disk;
|
|
blkdat->medium_changed = 1;
|
|
blkdat->request_queue->queuedata = tape_get_device(device);
|
|
|
|
add_disk(disk);
|
|
|
|
tape_get_device(device);
|
|
INIT_WORK(&blkdat->requeue_task, tapeblock_requeue);
|
|
|
|
return 0;
|
|
|
|
cleanup_queue:
|
|
blk_cleanup_queue(blkdat->request_queue);
|
|
blkdat->request_queue = NULL;
|
|
|
|
return rc;
|
|
}
|
|
|
|
void
|
|
tapeblock_cleanup_device(struct tape_device *device)
|
|
{
|
|
flush_work_sync(&device->blk_data.requeue_task);
|
|
tape_put_device(device);
|
|
|
|
if (!device->blk_data.disk) {
|
|
goto cleanup_queue;
|
|
}
|
|
|
|
del_gendisk(device->blk_data.disk);
|
|
device->blk_data.disk->private_data = NULL;
|
|
tape_put_device(device);
|
|
put_disk(device->blk_data.disk);
|
|
|
|
device->blk_data.disk = NULL;
|
|
cleanup_queue:
|
|
device->blk_data.request_queue->queuedata = NULL;
|
|
tape_put_device(device);
|
|
|
|
blk_cleanup_queue(device->blk_data.request_queue);
|
|
device->blk_data.request_queue = NULL;
|
|
}
|
|
|
|
/*
|
|
* Detect number of blocks of the tape.
|
|
* FIXME: can we extent this to detect the blocks size as well ?
|
|
*/
|
|
static int
|
|
tapeblock_revalidate_disk(struct gendisk *disk)
|
|
{
|
|
struct tape_device * device;
|
|
unsigned int nr_of_blks;
|
|
int rc;
|
|
|
|
device = (struct tape_device *) disk->private_data;
|
|
BUG_ON(!device);
|
|
|
|
if (!device->blk_data.medium_changed)
|
|
return 0;
|
|
|
|
rc = tape_mtop(device, MTFSFM, 1);
|
|
if (rc)
|
|
return rc;
|
|
|
|
rc = tape_mtop(device, MTTELL, 1);
|
|
if (rc < 0)
|
|
return rc;
|
|
|
|
pr_info("%s: Determining the size of the recorded area...\n",
|
|
dev_name(&device->cdev->dev));
|
|
DBF_LH(3, "Image file ends at %d\n", rc);
|
|
nr_of_blks = rc;
|
|
|
|
/* This will fail for the first file. Catch the error by checking the
|
|
* position. */
|
|
tape_mtop(device, MTBSF, 1);
|
|
|
|
rc = tape_mtop(device, MTTELL, 1);
|
|
if (rc < 0)
|
|
return rc;
|
|
|
|
if (rc > nr_of_blks)
|
|
return -EINVAL;
|
|
|
|
DBF_LH(3, "Image file starts at %d\n", rc);
|
|
device->bof = rc;
|
|
nr_of_blks -= rc;
|
|
|
|
pr_info("%s: The size of the recorded area is %i blocks\n",
|
|
dev_name(&device->cdev->dev), nr_of_blks);
|
|
set_capacity(device->blk_data.disk,
|
|
nr_of_blks*(TAPEBLOCK_HSEC_SIZE/512));
|
|
|
|
device->blk_data.block_position = 0;
|
|
device->blk_data.medium_changed = 0;
|
|
return 0;
|
|
}
|
|
|
|
static unsigned int
|
|
tapeblock_check_events(struct gendisk *disk, unsigned int clearing)
|
|
{
|
|
struct tape_device *device;
|
|
|
|
device = (struct tape_device *) disk->private_data;
|
|
DBF_LH(6, "tapeblock_medium_changed(%p) = %d\n",
|
|
device, device->blk_data.medium_changed);
|
|
|
|
return device->blk_data.medium_changed ? DISK_EVENT_MEDIA_CHANGE : 0;
|
|
}
|
|
|
|
/*
|
|
* Block frontend tape device open function.
|
|
*/
|
|
static int
|
|
tapeblock_open(struct block_device *bdev, fmode_t mode)
|
|
{
|
|
struct gendisk * disk = bdev->bd_disk;
|
|
struct tape_device * device;
|
|
int rc;
|
|
|
|
mutex_lock(&tape_block_mutex);
|
|
device = tape_get_device(disk->private_data);
|
|
|
|
if (device->required_tapemarks) {
|
|
DBF_EVENT(2, "TBLOCK: missing tapemarks\n");
|
|
pr_warning("%s: Opening the tape failed because of missing "
|
|
"end-of-file marks\n", dev_name(&device->cdev->dev));
|
|
rc = -EPERM;
|
|
goto put_device;
|
|
}
|
|
|
|
rc = tape_open(device);
|
|
if (rc)
|
|
goto put_device;
|
|
|
|
rc = tapeblock_revalidate_disk(disk);
|
|
if (rc)
|
|
goto release;
|
|
|
|
/*
|
|
* Note: The reference to <device> is hold until the release function
|
|
* is called.
|
|
*/
|
|
tape_state_set(device, TS_BLKUSE);
|
|
mutex_unlock(&tape_block_mutex);
|
|
return 0;
|
|
|
|
release:
|
|
tape_release(device);
|
|
put_device:
|
|
tape_put_device(device);
|
|
mutex_unlock(&tape_block_mutex);
|
|
return rc;
|
|
}
|
|
|
|
/*
|
|
* Block frontend tape device release function.
|
|
*
|
|
* Note: One reference to the tape device was made by the open function. So
|
|
* we just get the pointer here and release the reference.
|
|
*/
|
|
static int
|
|
tapeblock_release(struct gendisk *disk, fmode_t mode)
|
|
{
|
|
struct tape_device *device = disk->private_data;
|
|
|
|
mutex_lock(&tape_block_mutex);
|
|
tape_state_set(device, TS_IN_USE);
|
|
tape_release(device);
|
|
tape_put_device(device);
|
|
mutex_unlock(&tape_block_mutex);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Initialize block device frontend.
|
|
*/
|
|
int
|
|
tapeblock_init(void)
|
|
{
|
|
int rc;
|
|
|
|
/* Register the tape major number to the kernel */
|
|
rc = register_blkdev(tapeblock_major, "tBLK");
|
|
if (rc < 0)
|
|
return rc;
|
|
|
|
if (tapeblock_major == 0)
|
|
tapeblock_major = rc;
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Deregister major for block device frontend
|
|
*/
|
|
void
|
|
tapeblock_exit(void)
|
|
{
|
|
unregister_blkdev(tapeblock_major, "tBLK");
|
|
}
|