Commit 846785e6 authored by Christoph Hellwig's avatar Christoph Hellwig Committed by Jens Axboe

dm: don't return errnos from ->map

Instead use the special DM_MAPIO_KILL return value to return -EIO just
like we do for the request based path.  Note that dm-log-writes returned
-ENOMEM in a few places, which now becomes -EIO instead.  No consumer
treats -ENOMEM special so this shouldn't be an issue (and it should
use a mempool to start with to make guaranteed progress).
Signed-off-by: default avatarChristoph Hellwig <hch@lst.de>
Signed-off-by: default avatarMike Snitzer <snitzer@redhat.com>
Signed-off-by: default avatarJens Axboe <axboe@fb.com>
parent 14ef1e48
...@@ -2795,10 +2795,10 @@ static int crypt_map(struct dm_target *ti, struct bio *bio) ...@@ -2795,10 +2795,10 @@ static int crypt_map(struct dm_target *ti, struct bio *bio)
* and is aligned to this size as defined in IO hints. * and is aligned to this size as defined in IO hints.
*/ */
if (unlikely((bio->bi_iter.bi_sector & ((cc->sector_size >> SECTOR_SHIFT) - 1)) != 0)) if (unlikely((bio->bi_iter.bi_sector & ((cc->sector_size >> SECTOR_SHIFT) - 1)) != 0))
return -EIO; return DM_MAPIO_KILL;
if (unlikely(bio->bi_iter.bi_size & (cc->sector_size - 1))) if (unlikely(bio->bi_iter.bi_size & (cc->sector_size - 1)))
return -EIO; return DM_MAPIO_KILL;
io = dm_per_bio_data(bio, cc->per_bio_data_size); io = dm_per_bio_data(bio, cc->per_bio_data_size);
crypt_io_init(io, cc, bio, dm_target_offset(ti, bio->bi_iter.bi_sector)); crypt_io_init(io, cc, bio, dm_target_offset(ti, bio->bi_iter.bi_sector));
......
...@@ -321,7 +321,7 @@ static int flakey_map(struct dm_target *ti, struct bio *bio) ...@@ -321,7 +321,7 @@ static int flakey_map(struct dm_target *ti, struct bio *bio)
if (bio_data_dir(bio) == READ) { if (bio_data_dir(bio) == READ) {
if (!fc->corrupt_bio_byte && !test_bit(DROP_WRITES, &fc->flags) && if (!fc->corrupt_bio_byte && !test_bit(DROP_WRITES, &fc->flags) &&
!test_bit(ERROR_WRITES, &fc->flags)) !test_bit(ERROR_WRITES, &fc->flags))
return -EIO; return DM_MAPIO_KILL;
goto map_bio; goto map_bio;
} }
...@@ -349,7 +349,7 @@ static int flakey_map(struct dm_target *ti, struct bio *bio) ...@@ -349,7 +349,7 @@ static int flakey_map(struct dm_target *ti, struct bio *bio)
/* /*
* By default, error all I/O. * By default, error all I/O.
*/ */
return -EIO; return DM_MAPIO_KILL;
} }
map_bio: map_bio:
......
...@@ -1352,13 +1352,13 @@ static int dm_integrity_map(struct dm_target *ti, struct bio *bio) ...@@ -1352,13 +1352,13 @@ static int dm_integrity_map(struct dm_target *ti, struct bio *bio)
DMERR("Too big sector number: 0x%llx + 0x%x > 0x%llx", DMERR("Too big sector number: 0x%llx + 0x%x > 0x%llx",
(unsigned long long)dio->range.logical_sector, bio_sectors(bio), (unsigned long long)dio->range.logical_sector, bio_sectors(bio),
(unsigned long long)ic->provided_data_sectors); (unsigned long long)ic->provided_data_sectors);
return -EIO; return DM_MAPIO_KILL;
} }
if (unlikely((dio->range.logical_sector | bio_sectors(bio)) & (unsigned)(ic->sectors_per_block - 1))) { if (unlikely((dio->range.logical_sector | bio_sectors(bio)) & (unsigned)(ic->sectors_per_block - 1))) {
DMERR("Bio not aligned on %u sectors: 0x%llx, 0x%x", DMERR("Bio not aligned on %u sectors: 0x%llx, 0x%x",
ic->sectors_per_block, ic->sectors_per_block,
(unsigned long long)dio->range.logical_sector, bio_sectors(bio)); (unsigned long long)dio->range.logical_sector, bio_sectors(bio));
return -EIO; return DM_MAPIO_KILL;
} }
if (ic->sectors_per_block > 1) { if (ic->sectors_per_block > 1) {
...@@ -1368,7 +1368,7 @@ static int dm_integrity_map(struct dm_target *ti, struct bio *bio) ...@@ -1368,7 +1368,7 @@ static int dm_integrity_map(struct dm_target *ti, struct bio *bio)
if (unlikely((bv.bv_offset | bv.bv_len) & ((ic->sectors_per_block << SECTOR_SHIFT) - 1))) { if (unlikely((bv.bv_offset | bv.bv_len) & ((ic->sectors_per_block << SECTOR_SHIFT) - 1))) {
DMERR("Bio vector (%u,%u) is not aligned on %u-sector boundary", DMERR("Bio vector (%u,%u) is not aligned on %u-sector boundary",
bv.bv_offset, bv.bv_len, ic->sectors_per_block); bv.bv_offset, bv.bv_len, ic->sectors_per_block);
return -EIO; return DM_MAPIO_KILL;
} }
} }
} }
...@@ -1383,18 +1383,18 @@ static int dm_integrity_map(struct dm_target *ti, struct bio *bio) ...@@ -1383,18 +1383,18 @@ static int dm_integrity_map(struct dm_target *ti, struct bio *bio)
wanted_tag_size *= ic->tag_size; wanted_tag_size *= ic->tag_size;
if (unlikely(wanted_tag_size != bip->bip_iter.bi_size)) { if (unlikely(wanted_tag_size != bip->bip_iter.bi_size)) {
DMERR("Invalid integrity data size %u, expected %u", bip->bip_iter.bi_size, wanted_tag_size); DMERR("Invalid integrity data size %u, expected %u", bip->bip_iter.bi_size, wanted_tag_size);
return -EIO; return DM_MAPIO_KILL;
} }
} }
} else { } else {
if (unlikely(bip != NULL)) { if (unlikely(bip != NULL)) {
DMERR("Unexpected integrity data when using internal hash"); DMERR("Unexpected integrity data when using internal hash");
return -EIO; return DM_MAPIO_KILL;
} }
} }
if (unlikely(ic->mode == 'R') && unlikely(dio->write)) if (unlikely(ic->mode == 'R') && unlikely(dio->write))
return -EIO; return DM_MAPIO_KILL;
get_area_and_offset(ic, dio->range.logical_sector, &area, &offset); get_area_and_offset(ic, dio->range.logical_sector, &area, &offset);
dio->metadata_block = get_metadata_sector_and_offset(ic, area, offset, &dio->metadata_offset); dio->metadata_block = get_metadata_sector_and_offset(ic, area, offset, &dio->metadata_offset);
......
...@@ -586,7 +586,7 @@ static int log_writes_map(struct dm_target *ti, struct bio *bio) ...@@ -586,7 +586,7 @@ static int log_writes_map(struct dm_target *ti, struct bio *bio)
spin_lock_irq(&lc->blocks_lock); spin_lock_irq(&lc->blocks_lock);
lc->logging_enabled = false; lc->logging_enabled = false;
spin_unlock_irq(&lc->blocks_lock); spin_unlock_irq(&lc->blocks_lock);
return -ENOMEM; return DM_MAPIO_KILL;
} }
INIT_LIST_HEAD(&block->list); INIT_LIST_HEAD(&block->list);
pb->block = block; pb->block = block;
...@@ -639,7 +639,7 @@ static int log_writes_map(struct dm_target *ti, struct bio *bio) ...@@ -639,7 +639,7 @@ static int log_writes_map(struct dm_target *ti, struct bio *bio)
spin_lock_irq(&lc->blocks_lock); spin_lock_irq(&lc->blocks_lock);
lc->logging_enabled = false; lc->logging_enabled = false;
spin_unlock_irq(&lc->blocks_lock); spin_unlock_irq(&lc->blocks_lock);
return -ENOMEM; return DM_MAPIO_KILL;
} }
src = kmap_atomic(bv.bv_page); src = kmap_atomic(bv.bv_page);
......
...@@ -559,7 +559,7 @@ static int __multipath_map_bio(struct multipath *m, struct bio *bio, struct dm_m ...@@ -559,7 +559,7 @@ static int __multipath_map_bio(struct multipath *m, struct bio *bio, struct dm_m
if (test_bit(MPATHF_QUEUE_IF_NO_PATH, &m->flags)) if (test_bit(MPATHF_QUEUE_IF_NO_PATH, &m->flags))
return DM_MAPIO_REQUEUE; return DM_MAPIO_REQUEUE;
dm_report_EIO(m); dm_report_EIO(m);
return -EIO; return DM_MAPIO_KILL;
} }
mpio->pgpath = pgpath; mpio->pgpath = pgpath;
...@@ -621,11 +621,18 @@ static void process_queued_bios(struct work_struct *work) ...@@ -621,11 +621,18 @@ static void process_queued_bios(struct work_struct *work)
blk_start_plug(&plug); blk_start_plug(&plug);
while ((bio = bio_list_pop(&bios))) { while ((bio = bio_list_pop(&bios))) {
r = __multipath_map_bio(m, bio, get_mpio_from_bio(bio)); r = __multipath_map_bio(m, bio, get_mpio_from_bio(bio));
if (r < 0 || r == DM_MAPIO_REQUEUE) { switch (r) {
case DM_MAPIO_KILL:
r = -EIO;
/*FALLTHRU*/
case DM_MAPIO_REQUEUE:
bio->bi_error = r; bio->bi_error = r;
bio_endio(bio); bio_endio(bio);
} else if (r == DM_MAPIO_REMAPPED) break;
case DM_MAPIO_REMAPPED:
generic_make_request(bio); generic_make_request(bio);
break;
}
} }
blk_finish_plug(&plug); blk_finish_plug(&plug);
} }
......
...@@ -1207,14 +1207,14 @@ static int mirror_map(struct dm_target *ti, struct bio *bio) ...@@ -1207,14 +1207,14 @@ static int mirror_map(struct dm_target *ti, struct bio *bio)
r = log->type->in_sync(log, dm_rh_bio_to_region(ms->rh, bio), 0); r = log->type->in_sync(log, dm_rh_bio_to_region(ms->rh, bio), 0);
if (r < 0 && r != -EWOULDBLOCK) if (r < 0 && r != -EWOULDBLOCK)
return r; return DM_MAPIO_KILL;
/* /*
* If region is not in-sync queue the bio. * If region is not in-sync queue the bio.
*/ */
if (!r || (r == -EWOULDBLOCK)) { if (!r || (r == -EWOULDBLOCK)) {
if (bio->bi_opf & REQ_RAHEAD) if (bio->bi_opf & REQ_RAHEAD)
return -EIO; return DM_MAPIO_KILL;
queue_bio(ms, bio, rw); queue_bio(ms, bio, rw);
return DM_MAPIO_SUBMITTED; return DM_MAPIO_SUBMITTED;
...@@ -1226,7 +1226,7 @@ static int mirror_map(struct dm_target *ti, struct bio *bio) ...@@ -1226,7 +1226,7 @@ static int mirror_map(struct dm_target *ti, struct bio *bio)
*/ */
m = choose_mirror(ms, bio->bi_iter.bi_sector); m = choose_mirror(ms, bio->bi_iter.bi_sector);
if (unlikely(!m)) if (unlikely(!m))
return -EIO; return DM_MAPIO_KILL;
dm_bio_record(&bio_record->details, bio); dm_bio_record(&bio_record->details, bio);
bio_record->m = m; bio_record->m = m;
......
...@@ -1690,7 +1690,7 @@ static int snapshot_map(struct dm_target *ti, struct bio *bio) ...@@ -1690,7 +1690,7 @@ static int snapshot_map(struct dm_target *ti, struct bio *bio)
/* Full snapshots are not usable */ /* Full snapshots are not usable */
/* To get here the table must be live so s->active is always set. */ /* To get here the table must be live so s->active is always set. */
if (!s->valid) if (!s->valid)
return -EIO; return DM_MAPIO_KILL;
/* FIXME: should only take write lock if we need /* FIXME: should only take write lock if we need
* to copy an exception */ * to copy an exception */
...@@ -1698,7 +1698,7 @@ static int snapshot_map(struct dm_target *ti, struct bio *bio) ...@@ -1698,7 +1698,7 @@ static int snapshot_map(struct dm_target *ti, struct bio *bio)
if (!s->valid || (unlikely(s->snapshot_overflowed) && if (!s->valid || (unlikely(s->snapshot_overflowed) &&
bio_data_dir(bio) == WRITE)) { bio_data_dir(bio) == WRITE)) {
r = -EIO; r = DM_MAPIO_KILL;
goto out_unlock; goto out_unlock;
} }
...@@ -1723,7 +1723,7 @@ static int snapshot_map(struct dm_target *ti, struct bio *bio) ...@@ -1723,7 +1723,7 @@ static int snapshot_map(struct dm_target *ti, struct bio *bio)
if (!s->valid || s->snapshot_overflowed) { if (!s->valid || s->snapshot_overflowed) {
free_pending_exception(pe); free_pending_exception(pe);
r = -EIO; r = DM_MAPIO_KILL;
goto out_unlock; goto out_unlock;
} }
...@@ -1741,7 +1741,7 @@ static int snapshot_map(struct dm_target *ti, struct bio *bio) ...@@ -1741,7 +1741,7 @@ static int snapshot_map(struct dm_target *ti, struct bio *bio)
DMERR("Snapshot overflowed: Unable to allocate exception."); DMERR("Snapshot overflowed: Unable to allocate exception.");
} else } else
__invalidate_snapshot(s, -ENOMEM); __invalidate_snapshot(s, -ENOMEM);
r = -EIO; r = DM_MAPIO_KILL;
goto out_unlock; goto out_unlock;
} }
} }
......
...@@ -128,7 +128,7 @@ static void io_err_dtr(struct dm_target *tt) ...@@ -128,7 +128,7 @@ static void io_err_dtr(struct dm_target *tt)
static int io_err_map(struct dm_target *tt, struct bio *bio) static int io_err_map(struct dm_target *tt, struct bio *bio)
{ {
return -EIO; return DM_MAPIO_KILL;
} }
static int io_err_clone_and_map_rq(struct dm_target *ti, struct request *rq, static int io_err_clone_and_map_rq(struct dm_target *ti, struct request *rq,
......
...@@ -643,17 +643,17 @@ static int verity_map(struct dm_target *ti, struct bio *bio) ...@@ -643,17 +643,17 @@ static int verity_map(struct dm_target *ti, struct bio *bio)
if (((unsigned)bio->bi_iter.bi_sector | bio_sectors(bio)) & if (((unsigned)bio->bi_iter.bi_sector | bio_sectors(bio)) &
((1 << (v->data_dev_block_bits - SECTOR_SHIFT)) - 1)) { ((1 << (v->data_dev_block_bits - SECTOR_SHIFT)) - 1)) {
DMERR_LIMIT("unaligned io"); DMERR_LIMIT("unaligned io");
return -EIO; return DM_MAPIO_KILL;
} }
if (bio_end_sector(bio) >> if (bio_end_sector(bio) >>
(v->data_dev_block_bits - SECTOR_SHIFT) > v->data_blocks) { (v->data_dev_block_bits - SECTOR_SHIFT) > v->data_blocks) {
DMERR_LIMIT("io out of range"); DMERR_LIMIT("io out of range");
return -EIO; return DM_MAPIO_KILL;
} }
if (bio_data_dir(bio) == WRITE) if (bio_data_dir(bio) == WRITE)
return -EIO; return DM_MAPIO_KILL;
io = dm_per_bio_data(bio, ti->per_io_data_size); io = dm_per_bio_data(bio, ti->per_io_data_size);
io->v = v; io->v = v;
......
...@@ -39,7 +39,7 @@ static int zero_map(struct dm_target *ti, struct bio *bio) ...@@ -39,7 +39,7 @@ static int zero_map(struct dm_target *ti, struct bio *bio)
case REQ_OP_READ: case REQ_OP_READ:
if (bio->bi_opf & REQ_RAHEAD) { if (bio->bi_opf & REQ_RAHEAD) {
/* readahead of null bytes only wastes buffer cache */ /* readahead of null bytes only wastes buffer cache */
return -EIO; return DM_MAPIO_KILL;
} }
zero_fill_bio(bio); zero_fill_bio(bio);
break; break;
...@@ -47,7 +47,7 @@ static int zero_map(struct dm_target *ti, struct bio *bio) ...@@ -47,7 +47,7 @@ static int zero_map(struct dm_target *ti, struct bio *bio)
/* writes get silently dropped */ /* writes get silently dropped */
break; break;
default: default:
return -EIO; return DM_MAPIO_KILL;
} }
bio_endio(bio); bio_endio(bio);
......
...@@ -1084,18 +1084,24 @@ static void __map_bio(struct dm_target_io *tio) ...@@ -1084,18 +1084,24 @@ static void __map_bio(struct dm_target_io *tio)
r = ti->type->map(ti, clone); r = ti->type->map(ti, clone);
dm_offload_end(&o); dm_offload_end(&o);
if (r == DM_MAPIO_REMAPPED) { switch (r) {
case DM_MAPIO_SUBMITTED:
break;
case DM_MAPIO_REMAPPED:
/* the bio has been remapped so dispatch it */ /* the bio has been remapped so dispatch it */
trace_block_bio_remap(bdev_get_queue(clone->bi_bdev), clone, trace_block_bio_remap(bdev_get_queue(clone->bi_bdev), clone,
tio->io->bio->bi_bdev->bd_dev, sector); tio->io->bio->bi_bdev->bd_dev, sector);
generic_make_request(clone); generic_make_request(clone);
} else if (r < 0 || r == DM_MAPIO_REQUEUE) { break;
case DM_MAPIO_KILL:
r = -EIO;
/*FALLTHRU*/
case DM_MAPIO_REQUEUE:
/* error the io and bail out, or requeue it if needed */ /* error the io and bail out, or requeue it if needed */
dec_pending(tio->io, r); dec_pending(tio->io, r);
free_tio(tio); free_tio(tio);
} else if (r != DM_MAPIO_SUBMITTED) { break;
default:
DMWARN("unimplemented target map return value: %d", r); DMWARN("unimplemented target map return value: %d", r);
BUG(); BUG();
} }
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment