Loading drivers/md/dm-rq.c +16 −0 Original line number Diff line number Diff line Loading @@ -12,6 +12,22 @@ #define DM_MSG_PREFIX "core-rq" /* * One of these is allocated per request. */ struct dm_rq_target_io { struct mapped_device *md; struct dm_target *ti; struct request *orig, *clone; struct kthread_work work; blk_status_t error; union map_info info; struct dm_stats_aux stats_aux; unsigned long duration_jiffies; unsigned n_sectors; unsigned completed; }; #define DM_MQ_NR_HW_QUEUES 1 #define DM_MQ_QUEUE_DEPTH 2048 static unsigned dm_mq_nr_hw_queues = DM_MQ_NR_HW_QUEUES; Loading drivers/md/dm-rq.h +0 −16 Original line number Diff line number Diff line Loading @@ -16,22 +16,6 @@ struct mapped_device; /* * One of these is allocated per request. */ struct dm_rq_target_io { struct mapped_device *md; struct dm_target *ti; struct request *orig, *clone; struct kthread_work work; blk_status_t error; union map_info info; struct dm_stats_aux stats_aux; unsigned long duration_jiffies; unsigned n_sectors; unsigned completed; }; /* * For request-based dm - the bio clones we allocate are embedded in these * structs. Loading drivers/md/dm.c +2 −20 Original line number Diff line number Diff line Loading @@ -158,9 +158,6 @@ struct table_device { struct dm_dev dm_dev; }; static struct kmem_cache *_rq_tio_cache; static struct kmem_cache *_rq_cache; /* * Bio-based DM's mempools' reserved IOs set by the user. */ Loading Loading @@ -222,20 +219,11 @@ static unsigned dm_get_numa_node(void) static int __init local_init(void) { int r = -ENOMEM; _rq_tio_cache = KMEM_CACHE(dm_rq_target_io, 0); if (!_rq_tio_cache) return r; _rq_cache = kmem_cache_create("dm_old_clone_request", sizeof(struct request), __alignof__(struct request), 0, NULL); if (!_rq_cache) goto out_free_rq_tio_cache; int r; r = dm_uevent_init(); if (r) goto out_free_rq_cache; return r; deferred_remove_workqueue = alloc_workqueue("kdmremove", WQ_UNBOUND, 1); if (!deferred_remove_workqueue) { Loading @@ -257,10 +245,6 @@ static int __init local_init(void) destroy_workqueue(deferred_remove_workqueue); out_uevent_exit: dm_uevent_exit(); out_free_rq_cache: kmem_cache_destroy(_rq_cache); out_free_rq_tio_cache: kmem_cache_destroy(_rq_tio_cache); return r; } Loading @@ -270,8 +254,6 @@ static void local_exit(void) flush_scheduled_work(); destroy_workqueue(deferred_remove_workqueue); kmem_cache_destroy(_rq_cache); kmem_cache_destroy(_rq_tio_cache); unregister_blkdev(_major, _name); dm_uevent_exit(); Loading Loading
drivers/md/dm-rq.c +16 −0 Original line number Diff line number Diff line Loading @@ -12,6 +12,22 @@ #define DM_MSG_PREFIX "core-rq" /* * One of these is allocated per request. */ struct dm_rq_target_io { struct mapped_device *md; struct dm_target *ti; struct request *orig, *clone; struct kthread_work work; blk_status_t error; union map_info info; struct dm_stats_aux stats_aux; unsigned long duration_jiffies; unsigned n_sectors; unsigned completed; }; #define DM_MQ_NR_HW_QUEUES 1 #define DM_MQ_QUEUE_DEPTH 2048 static unsigned dm_mq_nr_hw_queues = DM_MQ_NR_HW_QUEUES; Loading
drivers/md/dm-rq.h +0 −16 Original line number Diff line number Diff line Loading @@ -16,22 +16,6 @@ struct mapped_device; /* * One of these is allocated per request. */ struct dm_rq_target_io { struct mapped_device *md; struct dm_target *ti; struct request *orig, *clone; struct kthread_work work; blk_status_t error; union map_info info; struct dm_stats_aux stats_aux; unsigned long duration_jiffies; unsigned n_sectors; unsigned completed; }; /* * For request-based dm - the bio clones we allocate are embedded in these * structs. Loading
drivers/md/dm.c +2 −20 Original line number Diff line number Diff line Loading @@ -158,9 +158,6 @@ struct table_device { struct dm_dev dm_dev; }; static struct kmem_cache *_rq_tio_cache; static struct kmem_cache *_rq_cache; /* * Bio-based DM's mempools' reserved IOs set by the user. */ Loading Loading @@ -222,20 +219,11 @@ static unsigned dm_get_numa_node(void) static int __init local_init(void) { int r = -ENOMEM; _rq_tio_cache = KMEM_CACHE(dm_rq_target_io, 0); if (!_rq_tio_cache) return r; _rq_cache = kmem_cache_create("dm_old_clone_request", sizeof(struct request), __alignof__(struct request), 0, NULL); if (!_rq_cache) goto out_free_rq_tio_cache; int r; r = dm_uevent_init(); if (r) goto out_free_rq_cache; return r; deferred_remove_workqueue = alloc_workqueue("kdmremove", WQ_UNBOUND, 1); if (!deferred_remove_workqueue) { Loading @@ -257,10 +245,6 @@ static int __init local_init(void) destroy_workqueue(deferred_remove_workqueue); out_uevent_exit: dm_uevent_exit(); out_free_rq_cache: kmem_cache_destroy(_rq_cache); out_free_rq_tio_cache: kmem_cache_destroy(_rq_tio_cache); return r; } Loading @@ -270,8 +254,6 @@ static void local_exit(void) flush_scheduled_work(); destroy_workqueue(deferred_remove_workqueue); kmem_cache_destroy(_rq_cache); kmem_cache_destroy(_rq_tio_cache); unregister_blkdev(_major, _name); dm_uevent_exit(); Loading