Commit 50e34d78 authored by Christoph Hellwig's avatar Christoph Hellwig Committed by Jens Axboe
Browse files

block: disable the elevator int del_gendisk



The elevator is only used for file system requests, which are stopped in
del_gendisk.  Move disabling the elevator and freeing the scheduler tags
to the end of del_gendisk instead of doing that work in disk_release and
blk_cleanup_queue to avoid a use after free on q->tag_set from
disk_release as the tag_set might not be alive at that point.

Move the blk_qos_exit call as well, as it just depends on the elevator
exit and would be the only reason to keep the not exactly cheap queue
freeze in disk_release.

Fixes: e155b0c2 ("blk-mq: Use shared tags for shared sbitmap support")
Reported-by: default avatar <syzbot+3e3f419f4a7816471838@syzkaller.appspotmail.com>
Signed-off-by: default avatarChristoph Hellwig <hch@lst.de>
Tested-by: default avatar <syzbot+3e3f419f4a7816471838@syzkaller.appspotmail.com>
Link: https://lore.kernel.org/r/20220614074827.458955-2-hch@lst.de


Signed-off-by: default avatarJens Axboe <axboe@kernel.dk>
parent b96f3cab
Loading
Loading
Loading
Loading
+0 −13
Original line number Diff line number Diff line
@@ -322,19 +322,6 @@ void blk_cleanup_queue(struct request_queue *q)
		blk_mq_exit_queue(q);
	}

	/*
	 * In theory, request pool of sched_tags belongs to request queue.
	 * However, the current implementation requires tag_set for freeing
	 * requests, so free the pool now.
	 *
	 * Queue has become frozen, there can't be any in-queue requests, so
	 * it is safe to free requests now.
	 */
	mutex_lock(&q->sysfs_lock);
	if (q->elevator)
		blk_mq_sched_free_rqs(q);
	mutex_unlock(&q->sysfs_lock);

	/* @q is and will stay empty, shutdown and put */
	blk_put_queue(q);
}
+11 −28
Original line number Diff line number Diff line
@@ -652,6 +652,17 @@ void del_gendisk(struct gendisk *disk)

	blk_sync_queue(q);
	blk_flush_integrity();
	blk_mq_cancel_work_sync(q);

	blk_mq_quiesce_queue(q);
	if (q->elevator) {
		mutex_lock(&q->sysfs_lock);
		elevator_exit(q);
		mutex_unlock(&q->sysfs_lock);
	}
	rq_qos_exit(q);
	blk_mq_unquiesce_queue(q);

	/*
	 * Allow using passthrough request again after the queue is torn down.
	 */
@@ -1120,31 +1131,6 @@ static const struct attribute_group *disk_attr_groups[] = {
	NULL
};

static void disk_release_mq(struct request_queue *q)
{
	blk_mq_cancel_work_sync(q);

	/*
	 * There can't be any non non-passthrough bios in flight here, but
	 * requests stay around longer, including passthrough ones so we
	 * still need to freeze the queue here.
	 */
	blk_mq_freeze_queue(q);

	/*
	 * Since the I/O scheduler exit code may access cgroup information,
	 * perform I/O scheduler exit before disassociating from the block
	 * cgroup controller.
	 */
	if (q->elevator) {
		mutex_lock(&q->sysfs_lock);
		elevator_exit(q);
		mutex_unlock(&q->sysfs_lock);
	}
	rq_qos_exit(q);
	__blk_mq_unfreeze_queue(q, true);
}

/**
 * disk_release - releases all allocated resources of the gendisk
 * @dev: the device representing this disk
@@ -1166,9 +1152,6 @@ static void disk_release(struct device *dev)
	might_sleep();
	WARN_ON_ONCE(disk_live(disk));

	if (queue_is_mq(disk->queue))
		disk_release_mq(disk->queue);

	blkcg_exit_queue(disk->queue);

	disk_release_events(disk);