Commit 67da05b3 authored by Colin Ian King's avatar Colin Ian King Committed by David Sterba
Browse files

btrfs: fix spelling mistakes found using codespell



There quite a few spelling mistakes as found using codespell. Fix them.

Signed-off-by: default avatarColin Ian King <colin.i.king@gmail.com>
Reviewed-by: default avatarDavid Sterba <dsterba@suse.com>
Signed-off-by: default avatarDavid Sterba <dsterba@suse.com>
parent e2fd8306
Loading
Loading
Loading
Loading
+1 −1
Original line number Diff line number Diff line
@@ -283,7 +283,7 @@ void btrfs_submit_bio(struct btrfs_fs_info *fs_info, struct bio *bio, int mirror
 * RAID setup.  Here we only want to write the one bad copy, so we do the
 * mapping ourselves and submit the bio directly.
 *
 * The I/O is issued sychronously to block the repair read completion from
 * The I/O is issued synchronously to block the repair read completion from
 * freeing the bio.
 */
int btrfs_repair_io_failure(struct btrfs_fs_info *fs_info, u64 ino, u64 start,
+2 −2
Original line number Diff line number Diff line
@@ -796,7 +796,7 @@ void btrfs_submit_compressed_read(struct inode *inode, struct bio *bio,
			 * Save the initial offset of this chunk, as there
			 * is no direct correlation between compressed pages and
			 * the original file offset.  The field is only used for
			 * priting error messages.
			 * printing error messages.
			 */
			btrfs_bio(comp_bio)->file_offset = file_offset;

@@ -1642,7 +1642,7 @@ static void heuristic_collect_sample(struct inode *inode, u64 start, u64 end,
 *
 * For now is's a naive and optimistic 'return true', we'll extend the logic to
 * quickly (compared to direct compression) detect data characteristics
 * (compressible/uncompressible) to avoid wasting CPU time on uncompressible
 * (compressible/incompressible) to avoid wasting CPU time on incompressible
 * data.
 *
 * The following types of analysis can be performed:
+2 −2
Original line number Diff line number Diff line
@@ -767,8 +767,8 @@ static void run_one_async_free(struct btrfs_work *work)
/*
 * Submit bio to an async queue.
 *
 * Retrun:
 * - true if the work has been succesfuly submitted
 * Return:
 * - true if the work has been successfully submitted
 * - false in case of error
 */
bool btrfs_wq_submit_bio(struct btrfs_inode *inode, struct bio *bio, int mirror_num,
+1 −1
Original line number Diff line number Diff line
@@ -1625,7 +1625,7 @@ u64 count_range_bits(struct extent_io_tree *tree,
}

/*
 * Searche a range in the state tree for a given mask.  If 'filled' == 1, this
 * Search a range in the state tree for a given mask.  If 'filled' == 1, this
 * returns 1 only if every extent in the tree has the bits set.  Otherwise, 1
 * is returned if any bit in the range is found set.
 */
+5 −5
Original line number Diff line number Diff line
@@ -998,7 +998,7 @@ static int alloc_rbio_parity_pages(struct btrfs_raid_bio *rbio)
}

/*
 * Return the total numer of errors found in the vertical stripe of @sector_nr.
 * Return the total number of errors found in the vertical stripe of @sector_nr.
 *
 * @faila and @failb will also be updated to the first and second stripe
 * number of the errors.
@@ -1183,7 +1183,7 @@ static void bio_get_trace_info(struct btrfs_raid_bio *rbio, struct bio *bio,
	trace_info->stripe_nr = -1;
}

/* Generate PQ for one veritical stripe. */
/* Generate PQ for one vertical stripe. */
static void generate_pq_vertical(struct btrfs_raid_bio *rbio, int sectornr)
{
	void **pointers = rbio->finish_pointers;
@@ -1357,7 +1357,7 @@ static void set_rbio_range_error(struct btrfs_raid_bio *rbio, struct bio *bio)
}

/*
 * For subpage case, we can no longer set page Uptodate directly for
 * For subpage case, we can no longer set page Up-to-date directly for
 * stripe_pages[], thus we need to locate the sector.
 */
static struct sector_ptr *find_stripe_sector(struct btrfs_raid_bio *rbio,
@@ -1529,7 +1529,7 @@ static int rmw_assemble_read_bios(struct btrfs_raid_bio *rbio,
	/*
	 * Build a list of bios to read all sectors (including data and P/Q).
	 *
	 * This behaviro is to compensate the later csum verification and
	 * This behavior is to compensate the later csum verification and
	 * recovery.
	 */
	for (total_sector_nr = 0; total_sector_nr < rbio->nr_sectors;
@@ -1773,7 +1773,7 @@ static int recover_vertical(struct btrfs_raid_bio *rbio, int sector_nr,
	found_errors = get_rbio_veritical_errors(rbio, sector_nr, &faila,
						 &failb);
	/*
	 * No errors in the veritical stripe, skip it.  Can happen for recovery
	 * No errors in the vertical stripe, skip it.  Can happen for recovery
	 * which only part of a stripe failed csum check.
	 */
	if (!found_errors)
Loading