Loading fs/btrfs/compression.c +3 −0 Original line number Diff line number Diff line Loading @@ -839,6 +839,9 @@ static void free_workspace(int type, struct list_head *workspace) btrfs_compress_op[idx]->free_workspace(workspace); atomic_dec(alloc_workspace); wake: /* * Make sure counter is updated before we wake up waiters. */ smp_mb(); if (waitqueue_active(workspace_wait)) wake_up(workspace_wait); Loading fs/btrfs/delayed-inode.c +4 −0 Original line number Diff line number Diff line Loading @@ -463,6 +463,10 @@ static int __btrfs_add_delayed_deletion_item(struct btrfs_delayed_node *node, static void finish_one_item(struct btrfs_delayed_root *delayed_root) { int seq = atomic_inc_return(&delayed_root->items_seq); /* * atomic_dec_return implies a barrier for waitqueue_active */ if ((atomic_dec_return(&delayed_root->items) < BTRFS_DELAYED_BACKGROUND || seq % BTRFS_DELAYED_BATCH == 0) && waitqueue_active(&delayed_root->wait)) Loading fs/btrfs/dev-replace.c +1 −2 Original line number Diff line number Diff line Loading @@ -454,7 +454,6 @@ static void btrfs_rm_dev_replace_blocked(struct btrfs_fs_info *fs_info) static void btrfs_rm_dev_replace_unblocked(struct btrfs_fs_info *fs_info) { clear_bit(BTRFS_FS_STATE_DEV_REPLACING, &fs_info->fs_state); if (waitqueue_active(&fs_info->replace_wait)) wake_up(&fs_info->replace_wait); } Loading fs/btrfs/disk-io.c +3 −0 Original line number Diff line number Diff line Loading @@ -801,6 +801,9 @@ static void run_one_async_done(struct btrfs_work *work) limit = btrfs_async_submit_limit(fs_info); limit = limit * 2 / 3; /* * atomic_dec_return implies a barrier for waitqueue_active */ if (atomic_dec_return(&fs_info->nr_async_submits) < limit && waitqueue_active(&fs_info->async_submit_wait)) wake_up(&fs_info->async_submit_wait); Loading fs/btrfs/extent-tree.c +1 −2 Original line number Diff line number Diff line Loading @@ -10372,8 +10372,7 @@ void btrfs_end_write_no_snapshoting(struct btrfs_root *root) { percpu_counter_dec(&root->subv_writers->counter); /* * Make sure counter is updated before we wake up * waiters. * Make sure counter is updated before we wake up waiters. */ smp_mb(); if (waitqueue_active(&root->subv_writers->wait)) Loading Loading
fs/btrfs/compression.c +3 −0 Original line number Diff line number Diff line Loading @@ -839,6 +839,9 @@ static void free_workspace(int type, struct list_head *workspace) btrfs_compress_op[idx]->free_workspace(workspace); atomic_dec(alloc_workspace); wake: /* * Make sure counter is updated before we wake up waiters. */ smp_mb(); if (waitqueue_active(workspace_wait)) wake_up(workspace_wait); Loading
fs/btrfs/delayed-inode.c +4 −0 Original line number Diff line number Diff line Loading @@ -463,6 +463,10 @@ static int __btrfs_add_delayed_deletion_item(struct btrfs_delayed_node *node, static void finish_one_item(struct btrfs_delayed_root *delayed_root) { int seq = atomic_inc_return(&delayed_root->items_seq); /* * atomic_dec_return implies a barrier for waitqueue_active */ if ((atomic_dec_return(&delayed_root->items) < BTRFS_DELAYED_BACKGROUND || seq % BTRFS_DELAYED_BATCH == 0) && waitqueue_active(&delayed_root->wait)) Loading
fs/btrfs/dev-replace.c +1 −2 Original line number Diff line number Diff line Loading @@ -454,7 +454,6 @@ static void btrfs_rm_dev_replace_blocked(struct btrfs_fs_info *fs_info) static void btrfs_rm_dev_replace_unblocked(struct btrfs_fs_info *fs_info) { clear_bit(BTRFS_FS_STATE_DEV_REPLACING, &fs_info->fs_state); if (waitqueue_active(&fs_info->replace_wait)) wake_up(&fs_info->replace_wait); } Loading
fs/btrfs/disk-io.c +3 −0 Original line number Diff line number Diff line Loading @@ -801,6 +801,9 @@ static void run_one_async_done(struct btrfs_work *work) limit = btrfs_async_submit_limit(fs_info); limit = limit * 2 / 3; /* * atomic_dec_return implies a barrier for waitqueue_active */ if (atomic_dec_return(&fs_info->nr_async_submits) < limit && waitqueue_active(&fs_info->async_submit_wait)) wake_up(&fs_info->async_submit_wait); Loading
fs/btrfs/extent-tree.c +1 −2 Original line number Diff line number Diff line Loading @@ -10372,8 +10372,7 @@ void btrfs_end_write_no_snapshoting(struct btrfs_root *root) { percpu_counter_dec(&root->subv_writers->counter); /* * Make sure counter is updated before we wake up * waiters. * Make sure counter is updated before we wake up waiters. */ smp_mb(); if (waitqueue_active(&root->subv_writers->wait)) Loading