Loading Makefile +1 −1 Original line number Diff line number Diff line # SPDX-License-Identifier: GPL-2.0 VERSION = 4 PATCHLEVEL = 19 SUBLEVEL = 63 SUBLEVEL = 64 EXTRAVERSION = NAME = "People's Front" Loading arch/arm64/include/asm/compat.h +1 −0 Original line number Diff line number Diff line Loading @@ -159,6 +159,7 @@ static inline compat_uptr_t ptr_to_compat(void __user *uptr) } #define compat_user_stack_pointer() (user_stack_pointer(task_pt_regs(current))) #define COMPAT_MINSIGSTKSZ 2048 static inline void __user *arch_compat_alloc_user_space(long len) { Loading arch/sh/boards/Kconfig +3 −11 Original line number Diff line number Diff line Loading @@ -8,27 +8,19 @@ config SH_ALPHA_BOARD bool config SH_DEVICE_TREE bool "Board Described by Device Tree" bool select OF select OF_EARLY_FLATTREE select TIMER_OF select COMMON_CLK select GENERIC_CALIBRATE_DELAY help Select Board Described by Device Tree to build a kernel that does not hard-code any board-specific knowledge but instead uses a device tree blob provided by the boot-loader. You must enable drivers for any hardware you want to use separately. At this time, only boards based on the open-hardware J-Core processors have sufficient driver coverage to use this option; do not select it if you are using original SuperH hardware. config SH_JCORE_SOC bool "J-Core SoC" depends on SH_DEVICE_TREE && (CPU_SH2 || CPU_J2) select SH_DEVICE_TREE select CLKSRC_JCORE_PIT select JCORE_AIC default y if CPU_J2 depends on CPU_J2 help Select this option to include drivers core components of the J-Core SoC, including interrupt controllers and timers. Loading block/blk-core.c +18 −17 Original line number Diff line number Diff line Loading @@ -421,24 +421,25 @@ void blk_sync_queue(struct request_queue *q) EXPORT_SYMBOL(blk_sync_queue); /** * blk_set_preempt_only - set QUEUE_FLAG_PREEMPT_ONLY * blk_set_pm_only - increment pm_only counter * @q: request queue pointer * * Returns the previous value of the PREEMPT_ONLY flag - 0 if the flag was not * set and 1 if the flag was already set. */ int blk_set_preempt_only(struct request_queue *q) void blk_set_pm_only(struct request_queue *q) { return blk_queue_flag_test_and_set(QUEUE_FLAG_PREEMPT_ONLY, q); atomic_inc(&q->pm_only); } EXPORT_SYMBOL_GPL(blk_set_preempt_only); EXPORT_SYMBOL_GPL(blk_set_pm_only); void blk_clear_preempt_only(struct request_queue *q) void blk_clear_pm_only(struct request_queue *q) { blk_queue_flag_clear(QUEUE_FLAG_PREEMPT_ONLY, q); int pm_only; pm_only = atomic_dec_return(&q->pm_only); WARN_ON_ONCE(pm_only < 0); if (pm_only == 0) wake_up_all(&q->mq_freeze_wq); } EXPORT_SYMBOL_GPL(blk_clear_preempt_only); EXPORT_SYMBOL_GPL(blk_clear_pm_only); /** * __blk_run_queue_uncond - run a queue whether or not it has been stopped Loading Loading @@ -916,7 +917,7 @@ EXPORT_SYMBOL(blk_alloc_queue); */ int blk_queue_enter(struct request_queue *q, blk_mq_req_flags_t flags) { const bool preempt = flags & BLK_MQ_REQ_PREEMPT; const bool pm = flags & BLK_MQ_REQ_PREEMPT; while (true) { bool success = false; Loading @@ -924,11 +925,11 @@ int blk_queue_enter(struct request_queue *q, blk_mq_req_flags_t flags) rcu_read_lock(); if (percpu_ref_tryget_live(&q->q_usage_counter)) { /* * The code that sets the PREEMPT_ONLY flag is * responsible for ensuring that that flag is globally * visible before the queue is unfrozen. * The code that increments the pm_only counter is * responsible for ensuring that that counter is * globally visible before the queue is unfrozen. */ if (preempt || !blk_queue_preempt_only(q)) { if (pm || !blk_queue_pm_only(q)) { success = true; } else { percpu_ref_put(&q->q_usage_counter); Loading @@ -953,7 +954,7 @@ int blk_queue_enter(struct request_queue *q, blk_mq_req_flags_t flags) wait_event(q->mq_freeze_wq, (atomic_read(&q->mq_freeze_depth) == 0 && (preempt || !blk_queue_preempt_only(q))) || (pm || !blk_queue_pm_only(q))) || blk_queue_dying(q)); if (blk_queue_dying(q)) return -ENODEV; Loading block/blk-mq-debugfs.c +9 −1 Original line number Diff line number Diff line Loading @@ -102,6 +102,14 @@ static int blk_flags_show(struct seq_file *m, const unsigned long flags, return 0; } static int queue_pm_only_show(void *data, struct seq_file *m) { struct request_queue *q = data; seq_printf(m, "%d\n", atomic_read(&q->pm_only)); return 0; } #define QUEUE_FLAG_NAME(name) [QUEUE_FLAG_##name] = #name static const char *const blk_queue_flag_name[] = { QUEUE_FLAG_NAME(QUEUED), Loading Loading @@ -132,7 +140,6 @@ static const char *const blk_queue_flag_name[] = { QUEUE_FLAG_NAME(REGISTERED), QUEUE_FLAG_NAME(SCSI_PASSTHROUGH), QUEUE_FLAG_NAME(QUIESCED), QUEUE_FLAG_NAME(PREEMPT_ONLY), }; #undef QUEUE_FLAG_NAME Loading Loading @@ -209,6 +216,7 @@ static ssize_t queue_write_hint_store(void *data, const char __user *buf, static const struct blk_mq_debugfs_attr blk_mq_debugfs_queue_attrs[] = { { "poll_stat", 0400, queue_poll_stat_show }, { "requeue_list", 0400, .seq_ops = &queue_requeue_list_seq_ops }, { "pm_only", 0600, queue_pm_only_show, NULL }, { "state", 0600, queue_state_show, queue_state_write }, { "write_hints", 0600, queue_write_hint_show, queue_write_hint_store }, { "zone_wlock", 0400, queue_zone_wlock_show, NULL }, Loading Loading
Makefile +1 −1 Original line number Diff line number Diff line # SPDX-License-Identifier: GPL-2.0 VERSION = 4 PATCHLEVEL = 19 SUBLEVEL = 63 SUBLEVEL = 64 EXTRAVERSION = NAME = "People's Front" Loading
arch/arm64/include/asm/compat.h +1 −0 Original line number Diff line number Diff line Loading @@ -159,6 +159,7 @@ static inline compat_uptr_t ptr_to_compat(void __user *uptr) } #define compat_user_stack_pointer() (user_stack_pointer(task_pt_regs(current))) #define COMPAT_MINSIGSTKSZ 2048 static inline void __user *arch_compat_alloc_user_space(long len) { Loading
arch/sh/boards/Kconfig +3 −11 Original line number Diff line number Diff line Loading @@ -8,27 +8,19 @@ config SH_ALPHA_BOARD bool config SH_DEVICE_TREE bool "Board Described by Device Tree" bool select OF select OF_EARLY_FLATTREE select TIMER_OF select COMMON_CLK select GENERIC_CALIBRATE_DELAY help Select Board Described by Device Tree to build a kernel that does not hard-code any board-specific knowledge but instead uses a device tree blob provided by the boot-loader. You must enable drivers for any hardware you want to use separately. At this time, only boards based on the open-hardware J-Core processors have sufficient driver coverage to use this option; do not select it if you are using original SuperH hardware. config SH_JCORE_SOC bool "J-Core SoC" depends on SH_DEVICE_TREE && (CPU_SH2 || CPU_J2) select SH_DEVICE_TREE select CLKSRC_JCORE_PIT select JCORE_AIC default y if CPU_J2 depends on CPU_J2 help Select this option to include drivers core components of the J-Core SoC, including interrupt controllers and timers. Loading
block/blk-core.c +18 −17 Original line number Diff line number Diff line Loading @@ -421,24 +421,25 @@ void blk_sync_queue(struct request_queue *q) EXPORT_SYMBOL(blk_sync_queue); /** * blk_set_preempt_only - set QUEUE_FLAG_PREEMPT_ONLY * blk_set_pm_only - increment pm_only counter * @q: request queue pointer * * Returns the previous value of the PREEMPT_ONLY flag - 0 if the flag was not * set and 1 if the flag was already set. */ int blk_set_preempt_only(struct request_queue *q) void blk_set_pm_only(struct request_queue *q) { return blk_queue_flag_test_and_set(QUEUE_FLAG_PREEMPT_ONLY, q); atomic_inc(&q->pm_only); } EXPORT_SYMBOL_GPL(blk_set_preempt_only); EXPORT_SYMBOL_GPL(blk_set_pm_only); void blk_clear_preempt_only(struct request_queue *q) void blk_clear_pm_only(struct request_queue *q) { blk_queue_flag_clear(QUEUE_FLAG_PREEMPT_ONLY, q); int pm_only; pm_only = atomic_dec_return(&q->pm_only); WARN_ON_ONCE(pm_only < 0); if (pm_only == 0) wake_up_all(&q->mq_freeze_wq); } EXPORT_SYMBOL_GPL(blk_clear_preempt_only); EXPORT_SYMBOL_GPL(blk_clear_pm_only); /** * __blk_run_queue_uncond - run a queue whether or not it has been stopped Loading Loading @@ -916,7 +917,7 @@ EXPORT_SYMBOL(blk_alloc_queue); */ int blk_queue_enter(struct request_queue *q, blk_mq_req_flags_t flags) { const bool preempt = flags & BLK_MQ_REQ_PREEMPT; const bool pm = flags & BLK_MQ_REQ_PREEMPT; while (true) { bool success = false; Loading @@ -924,11 +925,11 @@ int blk_queue_enter(struct request_queue *q, blk_mq_req_flags_t flags) rcu_read_lock(); if (percpu_ref_tryget_live(&q->q_usage_counter)) { /* * The code that sets the PREEMPT_ONLY flag is * responsible for ensuring that that flag is globally * visible before the queue is unfrozen. * The code that increments the pm_only counter is * responsible for ensuring that that counter is * globally visible before the queue is unfrozen. */ if (preempt || !blk_queue_preempt_only(q)) { if (pm || !blk_queue_pm_only(q)) { success = true; } else { percpu_ref_put(&q->q_usage_counter); Loading @@ -953,7 +954,7 @@ int blk_queue_enter(struct request_queue *q, blk_mq_req_flags_t flags) wait_event(q->mq_freeze_wq, (atomic_read(&q->mq_freeze_depth) == 0 && (preempt || !blk_queue_preempt_only(q))) || (pm || !blk_queue_pm_only(q))) || blk_queue_dying(q)); if (blk_queue_dying(q)) return -ENODEV; Loading
block/blk-mq-debugfs.c +9 −1 Original line number Diff line number Diff line Loading @@ -102,6 +102,14 @@ static int blk_flags_show(struct seq_file *m, const unsigned long flags, return 0; } static int queue_pm_only_show(void *data, struct seq_file *m) { struct request_queue *q = data; seq_printf(m, "%d\n", atomic_read(&q->pm_only)); return 0; } #define QUEUE_FLAG_NAME(name) [QUEUE_FLAG_##name] = #name static const char *const blk_queue_flag_name[] = { QUEUE_FLAG_NAME(QUEUED), Loading Loading @@ -132,7 +140,6 @@ static const char *const blk_queue_flag_name[] = { QUEUE_FLAG_NAME(REGISTERED), QUEUE_FLAG_NAME(SCSI_PASSTHROUGH), QUEUE_FLAG_NAME(QUIESCED), QUEUE_FLAG_NAME(PREEMPT_ONLY), }; #undef QUEUE_FLAG_NAME Loading Loading @@ -209,6 +216,7 @@ static ssize_t queue_write_hint_store(void *data, const char __user *buf, static const struct blk_mq_debugfs_attr blk_mq_debugfs_queue_attrs[] = { { "poll_stat", 0400, queue_poll_stat_show }, { "requeue_list", 0400, .seq_ops = &queue_requeue_list_seq_ops }, { "pm_only", 0600, queue_pm_only_show, NULL }, { "state", 0600, queue_state_show, queue_state_write }, { "write_hints", 0600, queue_write_hint_show, queue_write_hint_store }, { "zone_wlock", 0400, queue_zone_wlock_show, NULL }, Loading