Commit 8aa11dc8 authored by Ionela Voinescu's avatar Ionela Voinescu Committed by Zeng Heng
Browse files

cppc_cpufreq: clean up cpu, cpu_num and cpunum variable use

mainline inclusion
from mainline-v5.11-rc1
commit 48ad8dc9
category: bugfix
bugzilla: https://gitee.com/openeuler/kernel/issues/I8TPP6

Reference: https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/commit/?id=48ad8dc94032ab43f0655190d9687f6d65b98f7f



--------------------------------

In order to maintain the typical naming convention in the cpufreq
framework:

 - replace the use of "cpu" variable name for cppc_cpudata pointers
   with "cpu_data"
 - replace variable names "cpu_num" and "cpunum" with "cpu"
 - make cpu variables unsigned int

Where pertinent, also move the initialisation of cpu_data variable to
its declaration and make consistent use of the local "cpu" variable.

Signed-off-by: default avatarIonela Voinescu <ionela.voinescu@arm.com>
Acked-by: default avatarViresh Kumar <viresh.kumar@linaro.org>
Signed-off-by: default avatarRafael J. Wysocki <rafael.j.wysocki@intel.com>
Signed-off-by: default avatarZeng Heng <zengheng4@huawei.com>
parent 733bb2cd
Loading
Loading
Loading
Loading
+70 −75
Original line number Diff line number Diff line
@@ -102,11 +102,11 @@ static u64 cppc_get_dmi_max_khz(void)
 * and extrapolate the rest
 * For perf/freq > Nominal, we use the ratio perf:freq at Nominal for conversion
 */
static unsigned int cppc_cpufreq_perf_to_khz(struct cppc_cpudata *cpu,
static unsigned int cppc_cpufreq_perf_to_khz(struct cppc_cpudata *cpu_data,
					unsigned int perf)
{
	static u64 max_khz;
	struct cppc_perf_caps *caps = &cpu->perf_caps;
	struct cppc_perf_caps *caps = &cpu_data->perf_caps;
	u64 mul, div;

	if (caps->lowest_freq && caps->nominal_freq) {
@@ -126,11 +126,11 @@ static unsigned int cppc_cpufreq_perf_to_khz(struct cppc_cpudata *cpu,
	return (u64)perf * mul / div;
}

static unsigned int cppc_cpufreq_khz_to_perf(struct cppc_cpudata *cpu,
static unsigned int cppc_cpufreq_khz_to_perf(struct cppc_cpudata *cpu_data,
					unsigned int freq)
{
	static u64 max_khz;
	struct cppc_perf_caps *caps = &cpu->perf_caps;
	struct cppc_perf_caps *caps = &cpu_data->perf_caps;
	u64  mul, div;

	if (caps->lowest_freq && caps->nominal_freq) {
@@ -155,29 +155,27 @@ static int cppc_cpufreq_set_target(struct cpufreq_policy *policy,
		unsigned int target_freq,
		unsigned int relation)
{
	struct cppc_cpudata *cpu;
	struct cppc_cpudata *cpu_data = all_cpu_data[policy->cpu];
	struct cpufreq_freqs freqs;
	u32 desired_perf;
	int ret = 0;

	cpu = all_cpu_data[policy->cpu];

	desired_perf = cppc_cpufreq_khz_to_perf(cpu, target_freq);
	desired_perf = cppc_cpufreq_khz_to_perf(cpu_data, target_freq);
	/* Return if it is exactly the same perf */
	if (desired_perf == cpu->perf_ctrls.desired_perf)
	if (desired_perf == cpu_data->perf_ctrls.desired_perf)
		return ret;

	cpu->perf_ctrls.desired_perf = desired_perf;
	cpu_data->perf_ctrls.desired_perf = desired_perf;
	freqs.old = policy->cur;
	freqs.new = target_freq;

	cpufreq_freq_transition_begin(policy, &freqs);
	ret = cppc_set_perf(cpu->cpu, &cpu->perf_ctrls);
	ret = cppc_set_perf(cpu_data->cpu, &cpu_data->perf_ctrls);
	cpufreq_freq_transition_end(policy, &freqs, ret != 0);

	if (ret)
		pr_debug("Failed to set target on CPU:%d. ret:%d\n",
				cpu->cpu, ret);
				cpu_data->cpu, ret);

	return ret;
}
@@ -190,16 +188,16 @@ static int cppc_verify_policy(struct cpufreq_policy_data *policy)

static void cppc_cpufreq_stop_cpu(struct cpufreq_policy *policy)
{
	int cpu_num = policy->cpu;
	struct cppc_cpudata *cpu = all_cpu_data[cpu_num];
	struct cppc_cpudata *cpu_data = all_cpu_data[policy->cpu];
	unsigned int cpu = policy->cpu;
	int ret;

	cpu->perf_ctrls.desired_perf = cpu->perf_caps.lowest_perf;
	cpu_data->perf_ctrls.desired_perf = cpu_data->perf_caps.lowest_perf;

	ret = cppc_set_perf(cpu_num, &cpu->perf_ctrls);
	ret = cppc_set_perf(cpu, &cpu_data->perf_ctrls);
	if (ret)
		pr_debug("Err setting perf value:%d on CPU:%d. ret:%d\n",
				cpu->perf_caps.lowest_perf, cpu_num, ret);
				cpu_data->perf_caps.lowest_perf, cpu, ret);
}

/*
@@ -211,7 +209,7 @@ static void cppc_cpufreq_stop_cpu(struct cpufreq_policy *policy)
#ifdef CONFIG_ARM64
#include <asm/cputype.h>

static unsigned int cppc_cpufreq_get_transition_delay_us(int cpu)
static unsigned int cppc_cpufreq_get_transition_delay_us(unsigned int cpu)
{
	unsigned long implementor = read_cpuid_implementor();
	unsigned long part_num = read_cpuid_part_number();
@@ -239,7 +237,7 @@ static unsigned int cppc_cpufreq_get_transition_delay_us(int cpu)

#else

static unsigned int cppc_cpufreq_get_transition_delay_us(int cpu)
static unsigned int cppc_cpufreq_get_transition_delay_us(unsigned int cpu)
{
	return cppc_get_transition_latency(cpu) / NSEC_PER_USEC;
}
@@ -247,54 +245,52 @@ static unsigned int cppc_cpufreq_get_transition_delay_us(int cpu)

static int cppc_cpufreq_cpu_init(struct cpufreq_policy *policy)
{
	struct cppc_cpudata *cpu;
	unsigned int cpu_num = policy->cpu;
	struct cppc_cpudata *cpu_data = all_cpu_data[policy->cpu];
	unsigned int cpu = policy->cpu;
	int ret = 0;

	cpu = all_cpu_data[policy->cpu];

	cpu->cpu = cpu_num;
	ret = cppc_get_perf_caps(policy->cpu, &cpu->perf_caps);
	cpu_data->cpu = cpu;
	ret = cppc_get_perf_caps(policy->cpu, &cpu_data->perf_caps);

	if (ret) {
		pr_debug("Err reading CPU%d perf capabilities. ret:%d\n",
				cpu_num, ret);
				cpu, ret);
		return ret;
	}

	/* Convert the lowest and nominal freq from MHz to KHz */
	cpu->perf_caps.lowest_freq *= 1000;
	cpu->perf_caps.nominal_freq *= 1000;
	cpu_data->perf_caps.lowest_freq *= 1000;
	cpu_data->perf_caps.nominal_freq *= 1000;

	/*
	 * Set min to lowest nonlinear perf to avoid any efficiency penalty (see
	 * Section 8.4.7.1.1.5 of ACPI 6.1 spec)
	 */
	policy->min = cppc_cpufreq_perf_to_khz(cpu, cpu->perf_caps.lowest_nonlinear_perf);
	policy->max = cppc_cpufreq_perf_to_khz(cpu, cpu->perf_caps.nominal_perf);
	policy->min = cppc_cpufreq_perf_to_khz(cpu_data, cpu_data->perf_caps.lowest_nonlinear_perf);
	policy->max = cppc_cpufreq_perf_to_khz(cpu_data, cpu_data->perf_caps.nominal_perf);

	/*
	 * Set cpuinfo.min_freq to Lowest to make the full range of performance
	 * available if userspace wants to use any perf between lowest & lowest
	 * nonlinear perf
	 */
	policy->cpuinfo.min_freq = cppc_cpufreq_perf_to_khz(cpu, cpu->perf_caps.lowest_perf);
	policy->cpuinfo.max_freq = cppc_cpufreq_perf_to_khz(cpu, cpu->perf_caps.nominal_perf);
	policy->cpuinfo.min_freq = cppc_cpufreq_perf_to_khz(cpu_data, cpu_data->perf_caps.lowest_perf);
	policy->cpuinfo.max_freq = cppc_cpufreq_perf_to_khz(cpu_data, cpu_data->perf_caps.nominal_perf);

	policy->transition_delay_us = cppc_cpufreq_get_transition_delay_us(cpu_num);
	policy->shared_type = cpu->shared_type;
	policy->transition_delay_us = cppc_cpufreq_get_transition_delay_us(cpu);
	policy->shared_type = cpu_data->shared_type;

	if (policy->shared_type == CPUFREQ_SHARED_TYPE_ANY) {
		int i;

		cpumask_copy(policy->cpus, cpu->shared_cpu_map);
		cpumask_copy(policy->cpus, cpu_data->shared_cpu_map);

		for_each_cpu(i, policy->cpus) {
			if (unlikely(i == policy->cpu))
			if (unlikely(i == cpu))
				continue;

			memcpy(&all_cpu_data[i]->perf_caps, &cpu->perf_caps,
			       sizeof(cpu->perf_caps));
			memcpy(&all_cpu_data[i]->perf_caps, &cpu_data->perf_caps,
			       sizeof(cpu_data->perf_caps));
		}
	} else if (policy->shared_type == CPUFREQ_SHARED_TYPE_ALL) {
		/* Support only SW_ANY for now. */
@@ -302,24 +298,24 @@ static int cppc_cpufreq_cpu_init(struct cpufreq_policy *policy)
		return -EFAULT;
	}

	cpu->cur_policy = policy;
	cpu_data->cur_policy = policy;

	/*
	 * If 'highest_perf' is greater than 'nominal_perf', we assume CPU Boost
	 * is supported.
	 */
	if (cpu->perf_caps.highest_perf > cpu->perf_caps.nominal_perf)
	if (cpu_data->perf_caps.highest_perf > cpu_data->perf_caps.nominal_perf)
		boost_supported = true;

	/* Set policy->cur to max now. The governors will adjust later. */
	policy->cur = cppc_cpufreq_perf_to_khz(cpu,
					cpu->perf_caps.highest_perf);
	cpu->perf_ctrls.desired_perf = cpu->perf_caps.highest_perf;
	policy->cur = cppc_cpufreq_perf_to_khz(cpu_data,
					cpu_data->perf_caps.highest_perf);
	cpu_data->perf_ctrls.desired_perf = cpu_data->perf_caps.highest_perf;

	ret = cppc_set_perf(cpu_num, &cpu->perf_ctrls);
	ret = cppc_set_perf(cpu, &cpu_data->perf_ctrls);
	if (ret)
		pr_debug("Err setting perf value:%d on CPU:%d. ret:%d\n",
				cpu->perf_caps.highest_perf, cpu_num, ret);
				cpu_data->perf_caps.highest_perf, cpu, ret);

	return ret;
}
@@ -332,7 +328,7 @@ static inline u64 get_delta(u64 t1, u64 t0)
	return (u32)t1 - (u32)t0;
}

static int cppc_get_rate_from_fbctrs(struct cppc_cpudata *cpu,
static int cppc_get_rate_from_fbctrs(struct cppc_cpudata *cpu_data,
				     struct cppc_perf_fb_ctrs fb_ctrs_t0,
				     struct cppc_perf_fb_ctrs fb_ctrs_t1)
{
@@ -351,9 +347,9 @@ static int cppc_get_rate_from_fbctrs(struct cppc_cpudata *cpu,
		delivered_perf = (reference_perf * delta_delivered) /
					delta_reference;
	else
		delivered_perf = cpu->perf_ctrls.desired_perf;
		delivered_perf = cpu_data->perf_ctrls.desired_perf;

	return cppc_cpufreq_perf_to_khz(cpu, delivered_perf);
	return cppc_cpufreq_perf_to_khz(cpu_data, delivered_perf);
}

static int cppc_get_perf_ctrs_sample(void *val)
@@ -371,14 +367,14 @@ static int cppc_get_perf_ctrs_sample(void *val)
	return cppc_get_perf_ctrs(cpu, &fb_ctrs->fb_ctrs_t1);
}

static unsigned int cppc_cpufreq_get_rate(unsigned int cpunum)
static unsigned int cppc_cpufreq_get_rate(unsigned int cpu)
{
	struct fb_ctr_pair fb_ctrs = { .cpu = cpunum, };
	struct cppc_cpudata *cpu = all_cpu_data[cpunum];
	struct fb_ctr_pair fb_ctrs = { .cpu = cpu, };
	struct cppc_cpudata *cpu_data = all_cpu_data[cpu];
	int ret;

	if (cpu_has_amu_feat(cpunum))
		ret = smp_call_on_cpu(cpunum, cppc_get_perf_ctrs_sample,
	if (cpu_has_amu_feat(cpu))
		ret = smp_call_on_cpu(cpu, cppc_get_perf_ctrs_sample,
				      &fb_ctrs, false);
	else
		ret = cppc_get_perf_ctrs_sample(&fb_ctrs);
@@ -386,14 +382,14 @@ static unsigned int cppc_cpufreq_get_rate(unsigned int cpunum)
	if (ret)
		return ret;

	return cppc_get_rate_from_fbctrs(cpu,
	return cppc_get_rate_from_fbctrs(cpu_data,
					 fb_ctrs.fb_ctrs_t0,
					 fb_ctrs.fb_ctrs_t1);
}

static int cppc_cpufreq_set_boost(struct cpufreq_policy *policy, int state)
{
	struct cppc_cpudata *cpudata;
	struct cppc_cpudata *cpu_data = all_cpu_data[policy->cpu];
	int ret;

	if (!boost_supported) {
@@ -401,13 +397,12 @@ static int cppc_cpufreq_set_boost(struct cpufreq_policy *policy, int state)
		return -EINVAL;
	}

	cpudata = all_cpu_data[policy->cpu];
	if (state)
		policy->max = cppc_cpufreq_perf_to_khz(cpudata,
					cpudata->perf_caps.highest_perf);
		policy->max = cppc_cpufreq_perf_to_khz(cpu_data,
					cpu_data->perf_caps.highest_perf);
	else
		policy->max = cppc_cpufreq_perf_to_khz(cpudata,
					cpudata->perf_caps.nominal_perf);
		policy->max = cppc_cpufreq_perf_to_khz(cpu_data,
					cpu_data->perf_caps.nominal_perf);
	policy->cpuinfo.max_freq = policy->max;

	ret = freq_qos_update_request(policy->max_freq_req, policy->max);
@@ -434,17 +429,17 @@ static struct cpufreq_driver cppc_cpufreq_driver = {
 * platform specific mechanism. We reuse the desired performance register to
 * store the real performance calculated by the platform.
 */
static unsigned int hisi_cppc_cpufreq_get_rate(unsigned int cpunum)
static unsigned int hisi_cppc_cpufreq_get_rate(unsigned int cpu)
{
	struct cppc_cpudata *cpudata = all_cpu_data[cpunum];
	struct cppc_cpudata *cpu_data = all_cpu_data[cpu];
	u64 desired_perf;
	int ret;

	ret = cppc_get_desired_perf(cpunum, &desired_perf);
	ret = cppc_get_desired_perf(cpu, &desired_perf);
	if (ret < 0)
		return -EIO;

	return cppc_cpufreq_perf_to_khz(cpudata, desired_perf);
	return cppc_cpufreq_perf_to_khz(cpu_data, desired_perf);
}

static void cppc_check_hisi_workaround(void)
@@ -473,7 +468,7 @@ static void cppc_check_hisi_workaround(void)
static int __init cppc_cpufreq_init(void)
{
	int i, ret = 0;
	struct cppc_cpudata *cpu;
	struct cppc_cpudata *cpu_data;

	if (acpi_disabled)
		return -ENODEV;
@@ -488,8 +483,8 @@ static int __init cppc_cpufreq_init(void)
		if (!all_cpu_data[i])
			goto out;

		cpu = all_cpu_data[i];
		if (!zalloc_cpumask_var(&cpu->shared_cpu_map, GFP_KERNEL))
		cpu_data = all_cpu_data[i];
		if (!zalloc_cpumask_var(&cpu_data->shared_cpu_map, GFP_KERNEL))
			goto out;
	}

@@ -509,11 +504,11 @@ static int __init cppc_cpufreq_init(void)

out:
	for_each_possible_cpu(i) {
		cpu = all_cpu_data[i];
		if (!cpu)
		cpu_data = all_cpu_data[i];
		if (!cpu_data)
			break;
		free_cpumask_var(cpu->shared_cpu_map);
		kfree(cpu);
		free_cpumask_var(cpu_data->shared_cpu_map);
		kfree(cpu_data);
	}

	kfree(all_cpu_data);
@@ -522,15 +517,15 @@ static int __init cppc_cpufreq_init(void)

static void __exit cppc_cpufreq_exit(void)
{
	struct cppc_cpudata *cpu;
	struct cppc_cpudata *cpu_data;
	int i;

	cpufreq_unregister_driver(&cppc_cpufreq_driver);

	for_each_possible_cpu(i) {
		cpu = all_cpu_data[i];
		free_cpumask_var(cpu->shared_cpu_map);
		kfree(cpu);
		cpu_data = all_cpu_data[i];
		free_cpumask_var(cpu_data->shared_cpu_map);
		kfree(cpu_data);
	}

	kfree(all_cpu_data);