diff options
author | Tony Lambiris | 2018-07-31 00:53:31 -0400 |
---|---|---|
committer | Tony Lambiris | 2018-07-31 00:53:31 -0400 |
commit | 862248f1da1ff1d6e6579d652d62293e346fe303 (patch) | |
tree | 5bb64cc1ce968546b693e970b6198a0a7c30a081 /intel-pstate-backport.patch | |
parent | b9060af46c53aa2296480b763d34e0d1eb171c9b (diff) | |
download | aur-862248f1da1ff1d6e6579d652d62293e346fe303.tar.gz |
Version bump, sync with upstream
Diffstat (limited to 'intel-pstate-backport.patch')
-rw-r--r-- | intel-pstate-backport.patch | 1117 |
1 files changed, 0 insertions, 1117 deletions
diff --git a/intel-pstate-backport.patch b/intel-pstate-backport.patch deleted file mode 100644 index cf1427867261..000000000000 --- a/intel-pstate-backport.patch +++ /dev/null @@ -1,1117 +0,0 @@ ---- drivers/cpufreq/intel_pstate.c.orig 2017-02-19 17:34:00.000000000 -0500 -+++ drivers/cpufreq/intel_pstate.c 2017-03-23 23:47:19.095511820 -0400 -@@ -19,7 +19,7 @@ - #include <linux/hrtimer.h> - #include <linux/tick.h> - #include <linux/slab.h> --#include <linux/sched.h> -+#include <linux/sched/cpufreq.h> - #include <linux/list.h> - #include <linux/cpu.h> - #include <linux/cpufreq.h> -@@ -39,11 +39,6 @@ - - #define INTEL_CPUFREQ_TRANSITION_LATENCY 20000 - --#define ATOM_RATIOS 0x66a --#define ATOM_VIDS 0x66b --#define ATOM_TURBO_RATIOS 0x66c --#define ATOM_TURBO_VIDS 0x66d -- - #ifdef CONFIG_ACPI - #include <acpi/processor.h> - #include <acpi/cppc_acpi.h> -@@ -89,6 +84,11 @@ - return div64_u64(x << EXT_FRAC_BITS, y); - } - -+static inline int32_t percent_ext_fp(int percent) -+{ -+ return div_ext_fp(percent, 100); -+} -+ - /** - * struct sample - Store performance sample - * @core_avg_perf: Ratio of APERF/MPERF which is the actual average -@@ -358,42 +358,24 @@ - static int hwp_active __read_mostly; - static bool per_cpu_limits __read_mostly; - -+static bool driver_registered __read_mostly; -+ - #ifdef CONFIG_ACPI - static bool acpi_ppc; - #endif - --static struct perf_limits performance_limits = { -- .no_turbo = 0, -- .turbo_disabled = 0, -- .max_perf_pct = 100, -- .max_perf = int_ext_tofp(1), -- .min_perf_pct = 100, -- .min_perf = int_ext_tofp(1), -- .max_policy_pct = 100, -- .max_sysfs_pct = 100, -- .min_policy_pct = 0, -- .min_sysfs_pct = 0, --}; -- --static struct perf_limits powersave_limits = { -- .no_turbo = 0, -- .turbo_disabled = 0, -- .max_perf_pct = 100, -- .max_perf = int_ext_tofp(1), -- .min_perf_pct = 0, -- .min_perf = 0, -- .max_policy_pct = 100, -- .max_sysfs_pct = 100, -- .min_policy_pct = 0, -- .min_sysfs_pct = 0, --}; -+static struct perf_limits global; - --#ifdef CONFIG_CPU_FREQ_DEFAULT_GOV_PERFORMANCE --static struct perf_limits *limits = &performance_limits; --#else --static struct perf_limits *limits = &powersave_limits; --#endif -+static void intel_pstate_init_limits(struct perf_limits *limits) -+{ -+ memset(limits, 0, sizeof(*limits)); -+ limits->max_perf_pct = 100; -+ limits->max_perf = int_ext_tofp(1); -+ limits->max_policy_pct = 100; -+ limits->max_sysfs_pct = 100; -+} - -+static DEFINE_MUTEX(intel_pstate_driver_lock); - static DEFINE_MUTEX(intel_pstate_limits_lock); - - #ifdef CONFIG_ACPI -@@ -515,7 +497,7 @@ - * correct max turbo frequency based on the turbo state. - * Also need to convert to MHz as _PSS freq is in MHz. - */ -- if (!limits->turbo_disabled) -+ if (!global.turbo_disabled) - cpu->acpi_perf_data.states[0].core_frequency = - policy->cpuinfo.max_freq / 1000; - cpu->valid_pss_table = true; -@@ -538,7 +520,6 @@ - - acpi_processor_unregister_performance(policy->cpu); - } -- - #else - static inline void intel_pstate_init_acpi_perf_limits(struct cpufreq_policy *policy) - { -@@ -635,7 +616,7 @@ - - cpu = all_cpu_data[0]; - rdmsrl(MSR_IA32_MISC_ENABLE, misc_en); -- limits->turbo_disabled = -+ global.turbo_disabled = - (misc_en & MSR_IA32_MISC_ENABLE_TURBO_DISABLE || - cpu->pstate.max_pstate == cpu->pstate.turbo_pstate); - } -@@ -859,12 +840,11 @@ - - static void intel_pstate_hwp_set(struct cpufreq_policy *policy) - { -- int min, hw_min, max, hw_max, cpu, range, adj_range; -- struct perf_limits *perf_limits = limits; -+ int min, hw_min, max, hw_max, cpu; -+ struct perf_limits *perf_limits = &global; - u64 value, cap; - - for_each_cpu(cpu, policy->cpus) { -- int max_perf_pct, min_perf_pct; - struct cpudata *cpu_data = all_cpu_data[cpu]; - s16 epp; - -@@ -873,26 +853,22 @@ - - rdmsrl_on_cpu(cpu, MSR_HWP_CAPABILITIES, &cap); - hw_min = HWP_LOWEST_PERF(cap); -- hw_max = HWP_HIGHEST_PERF(cap); -- range = hw_max - hw_min; -+ if (global.no_turbo) -+ hw_max = HWP_GUARANTEED_PERF(cap); -+ else -+ hw_max = HWP_HIGHEST_PERF(cap); - -- max_perf_pct = perf_limits->max_perf_pct; -- min_perf_pct = perf_limits->min_perf_pct; -+ max = fp_ext_toint(hw_max * perf_limits->max_perf); -+ if (cpu_data->policy == CPUFREQ_POLICY_PERFORMANCE) -+ min = max; -+ else -+ min = fp_ext_toint(hw_max * perf_limits->min_perf); - - rdmsrl_on_cpu(cpu, MSR_HWP_REQUEST, &value); -- adj_range = min_perf_pct * range / 100; -- min = hw_min + adj_range; -+ - value &= ~HWP_MIN_PERF(~0L); - value |= HWP_MIN_PERF(min); - -- adj_range = max_perf_pct * range / 100; -- max = hw_min + adj_range; -- if (limits->no_turbo) { -- hw_max = HWP_GUARANTEED_PERF(cap); -- if (hw_max < max) -- max = hw_max; -- } -- - value &= ~HWP_MAX_PERF(~0L); - value |= HWP_MAX_PERF(max); - -@@ -996,6 +972,7 @@ - static int pid_param_set(void *data, u64 val) - { - *(u32 *)data = val; -+ pid_params.sample_rate_ns = pid_params.sample_rate_ms * NSEC_PER_MSEC; - intel_pstate_reset_all_pid(); - return 0; - } -@@ -1007,35 +984,57 @@ - } - DEFINE_SIMPLE_ATTRIBUTE(fops_pid_param, pid_param_get, pid_param_set, "%llu\n"); - -+static struct dentry *debugfs_parent; -+ - struct pid_param { - char *name; - void *value; -+ struct dentry *dentry; - }; - - static struct pid_param pid_files[] = { -- {"sample_rate_ms", &pid_params.sample_rate_ms}, -- {"d_gain_pct", &pid_params.d_gain_pct}, -- {"i_gain_pct", &pid_params.i_gain_pct}, -- {"deadband", &pid_params.deadband}, -- {"setpoint", &pid_params.setpoint}, -- {"p_gain_pct", &pid_params.p_gain_pct}, -- {NULL, NULL} -+ {"sample_rate_ms", &pid_params.sample_rate_ms, }, -+ {"d_gain_pct", &pid_params.d_gain_pct, }, -+ {"i_gain_pct", &pid_params.i_gain_pct, }, -+ {"deadband", &pid_params.deadband, }, -+ {"setpoint", &pid_params.setpoint, }, -+ {"p_gain_pct", &pid_params.p_gain_pct, }, -+ {NULL, NULL, } - }; - --static void __init intel_pstate_debug_expose_params(void) -+static void intel_pstate_debug_expose_params(void) - { -- struct dentry *debugfs_parent; -- int i = 0; -+ int i; - - debugfs_parent = debugfs_create_dir("pstate_snb", NULL); - if (IS_ERR_OR_NULL(debugfs_parent)) - return; -- while (pid_files[i].name) { -- debugfs_create_file(pid_files[i].name, 0660, -- debugfs_parent, pid_files[i].value, -- &fops_pid_param); -- i++; -+ -+ for (i = 0; pid_files[i].name; i++) { -+ struct dentry *dentry; -+ -+ dentry = debugfs_create_file(pid_files[i].name, 0660, -+ debugfs_parent, pid_files[i].value, -+ &fops_pid_param); -+ if (!IS_ERR(dentry)) -+ pid_files[i].dentry = dentry; -+ } -+} -+ -+static void intel_pstate_debug_hide_params(void) -+{ -+ int i; -+ -+ if (IS_ERR_OR_NULL(debugfs_parent)) -+ return; -+ -+ for (i = 0; pid_files[i].name; i++) { -+ debugfs_remove(pid_files[i].dentry); -+ pid_files[i].dentry = NULL; - } -+ -+ debugfs_remove(debugfs_parent); -+ debugfs_parent = NULL; - } - - /************************** debugfs end ************************/ -@@ -1045,9 +1044,37 @@ - static ssize_t show_##file_name \ - (struct kobject *kobj, struct attribute *attr, char *buf) \ - { \ -- return sprintf(buf, "%u\n", limits->object); \ -+ return sprintf(buf, "%u\n", global.object); \ - } - -+static ssize_t intel_pstate_show_status(char *buf); -+static int intel_pstate_update_status(const char *buf, size_t size); -+ -+static ssize_t show_status(struct kobject *kobj, -+ struct attribute *attr, char *buf) -+{ -+ ssize_t ret; -+ -+ mutex_lock(&intel_pstate_driver_lock); -+ ret = intel_pstate_show_status(buf); -+ mutex_unlock(&intel_pstate_driver_lock); -+ -+ return ret; -+} -+ -+static ssize_t store_status(struct kobject *a, struct attribute *b, -+ const char *buf, size_t count) -+{ -+ char *p = memchr(buf, '\n', count); -+ int ret; -+ -+ mutex_lock(&intel_pstate_driver_lock); -+ ret = intel_pstate_update_status(buf, p ? p - buf : count); -+ mutex_unlock(&intel_pstate_driver_lock); -+ -+ return ret < 0 ? ret : count; -+} -+ - static ssize_t show_turbo_pct(struct kobject *kobj, - struct attribute *attr, char *buf) - { -@@ -1055,12 +1082,22 @@ - int total, no_turbo, turbo_pct; - uint32_t turbo_fp; - -+ mutex_lock(&intel_pstate_driver_lock); -+ -+ if (!driver_registered) { -+ mutex_unlock(&intel_pstate_driver_lock); -+ return -EAGAIN; -+ } -+ - cpu = all_cpu_data[0]; - - total = cpu->pstate.turbo_pstate - cpu->pstate.min_pstate + 1; - no_turbo = cpu->pstate.max_pstate - cpu->pstate.min_pstate + 1; - turbo_fp = div_fp(no_turbo, total); - turbo_pct = 100 - fp_toint(mul_fp(turbo_fp, int_tofp(100))); -+ -+ mutex_unlock(&intel_pstate_driver_lock); -+ - return sprintf(buf, "%u\n", turbo_pct); - } - -@@ -1070,8 +1107,18 @@ - struct cpudata *cpu; - int total; - -+ mutex_lock(&intel_pstate_driver_lock); -+ -+ if (!driver_registered) { -+ mutex_unlock(&intel_pstate_driver_lock); -+ return -EAGAIN; -+ } -+ - cpu = all_cpu_data[0]; - total = cpu->pstate.turbo_pstate - cpu->pstate.min_pstate + 1; -+ -+ mutex_unlock(&intel_pstate_driver_lock); -+ - return sprintf(buf, "%u\n", total); - } - -@@ -1080,11 +1127,20 @@ - { - ssize_t ret; - -+ mutex_lock(&intel_pstate_driver_lock); -+ -+ if (!driver_registered) { -+ mutex_unlock(&intel_pstate_driver_lock); -+ return -EAGAIN; -+ } -+ - update_turbo_state(); -- if (limits->turbo_disabled) -- ret = sprintf(buf, "%u\n", limits->turbo_disabled); -+ if (global.turbo_disabled) -+ ret = sprintf(buf, "%u\n", global.turbo_disabled); - else -- ret = sprintf(buf, "%u\n", limits->no_turbo); -+ ret = sprintf(buf, "%u\n", global.no_turbo); -+ -+ mutex_unlock(&intel_pstate_driver_lock); - - return ret; - } -@@ -1099,21 +1155,31 @@ - if (ret != 1) - return -EINVAL; - -+ mutex_lock(&intel_pstate_driver_lock); -+ -+ if (!driver_registered) { -+ mutex_unlock(&intel_pstate_driver_lock); -+ return -EAGAIN; -+ } -+ - mutex_lock(&intel_pstate_limits_lock); - - update_turbo_state(); -- if (limits->turbo_disabled) { -+ if (global.turbo_disabled) { - pr_warn("Turbo disabled by BIOS or unavailable on processor\n"); - mutex_unlock(&intel_pstate_limits_lock); -+ mutex_unlock(&intel_pstate_driver_lock); - return -EPERM; - } - -- limits->no_turbo = clamp_t(int, input, 0, 1); -+ global.no_turbo = clamp_t(int, input, 0, 1); - - mutex_unlock(&intel_pstate_limits_lock); - - intel_pstate_update_policies(); - -+ mutex_unlock(&intel_pstate_driver_lock); -+ - return count; - } - -@@ -1127,21 +1193,27 @@ - if (ret != 1) - return -EINVAL; - -+ mutex_lock(&intel_pstate_driver_lock); -+ -+ if (!driver_registered) { -+ mutex_unlock(&intel_pstate_driver_lock); -+ return -EAGAIN; -+ } -+ - mutex_lock(&intel_pstate_limits_lock); - -- limits->max_sysfs_pct = clamp_t(int, input, 0 , 100); -- limits->max_perf_pct = min(limits->max_policy_pct, -- limits->max_sysfs_pct); -- limits->max_perf_pct = max(limits->min_policy_pct, -- limits->max_perf_pct); -- limits->max_perf_pct = max(limits->min_perf_pct, -- limits->max_perf_pct); -- limits->max_perf = div_ext_fp(limits->max_perf_pct, 100); -+ global.max_sysfs_pct = clamp_t(int, input, 0 , 100); -+ global.max_perf_pct = min(global.max_policy_pct, global.max_sysfs_pct); -+ global.max_perf_pct = max(global.min_policy_pct, global.max_perf_pct); -+ global.max_perf_pct = max(global.min_perf_pct, global.max_perf_pct); -+ global.max_perf = percent_ext_fp(global.max_perf_pct); - - mutex_unlock(&intel_pstate_limits_lock); - - intel_pstate_update_policies(); - -+ mutex_unlock(&intel_pstate_driver_lock); -+ - return count; - } - -@@ -1155,27 +1227,34 @@ - if (ret != 1) - return -EINVAL; - -+ mutex_lock(&intel_pstate_driver_lock); -+ -+ if (!driver_registered) { -+ mutex_unlock(&intel_pstate_driver_lock); -+ return -EAGAIN; -+ } -+ - mutex_lock(&intel_pstate_limits_lock); - -- limits->min_sysfs_pct = clamp_t(int, input, 0 , 100); -- limits->min_perf_pct = max(limits->min_policy_pct, -- limits->min_sysfs_pct); -- limits->min_perf_pct = min(limits->max_policy_pct, -- limits->min_perf_pct); -- limits->min_perf_pct = min(limits->max_perf_pct, -- limits->min_perf_pct); -- limits->min_perf = div_ext_fp(limits->min_perf_pct, 100); -+ global.min_sysfs_pct = clamp_t(int, input, 0 , 100); -+ global.min_perf_pct = max(global.min_policy_pct, global.min_sysfs_pct); -+ global.min_perf_pct = min(global.max_policy_pct, global.min_perf_pct); -+ global.min_perf_pct = min(global.max_perf_pct, global.min_perf_pct); -+ global.min_perf = percent_ext_fp(global.min_perf_pct); - - mutex_unlock(&intel_pstate_limits_lock); - - intel_pstate_update_policies(); - -+ mutex_unlock(&intel_pstate_driver_lock); -+ - return count; - } - - show_one(max_perf_pct, max_perf_pct); - show_one(min_perf_pct, min_perf_pct); - -+define_one_global_rw(status); - define_one_global_rw(no_turbo); - define_one_global_rw(max_perf_pct); - define_one_global_rw(min_perf_pct); -@@ -1183,6 +1262,7 @@ - define_one_global_ro(num_pstates); - - static struct attribute *intel_pstate_attributes[] = { -+ &status.attr, - &no_turbo.attr, - &turbo_pct.attr, - &num_pstates.attr, -@@ -1258,7 +1338,7 @@ - { - u64 value; - -- rdmsrl(ATOM_RATIOS, value); -+ rdmsrl(MSR_ATOM_CORE_RATIOS, value); - return (value >> 8) & 0x7F; - } - -@@ -1266,7 +1346,7 @@ - { - u64 value; - -- rdmsrl(ATOM_RATIOS, value); -+ rdmsrl(MSR_ATOM_CORE_RATIOS, value); - return (value >> 16) & 0x7F; - } - -@@ -1274,7 +1354,7 @@ - { - u64 value; - -- rdmsrl(ATOM_TURBO_RATIOS, value); -+ rdmsrl(MSR_ATOM_CORE_TURBO_RATIOS, value); - return value & 0x7F; - } - -@@ -1285,7 +1365,7 @@ - u32 vid; - - val = (u64)pstate << 8; -- if (limits->no_turbo && !limits->turbo_disabled) -+ if (global.no_turbo && !global.turbo_disabled) - val |= (u64)1 << 32; - - vid_fp = cpudata->vid.min + mul_fp( -@@ -1336,7 +1416,7 @@ - { - u64 value; - -- rdmsrl(ATOM_VIDS, value); -+ rdmsrl(MSR_ATOM_CORE_VIDS, value); - cpudata->vid.min = int_tofp((value >> 8) & 0x7f); - cpudata->vid.max = int_tofp((value >> 16) & 0x7f); - cpudata->vid.ratio = div_fp( -@@ -1344,7 +1424,7 @@ - int_tofp(cpudata->pstate.max_pstate - - cpudata->pstate.min_pstate)); - -- rdmsrl(ATOM_TURBO_VIDS, value); -+ rdmsrl(MSR_ATOM_CORE_TURBO_VIDS, value); - cpudata->vid.turbo = value & 0x7f; - } - -@@ -1364,48 +1444,71 @@ - return (value >> 8) & 0xFF; - } - -+static int core_get_tdp_ratio(u64 plat_info) -+{ -+ /* Check how many TDP levels present */ -+ if (plat_info & 0x600000000) { -+ u64 tdp_ctrl; -+ u64 tdp_ratio; -+ int tdp_msr; -+ int err; -+ -+ /* Get the TDP level (0, 1, 2) to get ratios */ -+ err = rdmsrl_safe(MSR_CONFIG_TDP_CONTROL, &tdp_ctrl); -+ if (err) -+ return err; -+ -+ /* TDP MSR are continuous starting at 0x648 */ -+ tdp_msr = MSR_CONFIG_TDP_NOMINAL + (tdp_ctrl & 0x03); -+ err = rdmsrl_safe(tdp_msr, &tdp_ratio); -+ if (err) -+ return err; -+ -+ /* For level 1 and 2, bits[23:16] contain the ratio */ -+ if (tdp_ctrl & 0x03) -+ tdp_ratio >>= 16; -+ -+ tdp_ratio &= 0xff; /* ratios are only 8 bits long */ -+ pr_debug("tdp_ratio %x\n", (int)tdp_ratio); -+ -+ return (int)tdp_ratio; -+ } -+ -+ return -ENXIO; -+} -+ - static int core_get_max_pstate(void) - { - u64 tar; - u64 plat_info; - int max_pstate; -+ int tdp_ratio; - int err; - - rdmsrl(MSR_PLATFORM_INFO, plat_info); - max_pstate = (plat_info >> 8) & 0xFF; - -+ tdp_ratio = core_get_tdp_ratio(plat_info); -+ if (tdp_ratio <= 0) -+ return max_pstate; -+ -+ if (hwp_active) { -+ /* Turbo activation ratio is not used on HWP platforms */ -+ return tdp_ratio; -+ } -+ - err = rdmsrl_safe(MSR_TURBO_ACTIVATION_RATIO, &tar); - if (!err) { -+ int tar_levels; -+ - /* Do some sanity checking for safety */ -- if (plat_info & 0x600000000) { -- u64 tdp_ctrl; -- u64 tdp_ratio; -- int tdp_msr; -- -- err = rdmsrl_safe(MSR_CONFIG_TDP_CONTROL, &tdp_ctrl); -- if (err) -- goto skip_tar; -- -- tdp_msr = MSR_CONFIG_TDP_NOMINAL + (tdp_ctrl & 0x3); -- err = rdmsrl_safe(tdp_msr, &tdp_ratio); -- if (err) -- goto skip_tar; -- -- /* For level 1 and 2, bits[23:16] contain the ratio */ -- if (tdp_ctrl) -- tdp_ratio >>= 16; -- -- tdp_ratio &= 0xff; /* ratios are only 8 bits long */ -- if (tdp_ratio - 1 == tar) { -- max_pstate = tar; -- pr_debug("max_pstate=TAC %x\n", max_pstate); -- } else { -- goto skip_tar; -- } -+ tar_levels = tar & 0xff; -+ if (tdp_ratio - 1 == tar_levels) { -+ max_pstate = tar_levels; -+ pr_debug("max_pstate=TAC %x\n", max_pstate); - } - } - --skip_tar: - return max_pstate; - } - -@@ -1432,7 +1535,7 @@ - u64 val; - - val = (u64)pstate << 8; -- if (limits->no_turbo && !limits->turbo_disabled) -+ if (global.no_turbo && !global.turbo_disabled) - val |= (u64)1 << 32; - - return val; -@@ -1558,9 +1661,9 @@ - int max_perf = cpu->pstate.turbo_pstate; - int max_perf_adj; - int min_perf; -- struct perf_limits *perf_limits = limits; -+ struct perf_limits *perf_limits = &global; - -- if (limits->no_turbo || limits->turbo_disabled) -+ if (global.no_turbo || global.turbo_disabled) - max_perf = cpu->pstate.max_pstate; - - if (per_cpu_limits) -@@ -1695,7 +1798,7 @@ - - sample->busy_scaled = busy_frac * 100; - -- target = limits->no_turbo || limits->turbo_disabled ? -+ target = global.no_turbo || global.turbo_disabled ? - cpu->pstate.max_pstate : cpu->pstate.turbo_pstate; - target += target >> 2; - target = mul_fp(target, busy_frac); -@@ -1759,13 +1862,11 @@ - - intel_pstate_get_min_max(cpu, &min_perf, &max_perf); - pstate = clamp_t(int, pstate, min_perf, max_perf); -- trace_cpu_frequency(pstate * cpu->pstate.scaling, cpu->cpu); - return pstate; - } - - static void intel_pstate_update_pstate(struct cpudata *cpu, int pstate) - { -- pstate = intel_pstate_prepare_request(cpu, pstate); - if (pstate == cpu->pstate.current_pstate) - return; - -@@ -1785,6 +1886,8 @@ - - update_turbo_state(); - -+ target_pstate = intel_pstate_prepare_request(cpu, target_pstate); -+ trace_cpu_frequency(target_pstate * cpu->pstate.scaling, cpu->cpu); - intel_pstate_update_pstate(cpu, target_pstate); - - sample = &cpu->sample; -@@ -1952,53 +2055,37 @@ - synchronize_sched(); - } - --static void intel_pstate_set_performance_limits(struct perf_limits *limits) --{ -- limits->no_turbo = 0; -- limits->turbo_disabled = 0; -- limits->max_perf_pct = 100; -- limits->max_perf = int_ext_tofp(1); -- limits->min_perf_pct = 100; -- limits->min_perf = int_ext_tofp(1); -- limits->max_policy_pct = 100; -- limits->max_sysfs_pct = 100; -- limits->min_policy_pct = 0; -- limits->min_sysfs_pct = 0; --} -- - static void intel_pstate_update_perf_limits(struct cpufreq_policy *policy, - struct perf_limits *limits) - { -+ int32_t max_policy_perf, min_policy_perf; - -- limits->max_policy_pct = DIV_ROUND_UP(policy->max * 100, -- policy->cpuinfo.max_freq); -- limits->max_policy_pct = clamp_t(int, limits->max_policy_pct, 0, 100); -+ max_policy_perf = div_ext_fp(policy->max, policy->cpuinfo.max_freq); -+ max_policy_perf = clamp_t(int32_t, max_policy_perf, 0, int_ext_tofp(1)); - if (policy->max == policy->min) { -- limits->min_policy_pct = limits->max_policy_pct; -+ min_policy_perf = max_policy_perf; - } else { -- limits->min_policy_pct = DIV_ROUND_UP(policy->min * 100, -- policy->cpuinfo.max_freq); -- limits->min_policy_pct = clamp_t(int, limits->min_policy_pct, -- 0, 100); -- } -- -- /* Normalize user input to [min_policy_pct, max_policy_pct] */ -- limits->min_perf_pct = max(limits->min_policy_pct, -- limits->min_sysfs_pct); -- limits->min_perf_pct = min(limits->max_policy_pct, -- limits->min_perf_pct); -- limits->max_perf_pct = min(limits->max_policy_pct, -- limits->max_sysfs_pct); -- limits->max_perf_pct = max(limits->min_policy_pct, -- limits->max_perf_pct); -+ min_policy_perf = div_ext_fp(policy->min, -+ policy->cpuinfo.max_freq); -+ min_policy_perf = clamp_t(int32_t, min_policy_perf, -+ 0, max_policy_perf); -+ } -+ -+ /* Normalize user input to [min_perf, max_perf] */ -+ limits->min_perf = max(min_policy_perf, -+ percent_ext_fp(limits->min_sysfs_pct)); -+ limits->min_perf = min(limits->min_perf, max_policy_perf); -+ limits->max_perf = min(max_policy_perf, -+ percent_ext_fp(limits->max_sysfs_pct)); -+ limits->max_perf = max(min_policy_perf, limits->max_perf); - -- /* Make sure min_perf_pct <= max_perf_pct */ -- limits->min_perf_pct = min(limits->max_perf_pct, limits->min_perf_pct); -+ /* Make sure min_perf <= max_perf */ -+ limits->min_perf = min(limits->min_perf, limits->max_perf); - -- limits->min_perf = div_ext_fp(limits->min_perf_pct, 100); -- limits->max_perf = div_ext_fp(limits->max_perf_pct, 100); - limits->max_perf = round_up(limits->max_perf, EXT_FRAC_BITS); - limits->min_perf = round_up(limits->min_perf, EXT_FRAC_BITS); -+ limits->max_perf_pct = fp_ext_toint(limits->max_perf * 100); -+ limits->min_perf_pct = fp_ext_toint(limits->min_perf * 100); - - pr_debug("cpu:%d max_perf_pct:%d min_perf_pct:%d\n", policy->cpu, - limits->max_perf_pct, limits->min_perf_pct); -@@ -2007,7 +2094,7 @@ - static int intel_pstate_set_policy(struct cpufreq_policy *policy) - { - struct cpudata *cpu; -- struct perf_limits *perf_limits = NULL; -+ struct perf_limits *perf_limits = &global; - - if (!policy->cpuinfo.max_freq) - return -ENODEV; -@@ -2030,28 +2117,8 @@ - - mutex_lock(&intel_pstate_limits_lock); - -- if (policy->policy == CPUFREQ_POLICY_PERFORMANCE) { -- if (!perf_limits) { -- limits = &performance_limits; -- perf_limits = limits; -- } -- if (policy->max >= policy->cpuinfo.max_freq && -- !limits->no_turbo) { -- pr_debug("set performance\n"); -- intel_pstate_set_performance_limits(perf_limits); -- goto out; -- } -- } else { -- pr_debug("set powersave\n"); -- if (!perf_limits) { -- limits = &powersave_limits; -- perf_limits = limits; -- } -- -- } -- - intel_pstate_update_perf_limits(policy, perf_limits); -- out: -+ - if (cpu->policy == CPUFREQ_POLICY_PERFORMANCE) { - /* - * NOHZ_FULL CPUs need this as the governor callback may not -@@ -2072,6 +2139,13 @@ - - static int intel_pstate_verify_policy(struct cpufreq_policy *policy) - { -+ struct cpudata *cpu = all_cpu_data[policy->cpu]; -+ -+ update_turbo_state(); -+ policy->cpuinfo.max_freq = global.turbo_disabled || global.no_turbo ? -+ cpu->pstate.max_freq : -+ cpu->pstate.turbo_freq; -+ - cpufreq_verify_within_cpu_limits(policy); - - if (policy->policy != CPUFREQ_POLICY_POWERSAVE && -@@ -2083,9 +2157,9 @@ - unsigned int max_freq, min_freq; - - max_freq = policy->cpuinfo.max_freq * -- limits->max_sysfs_pct / 100; -+ global.max_sysfs_pct / 100; - min_freq = policy->cpuinfo.max_freq * -- limits->min_sysfs_pct / 100; -+ global.min_sysfs_pct / 100; - cpufreq_verify_within_limits(policy, min_freq, max_freq); - } - -@@ -2128,13 +2202,8 @@ - - cpu = all_cpu_data[policy->cpu]; - -- /* -- * We need sane value in the cpu->perf_limits, so inherit from global -- * perf_limits limits, which are seeded with values based on the -- * CONFIG_CPU_FREQ_DEFAULT_GOV_*, during boot up. -- */ - if (per_cpu_limits) -- memcpy(cpu->perf_limits, limits, sizeof(struct perf_limits)); -+ intel_pstate_init_limits(cpu->perf_limits); - - policy->min = cpu->pstate.min_pstate * cpu->pstate.scaling; - policy->max = cpu->pstate.turbo_pstate * cpu->pstate.scaling; -@@ -2142,7 +2211,7 @@ - /* cpuinfo and default policy values */ - policy->cpuinfo.min_freq = cpu->pstate.min_pstate * cpu->pstate.scaling; - update_turbo_state(); -- policy->cpuinfo.max_freq = limits->turbo_disabled ? -+ policy->cpuinfo.max_freq = global.turbo_disabled ? - cpu->pstate.max_pstate : cpu->pstate.turbo_pstate; - policy->cpuinfo.max_freq *= cpu->pstate.scaling; - -@@ -2162,7 +2231,7 @@ - return ret; - - policy->cpuinfo.transition_latency = CPUFREQ_ETERNAL; -- if (limits->min_perf_pct == 100 && limits->max_perf_pct == 100) -+ if (IS_ENABLED(CONFIG_CPU_FREQ_DEFAULT_GOV_PERFORMANCE)) - policy->policy = CPUFREQ_POLICY_PERFORMANCE; - else - policy->policy = CPUFREQ_POLICY_POWERSAVE; -@@ -2186,46 +2255,16 @@ - static int intel_cpufreq_verify_policy(struct cpufreq_policy *policy) - { - struct cpudata *cpu = all_cpu_data[policy->cpu]; -- struct perf_limits *perf_limits = limits; - - update_turbo_state(); -- policy->cpuinfo.max_freq = limits->turbo_disabled ? -+ policy->cpuinfo.max_freq = global.no_turbo || global.turbo_disabled ? - cpu->pstate.max_freq : cpu->pstate.turbo_freq; - - cpufreq_verify_within_cpu_limits(policy); - -- if (per_cpu_limits) -- perf_limits = cpu->perf_limits; -- -- mutex_lock(&intel_pstate_limits_lock); -- -- intel_pstate_update_perf_limits(policy, perf_limits); -- -- mutex_unlock(&intel_pstate_limits_lock); -- - return 0; - } - --static unsigned int intel_cpufreq_turbo_update(struct cpudata *cpu, -- struct cpufreq_policy *policy, -- unsigned int target_freq) --{ -- unsigned int max_freq; -- -- update_turbo_state(); -- -- max_freq = limits->no_turbo || limits->turbo_disabled ? -- cpu->pstate.max_freq : cpu->pstate.turbo_freq; -- policy->cpuinfo.max_freq = max_freq; -- if (policy->max > max_freq) -- policy->max = max_freq; -- -- if (target_freq > max_freq) -- target_freq = max_freq; -- -- return target_freq; --} -- - static int intel_cpufreq_target(struct cpufreq_policy *policy, - unsigned int target_freq, - unsigned int relation) -@@ -2234,8 +2273,10 @@ - struct cpufreq_freqs freqs; - int target_pstate; - -+ update_turbo_state(); -+ - freqs.old = policy->cur; -- freqs.new = intel_cpufreq_turbo_update(cpu, policy, target_freq); -+ freqs.new = target_freq; - - cpufreq_freq_transition_begin(policy, &freqs); - switch (relation) { -@@ -2255,6 +2296,7 @@ - wrmsrl_on_cpu(policy->cpu, MSR_IA32_PERF_CTL, - pstate_funcs.get_val(cpu, target_pstate)); - } -+ freqs.new = target_pstate * cpu->pstate.scaling; - cpufreq_freq_transition_end(policy, &freqs, false); - - return 0; -@@ -2266,10 +2308,12 @@ - struct cpudata *cpu = all_cpu_data[policy->cpu]; - int target_pstate; - -- target_freq = intel_cpufreq_turbo_update(cpu, policy, target_freq); -+ update_turbo_state(); -+ - target_pstate = DIV_ROUND_UP(target_freq, cpu->pstate.scaling); -+ target_pstate = intel_pstate_prepare_request(cpu, target_pstate); - intel_pstate_update_pstate(cpu, target_pstate); -- return target_freq; -+ return target_pstate * cpu->pstate.scaling; - } - - static int intel_cpufreq_cpu_init(struct cpufreq_policy *policy) -@@ -2299,6 +2343,113 @@ - - static struct cpufreq_driver *intel_pstate_driver = &intel_pstate; - -+static void intel_pstate_driver_cleanup(void) -+{ -+ unsigned int cpu; -+ -+ get_online_cpus(); -+ for_each_online_cpu(cpu) { -+ if (all_cpu_data[cpu]) { -+ if (intel_pstate_driver == &intel_pstate) -+ intel_pstate_clear_update_util_hook(cpu); -+ -+ kfree(all_cpu_data[cpu]); -+ all_cpu_data[cpu] = NULL; -+ } -+ } -+ put_online_cpus(); -+} -+ -+static int intel_pstate_register_driver(void) -+{ -+ int ret; -+ -+ intel_pstate_init_limits(&global); -+ -+ ret = cpufreq_register_driver(intel_pstate_driver); -+ if (ret) { -+ intel_pstate_driver_cleanup(); -+ return ret; -+ } -+ -+ mutex_lock(&intel_pstate_limits_lock); -+ driver_registered = true; -+ mutex_unlock(&intel_pstate_limits_lock); -+ -+ if (intel_pstate_driver == &intel_pstate && !hwp_active && -+ pstate_funcs.get_target_pstate != get_target_pstate_use_cpu_load) -+ intel_pstate_debug_expose_params(); -+ -+ return 0; -+} -+ -+static int intel_pstate_unregister_driver(void) -+{ -+ if (hwp_active) -+ return -EBUSY; -+ -+ if (intel_pstate_driver == &intel_pstate && !hwp_active && -+ pstate_funcs.get_target_pstate != get_target_pstate_use_cpu_load) -+ intel_pstate_debug_hide_params(); -+ -+ mutex_lock(&intel_pstate_limits_lock); -+ driver_registered = false; -+ mutex_unlock(&intel_pstate_limits_lock); -+ -+ cpufreq_unregister_driver(intel_pstate_driver); -+ intel_pstate_driver_cleanup(); -+ -+ return 0; -+} -+ -+static ssize_t intel_pstate_show_status(char *buf) -+{ -+ if (!driver_registered) -+ return sprintf(buf, "off\n"); -+ -+ return sprintf(buf, "%s\n", intel_pstate_driver == &intel_pstate ? -+ "active" : "passive"); -+} -+ -+static int intel_pstate_update_status(const char *buf, size_t size) -+{ -+ int ret; -+ -+ if (size == 3 && !strncmp(buf, "off", size)) -+ return driver_registered ? -+ intel_pstate_unregister_driver() : -EINVAL; -+ -+ if (size == 6 && !strncmp(buf, "active", size)) { -+ if (driver_registered) { -+ if (intel_pstate_driver == &intel_pstate) -+ return 0; -+ -+ ret = intel_pstate_unregister_driver(); -+ if (ret) -+ return ret; -+ } -+ -+ intel_pstate_driver = &intel_pstate; -+ return intel_pstate_register_driver(); -+ } -+ -+ if (size == 7 && !strncmp(buf, "passive", size)) { -+ if (driver_registered) { -+ if (intel_pstate_driver != &intel_pstate) -+ return 0; -+ -+ ret = intel_pstate_unregister_driver(); -+ if (ret) -+ return ret; -+ } -+ -+ intel_pstate_driver = &intel_cpufreq; -+ return intel_pstate_register_driver(); -+ } -+ -+ return -EINVAL; -+} -+ - static int no_load __initdata; - static int no_hwp __initdata; - static int hwp_only __initdata; -@@ -2486,9 +2637,9 @@ - - static int __init intel_pstate_init(void) - { -- int cpu, rc = 0; - const struct x86_cpu_id *id; - struct cpu_defaults *cpu_def; -+ int rc = 0; - - if (no_load) - return -ENODEV; -@@ -2520,45 +2671,29 @@ - if (intel_pstate_platform_pwr_mgmt_exists()) - return -ENODEV; - -+ if (!hwp_active && hwp_only) -+ return -ENOTSUPP; -+ - pr_info("Intel P-state driver initializing\n"); - - all_cpu_data = vzalloc(sizeof(void *) * num_possible_cpus()); - if (!all_cpu_data) - return -ENOMEM; - -- if (!hwp_active && hwp_only) -- goto out; -- - intel_pstate_request_control_from_smm(); - -- rc = cpufreq_register_driver(intel_pstate_driver); -- if (rc) -- goto out; -- -- if (intel_pstate_driver == &intel_pstate && !hwp_active && -- pstate_funcs.get_target_pstate != get_target_pstate_use_cpu_load) -- intel_pstate_debug_expose_params(); -- - intel_pstate_sysfs_expose_params(); - -+ mutex_lock(&intel_pstate_driver_lock); -+ rc = intel_pstate_register_driver(); -+ mutex_unlock(&intel_pstate_driver_lock); -+ if (rc) -+ return rc; -+ - if (hwp_active) - pr_info("HWP enabled\n"); - -- return rc; --out: -- get_online_cpus(); -- for_each_online_cpu(cpu) { -- if (all_cpu_data[cpu]) { -- if (intel_pstate_driver == &intel_pstate) -- intel_pstate_clear_update_util_hook(cpu); -- -- kfree(all_cpu_data[cpu]); -- } -- } -- -- put_online_cpus(); -- vfree(all_cpu_data); -- return -ENODEV; -+ return 0; - } - device_initcall(intel_pstate_init); - |