356 lines
11 KiB
Diff
356 lines
11 KiB
Diff
From 300686c32b77583f45c6763535da85f2242bf820 Mon Sep 17 00:00:00 2001
|
|
From: Dhananjay Ugwekar <dhananjay.ugwekar@amd.com>
|
|
Date: Wed, 5 Feb 2025 11:25:17 +0000
|
|
Subject: cpufreq/amd-pstate: Convert all perf values to u8
|
|
|
|
All perf values are always within 0-255 range, hence convert their
|
|
datatype to u8 everywhere.
|
|
|
|
Signed-off-by: Dhananjay Ugwekar <dhananjay.ugwekar@amd.com>
|
|
Reviewed-by: Mario Limonciello <mario.limonciello@amd.com>
|
|
---
|
|
drivers/cpufreq/amd-pstate-trace.h | 46 +++++++++++------------
|
|
drivers/cpufreq/amd-pstate.c | 60 +++++++++++++++---------------
|
|
drivers/cpufreq/amd-pstate.h | 18 ++++-----
|
|
3 files changed, 62 insertions(+), 62 deletions(-)
|
|
|
|
--- a/drivers/cpufreq/amd-pstate-trace.h
|
|
+++ b/drivers/cpufreq/amd-pstate-trace.h
|
|
@@ -24,9 +24,9 @@
|
|
|
|
TRACE_EVENT(amd_pstate_perf,
|
|
|
|
- TP_PROTO(unsigned long min_perf,
|
|
- unsigned long target_perf,
|
|
- unsigned long capacity,
|
|
+ TP_PROTO(u8 min_perf,
|
|
+ u8 target_perf,
|
|
+ u8 capacity,
|
|
u64 freq,
|
|
u64 mperf,
|
|
u64 aperf,
|
|
@@ -47,9 +47,9 @@ TRACE_EVENT(amd_pstate_perf,
|
|
),
|
|
|
|
TP_STRUCT__entry(
|
|
- __field(unsigned long, min_perf)
|
|
- __field(unsigned long, target_perf)
|
|
- __field(unsigned long, capacity)
|
|
+ __field(u8, min_perf)
|
|
+ __field(u8, target_perf)
|
|
+ __field(u8, capacity)
|
|
__field(unsigned long long, freq)
|
|
__field(unsigned long long, mperf)
|
|
__field(unsigned long long, aperf)
|
|
@@ -70,10 +70,10 @@ TRACE_EVENT(amd_pstate_perf,
|
|
__entry->fast_switch = fast_switch;
|
|
),
|
|
|
|
- TP_printk("amd_min_perf=%lu amd_des_perf=%lu amd_max_perf=%lu freq=%llu mperf=%llu aperf=%llu tsc=%llu cpu_id=%u fast_switch=%s",
|
|
- (unsigned long)__entry->min_perf,
|
|
- (unsigned long)__entry->target_perf,
|
|
- (unsigned long)__entry->capacity,
|
|
+ TP_printk("amd_min_perf=%hhu amd_des_perf=%hhu amd_max_perf=%hhu freq=%llu mperf=%llu aperf=%llu tsc=%llu cpu_id=%u fast_switch=%s",
|
|
+ (u8)__entry->min_perf,
|
|
+ (u8)__entry->target_perf,
|
|
+ (u8)__entry->capacity,
|
|
(unsigned long long)__entry->freq,
|
|
(unsigned long long)__entry->mperf,
|
|
(unsigned long long)__entry->aperf,
|
|
@@ -86,10 +86,10 @@ TRACE_EVENT(amd_pstate_perf,
|
|
TRACE_EVENT(amd_pstate_epp_perf,
|
|
|
|
TP_PROTO(unsigned int cpu_id,
|
|
- unsigned int highest_perf,
|
|
- unsigned int epp,
|
|
- unsigned int min_perf,
|
|
- unsigned int max_perf,
|
|
+ u8 highest_perf,
|
|
+ u8 epp,
|
|
+ u8 min_perf,
|
|
+ u8 max_perf,
|
|
bool boost
|
|
),
|
|
|
|
@@ -102,10 +102,10 @@ TRACE_EVENT(amd_pstate_epp_perf,
|
|
|
|
TP_STRUCT__entry(
|
|
__field(unsigned int, cpu_id)
|
|
- __field(unsigned int, highest_perf)
|
|
- __field(unsigned int, epp)
|
|
- __field(unsigned int, min_perf)
|
|
- __field(unsigned int, max_perf)
|
|
+ __field(u8, highest_perf)
|
|
+ __field(u8, epp)
|
|
+ __field(u8, min_perf)
|
|
+ __field(u8, max_perf)
|
|
__field(bool, boost)
|
|
),
|
|
|
|
@@ -118,12 +118,12 @@ TRACE_EVENT(amd_pstate_epp_perf,
|
|
__entry->boost = boost;
|
|
),
|
|
|
|
- TP_printk("cpu%u: [%u<->%u]/%u, epp=%u, boost=%u",
|
|
+ TP_printk("cpu%u: [%hhu<->%hhu]/%hhu, epp=%hhu, boost=%u",
|
|
(unsigned int)__entry->cpu_id,
|
|
- (unsigned int)__entry->min_perf,
|
|
- (unsigned int)__entry->max_perf,
|
|
- (unsigned int)__entry->highest_perf,
|
|
- (unsigned int)__entry->epp,
|
|
+ (u8)__entry->min_perf,
|
|
+ (u8)__entry->max_perf,
|
|
+ (u8)__entry->highest_perf,
|
|
+ (u8)__entry->epp,
|
|
(bool)__entry->boost
|
|
)
|
|
);
|
|
--- a/drivers/cpufreq/amd-pstate.c
|
|
+++ b/drivers/cpufreq/amd-pstate.c
|
|
@@ -186,7 +186,7 @@ static inline int get_mode_idx_from_str(
|
|
static DEFINE_MUTEX(amd_pstate_limits_lock);
|
|
static DEFINE_MUTEX(amd_pstate_driver_lock);
|
|
|
|
-static s16 msr_get_epp(struct amd_cpudata *cpudata)
|
|
+static u8 msr_get_epp(struct amd_cpudata *cpudata)
|
|
{
|
|
u64 value;
|
|
int ret;
|
|
@@ -207,7 +207,7 @@ static inline s16 amd_pstate_get_epp(str
|
|
return static_call(amd_pstate_get_epp)(cpudata);
|
|
}
|
|
|
|
-static s16 shmem_get_epp(struct amd_cpudata *cpudata)
|
|
+static u8 shmem_get_epp(struct amd_cpudata *cpudata)
|
|
{
|
|
u64 epp;
|
|
int ret;
|
|
@@ -218,11 +218,11 @@ static s16 shmem_get_epp(struct amd_cpud
|
|
return ret;
|
|
}
|
|
|
|
- return (s16)(epp & 0xff);
|
|
+ return FIELD_GET(AMD_CPPC_EPP_PERF_MASK, epp);
|
|
}
|
|
|
|
-static int msr_update_perf(struct amd_cpudata *cpudata, u32 min_perf,
|
|
- u32 des_perf, u32 max_perf, u32 epp, bool fast_switch)
|
|
+static int msr_update_perf(struct amd_cpudata *cpudata, u8 min_perf,
|
|
+ u8 des_perf, u8 max_perf, u8 epp, bool fast_switch)
|
|
{
|
|
u64 value, prev;
|
|
|
|
@@ -257,15 +257,15 @@ static int msr_update_perf(struct amd_cp
|
|
DEFINE_STATIC_CALL(amd_pstate_update_perf, msr_update_perf);
|
|
|
|
static inline int amd_pstate_update_perf(struct amd_cpudata *cpudata,
|
|
- u32 min_perf, u32 des_perf,
|
|
- u32 max_perf, u32 epp,
|
|
+ u8 min_perf, u8 des_perf,
|
|
+ u8 max_perf, u8 epp,
|
|
bool fast_switch)
|
|
{
|
|
return static_call(amd_pstate_update_perf)(cpudata, min_perf, des_perf,
|
|
max_perf, epp, fast_switch);
|
|
}
|
|
|
|
-static int msr_set_epp(struct amd_cpudata *cpudata, u32 epp)
|
|
+static int msr_set_epp(struct amd_cpudata *cpudata, u8 epp)
|
|
{
|
|
u64 value, prev;
|
|
int ret;
|
|
@@ -292,12 +292,12 @@ static int msr_set_epp(struct amd_cpudat
|
|
|
|
DEFINE_STATIC_CALL(amd_pstate_set_epp, msr_set_epp);
|
|
|
|
-static inline int amd_pstate_set_epp(struct amd_cpudata *cpudata, u32 epp)
|
|
+static inline int amd_pstate_set_epp(struct amd_cpudata *cpudata, u8 epp)
|
|
{
|
|
return static_call(amd_pstate_set_epp)(cpudata, epp);
|
|
}
|
|
|
|
-static int shmem_set_epp(struct amd_cpudata *cpudata, u32 epp)
|
|
+static int shmem_set_epp(struct amd_cpudata *cpudata, u8 epp)
|
|
{
|
|
int ret;
|
|
struct cppc_perf_ctrls perf_ctrls;
|
|
@@ -320,7 +320,7 @@ static int amd_pstate_set_energy_pref_in
|
|
int pref_index)
|
|
{
|
|
struct amd_cpudata *cpudata = policy->driver_data;
|
|
- int epp;
|
|
+ u8 epp;
|
|
|
|
if (!pref_index)
|
|
epp = cpudata->epp_default;
|
|
@@ -479,8 +479,8 @@ static inline int amd_pstate_init_perf(s
|
|
return static_call(amd_pstate_init_perf)(cpudata);
|
|
}
|
|
|
|
-static int shmem_update_perf(struct amd_cpudata *cpudata, u32 min_perf,
|
|
- u32 des_perf, u32 max_perf, u32 epp, bool fast_switch)
|
|
+static int shmem_update_perf(struct amd_cpudata *cpudata, u8 min_perf,
|
|
+ u8 des_perf, u8 max_perf, u8 epp, bool fast_switch)
|
|
{
|
|
struct cppc_perf_ctrls perf_ctrls;
|
|
|
|
@@ -531,14 +531,14 @@ static inline bool amd_pstate_sample(str
|
|
return true;
|
|
}
|
|
|
|
-static void amd_pstate_update(struct amd_cpudata *cpudata, u32 min_perf,
|
|
- u32 des_perf, u32 max_perf, bool fast_switch, int gov_flags)
|
|
+static void amd_pstate_update(struct amd_cpudata *cpudata, u8 min_perf,
|
|
+ u8 des_perf, u8 max_perf, bool fast_switch, int gov_flags)
|
|
{
|
|
unsigned long max_freq;
|
|
struct cpufreq_policy *policy = cpufreq_cpu_get(cpudata->cpu);
|
|
- u32 nominal_perf = READ_ONCE(cpudata->nominal_perf);
|
|
+ u8 nominal_perf = READ_ONCE(cpudata->nominal_perf);
|
|
|
|
- des_perf = clamp_t(unsigned long, des_perf, min_perf, max_perf);
|
|
+ des_perf = clamp_t(u8, des_perf, min_perf, max_perf);
|
|
|
|
max_freq = READ_ONCE(cpudata->max_limit_freq);
|
|
policy->cur = div_u64(des_perf * max_freq, max_perf);
|
|
@@ -550,7 +550,7 @@ static void amd_pstate_update(struct amd
|
|
|
|
/* limit the max perf when core performance boost feature is disabled */
|
|
if (!cpudata->boost_supported)
|
|
- max_perf = min_t(unsigned long, nominal_perf, max_perf);
|
|
+ max_perf = min_t(u8, nominal_perf, max_perf);
|
|
|
|
if (trace_amd_pstate_perf_enabled() && amd_pstate_sample(cpudata)) {
|
|
trace_amd_pstate_perf(min_perf, des_perf, max_perf, cpudata->freq,
|
|
@@ -591,7 +591,8 @@ static int amd_pstate_verify(struct cpuf
|
|
|
|
static int amd_pstate_update_min_max_limit(struct cpufreq_policy *policy)
|
|
{
|
|
- u32 max_limit_perf, min_limit_perf, max_perf, max_freq;
|
|
+ u8 max_limit_perf, min_limit_perf, max_perf;
|
|
+ u32 max_freq;
|
|
struct amd_cpudata *cpudata = policy->driver_data;
|
|
|
|
max_perf = READ_ONCE(cpudata->highest_perf);
|
|
@@ -615,7 +616,7 @@ static int amd_pstate_update_freq(struct
|
|
{
|
|
struct cpufreq_freqs freqs;
|
|
struct amd_cpudata *cpudata = policy->driver_data;
|
|
- unsigned long des_perf, cap_perf;
|
|
+ u8 des_perf, cap_perf;
|
|
|
|
if (!cpudata->max_freq)
|
|
return -ENODEV;
|
|
@@ -670,8 +671,7 @@ static void amd_pstate_adjust_perf(unsig
|
|
unsigned long target_perf,
|
|
unsigned long capacity)
|
|
{
|
|
- unsigned long max_perf, min_perf, des_perf,
|
|
- cap_perf, min_limit_perf;
|
|
+ u8 max_perf, min_perf, des_perf, cap_perf, min_limit_perf;
|
|
struct cpufreq_policy *policy = cpufreq_cpu_get(cpu);
|
|
struct amd_cpudata *cpudata;
|
|
|
|
@@ -905,8 +905,8 @@ static int amd_pstate_init_freq(struct a
|
|
{
|
|
int ret;
|
|
u32 min_freq, max_freq;
|
|
- u32 highest_perf, nominal_perf, nominal_freq;
|
|
- u32 lowest_nonlinear_perf, lowest_nonlinear_freq;
|
|
+ u8 highest_perf, nominal_perf, lowest_nonlinear_perf;
|
|
+ u32 nominal_freq, lowest_nonlinear_freq;
|
|
struct cppc_perf_caps cppc_perf;
|
|
|
|
ret = cppc_get_perf_caps(cpudata->cpu, &cppc_perf);
|
|
@@ -1113,7 +1113,7 @@ static ssize_t show_amd_pstate_lowest_no
|
|
static ssize_t show_amd_pstate_highest_perf(struct cpufreq_policy *policy,
|
|
char *buf)
|
|
{
|
|
- u32 perf;
|
|
+ u8 perf;
|
|
struct amd_cpudata *cpudata = policy->driver_data;
|
|
|
|
perf = READ_ONCE(cpudata->highest_perf);
|
|
@@ -1124,7 +1124,7 @@ static ssize_t show_amd_pstate_highest_p
|
|
static ssize_t show_amd_pstate_prefcore_ranking(struct cpufreq_policy *policy,
|
|
char *buf)
|
|
{
|
|
- u32 perf;
|
|
+ u8 perf;
|
|
struct amd_cpudata *cpudata = policy->driver_data;
|
|
|
|
perf = READ_ONCE(cpudata->prefcore_ranking);
|
|
@@ -1187,7 +1187,7 @@ static ssize_t show_energy_performance_p
|
|
struct cpufreq_policy *policy, char *buf)
|
|
{
|
|
struct amd_cpudata *cpudata = policy->driver_data;
|
|
- int preference;
|
|
+ u8 preference;
|
|
|
|
switch (cpudata->epp_cached) {
|
|
case AMD_CPPC_EPP_PERFORMANCE:
|
|
@@ -1549,7 +1549,7 @@ static void amd_pstate_epp_cpu_exit(stru
|
|
static int amd_pstate_epp_update_limit(struct cpufreq_policy *policy)
|
|
{
|
|
struct amd_cpudata *cpudata = policy->driver_data;
|
|
- u32 epp;
|
|
+ u8 epp;
|
|
|
|
amd_pstate_update_min_max_limit(policy);
|
|
|
|
@@ -1598,7 +1598,7 @@ static int amd_pstate_epp_set_policy(str
|
|
static int amd_pstate_epp_reenable(struct cpufreq_policy *policy)
|
|
{
|
|
struct amd_cpudata *cpudata = policy->driver_data;
|
|
- u64 max_perf;
|
|
+ u8 max_perf;
|
|
int ret;
|
|
|
|
ret = amd_pstate_cppc_enable(true);
|
|
@@ -1635,7 +1635,7 @@ static int amd_pstate_epp_cpu_online(str
|
|
static int amd_pstate_epp_cpu_offline(struct cpufreq_policy *policy)
|
|
{
|
|
struct amd_cpudata *cpudata = policy->driver_data;
|
|
- int min_perf;
|
|
+ u8 min_perf;
|
|
|
|
if (cpudata->suspended)
|
|
return 0;
|
|
--- a/drivers/cpufreq/amd-pstate.h
|
|
+++ b/drivers/cpufreq/amd-pstate.h
|
|
@@ -70,13 +70,13 @@ struct amd_cpudata {
|
|
struct freq_qos_request req[2];
|
|
u64 cppc_req_cached;
|
|
|
|
- u32 highest_perf;
|
|
- u32 nominal_perf;
|
|
- u32 lowest_nonlinear_perf;
|
|
- u32 lowest_perf;
|
|
- u32 prefcore_ranking;
|
|
- u32 min_limit_perf;
|
|
- u32 max_limit_perf;
|
|
+ u8 highest_perf;
|
|
+ u8 nominal_perf;
|
|
+ u8 lowest_nonlinear_perf;
|
|
+ u8 lowest_perf;
|
|
+ u8 prefcore_ranking;
|
|
+ u8 min_limit_perf;
|
|
+ u8 max_limit_perf;
|
|
u32 min_limit_freq;
|
|
u32 max_limit_freq;
|
|
|
|
@@ -93,11 +93,11 @@ struct amd_cpudata {
|
|
bool hw_prefcore;
|
|
|
|
/* EPP feature related attributes*/
|
|
- s16 epp_cached;
|
|
+ u8 epp_cached;
|
|
u32 policy;
|
|
u64 cppc_cap1_cached;
|
|
bool suspended;
|
|
- s16 epp_default;
|
|
+ u8 epp_default;
|
|
};
|
|
|
|
/*
|