1
2
3
4
5
6
7#if defined(__i386__) || defined(__x86_64__)
8
9#include <stdio.h>
10#include <stdint.h>
11#include <stdlib.h>
12#include <string.h>
13#include <limits.h>
14
15#include <cpufreq.h>
16
17#include "helpers/helpers.h"
18#include "idle_monitor/cpupower-monitor.h"
19
20#define MSR_APERF 0xE8
21#define MSR_MPERF 0xE7
22
23#define MSR_TSC 0x10
24
25#define MSR_AMD_HWCR 0xc0010015
26
27enum mperf_id { C0 = 0, Cx, AVG_FREQ, MPERF_CSTATE_COUNT };
28
29static int mperf_get_count_percent(unsigned int self_id, double *percent,
30 unsigned int cpu);
31static int mperf_get_count_freq(unsigned int id, unsigned long long *count,
32 unsigned int cpu);
33static struct timespec time_start, time_end;
34
35static cstate_t mperf_cstates[MPERF_CSTATE_COUNT] = {
36 {
37 .name = "C0",
38 .desc = N_("Processor Core not idle"),
39 .id = C0,
40 .range = RANGE_THREAD,
41 .get_count_percent = mperf_get_count_percent,
42 },
43 {
44 .name = "Cx",
45 .desc = N_("Processor Core in an idle state"),
46 .id = Cx,
47 .range = RANGE_THREAD,
48 .get_count_percent = mperf_get_count_percent,
49 },
50
51 {
52 .name = "Freq",
53 .desc = N_("Average Frequency (including boost) in MHz"),
54 .id = AVG_FREQ,
55 .range = RANGE_THREAD,
56 .get_count = mperf_get_count_freq,
57 },
58};
59
60enum MAX_FREQ_MODE { MAX_FREQ_SYSFS, MAX_FREQ_TSC_REF };
61static int max_freq_mode;
62
63
64
65
66
67
68static unsigned long max_frequency;
69
70static unsigned long long tsc_at_measure_start;
71static unsigned long long tsc_at_measure_end;
72static unsigned long long *mperf_previous_count;
73static unsigned long long *aperf_previous_count;
74static unsigned long long *mperf_current_count;
75static unsigned long long *aperf_current_count;
76
77
78static int *is_valid;
79
80static int mperf_get_tsc(unsigned long long *tsc)
81{
82 int ret;
83
84 ret = read_msr(base_cpu, MSR_TSC, tsc);
85 if (ret)
86 dprint("Reading TSC MSR failed, returning %llu\n", *tsc);
87 return ret;
88}
89
90static int mperf_init_stats(unsigned int cpu)
91{
92 unsigned long long val;
93 int ret;
94
95 ret = read_msr(cpu, MSR_APERF, &val);
96 aperf_previous_count[cpu] = val;
97 ret |= read_msr(cpu, MSR_MPERF, &val);
98 mperf_previous_count[cpu] = val;
99 is_valid[cpu] = !ret;
100
101 return 0;
102}
103
104static int mperf_measure_stats(unsigned int cpu)
105{
106 unsigned long long val;
107 int ret;
108
109 ret = read_msr(cpu, MSR_APERF, &val);
110 aperf_current_count[cpu] = val;
111 ret |= read_msr(cpu, MSR_MPERF, &val);
112 mperf_current_count[cpu] = val;
113 is_valid[cpu] = !ret;
114
115 return 0;
116}
117
118static int mperf_get_count_percent(unsigned int id, double *percent,
119 unsigned int cpu)
120{
121 unsigned long long aperf_diff, mperf_diff, tsc_diff;
122 unsigned long long timediff;
123
124 if (!is_valid[cpu])
125 return -1;
126
127 if (id != C0 && id != Cx)
128 return -1;
129
130 mperf_diff = mperf_current_count[cpu] - mperf_previous_count[cpu];
131 aperf_diff = aperf_current_count[cpu] - aperf_previous_count[cpu];
132
133 if (max_freq_mode == MAX_FREQ_TSC_REF) {
134 tsc_diff = tsc_at_measure_end - tsc_at_measure_start;
135 *percent = 100.0 * mperf_diff / tsc_diff;
136 dprint("%s: TSC Ref - mperf_diff: %llu, tsc_diff: %llu\n",
137 mperf_cstates[id].name, mperf_diff, tsc_diff);
138 } else if (max_freq_mode == MAX_FREQ_SYSFS) {
139 timediff = max_frequency * timespec_diff_us(time_start, time_end);
140 *percent = 100.0 * mperf_diff / timediff;
141 dprint("%s: MAXFREQ - mperf_diff: %llu, time_diff: %llu\n",
142 mperf_cstates[id].name, mperf_diff, timediff);
143 } else
144 return -1;
145
146 if (id == Cx)
147 *percent = 100.0 - *percent;
148
149 dprint("%s: previous: %llu - current: %llu - (%u)\n",
150 mperf_cstates[id].name, mperf_diff, aperf_diff, cpu);
151 dprint("%s: %f\n", mperf_cstates[id].name, *percent);
152 return 0;
153}
154
155static int mperf_get_count_freq(unsigned int id, unsigned long long *count,
156 unsigned int cpu)
157{
158 unsigned long long aperf_diff, mperf_diff, time_diff, tsc_diff;
159
160 if (id != AVG_FREQ)
161 return 1;
162
163 if (!is_valid[cpu])
164 return -1;
165
166 mperf_diff = mperf_current_count[cpu] - mperf_previous_count[cpu];
167 aperf_diff = aperf_current_count[cpu] - aperf_previous_count[cpu];
168
169 if (max_freq_mode == MAX_FREQ_TSC_REF) {
170
171 tsc_diff = tsc_at_measure_end - tsc_at_measure_start;
172 time_diff = timespec_diff_us(time_start, time_end);
173 max_frequency = tsc_diff / time_diff;
174 }
175
176 *count = max_frequency * ((double)aperf_diff / mperf_diff);
177 dprint("%s: Average freq based on %s maximum frequency:\n",
178 mperf_cstates[id].name,
179 (max_freq_mode == MAX_FREQ_TSC_REF) ? "TSC calculated" : "sysfs read");
180 dprint("max_frequency: %lu\n", max_frequency);
181 dprint("aperf_diff: %llu\n", aperf_diff);
182 dprint("mperf_diff: %llu\n", mperf_diff);
183 dprint("avg freq: %llu\n", *count);
184 return 0;
185}
186
187static int mperf_start(void)
188{
189 int cpu;
190 unsigned long long dbg;
191
192 clock_gettime(CLOCK_REALTIME, &time_start);
193 mperf_get_tsc(&tsc_at_measure_start);
194
195 for (cpu = 0; cpu < cpu_count; cpu++)
196 mperf_init_stats(cpu);
197
198 mperf_get_tsc(&dbg);
199 dprint("TSC diff: %llu\n", dbg - tsc_at_measure_start);
200 return 0;
201}
202
203static int mperf_stop(void)
204{
205 unsigned long long dbg;
206 int cpu;
207
208 for (cpu = 0; cpu < cpu_count; cpu++)
209 mperf_measure_stats(cpu);
210
211 mperf_get_tsc(&tsc_at_measure_end);
212 clock_gettime(CLOCK_REALTIME, &time_end);
213
214 mperf_get_tsc(&dbg);
215 dprint("TSC diff: %llu\n", dbg - tsc_at_measure_end);
216
217 return 0;
218}
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235static int init_maxfreq_mode(void)
236{
237 int ret;
238 unsigned long long hwcr;
239 unsigned long min;
240
241 if (!(cpupower_cpu_info.caps & CPUPOWER_CAP_INV_TSC))
242 goto use_sysfs;
243
244 if (cpupower_cpu_info.vendor == X86_VENDOR_AMD ||
245 cpupower_cpu_info.vendor == X86_VENDOR_HYGON) {
246
247
248
249
250
251
252
253
254
255
256 ret = read_msr(0, MSR_AMD_HWCR, &hwcr);
257
258
259
260
261 if (ret != 0) {
262 dprint("TSC read 0x%x failed - assume TSC working\n",
263 MSR_AMD_HWCR);
264 return 0;
265 } else if (1 & (hwcr >> 24)) {
266 max_freq_mode = MAX_FREQ_TSC_REF;
267 return 0;
268 } else { }
269 } else if (cpupower_cpu_info.vendor == X86_VENDOR_INTEL) {
270
271
272
273
274 max_freq_mode = MAX_FREQ_TSC_REF;
275 return 0;
276 }
277use_sysfs:
278 if (cpufreq_get_hardware_limits(0, &min, &max_frequency)) {
279 dprint("Cannot retrieve max freq from cpufreq kernel "
280 "subsystem\n");
281 return -1;
282 }
283 max_freq_mode = MAX_FREQ_SYSFS;
284 max_frequency /= 1000;
285 return 0;
286}
287
288
289
290
291
292
293
294
295
296
297
298
299
300struct cpuidle_monitor mperf_monitor;
301struct cpuidle_monitor *mperf_register(void)
302{
303 if (!(cpupower_cpu_info.caps & CPUPOWER_CAP_APERF))
304 return NULL;
305
306 if (init_maxfreq_mode())
307 return NULL;
308
309
310 is_valid = calloc(cpu_count, sizeof(int));
311 mperf_previous_count = calloc(cpu_count, sizeof(unsigned long long));
312 aperf_previous_count = calloc(cpu_count, sizeof(unsigned long long));
313 mperf_current_count = calloc(cpu_count, sizeof(unsigned long long));
314 aperf_current_count = calloc(cpu_count, sizeof(unsigned long long));
315
316 mperf_monitor.name_len = strlen(mperf_monitor.name);
317 return &mperf_monitor;
318}
319
320void mperf_unregister(void)
321{
322 free(mperf_previous_count);
323 free(aperf_previous_count);
324 free(mperf_current_count);
325 free(aperf_current_count);
326 free(is_valid);
327}
328
329struct cpuidle_monitor mperf_monitor = {
330 .name = "Mperf",
331 .hw_states_num = MPERF_CSTATE_COUNT,
332 .hw_states = mperf_cstates,
333 .start = mperf_start,
334 .stop = mperf_stop,
335 .do_register = mperf_register,
336 .unregister = mperf_unregister,
337 .needs_root = 1,
338 .overflow_s = 922000000
339
340};
341#endif
342