Merge remote-tracking branch 'asoc/topic/cs42l73' into asoc-next
[GitHub/mt8127/android_kernel_alcatel_ttab.git] / arch / arm / kernel / perf_event_v7.c
1 /*
2 * ARMv7 Cortex-A8 and Cortex-A9 Performance Events handling code.
3 *
4 * ARMv7 support: Jean Pihet <jpihet@mvista.com>
5 * 2010 (c) MontaVista Software, LLC.
6 *
7 * Copied from ARMv6 code, with the low level code inspired
8 * by the ARMv7 Oprofile code.
9 *
10 * Cortex-A8 has up to 4 configurable performance counters and
11 * a single cycle counter.
12 * Cortex-A9 has up to 31 configurable performance counters and
13 * a single cycle counter.
14 *
15 * All counters can be enabled/disabled and IRQ masked separately. The cycle
16 * counter and all 4 performance counters together can be reset separately.
17 */
18
19 #ifdef CONFIG_CPU_V7
20
21 /*
22 * Common ARMv7 event types
23 *
24 * Note: An implementation may not be able to count all of these events
25 * but the encodings are considered to be `reserved' in the case that
26 * they are not available.
27 */
28 enum armv7_perf_types {
29 ARMV7_PERFCTR_PMNC_SW_INCR = 0x00,
30 ARMV7_PERFCTR_L1_ICACHE_REFILL = 0x01,
31 ARMV7_PERFCTR_ITLB_REFILL = 0x02,
32 ARMV7_PERFCTR_L1_DCACHE_REFILL = 0x03,
33 ARMV7_PERFCTR_L1_DCACHE_ACCESS = 0x04,
34 ARMV7_PERFCTR_DTLB_REFILL = 0x05,
35 ARMV7_PERFCTR_MEM_READ = 0x06,
36 ARMV7_PERFCTR_MEM_WRITE = 0x07,
37 ARMV7_PERFCTR_INSTR_EXECUTED = 0x08,
38 ARMV7_PERFCTR_EXC_TAKEN = 0x09,
39 ARMV7_PERFCTR_EXC_EXECUTED = 0x0A,
40 ARMV7_PERFCTR_CID_WRITE = 0x0B,
41
42 /*
43 * ARMV7_PERFCTR_PC_WRITE is equivalent to HW_BRANCH_INSTRUCTIONS.
44 * It counts:
45 * - all (taken) branch instructions,
46 * - instructions that explicitly write the PC,
47 * - exception generating instructions.
48 */
49 ARMV7_PERFCTR_PC_WRITE = 0x0C,
50 ARMV7_PERFCTR_PC_IMM_BRANCH = 0x0D,
51 ARMV7_PERFCTR_PC_PROC_RETURN = 0x0E,
52 ARMV7_PERFCTR_MEM_UNALIGNED_ACCESS = 0x0F,
53 ARMV7_PERFCTR_PC_BRANCH_MIS_PRED = 0x10,
54 ARMV7_PERFCTR_CLOCK_CYCLES = 0x11,
55 ARMV7_PERFCTR_PC_BRANCH_PRED = 0x12,
56
57 /* These events are defined by the PMUv2 supplement (ARM DDI 0457A). */
58 ARMV7_PERFCTR_MEM_ACCESS = 0x13,
59 ARMV7_PERFCTR_L1_ICACHE_ACCESS = 0x14,
60 ARMV7_PERFCTR_L1_DCACHE_WB = 0x15,
61 ARMV7_PERFCTR_L2_CACHE_ACCESS = 0x16,
62 ARMV7_PERFCTR_L2_CACHE_REFILL = 0x17,
63 ARMV7_PERFCTR_L2_CACHE_WB = 0x18,
64 ARMV7_PERFCTR_BUS_ACCESS = 0x19,
65 ARMV7_PERFCTR_MEM_ERROR = 0x1A,
66 ARMV7_PERFCTR_INSTR_SPEC = 0x1B,
67 ARMV7_PERFCTR_TTBR_WRITE = 0x1C,
68 ARMV7_PERFCTR_BUS_CYCLES = 0x1D,
69
70 ARMV7_PERFCTR_CPU_CYCLES = 0xFF
71 };
72
73 /* ARMv7 Cortex-A8 specific event types */
74 enum armv7_a8_perf_types {
75 ARMV7_A8_PERFCTR_L2_CACHE_ACCESS = 0x43,
76 ARMV7_A8_PERFCTR_L2_CACHE_REFILL = 0x44,
77 ARMV7_A8_PERFCTR_L1_ICACHE_ACCESS = 0x50,
78 ARMV7_A8_PERFCTR_STALL_ISIDE = 0x56,
79 };
80
81 /* ARMv7 Cortex-A9 specific event types */
82 enum armv7_a9_perf_types {
83 ARMV7_A9_PERFCTR_INSTR_CORE_RENAME = 0x68,
84 ARMV7_A9_PERFCTR_STALL_ICACHE = 0x60,
85 ARMV7_A9_PERFCTR_STALL_DISPATCH = 0x66,
86 };
87
88 /* ARMv7 Cortex-A5 specific event types */
89 enum armv7_a5_perf_types {
90 ARMV7_A5_PERFCTR_PREFETCH_LINEFILL = 0xc2,
91 ARMV7_A5_PERFCTR_PREFETCH_LINEFILL_DROP = 0xc3,
92 };
93
94 /* ARMv7 Cortex-A15 specific event types */
95 enum armv7_a15_perf_types {
96 ARMV7_A15_PERFCTR_L1_DCACHE_ACCESS_READ = 0x40,
97 ARMV7_A15_PERFCTR_L1_DCACHE_ACCESS_WRITE = 0x41,
98 ARMV7_A15_PERFCTR_L1_DCACHE_REFILL_READ = 0x42,
99 ARMV7_A15_PERFCTR_L1_DCACHE_REFILL_WRITE = 0x43,
100
101 ARMV7_A15_PERFCTR_DTLB_REFILL_L1_READ = 0x4C,
102 ARMV7_A15_PERFCTR_DTLB_REFILL_L1_WRITE = 0x4D,
103
104 ARMV7_A15_PERFCTR_L2_CACHE_ACCESS_READ = 0x50,
105 ARMV7_A15_PERFCTR_L2_CACHE_ACCESS_WRITE = 0x51,
106 ARMV7_A15_PERFCTR_L2_CACHE_REFILL_READ = 0x52,
107 ARMV7_A15_PERFCTR_L2_CACHE_REFILL_WRITE = 0x53,
108
109 ARMV7_A15_PERFCTR_PC_WRITE_SPEC = 0x76,
110 };
111
112 /*
113 * Cortex-A8 HW events mapping
114 *
115 * The hardware events that we support. We do support cache operations but
116 * we have harvard caches and no way to combine instruction and data
117 * accesses/misses in hardware.
118 */
119 static const unsigned armv7_a8_perf_map[PERF_COUNT_HW_MAX] = {
120 [PERF_COUNT_HW_CPU_CYCLES] = ARMV7_PERFCTR_CPU_CYCLES,
121 [PERF_COUNT_HW_INSTRUCTIONS] = ARMV7_PERFCTR_INSTR_EXECUTED,
122 [PERF_COUNT_HW_CACHE_REFERENCES] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
123 [PERF_COUNT_HW_CACHE_MISSES] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
124 [PERF_COUNT_HW_BRANCH_INSTRUCTIONS] = ARMV7_PERFCTR_PC_WRITE,
125 [PERF_COUNT_HW_BRANCH_MISSES] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
126 [PERF_COUNT_HW_BUS_CYCLES] = HW_OP_UNSUPPORTED,
127 [PERF_COUNT_HW_STALLED_CYCLES_FRONTEND] = ARMV7_A8_PERFCTR_STALL_ISIDE,
128 [PERF_COUNT_HW_STALLED_CYCLES_BACKEND] = HW_OP_UNSUPPORTED,
129 };
130
131 static const unsigned armv7_a8_perf_cache_map[PERF_COUNT_HW_CACHE_MAX]
132 [PERF_COUNT_HW_CACHE_OP_MAX]
133 [PERF_COUNT_HW_CACHE_RESULT_MAX] = {
134 [C(L1D)] = {
135 /*
136 * The performance counters don't differentiate between read
137 * and write accesses/misses so this isn't strictly correct,
138 * but it's the best we can do. Writes and reads get
139 * combined.
140 */
141 [C(OP_READ)] = {
142 [C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
143 [C(RESULT_MISS)] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
144 },
145 [C(OP_WRITE)] = {
146 [C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
147 [C(RESULT_MISS)] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
148 },
149 [C(OP_PREFETCH)] = {
150 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
151 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
152 },
153 },
154 [C(L1I)] = {
155 [C(OP_READ)] = {
156 [C(RESULT_ACCESS)] = ARMV7_A8_PERFCTR_L1_ICACHE_ACCESS,
157 [C(RESULT_MISS)] = ARMV7_PERFCTR_L1_ICACHE_REFILL,
158 },
159 [C(OP_WRITE)] = {
160 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
161 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
162 },
163 [C(OP_PREFETCH)] = {
164 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
165 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
166 },
167 },
168 [C(LL)] = {
169 [C(OP_READ)] = {
170 [C(RESULT_ACCESS)] = ARMV7_A8_PERFCTR_L2_CACHE_ACCESS,
171 [C(RESULT_MISS)] = ARMV7_A8_PERFCTR_L2_CACHE_REFILL,
172 },
173 [C(OP_WRITE)] = {
174 [C(RESULT_ACCESS)] = ARMV7_A8_PERFCTR_L2_CACHE_ACCESS,
175 [C(RESULT_MISS)] = ARMV7_A8_PERFCTR_L2_CACHE_REFILL,
176 },
177 [C(OP_PREFETCH)] = {
178 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
179 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
180 },
181 },
182 [C(DTLB)] = {
183 [C(OP_READ)] = {
184 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
185 [C(RESULT_MISS)] = ARMV7_PERFCTR_DTLB_REFILL,
186 },
187 [C(OP_WRITE)] = {
188 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
189 [C(RESULT_MISS)] = ARMV7_PERFCTR_DTLB_REFILL,
190 },
191 [C(OP_PREFETCH)] = {
192 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
193 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
194 },
195 },
196 [C(ITLB)] = {
197 [C(OP_READ)] = {
198 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
199 [C(RESULT_MISS)] = ARMV7_PERFCTR_ITLB_REFILL,
200 },
201 [C(OP_WRITE)] = {
202 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
203 [C(RESULT_MISS)] = ARMV7_PERFCTR_ITLB_REFILL,
204 },
205 [C(OP_PREFETCH)] = {
206 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
207 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
208 },
209 },
210 [C(BPU)] = {
211 [C(OP_READ)] = {
212 [C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
213 [C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
214 },
215 [C(OP_WRITE)] = {
216 [C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
217 [C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
218 },
219 [C(OP_PREFETCH)] = {
220 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
221 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
222 },
223 },
224 [C(NODE)] = {
225 [C(OP_READ)] = {
226 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
227 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
228 },
229 [C(OP_WRITE)] = {
230 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
231 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
232 },
233 [C(OP_PREFETCH)] = {
234 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
235 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
236 },
237 },
238 };
239
240 /*
241 * Cortex-A9 HW events mapping
242 */
243 static const unsigned armv7_a9_perf_map[PERF_COUNT_HW_MAX] = {
244 [PERF_COUNT_HW_CPU_CYCLES] = ARMV7_PERFCTR_CPU_CYCLES,
245 [PERF_COUNT_HW_INSTRUCTIONS] = ARMV7_A9_PERFCTR_INSTR_CORE_RENAME,
246 [PERF_COUNT_HW_CACHE_REFERENCES] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
247 [PERF_COUNT_HW_CACHE_MISSES] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
248 [PERF_COUNT_HW_BRANCH_INSTRUCTIONS] = ARMV7_PERFCTR_PC_WRITE,
249 [PERF_COUNT_HW_BRANCH_MISSES] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
250 [PERF_COUNT_HW_BUS_CYCLES] = HW_OP_UNSUPPORTED,
251 [PERF_COUNT_HW_STALLED_CYCLES_FRONTEND] = ARMV7_A9_PERFCTR_STALL_ICACHE,
252 [PERF_COUNT_HW_STALLED_CYCLES_BACKEND] = ARMV7_A9_PERFCTR_STALL_DISPATCH,
253 };
254
255 static const unsigned armv7_a9_perf_cache_map[PERF_COUNT_HW_CACHE_MAX]
256 [PERF_COUNT_HW_CACHE_OP_MAX]
257 [PERF_COUNT_HW_CACHE_RESULT_MAX] = {
258 [C(L1D)] = {
259 /*
260 * The performance counters don't differentiate between read
261 * and write accesses/misses so this isn't strictly correct,
262 * but it's the best we can do. Writes and reads get
263 * combined.
264 */
265 [C(OP_READ)] = {
266 [C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
267 [C(RESULT_MISS)] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
268 },
269 [C(OP_WRITE)] = {
270 [C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
271 [C(RESULT_MISS)] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
272 },
273 [C(OP_PREFETCH)] = {
274 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
275 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
276 },
277 },
278 [C(L1I)] = {
279 [C(OP_READ)] = {
280 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
281 [C(RESULT_MISS)] = ARMV7_PERFCTR_L1_ICACHE_REFILL,
282 },
283 [C(OP_WRITE)] = {
284 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
285 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
286 },
287 [C(OP_PREFETCH)] = {
288 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
289 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
290 },
291 },
292 [C(LL)] = {
293 [C(OP_READ)] = {
294 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
295 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
296 },
297 [C(OP_WRITE)] = {
298 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
299 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
300 },
301 [C(OP_PREFETCH)] = {
302 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
303 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
304 },
305 },
306 [C(DTLB)] = {
307 [C(OP_READ)] = {
308 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
309 [C(RESULT_MISS)] = ARMV7_PERFCTR_DTLB_REFILL,
310 },
311 [C(OP_WRITE)] = {
312 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
313 [C(RESULT_MISS)] = ARMV7_PERFCTR_DTLB_REFILL,
314 },
315 [C(OP_PREFETCH)] = {
316 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
317 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
318 },
319 },
320 [C(ITLB)] = {
321 [C(OP_READ)] = {
322 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
323 [C(RESULT_MISS)] = ARMV7_PERFCTR_ITLB_REFILL,
324 },
325 [C(OP_WRITE)] = {
326 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
327 [C(RESULT_MISS)] = ARMV7_PERFCTR_ITLB_REFILL,
328 },
329 [C(OP_PREFETCH)] = {
330 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
331 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
332 },
333 },
334 [C(BPU)] = {
335 [C(OP_READ)] = {
336 [C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
337 [C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
338 },
339 [C(OP_WRITE)] = {
340 [C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
341 [C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
342 },
343 [C(OP_PREFETCH)] = {
344 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
345 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
346 },
347 },
348 [C(NODE)] = {
349 [C(OP_READ)] = {
350 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
351 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
352 },
353 [C(OP_WRITE)] = {
354 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
355 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
356 },
357 [C(OP_PREFETCH)] = {
358 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
359 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
360 },
361 },
362 };
363
364 /*
365 * Cortex-A5 HW events mapping
366 */
367 static const unsigned armv7_a5_perf_map[PERF_COUNT_HW_MAX] = {
368 [PERF_COUNT_HW_CPU_CYCLES] = ARMV7_PERFCTR_CPU_CYCLES,
369 [PERF_COUNT_HW_INSTRUCTIONS] = ARMV7_PERFCTR_INSTR_EXECUTED,
370 [PERF_COUNT_HW_CACHE_REFERENCES] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
371 [PERF_COUNT_HW_CACHE_MISSES] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
372 [PERF_COUNT_HW_BRANCH_INSTRUCTIONS] = ARMV7_PERFCTR_PC_WRITE,
373 [PERF_COUNT_HW_BRANCH_MISSES] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
374 [PERF_COUNT_HW_BUS_CYCLES] = HW_OP_UNSUPPORTED,
375 [PERF_COUNT_HW_STALLED_CYCLES_FRONTEND] = HW_OP_UNSUPPORTED,
376 [PERF_COUNT_HW_STALLED_CYCLES_BACKEND] = HW_OP_UNSUPPORTED,
377 };
378
379 static const unsigned armv7_a5_perf_cache_map[PERF_COUNT_HW_CACHE_MAX]
380 [PERF_COUNT_HW_CACHE_OP_MAX]
381 [PERF_COUNT_HW_CACHE_RESULT_MAX] = {
382 [C(L1D)] = {
383 [C(OP_READ)] = {
384 [C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
385 [C(RESULT_MISS)] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
386 },
387 [C(OP_WRITE)] = {
388 [C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
389 [C(RESULT_MISS)] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
390 },
391 [C(OP_PREFETCH)] = {
392 [C(RESULT_ACCESS)] = ARMV7_A5_PERFCTR_PREFETCH_LINEFILL,
393 [C(RESULT_MISS)] = ARMV7_A5_PERFCTR_PREFETCH_LINEFILL_DROP,
394 },
395 },
396 [C(L1I)] = {
397 [C(OP_READ)] = {
398 [C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_ICACHE_ACCESS,
399 [C(RESULT_MISS)] = ARMV7_PERFCTR_L1_ICACHE_REFILL,
400 },
401 [C(OP_WRITE)] = {
402 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
403 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
404 },
405 /*
406 * The prefetch counters don't differentiate between the I
407 * side and the D side.
408 */
409 [C(OP_PREFETCH)] = {
410 [C(RESULT_ACCESS)] = ARMV7_A5_PERFCTR_PREFETCH_LINEFILL,
411 [C(RESULT_MISS)] = ARMV7_A5_PERFCTR_PREFETCH_LINEFILL_DROP,
412 },
413 },
414 [C(LL)] = {
415 [C(OP_READ)] = {
416 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
417 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
418 },
419 [C(OP_WRITE)] = {
420 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
421 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
422 },
423 [C(OP_PREFETCH)] = {
424 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
425 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
426 },
427 },
428 [C(DTLB)] = {
429 [C(OP_READ)] = {
430 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
431 [C(RESULT_MISS)] = ARMV7_PERFCTR_DTLB_REFILL,
432 },
433 [C(OP_WRITE)] = {
434 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
435 [C(RESULT_MISS)] = ARMV7_PERFCTR_DTLB_REFILL,
436 },
437 [C(OP_PREFETCH)] = {
438 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
439 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
440 },
441 },
442 [C(ITLB)] = {
443 [C(OP_READ)] = {
444 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
445 [C(RESULT_MISS)] = ARMV7_PERFCTR_ITLB_REFILL,
446 },
447 [C(OP_WRITE)] = {
448 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
449 [C(RESULT_MISS)] = ARMV7_PERFCTR_ITLB_REFILL,
450 },
451 [C(OP_PREFETCH)] = {
452 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
453 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
454 },
455 },
456 [C(BPU)] = {
457 [C(OP_READ)] = {
458 [C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
459 [C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
460 },
461 [C(OP_WRITE)] = {
462 [C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
463 [C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
464 },
465 [C(OP_PREFETCH)] = {
466 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
467 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
468 },
469 },
470 [C(NODE)] = {
471 [C(OP_READ)] = {
472 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
473 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
474 },
475 [C(OP_WRITE)] = {
476 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
477 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
478 },
479 [C(OP_PREFETCH)] = {
480 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
481 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
482 },
483 },
484 };
485
486 /*
487 * Cortex-A15 HW events mapping
488 */
489 static const unsigned armv7_a15_perf_map[PERF_COUNT_HW_MAX] = {
490 [PERF_COUNT_HW_CPU_CYCLES] = ARMV7_PERFCTR_CPU_CYCLES,
491 [PERF_COUNT_HW_INSTRUCTIONS] = ARMV7_PERFCTR_INSTR_EXECUTED,
492 [PERF_COUNT_HW_CACHE_REFERENCES] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
493 [PERF_COUNT_HW_CACHE_MISSES] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
494 [PERF_COUNT_HW_BRANCH_INSTRUCTIONS] = ARMV7_A15_PERFCTR_PC_WRITE_SPEC,
495 [PERF_COUNT_HW_BRANCH_MISSES] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
496 [PERF_COUNT_HW_BUS_CYCLES] = ARMV7_PERFCTR_BUS_CYCLES,
497 [PERF_COUNT_HW_STALLED_CYCLES_FRONTEND] = HW_OP_UNSUPPORTED,
498 [PERF_COUNT_HW_STALLED_CYCLES_BACKEND] = HW_OP_UNSUPPORTED,
499 };
500
501 static const unsigned armv7_a15_perf_cache_map[PERF_COUNT_HW_CACHE_MAX]
502 [PERF_COUNT_HW_CACHE_OP_MAX]
503 [PERF_COUNT_HW_CACHE_RESULT_MAX] = {
504 [C(L1D)] = {
505 [C(OP_READ)] = {
506 [C(RESULT_ACCESS)] = ARMV7_A15_PERFCTR_L1_DCACHE_ACCESS_READ,
507 [C(RESULT_MISS)] = ARMV7_A15_PERFCTR_L1_DCACHE_REFILL_READ,
508 },
509 [C(OP_WRITE)] = {
510 [C(RESULT_ACCESS)] = ARMV7_A15_PERFCTR_L1_DCACHE_ACCESS_WRITE,
511 [C(RESULT_MISS)] = ARMV7_A15_PERFCTR_L1_DCACHE_REFILL_WRITE,
512 },
513 [C(OP_PREFETCH)] = {
514 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
515 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
516 },
517 },
518 [C(L1I)] = {
519 /*
520 * Not all performance counters differentiate between read
521 * and write accesses/misses so we're not always strictly
522 * correct, but it's the best we can do. Writes and reads get
523 * combined in these cases.
524 */
525 [C(OP_READ)] = {
526 [C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_ICACHE_ACCESS,
527 [C(RESULT_MISS)] = ARMV7_PERFCTR_L1_ICACHE_REFILL,
528 },
529 [C(OP_WRITE)] = {
530 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
531 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
532 },
533 [C(OP_PREFETCH)] = {
534 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
535 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
536 },
537 },
538 [C(LL)] = {
539 [C(OP_READ)] = {
540 [C(RESULT_ACCESS)] = ARMV7_A15_PERFCTR_L2_CACHE_ACCESS_READ,
541 [C(RESULT_MISS)] = ARMV7_A15_PERFCTR_L2_CACHE_REFILL_READ,
542 },
543 [C(OP_WRITE)] = {
544 [C(RESULT_ACCESS)] = ARMV7_A15_PERFCTR_L2_CACHE_ACCESS_WRITE,
545 [C(RESULT_MISS)] = ARMV7_A15_PERFCTR_L2_CACHE_REFILL_WRITE,
546 },
547 [C(OP_PREFETCH)] = {
548 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
549 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
550 },
551 },
552 [C(DTLB)] = {
553 [C(OP_READ)] = {
554 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
555 [C(RESULT_MISS)] = ARMV7_A15_PERFCTR_DTLB_REFILL_L1_READ,
556 },
557 [C(OP_WRITE)] = {
558 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
559 [C(RESULT_MISS)] = ARMV7_A15_PERFCTR_DTLB_REFILL_L1_WRITE,
560 },
561 [C(OP_PREFETCH)] = {
562 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
563 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
564 },
565 },
566 [C(ITLB)] = {
567 [C(OP_READ)] = {
568 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
569 [C(RESULT_MISS)] = ARMV7_PERFCTR_ITLB_REFILL,
570 },
571 [C(OP_WRITE)] = {
572 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
573 [C(RESULT_MISS)] = ARMV7_PERFCTR_ITLB_REFILL,
574 },
575 [C(OP_PREFETCH)] = {
576 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
577 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
578 },
579 },
580 [C(BPU)] = {
581 [C(OP_READ)] = {
582 [C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
583 [C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
584 },
585 [C(OP_WRITE)] = {
586 [C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
587 [C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
588 },
589 [C(OP_PREFETCH)] = {
590 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
591 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
592 },
593 },
594 [C(NODE)] = {
595 [C(OP_READ)] = {
596 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
597 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
598 },
599 [C(OP_WRITE)] = {
600 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
601 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
602 },
603 [C(OP_PREFETCH)] = {
604 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
605 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
606 },
607 },
608 };
609
610 /*
611 * Cortex-A7 HW events mapping
612 */
613 static const unsigned armv7_a7_perf_map[PERF_COUNT_HW_MAX] = {
614 [PERF_COUNT_HW_CPU_CYCLES] = ARMV7_PERFCTR_CPU_CYCLES,
615 [PERF_COUNT_HW_INSTRUCTIONS] = ARMV7_PERFCTR_INSTR_EXECUTED,
616 [PERF_COUNT_HW_CACHE_REFERENCES] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
617 [PERF_COUNT_HW_CACHE_MISSES] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
618 [PERF_COUNT_HW_BRANCH_INSTRUCTIONS] = ARMV7_PERFCTR_PC_WRITE,
619 [PERF_COUNT_HW_BRANCH_MISSES] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
620 [PERF_COUNT_HW_BUS_CYCLES] = ARMV7_PERFCTR_BUS_CYCLES,
621 [PERF_COUNT_HW_STALLED_CYCLES_FRONTEND] = HW_OP_UNSUPPORTED,
622 [PERF_COUNT_HW_STALLED_CYCLES_BACKEND] = HW_OP_UNSUPPORTED,
623 };
624
625 static const unsigned armv7_a7_perf_cache_map[PERF_COUNT_HW_CACHE_MAX]
626 [PERF_COUNT_HW_CACHE_OP_MAX]
627 [PERF_COUNT_HW_CACHE_RESULT_MAX] = {
628 [C(L1D)] = {
629 /*
630 * The performance counters don't differentiate between read
631 * and write accesses/misses so this isn't strictly correct,
632 * but it's the best we can do. Writes and reads get
633 * combined.
634 */
635 [C(OP_READ)] = {
636 [C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
637 [C(RESULT_MISS)] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
638 },
639 [C(OP_WRITE)] = {
640 [C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_DCACHE_ACCESS,
641 [C(RESULT_MISS)] = ARMV7_PERFCTR_L1_DCACHE_REFILL,
642 },
643 [C(OP_PREFETCH)] = {
644 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
645 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
646 },
647 },
648 [C(L1I)] = {
649 [C(OP_READ)] = {
650 [C(RESULT_ACCESS)] = ARMV7_PERFCTR_L1_ICACHE_ACCESS,
651 [C(RESULT_MISS)] = ARMV7_PERFCTR_L1_ICACHE_REFILL,
652 },
653 [C(OP_WRITE)] = {
654 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
655 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
656 },
657 [C(OP_PREFETCH)] = {
658 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
659 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
660 },
661 },
662 [C(LL)] = {
663 [C(OP_READ)] = {
664 [C(RESULT_ACCESS)] = ARMV7_PERFCTR_L2_CACHE_ACCESS,
665 [C(RESULT_MISS)] = ARMV7_PERFCTR_L2_CACHE_REFILL,
666 },
667 [C(OP_WRITE)] = {
668 [C(RESULT_ACCESS)] = ARMV7_PERFCTR_L2_CACHE_ACCESS,
669 [C(RESULT_MISS)] = ARMV7_PERFCTR_L2_CACHE_REFILL,
670 },
671 [C(OP_PREFETCH)] = {
672 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
673 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
674 },
675 },
676 [C(DTLB)] = {
677 [C(OP_READ)] = {
678 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
679 [C(RESULT_MISS)] = ARMV7_PERFCTR_DTLB_REFILL,
680 },
681 [C(OP_WRITE)] = {
682 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
683 [C(RESULT_MISS)] = ARMV7_PERFCTR_DTLB_REFILL,
684 },
685 [C(OP_PREFETCH)] = {
686 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
687 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
688 },
689 },
690 [C(ITLB)] = {
691 [C(OP_READ)] = {
692 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
693 [C(RESULT_MISS)] = ARMV7_PERFCTR_ITLB_REFILL,
694 },
695 [C(OP_WRITE)] = {
696 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
697 [C(RESULT_MISS)] = ARMV7_PERFCTR_ITLB_REFILL,
698 },
699 [C(OP_PREFETCH)] = {
700 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
701 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
702 },
703 },
704 [C(BPU)] = {
705 [C(OP_READ)] = {
706 [C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
707 [C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
708 },
709 [C(OP_WRITE)] = {
710 [C(RESULT_ACCESS)] = ARMV7_PERFCTR_PC_BRANCH_PRED,
711 [C(RESULT_MISS)] = ARMV7_PERFCTR_PC_BRANCH_MIS_PRED,
712 },
713 [C(OP_PREFETCH)] = {
714 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
715 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
716 },
717 },
718 [C(NODE)] = {
719 [C(OP_READ)] = {
720 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
721 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
722 },
723 [C(OP_WRITE)] = {
724 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
725 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
726 },
727 [C(OP_PREFETCH)] = {
728 [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
729 [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
730 },
731 },
732 };
733
734 /*
735 * Perf Events' indices
736 */
737 #define ARMV7_IDX_CYCLE_COUNTER 0
738 #define ARMV7_IDX_COUNTER0 1
739 #define ARMV7_IDX_COUNTER_LAST(cpu_pmu) \
740 (ARMV7_IDX_CYCLE_COUNTER + cpu_pmu->num_events - 1)
741
742 #define ARMV7_MAX_COUNTERS 32
743 #define ARMV7_COUNTER_MASK (ARMV7_MAX_COUNTERS - 1)
744
745 /*
746 * ARMv7 low level PMNC access
747 */
748
749 /*
750 * Perf Event to low level counters mapping
751 */
752 #define ARMV7_IDX_TO_COUNTER(x) \
753 (((x) - ARMV7_IDX_COUNTER0) & ARMV7_COUNTER_MASK)
754
755 /*
756 * Per-CPU PMNC: config reg
757 */
758 #define ARMV7_PMNC_E (1 << 0) /* Enable all counters */
759 #define ARMV7_PMNC_P (1 << 1) /* Reset all counters */
760 #define ARMV7_PMNC_C (1 << 2) /* Cycle counter reset */
761 #define ARMV7_PMNC_D (1 << 3) /* CCNT counts every 64th cpu cycle */
762 #define ARMV7_PMNC_X (1 << 4) /* Export to ETM */
763 #define ARMV7_PMNC_DP (1 << 5) /* Disable CCNT if non-invasive debug*/
764 #define ARMV7_PMNC_N_SHIFT 11 /* Number of counters supported */
765 #define ARMV7_PMNC_N_MASK 0x1f
766 #define ARMV7_PMNC_MASK 0x3f /* Mask for writable bits */
767
768 /*
769 * FLAG: counters overflow flag status reg
770 */
771 #define ARMV7_FLAG_MASK 0xffffffff /* Mask for writable bits */
772 #define ARMV7_OVERFLOWED_MASK ARMV7_FLAG_MASK
773
774 /*
775 * PMXEVTYPER: Event selection reg
776 */
777 #define ARMV7_EVTYPE_MASK 0xc80000ff /* Mask for writable bits */
778 #define ARMV7_EVTYPE_EVENT 0xff /* Mask for EVENT bits */
779
780 /*
781 * Event filters for PMUv2
782 */
783 #define ARMV7_EXCLUDE_PL1 (1 << 31)
784 #define ARMV7_EXCLUDE_USER (1 << 30)
785 #define ARMV7_INCLUDE_HYP (1 << 27)
786
787 static inline u32 armv7_pmnc_read(void)
788 {
789 u32 val;
790 asm volatile("mrc p15, 0, %0, c9, c12, 0" : "=r"(val));
791 return val;
792 }
793
794 static inline void armv7_pmnc_write(u32 val)
795 {
796 val &= ARMV7_PMNC_MASK;
797 isb();
798 asm volatile("mcr p15, 0, %0, c9, c12, 0" : : "r"(val));
799 }
800
801 static inline int armv7_pmnc_has_overflowed(u32 pmnc)
802 {
803 return pmnc & ARMV7_OVERFLOWED_MASK;
804 }
805
806 static inline int armv7_pmnc_counter_valid(struct arm_pmu *cpu_pmu, int idx)
807 {
808 return idx >= ARMV7_IDX_CYCLE_COUNTER &&
809 idx <= ARMV7_IDX_COUNTER_LAST(cpu_pmu);
810 }
811
812 static inline int armv7_pmnc_counter_has_overflowed(u32 pmnc, int idx)
813 {
814 return pmnc & BIT(ARMV7_IDX_TO_COUNTER(idx));
815 }
816
817 static inline int armv7_pmnc_select_counter(int idx)
818 {
819 u32 counter = ARMV7_IDX_TO_COUNTER(idx);
820 asm volatile("mcr p15, 0, %0, c9, c12, 5" : : "r" (counter));
821 isb();
822
823 return idx;
824 }
825
826 static inline u32 armv7pmu_read_counter(struct perf_event *event)
827 {
828 struct arm_pmu *cpu_pmu = to_arm_pmu(event->pmu);
829 struct hw_perf_event *hwc = &event->hw;
830 int idx = hwc->idx;
831 u32 value = 0;
832
833 if (!armv7_pmnc_counter_valid(cpu_pmu, idx))
834 pr_err("CPU%u reading wrong counter %d\n",
835 smp_processor_id(), idx);
836 else if (idx == ARMV7_IDX_CYCLE_COUNTER)
837 asm volatile("mrc p15, 0, %0, c9, c13, 0" : "=r" (value));
838 else if (armv7_pmnc_select_counter(idx) == idx)
839 asm volatile("mrc p15, 0, %0, c9, c13, 2" : "=r" (value));
840
841 return value;
842 }
843
844 static inline void armv7pmu_write_counter(struct perf_event *event, u32 value)
845 {
846 struct arm_pmu *cpu_pmu = to_arm_pmu(event->pmu);
847 struct hw_perf_event *hwc = &event->hw;
848 int idx = hwc->idx;
849
850 if (!armv7_pmnc_counter_valid(cpu_pmu, idx))
851 pr_err("CPU%u writing wrong counter %d\n",
852 smp_processor_id(), idx);
853 else if (idx == ARMV7_IDX_CYCLE_COUNTER)
854 asm volatile("mcr p15, 0, %0, c9, c13, 0" : : "r" (value));
855 else if (armv7_pmnc_select_counter(idx) == idx)
856 asm volatile("mcr p15, 0, %0, c9, c13, 2" : : "r" (value));
857 }
858
859 static inline void armv7_pmnc_write_evtsel(int idx, u32 val)
860 {
861 if (armv7_pmnc_select_counter(idx) == idx) {
862 val &= ARMV7_EVTYPE_MASK;
863 asm volatile("mcr p15, 0, %0, c9, c13, 1" : : "r" (val));
864 }
865 }
866
867 static inline int armv7_pmnc_enable_counter(int idx)
868 {
869 u32 counter = ARMV7_IDX_TO_COUNTER(idx);
870 asm volatile("mcr p15, 0, %0, c9, c12, 1" : : "r" (BIT(counter)));
871 return idx;
872 }
873
874 static inline int armv7_pmnc_disable_counter(int idx)
875 {
876 u32 counter = ARMV7_IDX_TO_COUNTER(idx);
877 asm volatile("mcr p15, 0, %0, c9, c12, 2" : : "r" (BIT(counter)));
878 return idx;
879 }
880
881 static inline int armv7_pmnc_enable_intens(int idx)
882 {
883 u32 counter = ARMV7_IDX_TO_COUNTER(idx);
884 asm volatile("mcr p15, 0, %0, c9, c14, 1" : : "r" (BIT(counter)));
885 return idx;
886 }
887
888 static inline int armv7_pmnc_disable_intens(int idx)
889 {
890 u32 counter = ARMV7_IDX_TO_COUNTER(idx);
891 asm volatile("mcr p15, 0, %0, c9, c14, 2" : : "r" (BIT(counter)));
892 isb();
893 /* Clear the overflow flag in case an interrupt is pending. */
894 asm volatile("mcr p15, 0, %0, c9, c12, 3" : : "r" (BIT(counter)));
895 isb();
896
897 return idx;
898 }
899
900 static inline u32 armv7_pmnc_getreset_flags(void)
901 {
902 u32 val;
903
904 /* Read */
905 asm volatile("mrc p15, 0, %0, c9, c12, 3" : "=r" (val));
906
907 /* Write to clear flags */
908 val &= ARMV7_FLAG_MASK;
909 asm volatile("mcr p15, 0, %0, c9, c12, 3" : : "r" (val));
910
911 return val;
912 }
913
914 #ifdef DEBUG
915 static void armv7_pmnc_dump_regs(struct arm_pmu *cpu_pmu)
916 {
917 u32 val;
918 unsigned int cnt;
919
920 printk(KERN_INFO "PMNC registers dump:\n");
921
922 asm volatile("mrc p15, 0, %0, c9, c12, 0" : "=r" (val));
923 printk(KERN_INFO "PMNC =0x%08x\n", val);
924
925 asm volatile("mrc p15, 0, %0, c9, c12, 1" : "=r" (val));
926 printk(KERN_INFO "CNTENS=0x%08x\n", val);
927
928 asm volatile("mrc p15, 0, %0, c9, c14, 1" : "=r" (val));
929 printk(KERN_INFO "INTENS=0x%08x\n", val);
930
931 asm volatile("mrc p15, 0, %0, c9, c12, 3" : "=r" (val));
932 printk(KERN_INFO "FLAGS =0x%08x\n", val);
933
934 asm volatile("mrc p15, 0, %0, c9, c12, 5" : "=r" (val));
935 printk(KERN_INFO "SELECT=0x%08x\n", val);
936
937 asm volatile("mrc p15, 0, %0, c9, c13, 0" : "=r" (val));
938 printk(KERN_INFO "CCNT =0x%08x\n", val);
939
940 for (cnt = ARMV7_IDX_COUNTER0;
941 cnt <= ARMV7_IDX_COUNTER_LAST(cpu_pmu); cnt++) {
942 armv7_pmnc_select_counter(cnt);
943 asm volatile("mrc p15, 0, %0, c9, c13, 2" : "=r" (val));
944 printk(KERN_INFO "CNT[%d] count =0x%08x\n",
945 ARMV7_IDX_TO_COUNTER(cnt), val);
946 asm volatile("mrc p15, 0, %0, c9, c13, 1" : "=r" (val));
947 printk(KERN_INFO "CNT[%d] evtsel=0x%08x\n",
948 ARMV7_IDX_TO_COUNTER(cnt), val);
949 }
950 }
951 #endif
952
953 static void armv7pmu_enable_event(struct perf_event *event)
954 {
955 unsigned long flags;
956 struct hw_perf_event *hwc = &event->hw;
957 struct arm_pmu *cpu_pmu = to_arm_pmu(event->pmu);
958 struct pmu_hw_events *events = cpu_pmu->get_hw_events();
959 int idx = hwc->idx;
960
961 if (!armv7_pmnc_counter_valid(cpu_pmu, idx)) {
962 pr_err("CPU%u enabling wrong PMNC counter IRQ enable %d\n",
963 smp_processor_id(), idx);
964 return;
965 }
966
967 /*
968 * Enable counter and interrupt, and set the counter to count
969 * the event that we're interested in.
970 */
971 raw_spin_lock_irqsave(&events->pmu_lock, flags);
972
973 /*
974 * Disable counter
975 */
976 armv7_pmnc_disable_counter(idx);
977
978 /*
979 * Set event (if destined for PMNx counters)
980 * We only need to set the event for the cycle counter if we
981 * have the ability to perform event filtering.
982 */
983 if (cpu_pmu->set_event_filter || idx != ARMV7_IDX_CYCLE_COUNTER)
984 armv7_pmnc_write_evtsel(idx, hwc->config_base);
985
986 /*
987 * Enable interrupt for this counter
988 */
989 armv7_pmnc_enable_intens(idx);
990
991 /*
992 * Enable counter
993 */
994 armv7_pmnc_enable_counter(idx);
995
996 raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
997 }
998
999 static void armv7pmu_disable_event(struct perf_event *event)
1000 {
1001 unsigned long flags;
1002 struct hw_perf_event *hwc = &event->hw;
1003 struct arm_pmu *cpu_pmu = to_arm_pmu(event->pmu);
1004 struct pmu_hw_events *events = cpu_pmu->get_hw_events();
1005 int idx = hwc->idx;
1006
1007 if (!armv7_pmnc_counter_valid(cpu_pmu, idx)) {
1008 pr_err("CPU%u disabling wrong PMNC counter IRQ enable %d\n",
1009 smp_processor_id(), idx);
1010 return;
1011 }
1012
1013 /*
1014 * Disable counter and interrupt
1015 */
1016 raw_spin_lock_irqsave(&events->pmu_lock, flags);
1017
1018 /*
1019 * Disable counter
1020 */
1021 armv7_pmnc_disable_counter(idx);
1022
1023 /*
1024 * Disable interrupt for this counter
1025 */
1026 armv7_pmnc_disable_intens(idx);
1027
1028 raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
1029 }
1030
1031 static irqreturn_t armv7pmu_handle_irq(int irq_num, void *dev)
1032 {
1033 u32 pmnc;
1034 struct perf_sample_data data;
1035 struct arm_pmu *cpu_pmu = (struct arm_pmu *)dev;
1036 struct pmu_hw_events *cpuc = cpu_pmu->get_hw_events();
1037 struct pt_regs *regs;
1038 int idx;
1039
1040 /*
1041 * Get and reset the IRQ flags
1042 */
1043 pmnc = armv7_pmnc_getreset_flags();
1044
1045 /*
1046 * Did an overflow occur?
1047 */
1048 if (!armv7_pmnc_has_overflowed(pmnc))
1049 return IRQ_NONE;
1050
1051 /*
1052 * Handle the counter(s) overflow(s)
1053 */
1054 regs = get_irq_regs();
1055
1056 for (idx = 0; idx < cpu_pmu->num_events; ++idx) {
1057 struct perf_event *event = cpuc->events[idx];
1058 struct hw_perf_event *hwc;
1059
1060 /* Ignore if we don't have an event. */
1061 if (!event)
1062 continue;
1063
1064 /*
1065 * We have a single interrupt for all counters. Check that
1066 * each counter has overflowed before we process it.
1067 */
1068 if (!armv7_pmnc_counter_has_overflowed(pmnc, idx))
1069 continue;
1070
1071 hwc = &event->hw;
1072 armpmu_event_update(event);
1073 perf_sample_data_init(&data, 0, hwc->last_period);
1074 if (!armpmu_event_set_period(event))
1075 continue;
1076
1077 if (perf_event_overflow(event, &data, regs))
1078 cpu_pmu->disable(event);
1079 }
1080
1081 /*
1082 * Handle the pending perf events.
1083 *
1084 * Note: this call *must* be run with interrupts disabled. For
1085 * platforms that can have the PMU interrupts raised as an NMI, this
1086 * will not work.
1087 */
1088 irq_work_run();
1089
1090 return IRQ_HANDLED;
1091 }
1092
1093 static void armv7pmu_start(struct arm_pmu *cpu_pmu)
1094 {
1095 unsigned long flags;
1096 struct pmu_hw_events *events = cpu_pmu->get_hw_events();
1097
1098 raw_spin_lock_irqsave(&events->pmu_lock, flags);
1099 /* Enable all counters */
1100 armv7_pmnc_write(armv7_pmnc_read() | ARMV7_PMNC_E);
1101 raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
1102 }
1103
1104 static void armv7pmu_stop(struct arm_pmu *cpu_pmu)
1105 {
1106 unsigned long flags;
1107 struct pmu_hw_events *events = cpu_pmu->get_hw_events();
1108
1109 raw_spin_lock_irqsave(&events->pmu_lock, flags);
1110 /* Disable all counters */
1111 armv7_pmnc_write(armv7_pmnc_read() & ~ARMV7_PMNC_E);
1112 raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
1113 }
1114
1115 static int armv7pmu_get_event_idx(struct pmu_hw_events *cpuc,
1116 struct perf_event *event)
1117 {
1118 int idx;
1119 struct arm_pmu *cpu_pmu = to_arm_pmu(event->pmu);
1120 struct hw_perf_event *hwc = &event->hw;
1121 unsigned long evtype = hwc->config_base & ARMV7_EVTYPE_EVENT;
1122
1123 /* Always place a cycle counter into the cycle counter. */
1124 if (evtype == ARMV7_PERFCTR_CPU_CYCLES) {
1125 if (test_and_set_bit(ARMV7_IDX_CYCLE_COUNTER, cpuc->used_mask))
1126 return -EAGAIN;
1127
1128 return ARMV7_IDX_CYCLE_COUNTER;
1129 }
1130
1131 /*
1132 * For anything other than a cycle counter, try and use
1133 * the events counters
1134 */
1135 for (idx = ARMV7_IDX_COUNTER0; idx < cpu_pmu->num_events; ++idx) {
1136 if (!test_and_set_bit(idx, cpuc->used_mask))
1137 return idx;
1138 }
1139
1140 /* The counters are all in use. */
1141 return -EAGAIN;
1142 }
1143
1144 /*
1145 * Add an event filter to a given event. This will only work for PMUv2 PMUs.
1146 */
1147 static int armv7pmu_set_event_filter(struct hw_perf_event *event,
1148 struct perf_event_attr *attr)
1149 {
1150 unsigned long config_base = 0;
1151
1152 if (attr->exclude_idle)
1153 return -EPERM;
1154 if (attr->exclude_user)
1155 config_base |= ARMV7_EXCLUDE_USER;
1156 if (attr->exclude_kernel)
1157 config_base |= ARMV7_EXCLUDE_PL1;
1158 if (!attr->exclude_hv)
1159 config_base |= ARMV7_INCLUDE_HYP;
1160
1161 /*
1162 * Install the filter into config_base as this is used to
1163 * construct the event type.
1164 */
1165 event->config_base = config_base;
1166
1167 return 0;
1168 }
1169
1170 static void armv7pmu_reset(void *info)
1171 {
1172 struct arm_pmu *cpu_pmu = (struct arm_pmu *)info;
1173 u32 idx, nb_cnt = cpu_pmu->num_events;
1174
1175 /* The counter and interrupt enable registers are unknown at reset. */
1176 for (idx = ARMV7_IDX_CYCLE_COUNTER; idx < nb_cnt; ++idx) {
1177 armv7_pmnc_disable_counter(idx);
1178 armv7_pmnc_disable_intens(idx);
1179 }
1180
1181 /* Initialize & Reset PMNC: C and P bits */
1182 armv7_pmnc_write(ARMV7_PMNC_P | ARMV7_PMNC_C);
1183 }
1184
1185 static int armv7_a8_map_event(struct perf_event *event)
1186 {
1187 return armpmu_map_event(event, &armv7_a8_perf_map,
1188 &armv7_a8_perf_cache_map, 0xFF);
1189 }
1190
1191 static int armv7_a9_map_event(struct perf_event *event)
1192 {
1193 return armpmu_map_event(event, &armv7_a9_perf_map,
1194 &armv7_a9_perf_cache_map, 0xFF);
1195 }
1196
1197 static int armv7_a5_map_event(struct perf_event *event)
1198 {
1199 return armpmu_map_event(event, &armv7_a5_perf_map,
1200 &armv7_a5_perf_cache_map, 0xFF);
1201 }
1202
1203 static int armv7_a15_map_event(struct perf_event *event)
1204 {
1205 return armpmu_map_event(event, &armv7_a15_perf_map,
1206 &armv7_a15_perf_cache_map, 0xFF);
1207 }
1208
1209 static int armv7_a7_map_event(struct perf_event *event)
1210 {
1211 return armpmu_map_event(event, &armv7_a7_perf_map,
1212 &armv7_a7_perf_cache_map, 0xFF);
1213 }
1214
1215 static void armv7pmu_init(struct arm_pmu *cpu_pmu)
1216 {
1217 cpu_pmu->handle_irq = armv7pmu_handle_irq;
1218 cpu_pmu->enable = armv7pmu_enable_event;
1219 cpu_pmu->disable = armv7pmu_disable_event;
1220 cpu_pmu->read_counter = armv7pmu_read_counter;
1221 cpu_pmu->write_counter = armv7pmu_write_counter;
1222 cpu_pmu->get_event_idx = armv7pmu_get_event_idx;
1223 cpu_pmu->start = armv7pmu_start;
1224 cpu_pmu->stop = armv7pmu_stop;
1225 cpu_pmu->reset = armv7pmu_reset;
1226 cpu_pmu->max_period = (1LLU << 32) - 1;
1227 };
1228
1229 static u32 armv7_read_num_pmnc_events(void)
1230 {
1231 u32 nb_cnt;
1232
1233 /* Read the nb of CNTx counters supported from PMNC */
1234 nb_cnt = (armv7_pmnc_read() >> ARMV7_PMNC_N_SHIFT) & ARMV7_PMNC_N_MASK;
1235
1236 /* Add the CPU cycles counter and return */
1237 return nb_cnt + 1;
1238 }
1239
1240 static int armv7_a8_pmu_init(struct arm_pmu *cpu_pmu)
1241 {
1242 armv7pmu_init(cpu_pmu);
1243 cpu_pmu->name = "ARMv7 Cortex-A8";
1244 cpu_pmu->map_event = armv7_a8_map_event;
1245 cpu_pmu->num_events = armv7_read_num_pmnc_events();
1246 return 0;
1247 }
1248
1249 static int armv7_a9_pmu_init(struct arm_pmu *cpu_pmu)
1250 {
1251 armv7pmu_init(cpu_pmu);
1252 cpu_pmu->name = "ARMv7 Cortex-A9";
1253 cpu_pmu->map_event = armv7_a9_map_event;
1254 cpu_pmu->num_events = armv7_read_num_pmnc_events();
1255 return 0;
1256 }
1257
1258 static int armv7_a5_pmu_init(struct arm_pmu *cpu_pmu)
1259 {
1260 armv7pmu_init(cpu_pmu);
1261 cpu_pmu->name = "ARMv7 Cortex-A5";
1262 cpu_pmu->map_event = armv7_a5_map_event;
1263 cpu_pmu->num_events = armv7_read_num_pmnc_events();
1264 return 0;
1265 }
1266
1267 static int armv7_a15_pmu_init(struct arm_pmu *cpu_pmu)
1268 {
1269 armv7pmu_init(cpu_pmu);
1270 cpu_pmu->name = "ARMv7 Cortex-A15";
1271 cpu_pmu->map_event = armv7_a15_map_event;
1272 cpu_pmu->num_events = armv7_read_num_pmnc_events();
1273 cpu_pmu->set_event_filter = armv7pmu_set_event_filter;
1274 return 0;
1275 }
1276
1277 static int armv7_a7_pmu_init(struct arm_pmu *cpu_pmu)
1278 {
1279 armv7pmu_init(cpu_pmu);
1280 cpu_pmu->name = "ARMv7 Cortex-A7";
1281 cpu_pmu->map_event = armv7_a7_map_event;
1282 cpu_pmu->num_events = armv7_read_num_pmnc_events();
1283 cpu_pmu->set_event_filter = armv7pmu_set_event_filter;
1284 return 0;
1285 }
1286 #else
1287 static inline int armv7_a8_pmu_init(struct arm_pmu *cpu_pmu)
1288 {
1289 return -ENODEV;
1290 }
1291
1292 static inline int armv7_a9_pmu_init(struct arm_pmu *cpu_pmu)
1293 {
1294 return -ENODEV;
1295 }
1296
1297 static inline int armv7_a5_pmu_init(struct arm_pmu *cpu_pmu)
1298 {
1299 return -ENODEV;
1300 }
1301
1302 static inline int armv7_a15_pmu_init(struct arm_pmu *cpu_pmu)
1303 {
1304 return -ENODEV;
1305 }
1306
1307 static inline int armv7_a7_pmu_init(struct arm_pmu *cpu_pmu)
1308 {
1309 return -ENODEV;
1310 }
1311 #endif /* CONFIG_CPU_V7 */