perf_event_xscale.c 20 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823
  1. /*
  2. * ARMv5 [xscale] Performance counter handling code.
  3. *
  4. * Copyright (C) 2010, ARM Ltd., Will Deacon <will.deacon@arm.com>
  5. *
  6. * Based on the previous xscale OProfile code.
  7. *
  8. * There are two variants of the xscale PMU that we support:
  9. * - xscale1pmu: 2 event counters and a cycle counter
  10. * - xscale2pmu: 4 event counters and a cycle counter
  11. * The two variants share event definitions, but have different
  12. * PMU structures.
  13. */
  14. #ifdef CONFIG_CPU_XSCALE
  15. enum xscale_perf_types {
  16. XSCALE_PERFCTR_ICACHE_MISS = 0x00,
  17. XSCALE_PERFCTR_ICACHE_NO_DELIVER = 0x01,
  18. XSCALE_PERFCTR_DATA_STALL = 0x02,
  19. XSCALE_PERFCTR_ITLB_MISS = 0x03,
  20. XSCALE_PERFCTR_DTLB_MISS = 0x04,
  21. XSCALE_PERFCTR_BRANCH = 0x05,
  22. XSCALE_PERFCTR_BRANCH_MISS = 0x06,
  23. XSCALE_PERFCTR_INSTRUCTION = 0x07,
  24. XSCALE_PERFCTR_DCACHE_FULL_STALL = 0x08,
  25. XSCALE_PERFCTR_DCACHE_FULL_STALL_CONTIG = 0x09,
  26. XSCALE_PERFCTR_DCACHE_ACCESS = 0x0A,
  27. XSCALE_PERFCTR_DCACHE_MISS = 0x0B,
  28. XSCALE_PERFCTR_DCACHE_WRITE_BACK = 0x0C,
  29. XSCALE_PERFCTR_PC_CHANGED = 0x0D,
  30. XSCALE_PERFCTR_BCU_REQUEST = 0x10,
  31. XSCALE_PERFCTR_BCU_FULL = 0x11,
  32. XSCALE_PERFCTR_BCU_DRAIN = 0x12,
  33. XSCALE_PERFCTR_BCU_ECC_NO_ELOG = 0x14,
  34. XSCALE_PERFCTR_BCU_1_BIT_ERR = 0x15,
  35. XSCALE_PERFCTR_RMW = 0x16,
  36. /* XSCALE_PERFCTR_CCNT is not hardware defined */
  37. XSCALE_PERFCTR_CCNT = 0xFE,
  38. XSCALE_PERFCTR_UNUSED = 0xFF,
  39. };
  40. enum xscale_counters {
  41. XSCALE_CYCLE_COUNTER = 0,
  42. XSCALE_COUNTER0,
  43. XSCALE_COUNTER1,
  44. XSCALE_COUNTER2,
  45. XSCALE_COUNTER3,
  46. };
  47. static const unsigned xscale_perf_map[PERF_COUNT_HW_MAX] = {
  48. [PERF_COUNT_HW_CPU_CYCLES] = XSCALE_PERFCTR_CCNT,
  49. [PERF_COUNT_HW_INSTRUCTIONS] = XSCALE_PERFCTR_INSTRUCTION,
  50. [PERF_COUNT_HW_CACHE_REFERENCES] = HW_OP_UNSUPPORTED,
  51. [PERF_COUNT_HW_CACHE_MISSES] = HW_OP_UNSUPPORTED,
  52. [PERF_COUNT_HW_BRANCH_INSTRUCTIONS] = XSCALE_PERFCTR_BRANCH,
  53. [PERF_COUNT_HW_BRANCH_MISSES] = XSCALE_PERFCTR_BRANCH_MISS,
  54. [PERF_COUNT_HW_BUS_CYCLES] = HW_OP_UNSUPPORTED,
  55. };
  56. static const unsigned xscale_perf_cache_map[PERF_COUNT_HW_CACHE_MAX]
  57. [PERF_COUNT_HW_CACHE_OP_MAX]
  58. [PERF_COUNT_HW_CACHE_RESULT_MAX] = {
  59. [C(L1D)] = {
  60. [C(OP_READ)] = {
  61. [C(RESULT_ACCESS)] = XSCALE_PERFCTR_DCACHE_ACCESS,
  62. [C(RESULT_MISS)] = XSCALE_PERFCTR_DCACHE_MISS,
  63. },
  64. [C(OP_WRITE)] = {
  65. [C(RESULT_ACCESS)] = XSCALE_PERFCTR_DCACHE_ACCESS,
  66. [C(RESULT_MISS)] = XSCALE_PERFCTR_DCACHE_MISS,
  67. },
  68. [C(OP_PREFETCH)] = {
  69. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  70. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  71. },
  72. },
  73. [C(L1I)] = {
  74. [C(OP_READ)] = {
  75. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  76. [C(RESULT_MISS)] = XSCALE_PERFCTR_ICACHE_MISS,
  77. },
  78. [C(OP_WRITE)] = {
  79. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  80. [C(RESULT_MISS)] = XSCALE_PERFCTR_ICACHE_MISS,
  81. },
  82. [C(OP_PREFETCH)] = {
  83. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  84. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  85. },
  86. },
  87. [C(LL)] = {
  88. [C(OP_READ)] = {
  89. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  90. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  91. },
  92. [C(OP_WRITE)] = {
  93. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  94. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  95. },
  96. [C(OP_PREFETCH)] = {
  97. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  98. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  99. },
  100. },
  101. [C(DTLB)] = {
  102. [C(OP_READ)] = {
  103. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  104. [C(RESULT_MISS)] = XSCALE_PERFCTR_DTLB_MISS,
  105. },
  106. [C(OP_WRITE)] = {
  107. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  108. [C(RESULT_MISS)] = XSCALE_PERFCTR_DTLB_MISS,
  109. },
  110. [C(OP_PREFETCH)] = {
  111. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  112. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  113. },
  114. },
  115. [C(ITLB)] = {
  116. [C(OP_READ)] = {
  117. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  118. [C(RESULT_MISS)] = XSCALE_PERFCTR_ITLB_MISS,
  119. },
  120. [C(OP_WRITE)] = {
  121. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  122. [C(RESULT_MISS)] = XSCALE_PERFCTR_ITLB_MISS,
  123. },
  124. [C(OP_PREFETCH)] = {
  125. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  126. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  127. },
  128. },
  129. [C(BPU)] = {
  130. [C(OP_READ)] = {
  131. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  132. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  133. },
  134. [C(OP_WRITE)] = {
  135. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  136. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  137. },
  138. [C(OP_PREFETCH)] = {
  139. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  140. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  141. },
  142. },
  143. [C(NODE)] = {
  144. [C(OP_READ)] = {
  145. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  146. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  147. },
  148. [C(OP_WRITE)] = {
  149. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  150. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  151. },
  152. [C(OP_PREFETCH)] = {
  153. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  154. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  155. },
  156. },
  157. };
  158. #define XSCALE_PMU_ENABLE 0x001
  159. #define XSCALE_PMN_RESET 0x002
  160. #define XSCALE_CCNT_RESET 0x004
  161. #define XSCALE_PMU_RESET (CCNT_RESET | PMN_RESET)
  162. #define XSCALE_PMU_CNT64 0x008
  163. #define XSCALE1_OVERFLOWED_MASK 0x700
  164. #define XSCALE1_CCOUNT_OVERFLOW 0x400
  165. #define XSCALE1_COUNT0_OVERFLOW 0x100
  166. #define XSCALE1_COUNT1_OVERFLOW 0x200
  167. #define XSCALE1_CCOUNT_INT_EN 0x040
  168. #define XSCALE1_COUNT0_INT_EN 0x010
  169. #define XSCALE1_COUNT1_INT_EN 0x020
  170. #define XSCALE1_COUNT0_EVT_SHFT 12
  171. #define XSCALE1_COUNT0_EVT_MASK (0xff << XSCALE1_COUNT0_EVT_SHFT)
  172. #define XSCALE1_COUNT1_EVT_SHFT 20
  173. #define XSCALE1_COUNT1_EVT_MASK (0xff << XSCALE1_COUNT1_EVT_SHFT)
  174. static inline u32
  175. xscale1pmu_read_pmnc(void)
  176. {
  177. u32 val;
  178. asm volatile("mrc p14, 0, %0, c0, c0, 0" : "=r" (val));
  179. return val;
  180. }
  181. static inline void
  182. xscale1pmu_write_pmnc(u32 val)
  183. {
  184. /* upper 4bits and 7, 11 are write-as-0 */
  185. val &= 0xffff77f;
  186. asm volatile("mcr p14, 0, %0, c0, c0, 0" : : "r" (val));
  187. }
  188. static inline int
  189. xscale1_pmnc_counter_has_overflowed(unsigned long pmnc,
  190. enum xscale_counters counter)
  191. {
  192. int ret = 0;
  193. switch (counter) {
  194. case XSCALE_CYCLE_COUNTER:
  195. ret = pmnc & XSCALE1_CCOUNT_OVERFLOW;
  196. break;
  197. case XSCALE_COUNTER0:
  198. ret = pmnc & XSCALE1_COUNT0_OVERFLOW;
  199. break;
  200. case XSCALE_COUNTER1:
  201. ret = pmnc & XSCALE1_COUNT1_OVERFLOW;
  202. break;
  203. default:
  204. WARN_ONCE(1, "invalid counter number (%d)\n", counter);
  205. }
  206. return ret;
  207. }
  208. static irqreturn_t
  209. xscale1pmu_handle_irq(int irq_num, void *dev)
  210. {
  211. unsigned long pmnc;
  212. struct perf_sample_data data;
  213. struct cpu_hw_events *cpuc;
  214. struct pt_regs *regs;
  215. int idx;
  216. /*
  217. * NOTE: there's an A stepping erratum that states if an overflow
  218. * bit already exists and another occurs, the previous
  219. * Overflow bit gets cleared. There's no workaround.
  220. * Fixed in B stepping or later.
  221. */
  222. pmnc = xscale1pmu_read_pmnc();
  223. /*
  224. * Write the value back to clear the overflow flags. Overflow
  225. * flags remain in pmnc for use below. We also disable the PMU
  226. * while we process the interrupt.
  227. */
  228. xscale1pmu_write_pmnc(pmnc & ~XSCALE_PMU_ENABLE);
  229. if (!(pmnc & XSCALE1_OVERFLOWED_MASK))
  230. return IRQ_NONE;
  231. regs = get_irq_regs();
  232. perf_sample_data_init(&data, 0);
  233. cpuc = &__get_cpu_var(cpu_hw_events);
  234. for (idx = 0; idx < armpmu->num_events; ++idx) {
  235. struct perf_event *event = cpuc->events[idx];
  236. struct hw_perf_event *hwc;
  237. if (!xscale1_pmnc_counter_has_overflowed(pmnc, idx))
  238. continue;
  239. hwc = &event->hw;
  240. armpmu_event_update(event, hwc, idx, 1);
  241. data.period = event->hw.last_period;
  242. if (!armpmu_event_set_period(event, hwc, idx))
  243. continue;
  244. if (perf_event_overflow(event, &data, regs))
  245. armpmu->disable(hwc, idx);
  246. }
  247. irq_work_run();
  248. /*
  249. * Re-enable the PMU.
  250. */
  251. pmnc = xscale1pmu_read_pmnc() | XSCALE_PMU_ENABLE;
  252. xscale1pmu_write_pmnc(pmnc);
  253. return IRQ_HANDLED;
  254. }
  255. static void
  256. xscale1pmu_enable_event(struct hw_perf_event *hwc, int idx)
  257. {
  258. unsigned long val, mask, evt, flags;
  259. struct cpu_hw_events *events = armpmu->get_hw_events();
  260. switch (idx) {
  261. case XSCALE_CYCLE_COUNTER:
  262. mask = 0;
  263. evt = XSCALE1_CCOUNT_INT_EN;
  264. break;
  265. case XSCALE_COUNTER0:
  266. mask = XSCALE1_COUNT0_EVT_MASK;
  267. evt = (hwc->config_base << XSCALE1_COUNT0_EVT_SHFT) |
  268. XSCALE1_COUNT0_INT_EN;
  269. break;
  270. case XSCALE_COUNTER1:
  271. mask = XSCALE1_COUNT1_EVT_MASK;
  272. evt = (hwc->config_base << XSCALE1_COUNT1_EVT_SHFT) |
  273. XSCALE1_COUNT1_INT_EN;
  274. break;
  275. default:
  276. WARN_ONCE(1, "invalid counter number (%d)\n", idx);
  277. return;
  278. }
  279. raw_spin_lock_irqsave(&events->pmu_lock, flags);
  280. val = xscale1pmu_read_pmnc();
  281. val &= ~mask;
  282. val |= evt;
  283. xscale1pmu_write_pmnc(val);
  284. raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
  285. }
  286. static void
  287. xscale1pmu_disable_event(struct hw_perf_event *hwc, int idx)
  288. {
  289. unsigned long val, mask, evt, flags;
  290. struct cpu_hw_events *events = armpmu->get_hw_events();
  291. switch (idx) {
  292. case XSCALE_CYCLE_COUNTER:
  293. mask = XSCALE1_CCOUNT_INT_EN;
  294. evt = 0;
  295. break;
  296. case XSCALE_COUNTER0:
  297. mask = XSCALE1_COUNT0_INT_EN | XSCALE1_COUNT0_EVT_MASK;
  298. evt = XSCALE_PERFCTR_UNUSED << XSCALE1_COUNT0_EVT_SHFT;
  299. break;
  300. case XSCALE_COUNTER1:
  301. mask = XSCALE1_COUNT1_INT_EN | XSCALE1_COUNT1_EVT_MASK;
  302. evt = XSCALE_PERFCTR_UNUSED << XSCALE1_COUNT1_EVT_SHFT;
  303. break;
  304. default:
  305. WARN_ONCE(1, "invalid counter number (%d)\n", idx);
  306. return;
  307. }
  308. raw_spin_lock_irqsave(&events->pmu_lock, flags);
  309. val = xscale1pmu_read_pmnc();
  310. val &= ~mask;
  311. val |= evt;
  312. xscale1pmu_write_pmnc(val);
  313. raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
  314. }
  315. static int
  316. xscale1pmu_get_event_idx(struct cpu_hw_events *cpuc,
  317. struct hw_perf_event *event)
  318. {
  319. if (XSCALE_PERFCTR_CCNT == event->config_base) {
  320. if (test_and_set_bit(XSCALE_CYCLE_COUNTER, cpuc->used_mask))
  321. return -EAGAIN;
  322. return XSCALE_CYCLE_COUNTER;
  323. } else {
  324. if (!test_and_set_bit(XSCALE_COUNTER1, cpuc->used_mask))
  325. return XSCALE_COUNTER1;
  326. if (!test_and_set_bit(XSCALE_COUNTER0, cpuc->used_mask))
  327. return XSCALE_COUNTER0;
  328. return -EAGAIN;
  329. }
  330. }
  331. static void
  332. xscale1pmu_start(void)
  333. {
  334. unsigned long flags, val;
  335. struct cpu_hw_events *events = armpmu->get_hw_events();
  336. raw_spin_lock_irqsave(&events->pmu_lock, flags);
  337. val = xscale1pmu_read_pmnc();
  338. val |= XSCALE_PMU_ENABLE;
  339. xscale1pmu_write_pmnc(val);
  340. raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
  341. }
  342. static void
  343. xscale1pmu_stop(void)
  344. {
  345. unsigned long flags, val;
  346. struct cpu_hw_events *events = armpmu->get_hw_events();
  347. raw_spin_lock_irqsave(&events->pmu_lock, flags);
  348. val = xscale1pmu_read_pmnc();
  349. val &= ~XSCALE_PMU_ENABLE;
  350. xscale1pmu_write_pmnc(val);
  351. raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
  352. }
  353. static inline u32
  354. xscale1pmu_read_counter(int counter)
  355. {
  356. u32 val = 0;
  357. switch (counter) {
  358. case XSCALE_CYCLE_COUNTER:
  359. asm volatile("mrc p14, 0, %0, c1, c0, 0" : "=r" (val));
  360. break;
  361. case XSCALE_COUNTER0:
  362. asm volatile("mrc p14, 0, %0, c2, c0, 0" : "=r" (val));
  363. break;
  364. case XSCALE_COUNTER1:
  365. asm volatile("mrc p14, 0, %0, c3, c0, 0" : "=r" (val));
  366. break;
  367. }
  368. return val;
  369. }
  370. static inline void
  371. xscale1pmu_write_counter(int counter, u32 val)
  372. {
  373. switch (counter) {
  374. case XSCALE_CYCLE_COUNTER:
  375. asm volatile("mcr p14, 0, %0, c1, c0, 0" : : "r" (val));
  376. break;
  377. case XSCALE_COUNTER0:
  378. asm volatile("mcr p14, 0, %0, c2, c0, 0" : : "r" (val));
  379. break;
  380. case XSCALE_COUNTER1:
  381. asm volatile("mcr p14, 0, %0, c3, c0, 0" : : "r" (val));
  382. break;
  383. }
  384. }
  385. static struct arm_pmu xscale1pmu = {
  386. .id = ARM_PERF_PMU_ID_XSCALE1,
  387. .name = "xscale1",
  388. .handle_irq = xscale1pmu_handle_irq,
  389. .enable = xscale1pmu_enable_event,
  390. .disable = xscale1pmu_disable_event,
  391. .read_counter = xscale1pmu_read_counter,
  392. .write_counter = xscale1pmu_write_counter,
  393. .get_event_idx = xscale1pmu_get_event_idx,
  394. .start = xscale1pmu_start,
  395. .stop = xscale1pmu_stop,
  396. .cache_map = &xscale_perf_cache_map,
  397. .event_map = &xscale_perf_map,
  398. .raw_event_mask = 0xFF,
  399. .num_events = 3,
  400. .max_period = (1LLU << 32) - 1,
  401. };
  402. static struct arm_pmu *__init xscale1pmu_init(void)
  403. {
  404. return &xscale1pmu;
  405. }
  406. #define XSCALE2_OVERFLOWED_MASK 0x01f
  407. #define XSCALE2_CCOUNT_OVERFLOW 0x001
  408. #define XSCALE2_COUNT0_OVERFLOW 0x002
  409. #define XSCALE2_COUNT1_OVERFLOW 0x004
  410. #define XSCALE2_COUNT2_OVERFLOW 0x008
  411. #define XSCALE2_COUNT3_OVERFLOW 0x010
  412. #define XSCALE2_CCOUNT_INT_EN 0x001
  413. #define XSCALE2_COUNT0_INT_EN 0x002
  414. #define XSCALE2_COUNT1_INT_EN 0x004
  415. #define XSCALE2_COUNT2_INT_EN 0x008
  416. #define XSCALE2_COUNT3_INT_EN 0x010
  417. #define XSCALE2_COUNT0_EVT_SHFT 0
  418. #define XSCALE2_COUNT0_EVT_MASK (0xff << XSCALE2_COUNT0_EVT_SHFT)
  419. #define XSCALE2_COUNT1_EVT_SHFT 8
  420. #define XSCALE2_COUNT1_EVT_MASK (0xff << XSCALE2_COUNT1_EVT_SHFT)
  421. #define XSCALE2_COUNT2_EVT_SHFT 16
  422. #define XSCALE2_COUNT2_EVT_MASK (0xff << XSCALE2_COUNT2_EVT_SHFT)
  423. #define XSCALE2_COUNT3_EVT_SHFT 24
  424. #define XSCALE2_COUNT3_EVT_MASK (0xff << XSCALE2_COUNT3_EVT_SHFT)
  425. static inline u32
  426. xscale2pmu_read_pmnc(void)
  427. {
  428. u32 val;
  429. asm volatile("mrc p14, 0, %0, c0, c1, 0" : "=r" (val));
  430. /* bits 1-2 and 4-23 are read-unpredictable */
  431. return val & 0xff000009;
  432. }
  433. static inline void
  434. xscale2pmu_write_pmnc(u32 val)
  435. {
  436. /* bits 4-23 are write-as-0, 24-31 are write ignored */
  437. val &= 0xf;
  438. asm volatile("mcr p14, 0, %0, c0, c1, 0" : : "r" (val));
  439. }
  440. static inline u32
  441. xscale2pmu_read_overflow_flags(void)
  442. {
  443. u32 val;
  444. asm volatile("mrc p14, 0, %0, c5, c1, 0" : "=r" (val));
  445. return val;
  446. }
  447. static inline void
  448. xscale2pmu_write_overflow_flags(u32 val)
  449. {
  450. asm volatile("mcr p14, 0, %0, c5, c1, 0" : : "r" (val));
  451. }
  452. static inline u32
  453. xscale2pmu_read_event_select(void)
  454. {
  455. u32 val;
  456. asm volatile("mrc p14, 0, %0, c8, c1, 0" : "=r" (val));
  457. return val;
  458. }
  459. static inline void
  460. xscale2pmu_write_event_select(u32 val)
  461. {
  462. asm volatile("mcr p14, 0, %0, c8, c1, 0" : : "r"(val));
  463. }
  464. static inline u32
  465. xscale2pmu_read_int_enable(void)
  466. {
  467. u32 val;
  468. asm volatile("mrc p14, 0, %0, c4, c1, 0" : "=r" (val));
  469. return val;
  470. }
  471. static void
  472. xscale2pmu_write_int_enable(u32 val)
  473. {
  474. asm volatile("mcr p14, 0, %0, c4, c1, 0" : : "r" (val));
  475. }
  476. static inline int
  477. xscale2_pmnc_counter_has_overflowed(unsigned long of_flags,
  478. enum xscale_counters counter)
  479. {
  480. int ret = 0;
  481. switch (counter) {
  482. case XSCALE_CYCLE_COUNTER:
  483. ret = of_flags & XSCALE2_CCOUNT_OVERFLOW;
  484. break;
  485. case XSCALE_COUNTER0:
  486. ret = of_flags & XSCALE2_COUNT0_OVERFLOW;
  487. break;
  488. case XSCALE_COUNTER1:
  489. ret = of_flags & XSCALE2_COUNT1_OVERFLOW;
  490. break;
  491. case XSCALE_COUNTER2:
  492. ret = of_flags & XSCALE2_COUNT2_OVERFLOW;
  493. break;
  494. case XSCALE_COUNTER3:
  495. ret = of_flags & XSCALE2_COUNT3_OVERFLOW;
  496. break;
  497. default:
  498. WARN_ONCE(1, "invalid counter number (%d)\n", counter);
  499. }
  500. return ret;
  501. }
  502. static irqreturn_t
  503. xscale2pmu_handle_irq(int irq_num, void *dev)
  504. {
  505. unsigned long pmnc, of_flags;
  506. struct perf_sample_data data;
  507. struct cpu_hw_events *cpuc;
  508. struct pt_regs *regs;
  509. int idx;
  510. /* Disable the PMU. */
  511. pmnc = xscale2pmu_read_pmnc();
  512. xscale2pmu_write_pmnc(pmnc & ~XSCALE_PMU_ENABLE);
  513. /* Check the overflow flag register. */
  514. of_flags = xscale2pmu_read_overflow_flags();
  515. if (!(of_flags & XSCALE2_OVERFLOWED_MASK))
  516. return IRQ_NONE;
  517. /* Clear the overflow bits. */
  518. xscale2pmu_write_overflow_flags(of_flags);
  519. regs = get_irq_regs();
  520. perf_sample_data_init(&data, 0);
  521. cpuc = &__get_cpu_var(cpu_hw_events);
  522. for (idx = 0; idx < armpmu->num_events; ++idx) {
  523. struct perf_event *event = cpuc->events[idx];
  524. struct hw_perf_event *hwc;
  525. if (!xscale2_pmnc_counter_has_overflowed(pmnc, idx))
  526. continue;
  527. hwc = &event->hw;
  528. armpmu_event_update(event, hwc, idx, 1);
  529. data.period = event->hw.last_period;
  530. if (!armpmu_event_set_period(event, hwc, idx))
  531. continue;
  532. if (perf_event_overflow(event, &data, regs))
  533. armpmu->disable(hwc, idx);
  534. }
  535. irq_work_run();
  536. /*
  537. * Re-enable the PMU.
  538. */
  539. pmnc = xscale2pmu_read_pmnc() | XSCALE_PMU_ENABLE;
  540. xscale2pmu_write_pmnc(pmnc);
  541. return IRQ_HANDLED;
  542. }
  543. static void
  544. xscale2pmu_enable_event(struct hw_perf_event *hwc, int idx)
  545. {
  546. unsigned long flags, ien, evtsel;
  547. struct cpu_hw_events *events = armpmu->get_hw_events();
  548. ien = xscale2pmu_read_int_enable();
  549. evtsel = xscale2pmu_read_event_select();
  550. switch (idx) {
  551. case XSCALE_CYCLE_COUNTER:
  552. ien |= XSCALE2_CCOUNT_INT_EN;
  553. break;
  554. case XSCALE_COUNTER0:
  555. ien |= XSCALE2_COUNT0_INT_EN;
  556. evtsel &= ~XSCALE2_COUNT0_EVT_MASK;
  557. evtsel |= hwc->config_base << XSCALE2_COUNT0_EVT_SHFT;
  558. break;
  559. case XSCALE_COUNTER1:
  560. ien |= XSCALE2_COUNT1_INT_EN;
  561. evtsel &= ~XSCALE2_COUNT1_EVT_MASK;
  562. evtsel |= hwc->config_base << XSCALE2_COUNT1_EVT_SHFT;
  563. break;
  564. case XSCALE_COUNTER2:
  565. ien |= XSCALE2_COUNT2_INT_EN;
  566. evtsel &= ~XSCALE2_COUNT2_EVT_MASK;
  567. evtsel |= hwc->config_base << XSCALE2_COUNT2_EVT_SHFT;
  568. break;
  569. case XSCALE_COUNTER3:
  570. ien |= XSCALE2_COUNT3_INT_EN;
  571. evtsel &= ~XSCALE2_COUNT3_EVT_MASK;
  572. evtsel |= hwc->config_base << XSCALE2_COUNT3_EVT_SHFT;
  573. break;
  574. default:
  575. WARN_ONCE(1, "invalid counter number (%d)\n", idx);
  576. return;
  577. }
  578. raw_spin_lock_irqsave(&events->pmu_lock, flags);
  579. xscale2pmu_write_event_select(evtsel);
  580. xscale2pmu_write_int_enable(ien);
  581. raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
  582. }
  583. static void
  584. xscale2pmu_disable_event(struct hw_perf_event *hwc, int idx)
  585. {
  586. unsigned long flags, ien, evtsel;
  587. struct cpu_hw_events *events = armpmu->get_hw_events();
  588. ien = xscale2pmu_read_int_enable();
  589. evtsel = xscale2pmu_read_event_select();
  590. switch (idx) {
  591. case XSCALE_CYCLE_COUNTER:
  592. ien &= ~XSCALE2_CCOUNT_INT_EN;
  593. break;
  594. case XSCALE_COUNTER0:
  595. ien &= ~XSCALE2_COUNT0_INT_EN;
  596. evtsel &= ~XSCALE2_COUNT0_EVT_MASK;
  597. evtsel |= XSCALE_PERFCTR_UNUSED << XSCALE2_COUNT0_EVT_SHFT;
  598. break;
  599. case XSCALE_COUNTER1:
  600. ien &= ~XSCALE2_COUNT1_INT_EN;
  601. evtsel &= ~XSCALE2_COUNT1_EVT_MASK;
  602. evtsel |= XSCALE_PERFCTR_UNUSED << XSCALE2_COUNT1_EVT_SHFT;
  603. break;
  604. case XSCALE_COUNTER2:
  605. ien &= ~XSCALE2_COUNT2_INT_EN;
  606. evtsel &= ~XSCALE2_COUNT2_EVT_MASK;
  607. evtsel |= XSCALE_PERFCTR_UNUSED << XSCALE2_COUNT2_EVT_SHFT;
  608. break;
  609. case XSCALE_COUNTER3:
  610. ien &= ~XSCALE2_COUNT3_INT_EN;
  611. evtsel &= ~XSCALE2_COUNT3_EVT_MASK;
  612. evtsel |= XSCALE_PERFCTR_UNUSED << XSCALE2_COUNT3_EVT_SHFT;
  613. break;
  614. default:
  615. WARN_ONCE(1, "invalid counter number (%d)\n", idx);
  616. return;
  617. }
  618. raw_spin_lock_irqsave(&events->pmu_lock, flags);
  619. xscale2pmu_write_event_select(evtsel);
  620. xscale2pmu_write_int_enable(ien);
  621. raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
  622. }
  623. static int
  624. xscale2pmu_get_event_idx(struct cpu_hw_events *cpuc,
  625. struct hw_perf_event *event)
  626. {
  627. int idx = xscale1pmu_get_event_idx(cpuc, event);
  628. if (idx >= 0)
  629. goto out;
  630. if (!test_and_set_bit(XSCALE_COUNTER3, cpuc->used_mask))
  631. idx = XSCALE_COUNTER3;
  632. else if (!test_and_set_bit(XSCALE_COUNTER2, cpuc->used_mask))
  633. idx = XSCALE_COUNTER2;
  634. out:
  635. return idx;
  636. }
  637. static void
  638. xscale2pmu_start(void)
  639. {
  640. unsigned long flags, val;
  641. struct cpu_hw_events *events = armpmu->get_hw_events();
  642. raw_spin_lock_irqsave(&events->pmu_lock, flags);
  643. val = xscale2pmu_read_pmnc() & ~XSCALE_PMU_CNT64;
  644. val |= XSCALE_PMU_ENABLE;
  645. xscale2pmu_write_pmnc(val);
  646. raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
  647. }
  648. static void
  649. xscale2pmu_stop(void)
  650. {
  651. unsigned long flags, val;
  652. struct cpu_hw_events *events = armpmu->get_hw_events();
  653. raw_spin_lock_irqsave(&events->pmu_lock, flags);
  654. val = xscale2pmu_read_pmnc();
  655. val &= ~XSCALE_PMU_ENABLE;
  656. xscale2pmu_write_pmnc(val);
  657. raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
  658. }
  659. static inline u32
  660. xscale2pmu_read_counter(int counter)
  661. {
  662. u32 val = 0;
  663. switch (counter) {
  664. case XSCALE_CYCLE_COUNTER:
  665. asm volatile("mrc p14, 0, %0, c1, c1, 0" : "=r" (val));
  666. break;
  667. case XSCALE_COUNTER0:
  668. asm volatile("mrc p14, 0, %0, c0, c2, 0" : "=r" (val));
  669. break;
  670. case XSCALE_COUNTER1:
  671. asm volatile("mrc p14, 0, %0, c1, c2, 0" : "=r" (val));
  672. break;
  673. case XSCALE_COUNTER2:
  674. asm volatile("mrc p14, 0, %0, c2, c2, 0" : "=r" (val));
  675. break;
  676. case XSCALE_COUNTER3:
  677. asm volatile("mrc p14, 0, %0, c3, c2, 0" : "=r" (val));
  678. break;
  679. }
  680. return val;
  681. }
  682. static inline void
  683. xscale2pmu_write_counter(int counter, u32 val)
  684. {
  685. switch (counter) {
  686. case XSCALE_CYCLE_COUNTER:
  687. asm volatile("mcr p14, 0, %0, c1, c1, 0" : : "r" (val));
  688. break;
  689. case XSCALE_COUNTER0:
  690. asm volatile("mcr p14, 0, %0, c0, c2, 0" : : "r" (val));
  691. break;
  692. case XSCALE_COUNTER1:
  693. asm volatile("mcr p14, 0, %0, c1, c2, 0" : : "r" (val));
  694. break;
  695. case XSCALE_COUNTER2:
  696. asm volatile("mcr p14, 0, %0, c2, c2, 0" : : "r" (val));
  697. break;
  698. case XSCALE_COUNTER3:
  699. asm volatile("mcr p14, 0, %0, c3, c2, 0" : : "r" (val));
  700. break;
  701. }
  702. }
  703. static struct arm_pmu xscale2pmu = {
  704. .id = ARM_PERF_PMU_ID_XSCALE2,
  705. .name = "xscale2",
  706. .handle_irq = xscale2pmu_handle_irq,
  707. .enable = xscale2pmu_enable_event,
  708. .disable = xscale2pmu_disable_event,
  709. .read_counter = xscale2pmu_read_counter,
  710. .write_counter = xscale2pmu_write_counter,
  711. .get_event_idx = xscale2pmu_get_event_idx,
  712. .start = xscale2pmu_start,
  713. .stop = xscale2pmu_stop,
  714. .cache_map = &xscale_perf_cache_map,
  715. .event_map = &xscale_perf_map,
  716. .raw_event_mask = 0xFF,
  717. .num_events = 5,
  718. .max_period = (1LLU << 32) - 1,
  719. };
  720. static struct arm_pmu *__init xscale2pmu_init(void)
  721. {
  722. return &xscale2pmu;
  723. }
  724. #else
  725. static struct arm_pmu *__init xscale1pmu_init(void)
  726. {
  727. return NULL;
  728. }
  729. static struct arm_pmu *__init xscale2pmu_init(void)
  730. {
  731. return NULL;
  732. }
  733. #endif /* CONFIG_CPU_XSCALE */