perf_event_xscale.c 22 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840
  1. /*
  2. * ARMv5 [xscale] Performance counter handling code.
  3. *
  4. * Copyright (C) 2010, ARM Ltd., Will Deacon <will.deacon@arm.com>
  5. *
  6. * Based on the previous xscale OProfile code.
  7. *
  8. * There are two variants of the xscale PMU that we support:
  9. * - xscale1pmu: 2 event counters and a cycle counter
  10. * - xscale2pmu: 4 event counters and a cycle counter
  11. * The two variants share event definitions, but have different
  12. * PMU structures.
  13. */
  14. #ifdef CONFIG_CPU_XSCALE
  15. enum xscale_perf_types {
  16. XSCALE_PERFCTR_ICACHE_MISS = 0x00,
  17. XSCALE_PERFCTR_ICACHE_NO_DELIVER = 0x01,
  18. XSCALE_PERFCTR_DATA_STALL = 0x02,
  19. XSCALE_PERFCTR_ITLB_MISS = 0x03,
  20. XSCALE_PERFCTR_DTLB_MISS = 0x04,
  21. XSCALE_PERFCTR_BRANCH = 0x05,
  22. XSCALE_PERFCTR_BRANCH_MISS = 0x06,
  23. XSCALE_PERFCTR_INSTRUCTION = 0x07,
  24. XSCALE_PERFCTR_DCACHE_FULL_STALL = 0x08,
  25. XSCALE_PERFCTR_DCACHE_FULL_STALL_CONTIG = 0x09,
  26. XSCALE_PERFCTR_DCACHE_ACCESS = 0x0A,
  27. XSCALE_PERFCTR_DCACHE_MISS = 0x0B,
  28. XSCALE_PERFCTR_DCACHE_WRITE_BACK = 0x0C,
  29. XSCALE_PERFCTR_PC_CHANGED = 0x0D,
  30. XSCALE_PERFCTR_BCU_REQUEST = 0x10,
  31. XSCALE_PERFCTR_BCU_FULL = 0x11,
  32. XSCALE_PERFCTR_BCU_DRAIN = 0x12,
  33. XSCALE_PERFCTR_BCU_ECC_NO_ELOG = 0x14,
  34. XSCALE_PERFCTR_BCU_1_BIT_ERR = 0x15,
  35. XSCALE_PERFCTR_RMW = 0x16,
  36. /* XSCALE_PERFCTR_CCNT is not hardware defined */
  37. XSCALE_PERFCTR_CCNT = 0xFE,
  38. XSCALE_PERFCTR_UNUSED = 0xFF,
  39. };
  40. enum xscale_counters {
  41. XSCALE_CYCLE_COUNTER = 0,
  42. XSCALE_COUNTER0,
  43. XSCALE_COUNTER1,
  44. XSCALE_COUNTER2,
  45. XSCALE_COUNTER3,
  46. };
  47. static const unsigned xscale_perf_map[PERF_COUNT_HW_MAX] = {
  48. [PERF_COUNT_HW_CPU_CYCLES] = XSCALE_PERFCTR_CCNT,
  49. [PERF_COUNT_HW_INSTRUCTIONS] = XSCALE_PERFCTR_INSTRUCTION,
  50. [PERF_COUNT_HW_CACHE_REFERENCES] = HW_OP_UNSUPPORTED,
  51. [PERF_COUNT_HW_CACHE_MISSES] = HW_OP_UNSUPPORTED,
  52. [PERF_COUNT_HW_BRANCH_INSTRUCTIONS] = XSCALE_PERFCTR_BRANCH,
  53. [PERF_COUNT_HW_BRANCH_MISSES] = XSCALE_PERFCTR_BRANCH_MISS,
  54. [PERF_COUNT_HW_BUS_CYCLES] = HW_OP_UNSUPPORTED,
  55. [PERF_COUNT_HW_STALLED_CYCLES_FRONTEND] = XSCALE_PERFCTR_ICACHE_NO_DELIVER,
  56. [PERF_COUNT_HW_STALLED_CYCLES_BACKEND] = HW_OP_UNSUPPORTED,
  57. };
  58. static const unsigned xscale_perf_cache_map[PERF_COUNT_HW_CACHE_MAX]
  59. [PERF_COUNT_HW_CACHE_OP_MAX]
  60. [PERF_COUNT_HW_CACHE_RESULT_MAX] = {
  61. [C(L1D)] = {
  62. [C(OP_READ)] = {
  63. [C(RESULT_ACCESS)] = XSCALE_PERFCTR_DCACHE_ACCESS,
  64. [C(RESULT_MISS)] = XSCALE_PERFCTR_DCACHE_MISS,
  65. },
  66. [C(OP_WRITE)] = {
  67. [C(RESULT_ACCESS)] = XSCALE_PERFCTR_DCACHE_ACCESS,
  68. [C(RESULT_MISS)] = XSCALE_PERFCTR_DCACHE_MISS,
  69. },
  70. [C(OP_PREFETCH)] = {
  71. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  72. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  73. },
  74. },
  75. [C(L1I)] = {
  76. [C(OP_READ)] = {
  77. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  78. [C(RESULT_MISS)] = XSCALE_PERFCTR_ICACHE_MISS,
  79. },
  80. [C(OP_WRITE)] = {
  81. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  82. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  83. },
  84. [C(OP_PREFETCH)] = {
  85. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  86. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  87. },
  88. },
  89. [C(LL)] = {
  90. [C(OP_READ)] = {
  91. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  92. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  93. },
  94. [C(OP_WRITE)] = {
  95. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  96. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  97. },
  98. [C(OP_PREFETCH)] = {
  99. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  100. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  101. },
  102. },
  103. [C(DTLB)] = {
  104. [C(OP_READ)] = {
  105. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  106. [C(RESULT_MISS)] = XSCALE_PERFCTR_DTLB_MISS,
  107. },
  108. [C(OP_WRITE)] = {
  109. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  110. [C(RESULT_MISS)] = XSCALE_PERFCTR_DTLB_MISS,
  111. },
  112. [C(OP_PREFETCH)] = {
  113. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  114. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  115. },
  116. },
  117. [C(ITLB)] = {
  118. [C(OP_READ)] = {
  119. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  120. [C(RESULT_MISS)] = XSCALE_PERFCTR_ITLB_MISS,
  121. },
  122. [C(OP_WRITE)] = {
  123. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  124. [C(RESULT_MISS)] = XSCALE_PERFCTR_ITLB_MISS,
  125. },
  126. [C(OP_PREFETCH)] = {
  127. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  128. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  129. },
  130. },
  131. [C(BPU)] = {
  132. [C(OP_READ)] = {
  133. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  134. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  135. },
  136. [C(OP_WRITE)] = {
  137. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  138. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  139. },
  140. [C(OP_PREFETCH)] = {
  141. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  142. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  143. },
  144. },
  145. [C(NODE)] = {
  146. [C(OP_READ)] = {
  147. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  148. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  149. },
  150. [C(OP_WRITE)] = {
  151. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  152. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  153. },
  154. [C(OP_PREFETCH)] = {
  155. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  156. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  157. },
  158. },
  159. };
  160. #define XSCALE_PMU_ENABLE 0x001
  161. #define XSCALE_PMN_RESET 0x002
  162. #define XSCALE_CCNT_RESET 0x004
  163. #define XSCALE_PMU_RESET (CCNT_RESET | PMN_RESET)
  164. #define XSCALE_PMU_CNT64 0x008
  165. #define XSCALE1_OVERFLOWED_MASK 0x700
  166. #define XSCALE1_CCOUNT_OVERFLOW 0x400
  167. #define XSCALE1_COUNT0_OVERFLOW 0x100
  168. #define XSCALE1_COUNT1_OVERFLOW 0x200
  169. #define XSCALE1_CCOUNT_INT_EN 0x040
  170. #define XSCALE1_COUNT0_INT_EN 0x010
  171. #define XSCALE1_COUNT1_INT_EN 0x020
  172. #define XSCALE1_COUNT0_EVT_SHFT 12
  173. #define XSCALE1_COUNT0_EVT_MASK (0xff << XSCALE1_COUNT0_EVT_SHFT)
  174. #define XSCALE1_COUNT1_EVT_SHFT 20
  175. #define XSCALE1_COUNT1_EVT_MASK (0xff << XSCALE1_COUNT1_EVT_SHFT)
  176. static inline u32
  177. xscale1pmu_read_pmnc(void)
  178. {
  179. u32 val;
  180. asm volatile("mrc p14, 0, %0, c0, c0, 0" : "=r" (val));
  181. return val;
  182. }
  183. static inline void
  184. xscale1pmu_write_pmnc(u32 val)
  185. {
  186. /* upper 4bits and 7, 11 are write-as-0 */
  187. val &= 0xffff77f;
  188. asm volatile("mcr p14, 0, %0, c0, c0, 0" : : "r" (val));
  189. }
  190. static inline int
  191. xscale1_pmnc_counter_has_overflowed(unsigned long pmnc,
  192. enum xscale_counters counter)
  193. {
  194. int ret = 0;
  195. switch (counter) {
  196. case XSCALE_CYCLE_COUNTER:
  197. ret = pmnc & XSCALE1_CCOUNT_OVERFLOW;
  198. break;
  199. case XSCALE_COUNTER0:
  200. ret = pmnc & XSCALE1_COUNT0_OVERFLOW;
  201. break;
  202. case XSCALE_COUNTER1:
  203. ret = pmnc & XSCALE1_COUNT1_OVERFLOW;
  204. break;
  205. default:
  206. WARN_ONCE(1, "invalid counter number (%d)\n", counter);
  207. }
  208. return ret;
  209. }
  210. static irqreturn_t
  211. xscale1pmu_handle_irq(int irq_num, void *dev)
  212. {
  213. unsigned long pmnc;
  214. struct perf_sample_data data;
  215. struct arm_pmu *cpu_pmu = (struct arm_pmu *)dev;
  216. struct pmu_hw_events *cpuc = cpu_pmu->get_hw_events();
  217. struct pt_regs *regs;
  218. int idx;
  219. /*
  220. * NOTE: there's an A stepping erratum that states if an overflow
  221. * bit already exists and another occurs, the previous
  222. * Overflow bit gets cleared. There's no workaround.
  223. * Fixed in B stepping or later.
  224. */
  225. pmnc = xscale1pmu_read_pmnc();
  226. /*
  227. * Write the value back to clear the overflow flags. Overflow
  228. * flags remain in pmnc for use below. We also disable the PMU
  229. * while we process the interrupt.
  230. */
  231. xscale1pmu_write_pmnc(pmnc & ~XSCALE_PMU_ENABLE);
  232. if (!(pmnc & XSCALE1_OVERFLOWED_MASK))
  233. return IRQ_NONE;
  234. regs = get_irq_regs();
  235. for (idx = 0; idx < cpu_pmu->num_events; ++idx) {
  236. struct perf_event *event = cpuc->events[idx];
  237. struct hw_perf_event *hwc;
  238. if (!event)
  239. continue;
  240. if (!xscale1_pmnc_counter_has_overflowed(pmnc, idx))
  241. continue;
  242. hwc = &event->hw;
  243. armpmu_event_update(event);
  244. perf_sample_data_init(&data, 0, hwc->last_period);
  245. if (!armpmu_event_set_period(event))
  246. continue;
  247. if (perf_event_overflow(event, &data, regs))
  248. cpu_pmu->disable(event);
  249. }
  250. irq_work_run();
  251. /*
  252. * Re-enable the PMU.
  253. */
  254. pmnc = xscale1pmu_read_pmnc() | XSCALE_PMU_ENABLE;
  255. xscale1pmu_write_pmnc(pmnc);
  256. return IRQ_HANDLED;
  257. }
  258. static void xscale1pmu_enable_event(struct perf_event *event)
  259. {
  260. unsigned long val, mask, evt, flags;
  261. struct arm_pmu *cpu_pmu = to_arm_pmu(event->pmu);
  262. struct hw_perf_event *hwc = &event->hw;
  263. struct pmu_hw_events *events = cpu_pmu->get_hw_events();
  264. int idx = hwc->idx;
  265. switch (idx) {
  266. case XSCALE_CYCLE_COUNTER:
  267. mask = 0;
  268. evt = XSCALE1_CCOUNT_INT_EN;
  269. break;
  270. case XSCALE_COUNTER0:
  271. mask = XSCALE1_COUNT0_EVT_MASK;
  272. evt = (hwc->config_base << XSCALE1_COUNT0_EVT_SHFT) |
  273. XSCALE1_COUNT0_INT_EN;
  274. break;
  275. case XSCALE_COUNTER1:
  276. mask = XSCALE1_COUNT1_EVT_MASK;
  277. evt = (hwc->config_base << XSCALE1_COUNT1_EVT_SHFT) |
  278. XSCALE1_COUNT1_INT_EN;
  279. break;
  280. default:
  281. WARN_ONCE(1, "invalid counter number (%d)\n", idx);
  282. return;
  283. }
  284. raw_spin_lock_irqsave(&events->pmu_lock, flags);
  285. val = xscale1pmu_read_pmnc();
  286. val &= ~mask;
  287. val |= evt;
  288. xscale1pmu_write_pmnc(val);
  289. raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
  290. }
  291. static void xscale1pmu_disable_event(struct perf_event *event)
  292. {
  293. unsigned long val, mask, evt, flags;
  294. struct arm_pmu *cpu_pmu = to_arm_pmu(event->pmu);
  295. struct hw_perf_event *hwc = &event->hw;
  296. struct pmu_hw_events *events = cpu_pmu->get_hw_events();
  297. int idx = hwc->idx;
  298. switch (idx) {
  299. case XSCALE_CYCLE_COUNTER:
  300. mask = XSCALE1_CCOUNT_INT_EN;
  301. evt = 0;
  302. break;
  303. case XSCALE_COUNTER0:
  304. mask = XSCALE1_COUNT0_INT_EN | XSCALE1_COUNT0_EVT_MASK;
  305. evt = XSCALE_PERFCTR_UNUSED << XSCALE1_COUNT0_EVT_SHFT;
  306. break;
  307. case XSCALE_COUNTER1:
  308. mask = XSCALE1_COUNT1_INT_EN | XSCALE1_COUNT1_EVT_MASK;
  309. evt = XSCALE_PERFCTR_UNUSED << XSCALE1_COUNT1_EVT_SHFT;
  310. break;
  311. default:
  312. WARN_ONCE(1, "invalid counter number (%d)\n", idx);
  313. return;
  314. }
  315. raw_spin_lock_irqsave(&events->pmu_lock, flags);
  316. val = xscale1pmu_read_pmnc();
  317. val &= ~mask;
  318. val |= evt;
  319. xscale1pmu_write_pmnc(val);
  320. raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
  321. }
  322. static int
  323. xscale1pmu_get_event_idx(struct pmu_hw_events *cpuc,
  324. struct perf_event *event)
  325. {
  326. struct hw_perf_event *hwc = &event->hw;
  327. if (XSCALE_PERFCTR_CCNT == hwc->config_base) {
  328. if (test_and_set_bit(XSCALE_CYCLE_COUNTER, cpuc->used_mask))
  329. return -EAGAIN;
  330. return XSCALE_CYCLE_COUNTER;
  331. } else {
  332. if (!test_and_set_bit(XSCALE_COUNTER1, cpuc->used_mask))
  333. return XSCALE_COUNTER1;
  334. if (!test_and_set_bit(XSCALE_COUNTER0, cpuc->used_mask))
  335. return XSCALE_COUNTER0;
  336. return -EAGAIN;
  337. }
  338. }
  339. static void xscale1pmu_start(struct arm_pmu *cpu_pmu)
  340. {
  341. unsigned long flags, val;
  342. struct pmu_hw_events *events = cpu_pmu->get_hw_events();
  343. raw_spin_lock_irqsave(&events->pmu_lock, flags);
  344. val = xscale1pmu_read_pmnc();
  345. val |= XSCALE_PMU_ENABLE;
  346. xscale1pmu_write_pmnc(val);
  347. raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
  348. }
  349. static void xscale1pmu_stop(struct arm_pmu *cpu_pmu)
  350. {
  351. unsigned long flags, val;
  352. struct pmu_hw_events *events = cpu_pmu->get_hw_events();
  353. raw_spin_lock_irqsave(&events->pmu_lock, flags);
  354. val = xscale1pmu_read_pmnc();
  355. val &= ~XSCALE_PMU_ENABLE;
  356. xscale1pmu_write_pmnc(val);
  357. raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
  358. }
  359. static inline u32 xscale1pmu_read_counter(struct perf_event *event)
  360. {
  361. struct hw_perf_event *hwc = &event->hw;
  362. int counter = hwc->idx;
  363. u32 val = 0;
  364. switch (counter) {
  365. case XSCALE_CYCLE_COUNTER:
  366. asm volatile("mrc p14, 0, %0, c1, c0, 0" : "=r" (val));
  367. break;
  368. case XSCALE_COUNTER0:
  369. asm volatile("mrc p14, 0, %0, c2, c0, 0" : "=r" (val));
  370. break;
  371. case XSCALE_COUNTER1:
  372. asm volatile("mrc p14, 0, %0, c3, c0, 0" : "=r" (val));
  373. break;
  374. }
  375. return val;
  376. }
  377. static inline void xscale1pmu_write_counter(struct perf_event *event, u32 val)
  378. {
  379. struct hw_perf_event *hwc = &event->hw;
  380. int counter = hwc->idx;
  381. switch (counter) {
  382. case XSCALE_CYCLE_COUNTER:
  383. asm volatile("mcr p14, 0, %0, c1, c0, 0" : : "r" (val));
  384. break;
  385. case XSCALE_COUNTER0:
  386. asm volatile("mcr p14, 0, %0, c2, c0, 0" : : "r" (val));
  387. break;
  388. case XSCALE_COUNTER1:
  389. asm volatile("mcr p14, 0, %0, c3, c0, 0" : : "r" (val));
  390. break;
  391. }
  392. }
  393. static int xscale_map_event(struct perf_event *event)
  394. {
  395. return armpmu_map_event(event, &xscale_perf_map,
  396. &xscale_perf_cache_map, 0xFF);
  397. }
  398. static int xscale1pmu_init(struct arm_pmu *cpu_pmu)
  399. {
  400. cpu_pmu->name = "xscale1";
  401. cpu_pmu->handle_irq = xscale1pmu_handle_irq;
  402. cpu_pmu->enable = xscale1pmu_enable_event;
  403. cpu_pmu->disable = xscale1pmu_disable_event;
  404. cpu_pmu->read_counter = xscale1pmu_read_counter;
  405. cpu_pmu->write_counter = xscale1pmu_write_counter;
  406. cpu_pmu->get_event_idx = xscale1pmu_get_event_idx;
  407. cpu_pmu->start = xscale1pmu_start;
  408. cpu_pmu->stop = xscale1pmu_stop;
  409. cpu_pmu->map_event = xscale_map_event;
  410. cpu_pmu->num_events = 3;
  411. cpu_pmu->max_period = (1LLU << 32) - 1;
  412. return 0;
  413. }
  414. #define XSCALE2_OVERFLOWED_MASK 0x01f
  415. #define XSCALE2_CCOUNT_OVERFLOW 0x001
  416. #define XSCALE2_COUNT0_OVERFLOW 0x002
  417. #define XSCALE2_COUNT1_OVERFLOW 0x004
  418. #define XSCALE2_COUNT2_OVERFLOW 0x008
  419. #define XSCALE2_COUNT3_OVERFLOW 0x010
  420. #define XSCALE2_CCOUNT_INT_EN 0x001
  421. #define XSCALE2_COUNT0_INT_EN 0x002
  422. #define XSCALE2_COUNT1_INT_EN 0x004
  423. #define XSCALE2_COUNT2_INT_EN 0x008
  424. #define XSCALE2_COUNT3_INT_EN 0x010
  425. #define XSCALE2_COUNT0_EVT_SHFT 0
  426. #define XSCALE2_COUNT0_EVT_MASK (0xff << XSCALE2_COUNT0_EVT_SHFT)
  427. #define XSCALE2_COUNT1_EVT_SHFT 8
  428. #define XSCALE2_COUNT1_EVT_MASK (0xff << XSCALE2_COUNT1_EVT_SHFT)
  429. #define XSCALE2_COUNT2_EVT_SHFT 16
  430. #define XSCALE2_COUNT2_EVT_MASK (0xff << XSCALE2_COUNT2_EVT_SHFT)
  431. #define XSCALE2_COUNT3_EVT_SHFT 24
  432. #define XSCALE2_COUNT3_EVT_MASK (0xff << XSCALE2_COUNT3_EVT_SHFT)
  433. static inline u32
  434. xscale2pmu_read_pmnc(void)
  435. {
  436. u32 val;
  437. asm volatile("mrc p14, 0, %0, c0, c1, 0" : "=r" (val));
  438. /* bits 1-2 and 4-23 are read-unpredictable */
  439. return val & 0xff000009;
  440. }
  441. static inline void
  442. xscale2pmu_write_pmnc(u32 val)
  443. {
  444. /* bits 4-23 are write-as-0, 24-31 are write ignored */
  445. val &= 0xf;
  446. asm volatile("mcr p14, 0, %0, c0, c1, 0" : : "r" (val));
  447. }
  448. static inline u32
  449. xscale2pmu_read_overflow_flags(void)
  450. {
  451. u32 val;
  452. asm volatile("mrc p14, 0, %0, c5, c1, 0" : "=r" (val));
  453. return val;
  454. }
  455. static inline void
  456. xscale2pmu_write_overflow_flags(u32 val)
  457. {
  458. asm volatile("mcr p14, 0, %0, c5, c1, 0" : : "r" (val));
  459. }
  460. static inline u32
  461. xscale2pmu_read_event_select(void)
  462. {
  463. u32 val;
  464. asm volatile("mrc p14, 0, %0, c8, c1, 0" : "=r" (val));
  465. return val;
  466. }
  467. static inline void
  468. xscale2pmu_write_event_select(u32 val)
  469. {
  470. asm volatile("mcr p14, 0, %0, c8, c1, 0" : : "r"(val));
  471. }
  472. static inline u32
  473. xscale2pmu_read_int_enable(void)
  474. {
  475. u32 val;
  476. asm volatile("mrc p14, 0, %0, c4, c1, 0" : "=r" (val));
  477. return val;
  478. }
  479. static void
  480. xscale2pmu_write_int_enable(u32 val)
  481. {
  482. asm volatile("mcr p14, 0, %0, c4, c1, 0" : : "r" (val));
  483. }
  484. static inline int
  485. xscale2_pmnc_counter_has_overflowed(unsigned long of_flags,
  486. enum xscale_counters counter)
  487. {
  488. int ret = 0;
  489. switch (counter) {
  490. case XSCALE_CYCLE_COUNTER:
  491. ret = of_flags & XSCALE2_CCOUNT_OVERFLOW;
  492. break;
  493. case XSCALE_COUNTER0:
  494. ret = of_flags & XSCALE2_COUNT0_OVERFLOW;
  495. break;
  496. case XSCALE_COUNTER1:
  497. ret = of_flags & XSCALE2_COUNT1_OVERFLOW;
  498. break;
  499. case XSCALE_COUNTER2:
  500. ret = of_flags & XSCALE2_COUNT2_OVERFLOW;
  501. break;
  502. case XSCALE_COUNTER3:
  503. ret = of_flags & XSCALE2_COUNT3_OVERFLOW;
  504. break;
  505. default:
  506. WARN_ONCE(1, "invalid counter number (%d)\n", counter);
  507. }
  508. return ret;
  509. }
  510. static irqreturn_t
  511. xscale2pmu_handle_irq(int irq_num, void *dev)
  512. {
  513. unsigned long pmnc, of_flags;
  514. struct perf_sample_data data;
  515. struct arm_pmu *cpu_pmu = (struct arm_pmu *)dev;
  516. struct pmu_hw_events *cpuc = cpu_pmu->get_hw_events();
  517. struct pt_regs *regs;
  518. int idx;
  519. /* Disable the PMU. */
  520. pmnc = xscale2pmu_read_pmnc();
  521. xscale2pmu_write_pmnc(pmnc & ~XSCALE_PMU_ENABLE);
  522. /* Check the overflow flag register. */
  523. of_flags = xscale2pmu_read_overflow_flags();
  524. if (!(of_flags & XSCALE2_OVERFLOWED_MASK))
  525. return IRQ_NONE;
  526. /* Clear the overflow bits. */
  527. xscale2pmu_write_overflow_flags(of_flags);
  528. regs = get_irq_regs();
  529. for (idx = 0; idx < cpu_pmu->num_events; ++idx) {
  530. struct perf_event *event = cpuc->events[idx];
  531. struct hw_perf_event *hwc;
  532. if (!event)
  533. continue;
  534. if (!xscale2_pmnc_counter_has_overflowed(of_flags, idx))
  535. continue;
  536. hwc = &event->hw;
  537. armpmu_event_update(event);
  538. perf_sample_data_init(&data, 0, hwc->last_period);
  539. if (!armpmu_event_set_period(event))
  540. continue;
  541. if (perf_event_overflow(event, &data, regs))
  542. cpu_pmu->disable(event);
  543. }
  544. irq_work_run();
  545. /*
  546. * Re-enable the PMU.
  547. */
  548. pmnc = xscale2pmu_read_pmnc() | XSCALE_PMU_ENABLE;
  549. xscale2pmu_write_pmnc(pmnc);
  550. return IRQ_HANDLED;
  551. }
  552. static void xscale2pmu_enable_event(struct perf_event *event)
  553. {
  554. unsigned long flags, ien, evtsel;
  555. struct arm_pmu *cpu_pmu = to_arm_pmu(event->pmu);
  556. struct hw_perf_event *hwc = &event->hw;
  557. struct pmu_hw_events *events = cpu_pmu->get_hw_events();
  558. int idx = hwc->idx;
  559. ien = xscale2pmu_read_int_enable();
  560. evtsel = xscale2pmu_read_event_select();
  561. switch (idx) {
  562. case XSCALE_CYCLE_COUNTER:
  563. ien |= XSCALE2_CCOUNT_INT_EN;
  564. break;
  565. case XSCALE_COUNTER0:
  566. ien |= XSCALE2_COUNT0_INT_EN;
  567. evtsel &= ~XSCALE2_COUNT0_EVT_MASK;
  568. evtsel |= hwc->config_base << XSCALE2_COUNT0_EVT_SHFT;
  569. break;
  570. case XSCALE_COUNTER1:
  571. ien |= XSCALE2_COUNT1_INT_EN;
  572. evtsel &= ~XSCALE2_COUNT1_EVT_MASK;
  573. evtsel |= hwc->config_base << XSCALE2_COUNT1_EVT_SHFT;
  574. break;
  575. case XSCALE_COUNTER2:
  576. ien |= XSCALE2_COUNT2_INT_EN;
  577. evtsel &= ~XSCALE2_COUNT2_EVT_MASK;
  578. evtsel |= hwc->config_base << XSCALE2_COUNT2_EVT_SHFT;
  579. break;
  580. case XSCALE_COUNTER3:
  581. ien |= XSCALE2_COUNT3_INT_EN;
  582. evtsel &= ~XSCALE2_COUNT3_EVT_MASK;
  583. evtsel |= hwc->config_base << XSCALE2_COUNT3_EVT_SHFT;
  584. break;
  585. default:
  586. WARN_ONCE(1, "invalid counter number (%d)\n", idx);
  587. return;
  588. }
  589. raw_spin_lock_irqsave(&events->pmu_lock, flags);
  590. xscale2pmu_write_event_select(evtsel);
  591. xscale2pmu_write_int_enable(ien);
  592. raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
  593. }
  594. static void xscale2pmu_disable_event(struct perf_event *event)
  595. {
  596. unsigned long flags, ien, evtsel, of_flags;
  597. struct arm_pmu *cpu_pmu = to_arm_pmu(event->pmu);
  598. struct hw_perf_event *hwc = &event->hw;
  599. struct pmu_hw_events *events = cpu_pmu->get_hw_events();
  600. int idx = hwc->idx;
  601. ien = xscale2pmu_read_int_enable();
  602. evtsel = xscale2pmu_read_event_select();
  603. switch (idx) {
  604. case XSCALE_CYCLE_COUNTER:
  605. ien &= ~XSCALE2_CCOUNT_INT_EN;
  606. of_flags = XSCALE2_CCOUNT_OVERFLOW;
  607. break;
  608. case XSCALE_COUNTER0:
  609. ien &= ~XSCALE2_COUNT0_INT_EN;
  610. evtsel &= ~XSCALE2_COUNT0_EVT_MASK;
  611. evtsel |= XSCALE_PERFCTR_UNUSED << XSCALE2_COUNT0_EVT_SHFT;
  612. of_flags = XSCALE2_COUNT0_OVERFLOW;
  613. break;
  614. case XSCALE_COUNTER1:
  615. ien &= ~XSCALE2_COUNT1_INT_EN;
  616. evtsel &= ~XSCALE2_COUNT1_EVT_MASK;
  617. evtsel |= XSCALE_PERFCTR_UNUSED << XSCALE2_COUNT1_EVT_SHFT;
  618. of_flags = XSCALE2_COUNT1_OVERFLOW;
  619. break;
  620. case XSCALE_COUNTER2:
  621. ien &= ~XSCALE2_COUNT2_INT_EN;
  622. evtsel &= ~XSCALE2_COUNT2_EVT_MASK;
  623. evtsel |= XSCALE_PERFCTR_UNUSED << XSCALE2_COUNT2_EVT_SHFT;
  624. of_flags = XSCALE2_COUNT2_OVERFLOW;
  625. break;
  626. case XSCALE_COUNTER3:
  627. ien &= ~XSCALE2_COUNT3_INT_EN;
  628. evtsel &= ~XSCALE2_COUNT3_EVT_MASK;
  629. evtsel |= XSCALE_PERFCTR_UNUSED << XSCALE2_COUNT3_EVT_SHFT;
  630. of_flags = XSCALE2_COUNT3_OVERFLOW;
  631. break;
  632. default:
  633. WARN_ONCE(1, "invalid counter number (%d)\n", idx);
  634. return;
  635. }
  636. raw_spin_lock_irqsave(&events->pmu_lock, flags);
  637. xscale2pmu_write_event_select(evtsel);
  638. xscale2pmu_write_int_enable(ien);
  639. xscale2pmu_write_overflow_flags(of_flags);
  640. raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
  641. }
  642. static int
  643. xscale2pmu_get_event_idx(struct pmu_hw_events *cpuc,
  644. struct perf_event *event)
  645. {
  646. int idx = xscale1pmu_get_event_idx(cpuc, event);
  647. if (idx >= 0)
  648. goto out;
  649. if (!test_and_set_bit(XSCALE_COUNTER3, cpuc->used_mask))
  650. idx = XSCALE_COUNTER3;
  651. else if (!test_and_set_bit(XSCALE_COUNTER2, cpuc->used_mask))
  652. idx = XSCALE_COUNTER2;
  653. out:
  654. return idx;
  655. }
  656. static void xscale2pmu_start(struct arm_pmu *cpu_pmu)
  657. {
  658. unsigned long flags, val;
  659. struct pmu_hw_events *events = cpu_pmu->get_hw_events();
  660. raw_spin_lock_irqsave(&events->pmu_lock, flags);
  661. val = xscale2pmu_read_pmnc() & ~XSCALE_PMU_CNT64;
  662. val |= XSCALE_PMU_ENABLE;
  663. xscale2pmu_write_pmnc(val);
  664. raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
  665. }
  666. static void xscale2pmu_stop(struct arm_pmu *cpu_pmu)
  667. {
  668. unsigned long flags, val;
  669. struct pmu_hw_events *events = cpu_pmu->get_hw_events();
  670. raw_spin_lock_irqsave(&events->pmu_lock, flags);
  671. val = xscale2pmu_read_pmnc();
  672. val &= ~XSCALE_PMU_ENABLE;
  673. xscale2pmu_write_pmnc(val);
  674. raw_spin_unlock_irqrestore(&events->pmu_lock, flags);
  675. }
  676. static inline u32 xscale2pmu_read_counter(struct perf_event *event)
  677. {
  678. struct hw_perf_event *hwc = &event->hw;
  679. int counter = hwc->idx;
  680. u32 val = 0;
  681. switch (counter) {
  682. case XSCALE_CYCLE_COUNTER:
  683. asm volatile("mrc p14, 0, %0, c1, c1, 0" : "=r" (val));
  684. break;
  685. case XSCALE_COUNTER0:
  686. asm volatile("mrc p14, 0, %0, c0, c2, 0" : "=r" (val));
  687. break;
  688. case XSCALE_COUNTER1:
  689. asm volatile("mrc p14, 0, %0, c1, c2, 0" : "=r" (val));
  690. break;
  691. case XSCALE_COUNTER2:
  692. asm volatile("mrc p14, 0, %0, c2, c2, 0" : "=r" (val));
  693. break;
  694. case XSCALE_COUNTER3:
  695. asm volatile("mrc p14, 0, %0, c3, c2, 0" : "=r" (val));
  696. break;
  697. }
  698. return val;
  699. }
  700. static inline void xscale2pmu_write_counter(struct perf_event *event, u32 val)
  701. {
  702. struct hw_perf_event *hwc = &event->hw;
  703. int counter = hwc->idx;
  704. switch (counter) {
  705. case XSCALE_CYCLE_COUNTER:
  706. asm volatile("mcr p14, 0, %0, c1, c1, 0" : : "r" (val));
  707. break;
  708. case XSCALE_COUNTER0:
  709. asm volatile("mcr p14, 0, %0, c0, c2, 0" : : "r" (val));
  710. break;
  711. case XSCALE_COUNTER1:
  712. asm volatile("mcr p14, 0, %0, c1, c2, 0" : : "r" (val));
  713. break;
  714. case XSCALE_COUNTER2:
  715. asm volatile("mcr p14, 0, %0, c2, c2, 0" : : "r" (val));
  716. break;
  717. case XSCALE_COUNTER3:
  718. asm volatile("mcr p14, 0, %0, c3, c2, 0" : : "r" (val));
  719. break;
  720. }
  721. }
  722. static int xscale2pmu_init(struct arm_pmu *cpu_pmu)
  723. {
  724. cpu_pmu->name = "xscale2";
  725. cpu_pmu->handle_irq = xscale2pmu_handle_irq;
  726. cpu_pmu->enable = xscale2pmu_enable_event;
  727. cpu_pmu->disable = xscale2pmu_disable_event;
  728. cpu_pmu->read_counter = xscale2pmu_read_counter;
  729. cpu_pmu->write_counter = xscale2pmu_write_counter;
  730. cpu_pmu->get_event_idx = xscale2pmu_get_event_idx;
  731. cpu_pmu->start = xscale2pmu_start;
  732. cpu_pmu->stop = xscale2pmu_stop;
  733. cpu_pmu->map_event = xscale_map_event;
  734. cpu_pmu->num_events = 5;
  735. cpu_pmu->max_period = (1LLU << 32) - 1;
  736. return 0;
  737. }
  738. #else
  739. static inline int xscale1pmu_init(struct arm_pmu *cpu_pmu)
  740. {
  741. return -ENODEV;
  742. }
  743. static inline int xscale2pmu_init(struct arm_pmu *cpu_pmu)
  744. {
  745. return -ENODEV;
  746. }
  747. #endif /* CONFIG_CPU_XSCALE */