perf.h 1.5 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364
  1. #ifndef _PERF_PERF_H
  2. #define _PERF_PERF_H
  3. /*
  4. * prctl(PR_TASK_PERF_COUNTERS_DISABLE) will (cheaply) disable all
  5. * counters in the current task.
  6. */
  7. #define PR_TASK_PERF_COUNTERS_DISABLE 31
  8. #define PR_TASK_PERF_COUNTERS_ENABLE 32
  9. #define rdclock() \
  10. ({ \
  11. struct timespec ts; \
  12. \
  13. clock_gettime(CLOCK_MONOTONIC, &ts); \
  14. ts.tv_sec * 1000000000ULL + ts.tv_nsec; \
  15. })
  16. /*
  17. * Pick up some kernel type conventions:
  18. */
  19. #define __user
  20. #define asmlinkage
  21. #ifdef __x86_64__
  22. #define __NR_perf_counter_open 298
  23. #define rmb() asm volatile("lfence" ::: "memory")
  24. #define cpu_relax() asm volatile("rep; nop" ::: "memory");
  25. #endif
  26. #ifdef __i386__
  27. #define __NR_perf_counter_open 336
  28. #define rmb() asm volatile("lfence" ::: "memory")
  29. #define cpu_relax() asm volatile("rep; nop" ::: "memory");
  30. #endif
  31. #ifdef __powerpc__
  32. #define __NR_perf_counter_open 319
  33. #define rmb() asm volatile ("sync" ::: "memory")
  34. #define cpu_relax() asm volatile ("" ::: "memory");
  35. #endif
  36. #define unlikely(x) __builtin_expect(!!(x), 0)
  37. #define min(x, y) ({ \
  38. typeof(x) _min1 = (x); \
  39. typeof(y) _min2 = (y); \
  40. (void) (&_min1 == &_min2); \
  41. _min1 < _min2 ? _min1 : _min2; })
  42. static inline int
  43. sys_perf_counter_open(struct perf_counter_hw_event *hw_event_uptr,
  44. pid_t pid, int cpu, int group_fd,
  45. unsigned long flags)
  46. {
  47. return syscall(__NR_perf_counter_open, hw_event_uptr, pid, cpu,
  48. group_fd, flags);
  49. }
  50. #define MAX_COUNTERS 64
  51. #define MAX_NR_CPUS 256
  52. #define EID(type, id) (((__u64)(type) << PERF_COUNTER_TYPE_SHIFT) | (id))
  53. #endif