perf.h 5.3 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246
  1. #ifndef _PERF_PERF_H
  2. #define _PERF_PERF_H
  3. #include <asm/unistd.h>
  4. #if defined(__i386__)
  5. #define rmb() asm volatile("lock; addl $0,0(%%esp)" ::: "memory")
  6. #define cpu_relax() asm volatile("rep; nop" ::: "memory");
  7. #define CPUINFO_PROC "model name"
  8. #ifndef __NR_perf_event_open
  9. # define __NR_perf_event_open 336
  10. #endif
  11. #endif
  12. #if defined(__x86_64__)
  13. #define rmb() asm volatile("lfence" ::: "memory")
  14. #define cpu_relax() asm volatile("rep; nop" ::: "memory");
  15. #define CPUINFO_PROC "model name"
  16. #ifndef __NR_perf_event_open
  17. # define __NR_perf_event_open 298
  18. #endif
  19. #endif
  20. #ifdef __powerpc__
  21. #include "../../arch/powerpc/include/uapi/asm/unistd.h"
  22. #define rmb() asm volatile ("sync" ::: "memory")
  23. #define cpu_relax() asm volatile ("" ::: "memory");
  24. #define CPUINFO_PROC "cpu"
  25. #endif
  26. #ifdef __s390__
  27. #define rmb() asm volatile("bcr 15,0" ::: "memory")
  28. #define cpu_relax() asm volatile("" ::: "memory");
  29. #endif
  30. #ifdef __sh__
  31. #if defined(__SH4A__) || defined(__SH5__)
  32. # define rmb() asm volatile("synco" ::: "memory")
  33. #else
  34. # define rmb() asm volatile("" ::: "memory")
  35. #endif
  36. #define cpu_relax() asm volatile("" ::: "memory")
  37. #define CPUINFO_PROC "cpu type"
  38. #endif
  39. #ifdef __hppa__
  40. #define rmb() asm volatile("" ::: "memory")
  41. #define cpu_relax() asm volatile("" ::: "memory");
  42. #define CPUINFO_PROC "cpu"
  43. #endif
  44. #ifdef __sparc__
  45. #define rmb() asm volatile("":::"memory")
  46. #define cpu_relax() asm volatile("":::"memory")
  47. #define CPUINFO_PROC "cpu"
  48. #endif
  49. #ifdef __alpha__
  50. #define rmb() asm volatile("mb" ::: "memory")
  51. #define cpu_relax() asm volatile("" ::: "memory")
  52. #define CPUINFO_PROC "cpu model"
  53. #endif
  54. #ifdef __ia64__
  55. #define rmb() asm volatile ("mf" ::: "memory")
  56. #define cpu_relax() asm volatile ("hint @pause" ::: "memory")
  57. #define CPUINFO_PROC "model name"
  58. #endif
  59. #ifdef __arm__
  60. /*
  61. * Use the __kuser_memory_barrier helper in the CPU helper page. See
  62. * arch/arm/kernel/entry-armv.S in the kernel source for details.
  63. */
  64. #define rmb() ((void(*)(void))0xffff0fa0)()
  65. #define cpu_relax() asm volatile("":::"memory")
  66. #define CPUINFO_PROC "Processor"
  67. #endif
  68. #ifdef __aarch64__
  69. #define rmb() asm volatile("dmb ld" ::: "memory")
  70. #define cpu_relax() asm volatile("yield" ::: "memory")
  71. #endif
  72. #ifdef __mips__
  73. #define rmb() asm volatile( \
  74. ".set mips2\n\t" \
  75. "sync\n\t" \
  76. ".set mips0" \
  77. : /* no output */ \
  78. : /* no input */ \
  79. : "memory")
  80. #define cpu_relax() asm volatile("" ::: "memory")
  81. #define CPUINFO_PROC "cpu model"
  82. #endif
  83. #include <time.h>
  84. #include <unistd.h>
  85. #include <sys/types.h>
  86. #include <sys/syscall.h>
  87. #include <linux/perf_event.h>
  88. #include "util/types.h"
  89. #include <stdbool.h>
  90. struct perf_mmap {
  91. void *base;
  92. int mask;
  93. unsigned int prev;
  94. };
  95. static inline unsigned int perf_mmap__read_head(struct perf_mmap *mm)
  96. {
  97. struct perf_event_mmap_page *pc = mm->base;
  98. int head = pc->data_head;
  99. rmb();
  100. return head;
  101. }
  102. static inline void perf_mmap__write_tail(struct perf_mmap *md,
  103. unsigned long tail)
  104. {
  105. struct perf_event_mmap_page *pc = md->base;
  106. /*
  107. * ensure all reads are done before we write the tail out.
  108. */
  109. /* mb(); */
  110. pc->data_tail = tail;
  111. }
  112. /*
  113. * prctl(PR_TASK_PERF_EVENTS_DISABLE) will (cheaply) disable all
  114. * counters in the current task.
  115. */
  116. #define PR_TASK_PERF_EVENTS_DISABLE 31
  117. #define PR_TASK_PERF_EVENTS_ENABLE 32
  118. #ifndef NSEC_PER_SEC
  119. # define NSEC_PER_SEC 1000000000ULL
  120. #endif
  121. static inline unsigned long long rdclock(void)
  122. {
  123. struct timespec ts;
  124. clock_gettime(CLOCK_MONOTONIC, &ts);
  125. return ts.tv_sec * 1000000000ULL + ts.tv_nsec;
  126. }
  127. /*
  128. * Pick up some kernel type conventions:
  129. */
  130. #define __user
  131. #define asmlinkage
  132. #define unlikely(x) __builtin_expect(!!(x), 0)
  133. #define min(x, y) ({ \
  134. typeof(x) _min1 = (x); \
  135. typeof(y) _min2 = (y); \
  136. (void) (&_min1 == &_min2); \
  137. _min1 < _min2 ? _min1 : _min2; })
  138. extern bool test_attr__enabled;
  139. void test_attr__init(void);
  140. void test_attr__open(struct perf_event_attr *attr, pid_t pid, int cpu,
  141. int fd, int group_fd, unsigned long flags);
  142. static inline int
  143. sys_perf_event_open(struct perf_event_attr *attr,
  144. pid_t pid, int cpu, int group_fd,
  145. unsigned long flags)
  146. {
  147. int fd;
  148. fd = syscall(__NR_perf_event_open, attr, pid, cpu,
  149. group_fd, flags);
  150. if (unlikely(test_attr__enabled))
  151. test_attr__open(attr, pid, cpu, fd, group_fd, flags);
  152. return fd;
  153. }
  154. #define MAX_COUNTERS 256
  155. #define MAX_NR_CPUS 256
  156. struct ip_callchain {
  157. u64 nr;
  158. u64 ips[0];
  159. };
  160. struct branch_flags {
  161. u64 mispred:1;
  162. u64 predicted:1;
  163. u64 reserved:62;
  164. };
  165. struct branch_entry {
  166. u64 from;
  167. u64 to;
  168. struct branch_flags flags;
  169. };
  170. struct branch_stack {
  171. u64 nr;
  172. struct branch_entry entries[0];
  173. };
  174. extern const char *input_name;
  175. extern bool perf_host, perf_guest;
  176. extern const char perf_version_string[];
  177. void pthread__unblock_sigwinch(void);
  178. #include "util/target.h"
  179. enum perf_call_graph_mode {
  180. CALLCHAIN_NONE,
  181. CALLCHAIN_FP,
  182. CALLCHAIN_DWARF
  183. };
  184. struct perf_record_opts {
  185. struct perf_target target;
  186. int call_graph;
  187. bool group;
  188. bool inherit_stat;
  189. bool no_delay;
  190. bool no_inherit;
  191. bool no_samples;
  192. bool pipe_output;
  193. bool raw_samples;
  194. bool sample_address;
  195. bool sample_time;
  196. bool period;
  197. unsigned int freq;
  198. unsigned int mmap_pages;
  199. unsigned int user_freq;
  200. u64 branch_stack;
  201. u64 default_interval;
  202. u64 user_interval;
  203. u16 stack_dump_size;
  204. };
  205. #endif