cpuacct.c 6.4 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303
  1. #include <linux/cgroup.h>
  2. #include <linux/slab.h>
  3. #include <linux/percpu.h>
  4. #include <linux/spinlock.h>
  5. #include <linux/cpumask.h>
  6. #include <linux/seq_file.h>
  7. #include <linux/rcupdate.h>
  8. #include <linux/kernel_stat.h>
  9. #include "sched.h"
  10. /*
  11. * CPU accounting code for task groups.
  12. *
  13. * Based on the work by Paul Menage (menage@google.com) and Balbir Singh
  14. * (balbir@in.ibm.com).
  15. */
  16. /* Time spent by the tasks of the cpu accounting group executing in ... */
  17. enum cpuacct_stat_index {
  18. CPUACCT_STAT_USER, /* ... user mode */
  19. CPUACCT_STAT_SYSTEM, /* ... kernel mode */
  20. CPUACCT_STAT_NSTATS,
  21. };
  22. /* track cpu usage of a group of tasks and its child groups */
  23. struct cpuacct {
  24. struct cgroup_subsys_state css;
  25. /* cpuusage holds pointer to a u64-type object on every cpu */
  26. u64 __percpu *cpuusage;
  27. struct kernel_cpustat __percpu *cpustat;
  28. };
  29. /* return cpu accounting group corresponding to this container */
  30. static inline struct cpuacct *cgroup_ca(struct cgroup *cgrp)
  31. {
  32. return container_of(cgroup_subsys_state(cgrp, cpuacct_subsys_id),
  33. struct cpuacct, css);
  34. }
  35. /* return cpu accounting group to which this task belongs */
  36. static inline struct cpuacct *task_ca(struct task_struct *tsk)
  37. {
  38. return container_of(task_subsys_state(tsk, cpuacct_subsys_id),
  39. struct cpuacct, css);
  40. }
  41. static inline struct cpuacct *__parent_ca(struct cpuacct *ca)
  42. {
  43. return cgroup_ca(ca->css.cgroup->parent);
  44. }
  45. static inline struct cpuacct *parent_ca(struct cpuacct *ca)
  46. {
  47. if (!ca->css.cgroup->parent)
  48. return NULL;
  49. return cgroup_ca(ca->css.cgroup->parent);
  50. }
  51. static DEFINE_PER_CPU(u64, root_cpuacct_cpuusage);
  52. static struct cpuacct root_cpuacct;
  53. /* create a new cpu accounting group */
  54. static struct cgroup_subsys_state *cpuacct_css_alloc(struct cgroup *cgrp)
  55. {
  56. struct cpuacct *ca;
  57. if (!cgrp->parent)
  58. return &root_cpuacct.css;
  59. ca = kzalloc(sizeof(*ca), GFP_KERNEL);
  60. if (!ca)
  61. goto out;
  62. ca->cpuusage = alloc_percpu(u64);
  63. if (!ca->cpuusage)
  64. goto out_free_ca;
  65. ca->cpustat = alloc_percpu(struct kernel_cpustat);
  66. if (!ca->cpustat)
  67. goto out_free_cpuusage;
  68. return &ca->css;
  69. out_free_cpuusage:
  70. free_percpu(ca->cpuusage);
  71. out_free_ca:
  72. kfree(ca);
  73. out:
  74. return ERR_PTR(-ENOMEM);
  75. }
  76. /* destroy an existing cpu accounting group */
  77. static void cpuacct_css_free(struct cgroup *cgrp)
  78. {
  79. struct cpuacct *ca = cgroup_ca(cgrp);
  80. free_percpu(ca->cpustat);
  81. free_percpu(ca->cpuusage);
  82. kfree(ca);
  83. }
  84. static u64 cpuacct_cpuusage_read(struct cpuacct *ca, int cpu)
  85. {
  86. u64 *cpuusage = per_cpu_ptr(ca->cpuusage, cpu);
  87. u64 data;
  88. #ifndef CONFIG_64BIT
  89. /*
  90. * Take rq->lock to make 64-bit read safe on 32-bit platforms.
  91. */
  92. raw_spin_lock_irq(&cpu_rq(cpu)->lock);
  93. data = *cpuusage;
  94. raw_spin_unlock_irq(&cpu_rq(cpu)->lock);
  95. #else
  96. data = *cpuusage;
  97. #endif
  98. return data;
  99. }
  100. static void cpuacct_cpuusage_write(struct cpuacct *ca, int cpu, u64 val)
  101. {
  102. u64 *cpuusage = per_cpu_ptr(ca->cpuusage, cpu);
  103. #ifndef CONFIG_64BIT
  104. /*
  105. * Take rq->lock to make 64-bit write safe on 32-bit platforms.
  106. */
  107. raw_spin_lock_irq(&cpu_rq(cpu)->lock);
  108. *cpuusage = val;
  109. raw_spin_unlock_irq(&cpu_rq(cpu)->lock);
  110. #else
  111. *cpuusage = val;
  112. #endif
  113. }
  114. /* return total cpu usage (in nanoseconds) of a group */
  115. static u64 cpuusage_read(struct cgroup *cgrp, struct cftype *cft)
  116. {
  117. struct cpuacct *ca = cgroup_ca(cgrp);
  118. u64 totalcpuusage = 0;
  119. int i;
  120. for_each_present_cpu(i)
  121. totalcpuusage += cpuacct_cpuusage_read(ca, i);
  122. return totalcpuusage;
  123. }
  124. static int cpuusage_write(struct cgroup *cgrp, struct cftype *cftype,
  125. u64 reset)
  126. {
  127. struct cpuacct *ca = cgroup_ca(cgrp);
  128. int err = 0;
  129. int i;
  130. if (reset) {
  131. err = -EINVAL;
  132. goto out;
  133. }
  134. for_each_present_cpu(i)
  135. cpuacct_cpuusage_write(ca, i, 0);
  136. out:
  137. return err;
  138. }
  139. static int cpuacct_percpu_seq_read(struct cgroup *cgroup, struct cftype *cft,
  140. struct seq_file *m)
  141. {
  142. struct cpuacct *ca = cgroup_ca(cgroup);
  143. u64 percpu;
  144. int i;
  145. for_each_present_cpu(i) {
  146. percpu = cpuacct_cpuusage_read(ca, i);
  147. seq_printf(m, "%llu ", (unsigned long long) percpu);
  148. }
  149. seq_printf(m, "\n");
  150. return 0;
  151. }
  152. static const char * const cpuacct_stat_desc[] = {
  153. [CPUACCT_STAT_USER] = "user",
  154. [CPUACCT_STAT_SYSTEM] = "system",
  155. };
  156. static int cpuacct_stats_show(struct cgroup *cgrp, struct cftype *cft,
  157. struct cgroup_map_cb *cb)
  158. {
  159. struct cpuacct *ca = cgroup_ca(cgrp);
  160. int cpu;
  161. s64 val = 0;
  162. for_each_online_cpu(cpu) {
  163. struct kernel_cpustat *kcpustat = per_cpu_ptr(ca->cpustat, cpu);
  164. val += kcpustat->cpustat[CPUTIME_USER];
  165. val += kcpustat->cpustat[CPUTIME_NICE];
  166. }
  167. val = cputime64_to_clock_t(val);
  168. cb->fill(cb, cpuacct_stat_desc[CPUACCT_STAT_USER], val);
  169. val = 0;
  170. for_each_online_cpu(cpu) {
  171. struct kernel_cpustat *kcpustat = per_cpu_ptr(ca->cpustat, cpu);
  172. val += kcpustat->cpustat[CPUTIME_SYSTEM];
  173. val += kcpustat->cpustat[CPUTIME_IRQ];
  174. val += kcpustat->cpustat[CPUTIME_SOFTIRQ];
  175. }
  176. val = cputime64_to_clock_t(val);
  177. cb->fill(cb, cpuacct_stat_desc[CPUACCT_STAT_SYSTEM], val);
  178. return 0;
  179. }
  180. static struct cftype files[] = {
  181. {
  182. .name = "usage",
  183. .read_u64 = cpuusage_read,
  184. .write_u64 = cpuusage_write,
  185. },
  186. {
  187. .name = "usage_percpu",
  188. .read_seq_string = cpuacct_percpu_seq_read,
  189. },
  190. {
  191. .name = "stat",
  192. .read_map = cpuacct_stats_show,
  193. },
  194. { } /* terminate */
  195. };
  196. /*
  197. * charge this task's execution time to its accounting group.
  198. *
  199. * called with rq->lock held.
  200. */
  201. void cpuacct_charge(struct task_struct *tsk, u64 cputime)
  202. {
  203. struct cpuacct *ca;
  204. int cpu;
  205. if (unlikely(!cpuacct_subsys.active))
  206. return;
  207. cpu = task_cpu(tsk);
  208. rcu_read_lock();
  209. ca = task_ca(tsk);
  210. while (true) {
  211. u64 *cpuusage = per_cpu_ptr(ca->cpuusage, cpu);
  212. *cpuusage += cputime;
  213. ca = parent_ca(ca);
  214. if (!ca)
  215. break;
  216. }
  217. rcu_read_unlock();
  218. }
  219. /*
  220. * Add user/system time to cpuacct.
  221. *
  222. * Note: it's the caller that updates the account of the root cgroup.
  223. */
  224. void cpuacct_account_field(struct task_struct *p, int index, u64 val)
  225. {
  226. struct kernel_cpustat *kcpustat;
  227. struct cpuacct *ca;
  228. if (unlikely(!cpuacct_subsys.active))
  229. return;
  230. rcu_read_lock();
  231. ca = task_ca(p);
  232. while (ca != &root_cpuacct) {
  233. kcpustat = this_cpu_ptr(ca->cpustat);
  234. kcpustat->cpustat[index] += val;
  235. ca = __parent_ca(ca);
  236. }
  237. rcu_read_unlock();
  238. }
  239. void __init cpuacct_init(void)
  240. {
  241. root_cpuacct.cpustat = &kernel_cpustat;
  242. root_cpuacct.cpuusage = &root_cpuacct_cpuusage;
  243. }
  244. struct cgroup_subsys cpuacct_subsys = {
  245. .name = "cpuacct",
  246. .css_alloc = cpuacct_css_alloc,
  247. .css_free = cpuacct_css_free,
  248. .subsys_id = cpuacct_subsys_id,
  249. .base_cftypes = files,
  250. };