cpuacct.c 6.4 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303
  1. #include <linux/cgroup.h>
  2. #include <linux/slab.h>
  3. #include <linux/percpu.h>
  4. #include <linux/spinlock.h>
  5. #include <linux/cpumask.h>
  6. #include <linux/seq_file.h>
  7. #include <linux/rcupdate.h>
  8. #include <linux/kernel_stat.h>
  9. #include "sched.h"
  10. /*
  11. * CPU accounting code for task groups.
  12. *
  13. * Based on the work by Paul Menage (menage@google.com) and Balbir Singh
  14. * (balbir@in.ibm.com).
  15. */
  16. /* Time spent by the tasks of the cpu accounting group executing in ... */
  17. enum cpuacct_stat_index {
  18. CPUACCT_STAT_USER, /* ... user mode */
  19. CPUACCT_STAT_SYSTEM, /* ... kernel mode */
  20. CPUACCT_STAT_NSTATS,
  21. };
  22. /* track cpu usage of a group of tasks and its child groups */
  23. struct cpuacct {
  24. struct cgroup_subsys_state css;
  25. /* cpuusage holds pointer to a u64-type object on every cpu */
  26. u64 __percpu *cpuusage;
  27. struct kernel_cpustat __percpu *cpustat;
  28. };
  29. /* return cpu accounting group corresponding to this container */
  30. static inline struct cpuacct *cgroup_ca(struct cgroup *cgrp)
  31. {
  32. return container_of(cgroup_subsys_state(cgrp, cpuacct_subsys_id),
  33. struct cpuacct, css);
  34. }
  35. /* return cpu accounting group to which this task belongs */
  36. static inline struct cpuacct *task_ca(struct task_struct *tsk)
  37. {
  38. return container_of(task_subsys_state(tsk, cpuacct_subsys_id),
  39. struct cpuacct, css);
  40. }
  41. static inline struct cpuacct *__parent_ca(struct cpuacct *ca)
  42. {
  43. return cgroup_ca(ca->css.cgroup->parent);
  44. }
  45. static inline struct cpuacct *parent_ca(struct cpuacct *ca)
  46. {
  47. if (!ca->css.cgroup->parent)
  48. return NULL;
  49. return cgroup_ca(ca->css.cgroup->parent);
  50. }
  51. static struct cpuacct root_cpuacct;
  52. /* create a new cpu accounting group */
  53. static struct cgroup_subsys_state *cpuacct_css_alloc(struct cgroup *cgrp)
  54. {
  55. struct cpuacct *ca;
  56. if (!cgrp->parent)
  57. return &root_cpuacct.css;
  58. ca = kzalloc(sizeof(*ca), GFP_KERNEL);
  59. if (!ca)
  60. goto out;
  61. ca->cpuusage = alloc_percpu(u64);
  62. if (!ca->cpuusage)
  63. goto out_free_ca;
  64. ca->cpustat = alloc_percpu(struct kernel_cpustat);
  65. if (!ca->cpustat)
  66. goto out_free_cpuusage;
  67. return &ca->css;
  68. out_free_cpuusage:
  69. free_percpu(ca->cpuusage);
  70. out_free_ca:
  71. kfree(ca);
  72. out:
  73. return ERR_PTR(-ENOMEM);
  74. }
  75. /* destroy an existing cpu accounting group */
  76. static void cpuacct_css_free(struct cgroup *cgrp)
  77. {
  78. struct cpuacct *ca = cgroup_ca(cgrp);
  79. free_percpu(ca->cpustat);
  80. free_percpu(ca->cpuusage);
  81. kfree(ca);
  82. }
  83. static u64 cpuacct_cpuusage_read(struct cpuacct *ca, int cpu)
  84. {
  85. u64 *cpuusage = per_cpu_ptr(ca->cpuusage, cpu);
  86. u64 data;
  87. #ifndef CONFIG_64BIT
  88. /*
  89. * Take rq->lock to make 64-bit read safe on 32-bit platforms.
  90. */
  91. raw_spin_lock_irq(&cpu_rq(cpu)->lock);
  92. data = *cpuusage;
  93. raw_spin_unlock_irq(&cpu_rq(cpu)->lock);
  94. #else
  95. data = *cpuusage;
  96. #endif
  97. return data;
  98. }
  99. static void cpuacct_cpuusage_write(struct cpuacct *ca, int cpu, u64 val)
  100. {
  101. u64 *cpuusage = per_cpu_ptr(ca->cpuusage, cpu);
  102. #ifndef CONFIG_64BIT
  103. /*
  104. * Take rq->lock to make 64-bit write safe on 32-bit platforms.
  105. */
  106. raw_spin_lock_irq(&cpu_rq(cpu)->lock);
  107. *cpuusage = val;
  108. raw_spin_unlock_irq(&cpu_rq(cpu)->lock);
  109. #else
  110. *cpuusage = val;
  111. #endif
  112. }
  113. /* return total cpu usage (in nanoseconds) of a group */
  114. static u64 cpuusage_read(struct cgroup *cgrp, struct cftype *cft)
  115. {
  116. struct cpuacct *ca = cgroup_ca(cgrp);
  117. u64 totalcpuusage = 0;
  118. int i;
  119. for_each_present_cpu(i)
  120. totalcpuusage += cpuacct_cpuusage_read(ca, i);
  121. return totalcpuusage;
  122. }
  123. static int cpuusage_write(struct cgroup *cgrp, struct cftype *cftype,
  124. u64 reset)
  125. {
  126. struct cpuacct *ca = cgroup_ca(cgrp);
  127. int err = 0;
  128. int i;
  129. if (reset) {
  130. err = -EINVAL;
  131. goto out;
  132. }
  133. for_each_present_cpu(i)
  134. cpuacct_cpuusage_write(ca, i, 0);
  135. out:
  136. return err;
  137. }
  138. static int cpuacct_percpu_seq_read(struct cgroup *cgroup, struct cftype *cft,
  139. struct seq_file *m)
  140. {
  141. struct cpuacct *ca = cgroup_ca(cgroup);
  142. u64 percpu;
  143. int i;
  144. for_each_present_cpu(i) {
  145. percpu = cpuacct_cpuusage_read(ca, i);
  146. seq_printf(m, "%llu ", (unsigned long long) percpu);
  147. }
  148. seq_printf(m, "\n");
  149. return 0;
  150. }
  151. static const char * const cpuacct_stat_desc[] = {
  152. [CPUACCT_STAT_USER] = "user",
  153. [CPUACCT_STAT_SYSTEM] = "system",
  154. };
  155. static int cpuacct_stats_show(struct cgroup *cgrp, struct cftype *cft,
  156. struct cgroup_map_cb *cb)
  157. {
  158. struct cpuacct *ca = cgroup_ca(cgrp);
  159. int cpu;
  160. s64 val = 0;
  161. for_each_online_cpu(cpu) {
  162. struct kernel_cpustat *kcpustat = per_cpu_ptr(ca->cpustat, cpu);
  163. val += kcpustat->cpustat[CPUTIME_USER];
  164. val += kcpustat->cpustat[CPUTIME_NICE];
  165. }
  166. val = cputime64_to_clock_t(val);
  167. cb->fill(cb, cpuacct_stat_desc[CPUACCT_STAT_USER], val);
  168. val = 0;
  169. for_each_online_cpu(cpu) {
  170. struct kernel_cpustat *kcpustat = per_cpu_ptr(ca->cpustat, cpu);
  171. val += kcpustat->cpustat[CPUTIME_SYSTEM];
  172. val += kcpustat->cpustat[CPUTIME_IRQ];
  173. val += kcpustat->cpustat[CPUTIME_SOFTIRQ];
  174. }
  175. val = cputime64_to_clock_t(val);
  176. cb->fill(cb, cpuacct_stat_desc[CPUACCT_STAT_SYSTEM], val);
  177. return 0;
  178. }
  179. static struct cftype files[] = {
  180. {
  181. .name = "usage",
  182. .read_u64 = cpuusage_read,
  183. .write_u64 = cpuusage_write,
  184. },
  185. {
  186. .name = "usage_percpu",
  187. .read_seq_string = cpuacct_percpu_seq_read,
  188. },
  189. {
  190. .name = "stat",
  191. .read_map = cpuacct_stats_show,
  192. },
  193. { } /* terminate */
  194. };
  195. /*
  196. * charge this task's execution time to its accounting group.
  197. *
  198. * called with rq->lock held.
  199. */
  200. void cpuacct_charge(struct task_struct *tsk, u64 cputime)
  201. {
  202. struct cpuacct *ca;
  203. int cpu;
  204. if (unlikely(!cpuacct_subsys.active))
  205. return;
  206. cpu = task_cpu(tsk);
  207. rcu_read_lock();
  208. ca = task_ca(tsk);
  209. while (true) {
  210. u64 *cpuusage = per_cpu_ptr(ca->cpuusage, cpu);
  211. *cpuusage += cputime;
  212. ca = parent_ca(ca);
  213. if (!ca)
  214. break;
  215. }
  216. rcu_read_unlock();
  217. }
  218. /*
  219. * Add user/system time to cpuacct.
  220. *
  221. * Note: it's the caller that updates the account of the root cgroup.
  222. */
  223. void cpuacct_account_field(struct task_struct *p, int index, u64 val)
  224. {
  225. struct kernel_cpustat *kcpustat;
  226. struct cpuacct *ca;
  227. if (unlikely(!cpuacct_subsys.active))
  228. return;
  229. rcu_read_lock();
  230. ca = task_ca(p);
  231. while (ca != &root_cpuacct) {
  232. kcpustat = this_cpu_ptr(ca->cpustat);
  233. kcpustat->cpustat[index] += val;
  234. ca = __parent_ca(ca);
  235. }
  236. rcu_read_unlock();
  237. }
  238. void __init cpuacct_init(void)
  239. {
  240. root_cpuacct.cpustat = &kernel_cpustat;
  241. root_cpuacct.cpuusage = alloc_percpu(u64);
  242. BUG_ON(!root_cpuacct.cpuusage); /* Too early, not expected to fail */
  243. }
  244. struct cgroup_subsys cpuacct_subsys = {
  245. .name = "cpuacct",
  246. .css_alloc = cpuacct_css_alloc,
  247. .css_free = cpuacct_css_free,
  248. .subsys_id = cpuacct_subsys_id,
  249. .base_cftypes = files,
  250. };