cpu-hotplug.txt 14 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367
  1. CPU hotplug Support in Linux(tm) Kernel
  2. Maintainers:
  3. CPU Hotplug Core:
  4. Rusty Russell <rusty@rustycorp.com.au>
  5. Srivatsa Vaddagiri <vatsa@in.ibm.com>
  6. i386:
  7. Zwane Mwaikambo <zwane@arm.linux.org.uk>
  8. ppc64:
  9. Nathan Lynch <nathanl@austin.ibm.com>
  10. Joel Schopp <jschopp@austin.ibm.com>
  11. ia64/x86_64:
  12. Ashok Raj <ashok.raj@intel.com>
  13. Authors: Ashok Raj <ashok.raj@intel.com>
  14. Lots of feedback: Nathan Lynch <nathanl@austin.ibm.com>,
  15. Joel Schopp <jschopp@austin.ibm.com>
  16. Introduction
  17. Modern advances in system architectures have introduced advanced error
  18. reporting and correction capabilities in processors. CPU architectures permit
  19. partitioning support, where compute resources of a single CPU could be made
  20. available to virtual machine environments. There are couple OEMS that
  21. support NUMA hardware which are hot pluggable as well, where physical
  22. node insertion and removal require support for CPU hotplug.
  23. Such advances require CPUs available to a kernel to be removed either for
  24. provisioning reasons, or for RAS purposes to keep an offending CPU off
  25. system execution path. Hence the need for CPU hotplug support in the
  26. Linux kernel.
  27. A more novel use of CPU-hotplug support is its use today in suspend
  28. resume support for SMP. Dual-core and HT support makes even
  29. a laptop run SMP kernels which didn't support these methods. SMP support
  30. for suspend/resume is a work in progress.
  31. General Stuff about CPU Hotplug
  32. --------------------------------
  33. Command Line Switches
  34. ---------------------
  35. maxcpus=n Restrict boot time cpus to n. Say if you have 4 cpus, using
  36. maxcpus=2 will only boot 2. You can choose to bring the
  37. other cpus later online, read FAQ's for more info.
  38. additional_cpus*=n Use this to limit hotpluggable cpus. This option sets
  39. cpu_possible_map = cpu_present_map + additional_cpus
  40. (*) Option valid only for following architectures
  41. - x86_64, ia64
  42. ia64 and x86_64 use the number of disabled local apics in ACPI tables MADT
  43. to determine the number of potentially hot-pluggable cpus. The implementation
  44. should only rely on this to count the #of cpus, but *MUST* not rely on the
  45. apicid values in those tables for disabled apics. In the event BIOS doesnt
  46. mark such hot-pluggable cpus as disabled entries, one could use this
  47. parameter "additional_cpus=x" to represent those cpus in the cpu_possible_map.
  48. CPU maps and such
  49. -----------------
  50. [More on cpumaps and primitive to manipulate, please check
  51. include/linux/cpumask.h that has more descriptive text.]
  52. cpu_possible_map: Bitmap of possible CPUs that can ever be available in the
  53. system. This is used to allocate some boot time memory for per_cpu variables
  54. that aren't designed to grow/shrink as CPUs are made available or removed.
  55. Once set during boot time discovery phase, the map is static, i.e no bits
  56. are added or removed anytime. Trimming it accurately for your system needs
  57. upfront can save some boot time memory. See below for how we use heuristics
  58. in x86_64 case to keep this under check.
  59. cpu_online_map: Bitmap of all CPUs currently online. Its set in __cpu_up()
  60. after a cpu is available for kernel scheduling and ready to receive
  61. interrupts from devices. Its cleared when a cpu is brought down using
  62. __cpu_disable(), before which all OS services including interrupts are
  63. migrated to another target CPU.
  64. cpu_present_map: Bitmap of CPUs currently present in the system. Not all
  65. of them may be online. When physical hotplug is processed by the relevant
  66. subsystem (e.g ACPI) can change and new bit either be added or removed
  67. from the map depending on the event is hot-add/hot-remove. There are currently
  68. no locking rules as of now. Typical usage is to init topology during boot,
  69. at which time hotplug is disabled.
  70. You really dont need to manipulate any of the system cpu maps. They should
  71. be read-only for most use. When setting up per-cpu resources almost always use
  72. cpu_possible_map/for_each_cpu() to iterate.
  73. Never use anything other than cpumask_t to represent bitmap of CPUs.
  74. #include <linux/cpumask.h>
  75. for_each_cpu - Iterate over cpu_possible_map
  76. for_each_online_cpu - Iterate over cpu_online_map
  77. for_each_present_cpu - Iterate over cpu_present_map
  78. for_each_cpu_mask(x,mask) - Iterate over some random collection of cpu mask.
  79. #include <linux/cpu.h>
  80. lock_cpu_hotplug() and unlock_cpu_hotplug():
  81. The above calls are used to inhibit cpu hotplug operations. While holding the
  82. cpucontrol mutex, cpu_online_map will not change. If you merely need to avoid
  83. cpus going away, you could also use preempt_disable() and preempt_enable()
  84. for those sections. Just remember the critical section cannot call any
  85. function that can sleep or schedule this process away. The preempt_disable()
  86. will work as long as stop_machine_run() is used to take a cpu down.
  87. CPU Hotplug - Frequently Asked Questions.
  88. Q: How to i enable my kernel to support CPU hotplug?
  89. A: When doing make defconfig, Enable CPU hotplug support
  90. "Processor type and Features" -> Support for Hotpluggable CPUs
  91. Make sure that you have CONFIG_HOTPLUG, and CONFIG_SMP turned on as well.
  92. You would need to enable CONFIG_HOTPLUG_CPU for SMP suspend/resume support
  93. as well.
  94. Q: What architectures support CPU hotplug?
  95. A: As of 2.6.14, the following architectures support CPU hotplug.
  96. i386 (Intel), ppc, ppc64, parisc, s390, ia64 and x86_64
  97. Q: How to test if hotplug is supported on the newly built kernel?
  98. A: You should now notice an entry in sysfs.
  99. Check if sysfs is mounted, using the "mount" command. You should notice
  100. an entry as shown below in the output.
  101. ....
  102. none on /sys type sysfs (rw)
  103. ....
  104. if this is not mounted, do the following.
  105. #mkdir /sysfs
  106. #mount -t sysfs sys /sys
  107. now you should see entries for all present cpu, the following is an example
  108. in a 8-way system.
  109. #pwd
  110. #/sys/devices/system/cpu
  111. #ls -l
  112. total 0
  113. drwxr-xr-x 10 root root 0 Sep 19 07:44 .
  114. drwxr-xr-x 13 root root 0 Sep 19 07:45 ..
  115. drwxr-xr-x 3 root root 0 Sep 19 07:44 cpu0
  116. drwxr-xr-x 3 root root 0 Sep 19 07:44 cpu1
  117. drwxr-xr-x 3 root root 0 Sep 19 07:44 cpu2
  118. drwxr-xr-x 3 root root 0 Sep 19 07:44 cpu3
  119. drwxr-xr-x 3 root root 0 Sep 19 07:44 cpu4
  120. drwxr-xr-x 3 root root 0 Sep 19 07:44 cpu5
  121. drwxr-xr-x 3 root root 0 Sep 19 07:44 cpu6
  122. drwxr-xr-x 3 root root 0 Sep 19 07:48 cpu7
  123. Under each directory you would find an "online" file which is the control
  124. file to logically online/offline a processor.
  125. Q: Does hot-add/hot-remove refer to physical add/remove of cpus?
  126. A: The usage of hot-add/remove may not be very consistently used in the code.
  127. CONFIG_CPU_HOTPLUG enables logical online/offline capability in the kernel.
  128. To support physical addition/removal, one would need some BIOS hooks and
  129. the platform should have something like an attention button in PCI hotplug.
  130. CONFIG_ACPI_HOTPLUG_CPU enables ACPI support for physical add/remove of CPUs.
  131. Q: How do i logically offline a CPU?
  132. A: Do the following.
  133. #echo 0 > /sys/devices/system/cpu/cpuX/online
  134. once the logical offline is successful, check
  135. #cat /proc/interrupts
  136. you should now not see the CPU that you removed. Also online file will report
  137. the state as 0 when a cpu if offline and 1 when its online.
  138. #To display the current cpu state.
  139. #cat /sys/devices/system/cpu/cpuX/online
  140. Q: Why cant i remove CPU0 on some systems?
  141. A: Some architectures may have some special dependency on a certain CPU.
  142. For e.g in IA64 platforms we have ability to sent platform interrupts to the
  143. OS. a.k.a Corrected Platform Error Interrupts (CPEI). In current ACPI
  144. specifications, we didn't have a way to change the target CPU. Hence if the
  145. current ACPI version doesn't support such re-direction, we disable that CPU
  146. by making it not-removable.
  147. In such cases you will also notice that the online file is missing under cpu0.
  148. Q: How do i find out if a particular CPU is not removable?
  149. A: Depending on the implementation, some architectures may show this by the
  150. absence of the "online" file. This is done if it can be determined ahead of
  151. time that this CPU cannot be removed.
  152. In some situations, this can be a run time check, i.e if you try to remove the
  153. last CPU, this will not be permitted. You can find such failures by
  154. investigating the return value of the "echo" command.
  155. Q: What happens when a CPU is being logically offlined?
  156. A: The following happen, listed in no particular order :-)
  157. - A notification is sent to in-kernel registered modules by sending an event
  158. CPU_DOWN_PREPARE
  159. - All process is migrated away from this outgoing CPU to a new CPU
  160. - All interrupts targeted to this CPU is migrated to a new CPU
  161. - timers/bottom half/task lets are also migrated to a new CPU
  162. - Once all services are migrated, kernel calls an arch specific routine
  163. __cpu_disable() to perform arch specific cleanup.
  164. - Once this is successful, an event for successful cleanup is sent by an event
  165. CPU_DEAD.
  166. "It is expected that each service cleans up when the CPU_DOWN_PREPARE
  167. notifier is called, when CPU_DEAD is called its expected there is nothing
  168. running on behalf of this CPU that was offlined"
  169. Q: If i have some kernel code that needs to be aware of CPU arrival and
  170. departure, how to i arrange for proper notification?
  171. A: This is what you would need in your kernel code to receive notifications.
  172. #include <linux/cpu.h>
  173. static int __cpuinit foobar_cpu_callback(struct notifier_block *nfb,
  174. unsigned long action, void *hcpu)
  175. {
  176. unsigned int cpu = (unsigned long)hcpu;
  177. switch (action) {
  178. case CPU_ONLINE:
  179. foobar_online_action(cpu);
  180. break;
  181. case CPU_DEAD:
  182. foobar_dead_action(cpu);
  183. break;
  184. }
  185. return NOTIFY_OK;
  186. }
  187. static struct notifier_block foobar_cpu_notifer =
  188. {
  189. .notifier_call = foobar_cpu_callback,
  190. };
  191. In your init function,
  192. register_cpu_notifier(&foobar_cpu_notifier);
  193. You can fail PREPARE notifiers if something doesn't work to prepare resources.
  194. This will stop the activity and send a following CANCELED event back.
  195. CPU_DEAD should not be failed, its just a goodness indication, but bad
  196. things will happen if a notifier in path sent a BAD notify code.
  197. Q: I don't see my action being called for all CPUs already up and running?
  198. A: Yes, CPU notifiers are called only when new CPUs are on-lined or offlined.
  199. If you need to perform some action for each cpu already in the system, then
  200. for_each_online_cpu(i) {
  201. foobar_cpu_callback(&foobar_cpu_notifier, CPU_UP_PREPARE, i);
  202. foobar_cpu_callback(&foobar-cpu_notifier, CPU_ONLINE, i);
  203. }
  204. Q: If i would like to develop cpu hotplug support for a new architecture,
  205. what do i need at a minimum?
  206. A: The following are what is required for CPU hotplug infrastructure to work
  207. correctly.
  208. - Make sure you have an entry in Kconfig to enable CONFIG_HOTPLUG_CPU
  209. - __cpu_up() - Arch interface to bring up a CPU
  210. - __cpu_disable() - Arch interface to shutdown a CPU, no more interrupts
  211. can be handled by the kernel after the routine
  212. returns. Including local APIC timers etc are
  213. shutdown.
  214. - __cpu_die() - This actually supposed to ensure death of the CPU.
  215. Actually look at some example code in other arch
  216. that implement CPU hotplug. The processor is taken
  217. down from the idle() loop for that specific
  218. architecture. __cpu_die() typically waits for some
  219. per_cpu state to be set, to ensure the processor
  220. dead routine is called to be sure positively.
  221. Q: I need to ensure that a particular cpu is not removed when there is some
  222. work specific to this cpu is in progress.
  223. A: First switch the current thread context to preferred cpu
  224. int my_func_on_cpu(int cpu)
  225. {
  226. cpumask_t saved_mask, new_mask = CPU_MASK_NONE;
  227. int curr_cpu, err = 0;
  228. saved_mask = current->cpus_allowed;
  229. cpu_set(cpu, new_mask);
  230. err = set_cpus_allowed(current, new_mask);
  231. if (err)
  232. return err;
  233. /*
  234. * If we got scheduled out just after the return from
  235. * set_cpus_allowed() before running the work, this ensures
  236. * we stay locked.
  237. */
  238. curr_cpu = get_cpu();
  239. if (curr_cpu != cpu) {
  240. err = -EAGAIN;
  241. goto ret;
  242. } else {
  243. /*
  244. * Do work : But cant sleep, since get_cpu() disables preempt
  245. */
  246. }
  247. ret:
  248. put_cpu();
  249. set_cpus_allowed(current, saved_mask);
  250. return err;
  251. }
  252. Q: How do we determine how many CPUs are available for hotplug.
  253. A: There is no clear spec defined way from ACPI that can give us that
  254. information today. Based on some input from Natalie of Unisys,
  255. that the ACPI MADT (Multiple APIC Description Tables) marks those possible
  256. CPUs in a system with disabled status.
  257. Andi implemented some simple heuristics that count the number of disabled
  258. CPUs in MADT as hotpluggable CPUS. In the case there are no disabled CPUS
  259. we assume 1/2 the number of CPUs currently present can be hotplugged.
  260. Caveat: Today's ACPI MADT can only provide 256 entries since the apicid field
  261. in MADT is only 8 bits.
  262. User Space Notification
  263. Hotplug support for devices is common in Linux today. Its being used today to
  264. support automatic configuration of network, usb and pci devices. A hotplug
  265. event can be used to invoke an agent script to perform the configuration task.
  266. You can add /etc/hotplug/cpu.agent to handle hotplug notification user space
  267. scripts.
  268. #!/bin/bash
  269. # $Id: cpu.agent
  270. # Kernel hotplug params include:
  271. #ACTION=%s [online or offline]
  272. #DEVPATH=%s
  273. #
  274. cd /etc/hotplug
  275. . ./hotplug.functions
  276. case $ACTION in
  277. online)
  278. echo `date` ":cpu.agent" add cpu >> /tmp/hotplug.txt
  279. ;;
  280. offline)
  281. echo `date` ":cpu.agent" remove cpu >>/tmp/hotplug.txt
  282. ;;
  283. *)
  284. debug_mesg CPU $ACTION event not supported
  285. exit 1
  286. ;;
  287. esac