vmcore.c 22 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895
  1. /*
  2. * fs/proc/vmcore.c Interface for accessing the crash
  3. * dump from the system's previous life.
  4. * Heavily borrowed from fs/proc/kcore.c
  5. * Created by: Hariprasad Nellitheertha (hari@in.ibm.com)
  6. * Copyright (C) IBM Corporation, 2004. All rights reserved
  7. *
  8. */
  9. #include <linux/mm.h>
  10. #include <linux/kcore.h>
  11. #include <linux/user.h>
  12. #include <linux/elf.h>
  13. #include <linux/elfcore.h>
  14. #include <linux/export.h>
  15. #include <linux/slab.h>
  16. #include <linux/highmem.h>
  17. #include <linux/printk.h>
  18. #include <linux/bootmem.h>
  19. #include <linux/init.h>
  20. #include <linux/crash_dump.h>
  21. #include <linux/list.h>
  22. #include <asm/uaccess.h>
  23. #include <asm/io.h>
  24. #include "internal.h"
  25. /* List representing chunks of contiguous memory areas and their offsets in
  26. * vmcore file.
  27. */
  28. static LIST_HEAD(vmcore_list);
  29. /* Stores the pointer to the buffer containing kernel elf core headers. */
  30. static char *elfcorebuf;
  31. static size_t elfcorebuf_sz;
  32. static size_t elfcorebuf_sz_orig;
  33. static char *elfnotes_buf;
  34. static size_t elfnotes_sz;
  35. /* Total size of vmcore file. */
  36. static u64 vmcore_size;
  37. static struct proc_dir_entry *proc_vmcore = NULL;
  38. /*
  39. * Returns > 0 for RAM pages, 0 for non-RAM pages, < 0 on error
  40. * The called function has to take care of module refcounting.
  41. */
  42. static int (*oldmem_pfn_is_ram)(unsigned long pfn);
  43. int register_oldmem_pfn_is_ram(int (*fn)(unsigned long pfn))
  44. {
  45. if (oldmem_pfn_is_ram)
  46. return -EBUSY;
  47. oldmem_pfn_is_ram = fn;
  48. return 0;
  49. }
  50. EXPORT_SYMBOL_GPL(register_oldmem_pfn_is_ram);
  51. void unregister_oldmem_pfn_is_ram(void)
  52. {
  53. oldmem_pfn_is_ram = NULL;
  54. wmb();
  55. }
  56. EXPORT_SYMBOL_GPL(unregister_oldmem_pfn_is_ram);
  57. static int pfn_is_ram(unsigned long pfn)
  58. {
  59. int (*fn)(unsigned long pfn);
  60. /* pfn is ram unless fn() checks pagetype */
  61. int ret = 1;
  62. /*
  63. * Ask hypervisor if the pfn is really ram.
  64. * A ballooned page contains no data and reading from such a page
  65. * will cause high load in the hypervisor.
  66. */
  67. fn = oldmem_pfn_is_ram;
  68. if (fn)
  69. ret = fn(pfn);
  70. return ret;
  71. }
  72. /* Reads a page from the oldmem device from given offset. */
  73. static ssize_t read_from_oldmem(char *buf, size_t count,
  74. u64 *ppos, int userbuf)
  75. {
  76. unsigned long pfn, offset;
  77. size_t nr_bytes;
  78. ssize_t read = 0, tmp;
  79. if (!count)
  80. return 0;
  81. offset = (unsigned long)(*ppos % PAGE_SIZE);
  82. pfn = (unsigned long)(*ppos / PAGE_SIZE);
  83. do {
  84. if (count > (PAGE_SIZE - offset))
  85. nr_bytes = PAGE_SIZE - offset;
  86. else
  87. nr_bytes = count;
  88. /* If pfn is not ram, return zeros for sparse dump files */
  89. if (pfn_is_ram(pfn) == 0)
  90. memset(buf, 0, nr_bytes);
  91. else {
  92. tmp = copy_oldmem_page(pfn, buf, nr_bytes,
  93. offset, userbuf);
  94. if (tmp < 0)
  95. return tmp;
  96. }
  97. *ppos += nr_bytes;
  98. count -= nr_bytes;
  99. buf += nr_bytes;
  100. read += nr_bytes;
  101. ++pfn;
  102. offset = 0;
  103. } while (count);
  104. return read;
  105. }
  106. /* Read from the ELF header and then the crash dump. On error, negative value is
  107. * returned otherwise number of bytes read are returned.
  108. */
  109. static ssize_t read_vmcore(struct file *file, char __user *buffer,
  110. size_t buflen, loff_t *fpos)
  111. {
  112. ssize_t acc = 0, tmp;
  113. size_t tsz;
  114. u64 start;
  115. struct vmcore *m = NULL;
  116. if (buflen == 0 || *fpos >= vmcore_size)
  117. return 0;
  118. /* trim buflen to not go beyond EOF */
  119. if (buflen > vmcore_size - *fpos)
  120. buflen = vmcore_size - *fpos;
  121. /* Read ELF core header */
  122. if (*fpos < elfcorebuf_sz) {
  123. tsz = min(elfcorebuf_sz - (size_t)*fpos, buflen);
  124. if (copy_to_user(buffer, elfcorebuf + *fpos, tsz))
  125. return -EFAULT;
  126. buflen -= tsz;
  127. *fpos += tsz;
  128. buffer += tsz;
  129. acc += tsz;
  130. /* leave now if filled buffer already */
  131. if (buflen == 0)
  132. return acc;
  133. }
  134. /* Read Elf note segment */
  135. if (*fpos < elfcorebuf_sz + elfnotes_sz) {
  136. void *kaddr;
  137. tsz = min(elfcorebuf_sz + elfnotes_sz - (size_t)*fpos, buflen);
  138. kaddr = elfnotes_buf + *fpos - elfcorebuf_sz;
  139. if (copy_to_user(buffer, kaddr, tsz))
  140. return -EFAULT;
  141. buflen -= tsz;
  142. *fpos += tsz;
  143. buffer += tsz;
  144. acc += tsz;
  145. /* leave now if filled buffer already */
  146. if (buflen == 0)
  147. return acc;
  148. }
  149. list_for_each_entry(m, &vmcore_list, list) {
  150. if (*fpos < m->offset + m->size) {
  151. tsz = min_t(size_t, m->offset + m->size - *fpos, buflen);
  152. start = m->paddr + *fpos - m->offset;
  153. tmp = read_from_oldmem(buffer, tsz, &start, 1);
  154. if (tmp < 0)
  155. return tmp;
  156. buflen -= tsz;
  157. *fpos += tsz;
  158. buffer += tsz;
  159. acc += tsz;
  160. /* leave now if filled buffer already */
  161. if (buflen == 0)
  162. return acc;
  163. }
  164. }
  165. return acc;
  166. }
  167. static const struct file_operations proc_vmcore_operations = {
  168. .read = read_vmcore,
  169. .llseek = default_llseek,
  170. };
  171. static struct vmcore* __init get_new_element(void)
  172. {
  173. return kzalloc(sizeof(struct vmcore), GFP_KERNEL);
  174. }
  175. static u64 __init get_vmcore_size(size_t elfsz, size_t elfnotesegsz,
  176. struct list_head *vc_list)
  177. {
  178. u64 size;
  179. struct vmcore *m;
  180. size = elfsz + elfnotesegsz;
  181. list_for_each_entry(m, vc_list, list) {
  182. size += m->size;
  183. }
  184. return size;
  185. }
  186. /**
  187. * update_note_header_size_elf64 - update p_memsz member of each PT_NOTE entry
  188. *
  189. * @ehdr_ptr: ELF header
  190. *
  191. * This function updates p_memsz member of each PT_NOTE entry in the
  192. * program header table pointed to by @ehdr_ptr to real size of ELF
  193. * note segment.
  194. */
  195. static int __init update_note_header_size_elf64(const Elf64_Ehdr *ehdr_ptr)
  196. {
  197. int i, rc=0;
  198. Elf64_Phdr *phdr_ptr;
  199. Elf64_Nhdr *nhdr_ptr;
  200. phdr_ptr = (Elf64_Phdr *)(ehdr_ptr + 1);
  201. for (i = 0; i < ehdr_ptr->e_phnum; i++, phdr_ptr++) {
  202. void *notes_section;
  203. u64 offset, max_sz, sz, real_sz = 0;
  204. if (phdr_ptr->p_type != PT_NOTE)
  205. continue;
  206. max_sz = phdr_ptr->p_memsz;
  207. offset = phdr_ptr->p_offset;
  208. notes_section = kmalloc(max_sz, GFP_KERNEL);
  209. if (!notes_section)
  210. return -ENOMEM;
  211. rc = read_from_oldmem(notes_section, max_sz, &offset, 0);
  212. if (rc < 0) {
  213. kfree(notes_section);
  214. return rc;
  215. }
  216. nhdr_ptr = notes_section;
  217. while (real_sz < max_sz) {
  218. if (nhdr_ptr->n_namesz == 0)
  219. break;
  220. sz = sizeof(Elf64_Nhdr) +
  221. ((nhdr_ptr->n_namesz + 3) & ~3) +
  222. ((nhdr_ptr->n_descsz + 3) & ~3);
  223. real_sz += sz;
  224. nhdr_ptr = (Elf64_Nhdr*)((char*)nhdr_ptr + sz);
  225. }
  226. kfree(notes_section);
  227. phdr_ptr->p_memsz = real_sz;
  228. }
  229. return 0;
  230. }
  231. /**
  232. * get_note_number_and_size_elf64 - get the number of PT_NOTE program
  233. * headers and sum of real size of their ELF note segment headers and
  234. * data.
  235. *
  236. * @ehdr_ptr: ELF header
  237. * @nr_ptnote: buffer for the number of PT_NOTE program headers
  238. * @sz_ptnote: buffer for size of unique PT_NOTE program header
  239. *
  240. * This function is used to merge multiple PT_NOTE program headers
  241. * into a unique single one. The resulting unique entry will have
  242. * @sz_ptnote in its phdr->p_mem.
  243. *
  244. * It is assumed that program headers with PT_NOTE type pointed to by
  245. * @ehdr_ptr has already been updated by update_note_header_size_elf64
  246. * and each of PT_NOTE program headers has actual ELF note segment
  247. * size in its p_memsz member.
  248. */
  249. static int __init get_note_number_and_size_elf64(const Elf64_Ehdr *ehdr_ptr,
  250. int *nr_ptnote, u64 *sz_ptnote)
  251. {
  252. int i;
  253. Elf64_Phdr *phdr_ptr;
  254. *nr_ptnote = *sz_ptnote = 0;
  255. phdr_ptr = (Elf64_Phdr *)(ehdr_ptr + 1);
  256. for (i = 0; i < ehdr_ptr->e_phnum; i++, phdr_ptr++) {
  257. if (phdr_ptr->p_type != PT_NOTE)
  258. continue;
  259. *nr_ptnote += 1;
  260. *sz_ptnote += phdr_ptr->p_memsz;
  261. }
  262. return 0;
  263. }
  264. /**
  265. * copy_notes_elf64 - copy ELF note segments in a given buffer
  266. *
  267. * @ehdr_ptr: ELF header
  268. * @notes_buf: buffer into which ELF note segments are copied
  269. *
  270. * This function is used to copy ELF note segment in the 1st kernel
  271. * into the buffer @notes_buf in the 2nd kernel. It is assumed that
  272. * size of the buffer @notes_buf is equal to or larger than sum of the
  273. * real ELF note segment headers and data.
  274. *
  275. * It is assumed that program headers with PT_NOTE type pointed to by
  276. * @ehdr_ptr has already been updated by update_note_header_size_elf64
  277. * and each of PT_NOTE program headers has actual ELF note segment
  278. * size in its p_memsz member.
  279. */
  280. static int __init copy_notes_elf64(const Elf64_Ehdr *ehdr_ptr, char *notes_buf)
  281. {
  282. int i, rc=0;
  283. Elf64_Phdr *phdr_ptr;
  284. phdr_ptr = (Elf64_Phdr*)(ehdr_ptr + 1);
  285. for (i = 0; i < ehdr_ptr->e_phnum; i++, phdr_ptr++) {
  286. u64 offset;
  287. if (phdr_ptr->p_type != PT_NOTE)
  288. continue;
  289. offset = phdr_ptr->p_offset;
  290. rc = read_from_oldmem(notes_buf, phdr_ptr->p_memsz, &offset, 0);
  291. if (rc < 0)
  292. return rc;
  293. notes_buf += phdr_ptr->p_memsz;
  294. }
  295. return 0;
  296. }
  297. /* Merges all the PT_NOTE headers into one. */
  298. static int __init merge_note_headers_elf64(char *elfptr, size_t *elfsz,
  299. char **notes_buf, size_t *notes_sz)
  300. {
  301. int i, nr_ptnote=0, rc=0;
  302. char *tmp;
  303. Elf64_Ehdr *ehdr_ptr;
  304. Elf64_Phdr phdr;
  305. u64 phdr_sz = 0, note_off;
  306. struct vm_struct *vm;
  307. ehdr_ptr = (Elf64_Ehdr *)elfptr;
  308. rc = update_note_header_size_elf64(ehdr_ptr);
  309. if (rc < 0)
  310. return rc;
  311. rc = get_note_number_and_size_elf64(ehdr_ptr, &nr_ptnote, &phdr_sz);
  312. if (rc < 0)
  313. return rc;
  314. *notes_sz = roundup(phdr_sz, PAGE_SIZE);
  315. *notes_buf = vzalloc(*notes_sz);
  316. if (!*notes_buf)
  317. return -ENOMEM;
  318. /*
  319. * Allow users to remap ELF note segment buffer on vmalloc memory using
  320. * remap_vmalloc_range.()
  321. */
  322. vm = find_vm_area(*notes_buf);
  323. BUG_ON(!vm);
  324. vm->flags |= VM_USERMAP;
  325. rc = copy_notes_elf64(ehdr_ptr, *notes_buf);
  326. if (rc < 0)
  327. return rc;
  328. /* Prepare merged PT_NOTE program header. */
  329. phdr.p_type = PT_NOTE;
  330. phdr.p_flags = 0;
  331. note_off = sizeof(Elf64_Ehdr) +
  332. (ehdr_ptr->e_phnum - nr_ptnote +1) * sizeof(Elf64_Phdr);
  333. phdr.p_offset = roundup(note_off, PAGE_SIZE);
  334. phdr.p_vaddr = phdr.p_paddr = 0;
  335. phdr.p_filesz = phdr.p_memsz = phdr_sz;
  336. phdr.p_align = 0;
  337. /* Add merged PT_NOTE program header*/
  338. tmp = elfptr + sizeof(Elf64_Ehdr);
  339. memcpy(tmp, &phdr, sizeof(phdr));
  340. tmp += sizeof(phdr);
  341. /* Remove unwanted PT_NOTE program headers. */
  342. i = (nr_ptnote - 1) * sizeof(Elf64_Phdr);
  343. *elfsz = *elfsz - i;
  344. memmove(tmp, tmp+i, ((*elfsz)-sizeof(Elf64_Ehdr)-sizeof(Elf64_Phdr)));
  345. memset(elfptr + *elfsz, 0, i);
  346. *elfsz = roundup(*elfsz, PAGE_SIZE);
  347. /* Modify e_phnum to reflect merged headers. */
  348. ehdr_ptr->e_phnum = ehdr_ptr->e_phnum - nr_ptnote + 1;
  349. return 0;
  350. }
  351. /**
  352. * update_note_header_size_elf32 - update p_memsz member of each PT_NOTE entry
  353. *
  354. * @ehdr_ptr: ELF header
  355. *
  356. * This function updates p_memsz member of each PT_NOTE entry in the
  357. * program header table pointed to by @ehdr_ptr to real size of ELF
  358. * note segment.
  359. */
  360. static int __init update_note_header_size_elf32(const Elf32_Ehdr *ehdr_ptr)
  361. {
  362. int i, rc=0;
  363. Elf32_Phdr *phdr_ptr;
  364. Elf32_Nhdr *nhdr_ptr;
  365. phdr_ptr = (Elf32_Phdr *)(ehdr_ptr + 1);
  366. for (i = 0; i < ehdr_ptr->e_phnum; i++, phdr_ptr++) {
  367. void *notes_section;
  368. u64 offset, max_sz, sz, real_sz = 0;
  369. if (phdr_ptr->p_type != PT_NOTE)
  370. continue;
  371. max_sz = phdr_ptr->p_memsz;
  372. offset = phdr_ptr->p_offset;
  373. notes_section = kmalloc(max_sz, GFP_KERNEL);
  374. if (!notes_section)
  375. return -ENOMEM;
  376. rc = read_from_oldmem(notes_section, max_sz, &offset, 0);
  377. if (rc < 0) {
  378. kfree(notes_section);
  379. return rc;
  380. }
  381. nhdr_ptr = notes_section;
  382. while (real_sz < max_sz) {
  383. if (nhdr_ptr->n_namesz == 0)
  384. break;
  385. sz = sizeof(Elf32_Nhdr) +
  386. ((nhdr_ptr->n_namesz + 3) & ~3) +
  387. ((nhdr_ptr->n_descsz + 3) & ~3);
  388. real_sz += sz;
  389. nhdr_ptr = (Elf32_Nhdr*)((char*)nhdr_ptr + sz);
  390. }
  391. kfree(notes_section);
  392. phdr_ptr->p_memsz = real_sz;
  393. }
  394. return 0;
  395. }
  396. /**
  397. * get_note_number_and_size_elf32 - get the number of PT_NOTE program
  398. * headers and sum of real size of their ELF note segment headers and
  399. * data.
  400. *
  401. * @ehdr_ptr: ELF header
  402. * @nr_ptnote: buffer for the number of PT_NOTE program headers
  403. * @sz_ptnote: buffer for size of unique PT_NOTE program header
  404. *
  405. * This function is used to merge multiple PT_NOTE program headers
  406. * into a unique single one. The resulting unique entry will have
  407. * @sz_ptnote in its phdr->p_mem.
  408. *
  409. * It is assumed that program headers with PT_NOTE type pointed to by
  410. * @ehdr_ptr has already been updated by update_note_header_size_elf32
  411. * and each of PT_NOTE program headers has actual ELF note segment
  412. * size in its p_memsz member.
  413. */
  414. static int __init get_note_number_and_size_elf32(const Elf32_Ehdr *ehdr_ptr,
  415. int *nr_ptnote, u64 *sz_ptnote)
  416. {
  417. int i;
  418. Elf32_Phdr *phdr_ptr;
  419. *nr_ptnote = *sz_ptnote = 0;
  420. phdr_ptr = (Elf32_Phdr *)(ehdr_ptr + 1);
  421. for (i = 0; i < ehdr_ptr->e_phnum; i++, phdr_ptr++) {
  422. if (phdr_ptr->p_type != PT_NOTE)
  423. continue;
  424. *nr_ptnote += 1;
  425. *sz_ptnote += phdr_ptr->p_memsz;
  426. }
  427. return 0;
  428. }
  429. /**
  430. * copy_notes_elf32 - copy ELF note segments in a given buffer
  431. *
  432. * @ehdr_ptr: ELF header
  433. * @notes_buf: buffer into which ELF note segments are copied
  434. *
  435. * This function is used to copy ELF note segment in the 1st kernel
  436. * into the buffer @notes_buf in the 2nd kernel. It is assumed that
  437. * size of the buffer @notes_buf is equal to or larger than sum of the
  438. * real ELF note segment headers and data.
  439. *
  440. * It is assumed that program headers with PT_NOTE type pointed to by
  441. * @ehdr_ptr has already been updated by update_note_header_size_elf32
  442. * and each of PT_NOTE program headers has actual ELF note segment
  443. * size in its p_memsz member.
  444. */
  445. static int __init copy_notes_elf32(const Elf32_Ehdr *ehdr_ptr, char *notes_buf)
  446. {
  447. int i, rc=0;
  448. Elf32_Phdr *phdr_ptr;
  449. phdr_ptr = (Elf32_Phdr*)(ehdr_ptr + 1);
  450. for (i = 0; i < ehdr_ptr->e_phnum; i++, phdr_ptr++) {
  451. u64 offset;
  452. if (phdr_ptr->p_type != PT_NOTE)
  453. continue;
  454. offset = phdr_ptr->p_offset;
  455. rc = read_from_oldmem(notes_buf, phdr_ptr->p_memsz, &offset, 0);
  456. if (rc < 0)
  457. return rc;
  458. notes_buf += phdr_ptr->p_memsz;
  459. }
  460. return 0;
  461. }
  462. /* Merges all the PT_NOTE headers into one. */
  463. static int __init merge_note_headers_elf32(char *elfptr, size_t *elfsz,
  464. char **notes_buf, size_t *notes_sz)
  465. {
  466. int i, nr_ptnote=0, rc=0;
  467. char *tmp;
  468. Elf32_Ehdr *ehdr_ptr;
  469. Elf32_Phdr phdr;
  470. u64 phdr_sz = 0, note_off;
  471. struct vm_struct *vm;
  472. ehdr_ptr = (Elf32_Ehdr *)elfptr;
  473. rc = update_note_header_size_elf32(ehdr_ptr);
  474. if (rc < 0)
  475. return rc;
  476. rc = get_note_number_and_size_elf32(ehdr_ptr, &nr_ptnote, &phdr_sz);
  477. if (rc < 0)
  478. return rc;
  479. *notes_sz = roundup(phdr_sz, PAGE_SIZE);
  480. *notes_buf = vzalloc(*notes_sz);
  481. if (!*notes_buf)
  482. return -ENOMEM;
  483. /*
  484. * Allow users to remap ELF note segment buffer on vmalloc memory using
  485. * remap_vmalloc_range()
  486. */
  487. vm = find_vm_area(*notes_buf);
  488. BUG_ON(!vm);
  489. vm->flags |= VM_USERMAP;
  490. rc = copy_notes_elf32(ehdr_ptr, *notes_buf);
  491. if (rc < 0)
  492. return rc;
  493. /* Prepare merged PT_NOTE program header. */
  494. phdr.p_type = PT_NOTE;
  495. phdr.p_flags = 0;
  496. note_off = sizeof(Elf32_Ehdr) +
  497. (ehdr_ptr->e_phnum - nr_ptnote +1) * sizeof(Elf32_Phdr);
  498. phdr.p_offset = roundup(note_off, PAGE_SIZE);
  499. phdr.p_vaddr = phdr.p_paddr = 0;
  500. phdr.p_filesz = phdr.p_memsz = phdr_sz;
  501. phdr.p_align = 0;
  502. /* Add merged PT_NOTE program header*/
  503. tmp = elfptr + sizeof(Elf32_Ehdr);
  504. memcpy(tmp, &phdr, sizeof(phdr));
  505. tmp += sizeof(phdr);
  506. /* Remove unwanted PT_NOTE program headers. */
  507. i = (nr_ptnote - 1) * sizeof(Elf32_Phdr);
  508. *elfsz = *elfsz - i;
  509. memmove(tmp, tmp+i, ((*elfsz)-sizeof(Elf32_Ehdr)-sizeof(Elf32_Phdr)));
  510. memset(elfptr + *elfsz, 0, i);
  511. *elfsz = roundup(*elfsz, PAGE_SIZE);
  512. /* Modify e_phnum to reflect merged headers. */
  513. ehdr_ptr->e_phnum = ehdr_ptr->e_phnum - nr_ptnote + 1;
  514. return 0;
  515. }
  516. /* Add memory chunks represented by program headers to vmcore list. Also update
  517. * the new offset fields of exported program headers. */
  518. static int __init process_ptload_program_headers_elf64(char *elfptr,
  519. size_t elfsz,
  520. size_t elfnotes_sz,
  521. struct list_head *vc_list)
  522. {
  523. int i;
  524. Elf64_Ehdr *ehdr_ptr;
  525. Elf64_Phdr *phdr_ptr;
  526. loff_t vmcore_off;
  527. struct vmcore *new;
  528. ehdr_ptr = (Elf64_Ehdr *)elfptr;
  529. phdr_ptr = (Elf64_Phdr*)(elfptr + sizeof(Elf64_Ehdr)); /* PT_NOTE hdr */
  530. /* Skip Elf header, program headers and Elf note segment. */
  531. vmcore_off = elfsz + elfnotes_sz;
  532. for (i = 0; i < ehdr_ptr->e_phnum; i++, phdr_ptr++) {
  533. u64 paddr, start, end, size;
  534. if (phdr_ptr->p_type != PT_LOAD)
  535. continue;
  536. paddr = phdr_ptr->p_offset;
  537. start = rounddown(paddr, PAGE_SIZE);
  538. end = roundup(paddr + phdr_ptr->p_memsz, PAGE_SIZE);
  539. size = end - start;
  540. /* Add this contiguous chunk of memory to vmcore list.*/
  541. new = get_new_element();
  542. if (!new)
  543. return -ENOMEM;
  544. new->paddr = start;
  545. new->size = size;
  546. list_add_tail(&new->list, vc_list);
  547. /* Update the program header offset. */
  548. phdr_ptr->p_offset = vmcore_off + (paddr - start);
  549. vmcore_off = vmcore_off + size;
  550. }
  551. return 0;
  552. }
  553. static int __init process_ptload_program_headers_elf32(char *elfptr,
  554. size_t elfsz,
  555. size_t elfnotes_sz,
  556. struct list_head *vc_list)
  557. {
  558. int i;
  559. Elf32_Ehdr *ehdr_ptr;
  560. Elf32_Phdr *phdr_ptr;
  561. loff_t vmcore_off;
  562. struct vmcore *new;
  563. ehdr_ptr = (Elf32_Ehdr *)elfptr;
  564. phdr_ptr = (Elf32_Phdr*)(elfptr + sizeof(Elf32_Ehdr)); /* PT_NOTE hdr */
  565. /* Skip Elf header, program headers and Elf note segment. */
  566. vmcore_off = elfsz + elfnotes_sz;
  567. for (i = 0; i < ehdr_ptr->e_phnum; i++, phdr_ptr++) {
  568. u64 paddr, start, end, size;
  569. if (phdr_ptr->p_type != PT_LOAD)
  570. continue;
  571. paddr = phdr_ptr->p_offset;
  572. start = rounddown(paddr, PAGE_SIZE);
  573. end = roundup(paddr + phdr_ptr->p_memsz, PAGE_SIZE);
  574. size = end - start;
  575. /* Add this contiguous chunk of memory to vmcore list.*/
  576. new = get_new_element();
  577. if (!new)
  578. return -ENOMEM;
  579. new->paddr = start;
  580. new->size = size;
  581. list_add_tail(&new->list, vc_list);
  582. /* Update the program header offset */
  583. phdr_ptr->p_offset = vmcore_off + (paddr - start);
  584. vmcore_off = vmcore_off + size;
  585. }
  586. return 0;
  587. }
  588. /* Sets offset fields of vmcore elements. */
  589. static void __init set_vmcore_list_offsets(size_t elfsz, size_t elfnotes_sz,
  590. struct list_head *vc_list)
  591. {
  592. loff_t vmcore_off;
  593. struct vmcore *m;
  594. /* Skip Elf header, program headers and Elf note segment. */
  595. vmcore_off = elfsz + elfnotes_sz;
  596. list_for_each_entry(m, vc_list, list) {
  597. m->offset = vmcore_off;
  598. vmcore_off += m->size;
  599. }
  600. }
  601. static void free_elfcorebuf(void)
  602. {
  603. free_pages((unsigned long)elfcorebuf, get_order(elfcorebuf_sz_orig));
  604. elfcorebuf = NULL;
  605. vfree(elfnotes_buf);
  606. elfnotes_buf = NULL;
  607. }
  608. static int __init parse_crash_elf64_headers(void)
  609. {
  610. int rc=0;
  611. Elf64_Ehdr ehdr;
  612. u64 addr;
  613. addr = elfcorehdr_addr;
  614. /* Read Elf header */
  615. rc = read_from_oldmem((char*)&ehdr, sizeof(Elf64_Ehdr), &addr, 0);
  616. if (rc < 0)
  617. return rc;
  618. /* Do some basic Verification. */
  619. if (memcmp(ehdr.e_ident, ELFMAG, SELFMAG) != 0 ||
  620. (ehdr.e_type != ET_CORE) ||
  621. !vmcore_elf64_check_arch(&ehdr) ||
  622. ehdr.e_ident[EI_CLASS] != ELFCLASS64 ||
  623. ehdr.e_ident[EI_VERSION] != EV_CURRENT ||
  624. ehdr.e_version != EV_CURRENT ||
  625. ehdr.e_ehsize != sizeof(Elf64_Ehdr) ||
  626. ehdr.e_phentsize != sizeof(Elf64_Phdr) ||
  627. ehdr.e_phnum == 0) {
  628. pr_warn("Warning: Core image elf header is not sane\n");
  629. return -EINVAL;
  630. }
  631. /* Read in all elf headers. */
  632. elfcorebuf_sz_orig = sizeof(Elf64_Ehdr) +
  633. ehdr.e_phnum * sizeof(Elf64_Phdr);
  634. elfcorebuf_sz = elfcorebuf_sz_orig;
  635. elfcorebuf = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO,
  636. get_order(elfcorebuf_sz_orig));
  637. if (!elfcorebuf)
  638. return -ENOMEM;
  639. addr = elfcorehdr_addr;
  640. rc = read_from_oldmem(elfcorebuf, elfcorebuf_sz_orig, &addr, 0);
  641. if (rc < 0)
  642. goto fail;
  643. /* Merge all PT_NOTE headers into one. */
  644. rc = merge_note_headers_elf64(elfcorebuf, &elfcorebuf_sz,
  645. &elfnotes_buf, &elfnotes_sz);
  646. if (rc)
  647. goto fail;
  648. rc = process_ptload_program_headers_elf64(elfcorebuf, elfcorebuf_sz,
  649. elfnotes_sz, &vmcore_list);
  650. if (rc)
  651. goto fail;
  652. set_vmcore_list_offsets(elfcorebuf_sz, elfnotes_sz, &vmcore_list);
  653. return 0;
  654. fail:
  655. free_elfcorebuf();
  656. return rc;
  657. }
  658. static int __init parse_crash_elf32_headers(void)
  659. {
  660. int rc=0;
  661. Elf32_Ehdr ehdr;
  662. u64 addr;
  663. addr = elfcorehdr_addr;
  664. /* Read Elf header */
  665. rc = read_from_oldmem((char*)&ehdr, sizeof(Elf32_Ehdr), &addr, 0);
  666. if (rc < 0)
  667. return rc;
  668. /* Do some basic Verification. */
  669. if (memcmp(ehdr.e_ident, ELFMAG, SELFMAG) != 0 ||
  670. (ehdr.e_type != ET_CORE) ||
  671. !elf_check_arch(&ehdr) ||
  672. ehdr.e_ident[EI_CLASS] != ELFCLASS32||
  673. ehdr.e_ident[EI_VERSION] != EV_CURRENT ||
  674. ehdr.e_version != EV_CURRENT ||
  675. ehdr.e_ehsize != sizeof(Elf32_Ehdr) ||
  676. ehdr.e_phentsize != sizeof(Elf32_Phdr) ||
  677. ehdr.e_phnum == 0) {
  678. pr_warn("Warning: Core image elf header is not sane\n");
  679. return -EINVAL;
  680. }
  681. /* Read in all elf headers. */
  682. elfcorebuf_sz_orig = sizeof(Elf32_Ehdr) + ehdr.e_phnum * sizeof(Elf32_Phdr);
  683. elfcorebuf_sz = elfcorebuf_sz_orig;
  684. elfcorebuf = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO,
  685. get_order(elfcorebuf_sz_orig));
  686. if (!elfcorebuf)
  687. return -ENOMEM;
  688. addr = elfcorehdr_addr;
  689. rc = read_from_oldmem(elfcorebuf, elfcorebuf_sz_orig, &addr, 0);
  690. if (rc < 0)
  691. goto fail;
  692. /* Merge all PT_NOTE headers into one. */
  693. rc = merge_note_headers_elf32(elfcorebuf, &elfcorebuf_sz,
  694. &elfnotes_buf, &elfnotes_sz);
  695. if (rc)
  696. goto fail;
  697. rc = process_ptload_program_headers_elf32(elfcorebuf, elfcorebuf_sz,
  698. elfnotes_sz, &vmcore_list);
  699. if (rc)
  700. goto fail;
  701. set_vmcore_list_offsets(elfcorebuf_sz, elfnotes_sz, &vmcore_list);
  702. return 0;
  703. fail:
  704. free_elfcorebuf();
  705. return rc;
  706. }
  707. static int __init parse_crash_elf_headers(void)
  708. {
  709. unsigned char e_ident[EI_NIDENT];
  710. u64 addr;
  711. int rc=0;
  712. addr = elfcorehdr_addr;
  713. rc = read_from_oldmem(e_ident, EI_NIDENT, &addr, 0);
  714. if (rc < 0)
  715. return rc;
  716. if (memcmp(e_ident, ELFMAG, SELFMAG) != 0) {
  717. pr_warn("Warning: Core image elf header not found\n");
  718. return -EINVAL;
  719. }
  720. if (e_ident[EI_CLASS] == ELFCLASS64) {
  721. rc = parse_crash_elf64_headers();
  722. if (rc)
  723. return rc;
  724. } else if (e_ident[EI_CLASS] == ELFCLASS32) {
  725. rc = parse_crash_elf32_headers();
  726. if (rc)
  727. return rc;
  728. } else {
  729. pr_warn("Warning: Core image elf header is not sane\n");
  730. return -EINVAL;
  731. }
  732. /* Determine vmcore size. */
  733. vmcore_size = get_vmcore_size(elfcorebuf_sz, elfnotes_sz,
  734. &vmcore_list);
  735. return 0;
  736. }
  737. /* Init function for vmcore module. */
  738. static int __init vmcore_init(void)
  739. {
  740. int rc = 0;
  741. /* If elfcorehdr= has been passed in cmdline, then capture the dump.*/
  742. if (!(is_vmcore_usable()))
  743. return rc;
  744. rc = parse_crash_elf_headers();
  745. if (rc) {
  746. pr_warn("Kdump: vmcore not initialized\n");
  747. return rc;
  748. }
  749. proc_vmcore = proc_create("vmcore", S_IRUSR, NULL, &proc_vmcore_operations);
  750. if (proc_vmcore)
  751. proc_vmcore->size = vmcore_size;
  752. return 0;
  753. }
  754. module_init(vmcore_init)
  755. /* Cleanup function for vmcore module. */
  756. void vmcore_cleanup(void)
  757. {
  758. struct list_head *pos, *next;
  759. if (proc_vmcore) {
  760. proc_remove(proc_vmcore);
  761. proc_vmcore = NULL;
  762. }
  763. /* clear the vmcore list. */
  764. list_for_each_safe(pos, next, &vmcore_list) {
  765. struct vmcore *m;
  766. m = list_entry(pos, struct vmcore, list);
  767. list_del(&m->list);
  768. kfree(m);
  769. }
  770. free_elfcorebuf();
  771. }
  772. EXPORT_SYMBOL_GPL(vmcore_cleanup);