physmem.c 11 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486
  1. /*
  2. * Copyright (C) 2000 - 2003 Jeff Dike (jdike@addtoit.com)
  3. * Licensed under the GPL
  4. */
  5. #include "linux/mm.h"
  6. #include "linux/rbtree.h"
  7. #include "linux/slab.h"
  8. #include "linux/vmalloc.h"
  9. #include "linux/bootmem.h"
  10. #include "linux/module.h"
  11. #include "asm/types.h"
  12. #include "asm/pgtable.h"
  13. #include "kern_util.h"
  14. #include "user_util.h"
  15. #include "mode_kern.h"
  16. #include "mem.h"
  17. #include "mem_user.h"
  18. #include "os.h"
  19. #include "kern.h"
  20. #include "init.h"
  21. struct phys_desc {
  22. struct rb_node rb;
  23. int fd;
  24. __u64 offset;
  25. void *virt;
  26. unsigned long phys;
  27. struct list_head list;
  28. };
  29. static struct rb_root phys_mappings = RB_ROOT;
  30. static struct rb_node **find_rb(void *virt)
  31. {
  32. struct rb_node **n = &phys_mappings.rb_node;
  33. struct phys_desc *d;
  34. while(*n != NULL){
  35. d = rb_entry(*n, struct phys_desc, rb);
  36. if(d->virt == virt)
  37. return(n);
  38. if(d->virt > virt)
  39. n = &(*n)->rb_left;
  40. else
  41. n = &(*n)->rb_right;
  42. }
  43. return(n);
  44. }
  45. static struct phys_desc *find_phys_mapping(void *virt)
  46. {
  47. struct rb_node **n = find_rb(virt);
  48. if(*n == NULL)
  49. return(NULL);
  50. return(rb_entry(*n, struct phys_desc, rb));
  51. }
  52. static void insert_phys_mapping(struct phys_desc *desc)
  53. {
  54. struct rb_node **n = find_rb(desc->virt);
  55. if(*n != NULL)
  56. panic("Physical remapping for %p already present",
  57. desc->virt);
  58. rb_link_node(&desc->rb, (*n)->rb_parent, n);
  59. rb_insert_color(&desc->rb, &phys_mappings);
  60. }
  61. LIST_HEAD(descriptor_mappings);
  62. struct desc_mapping {
  63. int fd;
  64. struct list_head list;
  65. struct list_head pages;
  66. };
  67. static struct desc_mapping *find_mapping(int fd)
  68. {
  69. struct desc_mapping *desc;
  70. struct list_head *ele;
  71. list_for_each(ele, &descriptor_mappings){
  72. desc = list_entry(ele, struct desc_mapping, list);
  73. if(desc->fd == fd)
  74. return(desc);
  75. }
  76. return(NULL);
  77. }
  78. static struct desc_mapping *descriptor_mapping(int fd)
  79. {
  80. struct desc_mapping *desc;
  81. desc = find_mapping(fd);
  82. if(desc != NULL)
  83. return(desc);
  84. desc = kmalloc(sizeof(*desc), GFP_ATOMIC);
  85. if(desc == NULL)
  86. return(NULL);
  87. *desc = ((struct desc_mapping)
  88. { .fd = fd,
  89. .list = LIST_HEAD_INIT(desc->list),
  90. .pages = LIST_HEAD_INIT(desc->pages) });
  91. list_add(&desc->list, &descriptor_mappings);
  92. return(desc);
  93. }
  94. int physmem_subst_mapping(void *virt, int fd, __u64 offset, int w)
  95. {
  96. struct desc_mapping *fd_maps;
  97. struct phys_desc *desc;
  98. unsigned long phys;
  99. int err;
  100. fd_maps = descriptor_mapping(fd);
  101. if(fd_maps == NULL)
  102. return(-ENOMEM);
  103. phys = __pa(virt);
  104. desc = find_phys_mapping(virt);
  105. if(desc != NULL)
  106. panic("Address 0x%p is already substituted\n", virt);
  107. err = -ENOMEM;
  108. desc = kmalloc(sizeof(*desc), GFP_ATOMIC);
  109. if(desc == NULL)
  110. goto out;
  111. *desc = ((struct phys_desc)
  112. { .fd = fd,
  113. .offset = offset,
  114. .virt = virt,
  115. .phys = __pa(virt),
  116. .list = LIST_HEAD_INIT(desc->list) });
  117. insert_phys_mapping(desc);
  118. list_add(&desc->list, &fd_maps->pages);
  119. virt = (void *) ((unsigned long) virt & PAGE_MASK);
  120. err = os_map_memory(virt, fd, offset, PAGE_SIZE, 1, w, 0);
  121. if(!err)
  122. goto out;
  123. rb_erase(&desc->rb, &phys_mappings);
  124. kfree(desc);
  125. out:
  126. return(err);
  127. }
  128. static int physmem_fd = -1;
  129. static void remove_mapping(struct phys_desc *desc)
  130. {
  131. void *virt = desc->virt;
  132. int err;
  133. rb_erase(&desc->rb, &phys_mappings);
  134. list_del(&desc->list);
  135. kfree(desc);
  136. err = os_map_memory(virt, physmem_fd, __pa(virt), PAGE_SIZE, 1, 1, 0);
  137. if(err)
  138. panic("Failed to unmap block device page from physical memory, "
  139. "errno = %d", -err);
  140. }
  141. int physmem_remove_mapping(void *virt)
  142. {
  143. struct phys_desc *desc;
  144. virt = (void *) ((unsigned long) virt & PAGE_MASK);
  145. desc = find_phys_mapping(virt);
  146. if(desc == NULL)
  147. return(0);
  148. remove_mapping(desc);
  149. return(1);
  150. }
  151. void physmem_forget_descriptor(int fd)
  152. {
  153. struct desc_mapping *desc;
  154. struct phys_desc *page;
  155. struct list_head *ele, *next;
  156. __u64 offset;
  157. void *addr;
  158. int err;
  159. desc = find_mapping(fd);
  160. if(desc == NULL)
  161. return;
  162. list_for_each_safe(ele, next, &desc->pages){
  163. page = list_entry(ele, struct phys_desc, list);
  164. offset = page->offset;
  165. addr = page->virt;
  166. remove_mapping(page);
  167. err = os_seek_file(fd, offset);
  168. if(err)
  169. panic("physmem_forget_descriptor - failed to seek "
  170. "to %lld in fd %d, error = %d\n",
  171. offset, fd, -err);
  172. err = os_read_file(fd, addr, PAGE_SIZE);
  173. if(err < 0)
  174. panic("physmem_forget_descriptor - failed to read "
  175. "from fd %d to 0x%p, error = %d\n",
  176. fd, addr, -err);
  177. }
  178. list_del(&desc->list);
  179. kfree(desc);
  180. }
  181. EXPORT_SYMBOL(physmem_forget_descriptor);
  182. EXPORT_SYMBOL(physmem_remove_mapping);
  183. EXPORT_SYMBOL(physmem_subst_mapping);
  184. void arch_free_page(struct page *page, int order)
  185. {
  186. void *virt;
  187. int i;
  188. for(i = 0; i < (1 << order); i++){
  189. virt = __va(page_to_phys(page + i));
  190. physmem_remove_mapping(virt);
  191. }
  192. }
  193. int is_remapped(void *virt)
  194. {
  195. struct phys_desc *desc = find_phys_mapping(virt);
  196. return(desc != NULL);
  197. }
  198. /* Changed during early boot */
  199. unsigned long high_physmem;
  200. extern unsigned long physmem_size;
  201. void *to_virt(unsigned long phys)
  202. {
  203. return((void *) uml_physmem + phys);
  204. }
  205. unsigned long to_phys(void *virt)
  206. {
  207. return(((unsigned long) virt) - uml_physmem);
  208. }
  209. int init_maps(unsigned long physmem, unsigned long iomem, unsigned long highmem)
  210. {
  211. struct page *p, *map;
  212. unsigned long phys_len, phys_pages, highmem_len, highmem_pages;
  213. unsigned long iomem_len, iomem_pages, total_len, total_pages;
  214. int i;
  215. phys_pages = physmem >> PAGE_SHIFT;
  216. phys_len = phys_pages * sizeof(struct page);
  217. iomem_pages = iomem >> PAGE_SHIFT;
  218. iomem_len = iomem_pages * sizeof(struct page);
  219. highmem_pages = highmem >> PAGE_SHIFT;
  220. highmem_len = highmem_pages * sizeof(struct page);
  221. total_pages = phys_pages + iomem_pages + highmem_pages;
  222. total_len = phys_len + iomem_pages + highmem_len;
  223. if(kmalloc_ok){
  224. map = kmalloc(total_len, GFP_KERNEL);
  225. if(map == NULL)
  226. map = vmalloc(total_len);
  227. }
  228. else map = alloc_bootmem_low_pages(total_len);
  229. if(map == NULL)
  230. return(-ENOMEM);
  231. for(i = 0; i < total_pages; i++){
  232. p = &map[i];
  233. set_page_count(p, 0);
  234. SetPageReserved(p);
  235. INIT_LIST_HEAD(&p->lru);
  236. }
  237. max_mapnr = total_pages;
  238. return(0);
  239. }
  240. struct page *phys_to_page(const unsigned long phys)
  241. {
  242. return(&mem_map[phys >> PAGE_SHIFT]);
  243. }
  244. struct page *__virt_to_page(const unsigned long virt)
  245. {
  246. return(&mem_map[__pa(virt) >> PAGE_SHIFT]);
  247. }
  248. phys_t page_to_phys(struct page *page)
  249. {
  250. return((page - mem_map) << PAGE_SHIFT);
  251. }
  252. pte_t mk_pte(struct page *page, pgprot_t pgprot)
  253. {
  254. pte_t pte;
  255. pte_set_val(pte, page_to_phys(page), pgprot);
  256. if(pte_present(pte))
  257. pte_mknewprot(pte_mknewpage(pte));
  258. return(pte);
  259. }
  260. /* Changed during early boot */
  261. static unsigned long kmem_top = 0;
  262. unsigned long get_kmem_end(void)
  263. {
  264. if(kmem_top == 0)
  265. kmem_top = CHOOSE_MODE(kmem_end_tt, kmem_end_skas);
  266. return(kmem_top);
  267. }
  268. void map_memory(unsigned long virt, unsigned long phys, unsigned long len,
  269. int r, int w, int x)
  270. {
  271. __u64 offset;
  272. int fd, err;
  273. fd = phys_mapping(phys, &offset);
  274. err = os_map_memory((void *) virt, fd, offset, len, r, w, x);
  275. if(err) {
  276. if(err == -ENOMEM)
  277. printk("try increasing the host's "
  278. "/proc/sys/vm/max_map_count to <physical "
  279. "memory size>/4096\n");
  280. panic("map_memory(0x%lx, %d, 0x%llx, %ld, %d, %d, %d) failed, "
  281. "err = %d\n", virt, fd, offset, len, r, w, x, err);
  282. }
  283. }
  284. #define PFN_UP(x) (((x) + PAGE_SIZE-1) >> PAGE_SHIFT)
  285. extern int __syscall_stub_start, __binary_start;
  286. void setup_physmem(unsigned long start, unsigned long reserve_end,
  287. unsigned long len, unsigned long highmem)
  288. {
  289. unsigned long reserve = reserve_end - start;
  290. int pfn = PFN_UP(__pa(reserve_end));
  291. int delta = (len - reserve) >> PAGE_SHIFT;
  292. int err, offset, bootmap_size;
  293. physmem_fd = create_mem_file(len + highmem);
  294. offset = uml_reserved - uml_physmem;
  295. err = os_map_memory((void *) uml_reserved, physmem_fd, offset,
  296. len - offset, 1, 1, 0);
  297. if(err < 0){
  298. os_print_error(err, "Mapping memory");
  299. exit(1);
  300. }
  301. /* Special kludge - This page will be mapped in to userspace processes
  302. * from physmem_fd, so it needs to be written out there.
  303. */
  304. os_seek_file(physmem_fd, __pa(&__syscall_stub_start));
  305. os_write_file(physmem_fd, &__syscall_stub_start, PAGE_SIZE);
  306. bootmap_size = init_bootmem(pfn, pfn + delta);
  307. free_bootmem(__pa(reserve_end) + bootmap_size,
  308. len - bootmap_size - reserve);
  309. }
  310. int phys_mapping(unsigned long phys, __u64 *offset_out)
  311. {
  312. struct phys_desc *desc = find_phys_mapping(__va(phys & PAGE_MASK));
  313. int fd = -1;
  314. if(desc != NULL){
  315. fd = desc->fd;
  316. *offset_out = desc->offset;
  317. }
  318. else if(phys < physmem_size){
  319. fd = physmem_fd;
  320. *offset_out = phys;
  321. }
  322. else if(phys < __pa(end_iomem)){
  323. struct iomem_region *region = iomem_regions;
  324. while(region != NULL){
  325. if((phys >= region->phys) &&
  326. (phys < region->phys + region->size)){
  327. fd = region->fd;
  328. *offset_out = phys - region->phys;
  329. break;
  330. }
  331. region = region->next;
  332. }
  333. }
  334. else if(phys < __pa(end_iomem) + highmem){
  335. fd = physmem_fd;
  336. *offset_out = phys - iomem_size;
  337. }
  338. return(fd);
  339. }
  340. static int __init uml_mem_setup(char *line, int *add)
  341. {
  342. char *retptr;
  343. physmem_size = memparse(line,&retptr);
  344. return 0;
  345. }
  346. __uml_setup("mem=", uml_mem_setup,
  347. "mem=<Amount of desired ram>\n"
  348. " This controls how much \"physical\" memory the kernel allocates\n"
  349. " for the system. The size is specified as a number followed by\n"
  350. " one of 'k', 'K', 'm', 'M', which have the obvious meanings.\n"
  351. " This is not related to the amount of memory in the host. It can\n"
  352. " be more, and the excess, if it's ever used, will just be swapped out.\n"
  353. " Example: mem=64M\n\n"
  354. );
  355. unsigned long find_iomem(char *driver, unsigned long *len_out)
  356. {
  357. struct iomem_region *region = iomem_regions;
  358. while(region != NULL){
  359. if(!strcmp(region->driver, driver)){
  360. *len_out = region->size;
  361. return(region->virt);
  362. }
  363. }
  364. return(0);
  365. }
  366. int setup_iomem(void)
  367. {
  368. struct iomem_region *region = iomem_regions;
  369. unsigned long iomem_start = high_physmem + PAGE_SIZE;
  370. int err;
  371. while(region != NULL){
  372. err = os_map_memory((void *) iomem_start, region->fd, 0,
  373. region->size, 1, 1, 0);
  374. if(err)
  375. printk("Mapping iomem region for driver '%s' failed, "
  376. "errno = %d\n", region->driver, -err);
  377. else {
  378. region->virt = iomem_start;
  379. region->phys = __pa(region->virt);
  380. }
  381. iomem_start += region->size + PAGE_SIZE;
  382. region = region->next;
  383. }
  384. return(0);
  385. }
  386. __initcall(setup_iomem);
  387. /*
  388. * Overrides for Emacs so that we follow Linus's tabbing style.
  389. * Emacs will notice this stuff at the end of the file and automatically
  390. * adjust the settings for this buffer only. This must remain at the end
  391. * of the file.
  392. * ---------------------------------------------------------------------------
  393. * Local variables:
  394. * c-file-style: "linux"
  395. * End:
  396. */