physmem.c 9.9 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451
  1. /*
  2. * Copyright (C) 2000 - 2003 Jeff Dike (jdike@addtoit.com)
  3. * Licensed under the GPL
  4. */
  5. #include "linux/mm.h"
  6. #include "linux/rbtree.h"
  7. #include "linux/slab.h"
  8. #include "linux/vmalloc.h"
  9. #include "linux/bootmem.h"
  10. #include "linux/module.h"
  11. #include "asm/types.h"
  12. #include "asm/pgtable.h"
  13. #include "kern_util.h"
  14. #include "user_util.h"
  15. #include "mode_kern.h"
  16. #include "mem.h"
  17. #include "mem_user.h"
  18. #include "os.h"
  19. #include "kern.h"
  20. #include "init.h"
  21. struct phys_desc {
  22. struct rb_node rb;
  23. int fd;
  24. __u64 offset;
  25. void *virt;
  26. unsigned long phys;
  27. struct list_head list;
  28. };
  29. static struct rb_root phys_mappings = RB_ROOT;
  30. static struct rb_node **find_rb(void *virt)
  31. {
  32. struct rb_node **n = &phys_mappings.rb_node;
  33. struct phys_desc *d;
  34. while(*n != NULL){
  35. d = rb_entry(*n, struct phys_desc, rb);
  36. if(d->virt == virt)
  37. return(n);
  38. if(d->virt > virt)
  39. n = &(*n)->rb_left;
  40. else
  41. n = &(*n)->rb_right;
  42. }
  43. return(n);
  44. }
  45. static struct phys_desc *find_phys_mapping(void *virt)
  46. {
  47. struct rb_node **n = find_rb(virt);
  48. if(*n == NULL)
  49. return(NULL);
  50. return(rb_entry(*n, struct phys_desc, rb));
  51. }
  52. static void insert_phys_mapping(struct phys_desc *desc)
  53. {
  54. struct rb_node **n = find_rb(desc->virt);
  55. if(*n != NULL)
  56. panic("Physical remapping for %p already present",
  57. desc->virt);
  58. rb_link_node(&desc->rb, (*n)->rb_parent, n);
  59. rb_insert_color(&desc->rb, &phys_mappings);
  60. }
  61. LIST_HEAD(descriptor_mappings);
  62. struct desc_mapping {
  63. int fd;
  64. struct list_head list;
  65. struct list_head pages;
  66. };
  67. static struct desc_mapping *find_mapping(int fd)
  68. {
  69. struct desc_mapping *desc;
  70. struct list_head *ele;
  71. list_for_each(ele, &descriptor_mappings){
  72. desc = list_entry(ele, struct desc_mapping, list);
  73. if(desc->fd == fd)
  74. return(desc);
  75. }
  76. return(NULL);
  77. }
  78. static struct desc_mapping *descriptor_mapping(int fd)
  79. {
  80. struct desc_mapping *desc;
  81. desc = find_mapping(fd);
  82. if(desc != NULL)
  83. return(desc);
  84. desc = kmalloc(sizeof(*desc), GFP_ATOMIC);
  85. if(desc == NULL)
  86. return(NULL);
  87. *desc = ((struct desc_mapping)
  88. { .fd = fd,
  89. .list = LIST_HEAD_INIT(desc->list),
  90. .pages = LIST_HEAD_INIT(desc->pages) });
  91. list_add(&desc->list, &descriptor_mappings);
  92. return(desc);
  93. }
  94. int physmem_subst_mapping(void *virt, int fd, __u64 offset, int w)
  95. {
  96. struct desc_mapping *fd_maps;
  97. struct phys_desc *desc;
  98. unsigned long phys;
  99. int err;
  100. fd_maps = descriptor_mapping(fd);
  101. if(fd_maps == NULL)
  102. return(-ENOMEM);
  103. phys = __pa(virt);
  104. desc = find_phys_mapping(virt);
  105. if(desc != NULL)
  106. panic("Address 0x%p is already substituted\n", virt);
  107. err = -ENOMEM;
  108. desc = kmalloc(sizeof(*desc), GFP_ATOMIC);
  109. if(desc == NULL)
  110. goto out;
  111. *desc = ((struct phys_desc)
  112. { .fd = fd,
  113. .offset = offset,
  114. .virt = virt,
  115. .phys = __pa(virt),
  116. .list = LIST_HEAD_INIT(desc->list) });
  117. insert_phys_mapping(desc);
  118. list_add(&desc->list, &fd_maps->pages);
  119. virt = (void *) ((unsigned long) virt & PAGE_MASK);
  120. err = os_map_memory(virt, fd, offset, PAGE_SIZE, 1, w, 0);
  121. if(!err)
  122. goto out;
  123. rb_erase(&desc->rb, &phys_mappings);
  124. kfree(desc);
  125. out:
  126. return(err);
  127. }
  128. static int physmem_fd = -1;
  129. static void remove_mapping(struct phys_desc *desc)
  130. {
  131. void *virt = desc->virt;
  132. int err;
  133. rb_erase(&desc->rb, &phys_mappings);
  134. list_del(&desc->list);
  135. kfree(desc);
  136. err = os_map_memory(virt, physmem_fd, __pa(virt), PAGE_SIZE, 1, 1, 0);
  137. if(err)
  138. panic("Failed to unmap block device page from physical memory, "
  139. "errno = %d", -err);
  140. }
  141. int physmem_remove_mapping(void *virt)
  142. {
  143. struct phys_desc *desc;
  144. virt = (void *) ((unsigned long) virt & PAGE_MASK);
  145. desc = find_phys_mapping(virt);
  146. if(desc == NULL)
  147. return(0);
  148. remove_mapping(desc);
  149. return(1);
  150. }
  151. void physmem_forget_descriptor(int fd)
  152. {
  153. struct desc_mapping *desc;
  154. struct phys_desc *page;
  155. struct list_head *ele, *next;
  156. __u64 offset;
  157. void *addr;
  158. int err;
  159. desc = find_mapping(fd);
  160. if(desc == NULL)
  161. return;
  162. list_for_each_safe(ele, next, &desc->pages){
  163. page = list_entry(ele, struct phys_desc, list);
  164. offset = page->offset;
  165. addr = page->virt;
  166. remove_mapping(page);
  167. err = os_seek_file(fd, offset);
  168. if(err)
  169. panic("physmem_forget_descriptor - failed to seek "
  170. "to %lld in fd %d, error = %d\n",
  171. offset, fd, -err);
  172. err = os_read_file(fd, addr, PAGE_SIZE);
  173. if(err < 0)
  174. panic("physmem_forget_descriptor - failed to read "
  175. "from fd %d to 0x%p, error = %d\n",
  176. fd, addr, -err);
  177. }
  178. list_del(&desc->list);
  179. kfree(desc);
  180. }
  181. EXPORT_SYMBOL(physmem_forget_descriptor);
  182. EXPORT_SYMBOL(physmem_remove_mapping);
  183. EXPORT_SYMBOL(physmem_subst_mapping);
  184. void arch_free_page(struct page *page, int order)
  185. {
  186. void *virt;
  187. int i;
  188. for(i = 0; i < (1 << order); i++){
  189. virt = __va(page_to_phys(page + i));
  190. physmem_remove_mapping(virt);
  191. }
  192. }
  193. int is_remapped(void *virt)
  194. {
  195. struct phys_desc *desc = find_phys_mapping(virt);
  196. return(desc != NULL);
  197. }
  198. /* Changed during early boot */
  199. unsigned long high_physmem;
  200. extern unsigned long long physmem_size;
  201. int init_maps(unsigned long physmem, unsigned long iomem, unsigned long highmem)
  202. {
  203. struct page *p, *map;
  204. unsigned long phys_len, phys_pages, highmem_len, highmem_pages;
  205. unsigned long iomem_len, iomem_pages, total_len, total_pages;
  206. int i;
  207. phys_pages = physmem >> PAGE_SHIFT;
  208. phys_len = phys_pages * sizeof(struct page);
  209. iomem_pages = iomem >> PAGE_SHIFT;
  210. iomem_len = iomem_pages * sizeof(struct page);
  211. highmem_pages = highmem >> PAGE_SHIFT;
  212. highmem_len = highmem_pages * sizeof(struct page);
  213. total_pages = phys_pages + iomem_pages + highmem_pages;
  214. total_len = phys_len + iomem_len + highmem_len;
  215. if(kmalloc_ok){
  216. map = kmalloc(total_len, GFP_KERNEL);
  217. if(map == NULL)
  218. map = vmalloc(total_len);
  219. }
  220. else map = alloc_bootmem_low_pages(total_len);
  221. if(map == NULL)
  222. return(-ENOMEM);
  223. for(i = 0; i < total_pages; i++){
  224. p = &map[i];
  225. set_page_count(p, 0);
  226. SetPageReserved(p);
  227. INIT_LIST_HEAD(&p->lru);
  228. }
  229. max_mapnr = total_pages;
  230. return(0);
  231. }
  232. /* Changed during early boot */
  233. static unsigned long kmem_top = 0;
  234. unsigned long get_kmem_end(void)
  235. {
  236. if(kmem_top == 0)
  237. kmem_top = CHOOSE_MODE(kmem_end_tt, kmem_end_skas);
  238. return(kmem_top);
  239. }
  240. void map_memory(unsigned long virt, unsigned long phys, unsigned long len,
  241. int r, int w, int x)
  242. {
  243. __u64 offset;
  244. int fd, err;
  245. fd = phys_mapping(phys, &offset);
  246. err = os_map_memory((void *) virt, fd, offset, len, r, w, x);
  247. if(err) {
  248. if(err == -ENOMEM)
  249. printk("try increasing the host's "
  250. "/proc/sys/vm/max_map_count to <physical "
  251. "memory size>/4096\n");
  252. panic("map_memory(0x%lx, %d, 0x%llx, %ld, %d, %d, %d) failed, "
  253. "err = %d\n", virt, fd, offset, len, r, w, x, err);
  254. }
  255. }
  256. #define PFN_UP(x) (((x) + PAGE_SIZE-1) >> PAGE_SHIFT)
  257. extern int __syscall_stub_start, __binary_start;
  258. void setup_physmem(unsigned long start, unsigned long reserve_end,
  259. unsigned long len, unsigned long long highmem)
  260. {
  261. unsigned long reserve = reserve_end - start;
  262. int pfn = PFN_UP(__pa(reserve_end));
  263. int delta = (len - reserve) >> PAGE_SHIFT;
  264. int err, offset, bootmap_size;
  265. physmem_fd = create_mem_file(len + highmem);
  266. offset = uml_reserved - uml_physmem;
  267. err = os_map_memory((void *) uml_reserved, physmem_fd, offset,
  268. len - offset, 1, 1, 0);
  269. if(err < 0){
  270. os_print_error(err, "Mapping memory");
  271. exit(1);
  272. }
  273. /* Special kludge - This page will be mapped in to userspace processes
  274. * from physmem_fd, so it needs to be written out there.
  275. */
  276. os_seek_file(physmem_fd, __pa(&__syscall_stub_start));
  277. os_write_file(physmem_fd, &__syscall_stub_start, PAGE_SIZE);
  278. bootmap_size = init_bootmem(pfn, pfn + delta);
  279. free_bootmem(__pa(reserve_end) + bootmap_size,
  280. len - bootmap_size - reserve);
  281. }
  282. int phys_mapping(unsigned long phys, __u64 *offset_out)
  283. {
  284. struct phys_desc *desc = find_phys_mapping(__va(phys & PAGE_MASK));
  285. int fd = -1;
  286. if(desc != NULL){
  287. fd = desc->fd;
  288. *offset_out = desc->offset;
  289. }
  290. else if(phys < physmem_size){
  291. fd = physmem_fd;
  292. *offset_out = phys;
  293. }
  294. else if(phys < __pa(end_iomem)){
  295. struct iomem_region *region = iomem_regions;
  296. while(region != NULL){
  297. if((phys >= region->phys) &&
  298. (phys < region->phys + region->size)){
  299. fd = region->fd;
  300. *offset_out = phys - region->phys;
  301. break;
  302. }
  303. region = region->next;
  304. }
  305. }
  306. else if(phys < __pa(end_iomem) + highmem){
  307. fd = physmem_fd;
  308. *offset_out = phys - iomem_size;
  309. }
  310. return(fd);
  311. }
  312. static int __init uml_mem_setup(char *line, int *add)
  313. {
  314. char *retptr;
  315. physmem_size = memparse(line,&retptr);
  316. return 0;
  317. }
  318. __uml_setup("mem=", uml_mem_setup,
  319. "mem=<Amount of desired ram>\n"
  320. " This controls how much \"physical\" memory the kernel allocates\n"
  321. " for the system. The size is specified as a number followed by\n"
  322. " one of 'k', 'K', 'm', 'M', which have the obvious meanings.\n"
  323. " This is not related to the amount of memory in the host. It can\n"
  324. " be more, and the excess, if it's ever used, will just be swapped out.\n"
  325. " Example: mem=64M\n\n"
  326. );
  327. unsigned long find_iomem(char *driver, unsigned long *len_out)
  328. {
  329. struct iomem_region *region = iomem_regions;
  330. while(region != NULL){
  331. if(!strcmp(region->driver, driver)){
  332. *len_out = region->size;
  333. return(region->virt);
  334. }
  335. }
  336. return(0);
  337. }
  338. int setup_iomem(void)
  339. {
  340. struct iomem_region *region = iomem_regions;
  341. unsigned long iomem_start = high_physmem + PAGE_SIZE;
  342. int err;
  343. while(region != NULL){
  344. err = os_map_memory((void *) iomem_start, region->fd, 0,
  345. region->size, 1, 1, 0);
  346. if(err)
  347. printk("Mapping iomem region for driver '%s' failed, "
  348. "errno = %d\n", region->driver, -err);
  349. else {
  350. region->virt = iomem_start;
  351. region->phys = __pa(region->virt);
  352. }
  353. iomem_start += region->size + PAGE_SIZE;
  354. region = region->next;
  355. }
  356. return(0);
  357. }
  358. __initcall(setup_iomem);
  359. /*
  360. * Overrides for Emacs so that we follow Linus's tabbing style.
  361. * Emacs will notice this stuff at the end of the file and automatically
  362. * adjust the settings for this buffer only. This must remain at the end
  363. * of the file.
  364. * ---------------------------------------------------------------------------
  365. * Local variables:
  366. * c-file-style: "linux"
  367. * End:
  368. */