alloc.c 14 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520
  1. /*
  2. * alloc.c - NILFS dat/inode allocator
  3. *
  4. * Copyright (C) 2006-2008 Nippon Telegraph and Telephone Corporation.
  5. *
  6. * This program is free software; you can redistribute it and/or modify
  7. * it under the terms of the GNU General Public License as published by
  8. * the Free Software Foundation; either version 2 of the License, or
  9. * (at your option) any later version.
  10. *
  11. * This program is distributed in the hope that it will be useful,
  12. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  13. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
  14. * GNU General Public License for more details.
  15. *
  16. * You should have received a copy of the GNU General Public License
  17. * along with this program; if not, write to the Free Software
  18. * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
  19. *
  20. * Original code was written by Koji Sato <koji@osrg.net>.
  21. * Two allocators were unified by Ryusuke Konishi <ryusuke@osrg.net>,
  22. * Amagai Yoshiji <amagai@osrg.net>.
  23. */
  24. #include <linux/types.h>
  25. #include <linux/buffer_head.h>
  26. #include <linux/fs.h>
  27. #include <linux/bitops.h>
  28. #include "mdt.h"
  29. #include "alloc.h"
  30. static inline unsigned long
  31. nilfs_palloc_groups_per_desc_block(const struct inode *inode)
  32. {
  33. return (1UL << inode->i_blkbits) /
  34. sizeof(struct nilfs_palloc_group_desc);
  35. }
  36. static inline unsigned long
  37. nilfs_palloc_groups_count(const struct inode *inode)
  38. {
  39. return 1UL << (BITS_PER_LONG - (inode->i_blkbits + 3 /* log2(8) */));
  40. }
  41. int nilfs_palloc_init_blockgroup(struct inode *inode, unsigned entry_size)
  42. {
  43. struct nilfs_mdt_info *mi = NILFS_MDT(inode);
  44. mi->mi_bgl = kmalloc(sizeof(*mi->mi_bgl), GFP_NOFS);
  45. if (!mi->mi_bgl)
  46. return -ENOMEM;
  47. bgl_lock_init(mi->mi_bgl);
  48. nilfs_mdt_set_entry_size(inode, entry_size, 0);
  49. mi->mi_blocks_per_group =
  50. DIV_ROUND_UP(nilfs_palloc_entries_per_group(inode),
  51. mi->mi_entries_per_block) + 1;
  52. /* Number of blocks in a group including entry blocks and
  53. a bitmap block */
  54. mi->mi_blocks_per_desc_block =
  55. nilfs_palloc_groups_per_desc_block(inode) *
  56. mi->mi_blocks_per_group + 1;
  57. /* Number of blocks per descriptor including the
  58. descriptor block */
  59. return 0;
  60. }
  61. static unsigned long nilfs_palloc_group(const struct inode *inode, __u64 nr,
  62. unsigned long *offset)
  63. {
  64. __u64 group = nr;
  65. *offset = do_div(group, nilfs_palloc_entries_per_group(inode));
  66. return group;
  67. }
  68. static unsigned long
  69. nilfs_palloc_desc_blkoff(const struct inode *inode, unsigned long group)
  70. {
  71. unsigned long desc_block =
  72. group / nilfs_palloc_groups_per_desc_block(inode);
  73. return desc_block * NILFS_MDT(inode)->mi_blocks_per_desc_block;
  74. }
  75. static unsigned long
  76. nilfs_palloc_bitmap_blkoff(const struct inode *inode, unsigned long group)
  77. {
  78. unsigned long desc_offset =
  79. group % nilfs_palloc_groups_per_desc_block(inode);
  80. return nilfs_palloc_desc_blkoff(inode, group) + 1 +
  81. desc_offset * NILFS_MDT(inode)->mi_blocks_per_group;
  82. }
  83. static unsigned long
  84. nilfs_palloc_group_desc_nfrees(struct inode *inode, unsigned long group,
  85. const struct nilfs_palloc_group_desc *desc)
  86. {
  87. unsigned long nfree;
  88. spin_lock(nilfs_mdt_bgl_lock(inode, group));
  89. nfree = le32_to_cpu(desc->pg_nfrees);
  90. spin_unlock(nilfs_mdt_bgl_lock(inode, group));
  91. return nfree;
  92. }
  93. static void
  94. nilfs_palloc_group_desc_add_entries(struct inode *inode,
  95. unsigned long group,
  96. struct nilfs_palloc_group_desc *desc,
  97. u32 n)
  98. {
  99. spin_lock(nilfs_mdt_bgl_lock(inode, group));
  100. le32_add_cpu(&desc->pg_nfrees, n);
  101. spin_unlock(nilfs_mdt_bgl_lock(inode, group));
  102. }
  103. static unsigned long
  104. nilfs_palloc_entry_blkoff(const struct inode *inode, __u64 nr)
  105. {
  106. unsigned long group, group_offset;
  107. group = nilfs_palloc_group(inode, nr, &group_offset);
  108. return nilfs_palloc_bitmap_blkoff(inode, group) + 1 +
  109. group_offset / NILFS_MDT(inode)->mi_entries_per_block;
  110. }
  111. static void nilfs_palloc_desc_block_init(struct inode *inode,
  112. struct buffer_head *bh, void *kaddr)
  113. {
  114. struct nilfs_palloc_group_desc *desc = kaddr + bh_offset(bh);
  115. unsigned long n = nilfs_palloc_groups_per_desc_block(inode);
  116. __le32 nfrees;
  117. nfrees = cpu_to_le32(nilfs_palloc_entries_per_group(inode));
  118. while (n-- > 0) {
  119. desc->pg_nfrees = nfrees;
  120. desc++;
  121. }
  122. }
  123. static int nilfs_palloc_get_desc_block(struct inode *inode,
  124. unsigned long group,
  125. int create, struct buffer_head **bhp)
  126. {
  127. return nilfs_mdt_get_block(inode,
  128. nilfs_palloc_desc_blkoff(inode, group),
  129. create, nilfs_palloc_desc_block_init, bhp);
  130. }
  131. static int nilfs_palloc_get_bitmap_block(struct inode *inode,
  132. unsigned long group,
  133. int create, struct buffer_head **bhp)
  134. {
  135. return nilfs_mdt_get_block(inode,
  136. nilfs_palloc_bitmap_blkoff(inode, group),
  137. create, NULL, bhp);
  138. }
  139. int nilfs_palloc_get_entry_block(struct inode *inode, __u64 nr,
  140. int create, struct buffer_head **bhp)
  141. {
  142. return nilfs_mdt_get_block(inode, nilfs_palloc_entry_blkoff(inode, nr),
  143. create, NULL, bhp);
  144. }
  145. static struct nilfs_palloc_group_desc *
  146. nilfs_palloc_block_get_group_desc(const struct inode *inode,
  147. unsigned long group,
  148. const struct buffer_head *bh, void *kaddr)
  149. {
  150. return (struct nilfs_palloc_group_desc *)(kaddr + bh_offset(bh)) +
  151. group % nilfs_palloc_groups_per_desc_block(inode);
  152. }
  153. void *nilfs_palloc_block_get_entry(const struct inode *inode, __u64 nr,
  154. const struct buffer_head *bh, void *kaddr)
  155. {
  156. unsigned long entry_offset, group_offset;
  157. nilfs_palloc_group(inode, nr, &group_offset);
  158. entry_offset = group_offset % NILFS_MDT(inode)->mi_entries_per_block;
  159. return kaddr + bh_offset(bh) +
  160. entry_offset * NILFS_MDT(inode)->mi_entry_size;
  161. }
  162. static int nilfs_palloc_find_available_slot(struct inode *inode,
  163. unsigned long group,
  164. unsigned long target,
  165. unsigned char *bitmap,
  166. int bsize) /* size in bits */
  167. {
  168. int curr, pos, end, i;
  169. if (target > 0) {
  170. end = (target + BITS_PER_LONG - 1) & ~(BITS_PER_LONG - 1);
  171. if (end > bsize)
  172. end = bsize;
  173. pos = nilfs_find_next_zero_bit(bitmap, end, target);
  174. if (pos < end &&
  175. !nilfs_set_bit_atomic(
  176. nilfs_mdt_bgl_lock(inode, group), pos, bitmap))
  177. return pos;
  178. } else
  179. end = 0;
  180. for (i = 0, curr = end;
  181. i < bsize;
  182. i += BITS_PER_LONG, curr += BITS_PER_LONG) {
  183. /* wrap around */
  184. if (curr >= bsize)
  185. curr = 0;
  186. while (*((unsigned long *)bitmap + curr / BITS_PER_LONG)
  187. != ~0UL) {
  188. end = curr + BITS_PER_LONG;
  189. if (end > bsize)
  190. end = bsize;
  191. pos = nilfs_find_next_zero_bit(bitmap, end, curr);
  192. if ((pos < end) &&
  193. !nilfs_set_bit_atomic(
  194. nilfs_mdt_bgl_lock(inode, group), pos,
  195. bitmap))
  196. return pos;
  197. }
  198. }
  199. return -ENOSPC;
  200. }
  201. static unsigned long
  202. nilfs_palloc_rest_groups_in_desc_block(const struct inode *inode,
  203. unsigned long curr, unsigned long max)
  204. {
  205. return min_t(unsigned long,
  206. nilfs_palloc_groups_per_desc_block(inode) -
  207. curr % nilfs_palloc_groups_per_desc_block(inode),
  208. max - curr + 1);
  209. }
  210. int nilfs_palloc_prepare_alloc_entry(struct inode *inode,
  211. struct nilfs_palloc_req *req)
  212. {
  213. struct buffer_head *desc_bh, *bitmap_bh;
  214. struct nilfs_palloc_group_desc *desc;
  215. unsigned char *bitmap;
  216. void *desc_kaddr, *bitmap_kaddr;
  217. unsigned long group, maxgroup, ngroups;
  218. unsigned long group_offset, maxgroup_offset;
  219. unsigned long n, entries_per_group, groups_per_desc_block;
  220. unsigned long i, j;
  221. int pos, ret;
  222. ngroups = nilfs_palloc_groups_count(inode);
  223. maxgroup = ngroups - 1;
  224. group = nilfs_palloc_group(inode, req->pr_entry_nr, &group_offset);
  225. entries_per_group = nilfs_palloc_entries_per_group(inode);
  226. groups_per_desc_block = nilfs_palloc_groups_per_desc_block(inode);
  227. for (i = 0; i < ngroups; i += n) {
  228. if (group >= ngroups) {
  229. /* wrap around */
  230. group = 0;
  231. maxgroup = nilfs_palloc_group(inode, req->pr_entry_nr,
  232. &maxgroup_offset) - 1;
  233. }
  234. ret = nilfs_palloc_get_desc_block(inode, group, 1, &desc_bh);
  235. if (ret < 0)
  236. return ret;
  237. desc_kaddr = kmap(desc_bh->b_page);
  238. desc = nilfs_palloc_block_get_group_desc(
  239. inode, group, desc_bh, desc_kaddr);
  240. n = nilfs_palloc_rest_groups_in_desc_block(inode, group,
  241. maxgroup);
  242. for (j = 0; j < n; j++, desc++, group++) {
  243. if (nilfs_palloc_group_desc_nfrees(inode, group, desc)
  244. > 0) {
  245. ret = nilfs_palloc_get_bitmap_block(
  246. inode, group, 1, &bitmap_bh);
  247. if (ret < 0)
  248. goto out_desc;
  249. bitmap_kaddr = kmap(bitmap_bh->b_page);
  250. bitmap = bitmap_kaddr + bh_offset(bitmap_bh);
  251. pos = nilfs_palloc_find_available_slot(
  252. inode, group, group_offset, bitmap,
  253. entries_per_group);
  254. if (pos >= 0) {
  255. /* found a free entry */
  256. nilfs_palloc_group_desc_add_entries(
  257. inode, group, desc, -1);
  258. req->pr_entry_nr =
  259. entries_per_group * group + pos;
  260. kunmap(desc_bh->b_page);
  261. kunmap(bitmap_bh->b_page);
  262. req->pr_desc_bh = desc_bh;
  263. req->pr_bitmap_bh = bitmap_bh;
  264. return 0;
  265. }
  266. kunmap(bitmap_bh->b_page);
  267. brelse(bitmap_bh);
  268. }
  269. group_offset = 0;
  270. }
  271. kunmap(desc_bh->b_page);
  272. brelse(desc_bh);
  273. }
  274. /* no entries left */
  275. return -ENOSPC;
  276. out_desc:
  277. kunmap(desc_bh->b_page);
  278. brelse(desc_bh);
  279. return ret;
  280. }
  281. void nilfs_palloc_commit_alloc_entry(struct inode *inode,
  282. struct nilfs_palloc_req *req)
  283. {
  284. nilfs_mdt_mark_buffer_dirty(req->pr_bitmap_bh);
  285. nilfs_mdt_mark_buffer_dirty(req->pr_desc_bh);
  286. nilfs_mdt_mark_dirty(inode);
  287. brelse(req->pr_bitmap_bh);
  288. brelse(req->pr_desc_bh);
  289. }
  290. void nilfs_palloc_commit_free_entry(struct inode *inode,
  291. struct nilfs_palloc_req *req)
  292. {
  293. struct nilfs_palloc_group_desc *desc;
  294. unsigned long group, group_offset;
  295. unsigned char *bitmap;
  296. void *desc_kaddr, *bitmap_kaddr;
  297. group = nilfs_palloc_group(inode, req->pr_entry_nr, &group_offset);
  298. desc_kaddr = kmap(req->pr_desc_bh->b_page);
  299. desc = nilfs_palloc_block_get_group_desc(inode, group,
  300. req->pr_desc_bh, desc_kaddr);
  301. bitmap_kaddr = kmap(req->pr_bitmap_bh->b_page);
  302. bitmap = bitmap_kaddr + bh_offset(req->pr_bitmap_bh);
  303. if (!nilfs_clear_bit_atomic(nilfs_mdt_bgl_lock(inode, group),
  304. group_offset, bitmap))
  305. printk(KERN_WARNING "%s: entry number %llu already freed\n",
  306. __func__, (unsigned long long)req->pr_entry_nr);
  307. nilfs_palloc_group_desc_add_entries(inode, group, desc, 1);
  308. kunmap(req->pr_bitmap_bh->b_page);
  309. kunmap(req->pr_desc_bh->b_page);
  310. nilfs_mdt_mark_buffer_dirty(req->pr_desc_bh);
  311. nilfs_mdt_mark_buffer_dirty(req->pr_bitmap_bh);
  312. nilfs_mdt_mark_dirty(inode);
  313. brelse(req->pr_bitmap_bh);
  314. brelse(req->pr_desc_bh);
  315. }
  316. void nilfs_palloc_abort_alloc_entry(struct inode *inode,
  317. struct nilfs_palloc_req *req)
  318. {
  319. struct nilfs_palloc_group_desc *desc;
  320. void *desc_kaddr, *bitmap_kaddr;
  321. unsigned char *bitmap;
  322. unsigned long group, group_offset;
  323. group = nilfs_palloc_group(inode, req->pr_entry_nr, &group_offset);
  324. desc_kaddr = kmap(req->pr_desc_bh->b_page);
  325. desc = nilfs_palloc_block_get_group_desc(inode, group,
  326. req->pr_desc_bh, desc_kaddr);
  327. bitmap_kaddr = kmap(req->pr_bitmap_bh->b_page);
  328. bitmap = bitmap_kaddr + bh_offset(req->pr_bitmap_bh);
  329. if (!nilfs_clear_bit_atomic(nilfs_mdt_bgl_lock(inode, group),
  330. group_offset, bitmap))
  331. printk(KERN_WARNING "%s: entry numer %llu already freed\n",
  332. __func__, (unsigned long long)req->pr_entry_nr);
  333. nilfs_palloc_group_desc_add_entries(inode, group, desc, 1);
  334. kunmap(req->pr_bitmap_bh->b_page);
  335. kunmap(req->pr_desc_bh->b_page);
  336. brelse(req->pr_bitmap_bh);
  337. brelse(req->pr_desc_bh);
  338. req->pr_entry_nr = 0;
  339. req->pr_bitmap_bh = NULL;
  340. req->pr_desc_bh = NULL;
  341. }
  342. int nilfs_palloc_prepare_free_entry(struct inode *inode,
  343. struct nilfs_palloc_req *req)
  344. {
  345. struct buffer_head *desc_bh, *bitmap_bh;
  346. unsigned long group, group_offset;
  347. int ret;
  348. group = nilfs_palloc_group(inode, req->pr_entry_nr, &group_offset);
  349. ret = nilfs_palloc_get_desc_block(inode, group, 1, &desc_bh);
  350. if (ret < 0)
  351. return ret;
  352. ret = nilfs_palloc_get_bitmap_block(inode, group, 1, &bitmap_bh);
  353. if (ret < 0) {
  354. brelse(desc_bh);
  355. return ret;
  356. }
  357. req->pr_desc_bh = desc_bh;
  358. req->pr_bitmap_bh = bitmap_bh;
  359. return 0;
  360. }
  361. void nilfs_palloc_abort_free_entry(struct inode *inode,
  362. struct nilfs_palloc_req *req)
  363. {
  364. brelse(req->pr_bitmap_bh);
  365. brelse(req->pr_desc_bh);
  366. req->pr_entry_nr = 0;
  367. req->pr_bitmap_bh = NULL;
  368. req->pr_desc_bh = NULL;
  369. }
  370. static int
  371. nilfs_palloc_group_is_in(struct inode *inode, unsigned long group, __u64 nr)
  372. {
  373. __u64 first, last;
  374. first = group * nilfs_palloc_entries_per_group(inode);
  375. last = first + nilfs_palloc_entries_per_group(inode) - 1;
  376. return (nr >= first) && (nr <= last);
  377. }
  378. int nilfs_palloc_freev(struct inode *inode, __u64 *entry_nrs, size_t nitems)
  379. {
  380. struct buffer_head *desc_bh, *bitmap_bh;
  381. struct nilfs_palloc_group_desc *desc;
  382. unsigned char *bitmap;
  383. void *desc_kaddr, *bitmap_kaddr;
  384. unsigned long group, group_offset;
  385. int i, j, n, ret;
  386. for (i = 0; i < nitems; i += n) {
  387. group = nilfs_palloc_group(inode, entry_nrs[i], &group_offset);
  388. ret = nilfs_palloc_get_desc_block(inode, group, 0, &desc_bh);
  389. if (ret < 0)
  390. return ret;
  391. ret = nilfs_palloc_get_bitmap_block(inode, group, 0,
  392. &bitmap_bh);
  393. if (ret < 0) {
  394. brelse(desc_bh);
  395. return ret;
  396. }
  397. desc_kaddr = kmap(desc_bh->b_page);
  398. desc = nilfs_palloc_block_get_group_desc(
  399. inode, group, desc_bh, desc_kaddr);
  400. bitmap_kaddr = kmap(bitmap_bh->b_page);
  401. bitmap = bitmap_kaddr + bh_offset(bitmap_bh);
  402. for (j = i, n = 0;
  403. (j < nitems) && nilfs_palloc_group_is_in(inode, group,
  404. entry_nrs[j]);
  405. j++, n++) {
  406. nilfs_palloc_group(inode, entry_nrs[j], &group_offset);
  407. if (!nilfs_clear_bit_atomic(
  408. nilfs_mdt_bgl_lock(inode, group),
  409. group_offset, bitmap)) {
  410. printk(KERN_WARNING
  411. "%s: entry number %llu already freed\n",
  412. __func__,
  413. (unsigned long long)entry_nrs[j]);
  414. }
  415. }
  416. nilfs_palloc_group_desc_add_entries(inode, group, desc, n);
  417. kunmap(bitmap_bh->b_page);
  418. kunmap(desc_bh->b_page);
  419. nilfs_mdt_mark_buffer_dirty(desc_bh);
  420. nilfs_mdt_mark_buffer_dirty(bitmap_bh);
  421. nilfs_mdt_mark_dirty(inode);
  422. brelse(bitmap_bh);
  423. brelse(desc_bh);
  424. }
  425. return 0;
  426. }
  427. void nilfs_palloc_setup_cache(struct inode *inode,
  428. struct nilfs_palloc_cache *cache)
  429. {
  430. NILFS_MDT(inode)->mi_palloc_cache = cache;
  431. spin_lock_init(&cache->lock);
  432. }
  433. void nilfs_palloc_clear_cache(struct inode *inode)
  434. {
  435. struct nilfs_palloc_cache *cache = NILFS_MDT(inode)->mi_palloc_cache;
  436. spin_lock(&cache->lock);
  437. brelse(cache->prev_desc.bh);
  438. brelse(cache->prev_bitmap.bh);
  439. brelse(cache->prev_entry.bh);
  440. cache->prev_desc.bh = NULL;
  441. cache->prev_bitmap.bh = NULL;
  442. cache->prev_entry.bh = NULL;
  443. spin_unlock(&cache->lock);
  444. }
  445. void nilfs_palloc_destroy_cache(struct inode *inode)
  446. {
  447. nilfs_palloc_clear_cache(inode);
  448. NILFS_MDT(inode)->mi_palloc_cache = NULL;
  449. }