extent-tree.c 12 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514
  1. #include <stdio.h>
  2. #include <stdlib.h>
  3. #include "kerncompat.h"
  4. #include "radix-tree.h"
  5. #include "ctree.h"
  6. #include "disk-io.h"
  7. #include "print-tree.h"
  8. static int find_free_extent(struct ctree_root *orig_root, u64 num_blocks,
  9. u64 search_start, u64 search_end, struct key *ins);
  10. static int finish_current_insert(struct ctree_root *extent_root);
  11. static int run_pending(struct ctree_root *extent_root);
  12. /*
  13. * pending extents are blocks that we're trying to allocate in the extent
  14. * map while trying to grow the map because of other allocations. To avoid
  15. * recursing, they are tagged in the radix tree and cleaned up after
  16. * other allocations are done. The pending tag is also used in the same
  17. * manner for deletes.
  18. */
  19. #define CTREE_EXTENT_PENDING_DEL 0
  20. static int inc_block_ref(struct ctree_root *root, u64 blocknr)
  21. {
  22. struct ctree_path path;
  23. int ret;
  24. struct key key;
  25. struct leaf *l;
  26. struct extent_item *item;
  27. struct key ins;
  28. find_free_extent(root->extent_root, 0, 0, (u64)-1, &ins);
  29. init_path(&path);
  30. key.objectid = blocknr;
  31. key.flags = 0;
  32. key.offset = 1;
  33. ret = search_slot(root->extent_root, &key, &path, 0, 1);
  34. if (ret != 0)
  35. BUG();
  36. BUG_ON(ret != 0);
  37. l = &path.nodes[0]->leaf;
  38. item = (struct extent_item *)(l->data +
  39. l->items[path.slots[0]].offset);
  40. item->refs++;
  41. BUG_ON(list_empty(&path.nodes[0]->dirty));
  42. release_path(root->extent_root, &path);
  43. finish_current_insert(root->extent_root);
  44. run_pending(root->extent_root);
  45. return 0;
  46. }
  47. static int lookup_block_ref(struct ctree_root *root, u64 blocknr, u32 *refs)
  48. {
  49. struct ctree_path path;
  50. int ret;
  51. struct key key;
  52. struct leaf *l;
  53. struct extent_item *item;
  54. init_path(&path);
  55. key.objectid = blocknr;
  56. key.flags = 0;
  57. key.offset = 1;
  58. ret = search_slot(root->extent_root, &key, &path, 0, 0);
  59. if (ret != 0)
  60. BUG();
  61. l = &path.nodes[0]->leaf;
  62. item = (struct extent_item *)(l->data +
  63. l->items[path.slots[0]].offset);
  64. *refs = item->refs;
  65. release_path(root->extent_root, &path);
  66. return 0;
  67. }
  68. int btrfs_inc_ref(struct ctree_root *root, struct tree_buffer *buf)
  69. {
  70. u64 blocknr;
  71. int i;
  72. if (root == root->extent_root)
  73. return 0;
  74. if (btrfs_is_leaf(&buf->node))
  75. return 0;
  76. for (i = 0; i < btrfs_header_nritems(&buf->node.header); i++) {
  77. blocknr = buf->node.blockptrs[i];
  78. inc_block_ref(root, blocknr);
  79. }
  80. return 0;
  81. }
  82. int btrfs_finish_extent_commit(struct ctree_root *root)
  83. {
  84. struct ctree_root *extent_root = root->extent_root;
  85. unsigned long gang[8];
  86. int ret;
  87. int i;
  88. while(1) {
  89. ret = radix_tree_gang_lookup(&extent_root->pinned_radix,
  90. (void **)gang, 0,
  91. ARRAY_SIZE(gang));
  92. if (!ret)
  93. break;
  94. for (i = 0; i < ret; i++) {
  95. radix_tree_delete(&extent_root->pinned_radix, gang[i]);
  96. }
  97. }
  98. extent_root->last_insert.objectid = 0;
  99. extent_root->last_insert.offset = 0;
  100. return 0;
  101. }
  102. static int finish_current_insert(struct ctree_root *extent_root)
  103. {
  104. struct key ins;
  105. struct extent_item extent_item;
  106. int i;
  107. int ret;
  108. extent_item.refs = 1;
  109. extent_item.owner =
  110. btrfs_header_parentid(&extent_root->node->node.header);
  111. ins.offset = 1;
  112. ins.flags = 0;
  113. for (i = 0; i < extent_root->current_insert.flags; i++) {
  114. ins.objectid = extent_root->current_insert.objectid + i;
  115. ret = insert_item(extent_root, &ins, &extent_item,
  116. sizeof(extent_item));
  117. BUG_ON(ret);
  118. }
  119. extent_root->current_insert.offset = 0;
  120. return 0;
  121. }
  122. /*
  123. * remove an extent from the root, returns 0 on success
  124. */
  125. int __free_extent(struct ctree_root *root, u64 blocknr, u64 num_blocks)
  126. {
  127. struct ctree_path path;
  128. struct key key;
  129. struct ctree_root *extent_root = root->extent_root;
  130. int ret;
  131. struct item *item;
  132. struct extent_item *ei;
  133. struct key ins;
  134. key.objectid = blocknr;
  135. key.flags = 0;
  136. key.offset = num_blocks;
  137. find_free_extent(root, 0, 0, (u64)-1, &ins);
  138. init_path(&path);
  139. ret = search_slot(extent_root, &key, &path, -1, 1);
  140. if (ret) {
  141. printf("failed to find %Lu\n", key.objectid);
  142. print_tree(extent_root, extent_root->node);
  143. printf("failed to find %Lu\n", key.objectid);
  144. BUG();
  145. }
  146. item = path.nodes[0]->leaf.items + path.slots[0];
  147. ei = (struct extent_item *)(path.nodes[0]->leaf.data + item->offset);
  148. BUG_ON(ei->refs == 0);
  149. ei->refs--;
  150. if (ei->refs == 0) {
  151. if (root == extent_root) {
  152. int err;
  153. radix_tree_preload(GFP_KERNEL);
  154. err = radix_tree_insert(&extent_root->pinned_radix,
  155. blocknr, (void *)blocknr);
  156. BUG_ON(err);
  157. radix_tree_preload_end();
  158. }
  159. ret = del_item(extent_root, &path);
  160. if (root != extent_root &&
  161. extent_root->last_insert.objectid < blocknr)
  162. extent_root->last_insert.objectid = blocknr;
  163. if (ret)
  164. BUG();
  165. }
  166. release_path(extent_root, &path);
  167. finish_current_insert(extent_root);
  168. return ret;
  169. }
  170. /*
  171. * find all the blocks marked as pending in the radix tree and remove
  172. * them from the extent map
  173. */
  174. static int del_pending_extents(struct ctree_root *extent_root)
  175. {
  176. int ret;
  177. struct tree_buffer *gang[4];
  178. int i;
  179. while(1) {
  180. ret = radix_tree_gang_lookup_tag(&extent_root->cache_radix,
  181. (void **)gang, 0,
  182. ARRAY_SIZE(gang),
  183. CTREE_EXTENT_PENDING_DEL);
  184. if (!ret)
  185. break;
  186. for (i = 0; i < ret; i++) {
  187. ret = __free_extent(extent_root, gang[i]->blocknr, 1);
  188. radix_tree_tag_clear(&extent_root->cache_radix,
  189. gang[i]->blocknr,
  190. CTREE_EXTENT_PENDING_DEL);
  191. tree_block_release(extent_root, gang[i]);
  192. }
  193. }
  194. return 0;
  195. }
  196. static int run_pending(struct ctree_root *extent_root)
  197. {
  198. while(radix_tree_tagged(&extent_root->cache_radix,
  199. CTREE_EXTENT_PENDING_DEL))
  200. del_pending_extents(extent_root);
  201. return 0;
  202. }
  203. /*
  204. * remove an extent from the root, returns 0 on success
  205. */
  206. int free_extent(struct ctree_root *root, u64 blocknr, u64 num_blocks)
  207. {
  208. struct key key;
  209. struct ctree_root *extent_root = root->extent_root;
  210. struct tree_buffer *t;
  211. int pending_ret;
  212. int ret;
  213. if (root == extent_root) {
  214. t = find_tree_block(root, blocknr);
  215. radix_tree_tag_set(&root->cache_radix, blocknr,
  216. CTREE_EXTENT_PENDING_DEL);
  217. return 0;
  218. }
  219. key.objectid = blocknr;
  220. key.flags = 0;
  221. key.offset = num_blocks;
  222. ret = __free_extent(root, blocknr, num_blocks);
  223. pending_ret = run_pending(root->extent_root);
  224. return ret ? ret : pending_ret;
  225. }
  226. /*
  227. * walks the btree of allocated extents and find a hole of a given size.
  228. * The key ins is changed to record the hole:
  229. * ins->objectid == block start
  230. * ins->flags = 0
  231. * ins->offset == number of blocks
  232. * Any available blocks before search_start are skipped.
  233. */
  234. static int find_free_extent(struct ctree_root *orig_root, u64 num_blocks,
  235. u64 search_start, u64 search_end, struct key *ins)
  236. {
  237. struct ctree_path path;
  238. struct key *key;
  239. int ret;
  240. u64 hole_size = 0;
  241. int slot = 0;
  242. u64 last_block;
  243. u64 test_block;
  244. int start_found;
  245. struct leaf *l;
  246. struct ctree_root * root = orig_root->extent_root;
  247. int total_needed = num_blocks;
  248. total_needed += (btrfs_header_level(&root->node->node.header) + 1) * 3;
  249. if (root->last_insert.objectid > search_start)
  250. search_start = root->last_insert.objectid;
  251. check_failed:
  252. init_path(&path);
  253. ins->objectid = search_start;
  254. ins->offset = 0;
  255. ins->flags = 0;
  256. start_found = 0;
  257. ret = search_slot(root, ins, &path, 0, 0);
  258. if (ret < 0)
  259. goto error;
  260. if (path.slots[0] > 0)
  261. path.slots[0]--;
  262. while (1) {
  263. l = &path.nodes[0]->leaf;
  264. slot = path.slots[0];
  265. if (slot >= btrfs_header_nritems(&l->header)) {
  266. ret = next_leaf(root, &path);
  267. if (ret == 0)
  268. continue;
  269. if (ret < 0)
  270. goto error;
  271. if (!start_found) {
  272. ins->objectid = search_start;
  273. ins->offset = (u64)-1;
  274. start_found = 1;
  275. goto check_pending;
  276. }
  277. ins->objectid = last_block > search_start ?
  278. last_block : search_start;
  279. ins->offset = (u64)-1;
  280. goto check_pending;
  281. }
  282. key = &l->items[slot].key;
  283. if (key->objectid >= search_start) {
  284. if (start_found) {
  285. if (last_block < search_start)
  286. last_block = search_start;
  287. hole_size = key->objectid - last_block;
  288. if (hole_size > total_needed) {
  289. ins->objectid = last_block;
  290. ins->offset = hole_size;
  291. goto check_pending;
  292. }
  293. }
  294. }
  295. start_found = 1;
  296. last_block = key->objectid + key->offset;
  297. path.slots[0]++;
  298. }
  299. // FIXME -ENOSPC
  300. check_pending:
  301. /* we have to make sure we didn't find an extent that has already
  302. * been allocated by the map tree or the original allocation
  303. */
  304. release_path(root, &path);
  305. BUG_ON(ins->objectid < search_start);
  306. for (test_block = ins->objectid;
  307. test_block < ins->objectid + total_needed; test_block++) {
  308. if (radix_tree_lookup(&root->pinned_radix, test_block)) {
  309. search_start = test_block + 1;
  310. goto check_failed;
  311. }
  312. }
  313. BUG_ON(root->current_insert.offset);
  314. root->current_insert.offset = total_needed - num_blocks;
  315. root->current_insert.objectid = ins->objectid + num_blocks;
  316. root->current_insert.flags = 0;
  317. root->last_insert.objectid = ins->objectid;
  318. ins->offset = num_blocks;
  319. return 0;
  320. error:
  321. release_path(root, &path);
  322. return ret;
  323. }
  324. /*
  325. * finds a free extent and does all the dirty work required for allocation
  326. * returns the key for the extent through ins, and a tree buffer for
  327. * the first block of the extent through buf.
  328. *
  329. * returns 0 if everything worked, non-zero otherwise.
  330. */
  331. int alloc_extent(struct ctree_root *root, u64 num_blocks, u64 search_start,
  332. u64 search_end, u64 owner, struct key *ins)
  333. {
  334. int ret;
  335. int pending_ret;
  336. struct ctree_root *extent_root = root->extent_root;
  337. struct extent_item extent_item;
  338. extent_item.refs = 1;
  339. extent_item.owner = owner;
  340. if (root == extent_root) {
  341. BUG_ON(extent_root->current_insert.offset == 0);
  342. BUG_ON(num_blocks != 1);
  343. BUG_ON(extent_root->current_insert.flags ==
  344. extent_root->current_insert.offset);
  345. ins->offset = 1;
  346. ins->objectid = extent_root->current_insert.objectid +
  347. extent_root->current_insert.flags++;
  348. return 0;
  349. }
  350. ret = find_free_extent(root, num_blocks, search_start,
  351. search_end, ins);
  352. if (ret)
  353. return ret;
  354. ret = insert_item(extent_root, ins, &extent_item,
  355. sizeof(extent_item));
  356. finish_current_insert(extent_root);
  357. pending_ret = run_pending(extent_root);
  358. if (ret)
  359. return ret;
  360. if (pending_ret)
  361. return pending_ret;
  362. return 0;
  363. }
  364. /*
  365. * helper function to allocate a block for a given tree
  366. * returns the tree buffer or NULL.
  367. */
  368. struct tree_buffer *alloc_free_block(struct ctree_root *root)
  369. {
  370. struct key ins;
  371. int ret;
  372. struct tree_buffer *buf;
  373. ret = alloc_extent(root, 1, 0, (unsigned long)-1,
  374. btrfs_header_parentid(&root->node->node.header),
  375. &ins);
  376. if (ret) {
  377. BUG();
  378. return NULL;
  379. }
  380. buf = find_tree_block(root, ins.objectid);
  381. dirty_tree_block(root, buf);
  382. return buf;
  383. }
  384. int walk_down_tree(struct ctree_root *root, struct ctree_path *path, int *level)
  385. {
  386. struct tree_buffer *next;
  387. struct tree_buffer *cur;
  388. u64 blocknr;
  389. int ret;
  390. u32 refs;
  391. ret = lookup_block_ref(root, path->nodes[*level]->blocknr, &refs);
  392. BUG_ON(ret);
  393. if (refs > 1)
  394. goto out;
  395. while(*level > 0) {
  396. cur = path->nodes[*level];
  397. if (path->slots[*level] >=
  398. btrfs_header_nritems(&cur->node.header))
  399. break;
  400. blocknr = cur->node.blockptrs[path->slots[*level]];
  401. ret = lookup_block_ref(root, blocknr, &refs);
  402. if (refs != 1 || *level == 1) {
  403. path->slots[*level]++;
  404. ret = free_extent(root, blocknr, 1);
  405. BUG_ON(ret);
  406. continue;
  407. }
  408. BUG_ON(ret);
  409. next = read_tree_block(root, blocknr);
  410. if (path->nodes[*level-1])
  411. tree_block_release(root, path->nodes[*level-1]);
  412. path->nodes[*level-1] = next;
  413. *level = btrfs_header_level(&next->node.header);
  414. path->slots[*level] = 0;
  415. }
  416. out:
  417. ret = free_extent(root, path->nodes[*level]->blocknr, 1);
  418. tree_block_release(root, path->nodes[*level]);
  419. path->nodes[*level] = NULL;
  420. *level += 1;
  421. BUG_ON(ret);
  422. return 0;
  423. }
  424. int walk_up_tree(struct ctree_root *root, struct ctree_path *path, int *level)
  425. {
  426. int i;
  427. int slot;
  428. int ret;
  429. for(i = *level; i < MAX_LEVEL - 1 && path->nodes[i]; i++) {
  430. slot = path->slots[i];
  431. if (slot <
  432. btrfs_header_nritems(&path->nodes[i]->node.header)- 1) {
  433. path->slots[i]++;
  434. *level = i;
  435. return 0;
  436. } else {
  437. ret = free_extent(root,
  438. path->nodes[*level]->blocknr, 1);
  439. tree_block_release(root, path->nodes[*level]);
  440. path->nodes[*level] = NULL;
  441. *level = i + 1;
  442. BUG_ON(ret);
  443. }
  444. }
  445. return 1;
  446. }
  447. int btrfs_drop_snapshot(struct ctree_root *root, struct tree_buffer *snap)
  448. {
  449. int ret;
  450. int level;
  451. struct ctree_path path;
  452. int i;
  453. int orig_level;
  454. init_path(&path);
  455. level = btrfs_header_level(&snap->node.header);
  456. orig_level = level;
  457. path.nodes[level] = snap;
  458. path.slots[level] = 0;
  459. while(1) {
  460. ret = walk_down_tree(root, &path, &level);
  461. if (ret > 0)
  462. break;
  463. ret = walk_up_tree(root, &path, &level);
  464. if (ret > 0)
  465. break;
  466. }
  467. for (i = 0; i <= orig_level; i++) {
  468. if (path.nodes[i]) {
  469. tree_block_release(root, path.nodes[i]);
  470. }
  471. }
  472. return 0;
  473. }