|
@@ -100,7 +100,7 @@ void hfs_bnode_copy(struct hfs_bnode *dst_node, int dst,
|
|
|
struct hfs_btree *tree;
|
|
|
struct page *src_page, *dst_page;
|
|
|
|
|
|
- dprint(DBG_BNODE_MOD, "copybytes: %u,%u,%u\n", dst, src, len);
|
|
|
+ hfs_dbg(BNODE_MOD, "copybytes: %u,%u,%u\n", dst, src, len);
|
|
|
if (!len)
|
|
|
return;
|
|
|
tree = src_node->tree;
|
|
@@ -120,7 +120,7 @@ void hfs_bnode_move(struct hfs_bnode *node, int dst, int src, int len)
|
|
|
struct page *page;
|
|
|
void *ptr;
|
|
|
|
|
|
- dprint(DBG_BNODE_MOD, "movebytes: %u,%u,%u\n", dst, src, len);
|
|
|
+ hfs_dbg(BNODE_MOD, "movebytes: %u,%u,%u\n", dst, src, len);
|
|
|
if (!len)
|
|
|
return;
|
|
|
src += node->page_offset;
|
|
@@ -138,16 +138,16 @@ void hfs_bnode_dump(struct hfs_bnode *node)
|
|
|
__be32 cnid;
|
|
|
int i, off, key_off;
|
|
|
|
|
|
- dprint(DBG_BNODE_MOD, "bnode: %d\n", node->this);
|
|
|
+ hfs_dbg(BNODE_MOD, "bnode: %d\n", node->this);
|
|
|
hfs_bnode_read(node, &desc, 0, sizeof(desc));
|
|
|
- dprint(DBG_BNODE_MOD, "%d, %d, %d, %d, %d\n",
|
|
|
+ hfs_dbg(BNODE_MOD, "%d, %d, %d, %d, %d\n",
|
|
|
be32_to_cpu(desc.next), be32_to_cpu(desc.prev),
|
|
|
desc.type, desc.height, be16_to_cpu(desc.num_recs));
|
|
|
|
|
|
off = node->tree->node_size - 2;
|
|
|
for (i = be16_to_cpu(desc.num_recs); i >= 0; off -= 2, i--) {
|
|
|
key_off = hfs_bnode_read_u16(node, off);
|
|
|
- dprint(DBG_BNODE_MOD, " %d", key_off);
|
|
|
+ hfs_dbg_cont(BNODE_MOD, " %d", key_off);
|
|
|
if (i && node->type == HFS_NODE_INDEX) {
|
|
|
int tmp;
|
|
|
|
|
@@ -155,17 +155,18 @@ void hfs_bnode_dump(struct hfs_bnode *node)
|
|
|
tmp = (hfs_bnode_read_u8(node, key_off) | 1) + 1;
|
|
|
else
|
|
|
tmp = node->tree->max_key_len + 1;
|
|
|
- dprint(DBG_BNODE_MOD, " (%d,%d", tmp, hfs_bnode_read_u8(node, key_off));
|
|
|
+ hfs_dbg_cont(BNODE_MOD, " (%d,%d",
|
|
|
+ tmp, hfs_bnode_read_u8(node, key_off));
|
|
|
hfs_bnode_read(node, &cnid, key_off + tmp, 4);
|
|
|
- dprint(DBG_BNODE_MOD, ",%d)", be32_to_cpu(cnid));
|
|
|
+ hfs_dbg_cont(BNODE_MOD, ",%d)", be32_to_cpu(cnid));
|
|
|
} else if (i && node->type == HFS_NODE_LEAF) {
|
|
|
int tmp;
|
|
|
|
|
|
tmp = hfs_bnode_read_u8(node, key_off);
|
|
|
- dprint(DBG_BNODE_MOD, " (%d)", tmp);
|
|
|
+ hfs_dbg_cont(BNODE_MOD, " (%d)", tmp);
|
|
|
}
|
|
|
}
|
|
|
- dprint(DBG_BNODE_MOD, "\n");
|
|
|
+ hfs_dbg_cont(BNODE_MOD, "\n");
|
|
|
}
|
|
|
|
|
|
void hfs_bnode_unlink(struct hfs_bnode *node)
|
|
@@ -257,8 +258,8 @@ static struct hfs_bnode *__hfs_bnode_create(struct hfs_btree *tree, u32 cnid)
|
|
|
node->this = cnid;
|
|
|
set_bit(HFS_BNODE_NEW, &node->flags);
|
|
|
atomic_set(&node->refcnt, 1);
|
|
|
- dprint(DBG_BNODE_REFS, "new_node(%d:%d): 1\n",
|
|
|
- node->tree->cnid, node->this);
|
|
|
+ hfs_dbg(BNODE_REFS, "new_node(%d:%d): 1\n",
|
|
|
+ node->tree->cnid, node->this);
|
|
|
init_waitqueue_head(&node->lock_wq);
|
|
|
spin_lock(&tree->hash_lock);
|
|
|
node2 = hfs_bnode_findhash(tree, cnid);
|
|
@@ -301,7 +302,7 @@ void hfs_bnode_unhash(struct hfs_bnode *node)
|
|
|
{
|
|
|
struct hfs_bnode **p;
|
|
|
|
|
|
- dprint(DBG_BNODE_REFS, "remove_node(%d:%d): %d\n",
|
|
|
+ hfs_dbg(BNODE_REFS, "remove_node(%d:%d): %d\n",
|
|
|
node->tree->cnid, node->this, atomic_read(&node->refcnt));
|
|
|
for (p = &node->tree->node_hash[hfs_bnode_hash(node->this)];
|
|
|
*p && *p != node; p = &(*p)->next_hash)
|
|
@@ -443,8 +444,9 @@ void hfs_bnode_get(struct hfs_bnode *node)
|
|
|
{
|
|
|
if (node) {
|
|
|
atomic_inc(&node->refcnt);
|
|
|
- dprint(DBG_BNODE_REFS, "get_node(%d:%d): %d\n",
|
|
|
- node->tree->cnid, node->this, atomic_read(&node->refcnt));
|
|
|
+ hfs_dbg(BNODE_REFS, "get_node(%d:%d): %d\n",
|
|
|
+ node->tree->cnid, node->this,
|
|
|
+ atomic_read(&node->refcnt));
|
|
|
}
|
|
|
}
|
|
|
|
|
@@ -455,8 +457,9 @@ void hfs_bnode_put(struct hfs_bnode *node)
|
|
|
struct hfs_btree *tree = node->tree;
|
|
|
int i;
|
|
|
|
|
|
- dprint(DBG_BNODE_REFS, "put_node(%d:%d): %d\n",
|
|
|
- node->tree->cnid, node->this, atomic_read(&node->refcnt));
|
|
|
+ hfs_dbg(BNODE_REFS, "put_node(%d:%d): %d\n",
|
|
|
+ node->tree->cnid, node->this,
|
|
|
+ atomic_read(&node->refcnt));
|
|
|
BUG_ON(!atomic_read(&node->refcnt));
|
|
|
if (!atomic_dec_and_lock(&node->refcnt, &tree->hash_lock))
|
|
|
return;
|