summaryrefslogtreecommitdiff
path: root/src
diff options
context:
space:
mode:
authorMichael Cahill <michael.cahill@wiredtiger.com>2013-03-19 16:04:50 +1100
committerMichael Cahill <michael.cahill@wiredtiger.com>2013-03-19 16:04:50 +1100
commitc0370484f59e2abb9a82927eb81804f84369cdaa (patch)
treedfb345211e0df606a074518180f5813437f4a643 /src
parent8554c711d565bc00b45dcd681868c8e6bdf7a5fe (diff)
downloadmongo-c0370484f59e2abb9a82927eb81804f84369cdaa.tar.gz
Remove refsize from internal pages, figure out the amount of memory being transferred during merge.
Diffstat (limited to 'src')
-rw-r--r--src/btree/rec_merge.c73
-rw-r--r--src/btree/rec_write.c5
-rw-r--r--src/btree/row_key.c2
-rw-r--r--src/include/btmem.h1
-rw-r--r--src/include/btree.i8
5 files changed, 55 insertions, 34 deletions
diff --git a/src/btree/rec_merge.c b/src/btree/rec_merge.c
index 4dd9a71a749..fd0098be19d 100644
--- a/src/btree/rec_merge.c
+++ b/src/btree/rec_merge.c
@@ -31,7 +31,7 @@ typedef struct {
*/
static int
__merge_walk(WT_SESSION_IMPL *session, WT_PAGE *page, u_int depth,
- void (*visit)(WT_REF *, WT_VISIT_STATE *),
+ void (*visit)(WT_PAGE *, WT_REF *, WT_VISIT_STATE *),
WT_VISIT_STATE *state)
{
WT_PAGE *child;
@@ -61,7 +61,7 @@ __merge_walk(WT_SESSION_IMPL *session, WT_PAGE *page, u_int depth,
case WT_REF_DELETED:
case WT_REF_DISK:
- (*visit)(ref, state);
+ (*visit)(page, ref, state);
break;
case WT_REF_EVICT_FORCE:
@@ -80,8 +80,10 @@ __merge_walk(WT_SESSION_IMPL *session, WT_PAGE *page, u_int depth,
* the first/last "live" reference.
*/
static void
-__merge_count(WT_REF *ref, WT_VISIT_STATE *state)
+__merge_count(WT_PAGE *parent, WT_REF *ref, WT_VISIT_STATE *state)
{
+ WT_UNUSED(parent);
+
if (ref->state == WT_REF_LOCKED) {
/* Prevent eviction until it is hooked into the new tree. */
__wt_evict_list_clr_page(state->session, ref->page);
@@ -105,10 +107,12 @@ __merge_count(WT_REF *ref, WT_VISIT_STATE *state)
* Copy a child reference from the locked subtree to a new page.
*/
static void
-__merge_copy_ref(WT_REF *ref, WT_VISIT_STATE *state)
+__merge_copy_ref(WT_PAGE *parent, WT_REF *ref, WT_VISIT_STATE *state)
{
WT_REF *newref;
+ WT_UNUSED(parent);
+
if (state->split != 0 && state->refcnt++ == state->split)
state->ref = state->second_ref;
@@ -136,11 +140,27 @@ __merge_unlock(WT_PAGE *page)
}
/*
+ * __merge_transfer_footprint --
+ * Transfer the size of references from an old page to a new page.
+ *
+ * Note that both pages are locked and there is no net change, so avoid
+ * __wt_cache_page_inmem_incr.
+ */
+static void
+__merge_transfer_footprint(
+ WT_SESSION_IMPL *session, WT_PAGE *newpage, WT_PAGE *oldpage, size_t size)
+{
+ WT_ASSERT(session, size < (size_t)oldpage->memory_footprint);
+ oldpage->memory_footprint -= size;
+ newpage->memory_footprint += size;
+}
+
+/*
* __merge_switch_page --
* Switch a page from the locked tree into the new tree.
*/
static void
-__merge_switch_page(WT_REF *ref, WT_VISIT_STATE *state)
+__merge_switch_page(WT_PAGE *parent, WT_REF *ref, WT_VISIT_STATE *state)
{
WT_PAGE *child;
WT_PAGE_MODIFY *modify;
@@ -153,6 +173,16 @@ __merge_switch_page(WT_REF *ref, WT_VISIT_STATE *state)
newref = state->ref++;
+ if (ref->addr != NULL)
+ __merge_transfer_footprint(
+ state->session, state->page, parent,
+ sizeof(WT_ADDR) + ((WT_ADDR *)ref->addr)->size);
+
+ if (ref->u.key != NULL)
+ __merge_transfer_footprint(
+ state->session, state->page, parent,
+ sizeof(WT_IKEY) + ((WT_IKEY *)ref->u.key)->size);
+
if (ref->state == WT_REF_LOCKED) {
child = ref->page;
@@ -185,12 +215,13 @@ __merge_switch_page(WT_REF *ref, WT_VISIT_STATE *state)
WT_CLEAR(*ref);
}
+#ifdef HAVE_DIAGNOSTIC
/*
- * __merge_walk_discard --
+ * __merge_check_discard --
* Make sure we are only discarding split-merge pages.
*/
static void
-__merge_walk_discard(WT_SESSION_IMPL *session, WT_PAGE *mainpage, WT_PAGE *page)
+__merge_check_discard(WT_SESSION_IMPL *session, WT_PAGE *page)
{
WT_REF *ref;
uint32_t i;
@@ -200,23 +231,16 @@ __merge_walk_discard(WT_SESSION_IMPL *session, WT_PAGE *mainpage, WT_PAGE *page)
WT_ASSERT(session, page->modify != NULL &&
F_ISSET(page->modify, WT_PM_REC_SPLIT_MERGE));
- /* Transfer the size of references from an old page to a new page. */
- WT_ASSERT(session,
- page->u.intl.refsize < (size_t)page->memory_footprint);
- mainpage->u.intl.refsize += page->u.intl.refsize;
- mainpage->memory_footprint += page->u.intl.refsize;
- page->memory_footprint -= page->u.intl.refsize;
- page->u.intl.refsize = 0;
-
WT_REF_FOREACH(page, ref, i) {
if (ref->state == WT_REF_DISK ||
ref->state == WT_REF_DELETED)
continue;
WT_ASSERT(session, ref->state == WT_REF_LOCKED);
- __merge_walk_discard(session, mainpage, ref->page);
+ __merge_check_discard(session, ref->page);
}
}
+#endif
/*
* __merge_new_page --
@@ -290,7 +314,7 @@ int
__wt_merge_tree(WT_SESSION_IMPL *session, WT_PAGE *top)
{
WT_DECL_RET;
- WT_PAGE *lchild, *mainpage, *newtop, *rchild;
+ WT_PAGE *lchild, *newtop, *rchild;
WT_REF *newref;
WT_VISIT_STATE visit_state;
uint32_t refcnt, split;
@@ -300,7 +324,7 @@ __wt_merge_tree(WT_SESSION_IMPL *session, WT_PAGE *top)
WT_CLEAR(visit_state);
visit_state.session = session;
- lchild = mainpage = newtop = rchild = NULL;
+ lchild = newtop = rchild = NULL;
page_type = top->type;
WT_ASSERT(session, __wt_btree_mergeable(top));
@@ -379,7 +403,7 @@ __wt_merge_tree(WT_SESSION_IMPL *session, WT_PAGE *top)
else {
WT_ERR(__merge_new_page(session, page_type, split,
visit_state.first_live < split, &lchild));
- visit_state.first = mainpage = lchild;
+ visit_state.first = lchild;
}
/* Right split. */
@@ -390,7 +414,7 @@ __wt_merge_tree(WT_SESSION_IMPL *session, WT_PAGE *top)
WT_ERR(__merge_new_page(session, page_type,
refcnt - split, visit_state.last_live >= split,
&rchild));
- visit_state.second = mainpage = rchild;
+ visit_state.second = rchild;
visit_state.second_ref =
&visit_state.second->u.intl.t[0];
}
@@ -408,7 +432,7 @@ __wt_merge_tree(WT_SESSION_IMPL *session, WT_PAGE *top)
__wt_btree_mergeable(top->parent),
&newtop));
- visit_state.first = mainpage = newtop;
+ visit_state.first = newtop;
}
/*
@@ -459,12 +483,13 @@ __wt_merge_tree(WT_SESSION_IMPL *session, WT_PAGE *top)
newtop->parent = top->parent;
newtop->ref = top->ref;
+#ifdef HAVE_DIAGNOSTIC
/*
* Before swapping in the new tree, walk the pages we are discarding,
- * check everything looks right and transfer the size to the main page.
+ * check that everything looks right.
*/
- WT_ASSERT(session, mainpage != NULL);
- __merge_walk_discard(session, mainpage, top);
+ __merge_check_discard(session, top);
+#endif
/*
* Set up the new top-level page as a split so that it will be swapped
diff --git a/src/btree/rec_write.c b/src/btree/rec_write.c
index 1ce078680e6..a7056365cad 100644
--- a/src/btree/rec_write.c
+++ b/src/btree/rec_write.c
@@ -4127,9 +4127,8 @@ __rec_split_row(
ref->addr = addr;
ref->state = WT_REF_DISK;
}
-
- page->u.intl.refsize = r->bnd_next * sizeof(WT_ADDR) + size;
- __wt_cache_page_inmem_incr(session, page, page->u.intl.refsize);
+ __wt_cache_page_inmem_incr(
+ session, page, r->bnd_next * sizeof(WT_ADDR) + size);
*splitp = page;
return (0);
diff --git a/src/btree/row_key.c b/src/btree/row_key.c
index 8bafc8f05bb..75dc2f1ca75 100644
--- a/src/btree/row_key.c
+++ b/src/btree/row_key.c
@@ -391,8 +391,6 @@ __wt_row_ikey_incr(WT_SESSION_IMPL *session, WT_PAGE *page,
WT_RET(__wt_row_ikey(session, cell_offset, key, size, ikeyp));
__wt_cache_page_inmem_incr(session, page, sizeof(WT_IKEY) + size);
- WT_ASSERT(session, page->type == WT_PAGE_ROW_INT);
- page->u.intl.refsize += sizeof(WT_IKEY) + size;
return (0);
}
diff --git a/src/include/btmem.h b/src/include/btmem.h
index 88276af12ca..d6b2deb6255 100644
--- a/src/include/btmem.h
+++ b/src/include/btmem.h
@@ -233,7 +233,6 @@ struct __wt_page {
struct {
uint64_t recno; /* Starting recno */
WT_REF *t; /* Subtree */
- size_t refsize; /* Size of child addresses */
} intl;
/* Row-store leaf page. */
diff --git a/src/include/btree.i b/src/include/btree.i
index 8826a88ba0d..5ff4dcd2271 100644
--- a/src/include/btree.i
+++ b/src/include/btree.i
@@ -41,10 +41,10 @@ __wt_eviction_page_force(WT_SESSION_IMPL *session, WT_PAGE *page)
}
/*
- * Estimate the per-allocation overhead. All allocators have some kind of
- * header and pad for alignment. We can't know for sure what that adds up
- * to, but this is an estimate based on some measurements of heap size versus
- * bytes in use.
+ * Estimate the per-allocation overhead. All implementations of malloc / free
+ * have some kind of header and pad for alignment. We can't know for sure what
+ * that adds up to, but this is an estimate based on some measurements of heap
+ * size versus bytes in use.
*/
#define WT_ALLOC_OVERHEAD 32