summaryrefslogtreecommitdiff
path: root/innobase
diff options
context:
space:
mode:
Diffstat (limited to 'innobase')
-rw-r--r--innobase/btr/btr0sea.c74
-rw-r--r--innobase/buf/buf0buf.c1
-rw-r--r--innobase/buf/buf0lru.c1
-rw-r--r--innobase/ha/ha0ha.c125
-rw-r--r--innobase/include/buf0buf.h8
-rw-r--r--innobase/include/ha0ha.h18
-rw-r--r--innobase/include/hash0hash.h18
7 files changed, 114 insertions, 131 deletions
diff --git a/innobase/btr/btr0sea.c b/innobase/btr/btr0sea.c
index 468c5efd24d..f2f5d09fa0e 100644
--- a/innobase/btr/btr0sea.c
+++ b/innobase/btr/btr0sea.c
@@ -439,8 +439,8 @@ btr_search_update_hash_ref(
ha_insert_for_fold(btr_search_sys->hash_index, fold, rec);
}
-}
-
+}
+
/*************************************************************************
Updates the search info. */
@@ -926,6 +926,17 @@ btr_search_drop_page_hash_index(
{
hash_table_t* table;
buf_block_t* block;
+ ulint n_fields;
+ ulint n_bytes;
+ rec_t* rec;
+ rec_t* sup;
+ ulint fold;
+ ulint prev_fold;
+ dulint tree_id;
+ ulint n_cached;
+ ulint n_recs;
+ ulint* folds;
+ ulint i;
#ifdef UNIV_SYNC_DEBUG
ut_ad(!rw_lock_own(&btr_search_latch, RW_LOCK_SHARED));
@@ -951,18 +962,73 @@ btr_search_drop_page_hash_index(
|| (block->buf_fix_count == 0));
#endif /* UNIV_SYNC_DEBUG */
- ut_a(block->curr_n_fields + block->curr_n_bytes > 0);
+ n_fields = block->curr_n_fields;
+ n_bytes = block->curr_n_bytes;
+
+ ut_a(n_fields + n_bytes > 0);
rw_lock_s_unlock(&btr_search_latch);
+ n_recs = page_get_n_recs(page);
+
+ /* Calculate and cache fold values into an array for fast deletion
+ from the hash index */
+
+ folds = mem_alloc(n_recs * sizeof(ulint));
+
+ n_cached = 0;
+
+ sup = page_get_supremum_rec(page);
+
+ rec = page_get_infimum_rec(page);
+ rec = page_rec_get_next(rec);
+
+ if (rec != sup) {
+ ut_a(n_fields <= rec_get_n_fields(rec));
+
+ if (n_bytes > 0) {
+ ut_a(n_fields < rec_get_n_fields(rec));
+ }
+ }
+
+ tree_id = btr_page_get_index_id(page);
+
+ prev_fold = 0;
+
+ while (rec != sup) {
+ /* FIXME: in a mixed tree, not all records may have enough
+ ordering fields: */
+
+ fold = rec_fold(rec, n_fields, n_bytes, tree_id);
+
+ if (fold == prev_fold && prev_fold != 0) {
+
+ goto next_rec;
+ }
+
+ /* Remove all hash nodes pointing to this page from the
+ hash chain */
+
+ folds[n_cached] = fold;
+ n_cached++;
+next_rec:
+ rec = page_rec_get_next(rec);
+ prev_fold = fold;
+ }
+
rw_lock_x_lock(&btr_search_latch);
- ha_remove_all_nodes_to_page(table, page);
+ for (i = 0; i < n_cached; i++) {
+
+ ha_remove_all_nodes_to_page(table, folds[i], page);
+ }
block->is_hashed = FALSE;
block->index = NULL;
rw_lock_x_unlock(&btr_search_latch);
+
+ mem_free(folds);
}
/************************************************************************
diff --git a/innobase/buf/buf0buf.c b/innobase/buf/buf0buf.c
index 8ca353aa5fb..3930ea93889 100644
--- a/innobase/buf/buf0buf.c
+++ b/innobase/buf/buf0buf.c
@@ -466,7 +466,6 @@ buf_block_init(
block->in_LRU_list = FALSE;
block->n_pointers = 0;
- block->hash_nodes = NULL;
rw_lock_create(&(block->lock));
ut_ad(rw_lock_validate(&(block->lock)));
diff --git a/innobase/buf/buf0lru.c b/innobase/buf/buf0lru.c
index 985426a9e2b..8460a049d3e 100644
--- a/innobase/buf/buf0lru.c
+++ b/innobase/buf/buf0lru.c
@@ -823,7 +823,6 @@ buf_LRU_block_free_non_file_page(
|| (block->state == BUF_BLOCK_READY_FOR_USE));
ut_a(block->n_pointers == 0);
- ut_a(block->hash_nodes == NULL);
ut_a(!block->in_free_list);
block->state = BUF_BLOCK_NOT_USED;
diff --git a/innobase/ha/ha0ha.c b/innobase/ha/ha0ha.c
index 49b59882626..ad1391ff83e 100644
--- a/innobase/ha/ha0ha.c
+++ b/innobase/ha/ha0ha.c
@@ -66,52 +66,9 @@ ha_create(
}
/*****************************************************************
-Removes an adaptive hash index node from the doubly linked list of hash nodes
-for the buffer block. */
-UNIV_INLINE
-void
-ha_remove_buf_block_node(
-/*=====================*/
- buf_block_t* block, /* in: buffer block */
- ha_node_t* node) /* in: an adaptive hash index node */
-{
- if (node == block->hash_nodes) {
- block->hash_nodes = node->next_for_block;
- }
-
- if (node->prev_for_block != NULL) {
- (node->prev_for_block)->next_for_block = node->next_for_block;
- }
-
- if (node->next_for_block != NULL) {
- (node->next_for_block)->prev_for_block = node->prev_for_block;
- }
-}
-
-/*****************************************************************
-Adds an adaptive hash index node to the start of the doubly linked list of
-hash nodes for the buffer block. */
-UNIV_INLINE
-void
-ha_add_buf_block_node(
-/*==================*/
- buf_block_t* block, /* in: buffer block */
- ha_node_t* node) /* in: an adaptive hash index node */
-{
- node->next_for_block = block->hash_nodes;
- node->prev_for_block = NULL;
-
- block->hash_nodes = node;
-
- if (node->next_for_block != NULL) {
- (node->next_for_block)->prev_for_block = node;
- }
-}
-
-/*****************************************************************
Inserts an entry into a hash table. If an entry with the same fold number
is found, its node is updated to point to the new data, and no new node
-is inserted. This function is only used in the adaptive hash index. */
+is inserted. */
ibool
ha_insert_for_fold(
@@ -127,7 +84,6 @@ ha_insert_for_fold(
{
hash_cell_t* cell;
ha_node_t* node;
- buf_block_t* block;
ha_node_t* prev_node;
buf_block_t* prev_block;
ulint hash;
@@ -136,9 +92,6 @@ ha_insert_for_fold(
#ifdef UNIV_SYNC_DEBUG
ut_ad(!table->mutexes || mutex_own(hash_get_mutex(table, fold)));
#endif /* UNIV_SYNC_DEBUG */
-
- block = buf_block_align(data);
-
hash = hash_calc_hash(fold, table);
cell = hash_get_nth_cell(table, hash);
@@ -151,15 +104,7 @@ ha_insert_for_fold(
prev_block = buf_block_align(prev_node->data);
ut_a(prev_block->n_pointers > 0);
prev_block->n_pointers--;
-
- block->n_pointers++;
-
- if (prev_block != block) {
- ha_remove_buf_block_node(prev_block,
- prev_node);
- ha_add_buf_block_node(block,
- prev_node);
- }
+ buf_block_align(data)->n_pointers++;
}
prev_node->data = data;
@@ -186,9 +131,7 @@ ha_insert_for_fold(
ha_node_set_data(node, data);
if (table->adaptive) {
- block->n_pointers++;
-
- ha_add_buf_block_node(block, node);
+ buf_block_align(data)->n_pointers++;
}
node->fold = fold;
@@ -223,15 +166,9 @@ ha_delete_hash_node(
hash_table_t* table, /* in: hash table */
ha_node_t* del_node) /* in: node to be deleted */
{
- buf_block_t* block;
-
if (table->adaptive) {
- block = buf_block_align(del_node->data);
-
- ut_a(block->n_pointers > 0);
-
- block->n_pointers--;
- ha_remove_buf_block_node(block, del_node);
+ ut_a(buf_block_align(del_node->data)->n_pointers > 0);
+ buf_block_align(del_node->data)->n_pointers--;
}
HASH_DELETE_AND_COMPACT(ha_node_t, next, table, del_node);
@@ -272,8 +209,6 @@ ha_search_and_update_if_found(
void* data, /* in: pointer to the data */
void* new_data)/* in: new pointer to the data */
{
- buf_block_t* old_block;
- buf_block_t* block;
ha_node_t* node;
#ifdef UNIV_SYNC_DEBUG
@@ -285,15 +220,8 @@ ha_search_and_update_if_found(
if (node) {
if (table->adaptive) {
ut_a(buf_block_align(node->data)->n_pointers > 0);
-
- old_block = buf_block_align(node->data);
- ut_a(old_block->n_pointers > 0);
- old_block->n_pointers--;
- ha_remove_buf_block_node(old_block, node);
-
- block = buf_block_align(new_data);
- block->n_pointers++;
- ha_add_buf_block_node(block, node);
+ buf_block_align(node->data)->n_pointers--;
+ buf_block_align(new_data)->n_pointers++;
}
node->data = new_data;
@@ -308,25 +236,43 @@ void
ha_remove_all_nodes_to_page(
/*========================*/
hash_table_t* table, /* in: hash table */
+ ulint fold, /* in: fold value */
page_t* page) /* in: buffer page */
{
- buf_block_t* block;
ha_node_t* node;
- block = buf_block_align(page);
-
- node = block->hash_nodes;
+#ifdef UNIV_SYNC_DEBUG
+ ut_ad(!table->mutexes || mutex_own(hash_get_mutex(table, fold)));
+#endif /* UNIV_SYNC_DEBUG */
+ node = ha_chain_get_first(table, fold);
while (node) {
- /* Remove the hash node */
+ if (buf_frame_align(ha_node_get_data(node)) == page) {
+
+ /* Remove the hash node */
+
+ ha_delete_hash_node(table, node);
- ha_delete_hash_node(table, node);
+ /* Start again from the first node in the chain
+ because the deletion may compact the heap of
+ nodes and move other nodes! */
- node = block->hash_nodes;
+ node = ha_chain_get_first(table, fold);
+ } else {
+ node = ha_chain_get_next(node);
+ }
}
+#ifdef UNIV_DEBUG
+ /* Check that all nodes really got deleted */
- ut_a(block->n_pointers == 0);
- ut_a(block->hash_nodes == NULL);
+ node = ha_chain_get_first(table, fold);
+
+ while (node) {
+ ut_a(buf_frame_align(ha_node_get_data(node)) != page);
+
+ node = ha_chain_get_next(node);
+ }
+#endif
}
/*****************************************************************
@@ -406,7 +352,6 @@ ha_print_info(
n_bufs++;
}
- fprintf(file, ", node heap has %lu buffer(s)\n",
- (ulong) n_bufs);
+ fprintf(file, ", node heap has %lu buffer(s)\n", (ulong) n_bufs);
}
}
diff --git a/innobase/include/buf0buf.h b/innobase/include/buf0buf.h
index c116d7e0b3e..8df1211327f 100644
--- a/innobase/include/buf0buf.h
+++ b/innobase/include/buf0buf.h
@@ -29,7 +29,6 @@ Created 11/5/1995 Heikki Tuuri
#include "buf0types.h"
#include "sync0rw.h"
#include "hash0hash.h"
-#include "ha0ha.h"
#include "ut0byte.h"
#include "os0proc.h"
@@ -828,7 +827,7 @@ struct buf_block_struct{
records with the same prefix should be
indexed in the hash index */
- /* The following 6 fields are protected by btr_search_latch: */
+ /* The following 4 fields are protected by btr_search_latch: */
ibool is_hashed; /* TRUE if hash index has already been
built on this page; note that it does
@@ -845,11 +844,6 @@ struct buf_block_struct{
ulint curr_side; /* BTR_SEARCH_LEFT_SIDE or
BTR_SEARCH_RIGHT_SIDE in hash
indexing */
- ha_node_t* hash_nodes; /* a doubly linked list of hash nodes
- for this buffer block; this points to
- the first node in the list if any;
- note that we do not use UT_LST_ macros
- to manipulate this list */
/* 6. Debug fields */
#ifdef UNIV_SYNC_DEBUG
rw_lock_t debug_latch; /* in the debug version, each thread
diff --git a/innobase/include/ha0ha.h b/innobase/include/ha0ha.h
index 5e3af9c1869..bdaecfcc57a 100644
--- a/innobase/include/ha0ha.h
+++ b/innobase/include/ha0ha.h
@@ -54,7 +54,7 @@ ha_create(
/*****************************************************************
Inserts an entry into a hash table. If an entry with the same fold number
is found, its node is updated to point to the new data, and no new node
-is inserted. This function is only used in the adaptive hash index. */
+is inserted. */
ibool
ha_insert_for_fold(
@@ -111,6 +111,7 @@ void
ha_remove_all_nodes_to_page(
/*========================*/
hash_table_t* table, /* in: hash table */
+ ulint fold, /* in: fold value */
page_t* page); /* in: buffer page */
/*****************************************************************
Validates a hash table. */
@@ -133,18 +134,9 @@ ha_print_info(
typedef struct ha_node_struct ha_node_t;
struct ha_node_struct {
- ha_node_t* next; /* next chain node; NULL if none */
- void* data; /* pointer to the data */
- ulint fold; /* fold value for the data */
- ha_node_t* next_for_block;/* in an adaptive hash index
- (btr0sea.c), a doubly linked list of hash
- nodes for the buffer block; these nodes
- contain pointers to index records on the
- page; in the last node this field is NULL;
- note that we do not use UT_LST_ macros
- to manipulate this list */
- ha_node_t* prev_for_block;/* pointer to the previous node; in the
- first node NULL */
+ ha_node_t* next; /* next chain node or NULL if none */
+ void* data; /* pointer to the data */
+ ulint fold; /* fold value for the data */
};
#ifndef UNIV_NONINL
diff --git a/innobase/include/hash0hash.h b/innobase/include/hash0hash.h
index befe8a8d757..51315e40875 100644
--- a/innobase/include/hash0hash.h
+++ b/innobase/include/hash0hash.h
@@ -166,7 +166,7 @@ hash_get_n_cells(
/***********************************************************************
Deletes a struct which is stored in the heap of the hash table, and compacts
the heap. The fold value must be stored in the struct NODE in a field named
-'fold'. This macro is only used for the adaptive hash index. */
+'fold'. */
#define HASH_DELETE_AND_COMPACT(TYPE, NAME, TABLE, NODE)\
do {\
@@ -192,23 +192,11 @@ do {\
\
*(NODE) = *top_node111;\
\
- /* Update the adaptive hash list of the buffer block that\
- corresponds to the top node */\
- if (top_node111->next_for_block != NULL) {\
- (top_node111->next_for_block)->prev_for_block = NODE;\
- }\
-\
- if (top_node111->prev_for_block != NULL) {\
- (top_node111->prev_for_block)->next_for_block = NODE;\
- } else {\
- buf_block_align(top_node111->data)->hash_nodes = NODE;\
- }\
-\
- /* Look for the hash pointer to the top node, to update it */\
-\
cell111 = hash_get_nth_cell(TABLE,\
hash_calc_hash(top_node111->fold, TABLE));\
\
+ /* Look for the pointer to the top node, to update it */\
+\
if (cell111->node == top_node111) {\
/* The top node is the first in the chain */\
\