diff options
-rw-r--r-- | innobase/btr/btr0sea.c | 74 | ||||
-rw-r--r-- | innobase/buf/buf0buf.c | 1 | ||||
-rw-r--r-- | innobase/buf/buf0lru.c | 1 | ||||
-rw-r--r-- | innobase/ha/ha0ha.c | 125 | ||||
-rw-r--r-- | innobase/include/buf0buf.h | 8 | ||||
-rw-r--r-- | innobase/include/ha0ha.h | 18 | ||||
-rw-r--r-- | innobase/include/hash0hash.h | 18 |
7 files changed, 114 insertions, 131 deletions
diff --git a/innobase/btr/btr0sea.c b/innobase/btr/btr0sea.c index 468c5efd24d..f2f5d09fa0e 100644 --- a/innobase/btr/btr0sea.c +++ b/innobase/btr/btr0sea.c @@ -439,8 +439,8 @@ btr_search_update_hash_ref( ha_insert_for_fold(btr_search_sys->hash_index, fold, rec); } -} - +} + /************************************************************************* Updates the search info. */ @@ -926,6 +926,17 @@ btr_search_drop_page_hash_index( { hash_table_t* table; buf_block_t* block; + ulint n_fields; + ulint n_bytes; + rec_t* rec; + rec_t* sup; + ulint fold; + ulint prev_fold; + dulint tree_id; + ulint n_cached; + ulint n_recs; + ulint* folds; + ulint i; #ifdef UNIV_SYNC_DEBUG ut_ad(!rw_lock_own(&btr_search_latch, RW_LOCK_SHARED)); @@ -951,18 +962,73 @@ btr_search_drop_page_hash_index( || (block->buf_fix_count == 0)); #endif /* UNIV_SYNC_DEBUG */ - ut_a(block->curr_n_fields + block->curr_n_bytes > 0); + n_fields = block->curr_n_fields; + n_bytes = block->curr_n_bytes; + + ut_a(n_fields + n_bytes > 0); rw_lock_s_unlock(&btr_search_latch); + n_recs = page_get_n_recs(page); + + /* Calculate and cache fold values into an array for fast deletion + from the hash index */ + + folds = mem_alloc(n_recs * sizeof(ulint)); + + n_cached = 0; + + sup = page_get_supremum_rec(page); + + rec = page_get_infimum_rec(page); + rec = page_rec_get_next(rec); + + if (rec != sup) { + ut_a(n_fields <= rec_get_n_fields(rec)); + + if (n_bytes > 0) { + ut_a(n_fields < rec_get_n_fields(rec)); + } + } + + tree_id = btr_page_get_index_id(page); + + prev_fold = 0; + + while (rec != sup) { + /* FIXME: in a mixed tree, not all records may have enough + ordering fields: */ + + fold = rec_fold(rec, n_fields, n_bytes, tree_id); + + if (fold == prev_fold && prev_fold != 0) { + + goto next_rec; + } + + /* Remove all hash nodes pointing to this page from the + hash chain */ + + folds[n_cached] = fold; + n_cached++; +next_rec: + rec = page_rec_get_next(rec); + prev_fold = fold; + } + rw_lock_x_lock(&btr_search_latch); - ha_remove_all_nodes_to_page(table, page); + for (i = 0; i < n_cached; i++) { + + ha_remove_all_nodes_to_page(table, folds[i], page); + } block->is_hashed = FALSE; block->index = NULL; rw_lock_x_unlock(&btr_search_latch); + + mem_free(folds); } /************************************************************************ diff --git a/innobase/buf/buf0buf.c b/innobase/buf/buf0buf.c index 8ca353aa5fb..3930ea93889 100644 --- a/innobase/buf/buf0buf.c +++ b/innobase/buf/buf0buf.c @@ -466,7 +466,6 @@ buf_block_init( block->in_LRU_list = FALSE; block->n_pointers = 0; - block->hash_nodes = NULL; rw_lock_create(&(block->lock)); ut_ad(rw_lock_validate(&(block->lock))); diff --git a/innobase/buf/buf0lru.c b/innobase/buf/buf0lru.c index 985426a9e2b..8460a049d3e 100644 --- a/innobase/buf/buf0lru.c +++ b/innobase/buf/buf0lru.c @@ -823,7 +823,6 @@ buf_LRU_block_free_non_file_page( || (block->state == BUF_BLOCK_READY_FOR_USE)); ut_a(block->n_pointers == 0); - ut_a(block->hash_nodes == NULL); ut_a(!block->in_free_list); block->state = BUF_BLOCK_NOT_USED; diff --git a/innobase/ha/ha0ha.c b/innobase/ha/ha0ha.c index 49b59882626..ad1391ff83e 100644 --- a/innobase/ha/ha0ha.c +++ b/innobase/ha/ha0ha.c @@ -66,52 +66,9 @@ ha_create( } /***************************************************************** -Removes an adaptive hash index node from the doubly linked list of hash nodes -for the buffer block. */ -UNIV_INLINE -void -ha_remove_buf_block_node( -/*=====================*/ - buf_block_t* block, /* in: buffer block */ - ha_node_t* node) /* in: an adaptive hash index node */ -{ - if (node == block->hash_nodes) { - block->hash_nodes = node->next_for_block; - } - - if (node->prev_for_block != NULL) { - (node->prev_for_block)->next_for_block = node->next_for_block; - } - - if (node->next_for_block != NULL) { - (node->next_for_block)->prev_for_block = node->prev_for_block; - } -} - -/***************************************************************** -Adds an adaptive hash index node to the start of the doubly linked list of -hash nodes for the buffer block. */ -UNIV_INLINE -void -ha_add_buf_block_node( -/*==================*/ - buf_block_t* block, /* in: buffer block */ - ha_node_t* node) /* in: an adaptive hash index node */ -{ - node->next_for_block = block->hash_nodes; - node->prev_for_block = NULL; - - block->hash_nodes = node; - - if (node->next_for_block != NULL) { - (node->next_for_block)->prev_for_block = node; - } -} - -/***************************************************************** Inserts an entry into a hash table. If an entry with the same fold number is found, its node is updated to point to the new data, and no new node -is inserted. This function is only used in the adaptive hash index. */ +is inserted. */ ibool ha_insert_for_fold( @@ -127,7 +84,6 @@ ha_insert_for_fold( { hash_cell_t* cell; ha_node_t* node; - buf_block_t* block; ha_node_t* prev_node; buf_block_t* prev_block; ulint hash; @@ -136,9 +92,6 @@ ha_insert_for_fold( #ifdef UNIV_SYNC_DEBUG ut_ad(!table->mutexes || mutex_own(hash_get_mutex(table, fold))); #endif /* UNIV_SYNC_DEBUG */ - - block = buf_block_align(data); - hash = hash_calc_hash(fold, table); cell = hash_get_nth_cell(table, hash); @@ -151,15 +104,7 @@ ha_insert_for_fold( prev_block = buf_block_align(prev_node->data); ut_a(prev_block->n_pointers > 0); prev_block->n_pointers--; - - block->n_pointers++; - - if (prev_block != block) { - ha_remove_buf_block_node(prev_block, - prev_node); - ha_add_buf_block_node(block, - prev_node); - } + buf_block_align(data)->n_pointers++; } prev_node->data = data; @@ -186,9 +131,7 @@ ha_insert_for_fold( ha_node_set_data(node, data); if (table->adaptive) { - block->n_pointers++; - - ha_add_buf_block_node(block, node); + buf_block_align(data)->n_pointers++; } node->fold = fold; @@ -223,15 +166,9 @@ ha_delete_hash_node( hash_table_t* table, /* in: hash table */ ha_node_t* del_node) /* in: node to be deleted */ { - buf_block_t* block; - if (table->adaptive) { - block = buf_block_align(del_node->data); - - ut_a(block->n_pointers > 0); - - block->n_pointers--; - ha_remove_buf_block_node(block, del_node); + ut_a(buf_block_align(del_node->data)->n_pointers > 0); + buf_block_align(del_node->data)->n_pointers--; } HASH_DELETE_AND_COMPACT(ha_node_t, next, table, del_node); @@ -272,8 +209,6 @@ ha_search_and_update_if_found( void* data, /* in: pointer to the data */ void* new_data)/* in: new pointer to the data */ { - buf_block_t* old_block; - buf_block_t* block; ha_node_t* node; #ifdef UNIV_SYNC_DEBUG @@ -285,15 +220,8 @@ ha_search_and_update_if_found( if (node) { if (table->adaptive) { ut_a(buf_block_align(node->data)->n_pointers > 0); - - old_block = buf_block_align(node->data); - ut_a(old_block->n_pointers > 0); - old_block->n_pointers--; - ha_remove_buf_block_node(old_block, node); - - block = buf_block_align(new_data); - block->n_pointers++; - ha_add_buf_block_node(block, node); + buf_block_align(node->data)->n_pointers--; + buf_block_align(new_data)->n_pointers++; } node->data = new_data; @@ -308,25 +236,43 @@ void ha_remove_all_nodes_to_page( /*========================*/ hash_table_t* table, /* in: hash table */ + ulint fold, /* in: fold value */ page_t* page) /* in: buffer page */ { - buf_block_t* block; ha_node_t* node; - block = buf_block_align(page); - - node = block->hash_nodes; +#ifdef UNIV_SYNC_DEBUG + ut_ad(!table->mutexes || mutex_own(hash_get_mutex(table, fold))); +#endif /* UNIV_SYNC_DEBUG */ + node = ha_chain_get_first(table, fold); while (node) { - /* Remove the hash node */ + if (buf_frame_align(ha_node_get_data(node)) == page) { + + /* Remove the hash node */ + + ha_delete_hash_node(table, node); - ha_delete_hash_node(table, node); + /* Start again from the first node in the chain + because the deletion may compact the heap of + nodes and move other nodes! */ - node = block->hash_nodes; + node = ha_chain_get_first(table, fold); + } else { + node = ha_chain_get_next(node); + } } +#ifdef UNIV_DEBUG + /* Check that all nodes really got deleted */ - ut_a(block->n_pointers == 0); - ut_a(block->hash_nodes == NULL); + node = ha_chain_get_first(table, fold); + + while (node) { + ut_a(buf_frame_align(ha_node_get_data(node)) != page); + + node = ha_chain_get_next(node); + } +#endif } /***************************************************************** @@ -406,7 +352,6 @@ ha_print_info( n_bufs++; } - fprintf(file, ", node heap has %lu buffer(s)\n", - (ulong) n_bufs); + fprintf(file, ", node heap has %lu buffer(s)\n", (ulong) n_bufs); } } diff --git a/innobase/include/buf0buf.h b/innobase/include/buf0buf.h index c116d7e0b3e..8df1211327f 100644 --- a/innobase/include/buf0buf.h +++ b/innobase/include/buf0buf.h @@ -29,7 +29,6 @@ Created 11/5/1995 Heikki Tuuri #include "buf0types.h" #include "sync0rw.h" #include "hash0hash.h" -#include "ha0ha.h" #include "ut0byte.h" #include "os0proc.h" @@ -828,7 +827,7 @@ struct buf_block_struct{ records with the same prefix should be indexed in the hash index */ - /* The following 6 fields are protected by btr_search_latch: */ + /* The following 4 fields are protected by btr_search_latch: */ ibool is_hashed; /* TRUE if hash index has already been built on this page; note that it does @@ -845,11 +844,6 @@ struct buf_block_struct{ ulint curr_side; /* BTR_SEARCH_LEFT_SIDE or BTR_SEARCH_RIGHT_SIDE in hash indexing */ - ha_node_t* hash_nodes; /* a doubly linked list of hash nodes - for this buffer block; this points to - the first node in the list if any; - note that we do not use UT_LST_ macros - to manipulate this list */ /* 6. Debug fields */ #ifdef UNIV_SYNC_DEBUG rw_lock_t debug_latch; /* in the debug version, each thread diff --git a/innobase/include/ha0ha.h b/innobase/include/ha0ha.h index 5e3af9c1869..bdaecfcc57a 100644 --- a/innobase/include/ha0ha.h +++ b/innobase/include/ha0ha.h @@ -54,7 +54,7 @@ ha_create( /***************************************************************** Inserts an entry into a hash table. If an entry with the same fold number is found, its node is updated to point to the new data, and no new node -is inserted. This function is only used in the adaptive hash index. */ +is inserted. */ ibool ha_insert_for_fold( @@ -111,6 +111,7 @@ void ha_remove_all_nodes_to_page( /*========================*/ hash_table_t* table, /* in: hash table */ + ulint fold, /* in: fold value */ page_t* page); /* in: buffer page */ /***************************************************************** Validates a hash table. */ @@ -133,18 +134,9 @@ ha_print_info( typedef struct ha_node_struct ha_node_t; struct ha_node_struct { - ha_node_t* next; /* next chain node; NULL if none */ - void* data; /* pointer to the data */ - ulint fold; /* fold value for the data */ - ha_node_t* next_for_block;/* in an adaptive hash index - (btr0sea.c), a doubly linked list of hash - nodes for the buffer block; these nodes - contain pointers to index records on the - page; in the last node this field is NULL; - note that we do not use UT_LST_ macros - to manipulate this list */ - ha_node_t* prev_for_block;/* pointer to the previous node; in the - first node NULL */ + ha_node_t* next; /* next chain node or NULL if none */ + void* data; /* pointer to the data */ + ulint fold; /* fold value for the data */ }; #ifndef UNIV_NONINL diff --git a/innobase/include/hash0hash.h b/innobase/include/hash0hash.h index befe8a8d757..51315e40875 100644 --- a/innobase/include/hash0hash.h +++ b/innobase/include/hash0hash.h @@ -166,7 +166,7 @@ hash_get_n_cells( /*********************************************************************** Deletes a struct which is stored in the heap of the hash table, and compacts the heap. The fold value must be stored in the struct NODE in a field named -'fold'. This macro is only used for the adaptive hash index. */ +'fold'. */ #define HASH_DELETE_AND_COMPACT(TYPE, NAME, TABLE, NODE)\ do {\ @@ -192,23 +192,11 @@ do {\ \ *(NODE) = *top_node111;\ \ - /* Update the adaptive hash list of the buffer block that\ - corresponds to the top node */\ - if (top_node111->next_for_block != NULL) {\ - (top_node111->next_for_block)->prev_for_block = NODE;\ - }\ -\ - if (top_node111->prev_for_block != NULL) {\ - (top_node111->prev_for_block)->next_for_block = NODE;\ - } else {\ - buf_block_align(top_node111->data)->hash_nodes = NODE;\ - }\ -\ - /* Look for the hash pointer to the top node, to update it */\ -\ cell111 = hash_get_nth_cell(TABLE,\ hash_calc_hash(top_node111->fold, TABLE));\ \ + /* Look for the pointer to the top node, to update it */\ +\ if (cell111->node == top_node111) {\ /* The top node is the first in the chain */\ \ |