1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
|
/*-
* Copyright (c) 2008-2014 WiredTiger, Inc.
* All rights reserved.
*
* See the file LICENSE for redistribution information.
*/
/*******************************************
* Global per-process structure.
*******************************************/
/*
* WT_PROCESS --
* Per-process information for the library.
*/
struct __wt_process {
WT_SPINLOCK spinlock; /* Per-process spinlock */
/* Locked: connection queue */
TAILQ_HEAD(__wt_connection_impl_qh, __wt_connection_impl) connqh;
WT_CACHE_POOL *cache_pool;
};
extern WT_PROCESS __wt_process;
/*
* WT_NAMED_COLLATOR --
* A collator list entry
*/
struct __wt_named_collator {
const char *name; /* Name of collator */
WT_COLLATOR *collator; /* User supplied object */
TAILQ_ENTRY(__wt_named_collator) q; /* Linked list of collators */
};
/*
* WT_NAMED_COMPRESSOR --
* A compressor list entry
*/
struct __wt_named_compressor {
const char *name; /* Name of compressor */
WT_COMPRESSOR *compressor; /* User supplied callbacks */
/* Linked list of compressors */
TAILQ_ENTRY(__wt_named_compressor) q;
};
/*
* WT_NAMED_DATA_SOURCE --
* A data source list entry
*/
struct __wt_named_data_source {
const char *prefix; /* Name of data source */
WT_DATA_SOURCE *dsrc; /* User supplied callbacks */
/* Linked list of data sources */
TAILQ_ENTRY(__wt_named_data_source) q;
};
/*
* WT_NAMED_EXTRACTOR --
* An extractor list entry
*/
struct __wt_named_extractor {
const char *name; /* Name of extractor */
WT_EXTRACTOR *extractor; /* User supplied object */
TAILQ_ENTRY(__wt_named_extractor) q; /* Linked list of extractors */
};
/*
* Allocate some additional slots for internal sessions. There is a default
* session for each connection, plus a session for each server thread.
*/
#define WT_NUM_INTERNAL_SESSIONS 10
/*
* WT_CONN_CHECK_PANIC --
* Check if we've panicked and return the appropriate error.
*/
#define WT_CONN_CHECK_PANIC(conn) \
(F_ISSET(conn, WT_CONN_PANIC) ? WT_PANIC : 0)
#define WT_SESSION_CHECK_PANIC(session) \
WT_CONN_CHECK_PANIC(S2C(session))
/*
* Macros to ensure the dhandle is inserted or removed from both the
* main queue and the hashed queue.
*/
#define WT_CONN_DHANDLE_INSERT(conn, dhandle, bucket) do { \
SLIST_INSERT_HEAD(&(conn)->dhlh, dhandle, l); \
SLIST_INSERT_HEAD(&(conn)->dhhash[bucket], dhandle, hashl); \
} while (0)
#define WT_CONN_DHANDLE_REMOVE(conn, dhandle, bucket) do { \
SLIST_REMOVE(&(conn)->dhlh, dhandle, __wt_data_handle, l); \
SLIST_REMOVE(&(conn)->dhhash[bucket], \
dhandle, __wt_data_handle, hashl); \
} while (0)
/*
* WT_CONNECTION_IMPL --
* Implementation of WT_CONNECTION
*/
struct __wt_connection_impl {
WT_CONNECTION iface;
/* For operations without an application-supplied session */
WT_SESSION_IMPL *default_session;
WT_SESSION_IMPL dummy_session;
const char *cfg; /* Connection configuration */
WT_SPINLOCK api_lock; /* Connection API spinlock */
WT_SPINLOCK checkpoint_lock; /* Checkpoint spinlock */
WT_SPINLOCK dhandle_lock; /* Data handle list spinlock */
WT_SPINLOCK fh_lock; /* File handle queue spinlock */
WT_SPINLOCK reconfig_lock; /* Single thread reconfigure */
WT_SPINLOCK schema_lock; /* Schema operation spinlock */
WT_SPINLOCK table_lock; /* Table creation spinlock */
/*
* We distribute the btree page locks across a set of spin locks; it
* can't be an array, we impose cache-line alignment and gcc doesn't
* support that for arrays. Don't use too many: they are only held for
* very short operations, each one is 64 bytes, so 256 will fill the L1
* cache on most CPUs.
*/
#define WT_PAGE_LOCKS(conn) 16
WT_SPINLOCK *page_lock; /* Btree page spinlocks */
u_int page_lock_cnt; /* Next spinlock to use */
/* Connection queue */
TAILQ_ENTRY(__wt_connection_impl) q;
/* Cache pool queue */
TAILQ_ENTRY(__wt_connection_impl) cpq;
const char *home; /* Database home */
const char *error_prefix; /* Database error prefix */
int is_new; /* Connection created database */
WT_EXTENSION_API extension_api; /* Extension API */
/* Configuration */
const WT_CONFIG_ENTRY **config_entries;
void **foc; /* Free-on-close array */
size_t foc_cnt; /* Array entries */
size_t foc_size; /* Array size */
WT_FH *lock_fh; /* Lock file handle */
uint64_t split_gen; /* Generation number for splits */
/*
* The connection keeps a cache of data handles. The set of handles
* can grow quite large so we maintain both a simple list and a hash
* table of lists. The hash table key is based on a hash of the table
* URI.
*/
/* Locked: data handle hash array */
#define WT_HASH_ARRAY_SIZE 512
SLIST_HEAD(__wt_dhhash, __wt_data_handle) dhhash[WT_HASH_ARRAY_SIZE];
/* Locked: data handle list */
SLIST_HEAD(__wt_dhandle_lh, __wt_data_handle) dhlh;
/* Locked: LSM handle list. */
TAILQ_HEAD(__wt_lsm_qh, __wt_lsm_tree) lsmqh;
/* Locked: file list */
TAILQ_HEAD(__wt_fh_qh, __wt_fh) fhqh;
/* Locked: library list */
TAILQ_HEAD(__wt_dlh_qh, __wt_dlh) dlhqh;
WT_SPINLOCK block_lock; /* Locked: block manager list */
TAILQ_HEAD(__wt_block_qh, __wt_block) blockqh;
u_int open_btree_count; /* Locked: open writable btree count */
uint32_t next_file_id; /* Locked: file ID counter */
/*
* WiredTiger allocates space for 50 simultaneous sessions (threads of
* control) by default. Growing the number of threads dynamically is
* possible, but tricky since server threads are walking the array
* without locking it.
*
* There's an array of WT_SESSION_IMPL pointers that reference the
* allocated array; we do it that way because we want an easy way for
* the server thread code to avoid walking the entire array when only a
* few threads are running.
*/
WT_SESSION_IMPL *sessions; /* Session reference */
uint32_t session_size; /* Session array size */
uint32_t session_cnt; /* Session count */
/*
* WiredTiger allocates space for a fixed number of hazard pointers
* in each thread of control.
*/
uint32_t hazard_max; /* Hazard array size */
WT_CACHE *cache; /* Page cache */
uint64_t cache_size;
WT_TXN_GLOBAL txn_global; /* Global transaction state */
WT_SPINLOCK hot_backup_lock; /* Hot backup serialization */
int hot_backup;
WT_SESSION_IMPL *ckpt_session; /* Checkpoint thread session */
wt_thread_t ckpt_tid; /* Checkpoint thread */
int ckpt_tid_set; /* Checkpoint thread set */
WT_CONDVAR *ckpt_cond; /* Checkpoint wait mutex */
const char *ckpt_config; /* Checkpoint configuration */
#define WT_CKPT_LOGSIZE(conn) ((conn)->ckpt_logsize != 0)
wt_off_t ckpt_logsize; /* Checkpoint log size period */
uint32_t ckpt_signalled;/* Checkpoint signalled */
long ckpt_usecs; /* Checkpoint period */
int compact_in_memory_pass; /* Compaction serialization */
#define WT_CONN_STAT_ALL 0x01 /* "all" statistics configured */
#define WT_CONN_STAT_CLEAR 0x02 /* clear after gathering */
#define WT_CONN_STAT_FAST 0x04 /* "fast" statistics configured */
#define WT_CONN_STAT_NONE 0x08 /* don't gather statistics */
#define WT_CONN_STAT_ON_CLOSE 0x10 /* output statistics on close */
uint32_t stat_flags;
WT_CONNECTION_STATS stats; /* Connection statistics */
#if SPINLOCK_TYPE == SPINLOCK_PTHREAD_MUTEX_LOGGING
/*
* Spinlock registration, so we can track which spinlocks are heavily
* used, which are blocking and where.
*
* There's an array of spinlocks, and an array of blocking IDs.
*/
#define WT_SPINLOCK_MAX 1024
#define WT_SPINLOCK_MAX_LOCATION_ID 60
WT_SPINLOCK *spinlock_list[WT_SPINLOCK_MAX];
/* Spinlock blocking matrix */
struct __wt_connection_stats_spinlock {
const char *name; /* Mutex name */
const char *file; /* Caller's file/line, ID location */
int line;
u_int total; /* Count of total, blocked calls */
u_int blocked[WT_SPINLOCK_MAX_LOCATION_ID];
} spinlock_block[WT_SPINLOCK_MAX_LOCATION_ID];
#endif
WT_ASYNC *async; /* Async structure */
int async_cfg; /* Global async configuration */
uint32_t async_size; /* Async op array size */
uint32_t async_workers; /* Number of async workers */
WT_LSM_MANAGER lsm_manager; /* LSM worker thread information */
WT_SESSION_IMPL *evict_session; /* Eviction server sessions */
wt_thread_t evict_tid; /* Eviction server thread ID */
int evict_tid_set; /* Eviction server thread ID set */
uint32_t evict_workers_max;/* Max eviction workers */
uint32_t evict_workers_min;/* Min eviction workers */
uint32_t evict_workers; /* Number of eviction workers */
WT_EVICT_WORKER *evict_workctx; /* Eviction worker context */
WT_SESSION_IMPL *stat_session; /* Statistics log session */
wt_thread_t stat_tid; /* Statistics log thread */
int stat_tid_set; /* Statistics log thread set */
WT_CONDVAR *stat_cond; /* Statistics log wait mutex */
const char *stat_format; /* Statistics log timestamp format */
FILE *stat_fp; /* Statistics log file handle */
char *stat_path; /* Statistics log path format */
char **stat_sources; /* Statistics log list of objects */
const char *stat_stamp; /* Statistics log entry timestamp */
long stat_usecs; /* Statistics log period */
#define WT_CONN_LOG_ARCHIVE 0x01 /* Archive is enabled */
#define WT_CONN_LOG_ENABLED 0x02 /* Logging is enabled */
#define WT_CONN_LOG_PREALLOC 0x04 /* Pre-allocation is enabled */
uint32_t log_flags; /* Global logging configuration */
WT_CONDVAR *log_cond; /* Log archive wait mutex */
WT_SESSION_IMPL *log_session; /* Log archive session */
wt_thread_t log_tid; /* Log archive thread */
int log_tid_set; /* Log archive thread set */
WT_LOG *log; /* Logging structure */
WT_COMPRESSOR *log_compressor;/* Logging compressor */
wt_off_t log_file_max; /* Log file max size */
const char *log_path; /* Logging path format */
uint32_t log_prealloc; /* Log file pre-allocation */
uint32_t txn_logsync; /* Log sync configuration */
WT_SESSION_IMPL *sweep_session; /* Handle sweep session */
wt_thread_t sweep_tid; /* Handle sweep thread */
int sweep_tid_set; /* Handle sweep thread set */
WT_CONDVAR *sweep_cond; /* Handle sweep wait mutex */
/* Locked: collator list */
TAILQ_HEAD(__wt_coll_qh, __wt_named_collator) collqh;
/* Locked: compressor list */
TAILQ_HEAD(__wt_comp_qh, __wt_named_compressor) compqh;
/* Locked: data source list */
TAILQ_HEAD(__wt_dsrc_qh, __wt_named_data_source) dsrcqh;
/* Locked: extractor list */
TAILQ_HEAD(__wt_extractor_qh, __wt_named_extractor) extractorqh;
void *lang_private; /* Language specific private storage */
/* If non-zero, all buffers used for I/O will be aligned to this. */
size_t buffer_alignment;
uint32_t schema_gen; /* Schema generation number */
wt_off_t data_extend_len; /* file_extend data length */
wt_off_t log_extend_len; /* file_extend log length */
uint32_t direct_io; /* O_DIRECT file type flags */
int mmap; /* mmap configuration */
uint32_t verbose;
uint32_t flags;
};
|