1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
|
/*-
* See the file LICENSE for redistribution information.
*
* Copyright (c) 1996-2002
* Sleepycat Software. All rights reserved.
*
* Id: mp.h,v 11.44 2002/08/06 06:11:21 bostic Exp
*/
#ifndef _DB_MP_H_
#define _DB_MP_H_
struct __bh; typedef struct __bh BH;
struct __db_mpool_hash; typedef struct __db_mpool_hash DB_MPOOL_HASH;
struct __db_mpreg; typedef struct __db_mpreg DB_MPREG;
struct __mpool; typedef struct __mpool MPOOL;
/* We require at least 20KB of cache. */
#define DB_CACHESIZE_MIN (20 * 1024)
typedef enum {
DB_SYNC_ALLOC, /* Flush for allocation. */
DB_SYNC_CACHE, /* Checkpoint or flush entire cache. */
DB_SYNC_FILE, /* Flush file. */
DB_SYNC_TRICKLE /* Trickle sync. */
} db_sync_op;
/*
* DB_MPOOL --
* Per-process memory pool structure.
*/
struct __db_mpool {
/* These fields need to be protected for multi-threaded support. */
DB_MUTEX *mutexp; /* Structure thread lock. */
/* List of pgin/pgout routines. */
LIST_HEAD(__db_mpregh, __db_mpreg) dbregq;
/* List of DB_MPOOLFILE's. */
TAILQ_HEAD(__db_mpoolfileh, __db_mpoolfile) dbmfq;
/*
* The dbenv, nreg and reginfo fields are not thread protected,
* as they are initialized during mpool creation, and not modified
* again.
*/
DB_ENV *dbenv; /* Enclosing environment. */
u_int32_t nreg; /* N underlying cache regions. */
REGINFO *reginfo; /* Underlying cache regions. */
};
/*
* DB_MPREG --
* DB_MPOOL registry of pgin/pgout functions.
*/
struct __db_mpreg {
LIST_ENTRY(__db_mpreg) q; /* Linked list. */
int32_t ftype; /* File type. */
/* Pgin, pgout routines. */
int (*pgin) __P((DB_ENV *, db_pgno_t, void *, DBT *));
int (*pgout) __P((DB_ENV *, db_pgno_t, void *, DBT *));
};
/*
* NCACHE --
* Select a cache based on the file and the page number. Assumes accesses
* are uniform across pages, which is probably OK. What we really want to
* avoid is anything that puts all pages from any single file in the same
* cache, as we expect that file access will be bursty, and to avoid
* putting all page number N pages in the same cache as we expect access
* to the metapages (page 0) and the root of a btree (page 1) to be much
* more frequent than a random data page.
*/
#define NCACHE(mp, mf_offset, pgno) \
(((pgno) ^ ((mf_offset) >> 3)) % ((MPOOL *)mp)->nreg)
/*
* NBUCKET --
* We make the assumption that early pages of the file are more likely
* to be retrieved than the later pages, which means the top bits will
* be more interesting for hashing as they're less likely to collide.
* That said, as 512 8K pages represents a 4MB file, so only reasonably
* large files will have page numbers with any other than the bottom 9
* bits set. We XOR in the MPOOL offset of the MPOOLFILE that backs the
* page, since that should also be unique for the page. We don't want
* to do anything very fancy -- speed is more important to us than using
* good hashing.
*/
#define NBUCKET(mc, mf_offset, pgno) \
(((pgno) ^ ((mf_offset) << 9)) % (mc)->htab_buckets)
/*
* MPOOL --
* Shared memory pool region.
*/
struct __mpool {
/*
* The memory pool can be broken up into individual pieces/files.
* Not what we would have liked, but on Solaris you can allocate
* only a little more than 2GB of memory in a contiguous chunk,
* and I expect to see more systems with similar issues.
*
* While this structure is duplicated in each piece of the cache,
* the first of these pieces/files describes the entire pool, the
* second only describe a piece of the cache.
*/
/*
* The lsn field and list of underlying MPOOLFILEs are thread protected
* by the region lock.
*/
DB_LSN lsn; /* Maximum checkpoint LSN. */
SH_TAILQ_HEAD(__mpfq) mpfq; /* List of MPOOLFILEs. */
/*
* The nreg, regids and maint_off fields are not thread protected,
* as they are initialized during mpool creation, and not modified
* again.
*/
u_int32_t nreg; /* Number of underlying REGIONS. */
roff_t regids; /* Array of underlying REGION Ids. */
#ifdef HAVE_MUTEX_SYSTEM_RESOURCES
roff_t maint_off; /* Maintenance information offset */
#endif
/*
* The following structure fields only describe the per-cache portion
* of the region.
*
* The htab and htab_buckets fields are not thread protected as they
* are initialized during mpool creation, and not modified again.
*
* The last_checked and lru_count fields are thread protected by
* the region lock.
*/
int htab_buckets; /* Number of hash table entries. */
roff_t htab; /* Hash table offset. */
u_int32_t last_checked; /* Last bucket checked for free. */
u_int32_t lru_count; /* Counter for buffer LRU */
/*
* The stat fields are generally not thread protected, and cannot be
* trusted. Note that st_pages is an exception, and is always updated
* inside a region lock (although it is sometimes read outside of the
* region lock).
*/
DB_MPOOL_STAT stat; /* Per-cache mpool statistics. */
};
struct __db_mpool_hash {
DB_MUTEX hash_mutex; /* Per-bucket mutex. */
DB_HASHTAB hash_bucket; /* Head of bucket. */
u_int32_t hash_page_dirty;/* Count of dirty pages. */
u_int32_t hash_priority; /* Minimum priority of bucket buffer. */
};
/*
* The base mpool priority is 1/4th of the name space, or just under 2^30.
* When the LRU counter wraps, we shift everybody down to a base-relative
* value.
*/
#define MPOOL_BASE_DECREMENT (UINT32_T_MAX - (UINT32_T_MAX / 4))
/*
* Mpool priorities from low to high. Defined in terms of fractions of the
* buffers in the pool.
*/
#define MPOOL_PRI_VERY_LOW -1 /* Dead duck. Check and set to 0. */
#define MPOOL_PRI_LOW -2 /* Low. */
#define MPOOL_PRI_DEFAULT 0 /* No adjustment -- special case.*/
#define MPOOL_PRI_HIGH 10 /* With the dirty buffers. */
#define MPOOL_PRI_DIRTY 10 /* Dirty gets a 10% boost. */
#define MPOOL_PRI_VERY_HIGH 1 /* Add number of buffers in pool. */
/*
* MPOOLFILE_IGNORE --
* Discard an MPOOLFILE and any buffers it references: update the flags
* so we never try to write buffers associated with the file, nor can we
* find it when looking for files to join. In addition, clear the ftype
* field, there's no reason to post-process pages, they can be discarded
* by any thread.
*
* Expects the MPOOLFILE mutex to be held.
*/
#define MPOOLFILE_IGNORE(mfp) { \
(mfp)->ftype = 0; \
F_SET(mfp, MP_DEADFILE); \
}
/*
* MPOOLFILE --
* Shared DB_MPOOLFILE information.
*/
struct __mpoolfile {
DB_MUTEX mutex;
/* Protected by MPOOLFILE mutex. */
u_int32_t mpf_cnt; /* Ref count: DB_MPOOLFILEs. */
u_int32_t block_cnt; /* Ref count: blocks in cache. */
roff_t path_off; /* File name location. */
/* Protected by mpool cache 0 region lock. */
SH_TAILQ_ENTRY q; /* List of MPOOLFILEs */
db_pgno_t last_pgno; /* Last page in the file. */
db_pgno_t orig_last_pgno; /* Original last page in the file. */
/*
* None of the following fields are thread protected.
*
* There are potential races with the ftype field because it's read
* without holding a lock. However, it has to be set before adding
* any buffers to the cache that depend on it being set, so there
* would need to be incorrect operation ordering to have a problem.
*
* There are potential races with the priority field because it's read
* without holding a lock. However, a collision is unlikely and if it
* happens is of little consequence.
*
* We do not protect the statistics in "stat" because of the cost of
* the mutex in the get/put routines. There is a chance that a count
* will get lost.
*
* The remaining fields are initialized at open and never subsequently
* modified, except for the MP_DEADFILE, which is only set and never
* unset. (If there was more than one flag that was subsequently set,
* there might be a race, but with a single flag there can't be.)
*/
int32_t ftype; /* File type. */
int32_t priority; /* Priority when unpinning buffer. */
DB_MPOOL_FSTAT stat; /* Per-file mpool statistics. */
int32_t lsn_off; /* Page's LSN offset. */
u_int32_t clear_len; /* Bytes to clear on page create. */
roff_t fileid_off; /* File ID string location. */
roff_t pgcookie_len; /* Pgin/pgout cookie length. */
roff_t pgcookie_off; /* Pgin/pgout cookie location. */
#define MP_CAN_MMAP 0x01 /* If the file can be mmap'd. */
#define MP_DEADFILE 0x02 /* Dirty pages can simply be trashed. */
#define MP_DIRECT 0x04 /* No OS buffering. */
#define MP_EXTENT 0x08 /* Extent file. */
#define MP_TEMP 0x10 /* Backing file is a temporary. */
#define MP_UNLINK 0x20 /* Unlink file on last close. */
u_int32_t flags;
};
/*
* BH --
* Buffer header.
*/
struct __bh {
DB_MUTEX mutex; /* Buffer thread/process lock. */
u_int16_t ref; /* Reference count. */
u_int16_t ref_sync; /* Sync wait-for reference count. */
#define BH_CALLPGIN 0x001 /* Convert the page before use. */
#define BH_DIRTY 0x002 /* Page was modified. */
#define BH_DIRTY_CREATE 0x004 /* Page created, must be written. */
#define BH_DISCARD 0x008 /* Page is useless. */
#define BH_LOCKED 0x010 /* Page is locked (I/O in progress). */
#define BH_TRASH 0x020 /* Page is garbage. */
u_int16_t flags;
u_int32_t priority; /* LRU priority. */
SH_TAILQ_ENTRY hq; /* MPOOL hash bucket queue. */
db_pgno_t pgno; /* Underlying MPOOLFILE page number. */
roff_t mf_offset; /* Associated MPOOLFILE offset. */
/*
* !!!
* This array must be at least size_t aligned -- the DB access methods
* put PAGE and other structures into it, and then access them directly.
* (We guarantee size_t alignment to applications in the documentation,
* too.)
*/
u_int8_t buf[1]; /* Variable length data. */
};
#include "dbinc_auto/mp_ext.h"
#endif /* !_DB_MP_H_ */
|