1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293
|
/*-
* See the file LICENSE for redistribution information.
*
* Copyright (c) 1996-2002
* Sleepycat Software. All rights reserved.
*
* $Id: mp.h,v 1.1.1.1 2003/11/20 22:13:17 toshok Exp $
*/
#ifndef _DB_MP_H_
#define _DB_MP_H_
struct __bh; typedef struct __bh BH;
struct __db_mpool_hash; typedef struct __db_mpool_hash DB_MPOOL_HASH;
struct __db_mpreg; typedef struct __db_mpreg DB_MPREG;
struct __mpool; typedef struct __mpool MPOOL;
/* We require at least 20KB of cache. */
#define DB_CACHESIZE_MIN (20 * 1024)
typedef enum {
DB_SYNC_ALLOC, /* Flush for allocation. */
DB_SYNC_CACHE, /* Checkpoint or flush entire cache. */
DB_SYNC_FILE, /* Flush file. */
DB_SYNC_TRICKLE /* Trickle sync. */
} db_sync_op;
/*
* DB_MPOOL --
* Per-process memory pool structure.
*/
struct __db_mpool {
/* These fields need to be protected for multi-threaded support. */
DB_MUTEX *mutexp; /* Structure thread lock. */
/* List of pgin/pgout routines. */
LIST_HEAD(__db_mpregh, __db_mpreg) dbregq;
/* List of DB_MPOOLFILE's. */
TAILQ_HEAD(__db_mpoolfileh, __db_mpoolfile) dbmfq;
/*
* The dbenv, nreg and reginfo fields are not thread protected,
* as they are initialized during mpool creation, and not modified
* again.
*/
DB_ENV *dbenv; /* Enclosing environment. */
u_int32_t nreg; /* N underlying cache regions. */
REGINFO *reginfo; /* Underlying cache regions. */
};
/*
* DB_MPREG --
* DB_MPOOL registry of pgin/pgout functions.
*/
struct __db_mpreg {
LIST_ENTRY(__db_mpreg) q; /* Linked list. */
int32_t ftype; /* File type. */
/* Pgin, pgout routines. */
int (*pgin) __P((DB_ENV *, db_pgno_t, void *, DBT *));
int (*pgout) __P((DB_ENV *, db_pgno_t, void *, DBT *));
};
/*
* NCACHE --
* Select a cache based on the file and the page number. Assumes accesses
* are uniform across pages, which is probably OK. What we really want to
* avoid is anything that puts all pages from any single file in the same
* cache, as we expect that file access will be bursty, and to avoid
* putting all page number N pages in the same cache as we expect access
* to the metapages (page 0) and the root of a btree (page 1) to be much
* more frequent than a random data page.
*/
#define NCACHE(mp, mf_offset, pgno) \
(((pgno) ^ ((mf_offset) >> 3)) % ((MPOOL *)mp)->nreg)
/*
* NBUCKET --
* We make the assumption that early pages of the file are more likely
* to be retrieved than the later pages, which means the top bits will
* be more interesting for hashing as they're less likely to collide.
* That said, as 512 8K pages represents a 4MB file, so only reasonably
* large files will have page numbers with any other than the bottom 9
* bits set. We XOR in the MPOOL offset of the MPOOLFILE that backs the
* page, since that should also be unique for the page. We don't want
* to do anything very fancy -- speed is more important to us than using
* good hashing.
*/
#define NBUCKET(mc, mf_offset, pgno) \
(((pgno) ^ ((mf_offset) << 9)) % (mc)->htab_buckets)
/*
* MPOOL --
* Shared memory pool region.
*/
struct __mpool {
/*
* The memory pool can be broken up into individual pieces/files.
* Not what we would have liked, but on Solaris you can allocate
* only a little more than 2GB of memory in a contiguous chunk,
* and I expect to see more systems with similar issues.
*
* While this structure is duplicated in each piece of the cache,
* the first of these pieces/files describes the entire pool, the
* second only describe a piece of the cache.
*/
/*
* The lsn field and list of underlying MPOOLFILEs are thread protected
* by the region lock.
*/
DB_LSN lsn; /* Maximum checkpoint LSN. */
SH_TAILQ_HEAD(__mpfq) mpfq; /* List of MPOOLFILEs. */
/*
* The nreg, regids and maint_off fields are not thread protected,
* as they are initialized during mpool creation, and not modified
* again.
*/
u_int32_t nreg; /* Number of underlying REGIONS. */
roff_t regids; /* Array of underlying REGION Ids. */
#ifdef HAVE_MUTEX_SYSTEM_RESOURCES
roff_t maint_off; /* Maintenance information offset */
#endif
/*
* The following structure fields only describe the per-cache portion
* of the region.
*
* The htab and htab_buckets fields are not thread protected as they
* are initialized during mpool creation, and not modified again.
*
* The last_checked and lru_count fields are thread protected by
* the region lock.
*/
int htab_buckets; /* Number of hash table entries. */
roff_t htab; /* Hash table offset. */
u_int32_t last_checked; /* Last bucket checked for free. */
u_int32_t lru_count; /* Counter for buffer LRU */
/*
* The stat fields are generally not thread protected, and cannot be
* trusted. Note that st_pages is an exception, and is always updated
* inside a region lock (although it is sometimes read outside of the
* region lock).
*/
DB_MPOOL_STAT stat; /* Per-cache mpool statistics. */
};
struct __db_mpool_hash {
DB_MUTEX hash_mutex; /* Per-bucket mutex. */
DB_HASHTAB hash_bucket; /* Head of bucket. */
u_int32_t hash_page_dirty;/* Count of dirty pages. */
u_int32_t hash_priority; /* Minimum priority of bucket buffer. */
};
/*
* The base mpool priority is 1/4th of the name space, or just under 2^30.
* When the LRU counter wraps, we shift everybody down to a base-relative
* value.
*/
#define MPOOL_BASE_DECREMENT (UINT32_T_MAX - (UINT32_T_MAX / 4))
/*
* Mpool priorities from low to high. Defined in terms of fractions of the
* buffers in the pool.
*/
#define MPOOL_PRI_VERY_LOW -1 /* Dead duck. Check and set to 0. */
#define MPOOL_PRI_LOW -2 /* Low. */
#define MPOOL_PRI_DEFAULT 0 /* No adjustment -- special case.*/
#define MPOOL_PRI_HIGH 10 /* With the dirty buffers. */
#define MPOOL_PRI_DIRTY 10 /* Dirty gets a 10% boost. */
#define MPOOL_PRI_VERY_HIGH 1 /* Add number of buffers in pool. */
/*
* MPOOLFILE_IGNORE --
* Discard an MPOOLFILE and any buffers it references: update the flags
* so we never try to write buffers associated with the file, nor can we
* find it when looking for files to join. In addition, clear the ftype
* field, there's no reason to post-process pages, they can be discarded
* by any thread.
*
* Expects the MPOOLFILE mutex to be held.
*/
#define MPOOLFILE_IGNORE(mfp) { \
(mfp)->ftype = 0; \
F_SET(mfp, MP_DEADFILE); \
}
/*
* MPOOLFILE --
* Shared DB_MPOOLFILE information.
*/
struct __mpoolfile {
DB_MUTEX mutex;
/* Protected by MPOOLFILE mutex. */
u_int32_t mpf_cnt; /* Ref count: DB_MPOOLFILEs. */
u_int32_t block_cnt; /* Ref count: blocks in cache. */
roff_t path_off; /* File name location. */
/* Protected by mpool cache 0 region lock. */
SH_TAILQ_ENTRY q; /* List of MPOOLFILEs */
db_pgno_t last_pgno; /* Last page in the file. */
db_pgno_t orig_last_pgno; /* Original last page in the file. */
/*
* None of the following fields are thread protected.
*
* There are potential races with the ftype field because it's read
* without holding a lock. However, it has to be set before adding
* any buffers to the cache that depend on it being set, so there
* would need to be incorrect operation ordering to have a problem.
*
* There are potential races with the priority field because it's read
* without holding a lock. However, a collision is unlikely and if it
* happens is of little consequence.
*
* We do not protect the statistics in "stat" because of the cost of
* the mutex in the get/put routines. There is a chance that a count
* will get lost.
*
* The remaining fields are initialized at open and never subsequently
* modified, except for the MP_DEADFILE, which is only set and never
* unset. (If there was more than one flag that was subsequently set,
* there might be a race, but with a single flag there can't be.)
*/
int32_t ftype; /* File type. */
int32_t priority; /* Priority when unpinning buffer. */
DB_MPOOL_FSTAT stat; /* Per-file mpool statistics. */
int32_t lsn_off; /* Page's LSN offset. */
u_int32_t clear_len; /* Bytes to clear on page create. */
roff_t fileid_off; /* File ID string location. */
roff_t pgcookie_len; /* Pgin/pgout cookie length. */
roff_t pgcookie_off; /* Pgin/pgout cookie location. */
#define MP_CAN_MMAP 0x01 /* If the file can be mmap'd. */
#define MP_DEADFILE 0x02 /* Dirty pages can simply be trashed. */
#define MP_DIRECT 0x04 /* No OS buffering. */
#define MP_EXTENT 0x08 /* Extent file. */
#define MP_TEMP 0x10 /* Backing file is a temporary. */
#define MP_UNLINK 0x20 /* Unlink file on last close. */
u_int32_t flags;
};
/*
* BH --
* Buffer header.
*/
struct __bh {
DB_MUTEX mutex; /* Buffer thread/process lock. */
u_int16_t ref; /* Reference count. */
u_int16_t ref_sync; /* Sync wait-for reference count. */
#define BH_CALLPGIN 0x001 /* Convert the page before use. */
#define BH_DIRTY 0x002 /* Page was modified. */
#define BH_DIRTY_CREATE 0x004 /* Page created, must be written. */
#define BH_DISCARD 0x008 /* Page is useless. */
#define BH_LOCKED 0x010 /* Page is locked (I/O in progress). */
#define BH_TRASH 0x020 /* Page is garbage. */
u_int16_t flags;
u_int32_t priority; /* LRU priority. */
SH_TAILQ_ENTRY hq; /* MPOOL hash bucket queue. */
db_pgno_t pgno; /* Underlying MPOOLFILE page number. */
roff_t mf_offset; /* Associated MPOOLFILE offset. */
/*
* !!!
* This array must be at least size_t aligned -- the DB access methods
* put PAGE and other structures into it, and then access them directly.
* (We guarantee size_t alignment to applications in the documentation,
* too.)
*/
u_int8_t buf[1]; /* Variable length data. */
};
#include "dbinc_auto/mp_ext.h"
#endif /* !_DB_MP_H_ */
|