diff options
Diffstat (limited to 'storage/bdb/lock/lock_region.c')
-rw-r--r-- | storage/bdb/lock/lock_region.c | 417 |
1 files changed, 417 insertions, 0 deletions
diff --git a/storage/bdb/lock/lock_region.c b/storage/bdb/lock/lock_region.c new file mode 100644 index 00000000000..6df6937e873 --- /dev/null +++ b/storage/bdb/lock/lock_region.c @@ -0,0 +1,417 @@ +/*- + * See the file LICENSE for redistribution information. + * + * Copyright (c) 1996-2002 + * Sleepycat Software. All rights reserved. + */ + +#include "db_config.h" + +#ifndef lint +static const char revid[] = "$Id: lock_region.c,v 11.69 2002/08/06 05:05:22 bostic Exp $"; +#endif /* not lint */ + +#ifndef NO_SYSTEM_INCLUDES +#include <sys/types.h> + +#include <string.h> +#endif + +#include "db_int.h" +#include "dbinc/db_shash.h" +#include "dbinc/lock.h" + +static int __lock_init __P((DB_ENV *, DB_LOCKTAB *)); +static size_t + __lock_region_size __P((DB_ENV *)); + +#ifdef HAVE_MUTEX_SYSTEM_RESOURCES +static size_t __lock_region_maint __P((DB_ENV *)); +#endif + +/* + * The conflict arrays are set up such that the row is the lock you are + * holding and the column is the lock that is desired. + */ +#define DB_LOCK_RIW_N 9 +static const u_int8_t db_riw_conflicts[] = { +/* N R W WT IW IR RIW DR WW */ +/* N */ 0, 0, 0, 0, 0, 0, 0, 0, 0, +/* R */ 0, 0, 1, 0, 1, 0, 1, 0, 1, +/* W */ 0, 1, 1, 1, 1, 1, 1, 1, 1, +/* WT */ 0, 0, 0, 0, 0, 0, 0, 0, 0, +/* IW */ 0, 1, 1, 0, 0, 0, 0, 1, 1, +/* IR */ 0, 0, 1, 0, 0, 0, 0, 0, 1, +/* RIW */ 0, 1, 1, 0, 0, 0, 0, 1, 1, +/* DR */ 0, 0, 1, 0, 1, 0, 1, 0, 0, +/* WW */ 0, 1, 1, 0, 1, 1, 1, 0, 1 +}; + +/* + * This conflict array is used for concurrent db access (CDB). It uses + * the same locks as the db_riw_conflicts array, but adds an IW mode to + * be used for write cursors. + */ +#define DB_LOCK_CDB_N 5 +static const u_int8_t db_cdb_conflicts[] = { + /* N R W WT IW */ + /* N */ 0, 0, 0, 0, 0, + /* R */ 0, 0, 1, 0, 0, + /* W */ 0, 1, 1, 1, 1, + /* WT */ 0, 0, 0, 0, 0, + /* IW */ 0, 0, 1, 0, 1 +}; + +/* + * __lock_open -- + * Internal version of lock_open: only called from DB_ENV->open. + * + * PUBLIC: int __lock_open __P((DB_ENV *)); + */ +int +__lock_open(dbenv) + DB_ENV *dbenv; +{ + DB_LOCKREGION *region; + DB_LOCKTAB *lt; + size_t size; + int ret; + + /* Create the lock table structure. */ + if ((ret = __os_calloc(dbenv, 1, sizeof(DB_LOCKTAB), <)) != 0) + return (ret); + lt->dbenv = dbenv; + + /* Join/create the lock region. */ + lt->reginfo.type = REGION_TYPE_LOCK; + lt->reginfo.id = INVALID_REGION_ID; + lt->reginfo.mode = dbenv->db_mode; + lt->reginfo.flags = REGION_JOIN_OK; + if (F_ISSET(dbenv, DB_ENV_CREATE)) + F_SET(<->reginfo, REGION_CREATE_OK); + size = __lock_region_size(dbenv); + if ((ret = __db_r_attach(dbenv, <->reginfo, size)) != 0) + goto err; + + /* If we created the region, initialize it. */ + if (F_ISSET(<->reginfo, REGION_CREATE)) + if ((ret = __lock_init(dbenv, lt)) != 0) + goto err; + + /* Set the local addresses. */ + region = lt->reginfo.primary = + R_ADDR(<->reginfo, lt->reginfo.rp->primary); + + /* Check for incompatible automatic deadlock detection requests. */ + if (dbenv->lk_detect != DB_LOCK_NORUN) { + if (region->detect != DB_LOCK_NORUN && + dbenv->lk_detect != DB_LOCK_DEFAULT && + region->detect != dbenv->lk_detect) { + __db_err(dbenv, + "lock_open: incompatible deadlock detector mode"); + ret = EINVAL; + goto err; + } + + /* + * Upgrade if our caller wants automatic detection, and it + * was not currently being done, whether or not we created + * the region. + */ + if (region->detect == DB_LOCK_NORUN) + region->detect = dbenv->lk_detect; + } + + /* + * A process joining the region may have reset the lock and transaction + * timeouts. + */ + if (dbenv->lk_timeout != 0) + region->lk_timeout = dbenv->lk_timeout; + if (dbenv->tx_timeout != 0) + region->tx_timeout = dbenv->tx_timeout; + + /* Set remaining pointers into region. */ + lt->conflicts = (u_int8_t *)R_ADDR(<->reginfo, region->conf_off); + lt->obj_tab = (DB_HASHTAB *)R_ADDR(<->reginfo, region->obj_off); + lt->locker_tab = (DB_HASHTAB *)R_ADDR(<->reginfo, region->locker_off); + + R_UNLOCK(dbenv, <->reginfo); + + dbenv->lk_handle = lt; + return (0); + +err: if (lt->reginfo.addr != NULL) { + if (F_ISSET(<->reginfo, REGION_CREATE)) + ret = __db_panic(dbenv, ret); + R_UNLOCK(dbenv, <->reginfo); + (void)__db_r_detach(dbenv, <->reginfo, 0); + } + __os_free(dbenv, lt); + return (ret); +} + +/* + * __lock_init -- + * Initialize the lock region. + */ +static int +__lock_init(dbenv, lt) + DB_ENV *dbenv; + DB_LOCKTAB *lt; +{ + const u_int8_t *lk_conflicts; + struct __db_lock *lp; + DB_LOCKER *lidp; + DB_LOCKOBJ *op; + DB_LOCKREGION *region; +#ifdef HAVE_MUTEX_SYSTEM_RESOURCES + size_t maint_size; +#endif + u_int32_t i, lk_modes; + u_int8_t *addr; + int ret; + + if ((ret = __db_shalloc(lt->reginfo.addr, + sizeof(DB_LOCKREGION), 0, <->reginfo.primary)) != 0) + goto mem_err; + lt->reginfo.rp->primary = R_OFFSET(<->reginfo, lt->reginfo.primary); + region = lt->reginfo.primary; + memset(region, 0, sizeof(*region)); + + /* Select a conflict matrix if none specified. */ + if (dbenv->lk_modes == 0) + if (CDB_LOCKING(dbenv)) { + lk_modes = DB_LOCK_CDB_N; + lk_conflicts = db_cdb_conflicts; + } else { + lk_modes = DB_LOCK_RIW_N; + lk_conflicts = db_riw_conflicts; + } + else { + lk_modes = dbenv->lk_modes; + lk_conflicts = dbenv->lk_conflicts; + } + + region->need_dd = 0; + region->detect = DB_LOCK_NORUN; + region->lk_timeout = dbenv->lk_timeout; + region->tx_timeout = dbenv->tx_timeout; + region->locker_t_size = __db_tablesize(dbenv->lk_max_lockers); + region->object_t_size = __db_tablesize(dbenv->lk_max_objects); + memset(®ion->stat, 0, sizeof(region->stat)); + region->stat.st_id = 0; + region->stat.st_cur_maxid = DB_LOCK_MAXID; + region->stat.st_maxlocks = dbenv->lk_max; + region->stat.st_maxlockers = dbenv->lk_max_lockers; + region->stat.st_maxobjects = dbenv->lk_max_objects; + region->stat.st_nmodes = lk_modes; + + /* Allocate room for the conflict matrix and initialize it. */ + if ((ret = + __db_shalloc(lt->reginfo.addr, lk_modes * lk_modes, 0, &addr)) != 0) + goto mem_err; + memcpy(addr, lk_conflicts, lk_modes * lk_modes); + region->conf_off = R_OFFSET(<->reginfo, addr); + + /* Allocate room for the object hash table and initialize it. */ + if ((ret = __db_shalloc(lt->reginfo.addr, + region->object_t_size * sizeof(DB_HASHTAB), 0, &addr)) != 0) + goto mem_err; + __db_hashinit(addr, region->object_t_size); + region->obj_off = R_OFFSET(<->reginfo, addr); + + /* Allocate room for the locker hash table and initialize it. */ + if ((ret = __db_shalloc(lt->reginfo.addr, + region->locker_t_size * sizeof(DB_HASHTAB), 0, &addr)) != 0) + goto mem_err; + __db_hashinit(addr, region->locker_t_size); + region->locker_off = R_OFFSET(<->reginfo, addr); + +#ifdef HAVE_MUTEX_SYSTEM_RESOURCES + maint_size = __lock_region_maint(dbenv); + /* Allocate room for the locker maintenance info and initialize it. */ + if ((ret = __db_shalloc(lt->reginfo.addr, + sizeof(REGMAINT) + maint_size, 0, &addr)) != 0) + goto mem_err; + __db_maintinit(<->reginfo, addr, maint_size); + region->maint_off = R_OFFSET(<->reginfo, addr); +#endif + + /* + * Initialize locks onto a free list. Initialize and lock the mutex + * so that when we need to block, all we need do is try to acquire + * the mutex. + */ + SH_TAILQ_INIT(®ion->free_locks); + for (i = 0; i < region->stat.st_maxlocks; ++i) { + if ((ret = __db_shalloc(lt->reginfo.addr, + sizeof(struct __db_lock), MUTEX_ALIGN, &lp)) != 0) + goto mem_err; + lp->status = DB_LSTAT_FREE; + lp->gen = 0; + if ((ret = __db_mutex_setup(dbenv, <->reginfo, &lp->mutex, + MUTEX_NO_RLOCK | MUTEX_SELF_BLOCK)) != 0) + return (ret); + MUTEX_LOCK(dbenv, &lp->mutex); + SH_TAILQ_INSERT_HEAD(®ion->free_locks, lp, links, __db_lock); + } + + /* Initialize objects onto a free list. */ + SH_TAILQ_INIT(®ion->dd_objs); + SH_TAILQ_INIT(®ion->free_objs); + for (i = 0; i < region->stat.st_maxobjects; ++i) { + if ((ret = __db_shalloc(lt->reginfo.addr, + sizeof(DB_LOCKOBJ), 0, &op)) != 0) + goto mem_err; + SH_TAILQ_INSERT_HEAD( + ®ion->free_objs, op, links, __db_lockobj); + } + + /* Initialize lockers onto a free list. */ + SH_TAILQ_INIT(®ion->lockers); + SH_TAILQ_INIT(®ion->free_lockers); + for (i = 0; i < region->stat.st_maxlockers; ++i) { + if ((ret = __db_shalloc(lt->reginfo.addr, + sizeof(DB_LOCKER), 0, &lidp)) != 0) { +mem_err: __db_err(dbenv, + "Unable to allocate memory for the lock table"); + return (ret); + } + SH_TAILQ_INSERT_HEAD( + ®ion->free_lockers, lidp, links, __db_locker); + } + + return (0); +} + +/* + * __lock_dbenv_refresh -- + * Clean up after the lock system on a close or failed open. Called + * only from __dbenv_refresh. (Formerly called __lock_close.) + * + * PUBLIC: int __lock_dbenv_refresh __P((DB_ENV *)); + */ +int +__lock_dbenv_refresh(dbenv) + DB_ENV *dbenv; +{ + DB_LOCKTAB *lt; + int ret; + + lt = dbenv->lk_handle; + + /* Detach from the region. */ + ret = __db_r_detach(dbenv, <->reginfo, 0); + + __os_free(dbenv, lt); + + dbenv->lk_handle = NULL; + return (ret); +} + +/* + * __lock_region_size -- + * Return the region size. + */ +static size_t +__lock_region_size(dbenv) + DB_ENV *dbenv; +{ + size_t retval; + + /* + * Figure out how much space we're going to need. This list should + * map one-to-one with the __db_shalloc calls in __lock_init. + */ + retval = 0; + retval += __db_shalloc_size(sizeof(DB_LOCKREGION), 1); + retval += __db_shalloc_size(dbenv->lk_modes * dbenv->lk_modes, 1); + retval += __db_shalloc_size( + __db_tablesize(dbenv->lk_max_lockers) * (sizeof(DB_HASHTAB)), 1); + retval += __db_shalloc_size( + __db_tablesize(dbenv->lk_max_objects) * (sizeof(DB_HASHTAB)), 1); +#ifdef HAVE_MUTEX_SYSTEM_RESOURCES + retval += + __db_shalloc_size(sizeof(REGMAINT) + __lock_region_maint(dbenv), 1); +#endif + retval += __db_shalloc_size( + sizeof(struct __db_lock), MUTEX_ALIGN) * dbenv->lk_max; + retval += __db_shalloc_size( + sizeof(DB_LOCKOBJ), 1) * dbenv->lk_max_objects; + retval += __db_shalloc_size( + sizeof(DB_LOCKER), 1) * dbenv->lk_max_lockers; + + /* + * Include 16 bytes of string space per lock. DB doesn't use it + * because we pre-allocate lock space for DBTs in the structure. + */ + retval += __db_shalloc_size(dbenv->lk_max * 16, sizeof(size_t)); + + /* And we keep getting this wrong, let's be generous. */ + retval += retval / 4; + + return (retval); +} + +#ifdef HAVE_MUTEX_SYSTEM_RESOURCES +/* + * __lock_region_maint -- + * Return the amount of space needed for region maintenance info. + */ +static size_t +__lock_region_maint(dbenv) + DB_ENV *dbenv; +{ + size_t s; + + s = sizeof(DB_MUTEX *) * dbenv->lk_max; + return (s); +} +#endif + +/* + * __lock_region_destroy + * Destroy any region maintenance info. + * + * PUBLIC: void __lock_region_destroy __P((DB_ENV *, REGINFO *)); + */ +void +__lock_region_destroy(dbenv, infop) + DB_ENV *dbenv; + REGINFO *infop; +{ + __db_shlocks_destroy(infop, (REGMAINT *)R_ADDR(infop, + ((DB_LOCKREGION *)R_ADDR(infop, infop->rp->primary))->maint_off)); + + COMPQUIET(dbenv, NULL); + COMPQUIET(infop, NULL); +} + +#ifdef CONFIG_TEST +/* + * __lock_id_set -- + * Set the current locker ID and current maximum unused ID (for + * testing purposes only). + * + * PUBLIC: int __lock_id_set __P((DB_ENV *, u_int32_t, u_int32_t)); + */ +int +__lock_id_set(dbenv, cur_id, max_id) + DB_ENV *dbenv; + u_int32_t cur_id, max_id; +{ + DB_LOCKTAB *lt; + DB_LOCKREGION *region; + + ENV_REQUIRES_CONFIG(dbenv, + dbenv->lk_handle, "lock_id_set", DB_INIT_LOCK); + + lt = dbenv->lk_handle; + region = lt->reginfo.primary; + region->stat.st_id = cur_id; + region->stat.st_cur_maxid = max_id; + + return (0); +} +#endif |