You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

784 lines
21 KiB

/*-
* See the file LICENSE for redistribution information.
*
* Copyright (c) 1996, 1997, 1998, 1999
* Sleepycat Software. All rights reserved.
*/
/*
* Copyright (c) 1990, 1993, 1994
* The Regents of the University of California. All rights reserved.
*
* This code is derived from software contributed to Berkeley by
* Margo Seltzer.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
* 3. Neither the name of the University nor the names of its contributors
* may be used to endorse or promote products derived from this software
* without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
* ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
* SUCH DAMAGE.
*/
#include "db_config.h"
#ifndef lint
static const char sccsid[] = "@(#)hash_dup.c 11.17 (Sleepycat) 11/14/99";
#endif /* not lint */
/*
* PACKAGE: hashing
*
* DESCRIPTION:
* Manipulation of duplicates for the hash package.
*
* ROUTINES:
*
* External
* __add_dup
* Internal
*/
#ifndef NO_SYSTEM_INCLUDES
#include <sys/types.h>
#ifdef DIAGNOSTIC
#include <assert.h>
#endif
#include <errno.h>
#include <string.h>
#endif
#include "db_int.h"
#include "db_page.h"
#include "hash.h"
#include "btree.h"
static int CDB___ham_check_move __P((DBC *, u_int32_t));
/*
* Called from hash_access to add a duplicate key. nval is the new
* value that we want to add. The flags correspond to the flag values
* to cursor_put indicating where to add the new element.
* There are 4 cases.
* Case 1: The existing duplicate set already resides on a separate page.
* We can use common code for this.
* Case 2: The element is small enough to just be added to the existing set.
* Case 3: The element is large enough to be a big item, so we're going to
* have to push the set onto a new page.
* Case 4: The element is large enough to push the duplicate set onto a
* separate page.
*
* PUBLIC: int CDB___ham_add_dup __P((DBC *, DBT *, u_int32_t));
*/
int
CDB___ham_add_dup(dbc, nval, flags)
DBC *dbc;
DBT *nval;
u_int32_t flags;
{
DB *dbp;
HASH_CURSOR *hcp;
DBT pval, tmp_val;
u_int32_t add_bytes, new_size;
db_indx_t dndx;
int cmp, ret;
u_int8_t *hk;
dbp = dbc->dbp;
hcp = (HASH_CURSOR *)dbc->internal;
#ifdef DIAGNOSTIC
assert(flags != DB_CURRENT);
#endif
add_bytes = nval->size +
(F_ISSET(nval, DB_DBT_PARTIAL) ? nval->doff : 0);
add_bytes = DUP_SIZE(add_bytes);
if ((ret = CDB___ham_check_move(dbc, add_bytes)) != 0)
return (ret);
/*
* Check if resulting duplicate set is going to need to go
* onto a separate duplicate page. If so, convert the
* duplicate set and add the new one. After conversion,
* hcp->dndx is the first free ndx or the index of the
* current pointer into the duplicate set.
*/
hk = H_PAIRDATA(hcp->pagep, hcp->bndx);
new_size =
LEN_HKEYDATA(hcp->pagep, dbp->pgsize, H_DATAINDEX(hcp->bndx)) +
add_bytes;
/*
* We convert to off-page duplicates if the item is a big item,
* the addition of the new item will make the set large, or
* if there isn't enough room on this page to add the next item.
*/
if (HPAGE_PTYPE(hk) != H_OFFDUP &&
(HPAGE_PTYPE(hk) == H_OFFPAGE || ISBIG(hcp, new_size) ||
add_bytes > P_FREESPACE(hcp->pagep))) {
if ((ret = CDB___ham_dup_convert(dbc)) != 0)
return (ret);
hk = H_PAIRDATA(hcp->pagep, hcp->bndx);
}
/* There are two separate cases here: on page and off page. */
if (HPAGE_PTYPE(hk) != H_OFFDUP) {
if (HPAGE_PTYPE(hk) != H_DUPLICATE) {
pval.flags = 0;
pval.data = HKEYDATA_DATA(hk);
pval.size = LEN_HDATA(hcp->pagep, dbp->pgsize,
hcp->bndx);
if ((ret =
CDB___ham_make_dup(&pval, &tmp_val, &dbc->rdata.data,
&dbc->rdata.size)) != 0 || (ret =
CDB___ham_replpair(dbc, &tmp_val, 1)) != 0)
return (ret);
hk = H_PAIRDATA(hcp->pagep, hcp->bndx);
HPAGE_PTYPE(hk) = H_DUPLICATE;
/*
* Update the cursor position since we now are in
* duplicates.
*/
F_SET(hcp, H_ISDUP);
hcp->dup_off = 0;
hcp->dup_len = pval.size;
hcp->dup_tlen = DUP_SIZE(hcp->dup_len);
}
/* Now make the new entry a duplicate. */
if ((ret = CDB___ham_make_dup(nval,
&tmp_val, &dbc->rdata.data, &dbc->rdata.size)) != 0)
return (ret);
tmp_val.dlen = 0;
switch (flags) { /* On page. */
case DB_KEYFIRST:
case DB_KEYLAST:
if (dbp->dup_compare != NULL)
CDB___ham_dsearch(dbc, nval, &tmp_val.doff, &cmp);
else {
hcp->dup_tlen = LEN_HDATA(hcp->pagep,
dbp->pgsize, hcp->bndx);
hcp->dup_len = nval->size;
F_SET(hcp, H_ISDUP);
if (flags == DB_KEYFIRST)
hcp->dup_off = tmp_val.doff = 0;
else
hcp->dup_off =
tmp_val.doff = hcp->dup_tlen;
}
break;
case DB_BEFORE:
tmp_val.doff = hcp->dup_off;
break;
case DB_AFTER:
tmp_val.doff = hcp->dup_off + DUP_SIZE(hcp->dup_len);
break;
}
/* Add the duplicate. */
ret = CDB___ham_replpair(dbc, &tmp_val, 0);
if (ret == 0)
ret = CDB___ham_dirty_page(dbp, hcp->pagep);
/* Now, update the cursor if necessary. */
switch (flags) {
case DB_AFTER:
hcp->dup_off += DUP_SIZE(hcp->dup_len);
hcp->dup_len = nval->size;
hcp->dup_tlen += DUP_SIZE(nval->size);
break;
case DB_KEYFIRST:
case DB_KEYLAST:
case DB_BEFORE:
hcp->dup_tlen += DUP_SIZE(nval->size);
hcp->dup_len = nval->size;
break;
}
CDB___ham_c_update(hcp, hcp->pgno, tmp_val.size, 1, 1);
return (ret);
}
/* If we get here, then we're on duplicate pages. */
if (hcp->dpgno == PGNO_INVALID) {
memcpy(&hcp->dpgno, HOFFDUP_PGNO(hk), sizeof(db_pgno_t));
hcp->dndx = 0;
}
switch (flags) {
case DB_KEYFIRST:
if (dbp->dup_compare != NULL)
goto sorted_dups;
/*
* The only way that we are already on a dup page is
* if we just converted the on-page representation.
* In that case, we've only got one page of duplicates.
*/
if (hcp->dpagep == NULL && (ret =
CDB___db_dend(dbc, hcp->dpgno, &hcp->dpagep)) != 0)
return (ret);
hcp->dndx = 0;
break;
case DB_KEYLAST:
if (dbp->dup_compare != NULL) {
sorted_dups: if ((ret = CDB___db_dsearch(dbc, 1, nval,
hcp->dpgno, &hcp->dndx, &hcp->dpagep, &cmp)) != 0)
return (ret);
if (cmp == 0)
hcp->dpgno = PGNO(hcp->dpagep);
} else {
if (hcp->dpagep == NULL && (ret =
CDB___db_dend(dbc, hcp->dpgno, &hcp->dpagep)) != 0)
return (ret);
hcp->dpgno = PGNO(hcp->dpagep);
hcp->dndx = NUM_ENT(hcp->dpagep);
}
break;
case DB_BEFORE: /* The default behavior is correct. */
break;
case DB_AFTER:
hcp->dndx++;
break;
}
/*
* On page splits, the 4th parameter of db_dput returns the location
* the new item was put. We cannot pass in permanent fields from the
* cursor, they may have been updated in cursor adjustment.
*/
dndx = hcp->dndx;
ret = CDB___db_dput(dbc, nval, &hcp->dpagep, &dndx);
hcp->dpgno = PGNO(hcp->dpagep);
hcp->dndx = dndx;
CDB___ham_c_update(hcp, hcp->dpgno, nval->size, 1, 1);
return (ret);
}
/*
* Convert an on-page set of duplicates to an offpage set of duplicates.
*
* PUBLIC: int CDB___ham_dup_convert __P((DBC *));
*/
int
CDB___ham_dup_convert(dbc)
DBC *dbc;
{
DB *dbp;
HASH_CURSOR *hcp, **hcs;
BOVERFLOW bo;
DBT dbt;
HOFFPAGE ho;
db_indx_t dndx, i, len, off;
int c, ret;
u_int8_t *p, *pend;
/*
* Create a new page for the duplicates.
*/
dbp = dbc->dbp;
hcp = (HASH_CURSOR *)dbc->internal;
if ((ret = CDB___db_new(dbc, P_DUPLICATE, &hcp->dpagep)) != 0)
return (ret);
hcp->dpagep->type = P_DUPLICATE;
hcp->dpgno = PGNO(hcp->dpagep);
/*
* Now put the duplicates onto the new page.
*/
dndx = 0;
dbt.flags = 0;
switch (HPAGE_PTYPE(H_PAIRDATA(hcp->pagep, hcp->bndx))) {
case H_KEYDATA:
/* Simple case, one key on page; move it to dup page. */
dbt.size =
LEN_HDATA(hcp->pagep, dbp->pgsize, hcp->bndx);
dbt.data = HKEYDATA_DATA(H_PAIRDATA(hcp->pagep, hcp->bndx));
ret = CDB___db_pitem(dbc, hcp->dpagep,
(u_int32_t)dndx, BKEYDATA_SIZE(dbt.size), NULL, &dbt);
if (ret == 0)
CDB___ham_dirty_page(dbp, hcp->dpagep);
break;
case H_OFFPAGE:
/* Simple case, one key on page; move it to dup page. */
memcpy(&ho,
P_ENTRY(hcp->pagep, H_DATAINDEX(hcp->bndx)), HOFFPAGE_SIZE);
UMRW(bo.unused1);
B_TSET(bo.type, ho.type, 0);
UMRW(bo.unused2);
bo.pgno = ho.pgno;
bo.tlen = ho.tlen;
dbt.size = BOVERFLOW_SIZE;
dbt.data = &bo;
ret = CDB___db_pitem(dbc, hcp->dpagep,
(u_int32_t)dndx, dbt.size, &dbt, NULL);
if (ret == 0)
CDB___ham_dirty_page(dbp, hcp->dpagep);
break;
case H_DUPLICATE:
/*
* Get the list of cursors that may need to be updated.
*/
if ((ret = CDB___ham_get_clist(dbp,
PGNO(hcp->pagep), hcp->bndx, &hcs)) != 0)
break;
p = HKEYDATA_DATA(H_PAIRDATA(hcp->pagep, hcp->bndx));
pend = p +
LEN_HDATA(hcp->pagep, dbp->pgsize, hcp->bndx);
/*
* We need to maintain the duplicate cursor position.
* Keep track of where we are in the duplicate set via
* the offset, and when it matches the one in the cursor,
* set the off-page duplicate cursor index to the current
* index.
*/
for (off = 0, i = 0; p < pend; i++) {
if (off == hcp->dup_off)
dndx = i;
memcpy(&len, p, sizeof(db_indx_t));
dbt.size = len;
p += sizeof(db_indx_t);
dbt.data = p;
p += len + sizeof(db_indx_t);
if ((ret = CDB___db_dput(dbc, &dbt, &hcp->dpagep, &i)) != 0)
break;
/*
* Update any other cursors
*/
for (c = 0; hcs != NULL && hcs[c] != NULL; c++)
if (hcs[c]->dup_off == off) {
hcs[c]->dpgno = PGNO(hcp->dpagep);
hcs[c]->dndx = i;
}
off += len + 2 * sizeof(db_indx_t);
}
if (hcs != NULL)
CDB___os_free(hcs, 0);
break;
default:
ret = CDB___db_pgfmt(dbp, (u_long)hcp->pgno);
break;
}
if (ret == 0) {
/*
* Now attach this to the source page in place of
* the old duplicate item.
*/
CDB___ham_move_offpage(dbc, hcp->pagep,
(u_int32_t)H_DATAINDEX(hcp->bndx), hcp->dpgno);
/* Can probably just do a "put" here. */
ret = CDB___ham_dirty_page(dbp, hcp->pagep);
hcp->dndx = dndx;
} else {
(void)CDB___db_free(dbc, hcp->dpagep);
hcp->dpagep = NULL;
}
return (ret);
}
/*
* CDB___ham_make_dup
*
* Take a regular dbt and make it into a duplicate item with all the partial
* information set appropriately. If the incoming dbt is a partial, assume
* we are creating a new entry and make sure that we do any initial padding.
*
* PUBLIC: int CDB___ham_make_dup __P((const DBT *, DBT *d, void **, u_int32_t *));
*/
int
CDB___ham_make_dup(notdup, duplicate, bufp, sizep)
const DBT *notdup;
DBT *duplicate;
void **bufp;
u_int32_t *sizep;
{
db_indx_t tsize, item_size;
int ret;
u_int8_t *p;
item_size = (db_indx_t)notdup->size;
if (F_ISSET(notdup, DB_DBT_PARTIAL))
item_size += notdup->doff;
tsize = DUP_SIZE(item_size);
if ((ret = CDB___ham_init_dbt(duplicate, tsize, bufp, sizep)) != 0)
return (ret);
duplicate->dlen = 0;
duplicate->flags = notdup->flags;
F_SET(duplicate, DB_DBT_PARTIAL);
p = duplicate->data;
memcpy(p, &item_size, sizeof(db_indx_t));
p += sizeof(db_indx_t);
if (F_ISSET(notdup, DB_DBT_PARTIAL)) {
memset(p, 0, notdup->doff);
p += notdup->doff;
}
memcpy(p, notdup->data, notdup->size);
p += notdup->size;
memcpy(p, &item_size, sizeof(db_indx_t));
duplicate->doff = 0;
duplicate->dlen = notdup->size;
return (0);
}
/*
* CDB___ham_check_move --
*
* Check if we can do whatever we need to on this page. If not,
* then we'll have to move the current element to a new page.
*/
static int
CDB___ham_check_move(dbc, add_len)
DBC *dbc;
u_int32_t add_len;
{
DB *dbp;
HASH_CURSOR *hcp;
DBT k, d;
DB_LSN new_lsn;
PAGE *next_pagep;
db_pgno_t next_pgno;
u_int32_t new_datalen, old_len, rectype;
u_int8_t *hk;
int ret;
dbp = dbc->dbp;
hcp = (HASH_CURSOR *)dbc->internal;
hk = H_PAIRDATA(hcp->pagep, hcp->bndx);
/*
* If the item is already off page duplicates or an offpage item,
* then we know we can do whatever we need to do in-place
*/
if (HPAGE_PTYPE(hk) == H_OFFDUP || HPAGE_PTYPE(hk) == H_OFFPAGE)
return (0);
old_len = LEN_HITEM(hcp->pagep, dbp->pgsize, H_DATAINDEX(hcp->bndx));
new_datalen = old_len - HKEYDATA_SIZE(0) + add_len;
if (HPAGE_PTYPE(hk) != H_DUPLICATE)
new_datalen += DUP_SIZE(0);
/*
* We need to add a new page under two conditions:
* 1. The addition makes the total data length cross the BIG
* threshold and the OFFDUP structure won't fit on this page.
* 2. The addition does not make the total data cross the
* threshold, but the new data won't fit on the page.
* If neither of these is true, then we can return.
*/
if (ISBIG(hcp, new_datalen) && (old_len > HOFFDUP_SIZE ||
HOFFDUP_SIZE - old_len <= P_FREESPACE(hcp->pagep)))
return (0);
if (!ISBIG(hcp, new_datalen) && add_len <= P_FREESPACE(hcp->pagep))
return (0);
/*
* If we get here, then we need to move the item to a new page.
* Check if there are more pages in the chain.
*/
new_datalen = ISBIG(hcp, new_datalen) ?
HOFFDUP_SIZE : HKEYDATA_SIZE(new_datalen);
next_pagep = NULL;
for (next_pgno = NEXT_PGNO(hcp->pagep); next_pgno != PGNO_INVALID;
next_pgno = NEXT_PGNO(next_pagep)) {
if (next_pagep != NULL &&
(ret = CDB___ham_put_page(dbp, next_pagep, 0)) != 0)
return (ret);
if ((ret =
CDB___ham_get_page(dbp, next_pgno, &next_pagep)) != 0)
return (ret);
if (P_FREESPACE(next_pagep) >= new_datalen)
break;
}
/* No more pages, add one. */
if (next_pagep == NULL && (ret = CDB___ham_add_ovflpage(dbc,
hcp->pagep, 0, &next_pagep)) != 0)
return (ret);
/* Add new page at the end of the chain. */
if (P_FREESPACE(next_pagep) < new_datalen && (ret =
CDB___ham_add_ovflpage(dbc, next_pagep, 1, &next_pagep)) != 0) {
(void)CDB___ham_put_page(dbp, next_pagep, 0);
return (ret);
}
/* Copy the item to the new page. */
if (DB_LOGGING(hcp->dbc)) {
rectype = PUTPAIR;
k.flags = 0;
d.flags = 0;
if (HPAGE_PTYPE(
H_PAIRKEY(hcp->pagep, hcp->bndx)) == H_OFFPAGE) {
rectype |= PAIR_KEYMASK;
k.data = H_PAIRKEY(hcp->pagep, hcp->bndx);
k.size = HOFFPAGE_SIZE;
} else {
k.data =
HKEYDATA_DATA(H_PAIRKEY(hcp->pagep, hcp->bndx));
k.size = LEN_HKEY(hcp->pagep, dbp->pgsize, hcp->bndx);
}
if (HPAGE_PTYPE(hk) == H_OFFPAGE) {
rectype |= PAIR_DATAMASK;
d.data = H_PAIRDATA(hcp->pagep, hcp->bndx);
d.size = HOFFPAGE_SIZE;
} else {
d.data =
HKEYDATA_DATA(H_PAIRDATA(hcp->pagep, hcp->bndx));
d.size = LEN_HDATA(hcp->pagep, dbp->pgsize, hcp->bndx);
}
if ((ret = CDB___ham_insdel_log(dbp->dbenv,
dbc->txn, &new_lsn, 0, rectype,
dbp->log_fileid, PGNO(next_pagep),
(u_int32_t)H_NUMPAIRS(next_pagep), &LSN(next_pagep),
&k, &d)) != 0)
return (ret);
/* Move lsn onto page. */
LSN(next_pagep) = new_lsn; /* Structure assignment. */
}
CDB___ham_copy_item(dbp->pgsize,
hcp->pagep, H_KEYINDEX(hcp->bndx), next_pagep);
CDB___ham_copy_item(dbp->pgsize,
hcp->pagep, H_DATAINDEX(hcp->bndx), next_pagep);
/* Now delete the pair from the current page. */
ret = CDB___ham_del_pair(dbc, 0);
/*
* CDB___ham_del_pair decremented nelem. This is incorrect; we
* manually copied the element elsewhere, so the total number
* of elements hasn't changed. Increment it again.
*/
if (!F_ISSET(dbp->dbenv, DB_ENV_LOCKING))
hcp->hdr->nelem++;
(void)CDB___ham_put_page(dbp, hcp->pagep, 1);
hcp->pagep = next_pagep;
hcp->pgno = PGNO(hcp->pagep);
hcp->bndx = H_NUMPAIRS(hcp->pagep) - 1;
F_SET(hcp, H_EXPAND);
return (ret);
}
/*
* CDB___ham_move_offpage --
* Replace an onpage set of duplicates with the OFFDUP structure
* that references the duplicate page.
*
* XXX
* This is really just a special case of __onpage_replace; we should
* probably combine them.
*
* PUBLIC: void CDB___ham_move_offpage __P((DBC *, PAGE *, u_int32_t, db_pgno_t));
*/
void
CDB___ham_move_offpage(dbc, pagep, ndx, pgno)
DBC *dbc;
PAGE *pagep;
u_int32_t ndx;
db_pgno_t pgno;
{
DB *dbp;
HASH_CURSOR *hcp;
DBT new_dbt;
DBT old_dbt;
HOFFDUP od;
db_indx_t i;
int32_t shrink;
u_int8_t *src;
dbp = dbc->dbp;
hcp = (HASH_CURSOR *)dbc->internal;
od.type = H_OFFDUP;
UMRW(od.unused[0]);
UMRW(od.unused[1]);
UMRW(od.unused[2]);
od.pgno = pgno;
if (DB_LOGGING(dbc)) {
new_dbt.data = &od;
new_dbt.size = HOFFDUP_SIZE;
old_dbt.data = P_ENTRY(pagep, ndx);
old_dbt.size = LEN_HITEM(pagep, dbp->pgsize, ndx);
(void)CDB___ham_replace_log(dbp->dbenv,
dbc->txn, &LSN(pagep), 0, dbp->log_fileid,
PGNO(pagep), (u_int32_t)ndx, &LSN(pagep), -1,
&old_dbt, &new_dbt, 0);
}
shrink = LEN_HITEM(pagep, dbp->pgsize, ndx) - HOFFDUP_SIZE;
if (shrink != 0) {
/* Copy data. */
src = (u_int8_t *)(pagep) + HOFFSET(pagep);
memmove(src + shrink, src, pagep->inp[ndx] - HOFFSET(pagep));
HOFFSET(pagep) += shrink;
/* Update index table. */
for (i = ndx; i < NUM_ENT(pagep); i++)
pagep->inp[i] += shrink;
}
/* Now copy the offdup entry onto the page. */
memcpy(P_ENTRY(pagep, ndx), &od, HOFFDUP_SIZE);
}
/*
* CDB___ham_dsearch:
* Locate a particular duplicate in a duplicate set. Make sure that
* we exit with the cursor set appropriately.
*
* PUBLIC: void CDB___ham_dsearch __P((DBC *, DBT *, u_int32_t *, int *));
*/
void
CDB___ham_dsearch(dbc, dbt, offp, cmpp)
DBC *dbc;
DBT *dbt;
u_int32_t *offp;
int *cmpp;
{
DB *dbp;
HASH_CURSOR *hcp;
DBT cur;
db_indx_t i, len;
int (*func) __P((const DBT *, const DBT *));
u_int8_t *data;
dbp = dbc->dbp;
hcp = (HASH_CURSOR *)dbc->internal;
if (dbp->dup_compare == NULL)
func = CDB___bam_defcmp;
else
func = dbp->dup_compare;
i = F_ISSET(dbc, DBC_CONTINUE) ? hcp->dup_off: 0;
data = HKEYDATA_DATA(H_PAIRDATA(hcp->pagep, hcp->bndx)) + i;
hcp->dup_tlen = LEN_HDATA(hcp->pagep, dbp->pgsize, hcp->bndx);
while (i < hcp->dup_tlen) {
memcpy(&len, data, sizeof(db_indx_t));
data += sizeof(db_indx_t);
cur.data = data;
cur.size = (u_int32_t)len;
*cmpp = func(dbt, &cur);
if (*cmpp == 0 || (*cmpp < 0 && dbp->dup_compare != NULL))
break;
i += len + 2 * sizeof(db_indx_t);
data += len + sizeof(db_indx_t);
}
*offp = i;
hcp->dup_off = i;
hcp->dup_len = len;
F_SET(hcp, H_ISDUP);
}
/*
* CDB___ham_ca_split --
* Adjust the cursors when splitting a page.
*
* PUBLIC: void CDB___ham_ca_split __P((DB *,
* PUBLIC: db_pgno_t, db_pgno_t, db_pgno_t, u_int32_t, int));
*/
void
CDB___ham_ca_split(dbp, ppgno, lpgno, rpgno, split_indx, cleft)
DB *dbp;
db_pgno_t ppgno, lpgno, rpgno;
u_int32_t split_indx;
int cleft;
{
HASH_CURSOR *hcp;
DBC *dbc;
/*
* If splitting the page that a cursor was on, the cursor has to be
* adjusted to point to the same record as before the split. Most
* of the time we don't adjust pointers to the left page, because
* we're going to copy its contents back over the original page. If
* the cursor is on the right page, it is decremented by the number of
* records split to the left page.
*/
MUTEX_THREAD_LOCK(dbp->mutexp);
for (dbc = TAILQ_FIRST(&dbp->active_queue);
dbc != NULL; dbc = TAILQ_NEXT(dbc, links)) {
hcp = (HASH_CURSOR *)dbc->internal;
if (hcp->dpgno == ppgno) {
if (hcp->dndx < split_indx) {
if (cleft)
hcp->dpgno = lpgno;
} else {
hcp->dpgno = rpgno;
hcp->dndx -= split_indx;
}
}
}
MUTEX_THREAD_UNLOCK(dbp->mutexp);
}
#ifdef DEBUG
/*
* CDB___ham_cprint --
* Display the current cursor list.
*
* PUBLIC: int CDB___ham_cprint __P((DB *));
*/
int
CDB___ham_cprint(dbp)
DB *dbp;
{
HASH_CURSOR *cp;
DBC *dbc;
MUTEX_THREAD_LOCK(dbp->mutexp);
for (dbc = TAILQ_FIRST(&dbp->active_queue);
dbc != NULL; dbc = TAILQ_NEXT(dbc, links)) {
cp = (HASH_CURSOR *)dbc->internal;
fprintf(stderr,
"%#0x->%#0x: page: %lu index: %lu dpage %lu dindex: %lu",
(u_int)dbc, (u_int)cp, (u_long)cp->pgno, (u_long)cp->bndx,
(u_long)cp->dpgno, (u_long)cp->dndx);
if (F_ISSET(cp, H_DELETED))
fprintf(stderr, " (deleted)");
fprintf(stderr, "\n");
}
MUTEX_THREAD_UNLOCK(dbp->mutexp);
return (0);
}
#endif /* DEBUG */