/*- * See the file LICENSE for redistribution information. * * Copyright (c) 1996, 1997, 1998, 1999, 2000 * Sleepycat Software. All rights reserved. */ /* * Copyright (c) 1995, 1996 * Margo Seltzer. All rights reserved. */ /* * Copyright (c) 1995, 1996 * The President and Fellows of Harvard University. All rights reserved. * * This code is derived from software contributed to Berkeley by * Margo Seltzer. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * 3. Neither the name of the University nor the names of its contributors * may be used to endorse or promote products derived from this software * without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * * $Id: hash.src,v 10.24 2000/12/12 17:41:48 bostic Exp $ */ /* * This is the source file used to create the logging functions for the * hash package. Each access method (or set of routines wishing to register * record types with the transaction system) should have a file like this. * Each type of log record and its parameters is defined. The basic * format of a record definition is: * * BEGIN * ARG|STRING|POINTER * ... * END * ARG the argument is a simple parameter of the type * specified. * DBT the argument is a DBT (db.h) containing a length and pointer. * PTR the argument is a pointer to the data type specified; the entire * type should be logged. * * There are a set of shell scripts of the form xxx.sh that generate c * code and or h files to process these. (This is probably better done * in a single PERL script, but for now, this works.) * * The DB recovery system requires the following three fields appear in * every record, and will assign them to the per-record-type structures * as well as making them the first parameters to the appropriate logging * call. * rectype: record-type, identifies the structure and log/read call * txnid: transaction id, a DBT in this implementation * prev: the last LSN for this transaction */ /* * Use the argument of PREFIX as the prefix for all record types, * routines, id numbers, etc. */ PREFIX ham INCLUDE #include "db_config.h" INCLUDE INCLUDE #ifndef NO_SYSTEM_INCLUDES INCLUDE #include INCLUDE INCLUDE #include INCLUDE #include INCLUDE #include INCLUDE #endif INCLUDE INCLUDE #include "db_int.h" INCLUDE #include "db_page.h" INCLUDE #include "db_dispatch.h" INCLUDE #include "db_am.h" INCLUDE #include "hash.h" INCLUDE #include "txn.h" INCLUDE /* * HASH-insdel: used for hash to insert/delete a pair of entries onto a master * page. The pair might be regular key/data pairs or they might be the * structures that refer to off page items, duplicates or offpage duplicates. * opcode - PUTPAIR/DELPAIR + big masks * fileid - identifies the file referenced * pgno - page within file * ndx - index on the page of the item being added (item index) * pagelsn - lsn on the page before the update * key - the key being inserted * data - the data being inserted */ BEGIN insdel 21 ARG opcode u_int32_t lu ARG fileid int32_t ld ARG pgno db_pgno_t lu ARG ndx u_int32_t lu POINTER pagelsn DB_LSN * lu DBT key DBT s DBT data DBT s END /* * Used to add and remove overflow pages. * prev_pgno is the previous page that is going to get modified to * point to this one. If this is the first page in a chain * then prev_pgno should be PGNO_INVALID. * new_pgno is the page being allocated. * next_pgno is the page that follows this one. On allocation, * this should be PGNO_INVALID. For deletes, it may exist. * pagelsn is the old lsn on the page. */ BEGIN newpage 22 ARG opcode u_int32_t lu ARG fileid int32_t ld ARG prev_pgno db_pgno_t lu POINTER prevlsn DB_LSN * lu ARG new_pgno db_pgno_t lu POINTER pagelsn DB_LSN * lu ARG next_pgno db_pgno_t lu POINTER nextlsn DB_LSN * lu END /* * DEPRECATED in 3.0. * Superceded by metagroup which allocates a group of new pages. * * Splitting requires two types of log messages. The first logs the * meta-data of the split. * * For the meta-data split * bucket: max_bucket in table before split * ovflpoint: overflow point before split. * spares: spares[ovflpoint] before split. */ DEPRECATED splitmeta 23 ARG fileid int32_t ld ARG bucket u_int32_t lu ARG ovflpoint u_int32_t lu ARG spares u_int32_t lu POINTER metalsn DB_LSN * lu END /* * Splitting requires two types of log messages. The second logs the * data on the original page. To redo the split, we have to visit the * new page (pages) and add the items back on the page if they are not * yet there. */ BEGIN splitdata 24 ARG fileid int32_t ld ARG opcode u_int32_t lu ARG pgno db_pgno_t lu DBT pageimage DBT s POINTER pagelsn DB_LSN * lu END /* * HASH-replace: is used for hash to handle partial puts that only * affect a single master page. * fileid - identifies the file referenced * pgno - page within file * ndx - index on the page of the item being modified (item index) * pagelsn - lsn on the page before the update * off - offset in the old item where the new item is going. * olditem - DBT that describes the part of the item being replaced. * newitem - DBT of the new item. * makedup - this was a replacement that made an item a duplicate. */ BEGIN replace 25 ARG fileid int32_t ld ARG pgno db_pgno_t lu ARG ndx u_int32_t lu POINTER pagelsn DB_LSN * lu ARG off int32_t ld DBT olditem DBT s DBT newitem DBT s ARG makedup u_int32_t lu END /* * DEPRECATED in 3.0. * Hash now uses the btree allocation and deletion page routines. * * HASH-newpgno: is used to record getting/deleting a new page number. * This doesn't require much data modification, just modifying the * meta-data. * pgno is the page being allocated/freed. * free_pgno is the next_pgno on the free list. * old_type was the type of a page being deallocated. * old_pgno was the next page number before the deallocation. */ DEPRECATED newpgno 26 ARG opcode u_int32_t lu ARG fileid int32_t ld ARG pgno db_pgno_t lu ARG free_pgno db_pgno_t lu ARG old_type u_int32_t lu ARG old_pgno db_pgno_t lu ARG new_type u_int32_t lu POINTER pagelsn DB_LSN * lu POINTER metalsn DB_LSN * lu END /* * DEPRECATED in 3.0. * Since we now pre-allocate the contiguous chunk of pages for a doubling, * there is no big benefit to pre-allocating a few extra pages. It used * to be that the file was only physically as large as the current bucket, * so if you were on a doubling of 16K, but were only on the first bucket * of that 16K, the file was much shorter than it would be at the end of * the doubling, so we didn't want to force overflow pages at the end of the * 16K pages. Since we now must allocate the 16K pages (because of sub * databases), it's not a big deal to tack extra pages on at the end. * * ovfl: initialize a set of overflow pages. */ DEPRECATED ovfl 27 ARG fileid int32_t ld ARG start_pgno db_pgno_t lu ARG npages u_int32_t lu ARG free_pgno db_pgno_t lu ARG ovflpoint u_int32_t lu POINTER metalsn DB_LSN * lu END /* * Used when we empty the first page in a bucket and there are pages after * it. The page after it gets copied into the bucket page (since bucket * pages have to be in fixed locations). * pgno: the bucket page * pagelsn: the old LSN on the bucket page * next_pgno: the page number of the next page * nnext_pgno: page after next_pgno (may need to change its prev) * nnextlsn: the LSN of nnext_pgno. */ BEGIN copypage 28 ARG fileid int32_t ld ARG pgno db_pgno_t lu POINTER pagelsn DB_LSN * lu ARG next_pgno db_pgno_t lu POINTER nextlsn DB_LSN * lu ARG nnext_pgno db_pgno_t lu POINTER nnextlsn DB_LSN * lu DBT page DBT s END /* * This replaces the old splitmeta operation. It behaves largely the same * way, but it has enough information so that we can record a group allocation * which we do now because of sub databases. The number of pages allocated is * always bucket + 1 pgno is the page number of the first newly allocated * bucket. * bucket: Old maximum bucket number. * pgno: Page allocated to bucket + 1 (first newly allocated page) * metalsn: Lsn of the meta-data page. * pagelsn: Lsn of the maximum page allocated. */ BEGIN metagroup 29 ARG fileid int32_t ld ARG bucket u_int32_t lu ARG pgno db_pgno_t lu POINTER metalsn DB_LSN * lu POINTER pagelsn DB_LSN * lu END /* * groupalloc * * This is used in conjunction with MPOOL_NEW_GROUP when we are creating * a new database to make sure that we recreate or reclaim free pages * when we allocate a chunk of contiguous ones during database creation. * * pgno: meta-data page number * metalsn: meta-data lsn * start_pgno: starting page number * num: number of allocated pages */ DEPRECATED groupalloc1 30 ARG fileid int32_t ld ARG pgno db_pgno_t lu POINTER metalsn DB_LSN * lu POINTER mmetalsn DB_LSN * lu ARG start_pgno db_pgno_t lu ARG num u_int32_t lu END DEPRECATED groupalloc2 31 ARG fileid int32_t ld POINTER meta_lsn DB_LSN * lu POINTER alloc_lsn DB_LSN * lu ARG start_pgno db_pgno_t lu ARG num u_int32_t lu ARG free db_pgno_t lu END BEGIN groupalloc 32 ARG fileid int32_t ld POINTER meta_lsn DB_LSN * lu ARG start_pgno db_pgno_t lu ARG num u_int32_t lu ARG free db_pgno_t lu END /* * Records for backing out cursor adjustment. * curadj - added or deleted a record or a dup * within a record. * pgno - page that was effected * indx - indx of recrod effected. * len - if a dup its length. * dup_off - if a dup its offset * add - 1 if add 0 if delete * is_dup - 1 if dup 0 otherwise. * order - order assinged to this deleted record or dup. * * chgpg - rmoved a page, move the records to a new page * mode - CHGPG page was deleted or records move to new page. * - SPLIT we split a bucket * - DUP we convered to off page duplicates. * old_pgno, new_pgno - old and new page numbers. * old_index, new_index - old and new index numbers, NDX_INVALID if * it effects all records on the page. */ BEGIN curadj 33 ARG fileid int32_t ld ARG pgno db_pgno_t lu ARG indx u_int32_t lu ARG len u_int32_t lu ARG dup_off u_int32_t lu ARG add int ld ARG is_dup int ld ARG order u_int32_t lu END BEGIN chgpg 34 ARG fileid int32_t ld ARG mode db_ham_mode ld ARG old_pgno db_pgno_t lu ARG new_pgno db_pgno_t lu ARG old_indx u_int32_t lu ARG new_indx u_int32_t lu END