mirror of
https://github.com/HDFGroup/hdf5.git
synced 2024-11-27 02:10:55 +08:00
444 lines
15 KiB
C
444 lines
15 KiB
C
/* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *
|
||
* Copyright by The HDF Group. *
|
||
* Copyright by the Board of Trustees of the University of Illinois. *
|
||
* All rights reserved. *
|
||
* *
|
||
* This file is part of HDF5. The full HDF5 copyright notice, including *
|
||
* terms governing use, modification, and redistribution, is contained in *
|
||
* the files COPYING and Copyright.html. COPYING can be found at the root *
|
||
* of the source code distribution tree; Copyright.html can be found at the *
|
||
* root level of an installed copy of the electronic HDF5 document set and *
|
||
* is linked from the top-level documents page. It can also be found at *
|
||
* http://hdfgroup.org/HDF5/doc/Copyright.html. If you do not have *
|
||
* access to either file, you may request a copy from help@hdfgroup.org. *
|
||
* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
|
||
|
||
/*-------------------------------------------------------------------------
|
||
*
|
||
* Created: swmr_addrem_writer.c
|
||
*
|
||
* Purpose: Adds and removes data to a randomly selected subset of the
|
||
* datasets in the SWMR test file.
|
||
*
|
||
* This program is intended to run concurrently with the
|
||
* swmr_reader program. It is also run AFTER a sequential
|
||
* (not concurrent!) invoking of swmr_writer so the writer
|
||
* can dump a bunch of data into the datasets. Otherwise,
|
||
* there wouldn't be much to shrink :)
|
||
*
|
||
*-------------------------------------------------------------------------
|
||
*/
|
||
|
||
/***********/
|
||
/* Headers */
|
||
/***********/
|
||
|
||
#include "h5test.h"
|
||
#include "swmr_common.h"
|
||
|
||
/****************/
|
||
/* Local Macros */
|
||
/****************/
|
||
|
||
/* The maximum # of records to add/remove from the dataset in one step */
|
||
#define MAX_SIZE_CHANGE 10
|
||
|
||
/********************/
|
||
/* Local Prototypes */
|
||
/********************/
|
||
|
||
static hid_t open_skeleton(const char *filename, unsigned verbose);
|
||
static int addrem_records(hid_t fid, unsigned verbose, unsigned long nops,
|
||
unsigned long flush_count);
|
||
static void usage(void);
|
||
|
||
|
||
/*-------------------------------------------------------------------------
|
||
* Function: open_skeleton
|
||
*
|
||
* Purpose: Opens the SWMR HDF5 file and datasets.
|
||
*
|
||
* Parameters: const char *filename
|
||
* The filename of the SWMR HDF5 file to open
|
||
*
|
||
* unsigned verbose
|
||
* Whether or not to emit verbose console messages
|
||
*
|
||
* Return: Success: The file ID of the opened SWMR file
|
||
* The dataset IDs are stored in a global array
|
||
*
|
||
* Failure: -1
|
||
*
|
||
*-------------------------------------------------------------------------
|
||
*/
|
||
static hid_t
|
||
open_skeleton(const char *filename, unsigned verbose)
|
||
{
|
||
hid_t fid; /* File ID for new HDF5 file */
|
||
hid_t fapl; /* File access property list */
|
||
hid_t sid; /* Dataspace ID */
|
||
hsize_t dim[2]; /* Dataspace dimension */
|
||
unsigned u, v; /* Local index variable */
|
||
|
||
HDassert(filename);
|
||
|
||
/* Create file access property list */
|
||
if((fapl = h5_fileaccess()) < 0)
|
||
return -1;
|
||
|
||
/* Set to use the latest library format */
|
||
if(H5Pset_libver_bounds(fapl, H5F_LIBVER_LATEST, H5F_LIBVER_LATEST) < 0)
|
||
return -1;
|
||
|
||
#ifdef QAK
|
||
/* Increase the initial size of the metadata cache */
|
||
{
|
||
H5AC_cache_config_t mdc_config;
|
||
|
||
mdc_config.version = H5AC__CURR_CACHE_CONFIG_VERSION;
|
||
H5Pget_mdc_config(fapl, &mdc_config);
|
||
HDfprintf(stderr, "mdc_config.initial_size = %lu\n", (unsigned long)mdc_config.initial_size);
|
||
HDfprintf(stderr, "mdc_config.epoch_length = %lu\n", (unsigned long)mdc_config.epoch_length);
|
||
mdc_config.set_initial_size = 1;
|
||
mdc_config.initial_size = 16 * 1024 * 1024;
|
||
/* mdc_config.epoch_length = 5000; */
|
||
H5Pset_mdc_config(fapl, &mdc_config);
|
||
}
|
||
#endif /* QAK */
|
||
|
||
#ifdef QAK
|
||
H5Pset_fapl_log(fapl, "append.log", H5FD_LOG_ALL, (size_t)(512 * 1024 * 1024));
|
||
#endif /* QAK */
|
||
|
||
/* Open the file */
|
||
if((fid = H5Fopen(filename, H5F_ACC_RDWR | H5F_ACC_SWMR_WRITE, fapl)) < 0)
|
||
return -1;
|
||
|
||
/* Close file access property list */
|
||
if(H5Pclose(fapl) < 0)
|
||
return -1;
|
||
|
||
/* Emit informational message */
|
||
if(verbose)
|
||
HDfprintf(stderr, "Opening datasets\n");
|
||
|
||
/* Open the datasets */
|
||
for(u = 0; u < NLEVELS; u++)
|
||
for(v = 0; v < symbol_count[u]; v++) {
|
||
if((symbol_info[u][v].dsid = H5Dopen2(fid, symbol_info[u][v].name, H5P_DEFAULT)) < 0)
|
||
return -1;
|
||
if((sid = H5Dget_space(symbol_info[u][v].dsid)) < 0)
|
||
return -1;
|
||
if(2 != H5Sget_simple_extent_ndims(sid))
|
||
return -1;
|
||
if(H5Sget_simple_extent_dims(sid, dim, NULL) < 0)
|
||
return -1;
|
||
symbol_info[u][v].nrecords = dim[1];
|
||
} /* end for */
|
||
|
||
return fid;
|
||
}
|
||
|
||
|
||
/*-------------------------------------------------------------------------
|
||
* Function: addrem_records
|
||
*
|
||
* Purpose: Adds/removes a specified number of records to random datasets
|
||
* to the SWMR test file.
|
||
*
|
||
* Parameters: hid_t fid
|
||
* The file ID of the SWMR HDF5 file
|
||
*
|
||
* unsigned verbose
|
||
* Whether or not to emit verbose console messages
|
||
*
|
||
* unsigned long nops
|
||
* # of records to read/write in the datasets
|
||
*
|
||
* unsigned long flush_count
|
||
* # of records to write before flushing the file to disk
|
||
*
|
||
* Return: Success: 0
|
||
* Failure: -1
|
||
*
|
||
*-------------------------------------------------------------------------
|
||
*/
|
||
static int
|
||
addrem_records(hid_t fid, unsigned verbose, unsigned long nops, unsigned long flush_count)
|
||
{
|
||
hid_t tid; /* Datatype ID for records */
|
||
hid_t mem_sid; /* Memory dataspace ID */
|
||
hsize_t start[2] = {0, 0}, count[2] = {1, 1}; /* Hyperslab selection values */
|
||
hsize_t dim[2] = {1, 0}; /* Dataspace dimensions */
|
||
symbol_t buf[MAX_SIZE_CHANGE]; /* Write buffer */
|
||
unsigned long op_to_flush; /* # of operations before flush */
|
||
unsigned long u, v; /* Local index variables */
|
||
|
||
HDassert(fid > 0);
|
||
|
||
/* Reset the buffer */
|
||
HDmemset(&buf, 0, sizeof(buf));
|
||
|
||
/* Create a dataspace for the record to add */
|
||
if((mem_sid = H5Screate_simple(2, count, NULL)) < 0)
|
||
return -1;
|
||
|
||
/* Create datatype for appending records */
|
||
if((tid = create_symbol_datatype()) < 0)
|
||
return -1;
|
||
|
||
/* Add and remove records to random datasets, according to frequency
|
||
* distribution */
|
||
op_to_flush = flush_count;
|
||
for(u=0; u<nops; u++) {
|
||
symbol_info_t *symbol; /* Symbol to write record to */
|
||
hid_t file_sid; /* Dataset's space ID */
|
||
|
||
/* Get a random dataset, according to the symbol distribution */
|
||
symbol = choose_dataset();
|
||
|
||
/* Decide whether to shrink or expand, and by how much */
|
||
count[1] = (hsize_t)HDrandom() % (MAX_SIZE_CHANGE * 2) + 1;
|
||
|
||
if(count[1] > MAX_SIZE_CHANGE) {
|
||
/* Add records */
|
||
count[1] -= MAX_SIZE_CHANGE;
|
||
|
||
/* Set the buffer's IDs (equal to its position) */
|
||
for(v=0; v<count[1]; v++)
|
||
buf[v].rec_id = (uint64_t)symbol->nrecords + (uint64_t)v;
|
||
|
||
/* Set the memory space to the correct size */
|
||
if(H5Sset_extent_simple(mem_sid, 2, count, NULL) < 0)
|
||
return -1;
|
||
|
||
/* Get the coordinates to write */
|
||
start[1] = symbol->nrecords;
|
||
|
||
/* Cork the metadata cache, to prevent the object header from being
|
||
* flushed before the data has been written */
|
||
if(H5Odisable_mdc_flushes(symbol->dsid) < 0)
|
||
return -1;
|
||
|
||
/* Extend the dataset's dataspace to hold the new record */
|
||
symbol->nrecords+= count[1];
|
||
dim[1] = symbol->nrecords;
|
||
if(H5Dset_extent(symbol->dsid, dim) < 0)
|
||
return -1;
|
||
|
||
/* Get the dataset's dataspace */
|
||
if((file_sid = H5Dget_space(symbol->dsid)) < 0)
|
||
return -1;
|
||
|
||
/* Choose the last record in the dataset */
|
||
if(H5Sselect_hyperslab(file_sid, H5S_SELECT_SET, start, NULL, count, NULL) < 0)
|
||
return -1;
|
||
|
||
/* Write record to the dataset */
|
||
if(H5Dwrite(symbol->dsid, tid, mem_sid, file_sid, H5P_DEFAULT, &buf) < 0)
|
||
return -1;
|
||
|
||
/* Uncork the metadata cache */
|
||
if(H5Oenable_mdc_flushes(symbol->dsid) < 0)
|
||
return -1;
|
||
|
||
/* Close the dataset's dataspace */
|
||
if(H5Sclose(file_sid) < 0)
|
||
return -1;
|
||
} /* end if */
|
||
else {
|
||
/* Shrink the dataset's dataspace */
|
||
if(count[1] > symbol->nrecords)
|
||
symbol->nrecords = 0;
|
||
else
|
||
symbol->nrecords -= count[1];
|
||
dim[1] = symbol->nrecords;
|
||
if(H5Dset_extent(symbol->dsid, dim) < 0)
|
||
return -1;
|
||
} /* end else */
|
||
|
||
/* Check for flushing file */
|
||
if(flush_count > 0) {
|
||
/* Decrement count of records to write before flushing */
|
||
op_to_flush--;
|
||
|
||
/* Check for counter being reached */
|
||
if(0 == op_to_flush) {
|
||
/* Flush contents of file */
|
||
if(H5Fflush(fid, H5F_SCOPE_GLOBAL) < 0)
|
||
return -1;
|
||
|
||
/* Reset flush counter */
|
||
op_to_flush = flush_count;
|
||
} /* end if */
|
||
} /* end if */
|
||
} /* end for */
|
||
|
||
/* Emit informational message */
|
||
if(verbose)
|
||
HDfprintf(stderr, "Closing datasets\n");
|
||
|
||
/* Close the datasets */
|
||
for(u = 0; u < NLEVELS; u++)
|
||
for(v = 0; v < symbol_count[u]; v++)
|
||
if(H5Dclose(symbol_info[u][v].dsid) < 0)
|
||
return -1;
|
||
|
||
return 0;
|
||
}
|
||
|
||
static void
|
||
usage(void)
|
||
{
|
||
printf("\n");
|
||
printf("Usage error!\n");
|
||
printf("\n");
|
||
printf("Usage: swmr_addrem_writer [-q] [-f <# of operations between flushing\n");
|
||
printf(" file contents>] [-r <random seed>] <# of operations>\n");
|
||
printf("\n");
|
||
printf("<# of operations between flushing file contents> should be 0 (for\n");
|
||
printf("no flushing) or between 1 and (<# of operations> - 1).\n");
|
||
printf("\n");
|
||
printf("<# of operations> must be specified.\n");
|
||
printf("\n");
|
||
printf("Defaults to verbose (no '-q' given), flushing every 1000 operations\n");
|
||
printf("('-f 1000'), and will generate a random seed (no -r given).\n");
|
||
printf("\n");
|
||
HDexit(1);
|
||
}
|
||
|
||
int main(int argc, const char *argv[])
|
||
{
|
||
hid_t fid; /* File ID for file opened */
|
||
long nops = 0; /* # of times to grow or shrink the dataset */
|
||
long flush_count = 1000; /* # of records to write between flushing file */
|
||
unsigned verbose = 1; /* Whether to emit some informational messages */
|
||
unsigned use_seed = 0; /* Set to 1 if a seed was set on the command line */
|
||
unsigned random_seed = 0; /* Random # seed */
|
||
unsigned u; /* Local index variable */
|
||
int temp;
|
||
|
||
/* Parse command line options */
|
||
if(argc < 2)
|
||
usage();
|
||
if(argc > 1) {
|
||
u = 1;
|
||
while(u < (unsigned)argc) {
|
||
if(argv[u][0] == '-') {
|
||
switch(argv[u][1]) {
|
||
/* # of records to write between flushing file */
|
||
case 'f':
|
||
flush_count = HDatol(argv[u + 1]);
|
||
if(flush_count < 0)
|
||
usage();
|
||
u += 2;
|
||
break;
|
||
|
||
/* Be quiet */
|
||
case 'q':
|
||
verbose = 0;
|
||
u++;
|
||
break;
|
||
|
||
/* Random # seed */
|
||
case 'r':
|
||
use_seed = 1;
|
||
temp = HDatoi(argv[u + 1]);
|
||
if(temp < 0)
|
||
usage();
|
||
else
|
||
random_seed = (unsigned)temp;
|
||
u += 2;
|
||
break;
|
||
|
||
default:
|
||
usage();
|
||
break;
|
||
} /* end switch */
|
||
} /* end if */
|
||
else {
|
||
/* Get the number of records to append */
|
||
nops = HDatol(argv[u]);
|
||
if(nops <= 0)
|
||
usage();
|
||
|
||
u++;
|
||
} /* end else */
|
||
} /* end while */
|
||
} /* end if */
|
||
if(nops <= 0)
|
||
usage();
|
||
if(flush_count >= nops)
|
||
usage();
|
||
|
||
/* Emit informational message */
|
||
if(verbose) {
|
||
HDfprintf(stderr, "Parameters:\n");
|
||
HDfprintf(stderr, "\t# of operations between flushes = %ld\n", flush_count);
|
||
HDfprintf(stderr, "\t# of operations = %ld\n", nops);
|
||
} /* end if */
|
||
|
||
/* Set the random seed */
|
||
if(0 == use_seed) {
|
||
struct timeval t;
|
||
HDgettimeofday(&t, NULL);
|
||
random_seed = (unsigned)(t.tv_usec);
|
||
} /* end if */
|
||
HDsrandom(random_seed);
|
||
/* ALWAYS emit the random seed for possible debugging */
|
||
HDfprintf(stderr, "Using writer random seed: %u\n", random_seed);
|
||
|
||
/* Emit informational message */
|
||
if(verbose)
|
||
HDfprintf(stderr, "Generating symbol names\n");
|
||
|
||
/* Generate dataset names */
|
||
if(generate_symbols() < 0)
|
||
return -1;
|
||
|
||
/* Emit informational message */
|
||
if(verbose)
|
||
HDfprintf(stderr, "Opening skeleton file: %s\n", FILENAME);
|
||
|
||
/* Open file skeleton */
|
||
if((fid = open_skeleton(FILENAME, verbose)) < 0) {
|
||
HDfprintf(stderr, "Error opening skeleton file!\n");
|
||
HDexit(1);
|
||
} /* end if */
|
||
|
||
/* Send a message to indicate "H5Fopen" is complete--releasing the file lock */
|
||
h5_send_message(WRITER_MESSAGE, NULL, NULL);
|
||
|
||
/* Emit informational message */
|
||
if(verbose)
|
||
HDfprintf(stderr, "Adding and removing records\n");
|
||
|
||
/* Grow and shrink datasets */
|
||
if(addrem_records(fid, verbose, (unsigned long)nops, (unsigned long)flush_count) < 0) {
|
||
HDfprintf(stderr, "Error adding and removing records from datasets!\n");
|
||
HDexit(1);
|
||
} /* end if */
|
||
|
||
/* Emit informational message */
|
||
if(verbose)
|
||
HDfprintf(stderr, "Releasing symbols\n");
|
||
|
||
/* Clean up the symbols */
|
||
if(shutdown_symbols() < 0) {
|
||
HDfprintf(stderr, "Error releasing symbols!\n");
|
||
HDexit(1);
|
||
} /* end if */
|
||
|
||
/* Emit informational message */
|
||
if(verbose)
|
||
HDfprintf(stderr, "Closing objects\n");
|
||
|
||
/* Close objects opened */
|
||
if(H5Fclose(fid) < 0) {
|
||
HDfprintf(stderr, "Error closing file!\n");
|
||
HDexit(1);
|
||
} /* end if */
|
||
|
||
return 0;
|
||
}
|