[svn-r10953] Purpose: Updating C++ tests

Description:
    Updated various comments/headers.

Platforms tested:
    Linux 2.4 (heping)
    AIX 5.1 (copper)
This commit is contained in:
Binh-Minh Ribler 2005-06-19 16:02:21 -05:00
parent d641030436
commit 828c6646a1
3 changed files with 224 additions and 241 deletions

View File

@ -15,14 +15,12 @@
/*****************************************************************************
FILE
dsets.cpp - HDF5 C++ testing the functionalities associated with the
C dataset interface (H5D)
C dataset interface (H5D)
EXTERNAL ROUTINES/VARIABLES:
These routines are in the test directory of the C library:
h5_reset() -- in h5test.c, resets the library by closing it
h5_fileaccess() -- in h5test.c, returns a file access template
h5_fixname() -- in h5test.c, create a file name from a file base name
h5_cleanup() -- in h5test.c, cleanup temporary test files
h5_reset() -- in h5test.c, resets the library by closing it
h5_fileaccess() -- in h5test.c, returns a file access template
***************************************************************************/
@ -76,8 +74,9 @@ void cleanup_dsets(void);
static herr_t
test_create( H5File& file)
{
TESTING("create, open, close");
TESTING("create, open, close");
// Setting this to NULL for cleaning up in failure situations
DataSet *dataset = NULL;
try {
// Create a data space
@ -88,7 +87,7 @@ test_create( H5File& file)
// Create a dataset using the default dataset creation properties.
// We're not sure what they are, so we won't check.
dataset = new DataSet (file.createDataSet
dataset = new DataSet (file.createDataSet
(DSET_DEFAULT_NAME, PredType::NATIVE_DOUBLE, space));
// Add a comment to the dataset
@ -96,6 +95,7 @@ test_create( H5File& file)
// Close the dataset
delete dataset;
dataset = NULL;
// Try creating a dataset that already exists. This should fail since a
// dataset can only be created once. If an exception is not thrown for
@ -106,9 +106,9 @@ test_create( H5File& file)
// continuation here, that means no exception has been thrown
throw InvalidActionException("H5File::createDataSet", "Library allowed overwrite of existing dataset");
}
catch (FileIException E) // catching invalid creating dataset
{} // do nothing, exception expected
}
catch (FileIException E) // catching invalid creating dataset
{} // do nothing, exception expected
// Open the dataset we created above and then close it. This is one
// way to open an existing dataset for accessing.
@ -119,9 +119,9 @@ test_create( H5File& file)
// This is another way to open an existing dataset for accessing.
DataSet another_dataset(file.openDataSet (DSET_DEFAULT_NAME));
// Try opening a non-existent dataset. This should fail so if an
// exception is not thrown for this action by openDataSet, then
// exception is not thrown for this action by openDataSet, then
// display failure information and throw an exception.
try {
dataset = new DataSet (file.openDataSet( "does_not_exist" ));
@ -130,12 +130,10 @@ test_create( H5File& file)
throw InvalidActionException("H5File::openDataSet", "Attempted to open a non-existent dataset");
}
catch (FileIException E ) // catching creating non-existent dataset
{} // do nothing, exception expected
{} // do nothing, exception expected
/*
* Create a new dataset that uses chunked storage instead of the default
* layout.
*/
// Create a new dataset that uses chunked storage instead of the default
// layout.
DSetCreatPropList create_parms;
hsize_t csize[2];
csize[0] = 5;
@ -165,25 +163,26 @@ test_create( H5File& file)
}
// catch all other exceptions
catch (Exception E)
{
cerr << " FAILED" << endl;
cerr << " <<< " << E.getDetailMsg() << " >>>" << endl << endl;
{
issue_fail_msg(E.getCFuncName(), __LINE__, __FILE__);
// clean up and return with failure
if (dataset != NULL)
delete dataset;
return -1;
}
}
} // test_create
/*-------------------------------------------------------------------------
* Function: check_values
*
* Purpose: Checks a read value against the written value. If they are
* different, the function will
* print out a message and the different values. This function
* is made to reuse the code segment that is used in various
* places throughout test_compression and in test_simple_io.
* different, the function will print out a message and the
* different values. This function is made to reuse the code
* segment that is used in various places throughout
* test_compression and in test_simple_io. Where the C version
* of this code segment "goto error," this function will
* return -1.
*
* Return: Success: 0
*
@ -237,7 +236,7 @@ test_simple_io( H5File& file)
int check[100][200];
int i, j, n;
/* Initialize the dataset */
// Initialize the dataset
for (i = n = 0; i < 100; i++)
{
for (j = 0; j < 200; j++) {
@ -248,27 +247,27 @@ test_simple_io( H5File& file)
char* tconv_buf = new char [1000];
try
{
/* Create the data space */
// Create the data space
hsize_t dims[2];
dims[0] = 100;
dims[1] = 200;
DataSpace space (2, dims, NULL);
/* Create a small conversion buffer to test strip mining */
// Create a small conversion buffer to test strip mining
DSetMemXferPropList xfer;
xfer.setBuffer (1000, tconv_buf, NULL);
/* Create the dataset */
// Create the dataset
DataSet dataset (file.createDataSet (DSET_SIMPLE_IO_NAME, PredType::NATIVE_INT, space));
/* Write the data to the dataset */
// Write the data to the dataset
dataset.write ((void*) points, PredType::NATIVE_INT, DataSpace::ALL, DataSpace::ALL, xfer);
/* Read the dataset back */
// Read the dataset back
dataset.read ((void*) check, PredType::NATIVE_INT, DataSpace::ALL, DataSpace::ALL, xfer);
/* Check that the values read are the same as the values written */
// Check that the values read are the same as the values written
for (i = 0; i < 100; i++)
for (j = 0; j < 200; j++)
{
@ -282,9 +281,10 @@ test_simple_io( H5File& file)
PASSED();
return 0;
} // end try
// catch all dataset, space, plist exceptions
catch (Exception E)
{
{
cerr << " FAILED" << endl;
cerr << " <<< " << E.getDetailMsg() << " >>>" << endl << endl;
@ -293,7 +293,7 @@ test_simple_io( H5File& file)
delete [] tconv_buf;
return -1;
}
}
} // test_simple_io
/*-------------------------------------------------------------------------
* Function: test_tconv
@ -323,7 +323,7 @@ test_tconv( H5File& file)
TESTING("data type conversion");
/* Initialize the dataset */
// Initialize the dataset
for (int i = 0; i < 1000000; i++) {
out[i*4+0] = 0x11;
out[i*4+1] = 0x22;
@ -333,21 +333,21 @@ test_tconv( H5File& file)
try
{
/* Create the data space */
// Create the data space
hsize_t dims[1];
dims[0] = 1000000;
DataSpace space (1, dims, NULL);
/* Create the data set */
// Create the data set
DataSet dataset (file.createDataSet (DSET_TCONV_NAME, PredType::STD_I32LE, space));
/* Write the data to the dataset */
// Write the data to the dataset
dataset.write ((void*) out, PredType::STD_I32LE);
/* Read data with byte order conversion */
// Read data with byte order conversion
dataset.read ((void*) in, PredType::STD_I32BE);
/* Check */
// Check
for (int i = 0; i < 1000000; i++) {
if (in[4*i+0]!=out[4*i+3] ||
in[4*i+1]!=out[4*i+2] ||
@ -367,7 +367,7 @@ test_tconv( H5File& file)
// catch all dataset and space exceptions
catch (Exception E)
{
{
cerr << " FAILED" << endl;
cerr << " <<< " << E.getDetailMsg() << " >>>" << endl << endl;
@ -376,7 +376,7 @@ test_tconv( H5File& file)
delete [] in;
return -1;
}
}
} // test_tconv
/* This message derives from H5Z */
const H5Z_class_t H5Z_BOGUS[1] = {{
@ -399,7 +399,7 @@ const H5Z_class_t H5Z_BOGUS[1] = {{
* Failure: 0
*
* Programmer: Robb Matzke
* Tuesday, April 21, 1998
* Tuesday, April 21, 1998
*
* Modifications:
*
@ -437,7 +437,6 @@ bogus(unsigned int flags, size_t cd_nelmts,
*
*-------------------------------------------------------------------------
*/
static herr_t
test_compression(H5File& file)
{
@ -447,7 +446,7 @@ test_compression(H5File& file)
int check[100][200];
hsize_t i, j, n;
/* Initialize the dataset */
// Initialize the dataset
for (i = n = 0; i < 100; i++)
{
for (j = 0; j < 200; j++) {
@ -456,22 +455,18 @@ test_compression(H5File& file)
}
char* tconv_buf = new char [1000];
DataSet* dataset = NULL;
try
{
const hsize_t size[2] = {100, 200};
/* Create the data space */
// Create the data space
DataSpace space1(2, size, NULL);
/*
* Create a small conversion buffer to test strip mining. We
* might as well test all we can!
*/
// Create a small conversion buffer to test strip mining
DSetMemXferPropList xfer;
xfer.setBuffer (1000, tconv_buf, NULL);
/* Use chunked storage with compression */
// Use chunked storage with compression
DSetCreatPropList dscreatplist;
const hsize_t chunk_size[2] = {2, 25};
@ -479,9 +474,9 @@ test_compression(H5File& file)
dscreatplist.setDeflate (6);
#ifdef H5_HAVE_FILTER_DEFLATE
TESTING("compression (setup)");
TESTING("compression (setup)");
/* Create the dataset */
// Create the dataset
dataset = new DataSet (file.createDataSet
(DSET_COMPRESS_NAME, PredType::NATIVE_INT, space1, dscreatplist));
@ -491,7 +486,7 @@ test_compression(H5File& file)
* STEP 1: Read uninitialized data. It should be zero.
*----------------------------------------------------------------------
*/
TESTING("compression (uninitialized read)");
TESTING("compression (uninitialized read)");
dataset->read ((void*) check, PredType::NATIVE_INT, DataSpace::ALL, DataSpace::ALL, xfer);
@ -506,14 +501,14 @@ test_compression(H5File& file)
}
}
}
PASSED();
PASSED();
/*----------------------------------------------------------------------
* STEP 2: Test compression by setting up a chunked dataset and writing
* to it.
*----------------------------------------------------------------------
*/
TESTING("compression (write)");
TESTING("compression (write)");
for (i=n=0; i<size[0]; i++)
{
@ -525,18 +520,18 @@ test_compression(H5File& file)
dataset->write ((void*) points, PredType::NATIVE_INT, DataSpace::ALL, DataSpace::ALL, xfer);
PASSED();
PASSED();
/*----------------------------------------------------------------------
* STEP 3: Try to read the data we just wrote.
*----------------------------------------------------------------------
*/
TESTING("compression (read)");
TESTING("compression (read)");
/* Read the dataset back */
// Read the dataset back
dataset->read ((void*)check, PredType::NATIVE_INT, DataSpace::ALL, DataSpace::ALL, xfer);
/* Check that the values read are the same as the values written */
// Check that the values read are the same as the values written
for (i = 0; i < size[0]; i++)
for (j = 0; j < size[1]; j++)
{
@ -545,7 +540,7 @@ test_compression(H5File& file)
throw Exception("test_compression", "Failed in read");
}
PASSED();
PASSED();
/*----------------------------------------------------------------------
* STEP 4: Write new data over the top of the old data. The new data is
@ -554,7 +549,7 @@ test_compression(H5File& file)
* dataset although we rewrite the whole thing.
*----------------------------------------------------------------------
*/
TESTING("compression (modify)");
TESTING("compression (modify)");
for (i=0; i<size[0]; i++)
{
@ -565,10 +560,10 @@ test_compression(H5File& file)
}
dataset->write ((void*)points, PredType::NATIVE_INT, DataSpace::ALL, DataSpace::ALL, xfer);
/* Read the dataset back and check it */
// Read the dataset back and check it
dataset->read ((void*)check, PredType::NATIVE_INT, DataSpace::ALL, DataSpace::ALL, xfer);
/* Check that the values read are the same as the values written */
// Check that the values read are the same as the values written
for (i = 0; i < size[0]; i++)
for (j = 0; j < size[1]; j++)
{
@ -577,7 +572,7 @@ test_compression(H5File& file)
throw Exception("test_compression", "Failed in modify");
}
PASSED();
PASSED();
/*----------------------------------------------------------------------
* STEP 5: Close the dataset and then open it and read it again. This
@ -585,14 +580,15 @@ test_compression(H5File& file)
* object header.
*----------------------------------------------------------------------
*/
TESTING("compression (re-open)");
// close this dataset
TESTING("compression (re-open)");
// close this dataset to reuse the var
delete dataset;
dataset = new DataSet (file.openDataSet (DSET_COMPRESS_NAME));
dataset->read ((void*)check, PredType::NATIVE_INT, DataSpace::ALL, DataSpace::ALL, xfer);
/* Check that the values read are the same as the values written */
// Check that the values read are the same as the values written
for (i = 0; i < size[0]; i++)
for (j = 0; j < size[1]; j++)
{
@ -601,7 +597,7 @@ test_compression(H5File& file)
throw Exception("test_compression", "Failed in re-open");
}
PASSED();
PASSED();
/*----------------------------------------------------------------------
@ -610,7 +606,7 @@ test_compression(H5File& file)
* boundaries (we know that case already works from above tests).
*----------------------------------------------------------------------
*/
TESTING("compression (partial I/O)");
TESTING("compression (partial I/O)");
const hsize_t hs_size[2] = {4, 50};
const hsize_t hs_offset[2] = {7, 30};
@ -623,7 +619,7 @@ test_compression(H5File& file)
dataset->write ((void*)points, PredType::NATIVE_INT, space1, space1, xfer);
dataset->read ((void*)check, PredType::NATIVE_INT, space1, space1, xfer);
/* Check that the values read are the same as the values written */
// Check that the values read are the same as the values written
for (i=0; i<hs_size[0]; i++) {
for (j=0; j<hs_size[1]; j++) {
if (points[hs_offset[0]+i][hs_offset[1]+j] !=
@ -643,12 +639,12 @@ test_compression(H5File& file)
delete dataset;
dataset = NULL;
PASSED();
PASSED();
#else
TESTING("deflate filter");
SKIPPED();
cerr << not_supported << endl;
TESTING("deflate filter");
SKIPPED();
cerr << not_supported << endl;
#endif
/*----------------------------------------------------------------------
@ -656,12 +652,12 @@ test_compression(H5File& file)
* to write and then read the dataset.
*----------------------------------------------------------------------
*/
TESTING("compression (app-defined method)");
TESTING("compression (app-defined method)");
if (H5Zregister (H5Z_BOGUS)<0)
throw Exception("test_compression", "Failed in app-defined method");
if (H5Pset_filter (dscreatplist.getId(), H5Z_FILTER_BOGUS, 0, 0, NULL)<0)
throw Exception("test_compression", "Failed in app-defined method");
throw Exception("test_compression", "Failed in app-defined method");
dscreatplist.setFilter (H5Z_FILTER_BOGUS, 0, 0, NULL);
DataSpace space2 (2, size, NULL);
@ -670,7 +666,7 @@ test_compression(H5File& file)
dataset->write ((void*)points, PredType::NATIVE_INT, DataSpace::ALL, DataSpace::ALL, xfer);
dataset->read ((void*)check, PredType::NATIVE_INT, DataSpace::ALL, DataSpace::ALL, xfer);
/* Check that the values read are the same as the values written */
// Check that the values read are the same as the values written
for (i = 0; i < size[0]; i++)
for (j = 0; j < size[1]; j++)
{
@ -691,8 +687,8 @@ test_compression(H5File& file)
} // end try
// catch all dataset, file, space, and plist exceptions
catch (Exception E)
{
catch (Exception E)
{
cerr << " FAILED" << endl;
cerr << " <<< " << E.getDetailMsg() << " >>>" << endl << endl;
@ -703,7 +699,7 @@ test_compression(H5File& file)
delete [] tconv_buf;
return -1;
}
}
} // test_compression
/*-------------------------------------------------------------------------
* Function: test_multiopen
@ -758,7 +754,7 @@ test_multiopen (H5File& file)
cur_size[0] = 20;
dset1.extend (cur_size);
/* Get the size from the second handle */
// Get the size from the second handle
space = new DataSpace (dset2.getSpace());
hsize_t tmp_size[1];
@ -777,16 +773,17 @@ test_multiopen (H5File& file)
} // end try block
// catch all dataset, file, space, and plist exceptions
catch (Exception E)
catch (Exception E)
{
cerr << " FAILED" << endl;
cerr << " <<< " << E.getDetailMsg() << " >>>" << endl << endl;
// clean up and return with failure
if (space != NULL)
delete space;
return -1;
}
}
} // test_multiopen
/*-------------------------------------------------------------------------
@ -799,7 +796,7 @@ test_multiopen (H5File& file)
* Failure: -1
*
* Programmer: Binh-Minh Ribler (using C version)
* February 17, 2001
* February 17, 2001
*
* Modifications:
*
@ -852,8 +849,7 @@ test_types(H5File& file)
for (i=0; i<sizeof buf; i++)
buf[i] = (unsigned char)0xff ^ (unsigned char)i;
// Write data from buf using all default dataspaces and property
// list
// Write data from buf using all default dataspaces and property list
dset->write (buf, type);
// no failure in bitfield_1, close this dataset
@ -862,9 +858,9 @@ test_types(H5File& file)
// catch exceptions thrown in try block of bitfield_1
catch (Exception E)
{
{
cerr << " FAILED" << endl;
cerr << " <<< " << "bitfield_1: " << E.getFuncName()
cerr << " <<< " << "bitfield_1: " << E.getFuncName()
<< " - " << E.getDetailMsg() << " >>>" << endl << endl;
if (dset != NULL)
delete dset;
@ -882,7 +878,7 @@ test_types(H5File& file)
for (i=0; i<sizeof(buf); i++)
buf[i] = (unsigned char)0xff ^ (unsigned char)i;
// Write data from buf using all default dataspaces and property
// Write data from buf using all default dataspaces and property
// list; if writing fails, deallocate dset and return.
dset->write (buf, type);
@ -895,14 +891,14 @@ test_types(H5File& file)
// catch exceptions thrown in try block of bitfield_2
catch (Exception E) {
cerr << " FAILED" << endl;
cerr << " <<< " << "bitfield_2: " << E.getFuncName()
cerr << " <<< " << "bitfield_2: " << E.getFuncName()
<< " - " << E.getDetailMsg() << " >>>" << endl << endl;
if (dset != NULL)
delete dset;
delete dset;
throw E; // propagate the exception
}
/* opaque_1 */
/* opaque_1 */
DataType* optype = NULL;
try { // opaque_1 block
optype = new DataType(H5T_OPAQUE, 1);
@ -915,7 +911,7 @@ test_types(H5File& file)
for (i=0; i<sizeof buf; i++)
buf[i] = (unsigned char)0xff ^ (unsigned char)i;
// Write data from buf using all default dataspaces and property
// Write data from buf using all default dataspaces and property
// list; if writing fails, deallocate dset and return.
dset->write (buf, *optype);
@ -925,14 +921,14 @@ test_types(H5File& file)
} // end try block of opaque_1
// catch exceptions thrown in try block of opaque_1
catch (Exception E) {
catch (Exception E) {
cerr << " FAILED" << endl;
cerr << " <<< " << "opaque_1: " << E.getFuncName()
cerr << " <<< " << "opaque_1: " << E.getFuncName()
<< " - " << E.getDetailMsg() << " >>>" << endl << endl;
if (dset != NULL)
delete dset;
delete dset;
if (optype != NULL)
delete optype;
delete optype;
throw E; // propagate the exception
}
@ -946,9 +942,9 @@ test_types(H5File& file)
// Fill buffer
for (i=0; i<sizeof(buf); i++)
buf[i] = (unsigned char)0xff ^ (unsigned char)i;
buf[i] = (unsigned char)0xff ^ (unsigned char)i;
// Write data from buf using all default dataspaces and property
// Write data from buf using all default dataspaces and property
// list; if writing fails, deallocate dset and return.
dset->write (buf, *optype);
@ -957,26 +953,27 @@ test_types(H5File& file)
delete optype; optype = NULL;
} //end try block of opaque_2
// catch exceptions thrown in try block of opaque_2
catch (Exception E) {
cerr << " FAILED" << endl;
cerr << " <<< " << "opaque_2: " << E.getFuncName()
cerr << " <<< " << "opaque_2: " << E.getFuncName()
<< " - " << E.getDetailMsg() << " >>>" << endl << endl;
if (dset != NULL)
delete dset;
delete dset;
if (optype != NULL)
delete optype;
delete optype;
throw E; // propagate the exception
}
PASSED();
return 0;
PASSED();
return 0;
} // end top try block
catch (Exception E)
catch (Exception E)
{
return -1;
return -1;
}
}
} // test_types
/*-------------------------------------------------------------------------
* Function: main
@ -995,6 +992,8 @@ test_types(H5File& file)
* - moved h5_cleanup to outside of try block because
* dataset.h5 cannot be removed until "file" is out of
* scope and dataset.h5 is closed.
* Feb 20, 05:
* - cleanup_dsets took care of the cleanup now.
*
*-------------------------------------------------------------------------
*/
@ -1030,11 +1029,6 @@ main(void)
nerrors += test_compression(file)<0 ?1:0;
nerrors += test_multiopen (file)<0 ?1:0;
nerrors += test_types(file)<0 ?1:0;
// increment the ref count of this property list so that the
// property list id won't be closed when fapl goes out of scope.
// This is a bad hack, but I want to use existing routine h5_cleanup!
fapl.incRefCount();
}
catch (Exception E)
{
@ -1047,7 +1041,7 @@ main(void)
// Print out dsets test results
cerr << endl << endl;
return(test_report(nerrors, string(" Dataset")));
}
} // main
/*-------------------------------------------------------------------------
* Function: cleanup_dsets
@ -1066,5 +1060,5 @@ void
cleanup_dsets(void)
{
remove(FILE1.c_str());
} /* cleanup_dsets */
} // cleanup_dsets

View File

@ -79,7 +79,7 @@ const string FILE4("tfile4.h5");
* cases. Since there are no operator<< for 'long long'
* or int64 in VS C++ ostream, I casted the hsize_t values
* passed to verify_val to 'long' as well. If problems
* arises later, this will have to be specificly handled
* arises later, this will have to be specificly handled
* with a special routine.
*
*-------------------------------------------------------------------------
@ -87,14 +87,14 @@ const string FILE4("tfile4.h5");
static void
test_file_create(void)
{
/* Output message about test being performed */
// Output message about test being performed
MESSAGE(5, ("Testing File Creation I/O\n"));
/* Test create with various sequences of H5F_ACC_EXCL and */
/* H5F_ACC_TRUNC flags */
// Test create with various sequences of H5F_ACC_EXCL and
// H5F_ACC_TRUNC flags
/* Create with H5F_ACC_EXCL */
/* First ensure the file does not exist */
// Create with H5F_ACC_EXCL
// First ensure the file does not exist
remove(FILE1.c_str());
// Setting this to NULL for cleaning up in failure situations
@ -109,21 +109,22 @@ test_file_create(void)
H5File file2 (FILE1, H5F_ACC_TRUNC); // should throw E
// Should FAIL but didn't, so throw an invalid action exception
throw InvalidActionException("H5File constructor", "Attempted to create an existing file.");
throw InvalidActionException("H5File constructor", "Attempted to create an existing file.");
}
catch( FileIException E ) {} // do nothing, FAIL expected
catch( FileIException E ) // catch truncating existing file
{} // do nothing, FAIL expected
// Close file1
delete file1;
file1 = NULL;
// Try again with H5F_ACC_EXCL. This should fail because the file already
// exists from the previous steps.
// Try again with H5F_ACC_EXCL. This should fail because the file
// already exists from the previous steps.
try {
H5File file2(FILE1, H5F_ACC_EXCL); // should throw E
// Should FAIL but didn't, so throw an invalid action exception
throw InvalidActionException("H5File constructor", "File already exists.");
throw InvalidActionException("H5File constructor", "File already exists.");
}
catch( FileIException E ) // catching creating existing file
{} // do nothing, FAIL expected
@ -131,15 +132,13 @@ test_file_create(void)
// Test create with H5F_ACC_TRUNC. This will truncate the existing file.
file1 = new H5File (FILE1, H5F_ACC_TRUNC);
/*
* Try to truncate first file again. This should fail because file1 is the
* same file and is currently open.
*/
// Try to truncate first file again. This should fail because file1
// is the same file and is currently open.
try {
H5File file2 (FILE1, H5F_ACC_TRUNC); // should throw E
// Should FAIL but didn't, so throw an invalid action exception
throw InvalidActionException("H5File constructor", "H5F_ACC_TRUNC attempt on an opened file.");
throw InvalidActionException("H5File constructor", "H5F_ACC_TRUNC attempt on an opened file.");
}
catch( FileIException E ) // catching truncating opened file
{} // do nothing, FAIL expected
@ -150,7 +149,7 @@ test_file_create(void)
H5File file3 (FILE1, H5F_ACC_EXCL); // should throw E
// Should FAIL but didn't, so throw an invalid action exception
throw InvalidActionException("H5File constructor", "H5F_ACC_EXCL attempt on an existing file.");
throw InvalidActionException("H5File constructor", "H5F_ACC_EXCL attempt on an existing file.");
}
catch( FileIException E ) // catching H5F_ACC_EXCL on existing file
{} // do nothing, FAIL expected
@ -196,10 +195,10 @@ test_file_create(void)
FileCreatPropList* tmpl1 = NULL;
try
{
/* Create a new file with a non-standard file-creation template */
// Create a new file with a non-standard file-creation template
tmpl1 = new FileCreatPropList;
/* Set the new file-creation parameters */
// Set the new file-creation parameters
tmpl1->setUserblock (F2_USERBLOCK_SIZE);
tmpl1->setSizes( F2_OFFSET_SIZE, F2_LENGTH_SIZE );
tmpl1->setSymk( F2_SYM_INTERN_K, F2_SYM_LEAF_K );
@ -208,48 +207,46 @@ test_file_create(void)
// params.
H5File file2( FILE2, H5F_ACC_TRUNC, *tmpl1 );
/* Release file-creation template */
// Release file-creation template
delete tmpl1;
tmpl1 = NULL;
/* Get the file-creation template */
// Get the file-creation template
tmpl1 = new FileCreatPropList (file2.getCreatePlist());
/* Get the file-creation parameters */
// Get the file-creation parameters
hsize_t ublock = tmpl1->getUserblock();
verify_val((long)ublock, (long)F2_USERBLOCK_SIZE, "FileCreatPropList::getUserblock", __LINE__, __FILE__);
size_t parm1, parm2; /*file-creation parameters */
size_t parm1, parm2; // file-creation parameters
tmpl1->getSizes( parm1, parm2);
verify_val(parm1, F2_OFFSET_SIZE, "FileCreatPropList::getSizes", __LINE__, __FILE__);
verify_val(parm2, F2_LENGTH_SIZE, "FileCreatPropList::getSizes", __LINE__, __FILE__);
unsigned iparm1,iparm2; /*file-creation parameters */
unsigned iparm1,iparm2; // file-creation parameters
tmpl1->getSymk( iparm1, iparm2);
verify_val(iparm1, F2_SYM_INTERN_K, "FileCreatPropList::getSymk", __LINE__, __FILE__);
verify_val(iparm2, F2_SYM_LEAF_K, "FileCreatPropList::getSymk", __LINE__, __FILE__);
/* Clone the file-creation template */
// Clone the file-creation template
FileCreatPropList tmpl2;
tmpl2.copy (*tmpl1);
/* Dynamically release file-creation template */
// Release file-creation template
delete tmpl1;
tmpl1 = NULL;
/* Set the new file-creation parameter */
// Set the new file-creation parameter
tmpl2.setUserblock( F3_USERBLOCK_SIZE );
/*
* Try to create second file, with non-standard file-creation template
* params
*/
// Try to create second file, with non-standard file-creation template
// params
H5File file3( FILE3, H5F_ACC_TRUNC, tmpl2 );
/* Get the file-creation template */
// Get the file-creation template
tmpl1 = new FileCreatPropList (file3.getCreatePlist());
/* Get the file-creation parameters */
// Get the file-creation parameters
ublock = tmpl1->getUserblock();
verify_val((long)ublock, (long)F3_USERBLOCK_SIZE, "FileCreatPropList::getUserblock", __LINE__, __FILE__);
@ -261,17 +258,17 @@ test_file_create(void)
verify_val(iparm1, F3_SYM_INTERN_K, "FileCreatPropList::getSymk", __LINE__, __FILE__);
verify_val(iparm2, F3_SYM_LEAF_K, "FileCreatPropList::getSymk", __LINE__, __FILE__);
/* Dynamically release file-creation template */
// Release file-creation template
delete tmpl1;
}
// catch all exceptions
catch (Exception E)
{
issue_fail_msg(E.getCFuncName(), __LINE__, __FILE__, E.getCDetailMsg());
if (tmpl1 != NULL) // clean up
if (tmpl1 != NULL) // clean up
delete tmpl1;
}
} /* test_file_create() */
} // test_file_create()
/*-------------------------------------------------------------------------
@ -289,7 +286,7 @@ test_file_create(void)
* cases. Since there are no operator<< for 'long long'
* or int64 in VS C++ ostream, I casted the hsize_t values
* passed to verify_val to 'long' as well. If problems
* arises later, this will have to be specificly handled
* arises later, this will have to be specificly handled
* with a special routine.
*
*-------------------------------------------------------------------------
@ -297,27 +294,27 @@ test_file_create(void)
static void
test_file_open(void)
{
/* Output message about test being performed */
// Output message about test being performed
MESSAGE(5, ("Testing File Opening I/O\n"));
try {
/* Open first file */
// Open first file
H5File file1 (FILE2, H5F_ACC_RDWR );
/* Get the file-creation template */
// Get the file-creation template
FileCreatPropList tmpl1 = file1.getCreatePlist();
/* Get the file-creation parameters */
// Get the file-creation parameters
hsize_t ublock = tmpl1.getUserblock();
verify_val((long)ublock, (long)F2_USERBLOCK_SIZE, "FileCreatPropList::getUserblock", __LINE__, __FILE__);
size_t parm1, parm2; /*file-creation parameters */
size_t parm1, parm2; // file-creation parameters
tmpl1.getSizes( parm1, parm2);
verify_val(parm1, F2_OFFSET_SIZE, "FileCreatPropList::getSizes", __LINE__, __FILE__);
verify_val(parm2, F2_LENGTH_SIZE, "FileCreatPropList::getSizes", __LINE__, __FILE__);
unsigned iparm1,iparm2; /*file-creation parameters */
unsigned iparm1,iparm2; // file-creation parameters
tmpl1.getSymk( iparm1, iparm2);
verify_val(iparm1, F2_SYM_INTERN_K, "FileCreatPropList::getSymk", __LINE__, __FILE__);
verify_val(iparm2, F2_SYM_LEAF_K, "FileCreatPropList::getSymk", __LINE__, __FILE__);
@ -326,7 +323,7 @@ test_file_open(void)
catch( Exception E ) {
issue_fail_msg(E.getCFuncName(), __LINE__, __FILE__, E.getCDetailMsg());
}
} /* test_file_open() */
} // test_file_open()
/*-------------------------------------------------------------------------
@ -346,7 +343,7 @@ test_file_open(void)
static void
test_file_size(void)
{
/* Output message about test being performed */
// Output message about test being performed
MESSAGE(5, ("Testing File Size\n"));
hid_t fapl_id;
@ -360,7 +357,7 @@ test_file_size(void)
// Set to sec2 driver. Do we want to test other file drivers?
// They're not tested in C++.
// File drivers seem not implemented.
//fapl.setSec2();
// fapl.setSec2();
// Create a file
H5File file4( FILE4, H5F_ACC_TRUNC, FileCreatPropList::DEFAULT, fapl);
@ -380,7 +377,7 @@ test_file_size(void)
// use C test utility routine to close property list.
H5Pclose(fapl_id);
} /* test_file_size() */
} // test_file_size()
/*-------------------------------------------------------------------------
@ -405,7 +402,7 @@ const string DSETNAME ("dataset");
const string ATTRNAME ("attribute");
const string DTYPENAME ("compound");
/* Compound datatype */
// Compound datatype
typedef struct s1_t {
unsigned int a;
float b;
@ -414,7 +411,7 @@ typedef struct s1_t {
static void
test_file_name()
{
/* Output message about test being performed */
// Output message about test being performed
MESSAGE(5, ("Testing File Name\n"));
string file_name;
@ -426,42 +423,42 @@ test_file_name()
file_name = file4.getFileName();
verify_val(file_name, FILE4, "H5File::getFileName", __LINE__, __FILE__);
/* Create a group in the root group */
// Create a group in the root group
Group group(file4.createGroup(GROUPNAME, 0));
/* Get and verify file name */
// Get and verify file name
file_name = group.getFileName();
verify_val(file_name, FILE4, "Group::getFileName", __LINE__, __FILE__);
/* Create the data space */
// Create the data space
hsize_t dims[RANK] = {NX, NY};
DataSpace space(RANK, dims);
/* Create a new dataset */
// Create a new dataset
DataSet dataset(file4.createDataSet (DSETNAME, PredType::NATIVE_INT, space));
/* Get and verify file name */
// Get and verify file name
file_name = dataset.getFileName();
verify_val(file_name, FILE4, "DataSet::getFileName", __LINE__, __FILE__);
/* Create an attribute for the dataset */
// Create an attribute for the dataset
Attribute attr(dataset.createAttribute(ATTRNAME, PredType::NATIVE_INT, space));
/* Get and verify file name */
// Get and verify file name
file_name = attr.getFileName();
verify_val(file_name, FILE4, "Attribute::getFileName", __LINE__, __FILE__);
/* Create a compound datatype */
// Create a compound datatype
CompType comp_type (sizeof(s1_t));
/* Insert fields */
// Insert fields
comp_type.insertMember("a", HOFFSET(s1_t, a), PredType::NATIVE_INT);
comp_type.insertMember("b", HOFFSET(s1_t, b), PredType::NATIVE_FLOAT);
/* Save it on file */
// Save it on file
comp_type.commit(file4, DTYPENAME);
/* Get and verify file name */
// Get and verify file name
comp_type.getFileName();
verify_val(file_name, FILE4, "CompType::getFileName", __LINE__, __FILE__);
} // end of try block
@ -470,7 +467,7 @@ test_file_name()
issue_fail_msg(E.getCFuncName(), __LINE__, __FILE__, E.getCDetailMsg());
}
} /* test_file_name() */
} // test_file_name()
/*-------------------------------------------------------------------------
@ -490,14 +487,14 @@ test_file_name()
void
test_file(void)
{
/* Output message about test being performed */
// Output message about test being performed
MESSAGE(5, ("Testing File I/O operations\n"));
test_file_create(); /* Test file creation (also creation templates) */
test_file_open(); /* Test file opening */
test_file_size(); /* Test file size */
test_file_name(); /* Test getting file's name */
} /* test_file() */
test_file_create(); // Test file creation (also creation templates)
test_file_open(); // Test file opening
test_file_size(); // Test file size
test_file_name(); // Test getting file's name
} // test_file()
/*-------------------------------------------------------------------------
@ -520,4 +517,4 @@ cleanup_file(void)
remove(FILE2.c_str());
remove(FILE3.c_str());
remove(FILE4.c_str());
} /* cleanup_file */
} // cleanup_file

View File

@ -113,32 +113,29 @@ test_h5s_basic(void)
hsize_t dims3[H5S_MAX_RANK+1];
hsize_t tmax[4];
/* Output message about test being performed */
// Output message about test being performed
MESSAGE(5, ("Testing Dataspace Manipulation\n"));
try
{ // beginning of first try block
/* Create file - removed this since the following operations don't
need the file to be opened */
// Create simple dataspace sid1
DataSpace sid1 (SPACE1_RANK, dims1 );
// Get simple extent npoints of the dataspace sid1 and verify it
hssize_t n; /* Number of dataspace elements */
hssize_t n; // Number of dataspace elements
n = sid1.getSimpleExtentNpoints();
verify_val((long)n, (long)(SPACE1_DIM1 * SPACE1_DIM2 * SPACE1_DIM3),
"DataSpace::getSimpleExtentNpoints", __LINE__, __FILE__);
// Get the logical rank of dataspace sid1 and verify it
int rank; /* Logical rank of dataspace */
int rank; // Logical rank of dataspace
rank = sid1.getSimpleExtentNdims();
verify_val(rank, SPACE1_RANK, "DataSpace::getSimpleExtentNdims", __LINE__, __FILE__);
// Retrieves dimension size of dataspace sid1 and verify it
int ndims; /* Number of dimensions */
hsize_t tdims[4]; /* Dimension array to test with */
int ndims; // Number of dimensions
hsize_t tdims[4]; // Dimension array to test with
ndims = sid1.getSimpleExtentDims( tdims );
verify_val(HDmemcmp(tdims, dims1, SPACE1_RANK * sizeof(unsigned)), 0,
"DataSpace::getSimpleExtentDims", __LINE__, __FILE__);
@ -240,7 +237,7 @@ test_h5s_basic(void)
{
issue_fail_msg(E.getCFuncName(), __LINE__, __FILE__, E.getCDetailMsg());
}
} /* test_h5s_basic() */
} // test_h5s_basic()
/*-------------------------------------------------------------------------
*
@ -265,8 +262,7 @@ test_h5s_basic(void)
static void
test_h5s_scalar_write(void)
{
/* Output message about test being performed */
// Output message about test being performed
MESSAGE(5, ("Testing Scalar Dataspace Writing\n"));
try
@ -274,30 +270,30 @@ test_h5s_scalar_write(void)
// Create file
H5File fid1(DATAFILE, H5F_ACC_TRUNC);
/* Create scalar dataspace */
// Create scalar dataspace
DataSpace sid1(SPACE3_RANK, NULL);
//n = H5Sget_simple_extent_npoints(sid1);
hssize_t n; /* Number of dataspace elements */
hssize_t n; // Number of dataspace elements
n = sid1.getSimpleExtentNpoints();
verify_val((long)n, 1, "DataSpace::getSimpleExtentNpoints", __LINE__, __FILE__);
int rank; /* Logical rank of dataspace */
int rank; // Logical rank of dataspace
rank = sid1.getSimpleExtentNdims();
verify_val(rank, SPACE3_RANK, "DataSpace::getSimpleExtentNdims", __LINE__, __FILE__);
// Retrieves dimension size of dataspace sid1 and verify it
int ndims; /* Number of dimensions */
hsize_t tdims[4]; /* Dimension array to test with */
int ndims; // Number of dimensions
hsize_t tdims[4]; // Dimension array to test with
ndims = sid1.getSimpleExtentDims( tdims );
verify_val(ndims, 0, "DataSpace::getSimpleExtentDims", __LINE__, __FILE__);
/* Verify extent type */
H5S_class_t ext_type; /* Extent type */
// Verify extent type
H5S_class_t ext_type; // Extent type
ext_type = sid1.getSimpleExtentType();
verify_val(ext_type, H5S_SCALAR, "DataSpace::getSimpleExtentType", __LINE__, __FILE__);
/* Create a dataset */
// Create a dataset
DataSet dataset = fid1.createDataSet("Dataset1", PredType::NATIVE_UINT,sid1);
dataset.write(&space3_data, PredType::NATIVE_UINT);
@ -306,7 +302,7 @@ test_h5s_scalar_write(void)
{
issue_fail_msg(E.getCFuncName(), __LINE__, __FILE__, E.getCDetailMsg());
}
} /* test_h5s_scalar_write() */
} // test_h5s_scalar_write()
/*-------------------------------------------------------------------------
*
@ -331,17 +327,17 @@ test_h5s_scalar_write(void)
static void
test_h5s_scalar_read(void)
{
hsize_t tdims[4]; /* Dimension array to test with */
hsize_t tdims[4]; // Dimension array to test with
/* Output message about test being performed */
// Output message about test being performed
MESSAGE(5, ("Testing Scalar Dataspace Reading\n"));
try
{
/* Create file */
// Create file
H5File fid1(DATAFILE, H5F_ACC_RDWR);
/* Create a dataset */
// Create a dataset
DataSet dataset = fid1.openDataSet("Dataset1");
DataSpace sid1 = dataset.getSpace();
@ -357,7 +353,7 @@ test_h5s_scalar_read(void)
ndims = sid1.getSimpleExtentDims(tdims);
verify_val(ndims, 0, "DataSpace::getSimpleExtentDims", __LINE__, __FILE__);
unsigned rdata; /* Scalar data read in */
unsigned rdata; // Scalar data read in
dataset.read(&rdata, PredType::NATIVE_UINT);
verify_val(rdata, space3_data, "DataSet::read", __LINE__, __FILE__);
} // end of try block
@ -367,7 +363,7 @@ test_h5s_scalar_read(void)
issue_fail_msg(E.getCFuncName(), __LINE__, __FILE__, E.getCDetailMsg());
}
} /* test_h5s_scalar_read() */
} // test_h5s_scalar_read()
/*-------------------------------------------------------------------------
*
@ -392,8 +388,7 @@ test_h5s_scalar_read(void)
static void
test_h5s_null(void)
{
/* Output message about test being performed */
// Output message about test being performed
MESSAGE(5, ("Testing Null Dataspace Writing\n"));
try
@ -401,11 +396,10 @@ test_h5s_null(void)
// Create file
H5File fid1(DATAFILE, H5F_ACC_TRUNC);
/* Create scalar dataspace */
// Create scalar dataspace
DataSpace sid1(H5S_NULL);
//n = H5Sget_simple_extent_npoints(sid1);
hssize_t n; /* Number of dataspace elements */
hssize_t n; // Number of dataspace elements
n = sid1.getSimpleExtentNpoints();
verify_val((long)n, 0, "DataSpace::getSimpleExtentNpoints", __LINE__, __FILE__);
@ -423,7 +417,7 @@ test_h5s_null(void)
{
issue_fail_msg(E.getCFuncName(), __LINE__, __FILE__, E.getCDetailMsg());
}
} /* test_h5s_null() */
} // test_h5s_null()
/*-------------------------------------------------------------------------
*
@ -449,16 +443,15 @@ test_h5s_null(void)
static void
test_h5s_compound_scalar_write(void)
{
/* Output message about test being performed */
// Output message about test being performed
MESSAGE(5, ("Testing Compound Dataspace Writing\n"));
try
{
/* Create file */
// Create file
H5File fid1(DATAFILE, H5F_ACC_TRUNC);
/* Create the compound datatype. */
// Create the compound datatype.
CompType tid1(sizeof(struct space4_struct));
space4_field1_off=HOFFSET(struct space4_struct, c1);
tid1.insertMember(SPACE4_FIELDNAME1, space4_field1_off,
@ -473,7 +466,7 @@ test_h5s_compound_scalar_write(void)
tid1.insertMember(SPACE4_FIELDNAME4, space4_field4_off,
PredType::NATIVE_SCHAR);
/* Create scalar dataspace */
// Create scalar dataspace
DataSpace sid1(SPACE3_RANK, NULL);
// Get the number of dataspace elements
@ -484,11 +477,11 @@ test_h5s_compound_scalar_write(void)
int ndims = sid1.getSimpleExtentNdims();
verify_val(ndims, SPACE3_RANK, "DataSpace::getSimpleExtentNdims", __LINE__, __FILE__);
hsize_t tdims[4]; /* Dimension array to test with */
hsize_t tdims[4]; // Dimension array to test with
ndims = sid1.getSimpleExtentDims(tdims);
verify_val(ndims, 0, "DataSpace::getSimpleExtentDims", __LINE__, __FILE__);
/* Create a dataset */
// Create a dataset
DataSet dataset = fid1.createDataSet("Dataset1", tid1, sid1);
dataset.write(&space4_data, tid1);
@ -498,8 +491,7 @@ test_h5s_compound_scalar_write(void)
// all the exceptions caused by negative returned values by C APIs
issue_fail_msg(E.getCFuncName(), __LINE__, __FILE__, E.getCDetailMsg());
}
} /* test_h5s_compound_scalar_write() */
} // test_h5s_compound_scalar_write()
/*-------------------------------------------------------------------------
*
@ -525,16 +517,16 @@ test_h5s_compound_scalar_write(void)
static void
test_h5s_compound_scalar_read(void)
{
hsize_t tdims[4]; /* Dimension array to test with */
hsize_t tdims[4]; // Dimension array to test with
/* Output message about test being performed */
// Output message about test being performed
MESSAGE(5, ("Testing Compound Dataspace Reading\n"));
try
{
/* Create file */
// Create file
H5File fid1(DATAFILE, H5F_ACC_RDWR);
/* Create a dataset */
// Create a dataset
DataSet dataset = fid1.openDataSet("Dataset1");
DataSpace sid1 = dataset.getSpace();
@ -553,7 +545,7 @@ test_h5s_compound_scalar_read(void)
// Get the datatype of this dataset.
CompType type(dataset);
struct space4_struct rdata; /* Scalar data read in */
struct space4_struct rdata; // Scalar data read in
dataset.read(&rdata, type);
// Verify read data
@ -567,14 +559,14 @@ test_h5s_compound_scalar_read(void)
<< space4_data.f << ", read_data4.f=" << rdata.f << endl;
TestErrPrintf("scalar data different: space4_data.c1=%c, read_data4.c1=%c\n",
space4_data.c1, rdata.c2);
} /* end if */
} // end if
} // end of try block
catch (Exception E)
{
// all the exceptions caused by negative returned values by C APIs
issue_fail_msg(E.getCFuncName(), __LINE__, __FILE__, E.getCDetailMsg());
}
} /* test_h5s_compound_scalar_read() */
} // test_h5s_compound_scalar_read()
/*-------------------------------------------------------------------------
*
@ -593,16 +585,16 @@ test_h5s_compound_scalar_read(void)
void
test_h5s(void)
{
/* Output message about test being performed */
// Output message about test being performed
MESSAGE(5, ("Testing Dataspaces\n"));
test_h5s_basic(); /* Test basic H5S code */
test_h5s_scalar_write(); /* Test scalar H5S writing code */
test_h5s_scalar_read(); /* Test scalar H5S reading code */
test_h5s_null(); /* Test null H5S code */
test_h5s_compound_scalar_write(); /* Test compound datatype scalar H5S writing code */
test_h5s_compound_scalar_read(); /* Test compound datatype scalar H5S reading code */
} /* test_h5s() */
test_h5s_basic(); // Test basic H5S code
test_h5s_scalar_write(); // Test scalar H5S writing code
test_h5s_scalar_read(); // Test scalar H5S reading code
test_h5s_null(); // Test null H5S code
test_h5s_compound_scalar_write(); // Test compound datatype scalar H5S writing code
test_h5s_compound_scalar_read(); // Test compound datatype scalar H5S reading code
} // test_h5s()
/*-------------------------------------------------------------------------
@ -623,5 +615,5 @@ void
cleanup_h5s(void)
{
remove(DATAFILE.c_str());
}
} // cleanup_h5s