mirror of
https://github.com/HDFGroup/hdf5.git
synced 2025-01-18 15:15:56 +08:00
5b18a76318
Added new testframe.h header to document testing framework functions and split them away from h5test.h and from test programs that don't integrate with the testframe.c testing framework Added new test setup callback to testframe.c testing framework Added parameters to AddTest() to specify size of test parameters so they can be copied for later use Enabled HDF5 error stacks in testframe.c framework by default and added some error stack suppressions to some testhdf5 tests Added new maxthreads option to testframe.c framework to allow specifying the maximum number of threads a multi-threaded test can use Moved TestExpress functionality out of testframe.c and into more general h5test.c for wider use by tests through getter and setter Updated some tests to not mix and match functionality between h5test.c/h and testframe.c/h Moved some functionality from testphdf5.h into testpar.h for parallel tests that aren't part of testphdf5 Added new parallel test library that contains common shared functionality for parallel tests (similar to h5test library)
175 lines
6.1 KiB
C
175 lines
6.1 KiB
C
/* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *
|
|
* Copyright by The HDF Group. *
|
|
* All rights reserved. *
|
|
* *
|
|
* This file is part of HDF5. The full HDF5 copyright notice, including *
|
|
* terms governing use, modification, and redistribution, is contained in *
|
|
* the COPYING file, which can be found at the root of the source code *
|
|
* distribution tree, or in https://www.hdfgroup.org/licenses. *
|
|
* If you do not have access to either file, you may request a copy from *
|
|
* help@hdfgroup.org. *
|
|
* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
|
|
|
|
/*
|
|
* Purpose: This test creates a file and a bunch of objects in the
|
|
* file and then calls MPI_Finalize without closing anything. The
|
|
* library should exercise the attribute callback destroy attached to
|
|
* MPI_COMM_SELF and terminate the HDF5 library closing all open
|
|
* objects. The t_prestart test will read back the file and make sure
|
|
* all created objects are there.
|
|
*/
|
|
|
|
#include "testpar.h"
|
|
|
|
#define RANK 2
|
|
#define ROW_FACTOR 8 /* Nominal row factor for dataset size */
|
|
#define COL_FACTOR 16 /* Nominal column factor for dataset size */
|
|
|
|
/* Dataset data type. Int's can be easily octo dumped. */
|
|
typedef int DATATYPE;
|
|
|
|
int nerrors = 0; /* errors count */
|
|
|
|
static const char *FILENAME[] = {"shutdown", NULL};
|
|
|
|
int
|
|
main(int argc, char **argv)
|
|
{
|
|
hid_t file_id, dset_id, grp_id;
|
|
hid_t fapl, sid, mem_dataspace;
|
|
hsize_t dims[RANK], i;
|
|
herr_t ret;
|
|
char filename[1024];
|
|
int mpi_size, mpi_rank;
|
|
MPI_Comm comm = MPI_COMM_WORLD;
|
|
MPI_Info info = MPI_INFO_NULL;
|
|
hsize_t start[RANK];
|
|
hsize_t count[RANK];
|
|
hsize_t stride[RANK];
|
|
hsize_t block[RANK];
|
|
DATATYPE *data_array = NULL; /* data buffer */
|
|
int mpi_code;
|
|
#ifdef H5_HAVE_TEST_API
|
|
int required = MPI_THREAD_MULTIPLE;
|
|
int provided;
|
|
#endif
|
|
|
|
#ifdef H5_HAVE_TEST_API
|
|
/* Attempt to initialize with MPI_THREAD_MULTIPLE if possible */
|
|
if (MPI_SUCCESS != (mpi_code = MPI_Init_thread(&argc, &argv, required, &provided))) {
|
|
printf("MPI_Init_thread failed with error code %d\n", mpi_code);
|
|
return -1;
|
|
}
|
|
#else
|
|
if (MPI_SUCCESS != (mpi_code = MPI_Init(&argc, &argv))) {
|
|
printf("MPI_Init failed with error code %d\n", mpi_code);
|
|
return -1;
|
|
}
|
|
#endif
|
|
|
|
if (MPI_SUCCESS != (mpi_code = MPI_Comm_rank(comm, &mpi_rank))) {
|
|
printf("MPI_Comm_rank failed with error code %d\n", mpi_code);
|
|
MPI_Finalize();
|
|
return -1;
|
|
}
|
|
|
|
#ifdef H5_HAVE_TEST_API
|
|
/* Warn about missing MPI_THREAD_MULTIPLE support */
|
|
if ((provided < required) && MAINPROCESS)
|
|
printf("** MPI doesn't support MPI_Init_thread with MPI_THREAD_MULTIPLE **\n");
|
|
#endif
|
|
|
|
if (MPI_SUCCESS != (mpi_code = MPI_Comm_size(comm, &mpi_size))) {
|
|
if (MAINPROCESS)
|
|
printf("MPI_Comm_size failed with error code %d\n", mpi_code);
|
|
MPI_Finalize();
|
|
return -1;
|
|
}
|
|
|
|
if (MAINPROCESS)
|
|
TESTING("proper shutdown of HDF5 library");
|
|
|
|
/* Set up file access property list with parallel I/O access */
|
|
fapl = H5Pcreate(H5P_FILE_ACCESS);
|
|
VRFY((fapl >= 0), "H5Pcreate succeeded");
|
|
|
|
/* Get the capability flag of the VOL connector being used */
|
|
ret = H5Pget_vol_cap_flags(fapl, &vol_cap_flags_g);
|
|
VRFY((ret >= 0), "H5Pget_vol_cap_flags succeeded");
|
|
|
|
/* Make sure the connector supports the API functions being tested */
|
|
if (!(vol_cap_flags_g & H5VL_CAP_FLAG_FILE_BASIC) || !(vol_cap_flags_g & H5VL_CAP_FLAG_GROUP_BASIC) ||
|
|
!(vol_cap_flags_g & H5VL_CAP_FLAG_DATASET_BASIC)) {
|
|
if (MAINPROCESS) {
|
|
puts("SKIPPED");
|
|
printf(
|
|
" API functions for basic file, group, or dataset aren't supported with this connector\n");
|
|
fflush(stdout);
|
|
}
|
|
|
|
MPI_Finalize();
|
|
return 0;
|
|
}
|
|
|
|
ret = H5Pset_fapl_mpio(fapl, comm, info);
|
|
VRFY((ret >= 0), "");
|
|
|
|
h5_fixname(FILENAME[0], fapl, filename, sizeof filename);
|
|
file_id = H5Fcreate(filename, H5F_ACC_TRUNC, H5P_DEFAULT, fapl);
|
|
VRFY((file_id >= 0), "H5Fcreate succeeded");
|
|
grp_id = H5Gcreate2(file_id, "Group", H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT);
|
|
VRFY((grp_id >= 0), "H5Gcreate succeeded");
|
|
|
|
dims[0] = (hsize_t)ROW_FACTOR * (hsize_t)mpi_size;
|
|
dims[1] = (hsize_t)COL_FACTOR * (hsize_t)mpi_size;
|
|
sid = H5Screate_simple(RANK, dims, NULL);
|
|
VRFY((sid >= 0), "H5Screate_simple succeeded");
|
|
|
|
dset_id = H5Dcreate2(grp_id, "Dataset", H5T_NATIVE_INT, sid, H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT);
|
|
VRFY((dset_id >= 0), "H5Dcreate succeeded");
|
|
|
|
/* allocate memory for data buffer */
|
|
data_array = (DATATYPE *)malloc(dims[0] * dims[1] * sizeof(DATATYPE));
|
|
VRFY((data_array != NULL), "data_array malloc succeeded");
|
|
|
|
/* Each process takes a slabs of rows. */
|
|
block[0] = dims[0] / (hsize_t)mpi_size;
|
|
block[1] = dims[1];
|
|
stride[0] = block[0];
|
|
stride[1] = block[1];
|
|
count[0] = 1;
|
|
count[1] = 1;
|
|
start[0] = (hsize_t)mpi_rank * block[0];
|
|
start[1] = 0;
|
|
|
|
/* put some trivial data in the data_array */
|
|
for (i = 0; i < dims[0] * dims[1]; i++)
|
|
data_array[i] = mpi_rank + 1;
|
|
|
|
ret = H5Sselect_hyperslab(sid, H5S_SELECT_SET, start, stride, count, block);
|
|
VRFY((ret >= 0), "H5Sset_hyperslab succeeded");
|
|
|
|
/* create a memory dataspace independently */
|
|
mem_dataspace = H5Screate_simple(RANK, block, NULL);
|
|
VRFY((mem_dataspace >= 0), "");
|
|
|
|
/* write data independently */
|
|
ret = H5Dwrite(dset_id, H5T_NATIVE_INT, mem_dataspace, sid, H5P_DEFAULT, data_array);
|
|
VRFY((ret >= 0), "H5Dwrite succeeded");
|
|
|
|
/* release data buffers */
|
|
if (data_array)
|
|
free(data_array);
|
|
|
|
MPI_Finalize();
|
|
|
|
if (MAINPROCESS) {
|
|
if (0 == nerrors)
|
|
PASSED();
|
|
else
|
|
H5_FAILED();
|
|
}
|
|
|
|
return (nerrors != 0);
|
|
}
|