mirror of
https://github.com/HDFGroup/hdf5.git
synced 2024-11-27 02:10:55 +08:00
cc89b8a605
---------------------- ./MANIFEST ./src/H5FDmulti.c [NEW] ./src/H5FDmulti.h [NEW] ./src/Makefile.in ./src/hdf5.h The split driver was reimplemented as a more general "multi" driver which is capable of splitting data into multiple files like the family driver except the partioning is done by memory usage type instead of address. The H5Pset_fapl_split() function just calls H5Pset_fapl_multi() with arguments which prepare to split the address space into two files: meta and raw data. This is the first version. I plan to allow the open() call to relax a bit which would allow one to open an hdf5 file when only the meta-data file is present. This would allow a very large file to be split and stored on tape and the relatively small meta file to be mirrored on disk to allow limited browsing of the file (any request for raw data would fail). ./src/H5private.h ./src/H5F.c ./src/H5FD.c ./src/H5FDprivate.h ./src/H5FDpublic.h ./src/H5FDcore.c ./src/H5FDfamily.c ./src/H5FDmpio.c ./src/H5FDsec2.c Added the ability for a file driver to store information in the superblock which would be needed if the file were opened again later for reading. The format is driver-defined which allows users to extend it however they like. ./doc/html/H5.format.html Added information about the new driver information block of the superblock. This is where file drivers store information they need in order to reopen the file later. ./src/H5F.c ./src/H5Fprivate.h ./src/H5FD.c ./src/H5FDprivate.h ./src/H5FDpublic.h ./src/H5FDcore.c ./src/H5FDfamily.c ./src/H5FDmpio.c ./src/H5FDsec2.c ./src/H5Fistore.c ./src/H5R.c The file access properties and the file access property list were decoupled, which allows the property list to more cleanly contain properties for various levels of the file and which allows the property list to be modified more cleanly when opening files. ./src/H5.c ./src/H5FDpublic.h Removed H5FD_MEM_META and H5FD_MEM_GROUP since they're never used. ./src/H5D.c Changed the way we detect the MPIO driver in all these special cases. ./src/H5F.c ./src/H5Rpublic.h ./test/tfile.c The default file sizeof(offset) was changed to be a function of haddr_t instead of hsize_t. THE H5RPUBLIC.H DEFINITIONS WILL HAVE PROBLEMS IF THE USER CREATES A FILE WITH NON-DEFAULT OFFSET AND SIZE SIZES! ./src/H5F.c Fixed an uninitialized memory access bug in file closing related to the VFL. ./src/H5T.c ./src/H5Tpublic.h Added an H5T_NATIVE_HADDR predefined datatype which corresponds to the `haddr_t' type. ./test/Makefile.in Reformatted long lines. ./test/big.c ./test/cmpd_dset.c ./test/dsets.c ./test/dtypes.c ./test/extend.c ./test/external.c Removed the H5F_ACC_DEBUG flag from file creation/open calls. ./test/big.c Plugged a memory leak. ./test/h5test.c Added support for the `multi' driver. Removed #warning about not having the stdio driver. Plans are to not implement it since the sec2 driver serves the same purpose and testing didn't show any difference in execution times between the two.
714 lines
19 KiB
C
714 lines
19 KiB
C
/*
|
||
* Copyright (C) 1998 NCSA
|
||
* All rights reserved.
|
||
*
|
||
* Programmer: Robb Matzke <matzke@llnl.gov>
|
||
* Friday, January 23, 1998
|
||
*/
|
||
|
||
/* See H5private.h for how to include headers */
|
||
#undef NDEBUG
|
||
#include <h5test.h>
|
||
|
||
const char *FILENAME[] = {
|
||
"cmpd_dset",
|
||
NULL
|
||
};
|
||
|
||
/* The first dataset */
|
||
typedef struct s1_t {
|
||
unsigned int a;
|
||
unsigned int b;
|
||
unsigned int c[4];
|
||
unsigned int d;
|
||
unsigned int e;
|
||
} s1_t;
|
||
|
||
/* The second dataset (same as first) */
|
||
typedef s1_t s2_t;
|
||
|
||
/* The third dataset (reversed fields of s1) */
|
||
typedef struct s3_t {
|
||
unsigned int e;
|
||
unsigned int d;
|
||
unsigned int c[4];
|
||
unsigned int b;
|
||
unsigned int a;
|
||
} s3_t;
|
||
|
||
/* The fourth dataset (a subset of s1) */
|
||
typedef struct s4_t {
|
||
unsigned int b;
|
||
unsigned int d;
|
||
} s4_t;
|
||
|
||
/* The fifth dataset (a superset of s1) */
|
||
typedef struct s5_t {
|
||
unsigned int pre;
|
||
unsigned int a;
|
||
unsigned int b;
|
||
unsigned int mid1;
|
||
unsigned int c[4];
|
||
unsigned int mid2;
|
||
unsigned int d;
|
||
unsigned int e;
|
||
unsigned int post;
|
||
} s5_t;
|
||
|
||
|
||
#if 1
|
||
# define NX 100u
|
||
# define NY 2000u
|
||
#else
|
||
# define NX 12u
|
||
# define NY 9u
|
||
#endif
|
||
|
||
|
||
/*-------------------------------------------------------------------------
|
||
* Function: main
|
||
*
|
||
* Purpose: Creates a simple dataset of a compound type and then reads
|
||
* it back. The dataset is read back in various ways to
|
||
* exercise the I/O pipeline and compound type conversion.
|
||
*
|
||
* Return: Success: 0
|
||
*
|
||
* Failure: 1
|
||
*
|
||
* Programmer: Robb Matzke
|
||
* Friday, January 23, 1998
|
||
*
|
||
* Modifications:
|
||
* Robb Matzke, 1999-06-23
|
||
* If the command line switch `--noopt' is present then the fast
|
||
* compound datatype conversion is turned off.
|
||
*-------------------------------------------------------------------------
|
||
*/
|
||
int
|
||
main (int argc, char *argv[])
|
||
{
|
||
/* First dataset */
|
||
static s1_t s1[NX*NY];
|
||
hid_t s1_tid;
|
||
|
||
/* Second dataset */
|
||
static s2_t s2[NX*NY];
|
||
hid_t s2_tid;
|
||
|
||
/* Third dataset */
|
||
static s3_t s3[NX*NY];
|
||
hid_t s3_tid;
|
||
|
||
/* Fourth dataset */
|
||
static s4_t s4[NX*NY];
|
||
hid_t s4_tid;
|
||
|
||
/* Fifth dataset */
|
||
static s5_t s5[NX*NY];
|
||
hid_t s5_tid;
|
||
|
||
/* Sixth dataset */
|
||
|
||
/* Seventh dataset */
|
||
hid_t s7_sid;
|
||
|
||
/* Eighth dataset */
|
||
s1_t *s8 = NULL;
|
||
hid_t s8_f_sid; /*file data space */
|
||
hid_t s8_m_sid; /*memory data space */
|
||
|
||
/* Ninth dataset */
|
||
|
||
/* Tenth dataset */
|
||
|
||
/* Eleventh dataset */
|
||
s4_t *s11 = NULL;
|
||
|
||
/* Other variables */
|
||
unsigned int i, j;
|
||
hid_t file, dataset, space, PRESERVE, fapl;
|
||
static hsize_t dim[] = {NX, NY};
|
||
hssize_t f_offset[2]; /*offset of hyperslab in file */
|
||
hsize_t h_size[2]; /*size of hyperslab */
|
||
size_t memb_size[1] = {4};
|
||
char filename[256];
|
||
|
||
h5_reset();
|
||
|
||
/* Turn off optimized compound converter? */
|
||
if (argc>1) {
|
||
if (argc>2 || strcmp("--noopt", argv[1])) {
|
||
fprintf(stderr, "usage: %s [--noopt]\n", argv[0]);
|
||
exit(1);
|
||
}
|
||
H5Tunregister(H5T_PERS_DONTCARE, NULL, -1, -1, H5T_conv_struct_opt);
|
||
}
|
||
|
||
/* Create the file */
|
||
fapl = h5_fileaccess();
|
||
h5_fixname(FILENAME[0], fapl, filename, sizeof(filename));
|
||
if ((file = H5Fcreate (filename, H5F_ACC_TRUNC, H5P_DEFAULT, fapl))<0) {
|
||
goto error;
|
||
}
|
||
|
||
/* Create the data space */
|
||
if ((space = H5Screate_simple (2, dim, NULL))<0) goto error;
|
||
|
||
/* Create xfer properties to preserve initialized data */
|
||
if ((PRESERVE = H5Pcreate (H5P_DATA_XFER))<0) goto error;
|
||
if (H5Pset_preserve (PRESERVE, 1)<0) goto error;
|
||
|
||
/*
|
||
*######################################################################
|
||
* STEP 1: Save the original dataset natively.
|
||
*/
|
||
TESTING("basic compound write");
|
||
|
||
/* Initialize the dataset */
|
||
for (i=0; i<NX*NY; i++) {
|
||
s1[i].a = 8*i+0;
|
||
s1[i].b = 2000+2*i;
|
||
s1[i].c[0] = 8*i+2;
|
||
s1[i].c[1] = 8*i+3;
|
||
s1[i].c[2] = 8*i+4;
|
||
s1[i].c[3] = 8*i+5;
|
||
s1[i].d = 2001+2*i;
|
||
s1[i].e = 8*i+7;
|
||
}
|
||
|
||
/* Create the memory data type */
|
||
if ((s1_tid = H5Tcreate (H5T_COMPOUND, sizeof(s1_t)))<0) goto error;
|
||
if (H5Tinsert (s1_tid, "a", HOFFSET(s1_t,a), H5T_NATIVE_INT)<0 ||
|
||
H5Tinsert (s1_tid, "b", HOFFSET(s1_t,b), H5T_NATIVE_INT)<0 ||
|
||
H5Tinsert_array (s1_tid, "c", HOFFSET(s1_t,c), 1, memb_size, NULL,
|
||
H5T_NATIVE_INT)<0 ||
|
||
H5Tinsert (s1_tid, "d", HOFFSET(s1_t,d), H5T_NATIVE_INT)<0 ||
|
||
H5Tinsert (s1_tid, "e", HOFFSET(s1_t,e), H5T_NATIVE_INT)<0) {
|
||
goto error;
|
||
}
|
||
|
||
/* Create the dataset */
|
||
if ((dataset = H5Dcreate (file, "s1", s1_tid, space, H5P_DEFAULT))<0) {
|
||
goto error;
|
||
}
|
||
|
||
/* Write the data */
|
||
if (H5Dwrite (dataset, s1_tid, H5S_ALL, H5S_ALL, H5P_DEFAULT, s1)<0) {
|
||
goto error;
|
||
}
|
||
PASSED();
|
||
|
||
/*
|
||
*######################################################################
|
||
* STEP 2: We create a new type ID for the second dataset even though
|
||
* it's the same as the first just to test things better, but
|
||
* in fact, we could have used s1_tid.
|
||
*/
|
||
TESTING("basic compound read");
|
||
|
||
/* Create a data type for s2 */
|
||
if ((s2_tid = H5Tcreate (H5T_COMPOUND, sizeof(s2_t)))<0) goto error;
|
||
if (H5Tinsert (s2_tid, "a", HOFFSET(s2_t,a), H5T_NATIVE_INT)<0 ||
|
||
H5Tinsert (s2_tid, "b", HOFFSET(s2_t,b), H5T_NATIVE_INT)<0 ||
|
||
H5Tinsert_array (s2_tid, "c", HOFFSET(s2_t,c), 1, memb_size, NULL,
|
||
H5T_NATIVE_INT)<0 ||
|
||
H5Tinsert (s2_tid, "d", HOFFSET(s2_t,d), H5T_NATIVE_INT)<0 ||
|
||
H5Tinsert (s2_tid, "e", HOFFSET(s2_t,e), H5T_NATIVE_INT)<0) {
|
||
goto error;
|
||
}
|
||
|
||
/* Read the data */
|
||
if (H5Dread (dataset, s2_tid, H5S_ALL, H5S_ALL, H5P_DEFAULT, s2)<0) {
|
||
goto error;
|
||
}
|
||
|
||
/* Compare s2 with s1. They should be the same */
|
||
for (i=0; i<NX*NY; i++) {
|
||
if (s1[i].a!=s2[i].a ||
|
||
s1[i].b!=s2[i].b ||
|
||
s1[i].c[0]!=s2[i].c[0] ||
|
||
s1[i].c[1]!=s2[i].c[1] ||
|
||
s1[i].c[2]!=s2[i].c[2] ||
|
||
s1[i].c[3]!=s2[i].c[3] ||
|
||
s1[i].d!=s2[i].d ||
|
||
s1[i].e!=s2[i].e) {
|
||
FAILED();
|
||
puts(" Incorrect values read from the file");
|
||
goto error;
|
||
}
|
||
}
|
||
PASSED();
|
||
|
||
/*
|
||
*######################################################################
|
||
* STEP 3: Read the dataset back into a third memory buffer. This buffer
|
||
* has the same data space but the data type is different: the
|
||
* data type is a struct whose members are in the opposite order.
|
||
*/
|
||
TESTING("reversal of struct members");
|
||
|
||
/* Create a data type for s3 */
|
||
if ((s3_tid = H5Tcreate (H5T_COMPOUND, sizeof(s3_t)))<0) goto error;
|
||
if (H5Tinsert (s3_tid, "a", HOFFSET(s3_t,a), H5T_NATIVE_INT)<0 ||
|
||
H5Tinsert (s3_tid, "b", HOFFSET(s3_t,b), H5T_NATIVE_INT)<0 ||
|
||
H5Tinsert_array (s3_tid, "c", HOFFSET(s3_t,c), 1, memb_size, NULL,
|
||
H5T_NATIVE_INT)<0 ||
|
||
H5Tinsert (s3_tid, "d", HOFFSET(s3_t,d), H5T_NATIVE_INT)<0 ||
|
||
H5Tinsert (s3_tid, "e", HOFFSET(s3_t,e), H5T_NATIVE_INT)<0) {
|
||
goto error;
|
||
}
|
||
|
||
/* Read the data */
|
||
if (H5Dread (dataset, s3_tid, H5S_ALL, H5S_ALL, H5P_DEFAULT, s3)<0) {
|
||
goto error;
|
||
}
|
||
|
||
/* Compare s3 with s1. They should be the same */
|
||
for (i=0; i<NX*NY; i++) {
|
||
if (s1[i].a!=s3[i].a ||
|
||
s1[i].b!=s3[i].b ||
|
||
s1[i].c[0]!=s3[i].c[0] ||
|
||
s1[i].c[1]!=s3[i].c[1] ||
|
||
s1[i].c[2]!=s3[i].c[2] ||
|
||
s1[i].c[3]!=s3[i].c[3] ||
|
||
s1[i].d!=s3[i].d ||
|
||
s1[i].e!=s3[i].e) {
|
||
FAILED();
|
||
puts(" Incorrect values read from the file");
|
||
goto error;
|
||
}
|
||
}
|
||
PASSED();
|
||
|
||
/*
|
||
*######################################################################
|
||
* STEP 4: Read a subset of the members. Of the <a,b,c,d,e> members
|
||
* stored on disk we'll read <b,d>.
|
||
*/
|
||
TESTING("subset struct read");
|
||
|
||
/* Create a datatype for s4 */
|
||
if ((s4_tid = H5Tcreate (H5T_COMPOUND, sizeof(s4_t)))<0) goto error;
|
||
if (H5Tinsert (s4_tid, "b", HOFFSET(s4_t,b), H5T_NATIVE_INT)<0) goto error;
|
||
if (H5Tinsert (s4_tid, "d", HOFFSET(s4_t,d), H5T_NATIVE_INT)<0) goto error;
|
||
|
||
/* Read the data */
|
||
if (H5Dread (dataset, s4_tid, H5S_ALL, H5S_ALL, H5P_DEFAULT, s4)<0) {
|
||
goto error;
|
||
}
|
||
|
||
/* Compare s4 with s1 */
|
||
for (i=0; i<NX*NY; i++) {
|
||
if (s1[i].b!=s4[i].b ||
|
||
s1[i].d!=s4[i].d) {
|
||
FAILED();
|
||
puts(" Incorrect values read from the file");
|
||
goto error;
|
||
}
|
||
}
|
||
PASSED();
|
||
|
||
/*
|
||
*######################################################################
|
||
* STEP 5: Read all the members into a struct which has other members
|
||
* which have already been initialized.
|
||
*/
|
||
TESTING("partially initialized superset read");
|
||
|
||
/* Initialize some members */
|
||
for (i=0; i<NX*NY; i++) {
|
||
s5[i].pre = 1000+4*i;
|
||
s5[i].mid1 = 1001+4*i;
|
||
s5[i].mid2 = 1002+4*i;
|
||
s5[i].post = 1003+4*i;
|
||
}
|
||
|
||
/* Create a data type for s5 */
|
||
if ((s5_tid = H5Tcreate (H5T_COMPOUND, sizeof(s5_t)))<0) goto error;
|
||
if (H5Tinsert (s5_tid, "a", HOFFSET(s5_t,a), H5T_NATIVE_INT)<0 ||
|
||
H5Tinsert (s5_tid, "b", HOFFSET(s5_t,b), H5T_NATIVE_INT)<0 ||
|
||
H5Tinsert_array (s5_tid, "c", HOFFSET(s5_t,c), 1, memb_size, NULL,
|
||
H5T_NATIVE_INT)<0 ||
|
||
H5Tinsert (s5_tid, "d", HOFFSET(s5_t,d), H5T_NATIVE_INT)<0 ||
|
||
H5Tinsert (s5_tid, "e", HOFFSET(s5_t,e), H5T_NATIVE_INT)) {
|
||
goto error;
|
||
}
|
||
|
||
/* Read the data */
|
||
if (H5Dread (dataset, s5_tid, H5S_ALL, H5S_ALL, PRESERVE, s5)<0) {
|
||
goto error;
|
||
}
|
||
|
||
/* Check that the data was read properly */
|
||
for (i=0; i<NX*NY; i++) {
|
||
if (s1[i].a!=s5[i].a ||
|
||
s1[i].b!=s5[i].b ||
|
||
s1[i].c[0]!=s5[i].c[0] ||
|
||
s1[i].c[1]!=s5[i].c[1] ||
|
||
s1[i].c[2]!=s5[i].c[2] ||
|
||
s1[i].c[3]!=s5[i].c[3] ||
|
||
s1[i].d!=s5[i].d ||
|
||
s1[i].e!=s5[i].e) {
|
||
FAILED();
|
||
puts(" Incorrect values read from the file");
|
||
goto error;
|
||
}
|
||
}
|
||
|
||
/* Check that no previous values were clobbered */
|
||
for (i=0; i<NX*NY; i++) {
|
||
if (s5[i].pre != 1000+4*i ||
|
||
s5[i].mid1 != 1001+4*i ||
|
||
s5[i].mid2 != 1002+4*i ||
|
||
s5[i].post != 1003+4*i) {
|
||
FAILED();
|
||
puts(" Memory values were clobbered");
|
||
goto error;
|
||
}
|
||
}
|
||
PASSED();
|
||
|
||
/*
|
||
*######################################################################
|
||
* STEP 6: Update fields `b' and `d' on the file leaving the other
|
||
* fields unchanged. This tests member alignment and background
|
||
* buffers.
|
||
*/
|
||
TESTING("partially initialized superset write");
|
||
|
||
/* Initialize `s4' with new values */
|
||
for (i=0; i<NX*NY; i++) {
|
||
s4[i].b = 8*i+1;
|
||
s4[i].d = 8*i+6;
|
||
}
|
||
|
||
/* Write the data to file */
|
||
if (H5Dwrite (dataset, s4_tid, H5S_ALL, H5S_ALL, PRESERVE, s4)<0) {
|
||
goto error;
|
||
}
|
||
|
||
/* Read the data back */
|
||
if (H5Dread (dataset, s1_tid, H5S_ALL, H5S_ALL, H5P_DEFAULT, s1)<0) {
|
||
goto error;
|
||
}
|
||
|
||
/* Compare */
|
||
for (i=0; i<NX*NY; i++) {
|
||
if (s1[i].a != 8*i+0 ||
|
||
s1[i].b != 8*i+1 ||
|
||
s1[i].c[0] != 8*i+2 ||
|
||
s1[i].c[1] != 8*i+3 ||
|
||
s1[i].c[2] != 8*i+4 ||
|
||
s1[i].c[3] != 8*i+5 ||
|
||
s1[i].d != 8*i+6 ||
|
||
s1[i].e != 8*i+7) {
|
||
FAILED();
|
||
printf(" i==%u, row=%u, col=%u\n", i, i/NY, i%NY);
|
||
printf(" got: {%7d,%7d,[%7d,%7d,%7d,%7d],%7d,%7d}\n",
|
||
s1[i].a, s1[i].b, s1[i].c[0], s1[i].c[1], s1[i].c[2],
|
||
s1[i].c[3], s1[i].d, s1[i].e);
|
||
printf(" ans: {%7d,%7d,[%7d,%7d,%7d,%7d],%7d,%7d}\n",
|
||
8*i+0, 8*i+1, 8*i+2, 8*i+3, 8*i+4, 8*i+5, 8*i+6, 8*i+7);
|
||
goto error;
|
||
}
|
||
}
|
||
PASSED();
|
||
|
||
/*
|
||
*######################################################################
|
||
* STEP 7. Read the original dataset with an explicit data space. Even
|
||
* though these data spaces are equal it tests a different part of the
|
||
* library.
|
||
*/
|
||
TESTING("explicit data space");
|
||
|
||
/* Create the data space */
|
||
if ((s7_sid = H5Screate_simple (2, dim, NULL))<0) goto error;
|
||
|
||
/* Read the dataset */
|
||
if (H5Dread (dataset, s2_tid, s7_sid, H5S_ALL, H5P_DEFAULT, s2)<0) {
|
||
goto error;
|
||
}
|
||
|
||
/* Compare */
|
||
for (i=0; i<NX*NY; i++) {
|
||
if (s2[i].a != s1[i].a ||
|
||
s2[i].b != s1[i].b ||
|
||
s2[i].c[0] != s1[i].c[0] ||
|
||
s2[i].c[1] != s1[i].c[1] ||
|
||
s2[i].c[2] != s1[i].c[2] ||
|
||
s2[i].c[3] != s1[i].c[3] ||
|
||
s2[i].d != s1[i].d ||
|
||
s2[i].e != s1[i].e) {
|
||
FAILED();
|
||
puts(" Incorrect values read from file");
|
||
goto error;
|
||
}
|
||
}
|
||
PASSED();
|
||
|
||
|
||
/*
|
||
*######################################################################
|
||
* STEP 8. Read a hyperslab of the file into a complete array in memory.
|
||
* The hyperslab is the middle third of the array.
|
||
*/
|
||
TESTING("hyperslab partial read to array");
|
||
|
||
/* Create the file data space */
|
||
if ((s8_f_sid = H5Dget_space (dataset))<0) goto error;
|
||
f_offset[0] = NX/3;
|
||
f_offset[1] = NY/3;
|
||
h_size[0] = 2*NX/3 - f_offset[0];
|
||
h_size[1] = 2*NY/3 - f_offset[1];
|
||
if (H5Sselect_hyperslab (s8_f_sid, H5S_SELECT_SET, f_offset, NULL,
|
||
h_size, NULL)<0) goto error;
|
||
|
||
/* Create memory data space */
|
||
if ((s8_m_sid = H5Screate_simple (2, h_size, NULL))<0) goto error;
|
||
|
||
/* Read the dataset */
|
||
s8 = calloc ((size_t)(h_size[0]*h_size[1]), sizeof(s1_t));
|
||
assert (s8);
|
||
if (H5Dread (dataset, s1_tid, s8_m_sid, s8_f_sid, H5P_DEFAULT, s8)<0) {
|
||
goto error;
|
||
}
|
||
|
||
/* Compare */
|
||
for (i=0; i<h_size[0]; i++) {
|
||
for (j=0; j<h_size[1]; j++) {
|
||
s1_t *ps1 = s1 + (f_offset[0]+i)*NY + f_offset[1] + j;
|
||
s1_t *ps8 = s8 + i*h_size[1] + j;
|
||
|
||
if (ps8->a != ps1->a ||
|
||
ps8->b != ps1->b ||
|
||
ps8->c[0] != ps1->c[0] ||
|
||
ps8->c[1] != ps1->c[1] ||
|
||
ps8->c[2] != ps1->c[2] ||
|
||
ps8->c[3] != ps1->c[3] ||
|
||
ps8->d != ps1->d ||
|
||
ps8->e != ps1->e) {
|
||
FAILED();
|
||
puts(" Incorrect values read from file");
|
||
goto error;
|
||
}
|
||
}
|
||
}
|
||
|
||
free (s8);
|
||
s8 = NULL;
|
||
PASSED();
|
||
|
||
|
||
/*
|
||
*######################################################################
|
||
* STEP 9. Read a hyperslab of the file into a hyperslab of memory. The
|
||
* part of memory not read is already initialized and must not change.
|
||
*/
|
||
TESTING("hyperslab partial read to another hyperslab");
|
||
|
||
/* Initialize */
|
||
for (i=0; i<NX*NY; i++) {
|
||
s2[i].a = s2[i].b = s2[i].d = s2[i].e = (unsigned)(-1);
|
||
s2[i].c[0] = s2[i].c[1] = s2[i].c[2] = s2[i].c[3] = (unsigned)(-1);
|
||
}
|
||
|
||
/* Read the hyperslab */
|
||
if (H5Dread (dataset, s2_tid, s8_f_sid, s8_f_sid, H5P_DEFAULT, s2)<0) {
|
||
goto error;
|
||
}
|
||
|
||
/* Compare */
|
||
for (i=0; i<NX; i++) {
|
||
for (j=0; j<NY; j++) {
|
||
s1_t *ps1 = s1 + i*NY + j;
|
||
s2_t *ps2 = s2 + i*NY + j;
|
||
if ((hssize_t)i>=f_offset[0] &&
|
||
(hsize_t)i<f_offset[0]+h_size[0] &&
|
||
(hssize_t)j>=f_offset[1] &&
|
||
(hsize_t)j<f_offset[1]+h_size[1]) {
|
||
if (ps2->a != ps1->a ||
|
||
ps2->b != ps1->b ||
|
||
ps2->c[0] != ps1->c[0] ||
|
||
ps2->c[1] != ps1->c[1] ||
|
||
ps2->c[2] != ps1->c[2] ||
|
||
ps2->c[3] != ps1->c[3] ||
|
||
ps2->d != ps1->d ||
|
||
ps2->e != ps1->e) {
|
||
FAILED();
|
||
puts(" Memory values clobbered");
|
||
goto error;
|
||
}
|
||
} else {
|
||
if (ps2->a != (unsigned)(-1) ||
|
||
ps2->b != (unsigned)(-1) ||
|
||
ps2->c[0] != (unsigned)(-1) ||
|
||
ps2->c[1] != (unsigned)(-1) ||
|
||
ps2->c[2] != (unsigned)(-1) ||
|
||
ps2->c[3] != (unsigned)(-1) ||
|
||
ps2->d != (unsigned)(-1) ||
|
||
ps2->e != (unsigned)(-1)) {
|
||
FAILED();
|
||
puts(" Incorrect values read from file");
|
||
goto error;
|
||
}
|
||
}
|
||
}
|
||
}
|
||
PASSED();
|
||
|
||
/*
|
||
*######################################################################
|
||
* STEP 10. Same as step 9 except the memory array contains some members
|
||
* which are already initialized, like step 5.
|
||
*/
|
||
TESTING("hyperslab to hyperslab part initialized read");
|
||
|
||
/* Initialize */
|
||
for (i=0; i<NX*NY; i++) {
|
||
s5[i].a = s5[i].b = s5[i].d = s5[i].e = (unsigned)(-1);
|
||
s5[i].c[0] = s5[i].c[1] = s5[i].c[2] = s5[i].c[3] = (unsigned)(-1);
|
||
s5[i].pre = s5[i].mid1 = s5[i].mid2 = s5[i].post = (unsigned)(-1);
|
||
}
|
||
|
||
/* Read the hyperslab */
|
||
if (H5Dread (dataset, s5_tid, s8_f_sid, s8_f_sid, PRESERVE, s5)<0) {
|
||
goto error;
|
||
}
|
||
|
||
/* Compare */
|
||
for (i=0; i<NX; i++) {
|
||
for (j=0; j<NY; j++) {
|
||
s1_t *ps1 = s1 + i*NY + j;
|
||
s5_t *ps5 = s5 + i*NY + j;
|
||
if ((hssize_t)i>=f_offset[0] &&
|
||
(hsize_t)i<f_offset[0]+h_size[0] &&
|
||
(hssize_t)j>=f_offset[1] &&
|
||
(hsize_t)j<f_offset[1]+h_size[1]) {
|
||
if (ps5->pre != (unsigned)(-1) ||
|
||
ps5->a != ps1->a ||
|
||
ps5->b != ps1->b ||
|
||
ps5->mid1 != (unsigned)(-1) ||
|
||
ps5->c[0] != ps1->c[0] ||
|
||
ps5->c[1] != ps1->c[1] ||
|
||
ps5->c[2] != ps1->c[2] ||
|
||
ps5->c[3] != ps1->c[3] ||
|
||
ps5->mid2 != (unsigned)(-1) ||
|
||
ps5->d != ps1->d ||
|
||
ps5->e != ps1->e ||
|
||
ps5->post != (unsigned)(-1)) {
|
||
FAILED();
|
||
puts(" Memory values clobbered");
|
||
goto error;
|
||
}
|
||
} else {
|
||
if (ps5->pre != (unsigned)(-1) ||
|
||
ps5->a != (unsigned)(-1) ||
|
||
ps5->b != (unsigned)(-1) ||
|
||
ps5->mid1 != (unsigned)(-1) ||
|
||
ps5->c[0] != (unsigned)(-1) ||
|
||
ps5->c[1] != (unsigned)(-1) ||
|
||
ps5->c[2] != (unsigned)(-1) ||
|
||
ps5->c[3] != (unsigned)(-1) ||
|
||
ps5->mid2 != (unsigned)(-1) ||
|
||
ps5->d != (unsigned)(-1) ||
|
||
ps5->e != (unsigned)(-1) ||
|
||
ps5->post != (unsigned)(-1)) {
|
||
FAILED();
|
||
puts(" Incorrect values read from file");
|
||
goto error;
|
||
}
|
||
}
|
||
}
|
||
}
|
||
PASSED();
|
||
|
||
/*
|
||
*######################################################################
|
||
* Step 11: Write an array into the middle third of the dataset
|
||
* initializeing only members `b' and `d' to -1.
|
||
*/
|
||
TESTING("hyperslab part initialized write");
|
||
|
||
/* Create the memory array and initialize all fields to zero */
|
||
f_offset[0] = NX/3;
|
||
f_offset[1] = NY/3;
|
||
h_size[0] = 2*NX/3 - f_offset[0];
|
||
h_size[1] = 2*NY/3 - f_offset[1];
|
||
s11 = malloc ((size_t)h_size[0]*(size_t)h_size[1]*sizeof(s4_t));
|
||
assert (s11);
|
||
|
||
/* Initialize */
|
||
for (i=0; i<h_size[0]*h_size[1]; i++) {
|
||
s11[i].b = s11[i].d = (unsigned)(-1);
|
||
}
|
||
|
||
/* Write to disk */
|
||
if (H5Dwrite (dataset, s4_tid, s8_m_sid, s8_f_sid, PRESERVE, s11)<0) {
|
||
goto error;
|
||
}
|
||
free (s11);
|
||
s11=NULL;
|
||
|
||
/* Read the whole thing */
|
||
if (H5Dread (dataset, s1_tid, H5S_ALL, H5S_ALL, H5P_DEFAULT, s1)<0) {
|
||
goto error;
|
||
}
|
||
|
||
/* Compare */
|
||
for (i=0; i<NX; i++) {
|
||
for (j=0; j<NY; j++) {
|
||
s1_t *ps1 = s1 + i*NY + j;
|
||
|
||
if (ps1->a != 8*(i*NY+j)+0 ||
|
||
ps1->c[0] != 8*(i*NY+j)+2 ||
|
||
ps1->c[1] != 8*(i*NY+j)+3 ||
|
||
ps1->c[2] != 8*(i*NY+j)+4 ||
|
||
ps1->c[3] != 8*(i*NY+j)+5 ||
|
||
ps1->e != 8*(i*NY+j)+7) {
|
||
FAILED();
|
||
puts(" Write clobbered values");
|
||
goto error;
|
||
}
|
||
|
||
if ((hssize_t)i>=f_offset[0] &&
|
||
(hsize_t)i<f_offset[0]+h_size[0] &&
|
||
(hssize_t)j>=f_offset[1] &&
|
||
(hsize_t)j<f_offset[1]+h_size[1]) {
|
||
if (ps1->b != (unsigned)(-1) ||
|
||
ps1->d != (unsigned)(-1)) {
|
||
FAILED();
|
||
puts(" Wrong values written or read");
|
||
goto error;
|
||
}
|
||
} else {
|
||
if (ps1->b != 8*(i*NY+j)+1 ||
|
||
ps1->d != 8*(i*NY+j)+6) {
|
||
FAILED();
|
||
puts(" Write clobbered values");
|
||
goto error;
|
||
}
|
||
}
|
||
}
|
||
}
|
||
PASSED();
|
||
|
||
|
||
/*
|
||
* Release resources.
|
||
*/
|
||
H5Pclose (PRESERVE);
|
||
H5Dclose (dataset);
|
||
H5Fclose (file);
|
||
|
||
h5_cleanup(fapl);
|
||
puts("All compound dataset tests passed.");
|
||
return 0;
|
||
|
||
error:
|
||
puts("Remaining tests have been skipped.");
|
||
puts("*** DATASET TESTS FAILED ***");
|
||
return 1;
|
||
}
|