mirror of
https://github.com/HDFGroup/hdf5.git
synced 2024-12-21 07:51:46 +08:00
b5eed1b563
* Replaced several uses of sprintf with safer snprintf * Committing clang-format changes Co-authored-by: github-actions <41898282+github-actions[bot]@users.noreply.github.com>
2211 lines
82 KiB
C
2211 lines
82 KiB
C
/* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *
|
|
* Copyright by The HDF Group. *
|
|
* Copyright by the Board of Trustees of the University of Illinois. *
|
|
* All rights reserved. *
|
|
* *
|
|
* This file is part of HDF5. The full HDF5 copyright notice, including *
|
|
* terms governing use, modification, and redistribution, is contained in *
|
|
* the COPYING file, which can be found at the root of the source code *
|
|
* distribution tree, or in https://www.hdfgroup.org/licenses. *
|
|
* If you do not have access to either file, you may request a copy from *
|
|
* help@hdfgroup.org. *
|
|
* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
|
|
/*
|
|
*
|
|
* Purpose: Tests chunk query API functions
|
|
*
|
|
* Modification:
|
|
* Many tests were added for HDFFV-10677. -BMR, August 2019
|
|
*
|
|
* Test structure:
|
|
* main()
|
|
* test_basic_query()
|
|
* test_get_chunk_info_highest_v18()
|
|
* test_get_chunk_info_v110()
|
|
* test_chunk_info_single_chunk()
|
|
* test_chunk_info_implicit()
|
|
* test_chunk_info_fixed_array()
|
|
* test_chunk_info_extensible_array()
|
|
* test_chunk_info_version2_btrees()
|
|
* test_failed_attempts()
|
|
* test_flt_msk_with_skip_compress()
|
|
*
|
|
* Helper functions:
|
|
* verify_idx_nchunks()
|
|
* verify_get_chunk_info()
|
|
* verify_get_chunk_info_by_coord()
|
|
* verify_empty_chunk_info()
|
|
* index_type_str()
|
|
*
|
|
*/
|
|
#define H5D_FRIEND
|
|
#define H5D_TESTING /* to use H5D__ functions */
|
|
#include "H5Dpkg.h"
|
|
|
|
#include "testhdf5.h"
|
|
#ifdef H5_HAVE_FILTER_DEFLATE
|
|
#include "zlib.h"
|
|
#endif
|
|
|
|
/* Test file names, using H5F_libver_t as indices */
|
|
const char *FILENAME[] = {"tchunk_info_earliest", "tchunk_info_v18", "tchunk_info_v110",
|
|
"tchunk_info_v112", "tchunk_info_v114", NULL};
|
|
|
|
/* File to be used in test_failed_attempts */
|
|
#define FILTERMASK_FILE "tflt_msk"
|
|
#define BASIC_FILE "basic_query"
|
|
|
|
/* Parameters for testing chunk querying */
|
|
#define SIMPLE_CHUNKED_DSET_NAME "Chunked Dataset"
|
|
#define CONTIGUOUS_DSET_NAME "Contiguous Dataset"
|
|
#define EMPTY_DSET_NAME "Empty Dataset"
|
|
#define EMPTY_EARLY_ALLOC_DSET_NAME "Empty Dataset with ALLOC_TIME_EARLY"
|
|
#define SINGLE_CHUNK_DSET_NAME "Single Chunk Index Dataset"
|
|
#define IMPLICIT_INDEX_DSET_NAME "Implicit Index Dataset"
|
|
#define FIXED_ARR_INDEX_DSET_NAME "Fixed Array Index Dataset"
|
|
#define EXT_ARR_INDEX_DSET_NAME "Extensible Array Index Dataset"
|
|
#define V2_BTREE_INDEX_DSET_NAME "Version 2 B-Tree Index Dataset"
|
|
#define SKIP_FILTER_DSET_NAME "Dataset with Skipping One Filter"
|
|
#define FILENAME_BUF_SIZE 256 /* Size for file names */
|
|
#define RANK 2 /* Rank for datasets */
|
|
|
|
/* Dimension of the dataset */
|
|
#define NX 24
|
|
#define NY 16
|
|
|
|
/* Dimension of the chunk */
|
|
#define CHUNK_NX 6
|
|
#define CHUNK_NY 4
|
|
|
|
/* X/Y coords of first chunk written */
|
|
#define START_CHK_X 0
|
|
#define START_CHK_Y 2
|
|
|
|
/* X/Y coord of last chunk written */
|
|
#define END_CHK_X 2
|
|
#define END_CHK_Y 4
|
|
|
|
/* X and Y coords of an empty chunk */
|
|
#define EMPTY_CHK_X 0
|
|
#define EMPTY_CHK_Y 0
|
|
|
|
/* Size of a chunk when the entire dataset is a one single chunk */
|
|
#define SINGLE_CHK_SIZE (NX * NY * sizeof(int))
|
|
|
|
/* Size of a chunk */
|
|
#define CHK_SIZE (CHUNK_NX * CHUNK_NY * sizeof(int))
|
|
|
|
/* Size of an empty chunk */
|
|
#define EMPTY_CHK_SIZE 0
|
|
|
|
/* Number of maximum chunks without extending */
|
|
#define NUM_CHUNKS ((NX / CHUNK_NX) * (NY / CHUNK_NY))
|
|
|
|
/* Number of chunks that have been written */
|
|
#define NUM_CHUNKS_WRITTEN 4
|
|
#define ONE_CHUNK_WRITTEN 1
|
|
#define TWO_CHUNKS_WRITTEN 2
|
|
#define NO_CHUNK_WRITTEN 0
|
|
|
|
/* For testing invalid arguments */
|
|
#define NONEXIST_CHK_INDEX 3
|
|
#define OUTOFRANGE_CHK_INDEX 5
|
|
#define INVALID_CHK_INDEX 5
|
|
|
|
/* For compressed data */
|
|
#define DEFLATE_SIZE_ADJUST(s) (ceil(((double)(s)) * 1.001) + 12.0)
|
|
|
|
/* For use in error reporting */
|
|
#define MSG_CHK_ADDR "Chunk address should not be HADDR_UNDEF because of H5D_ALLOC_TIME_EARLY."
|
|
#define MSG_CHK_SIZE "Chunk size should not be 0 because of H5D_ALLOC_TIME_EARLY."
|
|
|
|
/* Utility function to initialize arguments */
|
|
void reinit_vars(unsigned *read_flt_msk, haddr_t *addr, hsize_t *size);
|
|
|
|
/* Helper function containing common code that verifies indexing type
|
|
and number of chunks */
|
|
static int verify_idx_nchunks(hid_t dset, hid_t dspace, H5D_chunk_index_t exp_idx_type,
|
|
hsize_t exp_num_chunks);
|
|
static int verify_get_chunk_info(hid_t dset, hid_t dspace, hsize_t chk_index, hsize_t exp_chk_size,
|
|
const hsize_t *exp_offset, unsigned exp_flt_msk);
|
|
static int verify_get_chunk_info_by_coord(hid_t dset, hsize_t *offset, hsize_t exp_chk_size,
|
|
unsigned exp_flt_msk);
|
|
static int verify_empty_chunk_info(hid_t dset, hsize_t *offset);
|
|
static const char *index_type_str(H5D_chunk_index_t idx_type);
|
|
|
|
/*-------------------------------------------------------------------------
|
|
* Function: reinit_vars (helper function)
|
|
*
|
|
* Purpose: Wipes out variables for the next use, used in various tests.
|
|
*
|
|
* Return: Won't fail
|
|
*
|
|
* Date: September 2018
|
|
*
|
|
*-------------------------------------------------------------------------
|
|
*/
|
|
void
|
|
reinit_vars(unsigned *read_flt_msk, haddr_t *addr, hsize_t *size)
|
|
{
|
|
if (read_flt_msk)
|
|
*read_flt_msk = 0;
|
|
if (addr)
|
|
*addr = 0;
|
|
if (size)
|
|
*size = 0;
|
|
}
|
|
|
|
/*-------------------------------------------------------------------------
|
|
* Function: verify_get_chunk_info (helper function)
|
|
*
|
|
* Purpose: Verifies that H5Dget_chunk_info returns correct
|
|
* values for a chunk.
|
|
*
|
|
* Return: Success: SUCCEED
|
|
* Failure: FAIL
|
|
*
|
|
* Date: August 2019
|
|
*
|
|
*-------------------------------------------------------------------------
|
|
*/
|
|
static int
|
|
verify_get_chunk_info(hid_t dset, hid_t dspace, hsize_t chk_index, hsize_t exp_chk_size,
|
|
const hsize_t *exp_offset, unsigned exp_flt_msk)
|
|
{
|
|
unsigned read_flt_msk = 0; /* Read filter mask */
|
|
hsize_t out_offset[2] = {0, 0}; /* Buffer to get offset coordinates */
|
|
hsize_t size = 0; /* Size of an allocated/written chunk */
|
|
haddr_t addr = 0; /* Address of an allocated/written chunk */
|
|
|
|
if (H5Dget_chunk_info(dset, dspace, chk_index, out_offset, &read_flt_msk, &addr, &size) < 0)
|
|
TEST_ERROR
|
|
CHECK(addr, HADDR_UNDEF, "H5Dget_chunk_info");
|
|
VERIFY(size, exp_chk_size, "H5Dget_chunk_info, chunk size");
|
|
VERIFY(read_flt_msk, exp_flt_msk, "H5Dget_chunk_info, filter mask");
|
|
VERIFY(out_offset[0], exp_offset[0], "H5Dget_chunk_info, offset[0]");
|
|
VERIFY(out_offset[1], exp_offset[1], "H5Dget_chunk_info, offset[1]");
|
|
return SUCCEED;
|
|
|
|
error:
|
|
return FAIL;
|
|
}
|
|
|
|
/*-------------------------------------------------------------------------
|
|
* Function: verify_get_chunk_info_by_coord (helper function)
|
|
*
|
|
* Purpose: Verifies that H5Dget_chunk_info_by_coord returns correct
|
|
* values for a chunk.
|
|
*
|
|
* Return: Success: SUCCEED
|
|
* Failure: FAIL
|
|
*
|
|
* Date: August 2019
|
|
*
|
|
*-------------------------------------------------------------------------
|
|
*/
|
|
static int
|
|
verify_get_chunk_info_by_coord(hid_t dset, hsize_t *offset, hsize_t exp_chk_size, unsigned exp_flt_msk)
|
|
{
|
|
unsigned read_flt_msk = 0; /* Read filter mask */
|
|
hsize_t size = 0; /* Size of an allocated/written chunk */
|
|
haddr_t addr = 0; /* Address of an allocated/written chunk */
|
|
|
|
/* Get info of the chunk at logical coordinates specified by offset */
|
|
if (H5Dget_chunk_info_by_coord(dset, offset, &read_flt_msk, &addr, &size) < 0)
|
|
TEST_ERROR
|
|
CHECK(addr, HADDR_UNDEF, "H5Dget_chunk_info_by_coord");
|
|
VERIFY(size, exp_chk_size, "H5Dget_chunk_info_by_coord, chunk size");
|
|
VERIFY(read_flt_msk, exp_flt_msk, "H5Dget_chunk_info_by_coord, filter mask");
|
|
return SUCCEED;
|
|
|
|
error:
|
|
return FAIL;
|
|
}
|
|
|
|
/*-------------------------------------------------------------------------
|
|
* Function: verify_empty_chunk_info (helper function)
|
|
*
|
|
* Purpose: Verifies that H5Dget_chunk_info_by_coord returns correct
|
|
* values for an empty chunk.
|
|
*
|
|
* Return: Success: SUCCEED
|
|
* Failure: FAIL
|
|
*
|
|
* Date: August 2018
|
|
*
|
|
*-------------------------------------------------------------------------
|
|
*/
|
|
static int
|
|
verify_empty_chunk_info(hid_t dset, hsize_t *offset)
|
|
{
|
|
unsigned read_flt_msk = 0; /* Read filter mask */
|
|
hsize_t size = 0; /* Size of an allocated/written chunk */
|
|
haddr_t addr = 0; /* Address of an allocated/written chunk */
|
|
|
|
/* Get info of the chunk at logical coordinates specified by offset */
|
|
if (H5Dget_chunk_info_by_coord(dset, offset, &read_flt_msk, &addr, &size) < 0)
|
|
TEST_ERROR
|
|
VERIFY(addr, HADDR_UNDEF, "H5Dget_chunk_info_by_coord, chunk address");
|
|
VERIFY(size, EMPTY_CHK_SIZE, "H5Dget_chunk_info_by_coord, chunk size");
|
|
return SUCCEED;
|
|
|
|
error:
|
|
return FAIL;
|
|
}
|
|
|
|
/*-------------------------------------------------------------------------
|
|
* Function: index_type_str (helper function)
|
|
*
|
|
* Purpose: Returns the string containing the text associated with the
|
|
* given indexing scheme. For use in error messages.
|
|
*
|
|
* Return: Success: a valid indexing scheme string
|
|
* Failure: a note indicating the indexing type is invalid
|
|
*
|
|
* Date: August 2019
|
|
*
|
|
*-------------------------------------------------------------------------
|
|
*/
|
|
static const char *
|
|
index_type_str(H5D_chunk_index_t idx_type)
|
|
{
|
|
switch (idx_type) {
|
|
case H5D_CHUNK_IDX_SINGLE:
|
|
return ("Single Chunk index type");
|
|
case H5D_CHUNK_IDX_NONE:
|
|
return ("Implicit index type");
|
|
case H5D_CHUNK_IDX_FARRAY:
|
|
return ("Fixed Array index type");
|
|
case H5D_CHUNK_IDX_EARRAY:
|
|
return ("Extensible Array index type");
|
|
case H5D_CHUNK_IDX_BT2:
|
|
return ("Version 2 B-tree index type");
|
|
case H5D_CHUNK_IDX_BTREE:
|
|
return ("Version 1 B-tree index type (default)");
|
|
case H5D_CHUNK_IDX_NTYPES:
|
|
default:
|
|
return ("invalid index type");
|
|
}
|
|
} /* index_type_str */
|
|
|
|
/*-------------------------------------------------------------------------
|
|
* Function: verify_selected_chunks (helper function)
|
|
*
|
|
* Purpose: Reads the chunks within the boundary {start,end} and verify
|
|
* the values against the populated data.
|
|
*
|
|
* Return: Success: SUCCEED
|
|
* Failure: FAIL
|
|
*
|
|
* Date: August 2019
|
|
*
|
|
*-------------------------------------------------------------------------
|
|
*/
|
|
static int
|
|
verify_selected_chunks(hid_t dset, hid_t plist, const hsize_t *start, const hsize_t *end)
|
|
{
|
|
int read_buf[CHUNK_NX][CHUNK_NY];
|
|
int expected_buf[NUM_CHUNKS][CHUNK_NX][CHUNK_NY]; /* Expected data */
|
|
unsigned read_flt_msk = 0; /* Filter mask read back */
|
|
hsize_t offset[2] = {0, 0}; /* Offset coordinates of a chunk */
|
|
hsize_t chk_index; /* Chunk index */
|
|
hsize_t ii, jj; /* Array indices */
|
|
int n;
|
|
|
|
HDmemset(&read_buf, 0, sizeof(read_buf));
|
|
|
|
/* Initialize the array of chunk data for all NUM_CHUNKS chunks, this is
|
|
the same as the written data and will be used to verify the read data */
|
|
for (n = 0; n < NUM_CHUNKS; n++)
|
|
for (ii = 0; ii < CHUNK_NX; ii++)
|
|
for (jj = 0; jj < CHUNK_NY; jj++)
|
|
expected_buf[n][ii][jj] = (int)(ii * jj) + 1;
|
|
|
|
/* Read each chunk within the boundary of {start,end} and verify the
|
|
values against the expected data */
|
|
chk_index = 0;
|
|
for (ii = start[0]; ii < end[0]; ii++)
|
|
for (jj = start[1]; jj < end[1]; jj++, chk_index++) {
|
|
offset[0] = ii * CHUNK_NX;
|
|
offset[1] = jj * CHUNK_NY;
|
|
|
|
/* Read the current chunk */
|
|
if (H5Dread_chunk(dset, plist, offset, &read_flt_msk, read_buf) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Verify that read chunk is the same as the corresponding written one */
|
|
if (HDmemcmp(expected_buf[chk_index], read_buf, CHUNK_NX * CHUNK_NY) != 0) {
|
|
HDfprintf(stderr,
|
|
"Read chunk differs from written chunk at offset (%" PRIuHSIZE ",%" PRIuHSIZE ")\n",
|
|
offset[0], offset[1]);
|
|
return FAIL;
|
|
}
|
|
}
|
|
|
|
return SUCCEED;
|
|
|
|
error:
|
|
return FAIL;
|
|
} /* verify_selected_chunks */
|
|
|
|
/*-------------------------------------------------------------------------
|
|
* Function: write_selected_chunks (helper function)
|
|
*
|
|
* Purpose: Verifies that chunk indexing scheme and number of chunks of
|
|
* the dataset matches the expected values, then write data to
|
|
* a subset of chunks. This function opens the dataset then
|
|
* closes it after writing.
|
|
*
|
|
* Return: Success: SUCCEED
|
|
* Failure: FAIL
|
|
*
|
|
* Date: August 2019
|
|
*
|
|
*-------------------------------------------------------------------------
|
|
*/
|
|
static int
|
|
write_selected_chunks(hid_t dset, hid_t plist, const hsize_t *start, const hsize_t *end, unsigned flt_msk)
|
|
{
|
|
int direct_buf[NUM_CHUNKS][CHUNK_NX][CHUNK_NY]; /* Data in chunks */
|
|
hsize_t offset[2]; /* Offset coordinates of a chunk */
|
|
hsize_t chk_index; /* Chunk index */
|
|
hsize_t ii, jj; /* Array indices */
|
|
int n;
|
|
|
|
/* Initialize the array of chunk data for all NUM_CHUNKS chunks */
|
|
for (n = 0; n < NUM_CHUNKS; n++)
|
|
for (ii = 0; ii < CHUNK_NX; ii++)
|
|
for (jj = 0; jj < CHUNK_NY; jj++)
|
|
direct_buf[n][ii][jj] = (int)(ii * jj) + 1;
|
|
|
|
/* Write NUM_CHUNKS_WRITTEN chunks at the following logical coords:
|
|
(0,2) (0,3) (1,2) (1,3) */
|
|
chk_index = 0;
|
|
for (ii = start[0]; ii < end[0]; ii++)
|
|
for (jj = start[1]; jj < end[1]; jj++, chk_index++) {
|
|
offset[0] = ii * CHUNK_NX;
|
|
offset[1] = jj * CHUNK_NY;
|
|
if (H5Dwrite_chunk(dset, plist, flt_msk, offset, CHK_SIZE, (void *)direct_buf[chk_index]) < 0)
|
|
TEST_ERROR
|
|
}
|
|
|
|
return SUCCEED;
|
|
|
|
error:
|
|
return FAIL;
|
|
} /* write_selected_chunks */
|
|
|
|
/*-------------------------------------------------------------------------
|
|
* Function: verify_idx_nchunks (helper function)
|
|
*
|
|
* Purpose: Verifies that chunk indexing scheme and number of chunks of
|
|
* the dataset match the expected values.
|
|
*
|
|
* Return: Success: SUCCEED
|
|
* Failure: FAIL
|
|
*
|
|
* Date: August 2019
|
|
*
|
|
*-------------------------------------------------------------------------
|
|
*/
|
|
static int
|
|
verify_idx_nchunks(hid_t dset, hid_t dspace, H5D_chunk_index_t exp_idx_type, hsize_t exp_num_chunks)
|
|
{
|
|
H5D_chunk_index_t idx_type; /* Dataset chunk index type */
|
|
hsize_t nchunks = 0; /* Number of chunks */
|
|
|
|
/* Get the chunk indexing type of the dataset */
|
|
if (H5Dget_chunk_index_type(dset, &idx_type) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Ensure the correct chunk indexing scheme is used */
|
|
if (idx_type != exp_idx_type) {
|
|
char msg[256];
|
|
HDsnprintf(msg, sizeof(msg), "Should be using %s.\n", index_type_str(idx_type));
|
|
FAIL_PUTS_ERROR(msg);
|
|
}
|
|
|
|
/* Get and verify the number of chunks */
|
|
if (H5Dget_num_chunks(dset, dspace, &nchunks) < 0)
|
|
TEST_ERROR
|
|
VERIFY(nchunks, exp_num_chunks, "H5Dget_num_chunks, number of chunks");
|
|
|
|
/* Get and verify the number of chunks again, passing in H5S_ALL */
|
|
if (H5Dget_num_chunks(dset, H5S_ALL, &nchunks) < 0)
|
|
TEST_ERROR
|
|
VERIFY(nchunks, exp_num_chunks, "H5Dget_num_chunks, number of chunks");
|
|
|
|
return SUCCEED;
|
|
|
|
error:
|
|
return FAIL;
|
|
} /* verify_idx_nchunks */
|
|
|
|
/*-------------------------------------------------------------------------
|
|
* Function: test_get_chunk_info_highest_v18
|
|
*
|
|
* Purpose: Test getting various chunk information
|
|
*
|
|
* Return: Success: SUCCEED
|
|
* Failure: FAIL
|
|
*
|
|
* Note: Note that the dataspace argument in these new functions is
|
|
* currently not used. The functionality involved the dataspace
|
|
* will be implemented in the next version.
|
|
*
|
|
* Description:
|
|
* This function tests the new API functions added for EED-343:
|
|
* H5Dget_num_chunks, H5Dget_chunk_info, and
|
|
* H5Dget_chunk_info_by_coord for high bound up to 1.8.
|
|
*
|
|
* Date: September 2018
|
|
*
|
|
*-------------------------------------------------------------------------
|
|
*/
|
|
static herr_t
|
|
test_get_chunk_info_highest_v18(hid_t fapl)
|
|
{
|
|
char filename[FILENAME_BUF_SIZE]; /* File name */
|
|
hid_t chunkfile = H5I_INVALID_HID; /* File ID */
|
|
hid_t dspace = H5I_INVALID_HID; /* Dataspace ID */
|
|
hid_t dset = H5I_INVALID_HID; /* Dataset ID */
|
|
hid_t cparms = H5I_INVALID_HID; /* Creation plist */
|
|
hsize_t chunk_dims[2] = {CHUNK_NX, CHUNK_NY}; /* Chunk dimensions */
|
|
int direct_buf[CHUNK_NX][CHUNK_NY]; /* Data chunk */
|
|
hsize_t maxdims[2] = {H5S_UNLIMITED, H5S_UNLIMITED}; /* 2 unlimited dims */
|
|
hsize_t out_offset[2]; /* Buffer to get offset coordinates */
|
|
hsize_t size = 0; /* Size of an allocated/written chunk */
|
|
hsize_t nchunks = 0; /* Number of chunks */
|
|
haddr_t addr = 0; /* Address of an allocated/written chunk */
|
|
hsize_t chk_index = 0; /* Index of a chunk */
|
|
hsize_t dims[2] = {NX, NY}; /* Dataset dimensions */
|
|
unsigned flt_msk = 0; /* Filter mask */
|
|
unsigned read_flt_msk = 0; /* Filter mask after direct read */
|
|
int fillvalue = -1; /* Fill value */
|
|
hsize_t offset[2] = {0, 0}; /* Offset coordinates of a chunk */
|
|
#ifdef H5_HAVE_FILTER_DEFLATE
|
|
int aggression = 9; /* Compression aggression setting */
|
|
const Bytef *z_src = (const Bytef *)(direct_buf);
|
|
Bytef * z_dst; /* Destination buffer */
|
|
uLongf z_dst_nbytes = (uLongf)DEFLATE_SIZE_ADJUST(CHK_SIZE);
|
|
uLong z_src_nbytes = (uLong)CHK_SIZE;
|
|
#endif
|
|
void * inbuf = NULL; /* Pointer to new buffer */
|
|
hsize_t chunk_size = CHK_SIZE; /* Size of a chunk, can be compressed or not */
|
|
hsize_t ii, jj; /* Array indices */
|
|
int n; /* Used as chunk index, but int to avoid conversion warning */
|
|
herr_t ret; /* Temporary returned value for verifying failure */
|
|
|
|
TESTING("getting chunk information in file with version prior to 1.10");
|
|
|
|
/* Create the file */
|
|
h5_fixname(FILENAME[H5F_LIBVER_V18], fapl, filename, sizeof filename);
|
|
|
|
/* Set version bounds for creating the file. High bound to V18 to test
|
|
chunked dataset that use B-tree v1 structures to index chunks. */
|
|
if (H5Pset_libver_bounds(fapl, H5F_LIBVER_EARLIEST, H5F_LIBVER_V18) < 0)
|
|
TEST_ERROR
|
|
|
|
chunkfile = H5Fcreate(filename, H5F_ACC_TRUNC, H5P_DEFAULT, fapl);
|
|
if (chunkfile < 0)
|
|
TEST_ERROR
|
|
|
|
/* Create the file and memory dataspace */
|
|
if ((dspace = H5Screate_simple(RANK, dims, maxdims)) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Set dset creation properties with chunking, compression, and fillvalue */
|
|
if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
|
|
TEST_ERROR
|
|
if (H5Pset_chunk(cparms, RANK, chunk_dims) < 0)
|
|
TEST_ERROR
|
|
|
|
#ifdef H5_HAVE_FILTER_DEFLATE
|
|
if (H5Pset_deflate(cparms, (unsigned)aggression) < 0)
|
|
TEST_ERROR
|
|
#endif /* end H5_HAVE_FILTER_DEFLATE */
|
|
|
|
/* Set fill value */
|
|
if (H5Pset_fill_value(cparms, H5T_NATIVE_INT, &fillvalue) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Create a new dataset using cparms creation properties */
|
|
dset = H5Dcreate2(chunkfile, SIMPLE_CHUNKED_DSET_NAME, H5T_NATIVE_INT, dspace, H5P_DEFAULT, cparms,
|
|
H5P_DEFAULT);
|
|
if (dset < 0)
|
|
TEST_ERROR
|
|
|
|
/* Initialize a chunk of data */
|
|
for (ii = 0; ii < CHUNK_NX; ii++)
|
|
for (jj = 0; jj < CHUNK_NY; jj++)
|
|
direct_buf[ii][jj] = (int)(ii * jj) + 1;
|
|
|
|
#ifdef H5_HAVE_FILTER_DEFLATE
|
|
/* Allocate input (compressed) buffer */
|
|
inbuf = HDcalloc(1, z_dst_nbytes);
|
|
|
|
/* zlib-friendly alias for the input buffer */
|
|
z_dst = (Bytef *)inbuf;
|
|
|
|
/* Perform compression from the source to the destination buffer */
|
|
ret = compress2(z_dst, &z_dst_nbytes, z_src, z_src_nbytes, aggression);
|
|
|
|
/* Set the chunk size to the compressed chunk size */
|
|
chunk_size = (hsize_t)z_dst_nbytes;
|
|
|
|
/* Check for various zlib errors */
|
|
if (Z_BUF_ERROR == ret) {
|
|
HDfprintf(stderr, "overflow");
|
|
TEST_ERROR
|
|
}
|
|
else if (Z_MEM_ERROR == ret) {
|
|
HDfprintf(stderr, "deflate memory error");
|
|
TEST_ERROR
|
|
}
|
|
else if (Z_OK != ret) {
|
|
HDfprintf(stderr, "other deflate error");
|
|
TEST_ERROR
|
|
}
|
|
#else
|
|
/* Allocate input (non-compressed) buffer */
|
|
if (NULL == (inbuf = HDcalloc(1, CHK_SIZE)))
|
|
TEST_ERROR
|
|
HDmemcpy(inbuf, direct_buf, CHK_SIZE);
|
|
#endif /* end H5_HAVE_FILTER_DEFLATE */
|
|
|
|
/* Write only NUM_CHUNKS_WRITTEN chunks at the following logical coords:
|
|
(0,2) (0,3) (1,2) (1,3) */
|
|
n = 0;
|
|
for (ii = START_CHK_X; ii < END_CHK_X; ii++)
|
|
for (jj = START_CHK_Y; jj < END_CHK_Y; jj++, n++) {
|
|
offset[0] = ii * CHUNK_NX;
|
|
offset[1] = jj * CHUNK_NY;
|
|
ret = H5Dwrite_chunk(dset, H5P_DEFAULT, flt_msk, offset, chunk_size, (void *)inbuf);
|
|
if (ret < 0)
|
|
TEST_ERROR
|
|
}
|
|
|
|
/* Free the input buffer */
|
|
if (inbuf)
|
|
HDfree(inbuf);
|
|
|
|
if (H5Fflush(dset, H5F_SCOPE_LOCAL) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Close the dataset */
|
|
if (H5Dclose(dset) < 0)
|
|
TEST_ERROR
|
|
|
|
/* ...open it again to test the chunk query functions */
|
|
if ((dset = H5Dopen2(chunkfile, SIMPLE_CHUNKED_DSET_NAME, H5P_DEFAULT)) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Get and verify the number of chunks written */
|
|
if (H5Dget_num_chunks(dset, dspace, &nchunks) < 0)
|
|
TEST_ERROR
|
|
VERIFY(nchunks, NUM_CHUNKS_WRITTEN, "H5Dget_num_chunks, number of chunks");
|
|
|
|
/* Get and verify info of the last written chunk again, passing in H5S_ALL
|
|
this time */
|
|
offset[0] = 6;
|
|
offset[1] = 12;
|
|
if (verify_get_chunk_info(dset, H5S_ALL, NUM_CHUNKS_WRITTEN - 1, chunk_size, offset, flt_msk) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info failed\n");
|
|
|
|
/* Attempt to get info of a non-existing chunk, should fail */
|
|
chk_index = OUTOFRANGE_CHK_INDEX;
|
|
H5E_BEGIN_TRY
|
|
{
|
|
ret = H5Dget_chunk_info(dset, H5S_ALL, chk_index, out_offset, &read_flt_msk, &addr, &size);
|
|
}
|
|
H5E_END_TRY;
|
|
if (ret != FAIL)
|
|
FAIL_PUTS_ERROR(" Attempt to get info of a non-existing chunk.");
|
|
|
|
/* Attempt to get info of empty chunks, verify the returned addr and size */
|
|
offset[0] = 0;
|
|
offset[1] = 0;
|
|
if (verify_empty_chunk_info(dset, offset) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info_by_coord on empty chunk failed\n");
|
|
|
|
offset[0] = 3 * CHUNK_NX;
|
|
offset[1] = 3 * CHUNK_NY;
|
|
if (verify_empty_chunk_info(dset, offset) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info_by_coord on empty chunk failed\n");
|
|
|
|
/* Go through all written chunks, get their info and verify the values */
|
|
chk_index = 0;
|
|
for (ii = START_CHK_X; ii < END_CHK_X; ii++)
|
|
for (jj = START_CHK_Y; jj < END_CHK_Y; jj++, chk_index++) {
|
|
offset[0] = ii * CHUNK_NX;
|
|
offset[1] = jj * CHUNK_NY;
|
|
|
|
if (verify_get_chunk_info(dset, dspace, chk_index, chunk_size, offset, flt_msk) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info failed\n");
|
|
|
|
/* Use the same offset to pass into the next ...by_coord function */
|
|
if (verify_get_chunk_info_by_coord(dset, offset, chunk_size, flt_msk) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info_by_coord failed\n");
|
|
}
|
|
|
|
/* Close the first dataset */
|
|
if (H5Dclose(dset) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Create an empty dataset and close it */
|
|
dset = H5Dcreate2(chunkfile, EMPTY_DSET_NAME, H5T_NATIVE_INT, dspace, H5P_DEFAULT, cparms, H5P_DEFAULT);
|
|
if (dset < 0)
|
|
TEST_ERROR
|
|
if (H5Dclose(dset) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Reopen the empty dataset to verify the chunk query functions on it */
|
|
if ((dset = H5Dopen2(chunkfile, EMPTY_DSET_NAME, H5P_DEFAULT)) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Verify that the number of chunks is 0 */
|
|
if (H5Dget_num_chunks(dset, dspace, &nchunks) < 0)
|
|
TEST_ERROR
|
|
VERIFY(nchunks, NO_CHUNK_WRITTEN, "H5Dget_num_chunks, number of chunks");
|
|
|
|
/* Attempt to get info of a chunk from an empty dataset, should fail */
|
|
chk_index = OUTOFRANGE_CHK_INDEX;
|
|
H5E_BEGIN_TRY
|
|
{
|
|
ret = H5Dget_chunk_info(dset, dspace, chk_index, out_offset, &read_flt_msk, &addr, &size);
|
|
}
|
|
H5E_END_TRY;
|
|
if (ret != FAIL)
|
|
FAIL_PUTS_ERROR(" Attempt to get info of a non-existing chunk.");
|
|
|
|
/* Attempt to get info of a chunk given its coords from an empty dataset,
|
|
should succeed with the returned address as HADDR_UNDEF and size as 0 */
|
|
offset[0] = EMPTY_CHK_X;
|
|
offset[1] = EMPTY_CHK_Y;
|
|
if (verify_empty_chunk_info(dset, offset) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info_by_coord on empty chunk failed\n");
|
|
|
|
if (H5Dclose(dset) < 0)
|
|
TEST_ERROR
|
|
|
|
/************************************************************************
|
|
* Test empty dataset with H5D_ALLOC_TIME_EARLY *
|
|
************************************************************************/
|
|
|
|
/* Set space allocation to early so that chunk query functions will
|
|
retrieve chunk information even though the dataset is empty */
|
|
if (H5Pset_alloc_time(cparms, H5D_ALLOC_TIME_EARLY) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Create an empty dataset and close it */
|
|
dset = H5Dcreate2(chunkfile, EMPTY_EARLY_ALLOC_DSET_NAME, H5T_NATIVE_INT, dspace, H5P_DEFAULT, cparms,
|
|
H5P_DEFAULT);
|
|
if (dset < 0)
|
|
TEST_ERROR
|
|
if (H5Dclose(dset) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Reopen the empty dataset to verify the chunk query functions on it */
|
|
if ((dset = H5Dopen2(chunkfile, EMPTY_EARLY_ALLOC_DSET_NAME, H5P_DEFAULT)) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Verify that the number of chunks is NUM_CHUNKS */
|
|
if (H5Dget_num_chunks(dset, dspace, &nchunks) < 0)
|
|
TEST_ERROR
|
|
if (nchunks != NUM_CHUNKS)
|
|
TEST_ERROR
|
|
|
|
/* Attempt to get info of a chunk from an empty dataset, verify the
|
|
returned address and size in the case of H5D_ALLOC_TIME_EARLY */
|
|
chk_index = NONEXIST_CHK_INDEX;
|
|
reinit_vars(&read_flt_msk, &addr, &size);
|
|
ret = H5Dget_chunk_info(dset, dspace, chk_index, out_offset, &read_flt_msk, &addr, &size);
|
|
if (ret < 0)
|
|
TEST_ERROR
|
|
|
|
/* Because of H5D_ALLOC_TIME_EARLY, addr cannot be HADDR_UNDEF and size not 0 */
|
|
if (addr == HADDR_UNDEF)
|
|
TEST_ERROR
|
|
if (size == EMPTY_CHK_SIZE)
|
|
TEST_ERROR
|
|
|
|
chk_index = 10;
|
|
reinit_vars(&read_flt_msk, &addr, &size);
|
|
ret = H5Dget_chunk_info(dset, dspace, chk_index, out_offset, &read_flt_msk, &addr, &size);
|
|
if (ret < 0)
|
|
TEST_ERROR
|
|
/* Because of H5D_ALLOC_TIME_EARLY, addr cannot be HADDR_UNDEF and size not 0 */
|
|
if (addr == HADDR_UNDEF)
|
|
TEST_ERROR
|
|
if (size == EMPTY_CHK_SIZE)
|
|
TEST_ERROR
|
|
|
|
/* Attempt to get info of a chunk given its coords from an empty dataset,
|
|
verify the returned address and size */
|
|
offset[0] = 0;
|
|
offset[1] = 0;
|
|
if (H5Dget_chunk_info_by_coord(dset, offset, &read_flt_msk, &addr, &size) < 0)
|
|
TEST_ERROR
|
|
/* Because of H5D_ALLOC_TIME_EARLY, addr cannot be HADDR_UNDEF and size not 0 */
|
|
if (addr == HADDR_UNDEF)
|
|
TEST_ERROR
|
|
if (size == 0)
|
|
TEST_ERROR
|
|
|
|
if (H5Dclose(dset) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Close/release resources. */
|
|
if (H5Sclose(dspace) < 0)
|
|
TEST_ERROR
|
|
if (H5Pclose(cparms) < 0)
|
|
TEST_ERROR
|
|
if (H5Fclose(chunkfile) < 0)
|
|
TEST_ERROR
|
|
|
|
PASSED();
|
|
return SUCCEED;
|
|
|
|
error:
|
|
H5E_BEGIN_TRY
|
|
{
|
|
H5Dclose(dset);
|
|
H5Sclose(dspace);
|
|
H5Pclose(cparms);
|
|
H5Fclose(chunkfile);
|
|
}
|
|
H5E_END_TRY;
|
|
|
|
H5_FAILED();
|
|
return FAIL;
|
|
} /* test_get_chunk_info_highest_v18() */
|
|
|
|
/*-------------------------------------------------------------------------
|
|
* Function: test_chunk_info_single_chunk
|
|
*
|
|
* Purpose: Test getting various chunk information when Single Chunk
|
|
* index type is used
|
|
*
|
|
* Return: Success: SUCCEED
|
|
* Failure: FAIL
|
|
*
|
|
* Note: Note that the dataspace argument in these new functions are
|
|
* currently not used. The functionality involved the dataspace
|
|
* will be implemented in the next version.
|
|
*
|
|
* Date: November 2018
|
|
*
|
|
*-------------------------------------------------------------------------
|
|
*/
|
|
static herr_t
|
|
test_chunk_info_single_chunk(const char *filename, hid_t fapl)
|
|
{
|
|
hid_t chunkfile = H5I_INVALID_HID; /* File ID */
|
|
hid_t dspace = H5I_INVALID_HID; /* Dataspace ID */
|
|
hid_t dset = H5I_INVALID_HID; /* Dataset ID */
|
|
hid_t cparms = H5I_INVALID_HID; /* Creation plist */
|
|
hsize_t dims[2] = {NX, NY}; /* Dataset dimensions */
|
|
hsize_t chunk_dims[2] = {NX, NY}; /* Chunk dimensions */
|
|
int data_buf[NX][NY]; /* Input buffer */
|
|
H5D_chunk_index_t idx_type; /* Dataset chunk index type */
|
|
unsigned flt_msk = 0; /* Filter mask */
|
|
unsigned read_flt_msk = 0; /* Filter mask after direct read */
|
|
hsize_t offset[2]; /* Offset coordinates of a chunk */
|
|
hsize_t out_offset[2] = {0, 0}; /* Buffer to get offset coordinates */
|
|
hsize_t size = 0; /* Size of an allocated/written chunk */
|
|
hsize_t nchunks = 0; /* Number of chunks */
|
|
haddr_t addr = 0; /* Address of an allocated/written chunk */
|
|
hsize_t chk_index = 0; /* Index of a chunk */
|
|
hsize_t ii, jj; /* Array indices */
|
|
herr_t ret; /* Temporary returned value for verifying failure */
|
|
|
|
TESTING(" Single Chunk index");
|
|
|
|
/* Open the file for reading/writing */
|
|
if ((chunkfile = H5Fopen(filename, H5F_ACC_RDWR, fapl)) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Create dataspace */
|
|
if ((dspace = H5Screate_simple(RANK, dims, NULL)) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Enable chunking */
|
|
if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
|
|
TEST_ERROR
|
|
|
|
if (H5Pset_chunk(cparms, RANK, chunk_dims) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Create a new dataset using cparms creation properties */
|
|
dset = H5Dcreate2(chunkfile, SINGLE_CHUNK_DSET_NAME, H5T_NATIVE_INT, dspace, H5P_DEFAULT, cparms,
|
|
H5P_DEFAULT);
|
|
if (dset < 0)
|
|
TEST_ERROR
|
|
|
|
/* Close the dataset */
|
|
if (H5Dclose(dset) < 0)
|
|
TEST_ERROR
|
|
|
|
/* ...open it again to test the chunk query functions on a single empty
|
|
chunk */
|
|
if ((dset = H5Dopen2(chunkfile, SINGLE_CHUNK_DSET_NAME, H5P_DEFAULT)) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Ensure the correct chunk indexing scheme is used */
|
|
if (H5Dget_chunk_index_type(dset, &idx_type) < 0)
|
|
TEST_ERROR
|
|
if (idx_type != H5D_CHUNK_IDX_SINGLE)
|
|
FAIL_PUTS_ERROR("Should be using Single Chunk index type");
|
|
|
|
/* Get the number of chunks and verify that no chunk has been written */
|
|
if (H5Dget_num_chunks(dset, dspace, &nchunks) < 0)
|
|
TEST_ERROR
|
|
if (nchunks != NO_CHUNK_WRITTEN)
|
|
TEST_ERROR
|
|
|
|
/* Initialize the array of chunk data for the single chunk */
|
|
for (ii = 0; ii < NX; ii++)
|
|
for (jj = 0; jj < NY; jj++)
|
|
data_buf[ii][jj] = (int)(ii * jj);
|
|
|
|
/* Write the chunk */
|
|
if (H5Dwrite(dset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, H5P_DEFAULT, data_buf) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Get and verify that one chunk had been written */
|
|
if (H5Dget_num_chunks(dset, dspace, &nchunks) < 0)
|
|
TEST_ERROR
|
|
if (nchunks != ONE_CHUNK_WRITTEN)
|
|
TEST_ERROR
|
|
|
|
/* Offset of the only chunk */
|
|
offset[0] = 0;
|
|
offset[1] = 0;
|
|
|
|
/* Get and verify info of the first and only chunk */
|
|
if (verify_get_chunk_info(dset, H5S_ALL, 0, SINGLE_CHK_SIZE, offset, flt_msk) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification H5Dget_chunk_info failed\n");
|
|
|
|
/* Get and verify info of the chunk at logical coordinates (0,0) */
|
|
if (verify_get_chunk_info_by_coord(dset, offset, SINGLE_CHK_SIZE, flt_msk) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info_by_coord failed\n");
|
|
|
|
/* Attempt to get chunk info given an invalid chunk index and verify
|
|
* that failure occurs */
|
|
chk_index = INVALID_CHK_INDEX;
|
|
reinit_vars(&read_flt_msk, &addr, &size);
|
|
H5E_BEGIN_TRY
|
|
{
|
|
ret = H5Dget_chunk_info(dset, dspace, chk_index, out_offset, &read_flt_msk, &addr, &size);
|
|
}
|
|
H5E_END_TRY;
|
|
if (ret != FAIL)
|
|
TEST_ERROR
|
|
|
|
/* Release resourse */
|
|
if (H5Dclose(dset) < 0)
|
|
TEST_ERROR
|
|
if (H5Sclose(dspace) < 0)
|
|
TEST_ERROR
|
|
if (H5Fclose(chunkfile) < 0)
|
|
TEST_ERROR
|
|
|
|
PASSED();
|
|
return SUCCEED;
|
|
|
|
error:
|
|
H5E_BEGIN_TRY
|
|
{
|
|
H5Dclose(dset);
|
|
H5Sclose(dspace);
|
|
H5Pclose(cparms);
|
|
H5Fclose(chunkfile);
|
|
}
|
|
H5E_END_TRY;
|
|
|
|
H5_FAILED();
|
|
return FAIL;
|
|
} /* test_chunk_info_single_chunk() */
|
|
|
|
/*-------------------------------------------------------------------------
|
|
* Function: test_chunk_info_implicit
|
|
*
|
|
* Purpose: Test getting various chunk information when Implicit
|
|
* index type is used
|
|
*
|
|
* Return: Success: SUCCEED
|
|
* Failure: FAIL
|
|
*
|
|
* Note: Note that the dataspace argument in these new functions are
|
|
* currently not used. The functionality involved the dataspace
|
|
* will be implemented in the next version.
|
|
*
|
|
* Date: November 2018
|
|
*
|
|
*-------------------------------------------------------------------------
|
|
*/
|
|
static herr_t
|
|
test_chunk_info_implicit(char *filename, hid_t fapl)
|
|
{
|
|
hid_t chunkfile = H5I_INVALID_HID; /* File ID */
|
|
hid_t dspace = H5I_INVALID_HID; /* Dataspace ID */
|
|
hid_t dset = H5I_INVALID_HID; /* Dataset ID */
|
|
hid_t cparms = H5I_INVALID_HID; /* Creation plist */
|
|
hsize_t dims[2] = {NX, NY}; /* Dataset dimensions */
|
|
hsize_t chunk_dims[2] = {CHUNK_NX, CHUNK_NY}; /* Chunk dimensions */
|
|
unsigned flt_msk = 0; /* Filter mask */
|
|
hsize_t chk_index = 0; /* Index of a chunk */
|
|
hsize_t ii, jj; /* Array indices */
|
|
hsize_t start[2] = {START_CHK_X, START_CHK_Y}; /* Start position */
|
|
hsize_t end[2] = {END_CHK_X, END_CHK_Y}; /* End position */
|
|
|
|
TESTING(" Implicit index");
|
|
|
|
/* Open the file for reading/writing */
|
|
if ((chunkfile = H5Fopen(filename, H5F_ACC_RDWR, fapl)) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Create dataspace */
|
|
if ((dspace = H5Screate_simple(RANK, dims, NULL)) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Enable chunking */
|
|
if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
|
|
TEST_ERROR
|
|
|
|
if (H5Pset_chunk(cparms, RANK, chunk_dims) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Set allocation time to early */
|
|
if (H5Pset_alloc_time(cparms, H5D_ALLOC_TIME_EARLY) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Create a new dataset using cparms creation properties */
|
|
dset = H5Dcreate2(chunkfile, IMPLICIT_INDEX_DSET_NAME, H5T_NATIVE_INT, dspace, H5P_DEFAULT, cparms,
|
|
H5P_DEFAULT);
|
|
if (dset < 0)
|
|
TEST_ERROR
|
|
|
|
/* Close the dataset */
|
|
if (H5Dclose(dset) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Open the dataset again to test getting chunk info */
|
|
if ((dset = H5Dopen2(chunkfile, IMPLICIT_INDEX_DSET_NAME, H5P_DEFAULT)) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Verify chunk indexing scheme and number of chunks */
|
|
if (verify_idx_nchunks(dset, dspace, H5D_CHUNK_IDX_NONE, NUM_CHUNKS) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification and write failed\n");
|
|
|
|
/* Write NUM_CHUNKS_WRITTEN chunks at the following logical coords:
|
|
(0,2) (0,3) (1,2) (1,3) */
|
|
if (write_selected_chunks(dset, H5P_DEFAULT, start, end, flt_msk) == FAIL)
|
|
FAIL_PUTS_ERROR("Writing to selected chunks failed\n");
|
|
|
|
/* Go through all chunks, and get their info and verify the values */
|
|
chk_index = 0;
|
|
for (ii = 0; ii < NX / CHUNK_NX; ii++)
|
|
for (jj = 0; jj < NY / CHUNK_NY; jj++, chk_index++) {
|
|
hsize_t offset[2] = {ii * CHUNK_NX, jj * CHUNK_NY};
|
|
|
|
if (verify_get_chunk_info(dset, H5S_ALL, chk_index, CHK_SIZE, offset, flt_msk) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info failed\n");
|
|
|
|
/* Get info of a chunk and verify its information. Note that
|
|
all chunks in this dataset are allocated because of the property
|
|
H5D_ALLOC_TIME_EARLY */
|
|
if (verify_get_chunk_info_by_coord(dset, offset, CHK_SIZE, flt_msk) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info_by_coord failed\n");
|
|
}
|
|
|
|
/* Release resourse */
|
|
if (H5Dclose(dset) < 0)
|
|
TEST_ERROR
|
|
if (H5Sclose(dspace) < 0)
|
|
TEST_ERROR
|
|
if (H5Pclose(cparms) < 0)
|
|
TEST_ERROR
|
|
if (H5Fclose(chunkfile) < 0)
|
|
TEST_ERROR
|
|
|
|
PASSED();
|
|
return SUCCEED;
|
|
|
|
error:
|
|
H5E_BEGIN_TRY
|
|
{
|
|
H5Dclose(dset);
|
|
H5Sclose(dspace);
|
|
H5Pclose(cparms);
|
|
H5Fclose(chunkfile);
|
|
}
|
|
H5E_END_TRY;
|
|
|
|
H5_FAILED();
|
|
return FAIL;
|
|
} /* test_chunk_info_implicit() */
|
|
|
|
/*-------------------------------------------------------------------------
|
|
* Function: test_chunk_info_fixed_array
|
|
*
|
|
* Purpose: Test getting various chunk information when Fixed Array
|
|
* index type is used
|
|
*
|
|
* Return: Success: SUCCEED
|
|
* Failure: FAIL
|
|
*
|
|
* Note: Note that the dataspace argument in these new functions are
|
|
* currently not used. The functionality involved the dataspace
|
|
* will be implemented in the next version.
|
|
*
|
|
* Date: November 2018
|
|
*
|
|
*-------------------------------------------------------------------------
|
|
*/
|
|
static herr_t
|
|
test_chunk_info_fixed_array(const char *filename, hid_t fapl)
|
|
{
|
|
hid_t chunkfile = H5I_INVALID_HID; /* File ID */
|
|
hid_t dspace = H5I_INVALID_HID; /* Dataspace ID */
|
|
hid_t dset = H5I_INVALID_HID; /* Dataset ID */
|
|
hid_t cparms = H5I_INVALID_HID; /* Creation plist */
|
|
hsize_t dims[2] = {NX, NY}; /* Dataset dimensions */
|
|
hsize_t chunk_dims[2] = {CHUNK_NX, CHUNK_NY}; /* Chunk dimensions */
|
|
unsigned flt_msk = 0; /* Filter mask */
|
|
unsigned read_flt_msk = 0; /* Filter mask after direct read */
|
|
hsize_t offset[2]; /* Offset coordinates of a chunk */
|
|
hsize_t start[2] = {START_CHK_X, START_CHK_Y}; /* Start position */
|
|
hsize_t end[2] = {END_CHK_X, END_CHK_Y}; /* End position */
|
|
hsize_t out_offset[2] = {0, 0}; /* Buffer to get offset coordinates */
|
|
hsize_t size = 0; /* Size of an allocated/written chunk */
|
|
hsize_t nchunks = 0; /* Number of chunks */
|
|
haddr_t addr = 0; /* Address of an allocated/written chunk */
|
|
hsize_t chk_index = 0; /* Index of a chunk */
|
|
hsize_t ii, jj; /* Array indices */
|
|
herr_t ret; /* Temporary returned value for verifying failure */
|
|
|
|
TESTING(" Fixed Array index");
|
|
|
|
/* Open the file for reading/writing */
|
|
if ((chunkfile = H5Fopen(filename, H5F_ACC_RDWR, fapl)) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Create dataspace */
|
|
if ((dspace = H5Screate_simple(RANK, dims, NULL)) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Enable chunking */
|
|
if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
|
|
TEST_ERROR
|
|
|
|
if (H5Pset_chunk(cparms, RANK, chunk_dims) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Create a new dataset using cparms creation properties */
|
|
dset = H5Dcreate2(chunkfile, FIXED_ARR_INDEX_DSET_NAME, H5T_NATIVE_INT, dspace, H5P_DEFAULT, cparms,
|
|
H5P_DEFAULT);
|
|
if (dset < 0)
|
|
TEST_ERROR
|
|
|
|
/* Close the dataset */
|
|
if (H5Dclose(dset) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Open the dataset again to test getting chunk info */
|
|
if ((dset = H5Dopen2(chunkfile, FIXED_ARR_INDEX_DSET_NAME, H5P_DEFAULT)) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Verify chunk indexing scheme and number of chunks */
|
|
if (verify_idx_nchunks(dset, dspace, H5D_CHUNK_IDX_FARRAY, NO_CHUNK_WRITTEN) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification and write failed\n");
|
|
|
|
/* Write NUM_CHUNKS_WRITTEN chunks at the following logical coords:
|
|
(0,2) (0,3) (1,2) (1,3) */
|
|
if (write_selected_chunks(dset, H5P_DEFAULT, start, end, flt_msk) == FAIL)
|
|
FAIL_PUTS_ERROR("Writing to selected chunks failed\n");
|
|
|
|
/* Get and verify the number of chunks written */
|
|
if (H5Dget_num_chunks(dset, dspace, &nchunks) < 0)
|
|
TEST_ERROR
|
|
VERIFY(nchunks, NUM_CHUNKS_WRITTEN, "H5Dget_num_chunks, number of chunks");
|
|
|
|
/* Get and verify info of each written chunk */
|
|
chk_index = 0;
|
|
for (ii = START_CHK_X; ii < END_CHK_X; ii++)
|
|
for (jj = START_CHK_Y; jj < END_CHK_Y; jj++, chk_index++) {
|
|
offset[0] = ii * CHUNK_NX;
|
|
offset[1] = jj * CHUNK_NY;
|
|
if (verify_get_chunk_info(dset, dspace, chk_index, CHK_SIZE, offset, flt_msk) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info failed\n");
|
|
}
|
|
|
|
/* Attempt to get info using an out-of-range index, chk_index is now > NUM_CHUNKS_WRITTEN. should fail */
|
|
H5E_BEGIN_TRY
|
|
{
|
|
ret = H5Dget_chunk_info(dset, dspace, chk_index, out_offset, &read_flt_msk, &addr, &size);
|
|
}
|
|
H5E_END_TRY;
|
|
if (ret != FAIL)
|
|
FAIL_PUTS_ERROR(" Attempted to get info of a chunk using an out-of-range index.");
|
|
|
|
/* Attempt to get info of empty chunks, verify the returned address and size */
|
|
offset[0] = 0;
|
|
offset[1] = 0;
|
|
if (verify_empty_chunk_info(dset, offset) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info_by_coord on empty chunk failed\n");
|
|
|
|
offset[0] = 3 * CHUNK_NX;
|
|
offset[1] = 3 * CHUNK_NY;
|
|
if (verify_empty_chunk_info(dset, offset) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info_by_coord on empty chunk failed\n");
|
|
|
|
/* Read and verify values of selected chunks */
|
|
if (verify_selected_chunks(dset, H5P_DEFAULT, start, end) < 0)
|
|
|
|
/* Release resourse */
|
|
if (H5Dclose(dset) < 0)
|
|
TEST_ERROR
|
|
if (H5Sclose(dspace) < 0)
|
|
TEST_ERROR
|
|
if (H5Fclose(chunkfile) < 0)
|
|
TEST_ERROR
|
|
|
|
PASSED();
|
|
return SUCCEED;
|
|
|
|
error:
|
|
H5E_BEGIN_TRY
|
|
{
|
|
H5Dclose(dset);
|
|
H5Sclose(dspace);
|
|
H5Pclose(cparms);
|
|
H5Fclose(chunkfile);
|
|
}
|
|
H5E_END_TRY;
|
|
|
|
H5_FAILED();
|
|
return FAIL;
|
|
} /* test_chunk_info_fixed_array() */
|
|
|
|
/*-------------------------------------------------------------------------
|
|
* Function: test_chunk_info_extensible_array
|
|
*
|
|
* Purpose: Test getting various chunk information when Extensible Array
|
|
* index type is used
|
|
*
|
|
* Return: Success: SUCCEED
|
|
* Failure: FAIL
|
|
*
|
|
* Note: Note that the dataspace argument in these new functions are
|
|
* currently not used. The functionality involved the dataspace
|
|
* will be implemented in the next version.
|
|
*
|
|
* Date: November 2018
|
|
*
|
|
*-------------------------------------------------------------------------
|
|
*/
|
|
static herr_t
|
|
test_chunk_info_extensible_array(const char *filename, hid_t fapl)
|
|
{
|
|
hid_t chunkfile = H5I_INVALID_HID; /* File ID */
|
|
hid_t dspace = H5I_INVALID_HID; /* Dataspace ID */
|
|
hid_t dset = H5I_INVALID_HID; /* Dataset ID */
|
|
hid_t cparms = H5I_INVALID_HID; /* Creation plist */
|
|
hsize_t dims[2] = {NX, NY}; /* Dataset dimensions */
|
|
hsize_t chunk_dims[2] = {CHUNK_NX, CHUNK_NY}; /* Chunk dimensions */
|
|
hsize_t maxdims[2] = {H5S_UNLIMITED, NY}; /* One unlimited dimension */
|
|
unsigned flt_msk = 0; /* Filter mask */
|
|
unsigned read_flt_msk = 0; /* Filter mask after direct read */
|
|
hsize_t offset[2]; /* Offset coordinates of a chunk */
|
|
hsize_t start[2] = {START_CHK_X, START_CHK_Y}; /* Start position */
|
|
hsize_t end[2] = {END_CHK_X, END_CHK_Y}; /* End position */
|
|
hsize_t out_offset[2] = {0, 0}; /* Buffer to get offset coordinates */
|
|
hsize_t size = 0; /* Size of an allocated/written chunk */
|
|
hsize_t nchunks = 0; /* Number of chunks */
|
|
haddr_t addr = 0; /* Address of an allocated/written chunk */
|
|
hsize_t chk_index = 0; /* Index of a chunk */
|
|
hsize_t ii, jj; /* Array indices */
|
|
herr_t ret; /* Temporary returned value for verifying failure */
|
|
|
|
TESTING(" Extensible Array index");
|
|
|
|
/* Open the file for reading/writing */
|
|
if ((chunkfile = H5Fopen(filename, H5F_ACC_RDWR, fapl)) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Create dataspace */
|
|
if ((dspace = H5Screate_simple(RANK, dims, maxdims)) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Enable chunking */
|
|
if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
|
|
TEST_ERROR
|
|
|
|
if (H5Pset_chunk(cparms, RANK, chunk_dims) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Create a new dataset using cparms creation properties */
|
|
dset = H5Dcreate2(chunkfile, EXT_ARR_INDEX_DSET_NAME, H5T_NATIVE_INT, dspace, H5P_DEFAULT, cparms,
|
|
H5P_DEFAULT);
|
|
if (dset < 0)
|
|
TEST_ERROR
|
|
|
|
/* Close the dataset */
|
|
if (H5Dclose(dset) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Open the dataset again to test getting chunk info */
|
|
if ((dset = H5Dopen2(chunkfile, EXT_ARR_INDEX_DSET_NAME, H5P_DEFAULT)) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Verify chunk indexing scheme and number of chunks */
|
|
if (verify_idx_nchunks(dset, dspace, H5D_CHUNK_IDX_EARRAY, NO_CHUNK_WRITTEN) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification and write failed\n");
|
|
|
|
/* Write NUM_CHUNKS_WRITTEN chunks at the following logical coords:
|
|
(0,2) (0,3) (1,2) (1,3) */
|
|
if (write_selected_chunks(dset, H5P_DEFAULT, start, end, flt_msk) == FAIL)
|
|
FAIL_PUTS_ERROR("Writing to selected chunks failed\n");
|
|
|
|
/* Get and verify the number of chunks written */
|
|
if (H5Dget_num_chunks(dset, dspace, &nchunks) < 0)
|
|
TEST_ERROR
|
|
VERIFY(nchunks, NUM_CHUNKS_WRITTEN, "H5Dget_num_chunks, number of chunks");
|
|
|
|
/* Get and verify info of each written chunk */
|
|
chk_index = 0;
|
|
for (ii = START_CHK_X; ii < END_CHK_X; ii++)
|
|
for (jj = START_CHK_Y; jj < END_CHK_Y; jj++, chk_index++) {
|
|
offset[0] = ii * CHUNK_NX;
|
|
offset[1] = jj * CHUNK_NY;
|
|
|
|
if (verify_get_chunk_info(dset, dspace, chk_index, CHK_SIZE, offset, flt_msk) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info failed\n");
|
|
|
|
if (verify_get_chunk_info_by_coord(dset, offset, CHK_SIZE, flt_msk) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info_by_coord failed\n");
|
|
}
|
|
|
|
/* Attempt to get info using an out-of-range index, should fail */
|
|
chk_index = OUTOFRANGE_CHK_INDEX;
|
|
H5E_BEGIN_TRY
|
|
{
|
|
ret = H5Dget_chunk_info(dset, dspace, chk_index, out_offset, &read_flt_msk, &addr, &size);
|
|
}
|
|
H5E_END_TRY;
|
|
if (ret != FAIL)
|
|
FAIL_PUTS_ERROR(" Attempted to get info of a chunk using an out-of-range index.");
|
|
|
|
/* Attempt to get info of empty chunks, verify the returned address and size */
|
|
offset[0] = 0;
|
|
offset[1] = 0;
|
|
if (verify_empty_chunk_info(dset, offset) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info_by_coord on empty chunk failed\n");
|
|
|
|
offset[0] = 3 * CHUNK_NX;
|
|
offset[1] = 3 * CHUNK_NY;
|
|
if (verify_empty_chunk_info(dset, offset) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info_by_coord on empty chunk failed\n");
|
|
|
|
/* Read and verify values of selected chunks */
|
|
if (verify_selected_chunks(dset, H5P_DEFAULT, start, end) < 0)
|
|
|
|
/* Release resourse */
|
|
if (H5Dclose(dset) < 0)
|
|
TEST_ERROR
|
|
if (H5Sclose(dspace) < 0)
|
|
TEST_ERROR
|
|
if (H5Fclose(chunkfile) < 0)
|
|
TEST_ERROR
|
|
|
|
PASSED();
|
|
return SUCCEED;
|
|
|
|
error:
|
|
H5E_BEGIN_TRY
|
|
{
|
|
H5Dclose(dset);
|
|
H5Sclose(dspace);
|
|
H5Pclose(cparms);
|
|
H5Fclose(chunkfile);
|
|
}
|
|
H5E_END_TRY;
|
|
|
|
H5_FAILED();
|
|
return FAIL;
|
|
} /* test_chunk_info_extensible_array() */
|
|
|
|
/*-------------------------------------------------------------------------
|
|
* Function: test_chunk_info_version2_btrees
|
|
*
|
|
* Purpose: Test getting various chunk information when Version 2 B-trees
|
|
* index type is used
|
|
*
|
|
* Return: Success: SUCCEED
|
|
* Failure: FAIL
|
|
*
|
|
* Note: Note that the dataspace argument in these new functions are
|
|
* currently not used. The functionality involved the dataspace
|
|
* will be implemented in the next version.
|
|
*
|
|
* Date: November 2018
|
|
*
|
|
*-------------------------------------------------------------------------
|
|
*/
|
|
static herr_t
|
|
test_chunk_info_version2_btrees(const char *filename, hid_t fapl)
|
|
{
|
|
hid_t chunkfile = H5I_INVALID_HID; /* File ID */
|
|
hid_t dspace = H5I_INVALID_HID; /* Dataspace ID */
|
|
hid_t dset = H5I_INVALID_HID; /* Dataset ID */
|
|
hid_t cparms = H5I_INVALID_HID; /* Creation plist */
|
|
hsize_t dims[2] = {NX, NY}; /* Dataset dimensions */
|
|
hsize_t chunk_dims[2] = {CHUNK_NX, CHUNK_NY}; /* Chunk dimensions */
|
|
hsize_t maxdims[2] = {H5S_UNLIMITED, H5S_UNLIMITED}; /* Two unlimited dims */
|
|
unsigned flt_msk = 0; /* Filter mask */
|
|
unsigned read_flt_msk = 0; /* Filter mask after direct read */
|
|
hsize_t offset[2]; /* Offset coordinates of a chunk */
|
|
hsize_t start[2] = {START_CHK_X, START_CHK_Y}; /* Start position */
|
|
hsize_t end[2] = {END_CHK_X, END_CHK_Y}; /* End position */
|
|
hsize_t out_offset[2] = {0, 0}; /* Buffer to get offset coordinates */
|
|
hsize_t size = 0; /* Size of an allocated/written chunk */
|
|
hsize_t nchunks = 0; /* Number of chunks */
|
|
haddr_t addr = 0; /* Address of an allocated/written chunk */
|
|
hsize_t chk_index = 0; /* Index of a chunk */
|
|
hsize_t ii, jj; /* Array indices */
|
|
herr_t ret; /* Temporary returned value for verifying failure */
|
|
|
|
TESTING(" Version 2 B-trees index");
|
|
|
|
/* Open the file for reading/writing */
|
|
if ((chunkfile = H5Fopen(filename, H5F_ACC_RDWR, fapl)) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Create dataspace */
|
|
if ((dspace = H5Screate_simple(RANK, dims, maxdims)) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Enable chunking */
|
|
if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
|
|
TEST_ERROR
|
|
|
|
if (H5Pset_chunk(cparms, RANK, chunk_dims) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Create a new dataset using cparms creation properties */
|
|
dset = H5Dcreate2(chunkfile, V2_BTREE_INDEX_DSET_NAME, H5T_NATIVE_INT, dspace, H5P_DEFAULT, cparms,
|
|
H5P_DEFAULT);
|
|
if (dset < 0)
|
|
TEST_ERROR
|
|
|
|
/* Close the dataset */
|
|
if (H5Dclose(dset) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Open the dataset again to test getting chunk info */
|
|
if ((dset = H5Dopen2(chunkfile, V2_BTREE_INDEX_DSET_NAME, H5P_DEFAULT)) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Verify chunk indexing scheme and number of chunks */
|
|
if (verify_idx_nchunks(dset, dspace, H5D_CHUNK_IDX_BT2, NO_CHUNK_WRITTEN) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification and write failed\n");
|
|
|
|
/* Write NUM_CHUNKS_WRITTEN chunks at the following logical coords:
|
|
(0,2) (0,3) (1,2) (1,3) */
|
|
if (write_selected_chunks(dset, H5P_DEFAULT, start, end, flt_msk) == FAIL)
|
|
FAIL_PUTS_ERROR("Writing to selected chunks failed\n");
|
|
|
|
/* Get and verify the number of chunks written */
|
|
if (H5Dget_num_chunks(dset, dspace, &nchunks) < 0)
|
|
TEST_ERROR
|
|
VERIFY(nchunks, NUM_CHUNKS_WRITTEN, "H5Dget_num_chunks, number of chunks");
|
|
|
|
/* Go through all written chunks, get their info and verify the values */
|
|
chk_index = 0;
|
|
for (ii = START_CHK_X; ii < END_CHK_X; ii++)
|
|
for (jj = START_CHK_Y; jj < END_CHK_Y; jj++, chk_index++) {
|
|
offset[0] = ii * CHUNK_NX;
|
|
offset[1] = jj * CHUNK_NY;
|
|
|
|
if (verify_get_chunk_info(dset, dspace, chk_index, CHK_SIZE, offset, flt_msk) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info failed\n");
|
|
|
|
if (verify_get_chunk_info_by_coord(dset, offset, CHK_SIZE, flt_msk) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info_by_coord failed\n");
|
|
}
|
|
|
|
/* Attempt to provide out-of-range offsets, should fail */
|
|
chk_index = OUTOFRANGE_CHK_INDEX;
|
|
H5E_BEGIN_TRY
|
|
{
|
|
ret = H5Dget_chunk_info(dset, dspace, chk_index, out_offset, &read_flt_msk, &addr, &size);
|
|
}
|
|
H5E_END_TRY;
|
|
if (ret != FAIL)
|
|
FAIL_PUTS_ERROR(" Attempted to get info of a chunk using an out-of-range index.");
|
|
|
|
/* Attempt to get info of empty chunks, verify the returned address and size */
|
|
offset[0] = 0;
|
|
offset[1] = 0;
|
|
if (verify_empty_chunk_info(dset, offset) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info_by_coord on empty chunk failed\n");
|
|
|
|
offset[0] = 3 * CHUNK_NX;
|
|
offset[1] = 3 * CHUNK_NY;
|
|
if (verify_empty_chunk_info(dset, offset) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info_by_coord on empty chunk failed\n");
|
|
|
|
/* Read and verify values of selected chunks */
|
|
if (verify_selected_chunks(dset, H5P_DEFAULT, start, end) < 0)
|
|
|
|
/* Release resourse */
|
|
if (H5Dclose(dset) < 0)
|
|
TEST_ERROR
|
|
if (H5Sclose(dspace) < 0)
|
|
TEST_ERROR
|
|
if (H5Fclose(chunkfile) < 0)
|
|
TEST_ERROR
|
|
|
|
PASSED();
|
|
return SUCCEED;
|
|
|
|
error:
|
|
H5E_BEGIN_TRY
|
|
{
|
|
H5Dclose(dset);
|
|
H5Sclose(dspace);
|
|
H5Pclose(cparms);
|
|
H5Fclose(chunkfile);
|
|
}
|
|
H5E_END_TRY;
|
|
|
|
H5_FAILED();
|
|
return FAIL;
|
|
} /* test_chunk_info_version2_btrees() */
|
|
|
|
typedef struct chunk_iter_info_t {
|
|
hsize_t offset[2];
|
|
uint32_t filter_mask;
|
|
haddr_t addr;
|
|
uint32_t nbytes;
|
|
} chunk_iter_info_t;
|
|
|
|
typedef struct chunk_iter_udata_t {
|
|
chunk_iter_info_t *chunk_info;
|
|
int last_index;
|
|
} chunk_iter_udata_t;
|
|
|
|
static int
|
|
iter_cb(const hsize_t *offset, uint32_t filter_mask, haddr_t addr, uint32_t nbytes, void *op_data)
|
|
{
|
|
chunk_iter_udata_t *cidata = (chunk_iter_udata_t *)op_data;
|
|
int idx = cidata->last_index + 1;
|
|
|
|
cidata->chunk_info[idx].offset[0] = offset[0];
|
|
cidata->chunk_info[idx].offset[1] = offset[1];
|
|
cidata->chunk_info[idx].filter_mask = filter_mask;
|
|
cidata->chunk_info[idx].addr = addr;
|
|
cidata->chunk_info[idx].nbytes = nbytes;
|
|
|
|
cidata->last_index++;
|
|
|
|
return H5_ITER_CONT;
|
|
}
|
|
|
|
static int
|
|
iter_cb_stop(const hsize_t H5_ATTR_UNUSED *offset, uint32_t H5_ATTR_UNUSED filter_mask,
|
|
haddr_t H5_ATTR_UNUSED addr, uint32_t H5_ATTR_UNUSED nbytes, void *op_data)
|
|
{
|
|
chunk_iter_info_t **chunk_info = (chunk_iter_info_t **)op_data;
|
|
*chunk_info += 1;
|
|
return H5_ITER_STOP;
|
|
}
|
|
|
|
static int
|
|
iter_cb_fail(const hsize_t H5_ATTR_UNUSED *offset, uint32_t H5_ATTR_UNUSED filter_mask,
|
|
haddr_t H5_ATTR_UNUSED addr, uint32_t H5_ATTR_UNUSED nbytes, void *op_data)
|
|
{
|
|
chunk_iter_info_t **chunk_info = (chunk_iter_info_t **)op_data;
|
|
*chunk_info += 1;
|
|
return H5_ITER_ERROR;
|
|
}
|
|
|
|
/*-------------------------------------------------------------------------
|
|
* Function: test_basic_query
|
|
*
|
|
* Purpose: Tests basic operations to ensure the chunk query functions
|
|
* work properly.
|
|
*
|
|
* Return: Success: SUCCEED
|
|
* Failure: FAIL
|
|
*
|
|
* Note: Note that the dataspace argument in these new functions are
|
|
* currently not used. The functionality involved the dataspace
|
|
* will be implemented in the next version.
|
|
*
|
|
* Date: August 2019
|
|
*
|
|
*-------------------------------------------------------------------------
|
|
*/
|
|
static herr_t
|
|
test_basic_query(hid_t fapl)
|
|
{
|
|
char filename[FILENAME_BUF_SIZE]; /* File name */
|
|
hid_t basicfile = H5I_INVALID_HID; /* File ID */
|
|
hid_t dspace = H5I_INVALID_HID; /* Dataspace ID */
|
|
hid_t dset = H5I_INVALID_HID; /* Dataset ID */
|
|
hid_t cparms = H5I_INVALID_HID; /* Creation plist */
|
|
hsize_t dims[2] = {NX, NY}; /* Dataset dimensions */
|
|
hsize_t chunk_dims[2] = {CHUNK_NX, CHUNK_NY}; /* Chunk dimensions */
|
|
int direct_buf[CHUNK_NX][CHUNK_NY]; /* Data in chunks */
|
|
unsigned flt_msk = 0; /* Filter mask */
|
|
unsigned read_flt_msk = 0; /* Filter mask after direct read */
|
|
hsize_t offset[2]; /* Offset coordinates of a chunk */
|
|
hsize_t out_offset[2] = {0, 0}; /* Buffer to get offset coordinates */
|
|
hsize_t size = 0; /* Size of an allocated/written chunk */
|
|
hsize_t nchunks = 0; /* Number of chunks */
|
|
haddr_t addr = 0; /* Address of an allocated/written chunk */
|
|
hsize_t chk_index = 0; /* Index of a chunk */
|
|
hsize_t ii, jj; /* Array indices */
|
|
chunk_iter_info_t chunk_infos[2]; /* Chunk infos filled up by iterator */
|
|
chunk_iter_info_t *cptr; /* Pointer to array of chunks */
|
|
chunk_iter_udata_t udata; /* udata for iteration */
|
|
herr_t ret; /* Temporary returned value for verifying failure */
|
|
|
|
TESTING("basic operations");
|
|
|
|
/* Create the file */
|
|
h5_fixname(BASIC_FILE, fapl, filename, sizeof filename);
|
|
|
|
/* Create a new file. */
|
|
if ((basicfile = H5Fcreate(filename, H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT)) < 0)
|
|
TEST_ERROR;
|
|
|
|
/* Create dataspace */
|
|
if ((dspace = H5Screate_simple(RANK, dims, NULL)) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Enable chunking */
|
|
if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
|
|
TEST_ERROR
|
|
|
|
if (H5Pset_chunk(cparms, RANK, chunk_dims) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Create a new dataset using cparms creation properties */
|
|
dset = H5Dcreate2(basicfile, SIMPLE_CHUNKED_DSET_NAME, H5T_NATIVE_INT, dspace, H5P_DEFAULT, cparms,
|
|
H5P_DEFAULT);
|
|
if (dset < 0)
|
|
TEST_ERROR
|
|
|
|
/* Get the number of chunks and verify that no chunk has been written */
|
|
if (H5Dget_num_chunks(dset, dspace, &nchunks) < 0)
|
|
TEST_ERROR
|
|
VERIFY(nchunks, NO_CHUNK_WRITTEN, "H5Dget_num_chunks, number of chunks");
|
|
|
|
/* Initialize the array of chunk data for the single chunk */
|
|
for (ii = 0; ii < CHUNK_NX; ii++)
|
|
for (jj = 0; jj < CHUNK_NY; jj++)
|
|
direct_buf[ii][jj] = (int)(ii * jj);
|
|
|
|
/* Write the chunk of data */
|
|
offset[0] = CHUNK_NX;
|
|
offset[1] = CHUNK_NY;
|
|
if (H5Dwrite_chunk(dset, H5P_DEFAULT, flt_msk, offset, CHK_SIZE, direct_buf) < 0)
|
|
TEST_ERROR;
|
|
|
|
/* Get and verify that one chunk had been written */
|
|
if (H5Dget_num_chunks(dset, dspace, &nchunks) < 0)
|
|
TEST_ERROR
|
|
VERIFY(nchunks, ONE_CHUNK_WRITTEN, "H5Dget_num_chunks, number of chunks");
|
|
|
|
/* Get and verify info of the first and only chunk */
|
|
if (verify_get_chunk_info(dset, H5S_ALL, 0, CHK_SIZE, offset, flt_msk) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification H5Dget_chunk_info failed\n");
|
|
|
|
/* Get and verify info of the chunk at the offset (CHUNK_NX,CHUNK_NY) */
|
|
if (verify_get_chunk_info_by_coord(dset, offset, CHK_SIZE, flt_msk) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info_by_coord failed\n");
|
|
|
|
/* Attempt to get chunk info given an invalid chunk index and verify
|
|
* that failure occurs */
|
|
chk_index = INVALID_CHK_INDEX;
|
|
reinit_vars(&read_flt_msk, &addr, &size);
|
|
H5E_BEGIN_TRY
|
|
{
|
|
ret = H5Dget_chunk_info(dset, dspace, chk_index, out_offset, &read_flt_msk, &addr, &size);
|
|
}
|
|
H5E_END_TRY;
|
|
if (ret != FAIL)
|
|
TEST_ERROR
|
|
|
|
/* Write the chunk of data to another location */
|
|
offset[0] = 0;
|
|
offset[1] = 0;
|
|
if (H5Dwrite_chunk(dset, H5P_DEFAULT, flt_msk, offset, CHK_SIZE, direct_buf) < 0)
|
|
TEST_ERROR;
|
|
|
|
/* Get and verify that two chunks had been written */
|
|
if (H5Dget_num_chunks(dset, dspace, &nchunks) < 0)
|
|
TEST_ERROR
|
|
VERIFY(nchunks, TWO_CHUNKS_WRITTEN, "H5Dget_num_chunks, number of chunks");
|
|
|
|
/* Get and verify info of the first written chunk in the dataset, its
|
|
offset should be (0,0) */
|
|
if (verify_get_chunk_info(dset, H5S_ALL, 0, CHK_SIZE, offset, flt_msk) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification H5Dget_chunk_info failed\n");
|
|
|
|
/* Get and verify info of the chunk at the offset (0,0) */
|
|
if (verify_get_chunk_info_by_coord(dset, offset, CHK_SIZE, flt_msk) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info_by_coord failed\n");
|
|
|
|
/* Get and verify info of the second written chunk in the dataset, its
|
|
offset should be (CHUNK_NX, CHUNK_NY) */
|
|
offset[0] = CHUNK_NX;
|
|
offset[1] = CHUNK_NY;
|
|
if (verify_get_chunk_info(dset, H5S_ALL, 1, CHK_SIZE, offset, flt_msk) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification H5Dget_chunk_info failed\n");
|
|
|
|
/* Get and verify info of the chunk at the offset (CHUNK_NX, CHUNK_NY) */
|
|
if (verify_get_chunk_info_by_coord(dset, offset, CHK_SIZE, flt_msk) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info_by_coord failed\n");
|
|
|
|
/* Get and verify info of an empty chunk, at offset
|
|
(2*CHUNK_NX, 2*CHUNK_NY) */
|
|
offset[0] = 2 * CHUNK_NX;
|
|
offset[1] = 2 * CHUNK_NY;
|
|
/* Get and verify info of the chunk at the offset (CHUNK_NX, CHUNK_NY) */
|
|
if (verify_empty_chunk_info(dset, offset) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info_by_coord on empty chunk failed\n");
|
|
|
|
/* Iterate over all chunks */
|
|
udata.chunk_info = chunk_infos;
|
|
udata.last_index = -1;
|
|
if (H5Dchunk_iter(dset, H5P_DEFAULT, &iter_cb, &udata) < 0)
|
|
TEST_ERROR;
|
|
|
|
VERIFY(udata.last_index, 1, "Iterator did not iterate all chunks");
|
|
VERIFY(chunk_infos[0].offset[0], 0, "Offset mismatch");
|
|
VERIFY(chunk_infos[0].offset[1], 0, "Offset mismatch");
|
|
VERIFY(chunk_infos[0].filter_mask, 0, "Filter mismatch");
|
|
VERIFY(chunk_infos[0].nbytes, 96, "Size mismatch");
|
|
|
|
VERIFY(chunk_infos[1].offset[0], 1, "Offset mismatch");
|
|
VERIFY(chunk_infos[1].offset[1], 1, "Offset mismatch");
|
|
|
|
/* Iterate and stop after one iteration */
|
|
cptr = &(chunk_infos[0]);
|
|
if (H5Dchunk_iter(dset, H5P_DEFAULT, &iter_cb_stop, &cptr) < 0)
|
|
TEST_ERROR;
|
|
VERIFY(cptr, &(chunk_infos[1]), "Verification of halted iterator failed\n");
|
|
|
|
/* Iterate and fail after one iteration */
|
|
cptr = &(chunk_infos[0]);
|
|
H5E_BEGIN_TRY
|
|
{
|
|
ret = H5Dchunk_iter(dset, H5P_DEFAULT, &iter_cb_fail, &cptr);
|
|
}
|
|
H5E_END_TRY;
|
|
if (ret >= 0)
|
|
TEST_ERROR;
|
|
VERIFY(cptr, &(chunk_infos[1]), "Verification of halted iterator failed\n");
|
|
|
|
/* Release resourse */
|
|
if (H5Dclose(dset) < 0)
|
|
TEST_ERROR
|
|
if (H5Sclose(dspace) < 0)
|
|
TEST_ERROR
|
|
if (H5Pclose(cparms) < 0)
|
|
TEST_ERROR
|
|
if (H5Fclose(basicfile) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Remove the test file */
|
|
HDremove(filename);
|
|
|
|
PASSED();
|
|
return SUCCEED;
|
|
|
|
error:
|
|
H5E_BEGIN_TRY
|
|
{
|
|
H5Dclose(dset);
|
|
H5Sclose(dspace);
|
|
H5Pclose(cparms);
|
|
H5Fclose(basicfile);
|
|
}
|
|
H5E_END_TRY;
|
|
|
|
H5_FAILED();
|
|
return FAIL;
|
|
} /* test_basic_query() */
|
|
|
|
/*-------------------------------------------------------------------------
|
|
* Function: test_failed_attempts
|
|
*
|
|
* Purpose: Test attempting to use chunk query functions incorrectly.
|
|
*
|
|
* Return: Success: SUCCEED
|
|
* Failure: FAIL
|
|
*
|
|
* Note: Note that the dataspace argument in these new functions are
|
|
* currently not used. The functionality involved the dataspace
|
|
* will be implemented in the next version.
|
|
*
|
|
* Date: August 2019
|
|
*
|
|
*-------------------------------------------------------------------------
|
|
*/
|
|
static herr_t
|
|
test_failed_attempts(const char *filename, hid_t fapl)
|
|
{
|
|
hid_t chunkfile = H5I_INVALID_HID; /* File ID */
|
|
hid_t dspace = H5I_INVALID_HID; /* Dataspace ID */
|
|
hid_t dset = H5I_INVALID_HID; /* Dataset ID */
|
|
hsize_t dims[2] = {NX, NY}; /* Dataset dimensions */
|
|
int data_buf[NX][NY]; /* Input buffer */
|
|
unsigned read_flt_msk = 0; /* Filter mask after direct read */
|
|
hsize_t offset[2]; /* Offset coordinates of a chunk */
|
|
hsize_t out_offset[2] = {0, 0}; /* Buffer to get offset coordinates */
|
|
hsize_t size = 0; /* Size of an allocated/written chunk */
|
|
hsize_t nchunks = 0; /* Number of chunks */
|
|
haddr_t addr = 0; /* Address of an allocated/written chunk */
|
|
hsize_t chk_index = 0; /* Index of a chunk */
|
|
hsize_t ii, jj; /* Array indices */
|
|
herr_t ret; /* Temporary returned value for verifying failure */
|
|
|
|
TESTING(" Invalid Operations");
|
|
|
|
/* Open the file for reading/writing */
|
|
if ((chunkfile = H5Fopen(filename, H5F_ACC_RDWR, fapl)) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Create dataspace */
|
|
if ((dspace = H5Screate_simple(RANK, dims, NULL)) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Create a contiguous dataset */
|
|
dset = H5Dcreate2(chunkfile, CONTIGUOUS_DSET_NAME, H5T_NATIVE_INT, dspace, H5P_DEFAULT, H5P_DEFAULT,
|
|
H5P_DEFAULT);
|
|
if (dset < 0)
|
|
TEST_ERROR
|
|
|
|
/* Initialize the array of data */
|
|
for (ii = 0; ii < NX; ii++)
|
|
for (jj = 0; jj < NY; jj++)
|
|
data_buf[ii][jj] = (int)(ii * jj);
|
|
|
|
/* Write the data */
|
|
if (H5Dwrite(dset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, H5P_DEFAULT, data_buf) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Close the dataset */
|
|
if (H5Dclose(dset) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Open it again to test the chunk query functions on contiguous dataset */
|
|
if ((dset = H5Dopen2(chunkfile, CONTIGUOUS_DSET_NAME, H5P_DEFAULT)) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Attempt to get the number of chunks on contiguous dataset, should fail */
|
|
H5E_BEGIN_TRY
|
|
{
|
|
ret = H5Dget_num_chunks(dset, dspace, &nchunks);
|
|
}
|
|
H5E_END_TRY;
|
|
if (ret != FAIL)
|
|
FAIL_PUTS_ERROR(" Attempt a chunk query function on a contiguous dataset.")
|
|
|
|
/* Attempt to get chunk info on contiguous data, should fail */
|
|
chk_index = 0;
|
|
reinit_vars(&read_flt_msk, &addr, &size);
|
|
H5E_BEGIN_TRY
|
|
{
|
|
ret = H5Dget_chunk_info(dset, dspace, chk_index, out_offset, &read_flt_msk, &addr, &size);
|
|
}
|
|
H5E_END_TRY;
|
|
if (ret != FAIL)
|
|
FAIL_PUTS_ERROR(" Attempt a chunk query function on a contiguous dataset.")
|
|
|
|
/* Attempt to get chunk info at logical coordinates (0,0) on contiguous
|
|
* dataset, should fail */
|
|
offset[0] = 0;
|
|
offset[1] = 0;
|
|
H5E_BEGIN_TRY
|
|
{
|
|
ret = H5Dget_chunk_info_by_coord(dset, offset, &read_flt_msk, &addr, &size);
|
|
}
|
|
H5E_END_TRY;
|
|
if (ret != FAIL)
|
|
FAIL_PUTS_ERROR(" Attempt a chunk query function on a contiguous dataset.")
|
|
|
|
/* Release resourse */
|
|
if (H5Dclose(dset) < 0)
|
|
TEST_ERROR
|
|
if (H5Sclose(dspace) < 0)
|
|
TEST_ERROR
|
|
if (H5Fclose(chunkfile) < 0)
|
|
TEST_ERROR
|
|
|
|
PASSED();
|
|
return SUCCEED;
|
|
|
|
error:
|
|
H5E_BEGIN_TRY
|
|
{
|
|
H5Dclose(dset);
|
|
H5Sclose(dspace);
|
|
H5Fclose(chunkfile);
|
|
}
|
|
H5E_END_TRY;
|
|
|
|
H5_FAILED();
|
|
return FAIL;
|
|
} /* test_failed_attempts() */
|
|
|
|
/*-------------------------------------------------------------------------
|
|
* Function: test_get_chunk_info_v110
|
|
*
|
|
* Purpose: Test getting various chunk information in version 1.10.
|
|
*
|
|
* Return: Success: SUCCEED
|
|
* Failure: FAIL
|
|
*
|
|
* Note: Note that the dataspace argument in these new functions are
|
|
* currently not used. The functionality involved the dataspace
|
|
* will be implemented in the next version.
|
|
*
|
|
* Description:
|
|
* This function tests the new API functions added for HDFFV-10677:
|
|
* H5Dget_num_chunks, H5Dget_chunk_info, and
|
|
* H5Dget_chunk_info_by_coord for low bound beyond 1.8.
|
|
*
|
|
* Date: October 2018
|
|
*
|
|
*-------------------------------------------------------------------------
|
|
*/
|
|
static herr_t
|
|
test_get_chunk_info_v110(hid_t fapl)
|
|
{
|
|
char filename[FILENAME_BUF_SIZE]; /* File name */
|
|
hid_t chunkfile = H5I_INVALID_HID; /* File ID */
|
|
H5F_libver_t low, high; /* File format bounds */
|
|
|
|
TESTING("getting chunk information in file with versions 1.10 and later");
|
|
HDprintf("\n"); /* to list sub-tests */
|
|
|
|
/* Set high bound to the current latest version */
|
|
high = H5F_LIBVER_LATEST;
|
|
|
|
/* Test getting info of chunked datasets in version combo up to 1.10 */
|
|
for (low = H5F_LIBVER_V110; low <= H5F_LIBVER_LATEST; low++) {
|
|
/* Set version bounds for creating file */
|
|
if (H5Pset_libver_bounds(fapl, low, high) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Create the file */
|
|
h5_fixname(FILENAME[low], fapl, filename, sizeof filename);
|
|
chunkfile = H5Fcreate(filename, H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT);
|
|
if (chunkfile < 0)
|
|
TEST_ERROR
|
|
|
|
/* Close the file, individual tests will re-open the file with different
|
|
libvers via the fapl */
|
|
if (H5Fclose(chunkfile) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Test getting chunk info when Single Chunk index type is used */
|
|
if (test_chunk_info_single_chunk(filename, fapl) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Test getting chunk info when Implicit index type is used */
|
|
if (test_chunk_info_implicit(filename, fapl) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Test getting chunk info when Fixed Array index type is used */
|
|
if (test_chunk_info_fixed_array(filename, fapl) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Test getting chunk info when Extensible Array index type is used */
|
|
if (test_chunk_info_extensible_array(filename, fapl) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Test getting chunk info when Version 2 B-trees index type is used */
|
|
if (test_chunk_info_version2_btrees(filename, fapl) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Test various attempts to use the functions incorrectly */
|
|
if (test_failed_attempts(filename, fapl) < 0)
|
|
TEST_ERROR
|
|
|
|
} /* for low libver bound */
|
|
|
|
return SUCCEED;
|
|
|
|
error:
|
|
H5_FAILED();
|
|
return FAIL;
|
|
} /* test_get_chunk_info_v110() */
|
|
|
|
/*-------------------------------------------------------------------------
|
|
* Function: test_flt_msk_with_skip_compress
|
|
*
|
|
* Purpose: Test getting chunk info when compression filter is skipped.
|
|
*
|
|
* Return: Success: SUCCEED
|
|
* Failure: FAIL
|
|
*
|
|
* Date: August 2019 (based on direct_chunk.c/test_skip_compress_write1)
|
|
*
|
|
*-------------------------------------------------------------------------
|
|
*/
|
|
static herr_t
|
|
test_flt_msk_with_skip_compress(hid_t fapl)
|
|
{
|
|
char filename[FILENAME_BUF_SIZE]; /* File name */
|
|
hid_t filter_file = H5I_INVALID_HID; /* File ID for filter mask */
|
|
hid_t dspace = H5I_INVALID_HID; /* Dataspace ID */
|
|
hid_t mem_space = H5I_INVALID_HID; /* Dataspace ID */
|
|
hid_t dset = H5I_INVALID_HID; /* Dataset ID */
|
|
hid_t cparms = H5I_INVALID_HID; /* Creation plist */
|
|
hid_t dxpl = H5I_INVALID_HID; /* Transfer plist */
|
|
hsize_t dims[2] = {NX, NY}; /* Dataset dimensions */
|
|
hsize_t maxdims[2] = {H5S_UNLIMITED, H5S_UNLIMITED}; /* 2 unlimited dims */
|
|
hsize_t chunk_dims[2] = {CHUNK_NX, CHUNK_NY}; /* Chunk dimensions */
|
|
int direct_buf[CHUNK_NX][CHUNK_NY]; /* One chunk of data */
|
|
int check_chunk[CHUNK_NX][CHUNK_NY]; /* Buffer to read data in */
|
|
int read_direct_buf[CHUNK_NX][CHUNK_NY]; /* Buffer to read a chunk */
|
|
hsize_t read_buf_size = 0; /* buf size */
|
|
unsigned flt_msk = 0; /* Filter mask */
|
|
unsigned read_flt_msk = 0; /* Filter mask after direct read */
|
|
hsize_t offset[2] = {0, 0}; /* Offset coordinates of a chunk */
|
|
hsize_t nchunks = 0; /* Number of chunks */
|
|
hsize_t chk_index = 0; /* Index of a chunk */
|
|
int aggression = 9; /* Compression aggression setting */
|
|
hsize_t start[2]; /* Start of hyperslab */
|
|
hsize_t stride[2]; /* Stride of hyperslab */
|
|
hsize_t count[2]; /* Block count */
|
|
hsize_t block[2]; /* Block sizes */
|
|
int ii, jj; /* Array indices */
|
|
|
|
TESTING("getting filter mask when compression filter is skipped");
|
|
|
|
/* Create the file */
|
|
h5_fixname(FILTERMASK_FILE, fapl, filename, sizeof filename);
|
|
|
|
/* Create a new file. */
|
|
if ((filter_file = H5Fcreate(filename, H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT)) < 0)
|
|
TEST_ERROR;
|
|
|
|
/* Create file data space with unlimited dimensions. */
|
|
if ((dspace = H5Screate_simple(RANK, dims, maxdims)) < 0)
|
|
TEST_ERROR;
|
|
|
|
/* Create memory data space. */
|
|
if ((mem_space = H5Screate_simple(RANK, chunk_dims, NULL)) < 0)
|
|
TEST_ERROR;
|
|
|
|
/* Create dataset create property list with chunking and compression
|
|
enabled. */
|
|
if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
|
|
TEST_ERROR;
|
|
|
|
if (H5Pset_chunk(cparms, RANK, chunk_dims) < 0)
|
|
TEST_ERROR;
|
|
|
|
if (H5Pset_deflate(cparms, (unsigned)aggression) < 0)
|
|
TEST_ERROR;
|
|
|
|
/* Create a new dataset using cparms creation properties. */
|
|
if ((dset = H5Dcreate2(filter_file, SKIP_FILTER_DSET_NAME, H5T_NATIVE_INT, dspace, H5P_DEFAULT, cparms,
|
|
H5P_DEFAULT)) < 0)
|
|
TEST_ERROR;
|
|
|
|
/* Create transfer property list for writing */
|
|
if ((dxpl = H5Pcreate(H5P_DATASET_XFER)) < 0)
|
|
TEST_ERROR;
|
|
|
|
/* Initialize data for one chunk */
|
|
for (ii = 0; ii < CHUNK_NX; ii++)
|
|
for (jj = 0; jj < CHUNK_NY; jj++)
|
|
direct_buf[ii][jj] = (int)(ii * jj);
|
|
|
|
/* Indicate the compression filter is to be skipped. */
|
|
flt_msk = 0x00000001;
|
|
|
|
/* Write a chunk of uncompressed data */
|
|
offset[0] = CHUNK_NX;
|
|
offset[1] = CHUNK_NY;
|
|
if (H5Dwrite_chunk(dset, H5P_DEFAULT, flt_msk, offset, CHK_SIZE, direct_buf) < 0)
|
|
TEST_ERROR;
|
|
|
|
if (H5Fflush(dset, H5F_SCOPE_LOCAL) < 0)
|
|
TEST_ERROR;
|
|
|
|
/* Close and re-open the dataset */
|
|
if (H5Dclose(dset) < 0)
|
|
TEST_ERROR;
|
|
if ((dset = H5Dopen2(filter_file, SKIP_FILTER_DSET_NAME, H5P_DEFAULT)) < 0)
|
|
TEST_ERROR;
|
|
|
|
/* Select hyperslab for the chunk just written in the file */
|
|
start[0] = CHUNK_NX;
|
|
start[1] = CHUNK_NY;
|
|
stride[0] = 1;
|
|
stride[1] = 1;
|
|
count[0] = 1;
|
|
count[1] = 1;
|
|
block[0] = CHUNK_NX;
|
|
block[1] = CHUNK_NY;
|
|
if (H5Sselect_hyperslab(dspace, H5S_SELECT_SET, start, stride, count, block) < 0)
|
|
TEST_ERROR;
|
|
|
|
/* Read the chunk back */
|
|
if (H5Dread(dset, H5T_NATIVE_INT, mem_space, dspace, H5P_DEFAULT, check_chunk) < 0)
|
|
TEST_ERROR;
|
|
|
|
/* Check that the values read are the same as the values written */
|
|
for (ii = 0; ii < CHUNK_NX; ii++)
|
|
for (jj = 0; jj < CHUNK_NY; jj++)
|
|
if (direct_buf[ii][jj] != check_chunk[ii][jj]) {
|
|
HDprintf(" 1. Read different values than written.");
|
|
HDprintf(" At index %d,%d\n", ii, jj);
|
|
HDprintf(" direct_buf=%d, check_chunk=%d\n", direct_buf[ii][jj], check_chunk[ii][jj]);
|
|
TEST_ERROR;
|
|
}
|
|
|
|
/* Query chunk storage size */
|
|
if (H5Dget_chunk_storage_size(dset, offset, &read_buf_size) < 0)
|
|
TEST_ERROR;
|
|
if (read_buf_size != CHK_SIZE)
|
|
TEST_ERROR;
|
|
|
|
/* Read the raw chunk back with H5Dread_chunk */
|
|
HDmemset(&read_direct_buf, 0, sizeof(read_direct_buf));
|
|
if (H5Dread_chunk(dset, H5P_DEFAULT, offset, &read_flt_msk, read_direct_buf) < 0)
|
|
TEST_ERROR;
|
|
if (read_flt_msk != flt_msk)
|
|
TEST_ERROR;
|
|
|
|
/* Check that the direct chunk read is the same as the chunk written */
|
|
for (ii = 0; ii < CHUNK_NX; ii++)
|
|
for (jj = 0; jj < CHUNK_NY; jj++)
|
|
if (direct_buf[ii][jj] != read_direct_buf[ii][jj]) {
|
|
HDprintf(" 1. Read different values than written.");
|
|
HDprintf(" At index %d,%d\n", ii, jj);
|
|
HDprintf(" direct_buf=%d, read_direct_buf=%d\n", direct_buf[ii][jj],
|
|
read_direct_buf[ii][jj]);
|
|
TEST_ERROR;
|
|
}
|
|
|
|
/* Get and verify the number of chunks written */
|
|
if (H5Dget_num_chunks(dset, H5S_ALL, &nchunks) < 0)
|
|
TEST_ERROR
|
|
VERIFY(nchunks, ONE_CHUNK_WRITTEN, "H5Dget_num_chunks, number of chunks");
|
|
|
|
/* Get and verify info of the first and only chunk */
|
|
chk_index = 0;
|
|
offset[0] = CHUNK_NX;
|
|
offset[1] = CHUNK_NY;
|
|
if (verify_get_chunk_info(dset, H5S_ALL, chk_index, CHK_SIZE, offset, flt_msk) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info failed\n");
|
|
|
|
/* Get info of the chunk at the specified offsets and verify its info */
|
|
if (verify_get_chunk_info_by_coord(dset, offset, CHK_SIZE, flt_msk) == FAIL)
|
|
FAIL_PUTS_ERROR("Verification of H5Dget_chunk_info_by_coord failed\n");
|
|
|
|
/* Release resourse */
|
|
if (H5Dclose(dset) < 0)
|
|
TEST_ERROR
|
|
if (H5Sclose(mem_space) < 0)
|
|
TEST_ERROR
|
|
if (H5Sclose(dspace) < 0)
|
|
TEST_ERROR
|
|
if (H5Pclose(cparms) < 0)
|
|
TEST_ERROR
|
|
if (H5Pclose(dxpl) < 0)
|
|
TEST_ERROR
|
|
if (H5Fclose(filter_file) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Remove the test file */
|
|
HDremove(filename);
|
|
|
|
PASSED();
|
|
return SUCCEED;
|
|
|
|
error:
|
|
H5E_BEGIN_TRY
|
|
{
|
|
H5Dclose(dset);
|
|
H5Sclose(mem_space);
|
|
H5Sclose(dspace);
|
|
H5Pclose(cparms);
|
|
H5Pclose(dxpl);
|
|
H5Fclose(filter_file);
|
|
}
|
|
H5E_END_TRY;
|
|
|
|
H5_FAILED();
|
|
return FAIL;
|
|
} /* test_flt_msk_with_skip_compress() */
|
|
|
|
/*-------------------------------------------------------------------------
|
|
* Function: main
|
|
*
|
|
* Purpose: Tests functions related to chunk information
|
|
*
|
|
* Return: EXIT_SUCCESS/EXIT_FAILURE
|
|
*
|
|
* Programmer: Binh-Minh Ribler
|
|
* November 5, 2018
|
|
*
|
|
*-------------------------------------------------------------------------
|
|
*/
|
|
int
|
|
main(void)
|
|
{
|
|
hid_t fapl = H5I_INVALID_HID; /* File access property list */
|
|
int nerrors = 0; /* Number of errors so far */
|
|
|
|
h5_reset();
|
|
|
|
/* Create a copy of file access property list */
|
|
if ((fapl = H5Pcreate(H5P_FILE_ACCESS)) < 0)
|
|
TEST_ERROR
|
|
|
|
/* Test basic operations on the chunk query functions */
|
|
nerrors += test_basic_query(fapl) < 0 ? 1 : 0;
|
|
|
|
/* Tests getting chunk information of version 1.8 and prior */
|
|
nerrors += test_get_chunk_info_highest_v18(fapl) < 0 ? 1 : 0;
|
|
|
|
/* Tests getting chunk information of version 1.10 */
|
|
nerrors += test_get_chunk_info_v110(fapl) < 0 ? 1 : 0;
|
|
|
|
/* Tests getting filter mask when compression filter is skipped */
|
|
nerrors += test_flt_msk_with_skip_compress(fapl) < 0 ? 1 : 0;
|
|
|
|
if (nerrors)
|
|
goto error;
|
|
|
|
HDprintf("All chunk query tests passed.\n");
|
|
|
|
h5_cleanup(FILENAME, fapl);
|
|
|
|
return EXIT_SUCCESS;
|
|
|
|
error:
|
|
nerrors = MAX(1, nerrors);
|
|
HDprintf("***** %d QUERY CHUNK INFO TEST%s FAILED! *****\n", nerrors, 1 == nerrors ? "" : "S");
|
|
return EXIT_FAILURE;
|
|
}
|
|
|
|
/****************************************************************************
|
|
Additional tests to be added:
|
|
- do the query when extending the dataset (shrink or expand)
|
|
- verify that invalid input parameters are handled properly
|
|
|
|
****************************************************************************/
|