hdf5/test/direct_chunk.c
Scot Breitenfeld f859cb732b
Fixed Spelling Errors (#1166)
* fixed missed closing of a dataset

* fixed missed closing of a dataset

* fixed typo in error return

* Committing clang-format changes

* minor edits

* code format

* Committing clang-format changes

* code format

* minor edit

* switched from using MPI_count, to actual bytes written for H5FD_mpio_debug rw debugging

* Committing clang-format changes

* changed size_i in printf to reflect the I/O.

* Committing clang-format changes

* Fixed seg fault with xlf on BE with -qintsize=8

* fixed error function string

* spelling corrections via codespell, added new spell check github actions

* Committing clang-format changes

* misc

* misc

* misc

* misc

* misc

* misc

* misc

* misc

* misc

* misc

* misc

* misc

* misc

* misc

* Committing clang-format changes

* misc

* misc

* misc

* misc

* misc

* misc

* Committing clang-format changes

* misc

* work around for https://github.com/codespell-project/codespell/issues/2137

* misc

* added missing file

* misc

* misc.

* misc

* switch to using Codespell with GitHub Actions

* misc.

* misc.

* fixed more sp errors

* Fix new typos found by codespell.

* fixed proceed with precede

* fixed variable in fortran test

* fixed minnum

* updated spelling list

Co-authored-by: github-actions <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: Larry Knox <lrknox@hdfgroup.org>
2021-12-07 08:27:29 -06:00

2356 lines
76 KiB
C

/* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *
* Copyright by The HDF Group. *
* Copyright by the Board of Trustees of the University of Illinois. *
* All rights reserved. *
* *
* This file is part of HDF5. The full HDF5 copyright notice, including *
* terms governing use, modification, and redistribution, is contained in *
* the COPYING file, which can be found at the root of the source code *
* distribution tree, or in https://www.hdfgroup.org/licenses. *
* If you do not have access to either file, you may request a copy from *
* help@hdfgroup.org. *
* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
#include "h5test.h"
#if defined(H5_HAVE_ZLIB_H) && !defined(H5_ZLIB_HEADER)
#define H5_ZLIB_HEADER "zlib.h"
#endif
#if defined(H5_ZLIB_HEADER)
#include H5_ZLIB_HEADER /* "zlib.h" */
#endif
#define FILE_NAME "direct_chunk.h5"
/* Datasets for Direct Write tests */
#ifdef H5_HAVE_FILTER_DEFLATE
#define DATASETNAME1 "direct_write"
#endif
#define DATASETNAME2 "skip_one_filter"
#define DATASETNAME3 "skip_two_filters"
#define DATASETNAME4 "data_conv"
#define DATASETNAME5 "contiguous_dset"
#define DATASETNAME6 "invalid_argue"
#define DATASETNAME7 "overwrite_chunk"
/* Datasets for Direct Read tests */
#ifdef H5_HAVE_FILTER_DEFLATE
#define DATASETNAME8 "disabled_chunk_cache"
#define DATASETNAME9 "flush_chunk_cache"
#define DATASETNAME10 "read_w_valid_cache"
#endif
#define DATASETNAME11 "unallocated_chunk"
#define DATASETNAME12 "unfiltered_data"
#define RANK 2
#define NX 16
#define NY 16
#define CHUNK_NX 4
#define CHUNK_NY 4
#ifdef H5_HAVE_FILTER_DEFLATE
#define DEFLATE_SIZE_ADJUST(s) (HDceil(((double)(s)) * 1.001) + 12.0)
#endif
/* Temporary filter IDs used for testing */
#define H5Z_FILTER_BOGUS1 305
#define H5Z_FILTER_BOGUS2 306
#define ADD_ON 7
#define FACTOR 3
/* Constants for the overwrite test */
#define OVERWRITE_NDIMS 3
#define OVERWRITE_CHUNK_NX 3
#define OVERWRITE_CHUNK_2NX 6
#define OVERWRITE_CHUNK_NY 2
#define OVERWRITE_VALUE 42
/* Test configurations */
#define CONFIG_LATEST 0x01
#define CONFIG_REOPEN_FILE 0x02
#define CONFIG_REOPEN_DSET 0x04
#define CONFIG_DIRECT_WRITE 0x08
#define CONFIG_DIRECT_READ 0x10
#define CONFIG_END 0x20
/* Defines used in test_single_chunk_latest() */
#define FILE "single_latest.h5"
#define DATASET "dataset"
#define DIM0 4
#define DIM1 32
#define CHUNK0 DIM0
#define CHUNK1 DIM1
/* Local prototypes for filter functions */
static size_t filter_bogus1(unsigned int flags, size_t cd_nelmts, const unsigned int *cd_values,
size_t nbytes, size_t *buf_size, void **buf);
static size_t filter_bogus2(unsigned int flags, size_t cd_nelmts, const unsigned int *cd_values,
size_t nbytes, size_t *buf_size, void **buf);
/* This message derives from H5Z */
const H5Z_class2_t H5Z_BOGUS1[1] = {{
H5Z_CLASS_T_VERS, /* H5Z_class_t version */
H5Z_FILTER_BOGUS1, /* Filter id number */
1, 1, /* Encoding and decoding enabled */
"bogus1", /* Filter name for debugging */
NULL, /* The "can apply" callback */
NULL, /* The "set local" callback */
filter_bogus1, /* The actual filter function */
}};
const H5Z_class2_t H5Z_BOGUS2[1] = {{
H5Z_CLASS_T_VERS, /* H5Z_class_t version */
H5Z_FILTER_BOGUS2, /* Filter id number */
1, 1, /* Encoding and decoding enabled */
"bogus2", /* Filter name for debugging */
NULL, /* The "can apply" callback */
NULL, /* The "set local" callback */
filter_bogus2, /* The actual filter function */
}};
/*-------------------------------------------------------------------------
* Function: test_direct_chunk_write
*
* Purpose: Test the basic functionality of H5Dwrite_chunk
*
* Return: Success: 0
* Failure: 1
*
* Programmer: Raymond Lu
* 30 November 2012
*
*-------------------------------------------------------------------------
*/
#ifdef H5_HAVE_FILTER_DEFLATE
static int
test_direct_chunk_write(hid_t file)
{
hid_t dataspace = -1, dataset = -1;
hid_t mem_space = -1;
hid_t cparms = -1, dxpl = -1;
hsize_t dims[2] = {NX, NY};
hsize_t maxdims[2] = {H5S_UNLIMITED, H5S_UNLIMITED};
hsize_t chunk_dims[2] = {CHUNK_NX, CHUNK_NY};
herr_t status;
int ret;
int data[NX][NY];
int i, j, n;
unsigned filter_mask = 0;
int direct_buf[CHUNK_NX][CHUNK_NY];
int check_chunk[CHUNK_NX][CHUNK_NY];
hsize_t offset[2] = {0, 0};
size_t buf_size = CHUNK_NX * CHUNK_NY * sizeof(int);
const Bytef *z_src = (const Bytef *)(direct_buf);
Bytef * z_dst = NULL; /*destination buffer */
uLongf z_dst_nbytes = (uLongf)DEFLATE_SIZE_ADJUST(buf_size);
uLong z_src_nbytes = (uLong)buf_size;
int aggression = 9; /* Compression aggression setting */
void * outbuf = NULL; /* Pointer to new buffer */
hsize_t start[2]; /* Start of hyperslab */
hsize_t stride[2]; /* Stride of hyperslab */
hsize_t count[2]; /* Block count */
hsize_t block[2]; /* Block sizes */
TESTING("basic functionality of H5Dwrite_chunk");
/*
* Create the data space with unlimited dimensions.
*/
if ((dataspace = H5Screate_simple(RANK, dims, maxdims)) < 0)
goto error;
if ((mem_space = H5Screate_simple(RANK, chunk_dims, NULL)) < 0)
goto error;
/*
* Modify dataset creation properties, i.e. enable chunking and compression
*/
if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
goto error;
if ((status = H5Pset_chunk(cparms, RANK, chunk_dims)) < 0)
goto error;
if ((status = H5Pset_deflate(cparms, (unsigned)aggression)) < 0)
goto error;
/*
* Create a new dataset within the file using cparms
* creation properties.
*/
if ((dataset =
H5Dcreate2(file, DATASETNAME1, H5T_NATIVE_INT, dataspace, H5P_DEFAULT, cparms, H5P_DEFAULT)) < 0)
goto error;
/* Initialize the dataset */
for (i = n = 0; i < NX; i++)
for (j = 0; j < NY; j++)
data[i][j] = n++;
if ((dxpl = H5Pcreate(H5P_DATASET_XFER)) < 0)
goto error;
/*
* Write the data for the dataset. It should stay in the chunk cache.
* It will be evicted from the cache by the H5Dwrite_chunk calls.
*/
if ((status = H5Dwrite(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, data)) < 0)
goto error;
/* Initialize data for one chunk */
for (i = n = 0; i < CHUNK_NX; i++)
for (j = 0; j < CHUNK_NY; j++)
direct_buf[i][j] = n++;
/* Allocate output (compressed) buffer */
outbuf = HDmalloc(z_dst_nbytes);
z_dst = (Bytef *)outbuf;
/* Perform compression from the source to the destination buffer */
ret = compress2(z_dst, &z_dst_nbytes, z_src, z_src_nbytes, aggression);
/* Check for various zlib errors */
if (Z_BUF_ERROR == ret) {
HDfprintf(stderr, "overflow");
goto error;
}
else if (Z_MEM_ERROR == ret) {
HDfprintf(stderr, "deflate memory error");
goto error;
}
else if (Z_OK != ret) {
HDfprintf(stderr, "other deflate error");
goto error;
}
/* Write the compressed chunk data repeatedly to cover all the chunks in the
* dataset, using the direct writing function. */
for (i = 0; i < NX / CHUNK_NX; i++) {
for (j = 0; j < NY / CHUNK_NY; j++) {
status = H5Dwrite_chunk(dataset, dxpl, filter_mask, offset, z_dst_nbytes, outbuf);
offset[1] += CHUNK_NY;
}
offset[0] += CHUNK_NX;
offset[1] = 0;
}
if (outbuf)
HDfree(outbuf);
if (H5Fflush(dataset, H5F_SCOPE_LOCAL) < 0)
goto error;
if (H5Dclose(dataset) < 0)
goto error;
if ((dataset = H5Dopen2(file, DATASETNAME1, H5P_DEFAULT)) < 0)
goto error;
/*
* Select hyperslab for one chunk in the file
*/
start[0] = CHUNK_NX;
start[1] = CHUNK_NY;
stride[0] = 1;
stride[1] = 1;
count[0] = 1;
count[1] = 1;
block[0] = CHUNK_NX;
block[1] = CHUNK_NY;
if ((status = H5Sselect_hyperslab(dataspace, H5S_SELECT_SET, start, stride, count, block)) < 0)
goto error;
/* Read the chunk back */
if ((status = H5Dread(dataset, H5T_NATIVE_INT, mem_space, dataspace, H5P_DEFAULT, check_chunk)) < 0)
goto error;
/* Check that the values read are the same as the values written */
for (i = 0; i < CHUNK_NX; i++) {
for (j = 0; j < CHUNK_NY; j++) {
if (direct_buf[i][j] != check_chunk[i][j]) {
HDprintf(" 1. Read different values than written.");
HDprintf(" At index %d,%d\n", i, j);
HDprintf(" direct_buf=%d, check_chunk=%d\n", direct_buf[i][j], check_chunk[i][j]);
goto error;
}
}
}
/* Reinitialize different data for one chunk */
for (i = 0; i < CHUNK_NX; i++)
for (j = 0; j < CHUNK_NY; j++)
direct_buf[i][j] = i + j;
/* Allocate output (compressed) buffer */
outbuf = HDmalloc(z_dst_nbytes);
z_dst = (Bytef *)outbuf;
/* Perform compression from the source to the destination buffer */
ret = compress2(z_dst, &z_dst_nbytes, z_src, z_src_nbytes, aggression);
/* Check for various zlib errors */
if (Z_BUF_ERROR == ret) {
HDfprintf(stderr, "overflow");
goto error;
}
else if (Z_MEM_ERROR == ret) {
HDfprintf(stderr, "deflate memory error");
goto error;
}
else if (Z_OK != ret) {
HDfprintf(stderr, "other deflate error");
goto error;
}
/* Rewrite the compressed chunk data repeatedly to cover all the chunks in the
* dataset, using the direct writing function. */
offset[0] = offset[1] = 0;
for (i = 0; i < NX / CHUNK_NX; i++) {
for (j = 0; j < NY / CHUNK_NY; j++) {
status = H5Dwrite_chunk(dataset, dxpl, filter_mask, offset, z_dst_nbytes, outbuf);
offset[1] += CHUNK_NY;
}
offset[0] += CHUNK_NX;
offset[1] = 0;
}
if (outbuf)
HDfree(outbuf);
if (H5Fflush(dataset, H5F_SCOPE_LOCAL) < 0)
goto error;
if (H5Dclose(dataset) < 0)
goto error;
if ((dataset = H5Dopen2(file, DATASETNAME1, H5P_DEFAULT)) < 0)
goto error;
/* Read the chunk back */
if ((status = H5Dread(dataset, H5T_NATIVE_INT, mem_space, dataspace, H5P_DEFAULT, check_chunk)) < 0)
goto error;
/* Check that the values read are the same as the values written */
for (i = 0; i < CHUNK_NX; i++) {
for (j = 0; j < CHUNK_NY; j++) {
if (direct_buf[i][j] != check_chunk[i][j]) {
HDprintf(" 2. Read different values than written.");
HDprintf(" At index %d,%d\n", i, j);
HDprintf(" direct_buf=%d, check_chunk=%d\n", direct_buf[i][j], check_chunk[i][j]);
goto error;
}
}
}
/*
* Close/release resources.
*/
H5Dclose(dataset);
H5Sclose(mem_space);
H5Sclose(dataspace);
H5Pclose(cparms);
H5Pclose(dxpl);
PASSED();
return 0;
error:
H5E_BEGIN_TRY
{
H5Dclose(dataset);
H5Sclose(mem_space);
H5Sclose(dataspace);
H5Pclose(cparms);
H5Pclose(dxpl);
}
H5E_END_TRY;
if (outbuf)
HDfree(outbuf);
H5_FAILED();
return 1;
} /* test_direct_chunk_write() */
#endif /* H5_HAVE_FILTER_DEFLATE */
/*-------------------------------------------------------------------------
* Function: test_direct_chunk_overwrite_data
*
* Purpose: Test overwriting a chunk with new data.
*
* Return: Success: 0
* Failure: 1
*
* Programmer: Dana Robinson
* Spring 2017
*
*-------------------------------------------------------------------------
*/
static int
test_direct_chunk_overwrite_data(hid_t fid)
{
size_t buf_size = OVERWRITE_CHUNK_NX * OVERWRITE_CHUNK_NY * sizeof(int16_t);
int16_t data_buf[OVERWRITE_CHUNK_NY][OVERWRITE_CHUNK_NX];
int16_t overwrite_buf[OVERWRITE_CHUNK_NY][OVERWRITE_CHUNK_NX];
uint32_t filter_mask = 0;
hid_t tid = H5T_NATIVE_UINT16;
hid_t dcpl_id = -1;
hid_t sid = -1;
hid_t did = -1;
uint16_t fill_value = 0;
hsize_t dset_dims[] = {1, OVERWRITE_CHUNK_NY, OVERWRITE_CHUNK_2NX};
hsize_t dset_max_dims[] = {H5S_UNLIMITED, OVERWRITE_CHUNK_NY, OVERWRITE_CHUNK_2NX};
hsize_t chunk_dims[] = {1, OVERWRITE_CHUNK_NY, OVERWRITE_CHUNK_NX};
hsize_t offset[] = {0, 0, 0};
hsize_t i, j;
int16_t n;
int16_t read_buf[OVERWRITE_CHUNK_NY][OVERWRITE_CHUNK_2NX];
TESTING("overwriting existing data with H5Dwrite_chunk");
/* Create the dataset's data space */
if ((sid = H5Screate_simple(OVERWRITE_NDIMS, dset_dims, dset_max_dims)) < 0)
FAIL_STACK_ERROR
/* Set chunk size and filll value */
if ((dcpl_id = H5Pcreate(H5P_DATASET_CREATE)) < 0)
FAIL_STACK_ERROR
if (H5Pset_fill_value(dcpl_id, tid, &fill_value) < 0)
FAIL_STACK_ERROR
if (H5Pset_chunk(dcpl_id, OVERWRITE_NDIMS, chunk_dims) < 0)
FAIL_STACK_ERROR
/* Create dataset */
if ((did = H5Dcreate2(fid, DATASETNAME7, tid, sid, H5P_DEFAULT, dcpl_id, H5P_DEFAULT)) < 0)
FAIL_STACK_ERROR
/* Initialize data buffers */
n = 0;
for (i = 0; i < OVERWRITE_CHUNK_NY; i++) {
for (j = 0; j < OVERWRITE_CHUNK_NX; j++) {
data_buf[i][j] = n++;
overwrite_buf[i][j] = OVERWRITE_VALUE;
}
}
/* Write chunk data using the direct write function. */
if (H5Dwrite_chunk(did, H5P_DEFAULT, filter_mask, offset, buf_size, data_buf) < 0)
FAIL_STACK_ERROR
/* Write second chunk. */
offset[2] = OVERWRITE_CHUNK_NX;
if (H5Dwrite_chunk(did, H5P_DEFAULT, filter_mask, offset, buf_size, data_buf) < 0)
FAIL_STACK_ERROR
/* Overwrite first chunk. */
offset[2] = 0;
if (H5Dwrite_chunk(did, H5P_DEFAULT, filter_mask, offset, buf_size, overwrite_buf) < 0)
FAIL_STACK_ERROR
/* Read the data back out */
if (H5Dread(did, tid, H5S_ALL, H5S_ALL, H5P_DEFAULT, read_buf) < 0)
FAIL_STACK_ERROR
/* Ensure that the data are correct in chunk 1 */
for (i = 0; i < OVERWRITE_CHUNK_NY; i++)
for (j = 0; j < OVERWRITE_CHUNK_NX; j++) {
if (read_buf[i][j] != OVERWRITE_VALUE)
TEST_ERROR
}
if (H5Pclose(dcpl_id) < 0)
FAIL_STACK_ERROR
if (H5Sclose(sid) < 0)
FAIL_STACK_ERROR
if (H5Dclose(did) < 0)
FAIL_STACK_ERROR
PASSED();
return 0;
error:
H5E_BEGIN_TRY
{
H5Pclose(dcpl_id);
H5Sclose(sid);
H5Dclose(did);
}
H5E_END_TRY;
H5_FAILED();
return 1;
} /* end test_direct_chunk_overwrite_data() */
/*-------------------------------------------------------------------------
* Function: test_skip_compress_write1
*
* Purpose: Test skipping compression filter when it is the only filter
* for the dataset
*
* Return: Success: 0
* Failure: 1
*
* Programmer: Raymond Lu
* 30 November 2012
*
*-------------------------------------------------------------------------
*/
static int
test_skip_compress_write1(hid_t file)
{
hid_t dataspace = -1, dataset = -1;
hid_t mem_space = -1;
hid_t cparms = -1, dxpl = -1;
hsize_t dims[2] = {NX, NY};
hsize_t maxdims[2] = {H5S_UNLIMITED, H5S_UNLIMITED};
hsize_t chunk_dims[2] = {CHUNK_NX, CHUNK_NY};
herr_t status;
int i, j, n;
unsigned filter_mask = 0;
int direct_buf[CHUNK_NX][CHUNK_NY];
int check_chunk[CHUNK_NX][CHUNK_NY];
hsize_t offset[2] = {0, 0};
size_t buf_size = CHUNK_NX * CHUNK_NY * sizeof(int);
int aggression = 9; /* Compression aggression setting */
unsigned read_filter_mask = 0; /* filter mask after direct read */
int read_direct_buf[CHUNK_NX][CHUNK_NY];
hsize_t read_buf_size = 0; /* buf size */
hsize_t start[2]; /* Start of hyperslab */
hsize_t stride[2]; /* Stride of hyperslab */
hsize_t count[2]; /* Block count */
hsize_t block[2]; /* Block sizes */
TESTING("skipping compression filter for H5Dwrite_chunk/H5Dread_chunk");
/*
* Create the data space with unlimited dimensions.
*/
if ((dataspace = H5Screate_simple(RANK, dims, maxdims)) < 0)
goto error;
if ((mem_space = H5Screate_simple(RANK, chunk_dims, NULL)) < 0)
goto error;
/*
* Modify dataset creation properties, i.e. enable chunking and compression
*/
if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
goto error;
if ((status = H5Pset_chunk(cparms, RANK, chunk_dims)) < 0)
goto error;
if ((status = H5Pset_deflate(cparms, (unsigned)aggression)) < 0)
goto error;
/*
* Create a new dataset within the file using cparms
* creation properties.
*/
if ((dataset =
H5Dcreate2(file, DATASETNAME2, H5T_NATIVE_INT, dataspace, H5P_DEFAULT, cparms, H5P_DEFAULT)) < 0)
goto error;
if ((dxpl = H5Pcreate(H5P_DATASET_XFER)) < 0)
goto error;
/* Initialize data for one chunk */
for (i = n = 0; i < CHUNK_NX; i++)
for (j = 0; j < CHUNK_NY; j++) {
direct_buf[i][j] = n++;
}
/* write the uncompressed chunk data repeatedly to dataset, using the direct writing function.
* Indicate skipping the compression filter. */
offset[0] = CHUNK_NX;
offset[1] = CHUNK_NY;
filter_mask = 0x00000001;
if ((status = H5Dwrite_chunk(dataset, dxpl, filter_mask, offset, buf_size, direct_buf)) < 0)
goto error;
if (H5Fflush(dataset, H5F_SCOPE_LOCAL) < 0)
goto error;
if (H5Dclose(dataset) < 0)
goto error;
if ((dataset = H5Dopen2(file, DATASETNAME2, H5P_DEFAULT)) < 0)
goto error;
/*
* Select hyperslab for the chunk just written in the file
*/
start[0] = CHUNK_NX;
start[1] = CHUNK_NY;
stride[0] = 1;
stride[1] = 1;
count[0] = 1;
count[1] = 1;
block[0] = CHUNK_NX;
block[1] = CHUNK_NY;
if ((status = H5Sselect_hyperslab(dataspace, H5S_SELECT_SET, start, stride, count, block)) < 0)
goto error;
/* Read the chunk back */
if ((status = H5Dread(dataset, H5T_NATIVE_INT, mem_space, dataspace, H5P_DEFAULT, check_chunk)) < 0)
goto error;
/* Check that the values read are the same as the values written */
for (i = 0; i < CHUNK_NX; i++) {
for (j = 0; j < CHUNK_NY; j++) {
if (direct_buf[i][j] != check_chunk[i][j]) {
HDprintf(" 1. Read different values than written.");
HDprintf(" At index %d,%d\n", i, j);
HDprintf(" direct_buf=%d, check_chunk=%d\n", direct_buf[i][j], check_chunk[i][j]);
goto error;
}
}
}
/* Query chunk storage size */
if ((status = H5Dget_chunk_storage_size(dataset, offset, &read_buf_size)) < 0)
goto error;
if (read_buf_size != buf_size)
goto error;
/* Read the raw chunk back */
HDmemset(&read_direct_buf, 0, sizeof(read_direct_buf));
if ((status = H5Dread_chunk(dataset, H5P_DEFAULT, offset, &read_filter_mask, read_direct_buf)) < 0)
goto error;
if (read_filter_mask != filter_mask)
goto error;
/* Check that the direct chunk read is the same as the chunk written */
for (i = 0; i < CHUNK_NX; i++) {
for (j = 0; j < CHUNK_NY; j++) {
if (direct_buf[i][j] != read_direct_buf[i][j]) {
HDprintf(" 1. Read different values than written.");
HDprintf(" At index %d,%d\n", i, j);
HDprintf(" direct_buf=%d, read_direct_buf=%d\n", direct_buf[i][j], read_direct_buf[i][j]);
goto error;
}
}
}
/*
* Close/release resources.
*/
H5Dclose(dataset);
H5Sclose(mem_space);
H5Sclose(dataspace);
H5Pclose(cparms);
H5Pclose(dxpl);
PASSED();
return 0;
error:
H5E_BEGIN_TRY
{
H5Dclose(dataset);
H5Sclose(mem_space);
H5Sclose(dataspace);
H5Pclose(cparms);
H5Pclose(dxpl);
}
H5E_END_TRY;
H5_FAILED();
return 1;
} /* test_skip_compress_write1() */
/*-------------------------------------------------------------------------
* Function: filter_bogus1
*
* Purpose: A bogus filter that adds ADD_ON to the original value
*
* Return: Success: Data chunk size
*
* Programmer: Raymond Lu
* 30 November 2012
*
*-------------------------------------------------------------------------
*/
static size_t
filter_bogus1(unsigned int flags, size_t H5_ATTR_UNUSED cd_nelmts,
const unsigned int H5_ATTR_UNUSED *cd_values, size_t nbytes, size_t *buf_size, void **buf)
{
int * int_ptr = (int *)*buf; /* Pointer to the data values */
size_t buf_left = *buf_size; /* Amount of data buffer left to process */
if (flags & H5Z_FLAG_REVERSE) { /* read */
/* Subtract the "add on" value to all the data values */
while (buf_left > 0) {
*int_ptr++ -= (int)ADD_ON;
buf_left -= sizeof(int);
} /* end while */
} /* end if */
else { /* write */
/* Add the "add on" value to all the data values */
while (buf_left > 0) {
*int_ptr++ += (int)ADD_ON;
buf_left -= sizeof(int);
} /* end while */
} /* end else */
return nbytes;
} /* filter_bogus1() */
/*-------------------------------------------------------------------------
* Function: filter_bogus2
*
* Purpose: A bogus filter that multiplies the original value by FACTOR.
*
* Return: Success: Data chunk size
*
* Programmer: Raymond Lu
* 30 November 2012
*-------------------------------------------------------------------------
*/
static size_t
filter_bogus2(unsigned int flags, size_t H5_ATTR_UNUSED cd_nelmts,
const unsigned int H5_ATTR_UNUSED *cd_values, size_t nbytes, size_t *buf_size, void **buf)
{
int * int_ptr = (int *)*buf; /* Pointer to the data values */
size_t buf_left = *buf_size; /* Amount of data buffer left to process */
if (flags & H5Z_FLAG_REVERSE) { /* read */
/* Subtract the "add on" value to all the data values */
while (buf_left > 0) {
*int_ptr++ /= (int)FACTOR;
buf_left -= sizeof(int);
} /* end while */
} /* end if */
else { /* write */
/* Add the "add on" value to all the data values */
while (buf_left > 0) {
*int_ptr++ *= (int)FACTOR;
buf_left -= sizeof(int);
} /* end while */
} /* end else */
return nbytes;
} /* filter_bogus2() */
/*-------------------------------------------------------------------------
* Function: test_skip_compress_write2
*
* Purpose: Test skipping compression filter when there are three filters
* for the dataset
*
* Return: Success: 0
* Failure: 1
*
* Programmer: Raymond Lu
* 30 November 2012
*
*-------------------------------------------------------------------------
*/
static int
test_skip_compress_write2(hid_t file)
{
hid_t dataspace = -1, dataset = -1;
hid_t mem_space = -1;
hid_t cparms = -1, dxpl = -1;
hsize_t dims[2] = {NX, NY};
hsize_t maxdims[2] = {H5S_UNLIMITED, H5S_UNLIMITED};
hsize_t chunk_dims[2] = {CHUNK_NX, CHUNK_NY};
herr_t status;
int i, j, n;
unsigned filter_mask = 0; /* orig filter mask */
int origin_direct_buf[CHUNK_NX][CHUNK_NY];
int direct_buf[CHUNK_NX][CHUNK_NY];
int check_chunk[CHUNK_NX][CHUNK_NY];
hsize_t offset[2] = {0, 0};
size_t buf_size = CHUNK_NX * CHUNK_NY * sizeof(int);
int aggression = 9; /* Compression aggression setting */
unsigned read_filter_mask = 0; /* filter mask after direct read */
int read_direct_buf[CHUNK_NX][CHUNK_NY];
hsize_t read_buf_size = 0; /* buf size */
hsize_t start[2]; /* Start of hyperslab */
hsize_t stride[2]; /* Stride of hyperslab */
hsize_t count[2]; /* Block count */
hsize_t block[2]; /* Block sizes */
TESTING("skipping compression filters but keep two other filters");
/*
* Create the data space with unlimited dimensions.
*/
if ((dataspace = H5Screate_simple(RANK, dims, maxdims)) < 0)
goto error;
if ((mem_space = H5Screate_simple(RANK, chunk_dims, NULL)) < 0)
goto error;
/*
* Modify dataset creation properties, i.e. enable chunking and compression.
* The order of filters is bogus 1 + deflate + bogus 2.
*/
if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
goto error;
if ((status = H5Pset_chunk(cparms, RANK, chunk_dims)) < 0)
goto error;
/* Register and enable first bogus filter */
if (H5Zregister(H5Z_BOGUS1) < 0)
goto error;
if (H5Pset_filter(cparms, H5Z_FILTER_BOGUS1, 0, (size_t)0, NULL) < 0)
goto error;
/* Enable compression filter */
if ((status = H5Pset_deflate(cparms, (unsigned)aggression)) < 0)
goto error;
/* Register and enable second bogus filter */
if (H5Zregister(H5Z_BOGUS2) < 0)
goto error;
if (H5Pset_filter(cparms, H5Z_FILTER_BOGUS2, 0, (size_t)0, NULL) < 0)
goto error;
/*
* Create a new dataset within the file using cparms
* creation properties.
*/
if ((dataset =
H5Dcreate2(file, DATASETNAME3, H5T_NATIVE_INT, dataspace, H5P_DEFAULT, cparms, H5P_DEFAULT)) < 0)
goto error;
if ((dxpl = H5Pcreate(H5P_DATASET_XFER)) < 0)
goto error;
/* Initialize data for one chunk. Apply operations of two bogus filters to the chunk */
for (i = n = 0; i < CHUNK_NX; i++)
for (j = 0; j < CHUNK_NY; j++) {
origin_direct_buf[i][j] = n++;
direct_buf[i][j] = (origin_direct_buf[i][j] + ADD_ON) * FACTOR;
}
/* write the uncompressed chunk data repeatedly to dataset, using the direct writing function.
* Indicate skipping the compression filter but keep the other two bogus filters */
offset[0] = CHUNK_NX;
offset[1] = CHUNK_NY;
/* compression filter is the middle one to be skipped */
filter_mask = 0x00000002;
if ((status = H5Dwrite_chunk(dataset, dxpl, filter_mask, offset, buf_size, direct_buf)) < 0)
goto error;
if (H5Fflush(dataset, H5F_SCOPE_LOCAL) < 0)
goto error;
if (H5Dclose(dataset) < 0)
goto error;
if ((dataset = H5Dopen2(file, DATASETNAME3, H5P_DEFAULT)) < 0)
goto error;
/*
* Select hyperslab for one chunk in the file
*/
start[0] = CHUNK_NX;
start[1] = CHUNK_NY;
stride[0] = 1;
stride[1] = 1;
count[0] = 1;
count[1] = 1;
block[0] = CHUNK_NX;
block[1] = CHUNK_NY;
if ((status = H5Sselect_hyperslab(dataspace, H5S_SELECT_SET, start, stride, count, block)) < 0)
goto error;
/* Read the chunk back */
if ((status = H5Dread(dataset, H5T_NATIVE_INT, mem_space, dataspace, H5P_DEFAULT, check_chunk)) < 0)
goto error;
/* Check that the values read are the same as the values written */
for (i = 0; i < CHUNK_NX; i++) {
for (j = 0; j < CHUNK_NY; j++) {
if (origin_direct_buf[i][j] != check_chunk[i][j]) {
HDprintf(" 1. Read different values than written.");
HDprintf(" At index %d,%d\n", i, j);
HDprintf(" origin_direct_buf=%d, check_chunk=%d\n", origin_direct_buf[i][j],
check_chunk[i][j]);
goto error;
}
}
}
/* Query chunk storage size */
if ((status = H5Dget_chunk_storage_size(dataset, offset, &read_buf_size)) < 0)
goto error;
if (read_buf_size != buf_size)
goto error;
/* Read the raw chunk back */
HDmemset(&read_direct_buf, 0, sizeof(read_direct_buf));
if ((status = H5Dread_chunk(dataset, H5P_DEFAULT, offset, &read_filter_mask, read_direct_buf)) < 0)
goto error;
if (read_filter_mask != filter_mask)
goto error;
/* Check that the direct chunk read is the same as the chunk written */
for (i = 0; i < CHUNK_NX; i++) {
for (j = 0; j < CHUNK_NY; j++) {
if (direct_buf[i][j] != read_direct_buf[i][j]) {
HDprintf(" 1. Read different values than written.");
HDprintf(" At index %d,%d\n", i, j);
HDprintf(" direct_buf=%d, read_direct_buf=%d\n", direct_buf[i][j], read_direct_buf[i][j]);
goto error;
}
}
}
/*
* Close/release resources.
*/
H5Dclose(dataset);
H5Sclose(mem_space);
H5Sclose(dataspace);
H5Pclose(cparms);
H5Pclose(dxpl);
PASSED();
return 0;
error:
H5E_BEGIN_TRY
{
H5Dclose(dataset);
H5Sclose(mem_space);
H5Sclose(dataspace);
H5Pclose(cparms);
H5Pclose(dxpl);
}
H5E_END_TRY;
H5_FAILED();
return 1;
} /* test_skip_compress_write2() */
/*-------------------------------------------------------------------------
* Function: test_data_conv
*
* Purpose: Test data conversion
*
* Return: Success: 0
* Failure: 1
*
* Programmer: Raymond Lu
* 30 November 2012
*
*-------------------------------------------------------------------------
*/
static int
test_data_conv(hid_t file)
{
typedef struct {
int a, b, c[4], d, e;
} src_type_t;
typedef struct {
int a, c[4], e;
} dst_type_t;
hid_t dataspace = -1, dataset = -1;
hid_t mem_space = -1;
hid_t cparms = -1, dxpl = -1;
hsize_t dims[2] = {NX, NY};
hsize_t maxdims[2] = {H5S_UNLIMITED, H5S_UNLIMITED};
hsize_t chunk_dims[2] = {CHUNK_NX, CHUNK_NY};
herr_t status;
int i, j, n;
const hsize_t four = 4;
hid_t st = -1, dt = -1;
hid_t array_dt;
unsigned filter_mask = 0;
src_type_t direct_buf[CHUNK_NX][CHUNK_NY];
dst_type_t check_chunk[CHUNK_NX][CHUNK_NY];
src_type_t read_chunk[CHUNK_NX][CHUNK_NY]; /* For H5Dread_chunk */
hsize_t offset[2] = {0, 0};
size_t buf_size = CHUNK_NX * CHUNK_NY * sizeof(src_type_t);
hsize_t start[2]; /* Start of hyperslab */
hsize_t stride[2]; /* Stride of hyperslab */
hsize_t count[2]; /* Block count */
hsize_t block[2]; /* Block sizes */
TESTING("data conversion for H5Dwrite_chunk/H5Dread_chunk");
/*
* Create the data space with unlimited dimensions.
*/
if ((dataspace = H5Screate_simple(RANK, dims, maxdims)) < 0)
goto error;
if ((mem_space = H5Screate_simple(RANK, chunk_dims, NULL)) < 0)
goto error;
/*
* Modify dataset creation properties, i.e. enable chunking
*/
if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
goto error;
if ((status = H5Pset_chunk(cparms, RANK, chunk_dims)) < 0)
goto error;
/* Build hdf5 datatypes */
array_dt = H5Tarray_create2(H5T_NATIVE_INT, 1, &four);
if ((st = H5Tcreate(H5T_COMPOUND, sizeof(src_type_t))) < 0 ||
H5Tinsert(st, "a", HOFFSET(src_type_t, a), H5T_NATIVE_INT) < 0 ||
H5Tinsert(st, "b", HOFFSET(src_type_t, b), H5T_NATIVE_INT) < 0 ||
H5Tinsert(st, "c", HOFFSET(src_type_t, c), array_dt) < 0 ||
H5Tinsert(st, "d", HOFFSET(src_type_t, d), H5T_NATIVE_INT) < 0 ||
H5Tinsert(st, "e", HOFFSET(src_type_t, e), H5T_NATIVE_INT) < 0)
goto error;
if (H5Tclose(array_dt) < 0)
goto error;
array_dt = H5Tarray_create2(H5T_NATIVE_INT, 1, &four);
if ((dt = H5Tcreate(H5T_COMPOUND, sizeof(dst_type_t))) < 0 ||
H5Tinsert(dt, "a", HOFFSET(dst_type_t, a), H5T_NATIVE_INT) < 0 ||
H5Tinsert(dt, "c", HOFFSET(dst_type_t, c), array_dt) < 0 ||
H5Tinsert(dt, "e", HOFFSET(dst_type_t, e), H5T_NATIVE_INT) < 0)
goto error;
if (H5Tclose(array_dt) < 0)
goto error;
/*
* Create a new dataset within the file using cparms
* creation properties.
*/
if ((dataset = H5Dcreate2(file, DATASETNAME4, st, dataspace, H5P_DEFAULT, cparms, H5P_DEFAULT)) < 0)
goto error;
if ((dxpl = H5Pcreate(H5P_DATASET_XFER)) < 0)
goto error;
/* Initialize data for one chunk */
for (i = n = 0; i < CHUNK_NX; i++) {
for (j = 0; j < CHUNK_NY; j++) {
(direct_buf[i][j]).a = i * j + 0;
(direct_buf[i][j]).b = i * j + 1;
(direct_buf[i][j]).c[0] = i * j + 2;
(direct_buf[i][j]).c[1] = i * j + 3;
(direct_buf[i][j]).c[2] = i * j + 4;
(direct_buf[i][j]).c[3] = i * j + 5;
(direct_buf[i][j]).d = i * j + 6;
(direct_buf[i][j]).e = i * j + 7;
}
}
/* write the chunk data to dataset, using the direct writing function.
* There should be no data conversion involved. */
offset[0] = CHUNK_NX;
offset[1] = CHUNK_NY;
if ((status = H5Dwrite_chunk(dataset, dxpl, filter_mask, offset, buf_size, direct_buf)) < 0)
goto error;
if (H5Fflush(dataset, H5F_SCOPE_LOCAL) < 0)
goto error;
if (H5Dclose(dataset) < 0)
goto error;
if ((dataset = H5Dopen2(file, DATASETNAME4, H5P_DEFAULT)) < 0)
goto error;
/* Use H5Dread_chunk() to read the uncompressed data */
if ((status = H5Dread_chunk(dataset, dxpl, offset, &filter_mask, read_chunk)) < 0)
goto error;
/* Check that the values read are the same as the values written */
for (i = 0; i < CHUNK_NX; i++) {
for (j = 0; j < CHUNK_NY; j++) {
if ((direct_buf[i][j]).a != (read_chunk[i][j]).a ||
(direct_buf[i][j]).b != (read_chunk[i][j]).b ||
(direct_buf[i][j]).c[0] != (read_chunk[i][j]).c[0] ||
(direct_buf[i][j]).c[1] != (read_chunk[i][j]).c[1] ||
(direct_buf[i][j]).c[2] != (read_chunk[i][j]).c[2] ||
(direct_buf[i][j]).c[3] != (read_chunk[i][j]).c[3] ||
(direct_buf[i][j]).d != (read_chunk[i][j]).d ||
(direct_buf[i][j]).e != (read_chunk[i][j]).e) {
HDprintf(" 1. Read different values than written.");
HDprintf(" At index %d,%d\n", i, j);
HDprintf(" src={a=%d, b=%d, c=[%d,%d,%d,%d], d=%d, e=%d\n", (direct_buf[i][j]).a,
(direct_buf[i][j]).b, (direct_buf[i][j]).c[0], (direct_buf[i][j]).c[1],
(direct_buf[i][j]).c[2], (direct_buf[i][j]).c[3], (direct_buf[i][j]).d,
(direct_buf[i][j]).e);
HDprintf(" dst={a=%d, b=%d, c=[%d,%d,%d,%d], d=%d, e=%d\n", (read_chunk[i][j]).a,
(read_chunk[i][j]).b, (read_chunk[i][j]).c[0], (read_chunk[i][j]).c[1],
(read_chunk[i][j]).c[2], (read_chunk[i][j]).c[3], (read_chunk[i][j]).d,
(read_chunk[i][j]).e);
goto error;
}
}
}
/*
* Select hyperslab for the chunk just written in the file
*/
start[0] = CHUNK_NX;
start[1] = CHUNK_NY;
stride[0] = 1;
stride[1] = 1;
count[0] = 1;
count[1] = 1;
block[0] = CHUNK_NX;
block[1] = CHUNK_NY;
if ((status = H5Sselect_hyperslab(dataspace, H5S_SELECT_SET, start, stride, count, block)) < 0)
goto error;
/* Read the chunk back. Data should be converted */
if ((status = H5Dread(dataset, dt, mem_space, dataspace, H5P_DEFAULT, check_chunk)) < 0)
goto error;
/* Check that the values read are the same as the values written */
for (i = 0; i < CHUNK_NX; i++) {
for (j = 0; j < CHUNK_NY; j++) {
if ((direct_buf[i][j]).a != (check_chunk[i][j]).a ||
(direct_buf[i][j]).c[0] != (check_chunk[i][j]).c[0] ||
(direct_buf[i][j]).c[1] != (check_chunk[i][j]).c[1] ||
(direct_buf[i][j]).c[2] != (check_chunk[i][j]).c[2] ||
(direct_buf[i][j]).c[3] != (check_chunk[i][j]).c[3] ||
(direct_buf[i][j]).e != (check_chunk[i][j]).e) {
HDprintf(" 1. Read different values than written.");
HDprintf(" At index %d,%d\n", i, j);
HDprintf(" src={a=%d, b=%d, c=[%d,%d,%d,%d], d=%d, e=%d\n", (direct_buf[i][j]).a,
(direct_buf[i][j]).b, (direct_buf[i][j]).c[0], (direct_buf[i][j]).c[1],
(direct_buf[i][j]).c[2], (direct_buf[i][j]).c[3], (direct_buf[i][j]).d,
(direct_buf[i][j]).e);
HDprintf(" dst={a=%d, c=[%d,%d,%d,%d], e=%d\n", (check_chunk[i][j]).a,
(check_chunk[i][j]).c[0], (check_chunk[i][j]).c[1], (check_chunk[i][j]).c[2],
(check_chunk[i][j]).c[3], (check_chunk[i][j]).e);
goto error;
}
}
}
/*
* Close/release resources.
*/
H5Dclose(dataset);
H5Sclose(mem_space);
H5Sclose(dataspace);
H5Pclose(cparms);
H5Pclose(dxpl);
H5Tclose(st);
H5Tclose(dt);
PASSED();
return 0;
error:
H5E_BEGIN_TRY
{
H5Dclose(dataset);
H5Sclose(mem_space);
H5Sclose(dataspace);
H5Pclose(cparms);
H5Pclose(dxpl);
H5Tclose(st);
H5Tclose(dt);
}
H5E_END_TRY;
H5_FAILED();
return 1;
} /* test_data_conv() */
/*-------------------------------------------------------------------------
* Function: test_invalid_parameters
*
* Purpose: Test invalid parameters for H5Dwrite_chunk and H5Dread_chunk
*
* Return: Success: 0
* Failure: 1
*
* Programmer: Raymond Lu
* 30 November 2012
*
*-------------------------------------------------------------------------
*/
static int
test_invalid_parameters(hid_t file)
{
hid_t dataspace = -1, dataset = -1;
hid_t mem_space = -1;
hid_t cparms = -1, dxpl = -1;
hsize_t dims[2] = {NX, NY};
hsize_t chunk_dims[2] = {CHUNK_NX, CHUNK_NY};
herr_t status;
int i, j, n;
unsigned filter_mask = 0;
int direct_buf[CHUNK_NX][CHUNK_NY];
hsize_t offset[2] = {0, 0};
size_t buf_size = CHUNK_NX * CHUNK_NY * sizeof(int);
int aggression = 9; /* Compression aggression setting */
hsize_t chunk_nbytes; /* Chunk size */
TESTING("invalid parameters for H5Dwrite_chunk/H5Dread_chunk");
/*
* Create the data space with unlimited dimensions.
*/
if ((dataspace = H5Screate_simple(RANK, dims, NULL)) < 0)
goto error;
if ((mem_space = H5Screate_simple(RANK, chunk_dims, NULL)) < 0)
goto error;
/*
* Modify dataset creation properties
*/
if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
goto error;
/*
* Create a new contiguous dataset to verify H5Dwrite_chunk/H5Dread_chunk doesn't work
*/
if ((dataset =
H5Dcreate2(file, DATASETNAME5, H5T_NATIVE_INT, dataspace, H5P_DEFAULT, cparms, H5P_DEFAULT)) < 0)
goto error;
if ((dxpl = H5Pcreate(H5P_DATASET_XFER)) < 0)
goto error;
/* Initialize data for one chunk */
for (i = n = 0; i < CHUNK_NX; i++)
for (j = 0; j < CHUNK_NY; j++) {
direct_buf[i][j] = n++;
}
/* Try to write the chunk data to contiguous dataset. It should fail */
offset[0] = CHUNK_NX;
offset[1] = CHUNK_NY;
H5E_BEGIN_TRY
{
if ((status = H5Dwrite_chunk(dataset, dxpl, filter_mask, offset, buf_size, direct_buf)) != FAIL)
goto error;
}
H5E_END_TRY;
/* Try to get chunk size for a contiguous dataset. It should fail */
H5E_BEGIN_TRY
{
if ((status = H5Dget_chunk_storage_size(dataset, offset, &chunk_nbytes)) != FAIL)
goto error;
}
H5E_END_TRY;
/* Try to H5Dread_chunk from the contiguous dataset. It should fail */
H5E_BEGIN_TRY
{
if ((status = H5Dread_chunk(dataset, dxpl, offset, &filter_mask, direct_buf)) != FAIL)
goto error;
}
H5E_END_TRY;
if (H5Dclose(dataset) < 0)
goto error;
/* Create a chunked dataset with compression filter */
if ((status = H5Pset_chunk(cparms, RANK, chunk_dims)) < 0)
goto error;
if ((status = H5Pset_deflate(cparms, (unsigned)aggression)) < 0)
goto error;
/*
* Create a new dataset within the file using cparms
* creation properties.
*/
if ((dataset =
H5Dcreate2(file, DATASETNAME6, H5T_NATIVE_INT, dataspace, H5P_DEFAULT, cparms, H5P_DEFAULT)) < 0)
goto error;
/* Check invalid dataset ID for H5Dwrite_chunk and H5Dread_chunk */
H5E_BEGIN_TRY
{
if ((status = H5Dwrite_chunk((hid_t)-1, dxpl, filter_mask, offset, buf_size, direct_buf)) != FAIL)
goto error;
}
H5E_END_TRY;
H5E_BEGIN_TRY
{
if ((status = H5Dread_chunk((hid_t)-1, dxpl, offset, &filter_mask, direct_buf)) != FAIL)
goto error;
}
H5E_END_TRY;
/* Check invalid DXPL ID for H5Dwrite_chunk and H5Dread_chunk */
H5E_BEGIN_TRY
{
if ((status = H5Dwrite_chunk(dataset, (hid_t)-1, filter_mask, offset, buf_size, direct_buf)) != FAIL)
goto error;
}
H5E_END_TRY;
H5E_BEGIN_TRY
{
if ((status = H5Dread_chunk(dataset, (hid_t)-1, offset, &filter_mask, direct_buf)) != FAIL)
goto error;
}
H5E_END_TRY;
/* Check invalid OFFSET for H5Dwrite_chunk and H5Dread_chunk */
H5E_BEGIN_TRY
{
if ((status = H5Dwrite_chunk(dataset, dxpl, filter_mask, NULL, buf_size, direct_buf)) != FAIL)
goto error;
}
H5E_END_TRY;
H5E_BEGIN_TRY
{
if ((status = H5Dread_chunk(dataset, dxpl, NULL, &filter_mask, direct_buf)) != FAIL)
goto error;
}
H5E_END_TRY;
/* Check when OFFSET is out of dataset range for H5Dwrite_chunk and H5Dread_chunk */
offset[0] = NX + 1;
offset[1] = NY;
H5E_BEGIN_TRY
{
if ((status = H5Dwrite_chunk(dataset, dxpl, filter_mask, offset, buf_size, direct_buf)) != FAIL)
goto error;
}
H5E_END_TRY;
H5E_BEGIN_TRY
{
if ((status = H5Dread_chunk(dataset, dxpl, offset, &filter_mask, direct_buf)) != FAIL)
goto error;
}
H5E_END_TRY;
/* Check when OFFSET is not on chunk boundary for H5Dwrite_chunk and H5Dread_chunk */
offset[0] = CHUNK_NX;
offset[1] = CHUNK_NY + 1;
H5E_BEGIN_TRY
{
if ((status = H5Dwrite_chunk(dataset, dxpl, filter_mask, offset, buf_size, direct_buf)) != FAIL)
goto error;
}
H5E_END_TRY;
H5E_BEGIN_TRY
{
if ((status = H5Dread_chunk(dataset, dxpl, offset, &filter_mask, direct_buf)) != FAIL)
goto error;
}
H5E_END_TRY;
/* Check invalid buffer size for H5Dwrite_chunk only */
offset[0] = CHUNK_NX;
offset[1] = CHUNK_NY;
buf_size = 0;
H5E_BEGIN_TRY
{
if ((status = H5Dwrite_chunk(dataset, dxpl, filter_mask, offset, buf_size, direct_buf)) != FAIL)
goto error;
}
H5E_END_TRY;
/* Check invalid data buffer for H5Dwrite_chunk and H5Dread_chunk */
buf_size = CHUNK_NX * CHUNK_NY * sizeof(int);
H5E_BEGIN_TRY
{
if ((status = H5Dwrite_chunk(dataset, dxpl, filter_mask, offset, buf_size, NULL)) != FAIL)
goto error;
}
H5E_END_TRY;
H5E_BEGIN_TRY
{
if ((status = H5Dread_chunk(dataset, dxpl, offset, &filter_mask, NULL)) != FAIL)
goto error;
}
H5E_END_TRY;
if (H5Dclose(dataset) < 0)
goto error;
/*
* Close/release resources.
*/
H5Sclose(mem_space);
H5Sclose(dataspace);
H5Pclose(cparms);
H5Pclose(dxpl);
PASSED();
return 0;
error:
H5E_BEGIN_TRY
{
H5Dclose(dataset);
H5Sclose(mem_space);
H5Sclose(dataspace);
H5Pclose(cparms);
H5Pclose(dxpl);
}
H5E_END_TRY;
H5_FAILED();
return 1;
} /* test_invalid_parameters() */
/*-------------------------------------------------------------------------
* Function: test_direct_chunk_read_no_cache
*
* Purpose: Test the basic functionality of H5Dread_chunk with the
* chunk cache disabled.
*
* Return: Success: 0
* Failure: 1
*
* Programmer: Matthew Strong (GE Healthcare)
* 14 February 2016
*
*-------------------------------------------------------------------------
*/
#ifdef H5_HAVE_FILTER_DEFLATE
static int
test_direct_chunk_read_no_cache(hid_t file)
{
hid_t dataspace = -1, dataset = -1;
hid_t mem_space = -1;
hid_t cparms = -1, dxpl = -1, dapl = -1;
hsize_t dims[2] = {NX, NY};
hsize_t maxdims[2] = {H5S_UNLIMITED, H5S_UNLIMITED};
hsize_t chunk_dims[2] = {CHUNK_NX, CHUNK_NY};
herr_t status; /* status from H5 function calls */
int ret; /* deflate return status */
int data[NX][NY];
int i, j, k, l, n; /* local index variables */
unsigned filter_mask = 0; /* filter mask returned from H5Dread_chunk */
int direct_buf[CHUNK_NX][CHUNK_NY]; /* chunk read with H5Dread and manually decompressed */
int check_chunk[CHUNK_NX][CHUNK_NY]; /* chunk read with H5Dread */
hsize_t offset[2]; /* chunk offset used for H5Dread_chunk */
size_t buf_size = CHUNK_NX * CHUNK_NY * sizeof(int);
Bytef *z_src = NULL; /* source buffer */
uLongf z_src_nbytes = (uLongf)DEFLATE_SIZE_ADJUST(buf_size);
Bytef *z_dst = (Bytef *)(direct_buf);
uLong z_dst_nbytes = (uLong)buf_size;
int aggression = 9; /* Compression aggression setting */
void * outbuf = NULL; /* Pointer to new buffer */
hsize_t start[2]; /* Start of hyperslab */
hsize_t stride[2]; /* Stride of hyperslab */
hsize_t count[2]; /* Block count */
hsize_t block[2]; /* Block sizes */
TESTING("basic functionality of H5Dread_chunk (chunk cache disabled)");
/* Create the data space with unlimited dimensions. */
if ((dataspace = H5Screate_simple(RANK, dims, maxdims)) < 0)
goto error;
if ((mem_space = H5Screate_simple(RANK, chunk_dims, NULL)) < 0)
goto error;
/* Modify dataset creation properties, i.e. enable chunking and compression */
if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
goto error;
if ((status = H5Pset_chunk(cparms, RANK, chunk_dims)) < 0)
goto error;
if ((status = H5Pset_deflate(cparms, (unsigned)aggression)) < 0)
goto error;
if ((dapl = H5Pcreate(H5P_DATASET_ACCESS)) < 0)
goto error;
/* Disable chunk cache by setting number of slots to 0 */
if ((status = H5Pset_chunk_cache(dapl, 0, H5D_CHUNK_CACHE_NBYTES_DEFAULT, H5D_CHUNK_CACHE_W0_DEFAULT)) <
0)
goto error;
/* Create a new dataset within the file using cparms creation properties. */
if ((dataset = H5Dcreate2(file, DATASETNAME8, H5T_NATIVE_INT, dataspace, H5P_DEFAULT, cparms, dapl)) < 0)
goto error;
/* Initialize the dataset */
for (i = n = 0; i < NX; i++)
for (j = 0; j < NY; j++)
data[i][j] = n++;
if ((dxpl = H5Pcreate(H5P_DATASET_XFER)) < 0)
goto error;
/* Write the data for the dataset.
* Data will skip chunk cache and go directly to disk. */
if ((status = H5Dwrite(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, data)) < 0)
goto error;
/* Allocate output (compressed) buffer */
outbuf = HDmalloc(z_src_nbytes);
z_src = (Bytef *)outbuf;
/* For each chunk in the dataset, compare the result of H5Dread and H5Dread_chunk. */
for (i = 0; i < NX / CHUNK_NX; i++) {
for (j = 0; j < NY / CHUNK_NY; j++) {
/* Select hyperslab for one chunk in the file */
start[0] = (hsize_t)i * CHUNK_NX;
start[1] = (hsize_t)j * CHUNK_NY;
stride[0] = 1;
stride[1] = 1;
count[0] = 1;
count[1] = 1;
block[0] = CHUNK_NX;
block[1] = CHUNK_NY;
/* Hyperslab selection equals single chunk */
if ((status = H5Sselect_hyperslab(dataspace, H5S_SELECT_SET, start, stride, count, block)) < 0)
goto error;
/* Read the chunk back */
if ((status = H5Dread(dataset, H5T_NATIVE_INT, mem_space, dataspace, H5P_DEFAULT, check_chunk)) <
0)
goto error;
offset[0] = (hsize_t)i * CHUNK_NX;
offset[1] = (hsize_t)j * CHUNK_NY;
/* Read the compressed chunk back using the direct read function. */
if ((status = H5Dread_chunk(dataset, dxpl, offset, &filter_mask, outbuf)) < 0)
goto error;
/* Check filter mask return value */
if (filter_mask != 0)
goto error;
/* Perform decompression from the source to the destination buffer */
ret = uncompress(z_dst, &z_dst_nbytes, z_src, z_src_nbytes);
/* Check for various zlib errors */
if (Z_BUF_ERROR == ret) {
HDfprintf(stderr, "overflow\n");
goto error;
}
else if (Z_MEM_ERROR == ret) {
HDfprintf(stderr, "deflate memory error\n");
goto error;
}
else if (Z_DATA_ERROR == ret) {
HDfprintf(stderr, "corrupted data\n");
goto error;
}
else if (Z_OK != ret) {
HDfprintf(stderr, "other deflate error\n");
goto error;
}
/* Check that the decompressed values match those read from H5Dread */
for (k = 0; k < CHUNK_NX; k++) {
for (l = 0; l < CHUNK_NY; l++) {
if (direct_buf[k][l] != check_chunk[k][l]) {
HDprintf("\n 1. Read different values than written.");
HDprintf(" At index %d,%d\n", k, l);
HDprintf(" direct_buf=%d, check_chunk=%d\n", direct_buf[k][l], check_chunk[k][l]);
goto error;
}
}
}
}
}
/* Close/release resources. */
H5Dclose(dataset);
H5Sclose(mem_space);
H5Sclose(dataspace);
H5Pclose(cparms);
H5Pclose(dxpl);
H5Pclose(dapl);
if (outbuf)
HDfree(outbuf);
PASSED();
return 0;
error:
H5E_BEGIN_TRY
{
H5Dclose(dataset);
H5Sclose(mem_space);
H5Sclose(dataspace);
H5Pclose(cparms);
H5Pclose(dxpl);
H5Pclose(dapl);
}
H5E_END_TRY;
if (outbuf)
HDfree(outbuf);
H5_FAILED();
return 1;
} /* test_direct_chunk_read_no_cache() */
#endif /* H5_HAVE_FILTER_DEFLATE */
#ifdef H5_HAVE_FILTER_DEFLATE
static int
test_direct_chunk_read_cache(hid_t file, hbool_t flush)
{
hid_t dataspace = -1, dataset = -1;
hid_t mem_space = -1;
hid_t cparms = -1, dxpl = -1;
hsize_t dims[2] = {NX, NY};
hsize_t maxdims[2] = {H5S_UNLIMITED, H5S_UNLIMITED};
hsize_t chunk_dims[2] = {CHUNK_NX, CHUNK_NY};
herr_t status; /* status from H5 function calls */
int ret; /* deflate return status */
int data[NX][NY];
int i, j, k, l, n; /* local index variables */
unsigned filter_mask = 0; /* filter mask returned from H5Dread_chunk */
int direct_buf[CHUNK_NX][CHUNK_NY]; /* chunk read with H5Dread and manually decompressed */
int check_chunk[CHUNK_NX][CHUNK_NY]; /* chunk read with H5Dread */
hsize_t offset[2]; /* chunk offset used for H5Dread_chunk */
size_t buf_size = CHUNK_NX * CHUNK_NY * sizeof(int);
Bytef * z_src = NULL; /* source buffer */
uLongf z_src_nbytes = (uLongf)DEFLATE_SIZE_ADJUST(buf_size);
Bytef * z_dst = (Bytef *)(direct_buf);
uLong z_dst_nbytes = (uLong)buf_size;
int aggression = 9; /* Compression aggression setting */
void * outbuf = NULL; /* Pointer to new buffer */
hsize_t read_buf_size = 0;
hsize_t start[2]; /* Start of hyperslab */
hsize_t stride[2]; /* Stride of hyperslab */
hsize_t count[2]; /* Block count */
hsize_t block[2]; /* Block sizes */
if (flush) {
TESTING("basic functionality of H5Dread_chunk (flush chunk cache)");
}
else {
TESTING("basic functionality of H5Dread_chunk (does not flush chunk cache)");
}
/* Create the data space with unlimited dimensions. */
if ((dataspace = H5Screate_simple(RANK, dims, maxdims)) < 0)
goto error;
if ((mem_space = H5Screate_simple(RANK, chunk_dims, NULL)) < 0)
goto error;
/* Modify dataset creation properties, i.e. enable chunking and compression */
if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
goto error;
if ((status = H5Pset_chunk(cparms, RANK, chunk_dims)) < 0)
goto error;
if ((status = H5Pset_deflate(cparms, (unsigned)aggression)) < 0)
goto error;
/* Create a new dataset within the file using cparms creation properties. */
if ((dataset = H5Dcreate2(file, flush ? DATASETNAME9 : DATASETNAME10, H5T_NATIVE_INT, dataspace,
H5P_DEFAULT, cparms, H5P_DEFAULT)) < 0)
goto error;
/* Initialize the dataset */
for (i = n = 0; i < NX; i++)
for (j = 0; j < NY; j++)
data[i][j] = n++;
if ((dxpl = H5Pcreate(H5P_DATASET_XFER)) < 0)
goto error;
/* Write the data for the dataset.
* It should stay in the chunk cache. */
if ((status = H5Dwrite(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, data)) < 0)
goto error;
if (flush) {
/* Flush the chunk cache to disk. Cache entry is not evicted. */
if (H5Fflush(dataset, H5F_SCOPE_LOCAL) < 0)
goto error;
}
/* Allocate output (compressed) buffer */
outbuf = HDmalloc(z_src_nbytes);
z_src = (Bytef *)outbuf;
/* For each chunk in the dataset, compare the result of H5Dread and H5Dread_chunk. */
for (i = 0; i < NX / CHUNK_NX; i++) {
for (j = 0; j < NY / CHUNK_NY; j++) {
/* Select hyperslab for one chunk in the file */
start[0] = (hsize_t)i * CHUNK_NX;
start[1] = (hsize_t)j * CHUNK_NY;
stride[0] = 1;
stride[1] = 1;
count[0] = 1;
count[1] = 1;
block[0] = CHUNK_NX;
block[1] = CHUNK_NY;
/* Hyperslab selection equals single chunk */
if ((status = H5Sselect_hyperslab(dataspace, H5S_SELECT_SET, start, stride, count, block)) < 0)
goto error;
/* Read the chunk back */
if ((status = H5Dread(dataset, H5T_NATIVE_INT, mem_space, dataspace, H5P_DEFAULT, check_chunk)) <
0)
goto error;
offset[0] = (hsize_t)i * CHUNK_NX;
offset[1] = (hsize_t)j * CHUNK_NY;
/* Query chunk storage size */
if ((status = H5Dget_chunk_storage_size(dataset, offset, &read_buf_size)) < 0)
goto error;
if (read_buf_size == 0)
goto error;
/* Read the compressed chunk back using the direct read function. */
if ((status = H5Dread_chunk(dataset, dxpl, offset, &filter_mask, outbuf)) < 0)
goto error;
/* Check filter mask return value */
if (filter_mask != 0)
goto error;
/* Perform decompression from the source to the destination buffer */
ret = uncompress(z_dst, &z_dst_nbytes, z_src, z_src_nbytes);
/* Check for various zlib errors */
if (Z_BUF_ERROR == ret) {
HDfprintf(stderr, "overflow\n");
goto error;
}
else if (Z_MEM_ERROR == ret) {
HDfprintf(stderr, "deflate memory error\n");
goto error;
}
else if (Z_DATA_ERROR == ret) {
HDfprintf(stderr, "corrupted data\n");
goto error;
}
else if (Z_OK != ret) {
HDfprintf(stderr, "other deflate error\n");
goto error;
}
/* Check that the decompressed values match those read from H5Dread */
for (k = 0; k < CHUNK_NX; k++) {
for (l = 0; l < CHUNK_NY; l++) {
if (direct_buf[k][l] != check_chunk[k][l]) {
HDprintf("\n 1. Read different values than written.");
HDprintf(" At index %d,%d\n", k, l);
HDprintf(" direct_buf=%d, check_chunk=%d\n", direct_buf[k][l], check_chunk[k][l]);
goto error;
}
}
}
}
}
/* Close/release resources. */
H5Dclose(dataset);
H5Sclose(mem_space);
H5Sclose(dataspace);
H5Pclose(cparms);
H5Pclose(dxpl);
if (outbuf)
HDfree(outbuf);
PASSED();
return 0;
error:
H5E_BEGIN_TRY
{
H5Dclose(dataset);
H5Sclose(mem_space);
H5Sclose(dataspace);
H5Pclose(cparms);
H5Pclose(dxpl);
}
H5E_END_TRY;
if (outbuf)
HDfree(outbuf);
H5_FAILED();
return 1;
} /* test_direct_chunk_read_cache() */
#endif /* H5_HAVE_FILTER_DEFLATE */
/*-------------------------------------------------------------------------
* Function: test_read_unfiltered_dset
*
* Purpose: Test the basic functionality of H5Dread_chunk on a dataset
* without no filters applied.
*
* Return: Success: 0
* Failure: 1
*
* Programmer: Matthew Strong (GE Healthcare)
* 30 November 2016
*
*-------------------------------------------------------------------------
*/
static int
test_read_unfiltered_dset(hid_t file)
{
hid_t dataspace = -1, dataset = -1;
hid_t mem_space = -1;
hid_t cparms = -1, dxpl = -1;
hsize_t dims[2] = {NX, NY};
hsize_t maxdims[2] = {H5S_UNLIMITED, H5S_UNLIMITED};
hsize_t chunk_dims[2] = {CHUNK_NX, CHUNK_NY};
herr_t status;
int data[NX][NY];
int i, j, k, l, n;
unsigned filter_mask = 0;
int direct_buf[CHUNK_NX][CHUNK_NY];
int check_chunk[CHUNK_NX][CHUNK_NY]; /* chunk read with H5Dread */
hsize_t offset[2] = {0, 0};
size_t buf_size = CHUNK_NX * CHUNK_NY * sizeof(int);
hsize_t read_buf_size = 0;
hsize_t start[2]; /* Start of hyperslab */
hsize_t stride[2]; /* Stride of hyperslab */
hsize_t count[2]; /* Block count */
hsize_t block[2]; /* Block sizes */
TESTING("basic functionality of H5Dread_chunk on unfiltered datasets");
/* Create the data space with unlimited dimensions. */
if ((dataspace = H5Screate_simple(RANK, dims, maxdims)) < 0)
goto error;
if ((mem_space = H5Screate_simple(RANK, chunk_dims, NULL)) < 0)
goto error;
/* Modify dataset creation properties, i.e. enable chunking, no compression */
if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
goto error;
if ((status = H5Pset_chunk(cparms, RANK, chunk_dims)) < 0)
goto error;
/* Create a new dataset within the file using cparms creation properties. */
if ((dataset = H5Dcreate2(file, DATASETNAME12, H5T_NATIVE_INT, dataspace, H5P_DEFAULT, cparms,
H5P_DEFAULT)) < 0)
goto error;
/* Initialize the dataset */
for (i = n = 0; i < NX; i++)
for (j = 0; j < NY; j++)
data[i][j] = n++;
if ((dxpl = H5Pcreate(H5P_DATASET_XFER)) < 0)
goto error;
/* Write the data for the dataset.
* It should stay in the chunk cache and will be evicted/flushed by
* the H5Dread_chunk function call. */
if ((status = H5Dwrite(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, data)) < 0)
goto error;
if (H5Fflush(dataset, H5F_SCOPE_LOCAL) < 0)
goto error;
/* For each chunk in the dataset, compare the result of H5Dread and H5Dread_chunk. */
for (i = 0; i < NX / CHUNK_NX; i++) {
for (j = 0; j < NY / CHUNK_NY; j++) {
/* Select hyperslab for one chunk in the file */
start[0] = (hsize_t)i * CHUNK_NX;
start[1] = (hsize_t)j * CHUNK_NY;
stride[0] = 1;
stride[1] = 1;
count[0] = 1;
count[1] = 1;
block[0] = CHUNK_NX;
block[1] = CHUNK_NY;
/* Hyperslab selection equals single chunk */
if ((status = H5Sselect_hyperslab(dataspace, H5S_SELECT_SET, start, stride, count, block)) < 0)
goto error;
/* Read the chunk back */
if ((status = H5Dread(dataset, H5T_NATIVE_INT, mem_space, dataspace, H5P_DEFAULT, check_chunk)) <
0)
goto error;
/* Query chunk storage size */
if ((status = H5Dget_chunk_storage_size(dataset, offset, &read_buf_size)) < 0)
goto error;
if (read_buf_size != buf_size)
goto error;
offset[0] = (hsize_t)i * CHUNK_NX;
offset[1] = (hsize_t)j * CHUNK_NY;
/* Read the raw chunk back */
HDmemset(&direct_buf, 0, sizeof(direct_buf));
filter_mask = UINT_MAX;
if ((status = H5Dread_chunk(dataset, dxpl, offset, &filter_mask, direct_buf)) < 0)
goto error;
/* Check filter mask return value */
if (filter_mask != 0)
goto error;
/* Check that the decompressed values match those read from H5Dread */
for (k = 0; k < CHUNK_NX; k++) {
for (l = 0; l < CHUNK_NY; l++) {
if (direct_buf[k][l] != check_chunk[k][l]) {
HDprintf("\n 1. Read different values than written.");
HDprintf(" At index %d,%d\n", k, l);
HDprintf(" direct_buf=%d, check_chunk=%d\n", direct_buf[k][l], check_chunk[k][l]);
goto error;
}
}
}
}
}
/* Close/release resources. */
H5Dclose(dataset);
H5Sclose(mem_space);
H5Sclose(dataspace);
H5Pclose(cparms);
H5Pclose(dxpl);
PASSED();
return 0;
error:
H5E_BEGIN_TRY
{
H5Dclose(dataset);
H5Sclose(mem_space);
H5Sclose(dataspace);
H5Pclose(cparms);
H5Pclose(dxpl);
}
H5E_END_TRY;
H5_FAILED();
return 1;
} /* test_read_unfiltered_dset() */
/*-------------------------------------------------------------------------
* Function: test_read_unallocated_chunk
*
* Purpose: Tests the H5Dread_chunk and H5Dget_chunk_storage_size with valid
* offsets to chunks that have not been written to the dataset and are
* not allocated in the chunk storage on disk.
*
* Return: Success: 0
* Failure: 1
*
* Programmer: Matthew Strong (GE Healthcare)
* 30 November 2016
*
*-------------------------------------------------------------------------
*/
static int
test_read_unallocated_chunk(hid_t file)
{
hid_t dataspace = -1, dataset = -1;
hid_t mem_space = -1;
hid_t cparms = -1, dxpl = -1;
hsize_t dims[2] = {NX, NY};
hsize_t maxdims[2] = {H5S_UNLIMITED, H5S_UNLIMITED};
hsize_t chunk_dims[2] = {CHUNK_NX, CHUNK_NY};
hsize_t chunk_nbytes = CHUNK_NX * CHUNK_NY * sizeof(int);
hsize_t direct_chunk_nbytes = 0; /* size (bytes) of the on-disk chunk */
herr_t status; /* status from H5 function calls */
hsize_t i, j; /* local index variables */
unsigned filter_mask = 0; /* filter mask returned from H5Dread_chunk */
int direct_buf[CHUNK_NX][CHUNK_NY]; /* chunk read with H5Dread and manually decompressed */
hsize_t offset[2]; /* chunk offset used for H5Dread_chunk */
TESTING("H5Dread_chunk with unallocated chunks");
/* Create the data space with unlimited dimensions. */
if ((dataspace = H5Screate_simple(RANK, dims, maxdims)) < 0)
FAIL_STACK_ERROR;
if ((mem_space = H5Screate_simple(RANK, chunk_dims, NULL)) < 0)
FAIL_STACK_ERROR;
/* Modify dataset creation properties, i.e. enable chunking, no compression */
if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
FAIL_STACK_ERROR;
if ((status = H5Pset_chunk(cparms, RANK, chunk_dims)) < 0)
FAIL_STACK_ERROR;
/* Create a new dataset within the file using cparms creation properties. */
if ((dataset = H5Dcreate2(file, DATASETNAME11, H5T_NATIVE_INT, dataspace, H5P_DEFAULT, cparms,
H5P_DEFAULT)) < 0)
FAIL_STACK_ERROR;
if ((dxpl = H5Pcreate(H5P_DATASET_XFER)) < 0)
FAIL_STACK_ERROR;
/* Write a single chunk to initialize the chunk storage */
HDmemset(direct_buf, 0, CHUNK_NX * CHUNK_NY * sizeof(int));
offset[0] = 0;
offset[1] = 0;
if (H5Dwrite_chunk(dataset, dxpl, filter_mask, offset, chunk_nbytes, direct_buf) < 0)
FAIL_STACK_ERROR;
/* Attempt to read each chunk in the dataset. Chunks are not allocated,
* therefore we expect the result of H5Dread_chunk to fail. Chunk idx starts
* at 1, since one chunk was written to init the chunk storage. */
for (i = 1; i < NX / CHUNK_NX; i++) {
for (j = 0; j < NY / CHUNK_NY; j++) {
offset[0] = i * CHUNK_NX;
offset[1] = j * CHUNK_NY;
/* Read a non-existent chunk using the direct read function. */
H5E_BEGIN_TRY
{
status = H5Dread_chunk(dataset, dxpl, offset, &filter_mask, &direct_buf);
}
H5E_END_TRY;
/* Check that the chunk read call does not succeed. */
if (status != -1)
TEST_ERROR
/* Query the size of the non-existent chunk */
direct_chunk_nbytes = ULONG_MAX;
H5E_BEGIN_TRY
{
status = H5Dget_chunk_storage_size(dataset, offset, &direct_chunk_nbytes);
}
H5E_END_TRY;
/* Check that the chunk storage size call does not succeed. */
if (status != -1)
TEST_ERROR
if (direct_chunk_nbytes != ULONG_MAX)
TEST_ERROR
}
}
/* Close/release resources. */
if (H5Dclose(dataset) < 0)
FAIL_STACK_ERROR;
if (H5Sclose(mem_space) < 0)
FAIL_STACK_ERROR;
if (H5Sclose(dataspace) < 0)
FAIL_STACK_ERROR;
if (H5Pclose(cparms) < 0)
FAIL_STACK_ERROR;
if (H5Pclose(dxpl) < 0)
FAIL_STACK_ERROR;
PASSED();
return 0;
error:
H5E_BEGIN_TRY
{
H5Dclose(dataset);
H5Sclose(mem_space);
H5Sclose(dataspace);
H5Pclose(cparms);
H5Pclose(dxpl);
}
H5E_END_TRY;
H5_FAILED();
return 1;
} /* test_read_unallocated_chunk() */
/*-------------------------------------------------------------------------
* Function: test_single_chunk
*
* Purpose: This is to verify the fix for jira issue HDFFV-10425.
* The problem was due to a bug in the internal ilbrary routine
* H5D__chunk_direct_write() which passed a null dataset
* pointer to the insert callback for the chunk index type.
* Currently, the single chunk index is the only one that
* used the dataset pointer in the insert callback.
*
* This routine is based on the test program attached to
* this jira issue:
* Create a file with the latest format and a chunked dataset
* with one single chunk. The library will use single chunk
* index for the dataset.
* Verify that the data read is the same as the written data.
*
* Since expanded to test multiple combinations of cases
* involving a single chunk
*
* Return: Success: 0
* Failure: 1
*
*-------------------------------------------------------------------------
*/
static int
test_single_chunk(unsigned config)
{
hid_t fid = H5I_INVALID_HID; /* File ID */
hid_t fapl = H5I_INVALID_HID; /* File access property list ID */
hid_t sid = H5I_INVALID_HID; /* Dataspace ID */
hid_t did = H5I_INVALID_HID; /* Dataset ID */
hid_t dcpl = H5I_INVALID_HID; /* Dataset creation property list */
hsize_t dims[2] = {DIM0, DIM1}; /* Dimension sizes */
hsize_t chunk[2] = {CHUNK0, CHUNK1}; /* Chunk dimension sizes */
hsize_t offset[2] = {0, 0}; /* Offset for writing */
uint32_t filters; /* Filter mask out */
int wdata[DIM0][DIM1]; /* Write buffer */
int rdata[DIM0][DIM1]; /* Read buffer */
int i, j; /* Local index variable */
TESTING("Single chunk I/O");
/* Initialize data */
for (i = 0; i < DIM0; i++)
for (j = 0; j < DIM1; j++)
wdata[i][j] = j / CHUNK0;
/* Create a new file with the latest format */
if ((fapl = H5Pcreate(H5P_FILE_ACCESS)) < 0)
FAIL_STACK_ERROR;
if (config & CONFIG_LATEST)
if (H5Pset_libver_bounds(fapl, H5F_LIBVER_LATEST, H5F_LIBVER_LATEST) < 0)
FAIL_STACK_ERROR;
if ((fid = H5Fcreate(FILE, H5F_ACC_TRUNC, H5P_DEFAULT, fapl)) < 0)
FAIL_STACK_ERROR;
/* Create dataspace */
if ((sid = H5Screate_simple(2, dims, NULL)) < 0)
FAIL_STACK_ERROR;
/* Create the dataset creation property list and set the chunk size */
if ((dcpl = H5Pcreate(H5P_DATASET_CREATE)) < 0)
FAIL_STACK_ERROR;
if (H5Pset_chunk(dcpl, 2, chunk) < 0)
FAIL_STACK_ERROR;
/* Create the dataset */
if ((did = H5Dcreate2(fid, DATASET, H5T_NATIVE_INT, sid, H5P_DEFAULT, dcpl, H5P_DEFAULT)) < 0)
FAIL_STACK_ERROR;
if (config & CONFIG_DIRECT_WRITE) {
/* Write the data directly to the dataset */
if (H5Dwrite_chunk(did, H5P_DEFAULT, 0, offset, CHUNK0 * CHUNK1 * 4, (void *)wdata) < 0)
FAIL_STACK_ERROR;
} /* end if */
else
/* Write the data to the dataset */
if (H5Dwrite(did, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, H5P_DEFAULT, (void *)wdata) < 0)
FAIL_STACK_ERROR;
/*
* Close and release resources.
*/
if (H5Pclose(dcpl) < 0)
FAIL_STACK_ERROR;
if (config & CONFIG_REOPEN_DSET)
if (H5Dclose(did) < 0)
FAIL_STACK_ERROR;
if (H5Sclose(sid) < 0)
FAIL_STACK_ERROR;
if (H5Pclose(fapl) < 0)
FAIL_STACK_ERROR;
if (config & CONFIG_REOPEN_FILE)
if (H5Fclose(fid) < 0)
FAIL_STACK_ERROR;
/* Open the file and dataset with default properties */
if (config & CONFIG_REOPEN_FILE)
if ((fid = H5Fopen(FILE, H5F_ACC_RDONLY, H5P_DEFAULT)) < 0)
FAIL_STACK_ERROR;
if (config & CONFIG_REOPEN_DSET)
if ((did = H5Dopen2(fid, DATASET, H5P_DEFAULT)) < 0)
FAIL_STACK_ERROR;
/* Retrieve dataset creation property list */
if ((dcpl = H5Dget_create_plist(did)) < 0)
FAIL_STACK_ERROR;
if (config & CONFIG_DIRECT_READ) {
/* Read the data directly */
if (H5Dread_chunk(did, H5P_DEFAULT, offset, &filters, rdata) < 0)
FAIL_STACK_ERROR;
/* Verify returned filter mask */
if (filters != 0)
TEST_ERROR
} /* end if */
else
/* Read the data */
if (H5Dread(did, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, H5P_DEFAULT, rdata) < 0)
FAIL_STACK_ERROR;
/* Verify that the data read was correct. */
for (i = 0; i < DIM0; i++)
for (j = 0; j < DIM1; j++)
if (rdata[i][j] != wdata[i][j])
TEST_ERROR
/*
* Close and release resources
*/
if (H5Pclose(dcpl) < 0)
FAIL_STACK_ERROR;
if (H5Dclose(did) < 0)
FAIL_STACK_ERROR;
if (H5Fclose(fid) < 0)
FAIL_STACK_ERROR;
PASSED();
return 0;
error:
H5E_BEGIN_TRY
{
H5Dclose(did);
H5Sclose(sid);
H5Pclose(dcpl);
H5Pclose(fapl);
H5Fclose(fid);
}
H5E_END_TRY;
H5_FAILED();
return 1;
} /* test_single_chunk_latest() */
/*-------------------------------------------------------------------------
* Function: Main function
*
* Purpose: Test direct chunk write function H5Dwrite_chunk and
* chunk direct read function H5Dread_chunk
*
* Return: Success: 0
* Failure: 1
*
* Programmer: Raymond Lu
* 30 November 2012
*
*-------------------------------------------------------------------------
*/
int
main(void)
{
hid_t file_id;
unsigned config;
int nerrors = 0;
/*
* Create a new file. If file exists its contents will be overwritten.
*/
if ((file_id = H5Fcreate(FILE_NAME, H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT)) < 0)
goto error;
/* Test direct chunk write and direct chunk read */
#ifdef H5_HAVE_FILTER_DEFLATE
nerrors += test_direct_chunk_write(file_id);
#endif /* H5_HAVE_FILTER_DEFLATE */
nerrors += test_direct_chunk_overwrite_data(file_id);
nerrors += test_skip_compress_write1(file_id);
nerrors += test_skip_compress_write2(file_id);
nerrors += test_data_conv(file_id);
nerrors += test_invalid_parameters(file_id);
/* Test direct chunk read */
#ifdef H5_HAVE_FILTER_DEFLATE
nerrors += test_direct_chunk_read_no_cache(file_id);
nerrors += test_direct_chunk_read_cache(file_id, TRUE);
nerrors += test_direct_chunk_read_cache(file_id, FALSE);
#endif /* H5_HAVE_FILTER_DEFLATE */
nerrors += test_read_unfiltered_dset(file_id);
nerrors += test_read_unallocated_chunk(file_id);
/* Loop over test configurations */
for (config = 0; config < CONFIG_END; config++) {
hbool_t need_comma = FALSE;
/* Check for invalid combinations */
if ((config & CONFIG_REOPEN_FILE) && !(config & CONFIG_REOPEN_DSET))
continue;
/* Print configuration */
HDprintf("Configuration: ");
if (config == 0)
HDprintf("<empty>");
if (config & CONFIG_LATEST) {
if (need_comma)
HDprintf(", ");
HDprintf("latest format");
need_comma = TRUE;
} /* end if */
if (config & CONFIG_REOPEN_FILE) {
if (need_comma)
HDprintf(", ");
HDprintf("reopen file");
need_comma = TRUE;
} /* end if */
else if (config & CONFIG_REOPEN_DSET) {
if (need_comma)
HDprintf(", ");
HDprintf("reopen dataset");
need_comma = TRUE;
} /* end if */
if (config & CONFIG_DIRECT_WRITE) {
if (need_comma)
HDprintf(", ");
HDprintf("direct write");
need_comma = TRUE;
} /* end if */
if (config & CONFIG_DIRECT_READ) {
if (need_comma)
HDprintf(", ");
HDprintf("direct read");
need_comma = TRUE;
} /* end if */
HDprintf(":\n");
HDfflush(stdout);
nerrors += test_single_chunk(config);
} /* end for */
if (H5Fclose(file_id) < 0)
goto error;
/* check for errors */
if (nerrors)
goto error;
HDputs("All direct chunk read/write tests passed.");
return EXIT_SUCCESS;
error:
HDputs("*** TESTS FAILED ***");
return EXIT_FAILURE;
}