/* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *
 * Copyright by The HDF Group.                                               *
 * Copyright by the Board of Trustees of the University of Illinois.         *
 * All rights reserved.                                                      *
 *                                                                           *
 * This file is part of HDF5.  The full HDF5 copyright notice, including     *
 * terms governing use, modification, and redistribution, is contained in    *
 * the COPYING file, which can be found at the root of the source code       *
 * distribution tree, or in https://www.hdfgroup.org/licenses.               *
 * If you do not have access to either file, you may request a copy from     *
 * help@hdfgroup.org.                                                        *
 * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */

#include "h5test.h"

#if defined(H5_HAVE_ZLIB_H) && !defined(H5_ZLIB_HEADER)
#define H5_ZLIB_HEADER "zlib.h"
#endif
#if defined(H5_ZLIB_HEADER)
#include H5_ZLIB_HEADER /* "zlib.h" */
#endif

#define FILE_NAME "direct_chunk.h5"

/* Datasets for Direct Write tests */
#ifdef H5_HAVE_FILTER_DEFLATE
#define DATASETNAME1 "direct_write"
#endif
#define DATASETNAME2 "skip_one_filter"
#define DATASETNAME3 "skip_two_filters"
#define DATASETNAME4 "data_conv"
#define DATASETNAME5 "contiguous_dset"
#define DATASETNAME6 "invalid_argue"
#define DATASETNAME7 "overwrite_chunk"
/* Datasets for Direct Read tests */
#ifdef H5_HAVE_FILTER_DEFLATE
#define DATASETNAME8  "disabled_chunk_cache"
#define DATASETNAME9  "flush_chunk_cache"
#define DATASETNAME10 "read_w_valid_cache"
#endif
#define DATASETNAME11 "unallocated_chunk"
#define DATASETNAME12 "unfiltered_data"

#define RANK     2
#define NX       16
#define NY       16
#define CHUNK_NX 4
#define CHUNK_NY 4

#ifdef H5_HAVE_FILTER_DEFLATE
#define DEFLATE_SIZE_ADJUST(s) (HDceil(((double)(s)) * 1.001) + 12.0)
#endif

/* Temporary filter IDs used for testing */
#define H5Z_FILTER_BOGUS1 305
#define H5Z_FILTER_BOGUS2 306
#define ADD_ON            7
#define FACTOR            3

/* Constants for the overwrite test */
#define OVERWRITE_NDIMS     3
#define OVERWRITE_CHUNK_NX  3
#define OVERWRITE_CHUNK_2NX 6
#define OVERWRITE_CHUNK_NY  2
#define OVERWRITE_VALUE     42

/* Test configurations */
#define CONFIG_LATEST       0x01
#define CONFIG_REOPEN_FILE  0x02
#define CONFIG_REOPEN_DSET  0x04
#define CONFIG_DIRECT_WRITE 0x08
#define CONFIG_DIRECT_READ  0x10
#define CONFIG_END          0x20

/* Defines used in test_single_chunk_latest() */
#define FILE    "single_latest.h5"
#define DATASET "dataset"
#define DIM0    4
#define DIM1    32
#define CHUNK0  DIM0
#define CHUNK1  DIM1

/* Local prototypes for filter functions */
static size_t filter_bogus1(unsigned int flags, size_t cd_nelmts, const unsigned int *cd_values,
                            size_t nbytes, size_t *buf_size, void **buf);
static size_t filter_bogus2(unsigned int flags, size_t cd_nelmts, const unsigned int *cd_values,
                            size_t nbytes, size_t *buf_size, void **buf);

/* This message derives from H5Z */
const H5Z_class2_t H5Z_BOGUS1[1] = {{
    H5Z_CLASS_T_VERS,  /* H5Z_class_t version */
    H5Z_FILTER_BOGUS1, /* Filter id number        */
    1, 1,              /* Encoding and decoding enabled */
    "bogus1",          /* Filter name for debugging    */
    NULL,              /* The "can apply" callback     */
    NULL,              /* The "set local" callback     */
    filter_bogus1,     /* The actual filter function    */
}};

const H5Z_class2_t H5Z_BOGUS2[1] = {{
    H5Z_CLASS_T_VERS,  /* H5Z_class_t version */
    H5Z_FILTER_BOGUS2, /* Filter id number        */
    1, 1,              /* Encoding and decoding enabled */
    "bogus2",          /* Filter name for debugging    */
    NULL,              /* The "can apply" callback     */
    NULL,              /* The "set local" callback     */
    filter_bogus2,     /* The actual filter function    */
}};

/*-------------------------------------------------------------------------
 * Function:    test_direct_chunk_write
 *
 * Purpose:        Test the basic functionality of H5Dwrite_chunk
 *
 * Return:        Success:    0
 *                Failure:    1
 *
 * Programmer:  Raymond Lu
 *              30 November 2012
 *
 *-------------------------------------------------------------------------
 */
#ifdef H5_HAVE_FILTER_DEFLATE
static int
test_direct_chunk_write(hid_t file)
{
    hid_t   dataspace = -1, dataset = -1;
    hid_t   mem_space = -1;
    hid_t   cparms = -1, dxpl = -1;
    hsize_t dims[2]       = {NX, NY};
    hsize_t maxdims[2]    = {H5S_UNLIMITED, H5S_UNLIMITED};
    hsize_t chunk_dims[2] = {CHUNK_NX, CHUNK_NY};
    herr_t  status;
    int     ret;
    int     data[NX][NY];
    int     i, j, n;

    unsigned filter_mask = 0;
    int      direct_buf[CHUNK_NX][CHUNK_NY];
    int      check_chunk[CHUNK_NX][CHUNK_NY];
    hsize_t  offset[2] = {0, 0};
    size_t   buf_size  = CHUNK_NX * CHUNK_NY * sizeof(int);

    const Bytef *z_src        = (const Bytef *)(direct_buf);
    Bytef *      z_dst        = NULL; /*destination buffer        */
    uLongf       z_dst_nbytes = (uLongf)DEFLATE_SIZE_ADJUST(buf_size);
    uLong        z_src_nbytes = (uLong)buf_size;
    int          aggression   = 9;    /* Compression aggression setting */
    void *       outbuf       = NULL; /* Pointer to new buffer */

    hsize_t start[2];  /* Start of hyperslab */
    hsize_t stride[2]; /* Stride of hyperslab */
    hsize_t count[2];  /* Block count */
    hsize_t block[2];  /* Block sizes */

    TESTING("basic functionality of H5Dwrite_chunk");

    /*
     * Create the data space with unlimited dimensions.
     */
    if ((dataspace = H5Screate_simple(RANK, dims, maxdims)) < 0)
        goto error;

    if ((mem_space = H5Screate_simple(RANK, chunk_dims, NULL)) < 0)
        goto error;

    /*
     * Modify dataset creation properties, i.e. enable chunking and compression
     */
    if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
        goto error;

    if ((status = H5Pset_chunk(cparms, RANK, chunk_dims)) < 0)
        goto error;

    if ((status = H5Pset_deflate(cparms, (unsigned)aggression)) < 0)
        goto error;

    /*
     * Create a new dataset within the file using cparms
     * creation properties.
     */
    if ((dataset =
             H5Dcreate2(file, DATASETNAME1, H5T_NATIVE_INT, dataspace, H5P_DEFAULT, cparms, H5P_DEFAULT)) < 0)
        goto error;

    /* Initialize the dataset */
    for (i = n = 0; i < NX; i++)
        for (j = 0; j < NY; j++)
            data[i][j] = n++;

    if ((dxpl = H5Pcreate(H5P_DATASET_XFER)) < 0)
        goto error;

    /*
     * Write the data for the dataset.  It should stay in the chunk cache.
     * It will be evicted from the cache by the H5Dwrite_chunk calls.
     */
    if ((status = H5Dwrite(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, data)) < 0)
        goto error;

    /* Initialize data for one chunk */
    for (i = n = 0; i < CHUNK_NX; i++)
        for (j = 0; j < CHUNK_NY; j++)
            direct_buf[i][j] = n++;

    /* Allocate output (compressed) buffer */
    outbuf = HDmalloc(z_dst_nbytes);
    z_dst  = (Bytef *)outbuf;

    /* Perform compression from the source to the destination buffer */
    ret = compress2(z_dst, &z_dst_nbytes, z_src, z_src_nbytes, aggression);

    /* Check for various zlib errors */
    if (Z_BUF_ERROR == ret) {
        HDfprintf(stderr, "overflow");
        goto error;
    }
    else if (Z_MEM_ERROR == ret) {
        HDfprintf(stderr, "deflate memory error");
        goto error;
    }
    else if (Z_OK != ret) {
        HDfprintf(stderr, "other deflate error");
        goto error;
    }

    /* Write the compressed chunk data repeatedly to cover all the chunks in the
     * dataset, using the direct writing function.     */
    for (i = 0; i < NX / CHUNK_NX; i++) {
        for (j = 0; j < NY / CHUNK_NY; j++) {
            status = H5Dwrite_chunk(dataset, dxpl, filter_mask, offset, z_dst_nbytes, outbuf);
            offset[1] += CHUNK_NY;
        }
        offset[0] += CHUNK_NX;
        offset[1] = 0;
    }

    if (outbuf)
        HDfree(outbuf);

    if (H5Fflush(dataset, H5F_SCOPE_LOCAL) < 0)
        goto error;

    if (H5Dclose(dataset) < 0)
        goto error;

    if ((dataset = H5Dopen2(file, DATASETNAME1, H5P_DEFAULT)) < 0)
        goto error;

    /*
     * Select hyperslab for one chunk in the file
     */
    start[0]  = CHUNK_NX;
    start[1]  = CHUNK_NY;
    stride[0] = 1;
    stride[1] = 1;
    count[0]  = 1;
    count[1]  = 1;
    block[0]  = CHUNK_NX;
    block[1]  = CHUNK_NY;
    if ((status = H5Sselect_hyperslab(dataspace, H5S_SELECT_SET, start, stride, count, block)) < 0)
        goto error;

    /* Read the chunk back */
    if ((status = H5Dread(dataset, H5T_NATIVE_INT, mem_space, dataspace, H5P_DEFAULT, check_chunk)) < 0)
        goto error;

    /* Check that the values read are the same as the values written */
    for (i = 0; i < CHUNK_NX; i++) {
        for (j = 0; j < CHUNK_NY; j++) {
            if (direct_buf[i][j] != check_chunk[i][j]) {
                HDprintf("    1. Read different values than written.");
                HDprintf("    At index %d,%d\n", i, j);
                HDprintf("    direct_buf=%d, check_chunk=%d\n", direct_buf[i][j], check_chunk[i][j]);
                goto error;
            }
        }
    }

    /* Reinitialize different data for one chunk */
    for (i = 0; i < CHUNK_NX; i++)
        for (j = 0; j < CHUNK_NY; j++)
            direct_buf[i][j] = i + j;

    /* Allocate output (compressed) buffer */
    outbuf = HDmalloc(z_dst_nbytes);
    z_dst  = (Bytef *)outbuf;

    /* Perform compression from the source to the destination buffer */
    ret = compress2(z_dst, &z_dst_nbytes, z_src, z_src_nbytes, aggression);

    /* Check for various zlib errors */
    if (Z_BUF_ERROR == ret) {
        HDfprintf(stderr, "overflow");
        goto error;
    }
    else if (Z_MEM_ERROR == ret) {
        HDfprintf(stderr, "deflate memory error");
        goto error;
    }
    else if (Z_OK != ret) {
        HDfprintf(stderr, "other deflate error");
        goto error;
    }

    /* Rewrite the compressed chunk data repeatedly to cover all the chunks in the
     * dataset, using the direct writing function.     */
    offset[0] = offset[1] = 0;
    for (i = 0; i < NX / CHUNK_NX; i++) {
        for (j = 0; j < NY / CHUNK_NY; j++) {
            status = H5Dwrite_chunk(dataset, dxpl, filter_mask, offset, z_dst_nbytes, outbuf);
            offset[1] += CHUNK_NY;
        }
        offset[0] += CHUNK_NX;
        offset[1] = 0;
    }

    if (outbuf)
        HDfree(outbuf);

    if (H5Fflush(dataset, H5F_SCOPE_LOCAL) < 0)
        goto error;

    if (H5Dclose(dataset) < 0)
        goto error;

    if ((dataset = H5Dopen2(file, DATASETNAME1, H5P_DEFAULT)) < 0)
        goto error;

    /* Read the chunk back */
    if ((status = H5Dread(dataset, H5T_NATIVE_INT, mem_space, dataspace, H5P_DEFAULT, check_chunk)) < 0)
        goto error;

    /* Check that the values read are the same as the values written */
    for (i = 0; i < CHUNK_NX; i++) {
        for (j = 0; j < CHUNK_NY; j++) {
            if (direct_buf[i][j] != check_chunk[i][j]) {
                HDprintf("    2. Read different values than written.");
                HDprintf("    At index %d,%d\n", i, j);
                HDprintf("    direct_buf=%d, check_chunk=%d\n", direct_buf[i][j], check_chunk[i][j]);
                goto error;
            }
        }
    }

    /*
     * Close/release resources.
     */
    H5Dclose(dataset);
    H5Sclose(mem_space);
    H5Sclose(dataspace);
    H5Pclose(cparms);
    H5Pclose(dxpl);

    PASSED();
    return 0;

error:
    H5E_BEGIN_TRY
    {
        H5Dclose(dataset);
        H5Sclose(mem_space);
        H5Sclose(dataspace);
        H5Pclose(cparms);
        H5Pclose(dxpl);
    }
    H5E_END_TRY;

    if (outbuf)
        HDfree(outbuf);

    H5_FAILED();
    return 1;
} /* test_direct_chunk_write() */
#endif /* H5_HAVE_FILTER_DEFLATE */

/*-------------------------------------------------------------------------
 * Function:    test_direct_chunk_overwrite_data
 *
 * Purpose:     Test overwriting a chunk with new data.
 *
 * Return:      Success:    0
 *              Failure:    1
 *
 * Programmer:  Dana Robinson
 *              Spring 2017
 *
 *-------------------------------------------------------------------------
 */
static int
test_direct_chunk_overwrite_data(hid_t fid)
{
    size_t   buf_size = OVERWRITE_CHUNK_NX * OVERWRITE_CHUNK_NY * sizeof(int16_t);
    int16_t  data_buf[OVERWRITE_CHUNK_NY][OVERWRITE_CHUNK_NX];
    int16_t  overwrite_buf[OVERWRITE_CHUNK_NY][OVERWRITE_CHUNK_NX];
    uint32_t filter_mask     = 0;
    hid_t    tid             = H5T_NATIVE_UINT16;
    hid_t    dcpl_id         = -1;
    hid_t    sid             = -1;
    hid_t    did             = -1;
    uint16_t fill_value      = 0;
    hsize_t  dset_dims[]     = {1, OVERWRITE_CHUNK_NY, OVERWRITE_CHUNK_2NX};
    hsize_t  dset_max_dims[] = {H5S_UNLIMITED, OVERWRITE_CHUNK_NY, OVERWRITE_CHUNK_2NX};
    hsize_t  chunk_dims[]    = {1, OVERWRITE_CHUNK_NY, OVERWRITE_CHUNK_NX};
    hsize_t  offset[]        = {0, 0, 0};
    hsize_t  i, j;
    int16_t  n;
    int16_t  read_buf[OVERWRITE_CHUNK_NY][OVERWRITE_CHUNK_2NX];

    TESTING("overwriting existing data with H5Dwrite_chunk");

    /* Create the dataset's data space */
    if ((sid = H5Screate_simple(OVERWRITE_NDIMS, dset_dims, dset_max_dims)) < 0)
        FAIL_STACK_ERROR

    /* Set chunk size and filll value */
    if ((dcpl_id = H5Pcreate(H5P_DATASET_CREATE)) < 0)
        FAIL_STACK_ERROR
    if (H5Pset_fill_value(dcpl_id, tid, &fill_value) < 0)
        FAIL_STACK_ERROR
    if (H5Pset_chunk(dcpl_id, OVERWRITE_NDIMS, chunk_dims) < 0)
        FAIL_STACK_ERROR

    /* Create dataset */
    if ((did = H5Dcreate2(fid, DATASETNAME7, tid, sid, H5P_DEFAULT, dcpl_id, H5P_DEFAULT)) < 0)
        FAIL_STACK_ERROR

    /* Initialize data buffers */
    n = 0;
    for (i = 0; i < OVERWRITE_CHUNK_NY; i++) {
        for (j = 0; j < OVERWRITE_CHUNK_NX; j++) {
            data_buf[i][j]      = n++;
            overwrite_buf[i][j] = OVERWRITE_VALUE;
        }
    }

    /* Write chunk data using the direct write function. */
    if (H5Dwrite_chunk(did, H5P_DEFAULT, filter_mask, offset, buf_size, data_buf) < 0)
        FAIL_STACK_ERROR

    /* Write second chunk. */
    offset[2] = OVERWRITE_CHUNK_NX;
    if (H5Dwrite_chunk(did, H5P_DEFAULT, filter_mask, offset, buf_size, data_buf) < 0)
        FAIL_STACK_ERROR

    /* Overwrite first chunk. */
    offset[2] = 0;
    if (H5Dwrite_chunk(did, H5P_DEFAULT, filter_mask, offset, buf_size, overwrite_buf) < 0)
        FAIL_STACK_ERROR

    /* Read the data back out */
    if (H5Dread(did, tid, H5S_ALL, H5S_ALL, H5P_DEFAULT, read_buf) < 0)
        FAIL_STACK_ERROR

    /* Ensure that the data are correct in chunk 1 */
    for (i = 0; i < OVERWRITE_CHUNK_NY; i++)
        for (j = 0; j < OVERWRITE_CHUNK_NX; j++) {
            if (read_buf[i][j] != OVERWRITE_VALUE)
                TEST_ERROR
        }

    if (H5Pclose(dcpl_id) < 0)
        FAIL_STACK_ERROR
    if (H5Sclose(sid) < 0)
        FAIL_STACK_ERROR
    if (H5Dclose(did) < 0)
        FAIL_STACK_ERROR

    PASSED();
    return 0;

error:
    H5E_BEGIN_TRY
    {
        H5Pclose(dcpl_id);
        H5Sclose(sid);
        H5Dclose(did);
    }
    H5E_END_TRY;

    H5_FAILED();
    return 1;
} /* end test_direct_chunk_overwrite_data() */

/*-------------------------------------------------------------------------
 * Function:    test_skip_compress_write1
 *
 * Purpose:    Test skipping compression filter when it is the only filter
 *              for the dataset
 *
 * Return:    Success:    0
 *            Failure:    1
 *
 * Programmer:  Raymond Lu
 *              30 November 2012
 *
 *-------------------------------------------------------------------------
 */
static int
test_skip_compress_write1(hid_t file)
{
    hid_t   dataspace = -1, dataset = -1;
    hid_t   mem_space = -1;
    hid_t   cparms = -1, dxpl = -1;
    hsize_t dims[2]       = {NX, NY};
    hsize_t maxdims[2]    = {H5S_UNLIMITED, H5S_UNLIMITED};
    hsize_t chunk_dims[2] = {CHUNK_NX, CHUNK_NY};
    herr_t  status;
    int     i, j, n;

    unsigned filter_mask = 0;
    int      direct_buf[CHUNK_NX][CHUNK_NY];
    int      check_chunk[CHUNK_NX][CHUNK_NY];
    hsize_t  offset[2]  = {0, 0};
    size_t   buf_size   = CHUNK_NX * CHUNK_NY * sizeof(int);
    int      aggression = 9; /* Compression aggression setting */

    unsigned read_filter_mask = 0; /* filter mask after direct read */
    int      read_direct_buf[CHUNK_NX][CHUNK_NY];
    hsize_t  read_buf_size = 0; /* buf size */

    hsize_t start[2];  /* Start of hyperslab */
    hsize_t stride[2]; /* Stride of hyperslab */
    hsize_t count[2];  /* Block count */
    hsize_t block[2];  /* Block sizes */

    TESTING("skipping compression filter for H5Dwrite_chunk/H5Dread_chunk");

    /*
     * Create the data space with unlimited dimensions.
     */
    if ((dataspace = H5Screate_simple(RANK, dims, maxdims)) < 0)
        goto error;

    if ((mem_space = H5Screate_simple(RANK, chunk_dims, NULL)) < 0)
        goto error;

    /*
     * Modify dataset creation properties, i.e. enable chunking and compression
     */
    if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
        goto error;

    if ((status = H5Pset_chunk(cparms, RANK, chunk_dims)) < 0)
        goto error;

    if ((status = H5Pset_deflate(cparms, (unsigned)aggression)) < 0)
        goto error;

    /*
     * Create a new dataset within the file using cparms
     * creation properties.
     */
    if ((dataset =
             H5Dcreate2(file, DATASETNAME2, H5T_NATIVE_INT, dataspace, H5P_DEFAULT, cparms, H5P_DEFAULT)) < 0)
        goto error;

    if ((dxpl = H5Pcreate(H5P_DATASET_XFER)) < 0)
        goto error;

    /* Initialize data for one chunk */
    for (i = n = 0; i < CHUNK_NX; i++)
        for (j = 0; j < CHUNK_NY; j++) {
            direct_buf[i][j] = n++;
        }

    /* write the uncompressed chunk data repeatedly to dataset, using the direct writing function.
     * Indicate skipping the compression filter.     */
    offset[0] = CHUNK_NX;
    offset[1] = CHUNK_NY;

    filter_mask = 0x00000001;

    if ((status = H5Dwrite_chunk(dataset, dxpl, filter_mask, offset, buf_size, direct_buf)) < 0)
        goto error;

    if (H5Fflush(dataset, H5F_SCOPE_LOCAL) < 0)
        goto error;

    if (H5Dclose(dataset) < 0)
        goto error;

    if ((dataset = H5Dopen2(file, DATASETNAME2, H5P_DEFAULT)) < 0)
        goto error;

    /*
     * Select hyperslab for the chunk just written in the file
     */
    start[0]  = CHUNK_NX;
    start[1]  = CHUNK_NY;
    stride[0] = 1;
    stride[1] = 1;
    count[0]  = 1;
    count[1]  = 1;
    block[0]  = CHUNK_NX;
    block[1]  = CHUNK_NY;
    if ((status = H5Sselect_hyperslab(dataspace, H5S_SELECT_SET, start, stride, count, block)) < 0)
        goto error;

    /* Read the chunk back */
    if ((status = H5Dread(dataset, H5T_NATIVE_INT, mem_space, dataspace, H5P_DEFAULT, check_chunk)) < 0)
        goto error;

    /* Check that the values read are the same as the values written */
    for (i = 0; i < CHUNK_NX; i++) {
        for (j = 0; j < CHUNK_NY; j++) {
            if (direct_buf[i][j] != check_chunk[i][j]) {
                HDprintf("    1. Read different values than written.");
                HDprintf("    At index %d,%d\n", i, j);
                HDprintf("    direct_buf=%d, check_chunk=%d\n", direct_buf[i][j], check_chunk[i][j]);
                goto error;
            }
        }
    }

    /* Query chunk storage size */
    if ((status = H5Dget_chunk_storage_size(dataset, offset, &read_buf_size)) < 0)
        goto error;
    if (read_buf_size != buf_size)
        goto error;

    /* Read the raw chunk back */
    HDmemset(&read_direct_buf, 0, sizeof(read_direct_buf));
    if ((status = H5Dread_chunk(dataset, H5P_DEFAULT, offset, &read_filter_mask, read_direct_buf)) < 0)
        goto error;
    if (read_filter_mask != filter_mask)
        goto error;

    /* Check that the direct chunk read is the same as the chunk written */
    for (i = 0; i < CHUNK_NX; i++) {
        for (j = 0; j < CHUNK_NY; j++) {
            if (direct_buf[i][j] != read_direct_buf[i][j]) {
                HDprintf("    1. Read different values than written.");
                HDprintf("    At index %d,%d\n", i, j);
                HDprintf("    direct_buf=%d, read_direct_buf=%d\n", direct_buf[i][j], read_direct_buf[i][j]);
                goto error;
            }
        }
    }

    /*
     * Close/release resources.
     */
    H5Dclose(dataset);
    H5Sclose(mem_space);
    H5Sclose(dataspace);
    H5Pclose(cparms);
    H5Pclose(dxpl);

    PASSED();
    return 0;

error:
    H5E_BEGIN_TRY
    {
        H5Dclose(dataset);
        H5Sclose(mem_space);
        H5Sclose(dataspace);
        H5Pclose(cparms);
        H5Pclose(dxpl);
    }
    H5E_END_TRY;

    H5_FAILED();
    return 1;
} /* test_skip_compress_write1() */

/*-------------------------------------------------------------------------
 * Function:    filter_bogus1
 *
 * Purpose:        A bogus filter that adds ADD_ON to the original value
 *
 * Return:        Success:    Data chunk size
 *
 * Programmer:  Raymond Lu
 *              30 November 2012
 *
 *-------------------------------------------------------------------------
 */
static size_t
filter_bogus1(unsigned int flags, size_t H5_ATTR_UNUSED cd_nelmts,
              const unsigned int H5_ATTR_UNUSED *cd_values, size_t nbytes, size_t *buf_size, void **buf)
{
    int *  int_ptr  = (int *)*buf; /* Pointer to the data values */
    size_t buf_left = *buf_size;   /* Amount of data buffer left to process */

    if (flags & H5Z_FLAG_REVERSE) { /* read */
        /* Substract the "add on" value to all the data values */
        while (buf_left > 0) {
            *int_ptr++ -= (int)ADD_ON;
            buf_left -= sizeof(int);
        }  /* end while */
    }      /* end if */
    else { /* write */
        /* Add the "add on" value to all the data values */
        while (buf_left > 0) {
            *int_ptr++ += (int)ADD_ON;
            buf_left -= sizeof(int);
        } /* end while */
    }     /* end else */

    return nbytes;
} /* filter_bogus1() */

/*-------------------------------------------------------------------------
 * Function:    filter_bogus2
 *
 * Purpose:    A bogus filter that multiplies the original value by FACTOR.
 *
 * Return:    Success:    Data chunk size
 *
 * Programmer:  Raymond Lu
 *              30 November 2012
 *-------------------------------------------------------------------------
 */
static size_t
filter_bogus2(unsigned int flags, size_t H5_ATTR_UNUSED cd_nelmts,
              const unsigned int H5_ATTR_UNUSED *cd_values, size_t nbytes, size_t *buf_size, void **buf)
{
    int *  int_ptr  = (int *)*buf; /* Pointer to the data values */
    size_t buf_left = *buf_size;   /* Amount of data buffer left to process */

    if (flags & H5Z_FLAG_REVERSE) { /* read */
        /* Substract the "add on" value to all the data values */
        while (buf_left > 0) {
            *int_ptr++ /= (int)FACTOR;
            buf_left -= sizeof(int);
        }  /* end while */
    }      /* end if */
    else { /* write */
        /* Add the "add on" value to all the data values */
        while (buf_left > 0) {
            *int_ptr++ *= (int)FACTOR;
            buf_left -= sizeof(int);
        } /* end while */
    }     /* end else */

    return nbytes;
} /* filter_bogus2() */

/*-------------------------------------------------------------------------
 * Function:    test_skip_compress_write2
 *
 * Purpose:    Test skipping compression filter when there are three filters
 *              for the dataset
 *
 * Return:    Success:    0
 *            Failure:    1
 *
 * Programmer:  Raymond Lu
 *              30 November 2012
 *
 *-------------------------------------------------------------------------
 */
static int
test_skip_compress_write2(hid_t file)
{
    hid_t   dataspace = -1, dataset = -1;
    hid_t   mem_space = -1;
    hid_t   cparms = -1, dxpl = -1;
    hsize_t dims[2]       = {NX, NY};
    hsize_t maxdims[2]    = {H5S_UNLIMITED, H5S_UNLIMITED};
    hsize_t chunk_dims[2] = {CHUNK_NX, CHUNK_NY};
    herr_t  status;
    int     i, j, n;

    unsigned filter_mask = 0; /* orig filter mask */
    int      origin_direct_buf[CHUNK_NX][CHUNK_NY];
    int      direct_buf[CHUNK_NX][CHUNK_NY];
    int      check_chunk[CHUNK_NX][CHUNK_NY];
    hsize_t  offset[2]  = {0, 0};
    size_t   buf_size   = CHUNK_NX * CHUNK_NY * sizeof(int);
    int      aggression = 9; /* Compression aggression setting */

    unsigned read_filter_mask = 0; /* filter mask after direct read */
    int      read_direct_buf[CHUNK_NX][CHUNK_NY];
    hsize_t  read_buf_size = 0; /* buf size */

    hsize_t start[2];  /* Start of hyperslab */
    hsize_t stride[2]; /* Stride of hyperslab */
    hsize_t count[2];  /* Block count */
    hsize_t block[2];  /* Block sizes */

    TESTING("skipping compression filters but keep two other filters");

    /*
     * Create the data space with unlimited dimensions.
     */
    if ((dataspace = H5Screate_simple(RANK, dims, maxdims)) < 0)
        goto error;

    if ((mem_space = H5Screate_simple(RANK, chunk_dims, NULL)) < 0)
        goto error;

    /*
     * Modify dataset creation properties, i.e. enable chunking and compression.
     * The order of filters is bogus 1 + deflate + bogus 2.
     */
    if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
        goto error;

    if ((status = H5Pset_chunk(cparms, RANK, chunk_dims)) < 0)
        goto error;

    /* Register and enable first bogus filter */
    if (H5Zregister(H5Z_BOGUS1) < 0)
        goto error;

    if (H5Pset_filter(cparms, H5Z_FILTER_BOGUS1, 0, (size_t)0, NULL) < 0)
        goto error;

    /* Enable compression filter */
    if ((status = H5Pset_deflate(cparms, (unsigned)aggression)) < 0)
        goto error;

    /* Register and enable second bogus filter */
    if (H5Zregister(H5Z_BOGUS2) < 0)
        goto error;

    if (H5Pset_filter(cparms, H5Z_FILTER_BOGUS2, 0, (size_t)0, NULL) < 0)
        goto error;

    /*
     * Create a new dataset within the file using cparms
     * creation properties.
     */
    if ((dataset =
             H5Dcreate2(file, DATASETNAME3, H5T_NATIVE_INT, dataspace, H5P_DEFAULT, cparms, H5P_DEFAULT)) < 0)
        goto error;

    if ((dxpl = H5Pcreate(H5P_DATASET_XFER)) < 0)
        goto error;

    /* Initialize data for one chunk. Apply operations of two bogus filters to the chunk */
    for (i = n = 0; i < CHUNK_NX; i++)
        for (j = 0; j < CHUNK_NY; j++) {
            origin_direct_buf[i][j] = n++;
            direct_buf[i][j]        = (origin_direct_buf[i][j] + ADD_ON) * FACTOR;
        }

    /* write the uncompressed chunk data repeatedly to dataset, using the direct writing function.
     * Indicate skipping the compression filter but keep the other two bogus filters */
    offset[0] = CHUNK_NX;
    offset[1] = CHUNK_NY;

    /* compression filter is the middle one to be skipped */
    filter_mask = 0x00000002;

    if ((status = H5Dwrite_chunk(dataset, dxpl, filter_mask, offset, buf_size, direct_buf)) < 0)
        goto error;

    if (H5Fflush(dataset, H5F_SCOPE_LOCAL) < 0)
        goto error;

    if (H5Dclose(dataset) < 0)
        goto error;

    if ((dataset = H5Dopen2(file, DATASETNAME3, H5P_DEFAULT)) < 0)
        goto error;

    /*
     * Select hyperslab for one chunk in the file
     */
    start[0]  = CHUNK_NX;
    start[1]  = CHUNK_NY;
    stride[0] = 1;
    stride[1] = 1;
    count[0]  = 1;
    count[1]  = 1;
    block[0]  = CHUNK_NX;
    block[1]  = CHUNK_NY;
    if ((status = H5Sselect_hyperslab(dataspace, H5S_SELECT_SET, start, stride, count, block)) < 0)
        goto error;

    /* Read the chunk back */
    if ((status = H5Dread(dataset, H5T_NATIVE_INT, mem_space, dataspace, H5P_DEFAULT, check_chunk)) < 0)
        goto error;

    /* Check that the values read are the same as the values written */
    for (i = 0; i < CHUNK_NX; i++) {
        for (j = 0; j < CHUNK_NY; j++) {
            if (origin_direct_buf[i][j] != check_chunk[i][j]) {
                HDprintf("    1. Read different values than written.");
                HDprintf("    At index %d,%d\n", i, j);
                HDprintf("    origin_direct_buf=%d, check_chunk=%d\n", origin_direct_buf[i][j],
                         check_chunk[i][j]);
                goto error;
            }
        }
    }

    /* Query chunk storage size */
    if ((status = H5Dget_chunk_storage_size(dataset, offset, &read_buf_size)) < 0)
        goto error;
    if (read_buf_size != buf_size)
        goto error;

    /* Read the raw chunk back */
    HDmemset(&read_direct_buf, 0, sizeof(read_direct_buf));
    if ((status = H5Dread_chunk(dataset, H5P_DEFAULT, offset, &read_filter_mask, read_direct_buf)) < 0)
        goto error;
    if (read_filter_mask != filter_mask)
        goto error;

    /* Check that the direct chunk read is the same as the chunk written */
    for (i = 0; i < CHUNK_NX; i++) {
        for (j = 0; j < CHUNK_NY; j++) {
            if (direct_buf[i][j] != read_direct_buf[i][j]) {
                HDprintf("    1. Read different values than written.");
                HDprintf("    At index %d,%d\n", i, j);
                HDprintf("    direct_buf=%d, read_direct_buf=%d\n", direct_buf[i][j], read_direct_buf[i][j]);
                goto error;
            }
        }
    }

    /*
     * Close/release resources.
     */
    H5Dclose(dataset);
    H5Sclose(mem_space);
    H5Sclose(dataspace);
    H5Pclose(cparms);
    H5Pclose(dxpl);

    PASSED();
    return 0;

error:
    H5E_BEGIN_TRY
    {
        H5Dclose(dataset);
        H5Sclose(mem_space);
        H5Sclose(dataspace);
        H5Pclose(cparms);
        H5Pclose(dxpl);
    }
    H5E_END_TRY;

    H5_FAILED();
    return 1;
} /* test_skip_compress_write2() */

/*-------------------------------------------------------------------------
 * Function:    test_data_conv
 *
 * Purpose:    Test data conversion
 *
 * Return:    Success:    0
 *            Failure:    1
 *
 * Programmer:  Raymond Lu
 *              30 November 2012
 *
 *-------------------------------------------------------------------------
 */
static int
test_data_conv(hid_t file)
{
    typedef struct {
        int a, b, c[4], d, e;
    } src_type_t;
    typedef struct {
        int a, c[4], e;
    } dst_type_t;

    hid_t         dataspace = -1, dataset = -1;
    hid_t         mem_space = -1;
    hid_t         cparms = -1, dxpl = -1;
    hsize_t       dims[2]       = {NX, NY};
    hsize_t       maxdims[2]    = {H5S_UNLIMITED, H5S_UNLIMITED};
    hsize_t       chunk_dims[2] = {CHUNK_NX, CHUNK_NY};
    herr_t        status;
    int           i, j, n;
    const hsize_t four = 4;
    hid_t         st = -1, dt = -1;
    hid_t         array_dt;

    unsigned   filter_mask = 0;
    src_type_t direct_buf[CHUNK_NX][CHUNK_NY];
    dst_type_t check_chunk[CHUNK_NX][CHUNK_NY];
    src_type_t read_chunk[CHUNK_NX][CHUNK_NY]; /* For H5Dread_chunk */

    hsize_t offset[2] = {0, 0};
    size_t  buf_size  = CHUNK_NX * CHUNK_NY * sizeof(src_type_t);

    hsize_t start[2];  /* Start of hyperslab */
    hsize_t stride[2]; /* Stride of hyperslab */
    hsize_t count[2];  /* Block count */
    hsize_t block[2];  /* Block sizes */

    TESTING("data conversion for H5Dwrite_chunk/H5Dread_chunk");

    /*
     * Create the data space with unlimited dimensions.
     */
    if ((dataspace = H5Screate_simple(RANK, dims, maxdims)) < 0)
        goto error;

    if ((mem_space = H5Screate_simple(RANK, chunk_dims, NULL)) < 0)
        goto error;

    /*
     * Modify dataset creation properties, i.e. enable chunking
     */
    if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
        goto error;

    if ((status = H5Pset_chunk(cparms, RANK, chunk_dims)) < 0)
        goto error;

    /* Build hdf5 datatypes */
    array_dt = H5Tarray_create2(H5T_NATIVE_INT, 1, &four);
    if ((st = H5Tcreate(H5T_COMPOUND, sizeof(src_type_t))) < 0 ||
        H5Tinsert(st, "a", HOFFSET(src_type_t, a), H5T_NATIVE_INT) < 0 ||
        H5Tinsert(st, "b", HOFFSET(src_type_t, b), H5T_NATIVE_INT) < 0 ||
        H5Tinsert(st, "c", HOFFSET(src_type_t, c), array_dt) < 0 ||
        H5Tinsert(st, "d", HOFFSET(src_type_t, d), H5T_NATIVE_INT) < 0 ||
        H5Tinsert(st, "e", HOFFSET(src_type_t, e), H5T_NATIVE_INT) < 0)
        goto error;

    if (H5Tclose(array_dt) < 0)
        goto error;

    array_dt = H5Tarray_create2(H5T_NATIVE_INT, 1, &four);
    if ((dt = H5Tcreate(H5T_COMPOUND, sizeof(dst_type_t))) < 0 ||
        H5Tinsert(dt, "a", HOFFSET(dst_type_t, a), H5T_NATIVE_INT) < 0 ||
        H5Tinsert(dt, "c", HOFFSET(dst_type_t, c), array_dt) < 0 ||
        H5Tinsert(dt, "e", HOFFSET(dst_type_t, e), H5T_NATIVE_INT) < 0)
        goto error;

    if (H5Tclose(array_dt) < 0)
        goto error;

    /*
     * Create a new dataset within the file using cparms
     * creation properties.
     */
    if ((dataset = H5Dcreate2(file, DATASETNAME4, st, dataspace, H5P_DEFAULT, cparms, H5P_DEFAULT)) < 0)
        goto error;

    if ((dxpl = H5Pcreate(H5P_DATASET_XFER)) < 0)
        goto error;

    /* Initialize data for one chunk */
    for (i = n = 0; i < CHUNK_NX; i++) {
        for (j = 0; j < CHUNK_NY; j++) {
            (direct_buf[i][j]).a    = i * j + 0;
            (direct_buf[i][j]).b    = i * j + 1;
            (direct_buf[i][j]).c[0] = i * j + 2;
            (direct_buf[i][j]).c[1] = i * j + 3;
            (direct_buf[i][j]).c[2] = i * j + 4;
            (direct_buf[i][j]).c[3] = i * j + 5;
            (direct_buf[i][j]).d    = i * j + 6;
            (direct_buf[i][j]).e    = i * j + 7;
        }
    }

    /* write the chunk data to dataset, using the direct writing function.
     * There should be no data conversion involved. */
    offset[0] = CHUNK_NX;
    offset[1] = CHUNK_NY;

    if ((status = H5Dwrite_chunk(dataset, dxpl, filter_mask, offset, buf_size, direct_buf)) < 0)
        goto error;

    if (H5Fflush(dataset, H5F_SCOPE_LOCAL) < 0)
        goto error;

    if (H5Dclose(dataset) < 0)
        goto error;

    if ((dataset = H5Dopen2(file, DATASETNAME4, H5P_DEFAULT)) < 0)
        goto error;

    /* Use H5Dread_chunk() to read the uncompressed data */
    if ((status = H5Dread_chunk(dataset, dxpl, offset, &filter_mask, read_chunk)) < 0)
        goto error;

    /* Check that the values read are the same as the values written */
    for (i = 0; i < CHUNK_NX; i++) {
        for (j = 0; j < CHUNK_NY; j++) {
            if ((direct_buf[i][j]).a != (read_chunk[i][j]).a ||
                (direct_buf[i][j]).b != (read_chunk[i][j]).b ||
                (direct_buf[i][j]).c[0] != (read_chunk[i][j]).c[0] ||
                (direct_buf[i][j]).c[1] != (read_chunk[i][j]).c[1] ||
                (direct_buf[i][j]).c[2] != (read_chunk[i][j]).c[2] ||
                (direct_buf[i][j]).c[3] != (read_chunk[i][j]).c[3] ||
                (direct_buf[i][j]).d != (read_chunk[i][j]).d ||
                (direct_buf[i][j]).e != (read_chunk[i][j]).e) {
                HDprintf("    1. Read different values than written.");
                HDprintf("    At index %d,%d\n", i, j);
                HDprintf("    src={a=%d, b=%d, c=[%d,%d,%d,%d], d=%d, e=%d\n", (direct_buf[i][j]).a,
                         (direct_buf[i][j]).b, (direct_buf[i][j]).c[0], (direct_buf[i][j]).c[1],
                         (direct_buf[i][j]).c[2], (direct_buf[i][j]).c[3], (direct_buf[i][j]).d,
                         (direct_buf[i][j]).e);
                HDprintf("    dst={a=%d, b=%d, c=[%d,%d,%d,%d], d=%d, e=%d\n", (read_chunk[i][j]).a,
                         (read_chunk[i][j]).b, (read_chunk[i][j]).c[0], (read_chunk[i][j]).c[1],
                         (read_chunk[i][j]).c[2], (read_chunk[i][j]).c[3], (read_chunk[i][j]).d,
                         (read_chunk[i][j]).e);

                goto error;
            }
        }
    }

    /*
     * Select hyperslab for the chunk just written in the file
     */
    start[0]  = CHUNK_NX;
    start[1]  = CHUNK_NY;
    stride[0] = 1;
    stride[1] = 1;
    count[0]  = 1;
    count[1]  = 1;
    block[0]  = CHUNK_NX;
    block[1]  = CHUNK_NY;
    if ((status = H5Sselect_hyperslab(dataspace, H5S_SELECT_SET, start, stride, count, block)) < 0)
        goto error;

    /* Read the chunk back. Data should be converted */
    if ((status = H5Dread(dataset, dt, mem_space, dataspace, H5P_DEFAULT, check_chunk)) < 0)
        goto error;

    /* Check that the values read are the same as the values written */
    for (i = 0; i < CHUNK_NX; i++) {
        for (j = 0; j < CHUNK_NY; j++) {
            if ((direct_buf[i][j]).a != (check_chunk[i][j]).a ||
                (direct_buf[i][j]).c[0] != (check_chunk[i][j]).c[0] ||
                (direct_buf[i][j]).c[1] != (check_chunk[i][j]).c[1] ||
                (direct_buf[i][j]).c[2] != (check_chunk[i][j]).c[2] ||
                (direct_buf[i][j]).c[3] != (check_chunk[i][j]).c[3] ||
                (direct_buf[i][j]).e != (check_chunk[i][j]).e) {
                HDprintf("    1. Read different values than written.");
                HDprintf("    At index %d,%d\n", i, j);
                HDprintf("    src={a=%d, b=%d, c=[%d,%d,%d,%d], d=%d, e=%d\n", (direct_buf[i][j]).a,
                         (direct_buf[i][j]).b, (direct_buf[i][j]).c[0], (direct_buf[i][j]).c[1],
                         (direct_buf[i][j]).c[2], (direct_buf[i][j]).c[3], (direct_buf[i][j]).d,
                         (direct_buf[i][j]).e);
                HDprintf("    dst={a=%d, c=[%d,%d,%d,%d], e=%d\n", (check_chunk[i][j]).a,
                         (check_chunk[i][j]).c[0], (check_chunk[i][j]).c[1], (check_chunk[i][j]).c[2],
                         (check_chunk[i][j]).c[3], (check_chunk[i][j]).e);

                goto error;
            }
        }
    }

    /*
     * Close/release resources.
     */
    H5Dclose(dataset);
    H5Sclose(mem_space);
    H5Sclose(dataspace);
    H5Pclose(cparms);
    H5Pclose(dxpl);
    H5Tclose(st);
    H5Tclose(dt);

    PASSED();
    return 0;

error:
    H5E_BEGIN_TRY
    {
        H5Dclose(dataset);
        H5Sclose(mem_space);
        H5Sclose(dataspace);
        H5Pclose(cparms);
        H5Pclose(dxpl);
        H5Tclose(st);
        H5Tclose(dt);
    }
    H5E_END_TRY;

    H5_FAILED();
    return 1;
} /* test_data_conv() */

/*-------------------------------------------------------------------------
 * Function:    test_invalid_parameters
 *
 * Purpose:     Test invalid parameters for H5Dwrite_chunk and H5Dread_chunk
 *
 * Return:      Success:    0
 *              Failure:    1
 *
 * Programmer:  Raymond Lu
 *              30 November 2012
 *
 *-------------------------------------------------------------------------
 */
static int
test_invalid_parameters(hid_t file)
{
    hid_t   dataspace = -1, dataset = -1;
    hid_t   mem_space = -1;
    hid_t   cparms = -1, dxpl = -1;
    hsize_t dims[2]       = {NX, NY};
    hsize_t chunk_dims[2] = {CHUNK_NX, CHUNK_NY};
    herr_t  status;
    int     i, j, n;

    unsigned filter_mask = 0;
    int      direct_buf[CHUNK_NX][CHUNK_NY];
    hsize_t  offset[2]  = {0, 0};
    size_t   buf_size   = CHUNK_NX * CHUNK_NY * sizeof(int);
    int      aggression = 9; /* Compression aggression setting */

    hsize_t chunk_nbytes; /* Chunk size */

    TESTING("invalid parameters for H5Dwrite_chunk/H5Dread_chunk");

    /*
     * Create the data space with unlimited dimensions.
     */
    if ((dataspace = H5Screate_simple(RANK, dims, NULL)) < 0)
        goto error;

    if ((mem_space = H5Screate_simple(RANK, chunk_dims, NULL)) < 0)
        goto error;

    /*
     * Modify dataset creation properties
     */
    if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
        goto error;

    /*
     * Create a new contiguous dataset to verify H5Dwrite_chunk/H5Dread_chunk doesn't work
     */
    if ((dataset =
             H5Dcreate2(file, DATASETNAME5, H5T_NATIVE_INT, dataspace, H5P_DEFAULT, cparms, H5P_DEFAULT)) < 0)
        goto error;

    if ((dxpl = H5Pcreate(H5P_DATASET_XFER)) < 0)
        goto error;

    /* Initialize data for one chunk */
    for (i = n = 0; i < CHUNK_NX; i++)
        for (j = 0; j < CHUNK_NY; j++) {
            direct_buf[i][j] = n++;
        }

    /* Try to write the chunk data to contiguous dataset.  It should fail */
    offset[0] = CHUNK_NX;
    offset[1] = CHUNK_NY;

    H5E_BEGIN_TRY
    {
        if ((status = H5Dwrite_chunk(dataset, dxpl, filter_mask, offset, buf_size, direct_buf)) != FAIL)
            goto error;
    }
    H5E_END_TRY;

    /* Try to get chunk size for a contiguous dataset.  It should fail */
    H5E_BEGIN_TRY
    {
        if ((status = H5Dget_chunk_storage_size(dataset, offset, &chunk_nbytes)) != FAIL)
            goto error;
    }
    H5E_END_TRY;

    /* Try to H5Dread_chunk from the contiguous dataset.  It should fail */
    H5E_BEGIN_TRY
    {
        if ((status = H5Dread_chunk(dataset, dxpl, offset, &filter_mask, direct_buf)) != FAIL)
            goto error;
    }
    H5E_END_TRY;

    if (H5Dclose(dataset) < 0)
        goto error;

    /* Create a chunked dataset with compression filter */
    if ((status = H5Pset_chunk(cparms, RANK, chunk_dims)) < 0)
        goto error;

    if ((status = H5Pset_deflate(cparms, (unsigned)aggression)) < 0)
        goto error;

    /*
     * Create a new dataset within the file using cparms
     * creation properties.
     */
    if ((dataset =
             H5Dcreate2(file, DATASETNAME6, H5T_NATIVE_INT, dataspace, H5P_DEFAULT, cparms, H5P_DEFAULT)) < 0)
        goto error;

    /* Check invalid dataset ID for H5Dwrite_chunk and H5Dread_chunk */
    H5E_BEGIN_TRY
    {
        if ((status = H5Dwrite_chunk((hid_t)-1, dxpl, filter_mask, offset, buf_size, direct_buf)) != FAIL)
            goto error;
    }
    H5E_END_TRY;

    H5E_BEGIN_TRY
    {
        if ((status = H5Dread_chunk((hid_t)-1, dxpl, offset, &filter_mask, direct_buf)) != FAIL)
            goto error;
    }
    H5E_END_TRY;

    /* Check invalid DXPL ID for H5Dwrite_chunk and H5Dread_chunk */
    H5E_BEGIN_TRY
    {
        if ((status = H5Dwrite_chunk(dataset, (hid_t)-1, filter_mask, offset, buf_size, direct_buf)) != FAIL)
            goto error;
    }
    H5E_END_TRY;

    H5E_BEGIN_TRY
    {
        if ((status = H5Dread_chunk(dataset, (hid_t)-1, offset, &filter_mask, direct_buf)) != FAIL)
            goto error;
    }
    H5E_END_TRY;

    /* Check invalid OFFSET for H5Dwrite_chunk and H5Dread_chunk */
    H5E_BEGIN_TRY
    {
        if ((status = H5Dwrite_chunk(dataset, dxpl, filter_mask, NULL, buf_size, direct_buf)) != FAIL)
            goto error;
    }
    H5E_END_TRY;

    H5E_BEGIN_TRY
    {
        if ((status = H5Dread_chunk(dataset, dxpl, NULL, &filter_mask, direct_buf)) != FAIL)
            goto error;
    }
    H5E_END_TRY;

    /* Check when OFFSET is out of dataset range for H5Dwrite_chunk and H5Dread_chunk */
    offset[0] = NX + 1;
    offset[1] = NY;
    H5E_BEGIN_TRY
    {
        if ((status = H5Dwrite_chunk(dataset, dxpl, filter_mask, offset, buf_size, direct_buf)) != FAIL)
            goto error;
    }
    H5E_END_TRY;

    H5E_BEGIN_TRY
    {
        if ((status = H5Dread_chunk(dataset, dxpl, offset, &filter_mask, direct_buf)) != FAIL)
            goto error;
    }
    H5E_END_TRY;

    /* Check when OFFSET is not on chunk boundary for H5Dwrite_chunk and H5Dread_chunk */
    offset[0] = CHUNK_NX;
    offset[1] = CHUNK_NY + 1;
    H5E_BEGIN_TRY
    {
        if ((status = H5Dwrite_chunk(dataset, dxpl, filter_mask, offset, buf_size, direct_buf)) != FAIL)
            goto error;
    }
    H5E_END_TRY;

    H5E_BEGIN_TRY
    {
        if ((status = H5Dread_chunk(dataset, dxpl, offset, &filter_mask, direct_buf)) != FAIL)
            goto error;
    }
    H5E_END_TRY;

    /* Check invalid buffer size for H5Dwrite_chunk only */
    offset[0] = CHUNK_NX;
    offset[1] = CHUNK_NY;
    buf_size  = 0;
    H5E_BEGIN_TRY
    {
        if ((status = H5Dwrite_chunk(dataset, dxpl, filter_mask, offset, buf_size, direct_buf)) != FAIL)
            goto error;
    }
    H5E_END_TRY;

    /* Check invalid data buffer for H5Dwrite_chunk and H5Dread_chunk */
    buf_size = CHUNK_NX * CHUNK_NY * sizeof(int);
    H5E_BEGIN_TRY
    {
        if ((status = H5Dwrite_chunk(dataset, dxpl, filter_mask, offset, buf_size, NULL)) != FAIL)
            goto error;
    }
    H5E_END_TRY;

    H5E_BEGIN_TRY
    {
        if ((status = H5Dread_chunk(dataset, dxpl, offset, &filter_mask, NULL)) != FAIL)
            goto error;
    }
    H5E_END_TRY;

    if (H5Dclose(dataset) < 0)
        goto error;

    /*
     * Close/release resources.
     */
    H5Sclose(mem_space);
    H5Sclose(dataspace);
    H5Pclose(cparms);
    H5Pclose(dxpl);

    PASSED();
    return 0;

error:
    H5E_BEGIN_TRY
    {
        H5Dclose(dataset);
        H5Sclose(mem_space);
        H5Sclose(dataspace);
        H5Pclose(cparms);
        H5Pclose(dxpl);
    }
    H5E_END_TRY;

    H5_FAILED();
    return 1;
} /* test_invalid_parameters() */

/*-------------------------------------------------------------------------
 * Function:    test_direct_chunk_read_no_cache
 *
 * Purpose:     Test the basic functionality of H5Dread_chunk with the
 *              chunk cache diabled.
 *
 * Return:      Success:        0
 *              Failure:        1
 *
 * Programmer:  Matthew Strong (GE Healthcare)
 *              14 February 2016
 *
 *-------------------------------------------------------------------------
 */
#ifdef H5_HAVE_FILTER_DEFLATE
static int
test_direct_chunk_read_no_cache(hid_t file)
{
    hid_t   dataspace = -1, dataset = -1;
    hid_t   mem_space = -1;
    hid_t   cparms = -1, dxpl = -1, dapl = -1;
    hsize_t dims[2]       = {NX, NY};
    hsize_t maxdims[2]    = {H5S_UNLIMITED, H5S_UNLIMITED};
    hsize_t chunk_dims[2] = {CHUNK_NX, CHUNK_NY};
    herr_t  status; /* status from H5 function calls */
    int     ret;    /* deflate return status */
    int     data[NX][NY];
    int     i, j, k, l, n; /* local index variables */

    unsigned filter_mask = 0;                 /* filter mask returned from H5Dread_chunk */
    int      direct_buf[CHUNK_NX][CHUNK_NY];  /* chunk read with H5Dread and manually decompressed */
    int      check_chunk[CHUNK_NX][CHUNK_NY]; /* chunk read with H5Dread */
    hsize_t  offset[2];                       /* chunk offset used for H5Dread_chunk */
    size_t   buf_size = CHUNK_NX * CHUNK_NY * sizeof(int);

    Bytef *z_src        = NULL; /* source buffer        */
    uLongf z_src_nbytes = (uLongf)DEFLATE_SIZE_ADJUST(buf_size);
    Bytef *z_dst        = (Bytef *)(direct_buf);
    uLong  z_dst_nbytes = (uLong)buf_size;
    int    aggression   = 9;    /* Compression aggression setting */
    void * outbuf       = NULL; /* Pointer to new buffer */

    hsize_t start[2];  /* Start of hyperslab */
    hsize_t stride[2]; /* Stride of hyperslab */
    hsize_t count[2];  /* Block count */
    hsize_t block[2];  /* Block sizes */

    TESTING("basic functionality of H5Dread_chunk (chunk cache disabled)");

    /* Create the data space with unlimited dimensions. */
    if ((dataspace = H5Screate_simple(RANK, dims, maxdims)) < 0)
        goto error;
    if ((mem_space = H5Screate_simple(RANK, chunk_dims, NULL)) < 0)
        goto error;

    /* Modify dataset creation properties, i.e. enable chunking and compression */
    if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
        goto error;
    if ((status = H5Pset_chunk(cparms, RANK, chunk_dims)) < 0)
        goto error;
    if ((status = H5Pset_deflate(cparms, (unsigned)aggression)) < 0)
        goto error;
    if ((dapl = H5Pcreate(H5P_DATASET_ACCESS)) < 0)
        goto error;

    /* Disable chunk cache by setting number of slots to 0 */
    if ((status = H5Pset_chunk_cache(dapl, 0, H5D_CHUNK_CACHE_NBYTES_DEFAULT, H5D_CHUNK_CACHE_W0_DEFAULT)) <
        0)
        goto error;

    /* Create a new dataset within the file using cparms creation properties. */
    if ((dataset = H5Dcreate2(file, DATASETNAME8, H5T_NATIVE_INT, dataspace, H5P_DEFAULT, cparms, dapl)) < 0)
        goto error;

    /* Initialize the dataset */
    for (i = n = 0; i < NX; i++)
        for (j = 0; j < NY; j++)
            data[i][j] = n++;

    if ((dxpl = H5Pcreate(H5P_DATASET_XFER)) < 0)
        goto error;

    /* Write the data for the dataset.
     * Data will skip chunk cache and go directly to disk. */
    if ((status = H5Dwrite(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, data)) < 0)
        goto error;

    /* Allocate output (compressed) buffer */
    outbuf = HDmalloc(z_src_nbytes);
    z_src  = (Bytef *)outbuf;

    /* For each chunk in the dataset, compare the result of H5Dread and H5Dread_chunk. */
    for (i = 0; i < NX / CHUNK_NX; i++) {
        for (j = 0; j < NY / CHUNK_NY; j++) {
            /* Select hyperslab for one chunk in the file */
            start[0]  = (hsize_t)i * CHUNK_NX;
            start[1]  = (hsize_t)j * CHUNK_NY;
            stride[0] = 1;
            stride[1] = 1;
            count[0]  = 1;
            count[1]  = 1;
            block[0]  = CHUNK_NX;
            block[1]  = CHUNK_NY;

            /* Hyperslab selection equals single chunk */
            if ((status = H5Sselect_hyperslab(dataspace, H5S_SELECT_SET, start, stride, count, block)) < 0)
                goto error;

            /* Read the chunk back */
            if ((status = H5Dread(dataset, H5T_NATIVE_INT, mem_space, dataspace, H5P_DEFAULT, check_chunk)) <
                0)
                goto error;

            offset[0] = (hsize_t)i * CHUNK_NX;
            offset[1] = (hsize_t)j * CHUNK_NY;
            /* Read the compressed chunk back using the direct read function. */
            if ((status = H5Dread_chunk(dataset, dxpl, offset, &filter_mask, outbuf)) < 0)
                goto error;

            /* Check filter mask return value */
            if (filter_mask != 0)
                goto error;

            /* Perform decompression from the source to the destination buffer */
            ret = uncompress(z_dst, &z_dst_nbytes, z_src, z_src_nbytes);

            /* Check for various zlib errors */
            if (Z_BUF_ERROR == ret) {
                HDfprintf(stderr, "overflow\n");
                goto error;
            }
            else if (Z_MEM_ERROR == ret) {
                HDfprintf(stderr, "deflate memory error\n");
                goto error;
            }
            else if (Z_DATA_ERROR == ret) {
                HDfprintf(stderr, "corrupted data\n");
                goto error;
            }
            else if (Z_OK != ret) {
                HDfprintf(stderr, "other deflate error\n");
                goto error;
            }

            /* Check that the decompressed values match those read from H5Dread */
            for (k = 0; k < CHUNK_NX; k++) {
                for (l = 0; l < CHUNK_NY; l++) {
                    if (direct_buf[k][l] != check_chunk[k][l]) {
                        HDprintf("\n    1. Read different values than written.");
                        HDprintf("    At index %d,%d\n", k, l);
                        HDprintf("    direct_buf=%d, check_chunk=%d\n", direct_buf[k][l], check_chunk[k][l]);
                        goto error;
                    }
                }
            }
        }
    }

    /* Close/release resources. */
    H5Dclose(dataset);
    H5Sclose(mem_space);
    H5Sclose(dataspace);
    H5Pclose(cparms);
    H5Pclose(dxpl);
    H5Pclose(dapl);

    if (outbuf)
        HDfree(outbuf);

    PASSED();
    return 0;

error:
    H5E_BEGIN_TRY
    {
        H5Dclose(dataset);
        H5Sclose(mem_space);
        H5Sclose(dataspace);
        H5Pclose(cparms);
        H5Pclose(dxpl);
        H5Pclose(dapl);
    }
    H5E_END_TRY;

    if (outbuf)
        HDfree(outbuf);

    H5_FAILED();
    return 1;
} /* test_direct_chunk_read_no_cache() */
#endif /* H5_HAVE_FILTER_DEFLATE */

#ifdef H5_HAVE_FILTER_DEFLATE
static int
test_direct_chunk_read_cache(hid_t file, hbool_t flush)
{
    hid_t   dataspace = -1, dataset = -1;
    hid_t   mem_space = -1;
    hid_t   cparms = -1, dxpl = -1;
    hsize_t dims[2]       = {NX, NY};
    hsize_t maxdims[2]    = {H5S_UNLIMITED, H5S_UNLIMITED};
    hsize_t chunk_dims[2] = {CHUNK_NX, CHUNK_NY};
    herr_t  status; /* status from H5 function calls */
    int     ret;    /* deflate return status */
    int     data[NX][NY];
    int     i, j, k, l, n; /* local index variables */

    unsigned filter_mask = 0;                 /* filter mask returned from H5Dread_chunk */
    int      direct_buf[CHUNK_NX][CHUNK_NY];  /* chunk read with H5Dread and manually decompressed */
    int      check_chunk[CHUNK_NX][CHUNK_NY]; /* chunk read with H5Dread */
    hsize_t  offset[2];                       /* chunk offset used for H5Dread_chunk */
    size_t   buf_size = CHUNK_NX * CHUNK_NY * sizeof(int);

    Bytef * z_src         = NULL; /* source buffer        */
    uLongf  z_src_nbytes  = (uLongf)DEFLATE_SIZE_ADJUST(buf_size);
    Bytef * z_dst         = (Bytef *)(direct_buf);
    uLong   z_dst_nbytes  = (uLong)buf_size;
    int     aggression    = 9;    /* Compression aggression setting */
    void *  outbuf        = NULL; /* Pointer to new buffer */
    hsize_t read_buf_size = 0;

    hsize_t start[2];  /* Start of hyperslab */
    hsize_t stride[2]; /* Stride of hyperslab */
    hsize_t count[2];  /* Block count */
    hsize_t block[2];  /* Block sizes */

    if (flush) {
        TESTING("basic functionality of H5Dread_chunk (flush chunk cache)");
    }
    else {
        TESTING("basic functionality of H5Dread_chunk (does not flush chunk cache)");
    }

    /* Create the data space with unlimited dimensions. */
    if ((dataspace = H5Screate_simple(RANK, dims, maxdims)) < 0)
        goto error;
    if ((mem_space = H5Screate_simple(RANK, chunk_dims, NULL)) < 0)
        goto error;

    /* Modify dataset creation properties, i.e. enable chunking and compression */
    if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
        goto error;
    if ((status = H5Pset_chunk(cparms, RANK, chunk_dims)) < 0)
        goto error;
    if ((status = H5Pset_deflate(cparms, (unsigned)aggression)) < 0)
        goto error;

    /* Create a new dataset within the file using cparms creation properties. */
    if ((dataset = H5Dcreate2(file, flush ? DATASETNAME9 : DATASETNAME10, H5T_NATIVE_INT, dataspace,
                              H5P_DEFAULT, cparms, H5P_DEFAULT)) < 0)
        goto error;

    /* Initialize the dataset */
    for (i = n = 0; i < NX; i++)
        for (j = 0; j < NY; j++)
            data[i][j] = n++;

    if ((dxpl = H5Pcreate(H5P_DATASET_XFER)) < 0)
        goto error;

    /* Write the data for the dataset.
     * It should stay in the chunk cache. */
    if ((status = H5Dwrite(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, data)) < 0)
        goto error;

    if (flush) {
        /* Flush the chunk cache to disk. Cache entry is not evicted. */
        if (H5Fflush(dataset, H5F_SCOPE_LOCAL) < 0)
            goto error;
    }

    /* Allocate output (compressed) buffer */
    outbuf = HDmalloc(z_src_nbytes);
    z_src  = (Bytef *)outbuf;

    /* For each chunk in the dataset, compare the result of H5Dread and H5Dread_chunk. */
    for (i = 0; i < NX / CHUNK_NX; i++) {
        for (j = 0; j < NY / CHUNK_NY; j++) {
            /* Select hyperslab for one chunk in the file */
            start[0]  = (hsize_t)i * CHUNK_NX;
            start[1]  = (hsize_t)j * CHUNK_NY;
            stride[0] = 1;
            stride[1] = 1;
            count[0]  = 1;
            count[1]  = 1;
            block[0]  = CHUNK_NX;
            block[1]  = CHUNK_NY;

            /* Hyperslab selection equals single chunk */
            if ((status = H5Sselect_hyperslab(dataspace, H5S_SELECT_SET, start, stride, count, block)) < 0)
                goto error;

            /* Read the chunk back */
            if ((status = H5Dread(dataset, H5T_NATIVE_INT, mem_space, dataspace, H5P_DEFAULT, check_chunk)) <
                0)
                goto error;

            offset[0] = (hsize_t)i * CHUNK_NX;
            offset[1] = (hsize_t)j * CHUNK_NY;

            /* Query chunk storage size */
            if ((status = H5Dget_chunk_storage_size(dataset, offset, &read_buf_size)) < 0)
                goto error;
            if (read_buf_size == 0)
                goto error;

            /* Read the compressed chunk back using the direct read function. */
            if ((status = H5Dread_chunk(dataset, dxpl, offset, &filter_mask, outbuf)) < 0)
                goto error;

            /* Check filter mask return value */
            if (filter_mask != 0)
                goto error;

            /* Perform decompression from the source to the destination buffer */
            ret = uncompress(z_dst, &z_dst_nbytes, z_src, z_src_nbytes);

            /* Check for various zlib errors */
            if (Z_BUF_ERROR == ret) {
                HDfprintf(stderr, "overflow\n");
                goto error;
            }
            else if (Z_MEM_ERROR == ret) {
                HDfprintf(stderr, "deflate memory error\n");
                goto error;
            }
            else if (Z_DATA_ERROR == ret) {
                HDfprintf(stderr, "corrupted data\n");
                goto error;
            }
            else if (Z_OK != ret) {
                HDfprintf(stderr, "other deflate error\n");
                goto error;
            }

            /* Check that the decompressed values match those read from H5Dread */
            for (k = 0; k < CHUNK_NX; k++) {
                for (l = 0; l < CHUNK_NY; l++) {
                    if (direct_buf[k][l] != check_chunk[k][l]) {
                        HDprintf("\n    1. Read different values than written.");
                        HDprintf("    At index %d,%d\n", k, l);
                        HDprintf("    direct_buf=%d, check_chunk=%d\n", direct_buf[k][l], check_chunk[k][l]);
                        goto error;
                    }
                }
            }
        }
    }

    /* Close/release resources. */
    H5Dclose(dataset);
    H5Sclose(mem_space);
    H5Sclose(dataspace);
    H5Pclose(cparms);
    H5Pclose(dxpl);

    if (outbuf)
        HDfree(outbuf);

    PASSED();
    return 0;

error:
    H5E_BEGIN_TRY
    {
        H5Dclose(dataset);
        H5Sclose(mem_space);
        H5Sclose(dataspace);
        H5Pclose(cparms);
        H5Pclose(dxpl);
    }
    H5E_END_TRY;

    if (outbuf)
        HDfree(outbuf);

    H5_FAILED();
    return 1;
} /* test_direct_chunk_read_cache() */
#endif /* H5_HAVE_FILTER_DEFLATE */

/*-------------------------------------------------------------------------
 * Function:    test_read_unfiltered_dset
 *
 * Purpose:     Test the basic functionality of H5Dread_chunk on a dataset
 *              without no filters applied.
 *
 * Return:      Success:        0
 *              Failure:        1
 *
 * Programmer:  Matthew Strong (GE Healthcare)
 *              30 November 2016
 *
 *-------------------------------------------------------------------------
 */
static int
test_read_unfiltered_dset(hid_t file)
{
    hid_t   dataspace = -1, dataset = -1;
    hid_t   mem_space = -1;
    hid_t   cparms = -1, dxpl = -1;
    hsize_t dims[2]       = {NX, NY};
    hsize_t maxdims[2]    = {H5S_UNLIMITED, H5S_UNLIMITED};
    hsize_t chunk_dims[2] = {CHUNK_NX, CHUNK_NY};
    herr_t  status;
    int     data[NX][NY];
    int     i, j, k, l, n;

    unsigned filter_mask = 0;
    int      direct_buf[CHUNK_NX][CHUNK_NY];
    int      check_chunk[CHUNK_NX][CHUNK_NY]; /* chunk read with H5Dread */
    hsize_t  offset[2]     = {0, 0};
    size_t   buf_size      = CHUNK_NX * CHUNK_NY * sizeof(int);
    hsize_t  read_buf_size = 0;

    hsize_t start[2];  /* Start of hyperslab */
    hsize_t stride[2]; /* Stride of hyperslab */
    hsize_t count[2];  /* Block count */
    hsize_t block[2];  /* Block sizes */

    TESTING("basic functionality of H5Dread_chunk on unfiltered datasets");

    /* Create the data space with unlimited dimensions. */
    if ((dataspace = H5Screate_simple(RANK, dims, maxdims)) < 0)
        goto error;
    if ((mem_space = H5Screate_simple(RANK, chunk_dims, NULL)) < 0)
        goto error;

    /* Modify dataset creation properties, i.e. enable chunking, no compression */
    if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
        goto error;
    if ((status = H5Pset_chunk(cparms, RANK, chunk_dims)) < 0)
        goto error;

    /* Create a new dataset within the file using cparms creation properties. */
    if ((dataset = H5Dcreate2(file, DATASETNAME12, H5T_NATIVE_INT, dataspace, H5P_DEFAULT, cparms,
                              H5P_DEFAULT)) < 0)
        goto error;

    /* Initialize the dataset */
    for (i = n = 0; i < NX; i++)
        for (j = 0; j < NY; j++)
            data[i][j] = n++;

    if ((dxpl = H5Pcreate(H5P_DATASET_XFER)) < 0)
        goto error;

    /* Write the data for the dataset.
     * It should stay in the chunk cache and will be evicted/flushed by
     * the H5Dread_chunk function call. */
    if ((status = H5Dwrite(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, data)) < 0)
        goto error;

    if (H5Fflush(dataset, H5F_SCOPE_LOCAL) < 0)
        goto error;

    /* For each chunk in the dataset, compare the result of H5Dread and H5Dread_chunk. */
    for (i = 0; i < NX / CHUNK_NX; i++) {
        for (j = 0; j < NY / CHUNK_NY; j++) {
            /* Select hyperslab for one chunk in the file */
            start[0]  = (hsize_t)i * CHUNK_NX;
            start[1]  = (hsize_t)j * CHUNK_NY;
            stride[0] = 1;
            stride[1] = 1;
            count[0]  = 1;
            count[1]  = 1;
            block[0]  = CHUNK_NX;
            block[1]  = CHUNK_NY;

            /* Hyperslab selection equals single chunk */
            if ((status = H5Sselect_hyperslab(dataspace, H5S_SELECT_SET, start, stride, count, block)) < 0)
                goto error;

            /* Read the chunk back */
            if ((status = H5Dread(dataset, H5T_NATIVE_INT, mem_space, dataspace, H5P_DEFAULT, check_chunk)) <
                0)
                goto error;

            /* Query chunk storage size */
            if ((status = H5Dget_chunk_storage_size(dataset, offset, &read_buf_size)) < 0)
                goto error;

            if (read_buf_size != buf_size)
                goto error;

            offset[0] = (hsize_t)i * CHUNK_NX;
            offset[1] = (hsize_t)j * CHUNK_NY;
            /* Read the raw chunk back */
            HDmemset(&direct_buf, 0, sizeof(direct_buf));
            filter_mask = UINT_MAX;
            if ((status = H5Dread_chunk(dataset, dxpl, offset, &filter_mask, direct_buf)) < 0)
                goto error;

            /* Check filter mask return value */
            if (filter_mask != 0)
                goto error;

            /* Check that the decompressed values match those read from H5Dread */
            for (k = 0; k < CHUNK_NX; k++) {
                for (l = 0; l < CHUNK_NY; l++) {
                    if (direct_buf[k][l] != check_chunk[k][l]) {
                        HDprintf("\n    1. Read different values than written.");
                        HDprintf("    At index %d,%d\n", k, l);
                        HDprintf("    direct_buf=%d, check_chunk=%d\n", direct_buf[k][l], check_chunk[k][l]);
                        goto error;
                    }
                }
            }
        }
    }

    /* Close/release resources. */
    H5Dclose(dataset);
    H5Sclose(mem_space);
    H5Sclose(dataspace);
    H5Pclose(cparms);
    H5Pclose(dxpl);

    PASSED();
    return 0;

error:
    H5E_BEGIN_TRY
    {
        H5Dclose(dataset);
        H5Sclose(mem_space);
        H5Sclose(dataspace);
        H5Pclose(cparms);
        H5Pclose(dxpl);
    }
    H5E_END_TRY;

    H5_FAILED();
    return 1;
} /* test_read_unfiltered_dset() */

/*-------------------------------------------------------------------------
 * Function:    test_read_unallocated_chunk
 *
 * Purpose:     Tests the H5Dread_chunk and H5Dget_chunk_storage_size with valid
 *              offets to chunks that have not been written to the dataset and are
 *              not allocated in the chunk storage on disk.
 *
 * Return:      Success:        0
 *              Failure:        1
 *
 * Programmer:  Matthew Strong (GE Healthcare)
 *              30 November 2016
 *
 *-------------------------------------------------------------------------
 */
static int
test_read_unallocated_chunk(hid_t file)
{
    hid_t   dataspace = -1, dataset = -1;
    hid_t   mem_space = -1;
    hid_t   cparms = -1, dxpl = -1;
    hsize_t dims[2]             = {NX, NY};
    hsize_t maxdims[2]          = {H5S_UNLIMITED, H5S_UNLIMITED};
    hsize_t chunk_dims[2]       = {CHUNK_NX, CHUNK_NY};
    hsize_t chunk_nbytes        = CHUNK_NX * CHUNK_NY * sizeof(int);
    hsize_t direct_chunk_nbytes = 0; /* size (bytes) of the on-disk chunk */
    herr_t  status;                  /* status from H5 function calls */
    hsize_t i, j;                    /* local index variables */

    unsigned filter_mask = 0;                /* filter mask returned from H5Dread_chunk */
    int      direct_buf[CHUNK_NX][CHUNK_NY]; /* chunk read with H5Dread and manually decompressed */
    hsize_t  offset[2];                      /* chunk offset used for H5Dread_chunk */

    TESTING("H5Dread_chunk with unallocated chunks");

    /* Create the data space with unlimited dimensions. */
    if ((dataspace = H5Screate_simple(RANK, dims, maxdims)) < 0)
        FAIL_STACK_ERROR;
    if ((mem_space = H5Screate_simple(RANK, chunk_dims, NULL)) < 0)
        FAIL_STACK_ERROR;

    /* Modify dataset creation properties, i.e. enable chunking, no compression */
    if ((cparms = H5Pcreate(H5P_DATASET_CREATE)) < 0)
        FAIL_STACK_ERROR;
    if ((status = H5Pset_chunk(cparms, RANK, chunk_dims)) < 0)
        FAIL_STACK_ERROR;

    /* Create a new dataset within the file using cparms creation properties. */
    if ((dataset = H5Dcreate2(file, DATASETNAME11, H5T_NATIVE_INT, dataspace, H5P_DEFAULT, cparms,
                              H5P_DEFAULT)) < 0)
        FAIL_STACK_ERROR;

    if ((dxpl = H5Pcreate(H5P_DATASET_XFER)) < 0)
        FAIL_STACK_ERROR;

    /* Write a single chunk to intialize the chunk storage */
    HDmemset(direct_buf, 0, CHUNK_NX * CHUNK_NY * sizeof(int));
    offset[0] = 0;
    offset[1] = 0;

    if (H5Dwrite_chunk(dataset, dxpl, filter_mask, offset, chunk_nbytes, direct_buf) < 0)
        FAIL_STACK_ERROR;

    /* Attempt to read each chunk in the dataset. Chunks are not allocated,
     * therefore we expect the result of H5Dread_chunk to fail. Chunk idx starts
     * at 1, since one chunk was written to init the chunk storage. */
    for (i = 1; i < NX / CHUNK_NX; i++) {
        for (j = 0; j < NY / CHUNK_NY; j++) {

            offset[0] = i * CHUNK_NX;
            offset[1] = j * CHUNK_NY;

            /* Read a non-existant chunk using the direct read function. */
            H5E_BEGIN_TRY
            {
                status = H5Dread_chunk(dataset, dxpl, offset, &filter_mask, &direct_buf);
            }
            H5E_END_TRY;

            /* Check that the chunk read call does not succeed. */
            if (status != -1)
                TEST_ERROR

            /* Query the size of the non-existant chunk */
            direct_chunk_nbytes = ULONG_MAX;
            H5E_BEGIN_TRY
            {
                status = H5Dget_chunk_storage_size(dataset, offset, &direct_chunk_nbytes);
            }
            H5E_END_TRY;

            /* Check that the chunk storage size call does not succeed. */
            if (status != -1)
                TEST_ERROR
            if (direct_chunk_nbytes != ULONG_MAX)
                TEST_ERROR
        }
    }

    /* Close/release resources. */
    if (H5Dclose(dataset) < 0)
        FAIL_STACK_ERROR;
    if (H5Sclose(mem_space) < 0)
        FAIL_STACK_ERROR;
    if (H5Sclose(dataspace) < 0)
        FAIL_STACK_ERROR;
    if (H5Pclose(cparms) < 0)
        FAIL_STACK_ERROR;
    if (H5Pclose(dxpl) < 0)
        FAIL_STACK_ERROR;

    PASSED();
    return 0;

error:
    H5E_BEGIN_TRY
    {
        H5Dclose(dataset);
        H5Sclose(mem_space);
        H5Sclose(dataspace);
        H5Pclose(cparms);
        H5Pclose(dxpl);
    }
    H5E_END_TRY;

    H5_FAILED();
    return 1;
} /* test_read_unallocated_chunk() */

/*-------------------------------------------------------------------------
 * Function:    test_single_chunk
 *
 * Purpose:     This is to verify the fix for jira issue HDFFV-10425.
 *              The problem was due to a bug in the internal ilbrary routine
 *              H5D__chunk_direct_write() which passed a null dataset
 *              pointer to the insert callback for the chunk index type.
 *              Currently, the single chunk index is the only one that
 *              used the dataset pointer in the insert callback.
 *
 *              This routine is based on the test program attached to
 *              this jira issue:
 *                  Create a file with the latest format and a chunked dataset
 *                  with one single chunk.  The library will use single chunk
 *                  index for the dataset.
 *                  Verify that the data read is the same as the written data.
 *
 *              Since expanded to test multiple combinations of cases
 *              involving a single chunk
 *
 * Return:      Success:        0
 *              Failure:        1
 *
 *-------------------------------------------------------------------------
 */
static int
test_single_chunk(unsigned config)
{
    hid_t    fid       = H5I_INVALID_HID;  /* File ID */
    hid_t    fapl      = H5I_INVALID_HID;  /* File access property list ID */
    hid_t    sid       = H5I_INVALID_HID;  /* Dataspace ID */
    hid_t    did       = H5I_INVALID_HID;  /* Dataset ID */
    hid_t    dcpl      = H5I_INVALID_HID;  /* Dataset creation property list */
    hsize_t  dims[2]   = {DIM0, DIM1};     /* Dimension sizes */
    hsize_t  chunk[2]  = {CHUNK0, CHUNK1}; /* Chunk dimension sizes */
    hsize_t  offset[2] = {0, 0};           /* Offset for writing */
    uint32_t filters;                      /* Filter mask out */
    int      wdata[DIM0][DIM1];            /* Write buffer */
    int      rdata[DIM0][DIM1];            /* Read buffer */
    int      i, j;                         /* Local index variable */

    TESTING("Single chunk I/O");

    /* Initialize data */
    for (i = 0; i < DIM0; i++)
        for (j = 0; j < DIM1; j++)
            wdata[i][j] = j / CHUNK0;

    /* Create a new file with the latest format  */
    if ((fapl = H5Pcreate(H5P_FILE_ACCESS)) < 0)
        FAIL_STACK_ERROR;
    if (config & CONFIG_LATEST)
        if (H5Pset_libver_bounds(fapl, H5F_LIBVER_LATEST, H5F_LIBVER_LATEST) < 0)
            FAIL_STACK_ERROR;
    if ((fid = H5Fcreate(FILE, H5F_ACC_TRUNC, H5P_DEFAULT, fapl)) < 0)
        FAIL_STACK_ERROR;

    /* Create dataspace */
    if ((sid = H5Screate_simple(2, dims, NULL)) < 0)
        FAIL_STACK_ERROR;

    /* Create the dataset creation property list and set the chunk size */
    if ((dcpl = H5Pcreate(H5P_DATASET_CREATE)) < 0)
        FAIL_STACK_ERROR;
    if (H5Pset_chunk(dcpl, 2, chunk) < 0)
        FAIL_STACK_ERROR;

    /* Create the dataset */
    if ((did = H5Dcreate2(fid, DATASET, H5T_NATIVE_INT, sid, H5P_DEFAULT, dcpl, H5P_DEFAULT)) < 0)
        FAIL_STACK_ERROR;

    if (config & CONFIG_DIRECT_WRITE) {
        /* Write the data directly to the dataset */
        if (H5Dwrite_chunk(did, H5P_DEFAULT, 0, offset, CHUNK0 * CHUNK1 * 4, (void *)wdata) < 0)
            FAIL_STACK_ERROR;
    } /* end if */
    else
        /* Write the data to the dataset */
        if (H5Dwrite(did, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, H5P_DEFAULT, (void *)wdata) < 0)
        FAIL_STACK_ERROR;

    /*
     * Close and release resources.
     */
    if (H5Pclose(dcpl) < 0)
        FAIL_STACK_ERROR;
    if (config & CONFIG_REOPEN_DSET)
        if (H5Dclose(did) < 0)
            FAIL_STACK_ERROR;
    if (H5Sclose(sid) < 0)
        FAIL_STACK_ERROR;
    if (H5Pclose(fapl) < 0)
        FAIL_STACK_ERROR;
    if (config & CONFIG_REOPEN_FILE)
        if (H5Fclose(fid) < 0)
            FAIL_STACK_ERROR;

    /* Open the file and dataset with default properties  */
    if (config & CONFIG_REOPEN_FILE)
        if ((fid = H5Fopen(FILE, H5F_ACC_RDONLY, H5P_DEFAULT)) < 0)
            FAIL_STACK_ERROR;
    if (config & CONFIG_REOPEN_DSET)
        if ((did = H5Dopen2(fid, DATASET, H5P_DEFAULT)) < 0)
            FAIL_STACK_ERROR;

    /* Retrieve dataset creation property list */
    if ((dcpl = H5Dget_create_plist(did)) < 0)
        FAIL_STACK_ERROR;

    if (config & CONFIG_DIRECT_READ) {
        /* Read the data directly */
        if (H5Dread_chunk(did, H5P_DEFAULT, offset, &filters, rdata) < 0)
            FAIL_STACK_ERROR;

        /* Verify returned filter mask */
        if (filters != 0)
            TEST_ERROR
    } /* end if */
    else
        /* Read the data */
        if (H5Dread(did, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, H5P_DEFAULT, rdata) < 0)
        FAIL_STACK_ERROR;

    /* Verify that the data read was correct.  */
    for (i = 0; i < DIM0; i++)
        for (j = 0; j < DIM1; j++)
            if (rdata[i][j] != wdata[i][j])
                TEST_ERROR

    /*
     * Close and release resources
     */
    if (H5Pclose(dcpl) < 0)
        FAIL_STACK_ERROR;
    if (H5Dclose(did) < 0)
        FAIL_STACK_ERROR;
    if (H5Fclose(fid) < 0)
        FAIL_STACK_ERROR;

    PASSED();
    return 0;

error:
    H5E_BEGIN_TRY
    {
        H5Dclose(did);
        H5Sclose(sid);
        H5Pclose(dcpl);
        H5Pclose(fapl);
        H5Fclose(fid);
    }
    H5E_END_TRY;

    H5_FAILED();
    return 1;
} /* test_single_chunk_latest() */

/*-------------------------------------------------------------------------
 * Function:    Main function
 *
 * Purpose:        Test direct chunk write function H5Dwrite_chunk and
 *              chunk direct read function H5Dread_chunk
 *
 * Return:        Success:    0
 *                Failure:    1
 *
 * Programmer:  Raymond Lu
 *              30 November 2012
 *
 *-------------------------------------------------------------------------
 */
int
main(void)
{
    hid_t    file_id;
    unsigned config;
    int      nerrors = 0;

    /*
     * Create a new file. If file exists its contents will be overwritten.
     */
    if ((file_id = H5Fcreate(FILE_NAME, H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT)) < 0)
        goto error;

        /* Test direct chunk write and direct chunk read */
#ifdef H5_HAVE_FILTER_DEFLATE
    nerrors += test_direct_chunk_write(file_id);
#endif /* H5_HAVE_FILTER_DEFLATE */
    nerrors += test_direct_chunk_overwrite_data(file_id);
    nerrors += test_skip_compress_write1(file_id);
    nerrors += test_skip_compress_write2(file_id);
    nerrors += test_data_conv(file_id);
    nerrors += test_invalid_parameters(file_id);

    /* Test direct chunk read */
#ifdef H5_HAVE_FILTER_DEFLATE
    nerrors += test_direct_chunk_read_no_cache(file_id);
    nerrors += test_direct_chunk_read_cache(file_id, TRUE);
    nerrors += test_direct_chunk_read_cache(file_id, FALSE);
#endif /* H5_HAVE_FILTER_DEFLATE */
    nerrors += test_read_unfiltered_dset(file_id);
    nerrors += test_read_unallocated_chunk(file_id);

    /* Loop over test configurations */
    for (config = 0; config < CONFIG_END; config++) {
        hbool_t need_comma = FALSE;

        /* Check for invalid combinations */
        if ((config & CONFIG_REOPEN_FILE) && !(config & CONFIG_REOPEN_DSET))
            continue;

        /* Print configuration */
        HDprintf("Configuration: ");
        if (config == 0)
            HDprintf("<empty>");
        if (config & CONFIG_LATEST) {
            if (need_comma)
                HDprintf(", ");
            HDprintf("latest format");
            need_comma = TRUE;
        } /* end if */
        if (config & CONFIG_REOPEN_FILE) {
            if (need_comma)
                HDprintf(", ");
            HDprintf("reopen file");
            need_comma = TRUE;
        } /* end if */
        else if (config & CONFIG_REOPEN_DSET) {
            if (need_comma)
                HDprintf(", ");
            HDprintf("reopen dataset");
            need_comma = TRUE;
        } /* end if */
        if (config & CONFIG_DIRECT_WRITE) {
            if (need_comma)
                HDprintf(", ");
            HDprintf("direct write");
            need_comma = TRUE;
        } /* end if */
        if (config & CONFIG_DIRECT_READ) {
            if (need_comma)
                HDprintf(", ");
            HDprintf("direct read");
            need_comma = TRUE;
        } /* end if */
        HDprintf(":\n");
        HDfflush(stdout);

        nerrors += test_single_chunk(config);
    } /* end for */

    if (H5Fclose(file_id) < 0)
        goto error;

    /* check for errors */
    if (nerrors)
        goto error;

    HDputs("All direct chunk read/write tests passed.");
    return EXIT_SUCCESS;

error:
    HDputs("*** TESTS FAILED ***");
    return EXIT_FAILURE;
}