/* Copyright 2019, UCAR/Unidata See COPYRIGHT file for copying and redistribution conditions. This program tests netcdf-4 parallel I/O using zlib compression while writing. Ed Hartnett, 12/19/2019 */ #include #include "err_macros.h" #include #define FILE_NAME "tst_parallel_zlib2.nc" #define NDIMS 3 #define DIMSIZE 24 #define QTR_DATA (DIMSIZE * DIMSIZE / 4) #define NUM_PROC 4 #define NUM_SLABS 10 int main(int argc, char **argv) { /* MPI stuff. */ int mpi_size, mpi_rank; MPI_Comm comm = MPI_COMM_WORLD; MPI_Info info = MPI_INFO_NULL; /* Netcdf-4 stuff. */ int ncid, v1id, dimids[NDIMS]; size_t start[NDIMS], count[NDIMS]; int i, res; int slab_data[DIMSIZE * DIMSIZE / 4]; /* one slab */ /* Initialize MPI. */ MPI_Init(&argc, &argv); MPI_Comm_size(MPI_COMM_WORLD, &mpi_size); MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); /* Create phony data. We're going to write a 24x24 array of ints, in 4 sets of 144. */ for (i = 0; i < DIMSIZE * DIMSIZE / 4; i++) slab_data[i] = mpi_rank; if (!mpi_rank) { printf("\n*** Testing parallel writes with zlib some more.\n"); printf("*** testing simple write with zlib..."); } { nc_set_log_level(3); /* Create a parallel netcdf-4 file. */ if (nc_create_par(FILE_NAME, NC_NETCDF4, comm, info, &ncid)) ERR; /* Create three dimensions. */ if (nc_def_dim(ncid, "d1", DIMSIZE, dimids)) ERR; if (nc_def_dim(ncid, "d2", DIMSIZE, &dimids[1])) ERR; if (nc_def_dim(ncid, "d3", NUM_SLABS, &dimids[2])) ERR; /* Create one var. Turn on deflation. */ if ((res = nc_def_var(ncid, "v1", NC_INT, NDIMS, dimids, &v1id))) ERR; /* Setting deflate only will work for HDF5-1.10.2 and later * versions. */ res = nc_def_var_deflate(ncid, 0, 0, 1, 1); #ifdef HDF5_1_10_2 if (res) ERR; #else if (res != NC_EINVAL) ERR; #endif /* Setting fletcher32 only will work for HDF5-1.10.2 and later * versions. */ res = nc_def_var_fletcher32(ncid, 0, 1); #ifdef HDF5_1_10_2 if (res) ERR; #else if (res != NC_EINVAL) ERR; #endif /* Write metadata to file. */ if (nc_enddef(ncid)) ERR; /* Set up slab for this process. */ start[0] = mpi_rank * DIMSIZE/mpi_size; start[1] = 0; count[0] = DIMSIZE/mpi_size; count[1] = DIMSIZE; count[2] = 1; /*printf("mpi_rank=%d start[0]=%d start[1]=%d count[0]=%d count[1]=%d\n", mpi_rank, start[0], start[1], count[0], count[1]);*/ /* Should not be allowed to change access to independent, * because filters are in use. */ if (nc_var_par_access(ncid, v1id, NC_INDEPENDENT) != NC_EINVAL) ERR; /* Write slabs of data. */ for (start[2] = 0; start[2] < NUM_SLABS; start[2]++) if (nc_put_vara_int(ncid, v1id, start, count, slab_data)) ERR; /* Close the netcdf file. */ if (nc_close(ncid)) ERR; /* Check file. */ { int shuffle_in, deflate_in, deflate_level_in; /* Reopen the file for parallel access. */ if (nc_open_par(FILE_NAME, NC_NOWRITE, comm, info, &ncid)) ERR; /* Check state of deflate. */ if (nc_inq_var_deflate(ncid, 0, &shuffle_in, &deflate_in, &deflate_level_in)) ERR; if (shuffle_in || !deflate_in || deflate_level_in != 1) ERR; /* Close the netcdf file. */ if (nc_close(ncid)) ERR; } } if (!mpi_rank) SUMMARIZE_ERR; /* Shut down MPI. */ MPI_Finalize(); if (!mpi_rank) FINAL_RESULTS; return 0; }