Commit Graph

7 Commits

Author SHA1 Message Date
Albert Cheng
bb2b806485 [svn-r12407] Purpose:
slight cleanup.

Description:
Changed the name of write type from write to write_pattern.

Platforms tested:
h5committested.
2006-06-06 10:24:38 -05:00
Albert Cheng
fbb0267346 [svn-r12405] Purpose:
Bug fix.

Description:
It failed when only 1 is used to test.

Solution:
Cleanup the code a little so that it works for any number
of processes to invoke it.

Platforms tested:
h5committested,
2006-06-06 01:55:26 -05:00
Albert Cheng
849a6c299d [svn-r12396] Purpose:
Bug fix.

Description:
Sometimes the parallel prefix is given in the form of nfs:/mnt/pfs which
if given to a non-MPIO VFD like the default H5Fcreate, it would fail.
Called h5_rmprefix which would return the non-prefix component
part of the file name which would be okay for the default H5Fcreate and
such.

Platforms tested:
Tested in heping parallel.
2006-06-01 17:57:29 -05:00
Albert Cheng
8698c5ce5e [svn-r12380] Purpose:
Feature
(Code and tests are done by Christian.  I just committed the code.)

Description:
Rewrote the purpose of this test. Now it tests these three cases,
/*
 * Test following possible scenarios,
 * Case 1:
 * Sequential create a file and dataset with H5D_ALLOC_TIME_EARLY and large
 * size, no write, close, reopen in parallel, read to verify all return
 * the fill value.
 * Case 2:
 * Sequential create a file and dataset with H5D_ALLOC_TIME_EARLY but small
 * size, no write, close, reopen in parallel, extend to large size, then close,
 * then reopen in parallel and read to verify all return the fill value.
 * Case 3:
 * Sequential create a file and dataset with H5D_ALLOC_TIME_EARLY and large
 * size, write just a small part of the dataset (second to the last), close,
 * then reopen in parallel, read to verify all return the fill value except
 * those small portion that has been written.  Without closing it, writes
 * all parts of the dataset in a interleave pattern, close it, and reopen
 * it, read to verify all data are as written.
 */

Platforms tested:
Tested in copper, tg-ncsa and heping, all in parallel mode.
2006-05-26 17:11:00 -05:00
Albert Cheng
c12eeffd9f [svn-r11349] Purpose:
Bug fix.

Description:
Some tests showed the filesize was not as expected.  But the error was
intermittent.  This was a racing condition as some processes finish
extend_chunked_dataset() sooner than others and return to the main
body which proceeds to call the next test which also uses the same
test data file and alters it.  That messes up the "slower" processes
which then see unexpected filesize.

Also, the routine create_chunked_dataset() which creates test data file
actually was executed by all processes.  That is wrong.

Solution:
Added a barrier at the end of extend_chunked_dataset to make sure all
processes are done with the test data file before returning.

Changed create_chunked_dataset such that only one process would create
the test data file.  The rest does nothing but just wait for it to finish.

Platforms tested:
Tested in TG-NCSA in which the errors were detected.

Misc. update:
2005-09-05 15:23:02 -05:00
Albert Cheng
40a1e57caa [svn-r11339] Purpose:
Bug fix #281

Description:
Committed a wrong copy in the previous checkin.

Solution:
Checked in the right one and did some code cleanup, rearrangment.

Platforms tested:
heping pp.

Misc. update:
2005-09-02 09:44:40 -05:00
Albert Cheng
3182a8a040 [svn-r11312] Purpose:
Bug #281

Description:
A dataset created in serial mode with H5D_ALLOC_TIME_INCR allocation setting
was not extendible, either explicitly by H5Dextend or implicitly by writing
to unallocated chunks.  This was because parallel mode expects the allocation
mode be H5D_ALLOC_TIME_INCR only.

Solution:
Modified library to allocate more space when needed or directed if the
file is opened by parallel mode, independent of what the dataset allocation
mode is.

Platforms tested:
Heping pp.
2005-08-29 23:35:17 -05:00