Description: After I added H5S_NULL dataspace to dumper test, the output violate
XML schema for attribute data.
Solution: changed.
Platforms tested: RH8(fuss), minor change, only dumper is involved.
h5dump new feature
Description:
add processing of tab characters (in the context of the new option process CR/LF)
Solution:
Platforms tested:
linux
solaris
AIX
Misc. update:
bug fix
Description:
a prototype for a new function used by h5tools_str.c was accidently left in that file
the c++ treats this as an error
Solution:
removed the prototype, and pointed to the correct one (in h5tools_ref.h )
Platforms tested:
Free BSD (c++)
linux
Misc. update:
Code cleanup
Description:
Add more files that should be removed to the clean & distclean targets.
Platforms tested:
FreeBSD 4.10 (sleipnir)
Too minor to require h5committest
bug fix
Description:
the file tfilters.h5 was missing the szip dataset, due to a absent szip configuration when
it was generated
Solution:
generated it it again with szip on
Platforms tested:
linux (smalll change )
Misc. update:
h5dump update
Description:
updated the h5dump test files generator , so that it coincides with the 1.6 branch
Solution:
Platforms tested:
linux (small)
Misc. update:
Description:
added the code for print strings with new line and display the path of references (new source files h5tools_ref.c and .h )
added a test suite in testh5dump.sh.in for
( note : to create testh5dump.sh , one must redo ./configure; this detects the availability of filters
and generates testh5dump.sh accordingly)
1) storage layout
2) fill value
3) print reference with path
4) print strings with new lines
5) filters
Solution:
Platforms tested:
linux
solaris
AIX
Misc. update:
Bug fix
Description:
Correct path to new testh5dump.sh, now that it's generated.
Platforms tested:
FreeBSD 4.10 (sleipnir) w & w/o srcdir
Too minor to require h5committest
Update shell scripts
Description:
Switch to generating the testh5dump.sh script at configure time, so we can
determine which filters are available to test.
Platforms tested:
FreeBSD 4.9 (sleipnir)
too small to require h5committest
h5dump new version
Description:
added the changes already made for 1.6
support for dumping of
1) filters
2) storage layout
3) fill value
4) comments
5) superblock
6) file contents
7) array indices
Solution:
Platforms tested:
linux
solaris
AIX
Misc. update:
Refactor code
Description:
Move chunk and contiguous cached raw data from file information to dataset
information. This simplifies a number of internal interfaces, aligns the
code with it's purpose better and should allow more optimizations to the
chunked data I/O performance.
Platforms tested:
Solaris 2.7 (arabica)
FreeBSD 4.10 (sleipnir)
h5committest
bug fix
Description:
Currently h5dump test on windows won't separate error form standard output; which shows at function error_message.
A #ifdef WIN32 macro at this file make stderr to be the same as stdout on windows.
Solution:
Separate standard error from standard output.
Platforms tested:
windows xp VS 6.0(by xuan bai), no needs to test on other platforms
Misc. update:
Bug fixes
Description:
Updated dependencies
Fixed error with C++ compiler builds of main library
Added H5Pset_data_transform to MPE info
Platforms tested:
FreeBSD 4.9 (sleipnir)
Code cleanup
Description:
Clean up lots of warnings based on those reported from the SGI compilers
as well as gcc.
Platforms tested:
SGI O3900, IRIX64 6.5 (Cheryl's SGI machine)
FreeBSD 4.9 (sleipnir) w/ & w/o parallel
h5committest
Code cleanup
Description:
Regenerated with older version of dataspace information, since these files
don't contain null dataspaces.
Platforms tested:
FreeBSD 4.9 (sleipnir)
h5committest
bug fix
Description:
the routine for reading the options from a file had the "old" parse syntax
Solution:
Platforms tested:
linux
solaris
AIX
Misc. update:
Bug fix.
Description:
Temporary output files were created in the source directory.
If multiple build/tests using the same source directory occur
at the same time, they would conflict because of the same file
names.
Solution:
Changed it to create temporary output file in the current build
directory.
Platforms tested:
sol.
Misc. update:
new test
Description:
add a test that generates a file with "holes" and then uses h5repack to clean
the empty space. it works
Solution:
Platforms tested:
linux
Misc. update:
Description: Mainly are header message changes for dataspace. In last round
of check-in, a new header message for dataspace to created, which is not a good
way. Now, there will be no new message for dataspace, but just add the type of
dataspace in the message while increment its version number. Backward compatibility
is addressed. The attribute design is modified accordingly. Took out Null
dataspace test from tmisc.c and put it in th5s.c. These files are re-generated
since the header message has been changed.
Platforms tested: h5committest
bug fix, new tests
Description:
added a couple of #ifdefs to check if a specific filter is available for read
added more tests
the h5tools check filters had a mismatched filter variable
Solution:
Platforms tested:
linux, with several filter configurations (all, some , none )
Misc. update:
new tests for h5repack
Description:
added more tests both to the test program and shell script that test
a variation of different filter converssions
Solution:
Platforms tested:
linux
Misc. update:
new tests for h5repack
Description:
added tests that do layout type to layout type conversion in a matrix of 9 between compact, contiguous and chunking
Solution:
Platforms tested:
linux
afs has problems; I could not telnet to sol and copper, arabica is really slow (meaning
waiting 1 minute for a typed character) and the writing of a file gave an error
arabica 181% afs: failed to store file (145)
afs: failed to store file (145)
Misc. update:
bug fix
Description:
the synntax of the input of h5repack conatined double quotes and spaces, which
were causing problems on the parsing in AIX paralell
Solution:
replaced the spaces by =
that is, instead of -f "GZIP 6"
we have now
-f GZIP=6
Platforms tested:
linux
solaris
AIX paralell
Misc. update:
Bug fix.
Description:
nerror was not updated when errors were detected.
Added the nerror update and print a message at
the end if all is well.
Platforms tested:
Tested in copper, both parallel and serial.
bug fix continuation
Description:
the atof return value on a hexadecimal input is different
on some systems; before checking for the atof return value
do a character check for the first 2 characters of the string input
Solution:
Platforms tested:
linux
solaris
Misc. update:
bug fix (sort of)
Description:
apparently linux and other systems (solaris) return a different value from atof if the
argument is in hexadecimal .
this return value was used to test if the argument to -p and -d was a valid floating point
number
for now , commented the calls in the script with -p <hexa> until we find a portable solution
Solution:
Platforms tested:
linux
solaris
windows
Misc. update:
bug fix
Description:
the -p option was not parsing correctly if the input was in the exponential format
Solution:
add a new parse test for floating point numbers
Platforms tested:
linux
Misc. update:
code clean, change test script
Description:
removed soem debugging messages
changed the chunking command line call to a non square dimensionality
Solution:
Platforms tested:
linux
Misc. update:
bug fix
new test
Description:
the fletcher filter used a temporary 2 byte word buffer to compute the checksum.
this is non portable between big-endian/little endian.
added a test that reads 2 pre-saved files (one LE, other BE) with that filter enabled
Solution:
replaced with a buffer of 1 byte type
Platforms tested:
linux
solaris
solaris 64 bit
AIX
windows
Misc. update:
new feature
Description:
added a check for the return value of the tools h5repack and h5diff
in the run script.
the return value is used to print PASSED (tool returns 0) or FAILED
Solution:
Platforms tested:
linux
Misc. update:
1) new function for tools library
2) new test script for h5repack
Description:
1) currently all the tools (h5dump, h5diff, etc) do not check if a filter is available
for reading some dataset that might have a filter not available on the current configuration (the behaviour
of the tools until now was to trigger a library error, saying that the dataset cannot be read
due to the lack of the filter)
Solution:
1) added a new function h5tools_canreadf that checks if a dataset can be read
depending on the availability of filters.
this function was added in calls for h5diff and h5repack.
instead of triggering the library error, a message is printed, saying that the dataset
cannot be read (the print is optional, it is on on verbose mode)
2) added a shell script that tests the commannd line tool behaviour of h5repack
the script does a series of runs of h5repack with several options on the same file (this file test4.h5
was added to the testfiles dir).
then, it runs the h5diff tool, with the input and output files , in each run.
the goal of the test is also to check item 1) . the binary file was saved with filters
that might not be available on other configurations
Platforms tested:
linux (all filters enabled)
linux (some filters disabled)
solaris (some filters disabled)
AIX (some filters disabled)
windows (all filters on and off )
Misc. update:
bug fix
Description:
I introduced some #ifdef <have_filter> on the check in yesterday on the places
that generated datasets, that had no corresponding on the read part.
this caused failures in the configurations where a filter was not defined
Solution:
just removed the #ifdefs
Platforms tested:
linux
Misc. update:
bug fix
Description:
generating a dataset in linux 2.4 (verbena) with the pgcc compiler gave an error with the following
sequence
1) make a space id1 and a dcpl
2) make a space id2, create and write a dataset with space id2, close
3) attempt to create a dataset with space id1 fails and calling a set_filter function with the dcpl in 1) fails
Solution:
just changed the order 2) to 1)
Platforms tested:
linux 2.4 (verbena) with pgcc compiler
Misc. update:
bug fix in H5Zshuffle.c
add more tests to h5repack that exposed the bug
Description:
when creating a dataset with the shuffle filter and duplicating it in a new dataset (file)
the call to H5Z_set_local_shuffle failed. this is because the value of cd_nelmts of the filter
structure is set to 1 (H5Z_SHUFFLE_TOTAL_NPARMS) when the original dataset is created, but when
the new dataset is created there is a checking instruction that fails if the value of
cd_nelmts is not 0 (its original value, H5Z_SHUFFLE_USER_NPARMS)
Solution:
just remove that check condition, since the value of cd_nelmts is not used anyway.
if we decide that the value of cd_nelmts is necessary, then the H5O_pline_copy function
must be changed to update this value (a different update for each filter)
Platforms tested:
linux
solaris
AIX
Misc. update:
small bug fix
Description:
a dataset creation property list was not closed
Solution:
closed it, calling H5Pclose(dcpl);
Platforms tested:
linux (small fix)
Misc. update:
new feature, bug fix, changed function
Description:
1) implemented the option that says if the dataset is too small , do not compress it
2) bug fix in the SZIP checking . only apply szip to atomic datatypes
3) made the apply_filters function more compact
Solution:
Platforms tested:
linux
AIX
solaris
Misc. update:
Bug fix
Description:
Add ifdef around most of this file, to protect it from the Intel 7.1 C
compiler which seems to have bugs in its -MG flag for creating dependency
information.
Platforms tested:
Linux IA64 (titan)
too obscure to require testing on other platforms
Bug fix
Description:
Fix h5tools routines to not try to call MPI_Init() unless an MPI-based
VFD is actually used.
Platforms tested:
FreeBSD 4.9 (sleipnir) w/parallel
Linux 2.4 (verbena) w/parallel