* Fix for github issue #2414: segfault when copying dataset with attributes.
This also fixes github issue #3241: segfault when copying dataset.
Need to set the location via H5T_set_loc() of the src datatype
when copying dense attributes.
Otherwise the vlen callbacks are not set up therefore causing seg fault
when doing H5T_convert() -> H5T__conv_vlen().
* Add binary test job
* Correct names of jobs
* Add build mode to binary names
* Limit packaging execution
* Make mac DragNDrop optional
* Format if checks
* fix checks for TS
* Allow examples to be packaged
* Correct if NOT format
* Correct configure and workflow call
* Separate out CMake/Autotools Parallel job
* Correct binary examples path
* Correct dir path for windows test
* Correct paths and add extlibs
* Windows binaries do not have a share folder
* The issue was that the "allargs" variable was not being used in the final command of the compiler wrapper. Any entries containing an escaped quote (\", \') or other non-matching argument (*) would not be passed to the compile line. I have fixed this problem by ensuring all arguments passed to the compiler wrapper are now included in the compile line.
* added testing for compiler wrappers
* Switch warnings as errors to default OFF
* Enable mac docs
* Add doxygen action uses step
* Use html div around snippet
* Allow preset name to be an argument to cmake-ctest.yml
ATM configuration was embedded within the CI workflow but that makes impossible
to use codespell locally since the tool would not find those settings. I moved them
into .codespellrc but then also removed codespell.yml workflow from skips, added .git and
in general makde it look for .dotfiles too
Remove cached datatype conversion path table entries on file close
When performing datatype conversions during I/O, the library
checks to see whether it can re-use a cached datatype conversion
pathway by performing comparisons between the source and destination
datatypes of the current operation and the source and destination
datatypes associated with each cached datatype conversion pathway.
For variable-length and reference datatypes, a comparison is made
between the VOL object for the file associated with these datatypes,
which may change as a file is closed and reopened. In workflows
involving a loop that opens a file, performs I/O on an object with a
variable-length or reference datatype and then closes the file, this
can lead to constant memory usage growth as the library compares the
file VOL objects between the datatypes as different and adds a new
cached conversion pathway entry on each iteration during I/O. This is
now fixed by clearing out any cached conversion pathway entries for
variable-length or reference datatypes associated with a particular
file when that file is closed.
off_t is a 32-bit signed value on Windows, so we should use HDoff_t
(which is __int64 on Windows) internally instead.
Also defines HDftell on Windows to be _ftelli64().
* autogen.sh: check directly for libtoolize instead of libtool [as libtool is not used here]
* autogen.sh: simplify (g)libtoolize check - look for glibtoolize first - and then libtoolize
* Reorg and update options for doc and cmake config
* Correct subfiling depends
* Add linux S3 binaries to GH published files
* Correct path
* Try reducing parallel jobs in tests
* Change examples jobs for test from 8 to 4
The tools library no longer checks for whether MPI is initialized when attempting to use a parallel VFD, such as the MPI I/O VFD or Subfiling VFD. This requires that parallel VFDs check for this situation and initialize MPI on their own if need be, but also allows parallel VFDs to be used with non-parallel HDF5 tools, such as h5dump.