mirror of
https://github.com/HDFGroup/hdf5.git
synced 2025-01-30 15:32:37 +08:00
[svn-r26683] Remove internal filter checks
This commit is contained in:
parent
e9db2519b5
commit
db2e8ff371
@ -1343,21 +1343,13 @@
|
||||
set (USE_FILTER_SZIP "true")
|
||||
endif (H5_HAVE_FILTER_SZIP)
|
||||
|
||||
if (H5_HAVE_FILTER_SHUFFLE)
|
||||
set (USE_FILTER_SHUFFLE "true")
|
||||
endif (H5_HAVE_FILTER_SHUFFLE)
|
||||
set (USE_FILTER_SHUFFLE "true")
|
||||
|
||||
if (H5_HAVE_FILTER_FLETCHER32)
|
||||
set (USE_FILTER_FLETCHER32 "true")
|
||||
endif (H5_HAVE_FILTER_FLETCHER32)
|
||||
set (USE_FILTER_FLETCHER32 "true")
|
||||
|
||||
if (H5_HAVE_FILTER_NBIT)
|
||||
set (USE_FILTER_NBIT "true")
|
||||
endif (H5_HAVE_FILTER_NBIT)
|
||||
set (USE_FILTER_NBIT "true")
|
||||
|
||||
if (H5_HAVE_FILTER_SCALEOFFSET)
|
||||
set (USE_FILTER_SCALEOFFSET "true")
|
||||
endif (H5_HAVE_FILTER_SCALEOFFSET)
|
||||
set (USE_FILTER_SCALEOFFSET "true")
|
||||
|
||||
if (USE_FILTER_DEFLATE AND USE_FILTER_SHUFFLE AND USE_FILTER_FLETCHER32 AND USE_FILTER_NBIT AND USE_FILTER_SCALEOFFSET)
|
||||
# data read internal filters
|
||||
|
@ -649,21 +649,13 @@
|
||||
set (USE_FILTER_SZIP "true")
|
||||
endif (H5_HAVE_FILTER_SZIP)
|
||||
|
||||
if (H5_HAVE_FILTER_SHUFFLE)
|
||||
set (USE_FILTER_SHUFFLE "true")
|
||||
endif (H5_HAVE_FILTER_SHUFFLE)
|
||||
set (USE_FILTER_SHUFFLE "true")
|
||||
|
||||
if (H5_HAVE_FILTER_FLETCHER32)
|
||||
set (USE_FILTER_FLETCHER32 "true")
|
||||
endif (H5_HAVE_FILTER_FLETCHER32)
|
||||
set (USE_FILTER_FLETCHER32 "true")
|
||||
|
||||
if (H5_HAVE_FILTER_NBIT)
|
||||
set (USE_FILTER_NBIT "true")
|
||||
endif (H5_HAVE_FILTER_NBIT)
|
||||
set (USE_FILTER_NBIT "true")
|
||||
|
||||
if (H5_HAVE_FILTER_SCALEOFFSET)
|
||||
set (USE_FILTER_SCALEOFFSET "true")
|
||||
endif (H5_HAVE_FILTER_SCALEOFFSET)
|
||||
set (USE_FILTER_SCALEOFFSET "true")
|
||||
|
||||
# copy files (these files have no filters)
|
||||
ADD_H5_TEST (fill "TEST" ${FILE0})
|
||||
@ -943,14 +935,14 @@
|
||||
# Use first dset to test.
|
||||
#---------------------------------------------------------------------------
|
||||
# chunk to chunk - specify chunk dim bigger than any current dim
|
||||
ADD_H5_VERIFY_TEST (chunk2chunk "TEST" 0 h5repack_layout3.h5 chunk_unlimit1 CHUNK -l chunk_unlimit1:CHUNK=100x300)
|
||||
ADD_H5_VERIFY_TEST (chunk2chunk "TEST" 0 h5repack_layout3.h5 chunk_unlimit1 CHUNK -l chunk_unlimit1:CHUNK=100x300)
|
||||
|
||||
# chunk to contiguous
|
||||
ADD_H5_VERIFY_TEST (chunk2conti "TEST" 0 h5repack_layout3.h5 chunk_unlimit1 CONTI -l chunk_unlimit1:CONTI)
|
||||
ADD_H5_VERIFY_TEST (chunk2conti "TEST" 0 h5repack_layout3.h5 chunk_unlimit1 CONTI -l chunk_unlimit1:CONTI)
|
||||
|
||||
# chunk to compact - convert big dataset (should be > 64k) for this purpose,
|
||||
# should remain as original layout (chunk)
|
||||
ADD_H5_VERIFY_TEST (chunk2compa "TEST" 0 h5repack_layout3.h5 chunk_unlimit1 CHUNK -l chunk_unlimit1:COMPA)
|
||||
ADD_H5_VERIFY_TEST (chunk2compa "TEST" 0 h5repack_layout3.h5 chunk_unlimit1 CHUNK -l chunk_unlimit1:COMPA)
|
||||
|
||||
#--------------------------------------------------------------------------
|
||||
# Test -f for some specific cases. Chunked dataset with unlimited max dims.
|
||||
@ -960,16 +952,16 @@ ADD_H5_VERIFY_TEST (chunk2compa "TEST" 0 h5repack_layout3.h5 chunk_unlimit1 CHUN
|
||||
# - should not change max dims from unlimit
|
||||
|
||||
# chunk dim is bigger than dataset dim. ( dset size < 64k )
|
||||
ADD_H5_VERIFY_TEST (error1 "TEST" 0 h5repack_layout3.h5 chunk_unlimit1 H5S_UNLIMITED -f chunk_unlimit1:NONE)
|
||||
ADD_H5_VERIFY_TEST (error1 "TEST" 0 h5repack_layout3.h5 chunk_unlimit1 H5S_UNLIMITED -f chunk_unlimit1:NONE)
|
||||
|
||||
# chunk dim is bigger than dataset dim. ( dset size > 64k )
|
||||
ADD_H5_VERIFY_TEST (error2 "TEST" 0 h5repack_layout3.h5 chunk_unlimit2 H5S_UNLIMITED -f chunk_unlimit2:NONE)
|
||||
ADD_H5_VERIFY_TEST (error2 "TEST" 0 h5repack_layout3.h5 chunk_unlimit2 H5S_UNLIMITED -f chunk_unlimit2:NONE)
|
||||
|
||||
# chunk dims are smaller than dataset dims. ( dset size < 64k )
|
||||
ADD_H5_VERIFY_TEST (error3 "TEST" 0 h5repack_layout3.h5 chunk_unlimit3 H5S_UNLIMITED -f chunk_unlimit3:NONE)
|
||||
ADD_H5_VERIFY_TEST (error3 "TEST" 0 h5repack_layout3.h5 chunk_unlimit3 H5S_UNLIMITED -f chunk_unlimit3:NONE)
|
||||
|
||||
# file input - should not fail
|
||||
ADD_H5_TEST (error4 "TEST" h5repack_layout3.h5 -f NONE)
|
||||
ADD_H5_TEST (error4 "TEST" h5repack_layout3.h5 -f NONE)
|
||||
|
||||
#--------------------------------------------------------------------------
|
||||
# Test base: Convert CHUNK to CONTI for a chunked dataset with small dataset
|
||||
@ -977,9 +969,9 @@ ADD_H5_TEST (error4 "TEST" h5repack_layout3.h5 -f NONE)
|
||||
# (HDFFV-8214)
|
||||
#--------------------------------------------------------------------------
|
||||
# chunk dim is bigger than dataset dim. should succeed.
|
||||
ADD_H5_VERIFY_TEST (ckdim_biger "TEST" 0 h5repack_layout3.h5 chunk_unlimit2 CONTI -l chunk_unlimit2:CONTI)
|
||||
ADD_H5_VERIFY_TEST (ckdim_biger "TEST" 0 h5repack_layout3.h5 chunk_unlimit2 CONTI -l chunk_unlimit2:CONTI)
|
||||
# chunk dim is smaller than dataset dim. should succeed.
|
||||
ADD_H5_VERIFY_TEST (ckdim_smaller "TEST" 0 h5repack_layout3.h5 chunk_unlimit3 CONTI -l chunk_unlimit3:CONTI)
|
||||
ADD_H5_VERIFY_TEST (ckdim_smaller "TEST" 0 h5repack_layout3.h5 chunk_unlimit3 CONTI -l chunk_unlimit3:CONTI)
|
||||
|
||||
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user