This question is somehow related to "Concatenating a large number of HDF5 files . "
I have some huge HDF5 files (~ 20 GB compression) that cannot fit in RAM. Each of them stores several pandas.DataFrame same format and indexes that do not overlap.
I would like to merge them into a single HDF5 file with all DataFrames data that would be merged correctly. One way to do this is to read each of them chunk-by-chunk and then save it in a single file, but actually it will take quite a lot of time.
Are there any special tools or methods for doing this without iterating through files?
python pandas hdf5 pytables
Vladimir
source share