-
Hello, I have a very large 2d numpy (cupy) array, Rt, whose dimension is 512512/32 by 240300*200/40, and now I want to achieve fast reading and writing of the array between local disk and memory. I have tried many methods, including cupy.save, cupy.savez, etc. It turns out that all the functions do not have enough compressive ratio, which result in 24GB disk space consumption. Only using cupy.savez_compressed and
can compress Rt down to hundreds of MBs. But reading and writing Rt from and to local drive costs 60 seconds, which is very slow. We want to achieve 10 times speed up of reading and writing between RAM and local disk compared with pandas.to_pickle() function and maintain same degree of compressive ratio as cupy.savez_compressed. Does anyone have a solution or provide zarr code examples showing how to read/write the Rt between local drive and memory? Thanks!!!! |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment
-
Would suggest checking out KvikIO, which can enable GDS backed IO:
Also here's a blogpost using KvikIO with XArray: cc @madsbk |
Beta Was this translation helpful? Give feedback.
Would suggest checking out KvikIO, which can enable GDS backed IO:
Also here's a blogpost using KvikIO with XArray:
cc @madsbk