You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
So 12 minutes versus 1 seconds for this example. Yes, I know that I can bypass DiskArrays by providing my own batchgetindex. But I don't think that DiskArray should make the current assumptions if it aims to be generic (if this is the goal).
Can we have the current assumptions (i.e. reading a chunk similarly fast than reading a subset of chunk) documented?
Concerning the API, another think I am wondering if we need a function named batchgetindex at all. Why not having getindex simply pass the indices to DiskArrays.readblock! and let DiskArrays.readblock! figure out how to best load the data. DiskArrays.readblock! is already specific to the storage format.
The text was updated successfully, but these errors were encountered:
Indexing with a vector can be incredibly slow with the current implementation of
batchgetindex
.From this comment by @Alexander-Barth
#131 (comment)
NetCDF.jl + DiskArray
NCDatasets 0.12 without DiskArray
The text was updated successfully, but these errors were encountered: