Optimal window size to reduce memory usage #132
-
Hi, I am analysing a data set of 18+17 sample size and ~1.5 GB genome. I used block length 145bp because the genetic diversity of my species is lower. Then I want to use a window size of 500 as in your original paper, so run gimble windows with -w 150000 -s 30000. it turns out giving error message because of out of memory usage (there is 512 GB of memory in total!). Shall I increase the window size or decrease it in order to reduce memory usage? Alternatively, is there a way to reduce memory usage while keeping the desired sample size? Best, |
Beta Was this translation helpful? Give feedback.
Replies: 3 comments 1 reply
-
The error message is: |
Beta Was this translation helpful? Give feedback.
-
@DRL . I have no intuition here on whether there is a window size setting that will increase or reduce the memory usage. I feel like memory usage will rather depend on the number of blocks in total. What I do see when looking at the code is the fact that we concatenate all blocks (here) with a window index effectively producing a copy of an existing array. We could easily process the array of blocks using for loops (without |
Beta Was this translation helpful? Give feedback.
-
@XieHongX There is no quick fix to this however. Cleverly reducing memory usage would require some work from our end. @KLohse is splitting your genomes prior to making blocks a potential issue for this? Ideally we would have function then that allows you to combine bSFS counts once counts have been generated for each of the separate genome chunks. |
Beta Was this translation helpful? Give feedback.
@XieHongX There is no quick fix to this however. Cleverly reducing memory usage would require some work from our end.
@KLohse is splitting your genomes prior to making blocks a potential issue for this? Ideally we would have function then that allows you to combine bSFS counts once counts have been generated for each of the separate genome chunks.