Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

FRASER maxing out memory on sge cluster #29

Open
Jessen-Erik opened this issue Sep 24, 2021 · 3 comments
Open

FRASER maxing out memory on sge cluster #29

Jessen-Erik opened this issue Sep 24, 2021 · 3 comments

Comments

@Jessen-Erik
Copy link

Hello!

We are attempting to use FRASER on a cohort of 400 samples. We've been experiencing issues completing FRASER when sending the job to our sge queue, even when providing 1.5TB of memory. It seems the parallelization of the PSI calculation (fds <- calculatePSIValues(fds,BPPARAM=BPPARAM)) is causing the job to go over our h_vmem allocation. We've attempted to force FRASER to run in serial (BPPARAM=SerialParam()), but encounter the same issue maxing out of the memory.

Is it possible that FRASER is ignoring the serial setting?

Is there a quick fix for this? Or does a solution similar to the link below have to be implemented:
gagneurlab/OUTRIDER#11

Thank you for the tool, we've really enjoyed running it on some previous cohorts.

@ischeller
Copy link
Collaborator

Hi @Jessen-Erik ,
thanks for trying out FRASER!
Regarding your problem, can you check the dimension of your fds object that you use as input for this step? As we typically run this step before filtering, I suspect that you could have a quite large fds object and that this is causing the problem rather than the parallelization itself. If this is indeed the case, you could try applying the minExpressionInOneSample filter before the PSI calculation step (we provide the option to do this as part of the countRNAData function), as this typically reduces the number of junctions inside the fds object a lot.

@Jessen-Erik
Copy link
Author

Jessen-Erik commented Sep 30, 2021 via email

@c-mertes
Copy link
Collaborator

We are sorry that we did not reply anymore. Thanks for sharing the dimensions. 3mio junctions is big, but should not require 1.5Tb memory. For this purpose 1 read is enough as it is used only to remove random alignments.

Since there was no further response, I assume that the minExpressionInOneSample filter step helped here.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants