You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
currently, in the sc_demultiplex step, the program will keep all the reads in memory and then open CSV and write to them one by one (because we cant open too many files at the same time). it will become a problem when we have a very large dataset.
to solve this. we need to write to CSV file during we process the bam file. it can be that every 100000 reads we trigger a write_to_csv thing and clear the cache.
The text was updated successfully, but these errors were encountered:
I currently have a massive dataset (6 billion reads to start with). Memory was an issue but I managed to get it running with 900Gb RAM... It completed the task, got to the next line but the output was completely empty. Any idea why?
currently, in the sc_demultiplex step, the program will keep all the reads in memory and then open CSV and write to them one by one (because we cant open too many files at the same time). it will become a problem when we have a very large dataset.
to solve this. we need to write to CSV file during we process the bam file. it can be that every 100000 reads we trigger a
write_to_csv
thing and clear the cache.The text was updated successfully, but these errors were encountered: