You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have a big cohort of scATACseq data, totaling 1 billion fragments from 24 pools of experiments (24 billion total). I've been trying to run hmmratac on these 24 fragment files with --cutoff-analysis-only, but it has been a week. The log says it downsampled to 800 million fragments for training, but the step to generate short, mono-, di-, and tri-nucleosomal signals has run for 3 days. Should I downsample the fragments files before hmmratac? There must be diminishing returns with more fragments, but I'm not sure how to decide the degree of downsampling.
Appreciate any suggestions!
The text was updated successfully, but these errors were encountered:
@taoliu
Thanks for your reply! I'm trying downsampling and see what happens. Is there any downside with using too many reads other than memory/time consumption, such as increased noise?
@leqi0001 First, it will be a waste of $ if you sample is already saturated. Secondly, it will take more time and memory to process -- especially for hmmratac. Lastly, for method calling peaks based on p-value cutoff, such as callpeak, p-value will be overly optimistic when the sample size is large. In this case, effective size (using foldchange) should be considered together with p-value cutoff.
Hi,
I have a big cohort of scATACseq data, totaling 1 billion fragments from 24 pools of experiments (24 billion total). I've been trying to run hmmratac on these 24 fragment files with --cutoff-analysis-only, but it has been a week. The log says it downsampled to 800 million fragments for training, but the step to generate short, mono-, di-, and tri-nucleosomal signals has run for 3 days. Should I downsample the fragments files before hmmratac? There must be diminishing returns with more fragments, but I'm not sure how to decide the degree of downsampling.
Appreciate any suggestions!
The text was updated successfully, but these errors were encountered: