We have already scoured the forum​ but we could not find something similar…
We are dealing with 454 pyrosequencing analysis of 18S rRNA genes and investigate the diversity of eukaryotic protists (mainly phytpolankton) in the Arctic Ocean using the V4 region (~600bp fragments).
To denoise different samples (size of ~30 000 - 80 000 sequences), we tried to run shhh.flows in mothur (Version 1.22.2) with 8 processors for each sample. However, the denoising took more than 1 month! So, we canceld this. The same holds true for mothur v 1.29.2. And we also tried to use only one processor.
As an alternative way, we used the large option to split the dataset (large=1000). This job was done in some minutes. However, it would be great if we could find a way to denoise the whole dataset. Is there any way to denoise large 18S datasets without splitting?
Thanks for this forum!
Kristin