[ERROR]: std::bad_alloc has occurred in the ShhherCommand class function getOTUD
ata. Please contact Pat Schloss at mothur.bugs@gmail.com, and be sure to include
the mothur.logFile with your inquiry.
The error you are getting indicates you are running out of memory. The shhh.flows command is very memory intensive. Here are 2 options that may help reduce the memory needed.
Run trim.flows with an oligos file before you run shhh.flows. When you run trim.flows with an oligos file, mothur will separate your flows by sample which allows you to run smaller sets through shhh.flows. Trim.flows will create a .flow.files file that can be used with shhh.flows file parameter to run each sample separately and combine the results into one .shhh.fasta file for you when complete.
Use 1 processor. The more processors you use the more memory is needed.
I used 1 processor and I got the same error. And when I set the oligo file in trim.flow command almost all the flows ended up in scrap file. I checked the oligo file, and the samples have a reasonable number of reads when I run trim.seqs command with oligo file set. Maybe I should mention that in my oligo file primer+tag were used in combination as barcode and no primer was defined.
Yes, barcode and length are the reason it failed. If you want to send your logfile, flow and oligos files to mothur.bugs@gmail.com, I can try to troubleshoot it for you.
If you translate this to DNA sequence, the beginning is:
GACTACGTACACACTACTACTATGTTCTGGAC [The initial GACT is the test sequence]
This isn’t close to anything in your oligos file. A bigger problem for you is that your sequences are very short and noisy. I’d figure out what’s wrong with your sequencer and go from there…
Hi again,
Thank you for your help. I figured out the problem: I had to include the linker and sequencing primer in the oligos file. As you mentioned the general quality of my reads is low. I decided to trim the flows with the setting of noise=0.5, and I got ~240,000 reads (out of ~400,000) in the output file (noise=0.7 gave only ~100,000); is this a wise thing to do?
Also I would like to get a consensus sequence for each OTU, but the function “Consensus.seqs” requires aligned sequences. The problem is that my reads are fungal ITS sequences which as you know are not alienable across the kingdom. Do you have any idea how to solve this?
Thank you for your help. I figured out the problem: I had to include the linker and sequencing primer in the oligos file. As you mentioned the general quality of my reads is low. I decided to trim the flows with the setting of noise=0.5, and I got ~240,000 reads (out of ~400,000) in the output file (noise=0.7 gave only ~100,000); is this a wise thing to do?
Um, probably not - but do you have mock community data? Without mock community data showing no bad effects, I wouldn’t trust it. Also, the lookup files were originally developed for 454, not IonTorrent - so that might be another hickup in the works. We’re working on creating a workup file but it might be a few weeks.