As you can see, in the first OTU the mean of group 1 and 2 are 153 and 114, mean of counts are 359 and 423. In OTU2 we get means of 40 and 29 and mean of counts of 88 and 114.
I assume the significance is based on the difference in unique sequences in the analysis and the mean of counts is a report on the mean of total sequences. What can I say about these OTU’s given that the mean of counts shows an opposite trend than the mean(groups)?
So the the first - mean(group1) - is a normalized mean so that each group in the analysis has the same number of sequences. I think the second - mean_of_counts(group1) - is the raw mean without normalization. The analysis is performed on the first column.
Hi there. I finally got around to making a design file and using the metastats command in mothur. I was blown away by all of the near-instant comparisons that used to take me hours to upload. However, the results of mothur’s implementation of metastats differ from the results I get by manually uploading a pairwise comparison to the metastats website. Here are the results of a given comparison computed by the metastats server (I should note that these are phylotype data):
The means are reported differently by each program, but as you can see from the p and q values the results from mothur are much less stringent/much more noisy. Do you have any explanation for this? I am 99% sure that it is not an error in the design file because the means reported in the mothur metastats outputs are correct for a given OTU in a given comparison, indicating that mothur analyzed the samples I told her to analyze. I am inclined to “believe” the results from the metastats server based on looking at the input data (a fair bit of variance per phylotype across samples).
My questions are, why the discrepancy, and how do I fix it? I appreciate you thinking about this.
Thanks for bringing this to our attention! The C code we use in mothur was provided by the original authors (White and Pop); however, it does seem that there are differences in pvalues for low frequency OTUs when you compare the C and R code. The next version that we post will have metastats code that more faithfully corresponds to the original R code.
I’m so glad that you found the source of the problem, and that it’s not me or my data!
Thank you for addressing this problem, and as always I’m looking forward to the next release of mothur.
Heather