-
Notifications
You must be signed in to change notification settings - Fork 25
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Acceptable data to use as DNA concentrations and what to do with non-quantifiable samples #155
Comments
I have not every tested with this type of DNA quantitation information. If they become "dubious" over a significant range of the DNA concentrations in your study, then I would be hesitant to use them.
Yes. One idea that might be nice is to use both types of DNA concentration measurements and then compare results to see if they at least roughly line up.
I would recommend excluding the controls when performing the frequency identification method. This actually happens automatically when you use the "combined" method, which does both frequency (without controls) and prevalence (using the controls) and then creates a combined decontam score from the two methods. |
Hi @benjjneb Thanks for your feedback! |
My last question is answered in #113 and #38 Indeed the frequency method "is effective all on its own", and in the "combined" method it separates out the negative controls, to do the contaminant labeling only using true samples. So then my question is: why ever use the "minimum", "either" or "both" methods, as it seems that the "combined" method is the most complete approach ? |
Those other variants are just different ways of using the scores from the prevalence + frequency methods to classify contaminants. One might use these variants depending on study goals (e.g. is it of utmost importance to rule out all possible contaminants? or instead to only rule out obvious contaminants?). |
Ok, so here I have some P score histograms using 1. different DNA concentration estimates for the frequency method (qubit of extracts and qiaxcel of PCR products), 2. the prevalence method and 3. the combined method Frequency method: qubit of eDNA extractsscore histogram looks strange - maybe the DNA concentrations are not appropriately describing input DNA - see #125 Frequency method: qiaxcel of PCR productswhich score to choose? prevalence methodcombined methodand the prevalence plot using the combined method and a threshold of 0.3: Based on your feedback and issue 125, I would not use the qubit readings of DNA extracts, but rather the qiaxcel readings of amplicons. I would stick to the combined method, but then, which threshold to choose? |
Based on these diganostic outputs, I would not use decontam to remove contaminants at all. What kind of samples are you measuring? The output of "combined" could be consistent with there being little-to-no contamination in your samples. |
And the "qubit of DNA extracts" measure is producing output that looks like one would get if a random number was assigned to the DNA concentration. I almost wonder if those measurements are being merged correctly with the taxonomic profiles. |
The samples are eDNA extracts of sediment samples, processed in an ancient DNA lab, so with expected very low contamination. It makes sense what you say about the "qubit of DNA extracts" plot, it seems that the QuBit measure of the initial eDNA extracts does not appropriately describe PCR input DNA concentration... What do you mean by "I almost wonder if those measurements are being merged correctly with the taxonomic profiles." |
Hi @benjjneb
Thanks a lot for your packages and the very detailed tutorials!
I am reaching out because I am unclear about what to do relating to the "conc" parameter in decontam.
It seems that you designed the package to use Qubit-like quantification values of the PCR products before pooling them equimolarly.
Now,
Thanks for your feedback!
The text was updated successfully, but these errors were encountered: