Excercise: Given a positive result from screening, what is the probability of having BC?
PPV - positive predictive value = posterior probability that a reported finding is true
True | False | |
---|---|---|
Accept | 1-β | α |
Reject | β | 1-α |
sum | 1 | 1 |
Note: PPV depends on α and β!
Like diagnostics, we need to consider the ratio R of true and false hypotheses that are tested:
True | False | |
---|---|---|
Accept | (1-β) R/(R+1) | α /(R+1) |
Reject | β R/(R+1) | (1-α) /(R+1) |
sum | R/R+1 | 1/R+1 |
If R is less than one (more false than true hypotheses are tested), this inflates the False/Accept, relative to True/Accept.
(Including Bonferroni correction.)
But: reducing α inflates β.
Multiply by a factor of c:
True | False | |
---|---|---|
Accept | c (1-β)R/(R+1) | c α/(R+1) |
Reject | c β R/(R+1) | c (1-α)/(R+1) |
sum | cR/(R+1) | c/(R+1) |
Moves a fraction, u, from Reject to Accept.
True | False | |
---|---|---|
Accept | (c(1-β)R +ucβ R)/(R+1) | (cα +uc(1-α))/(R+1) |
Reject | (1- u)cβ R/(R+1) | (1- u)c(1-α)/(R+1) |
sum | cR/(R+1) | c/(R+1) |
Type II: β = prob of one team missing a true result
Independent teams, prob all miss it: βn
I.e. prob none miss it: 1-βn
Type I: α = prob of rejecting a false null hypothesis.
Probability that nobody rejects a false null: (1-α)n
Ignoring bias, we substitute:
True | False | |
---|---|---|
Accept | c(1- βn)R/(R+1) | c(1- (1-α)n)/(R+1) |
Reject | c βn R/(R+1) | c (1-α)n /(R+1) |
sum | cR/(R+1) | c/(R+1) |
The smaller the studies conducted in a scientific field, the less likely the research findings are to be true.
We knew that. How many fish can you fit in a tank?
The smaller the effect sizes in a scientific field, the less likely the research findings are to be true.
Reduces power. What are the typical effect sizes we look for?
The greater the number and the lesser the selection of tested relationships in a scientific field, the less likely the research findings are to be true
How many genes affect a genotype? How many did we test? How common are the different variants?
The greater the flexibility in designs, definitions, outcomes, and analytical modes in a scientific field, the less likely the research findings are to be true
How do we properly analyse RNAseq data? GWAS? Epigenetics? Who wrote that program you use to predict salmon SNPs - a statistician?
Simulation suggests that false positive rates can jump to over 50% for a given dataset just by letting researchers try different statistical analyses until one works. (Simmons, et al, 2011)
The greater the financial and other interests and prejudices in a scientific field, the less likely the research findings are to be true.
Fiskaren: Research error cost one billion.
The hotter a scientific field (with more scientific teams involved), the less likely the research findings are to be true
Empirical evidence suggests that [publishing of dramatic results, followed by rapid refutations] is very common in molecular genetics
Larger studies
Formalized methods
(I.e. no data dredging)
Measurements of productivity:
It's a win-win situation!
(Except if you care about, you know, science)