"And that takes us to the principle I wrote earlier on my todo list:"
#2 - Separate experiments are usually supposed to avert 'conditional-dependencies', watch out for when that isn't true
"What I mean here is that - when you are otherwise doing things correctly - it should usually be the case that, for the likelihoods that the 'published-experimental-report' is summarizing for different hypotheses, if some replicators came along and did their own experiment, their likelihoods should be something they can calculate independently of your data. It shouldn't be the case that they have to look at your data, to figure out the likelihoods given their hypotheses."
"This in fact is the property that lets us compute the joint likelihood of a hypothesis across two experiments, by multiplying the likelihoods together from the separate experiments. Symbolically:"
P( data from first and second experiments ◁ the hypothesis )
= P( data from 1st experiment ◁ the hypothesis ) * P( data from 2nd experiment ◁ the hypothesis)
if and only if
P ( data from 2nd experiment ◁ the hypothesis ) = P( data from 2nd experiment ◁ the hypothesis & data from 1st experiment)
"When you say, 'maybe all-14s have 60% propensity to solve in time, and all-10s have 10% propensity to solve in time', you're describing a way reality can be, where the likelihood of my found pattern of YES and NO responses, if that's true, is just the same no matter what data you found in your own experiment. Maybe your data made that world look incredibly improbable, but that doesn't matter; I can still answer the question of how likely my data would be, if that world was the case, without looking at your data."
"When you say, 'maybe all-14s have at least five times the propensity of all-10s to solve 2-4-6 in 30 minutes', that is a way the world can be; but it's a way the world can be, where calculating the likelihood of my data in that world, requires me to make up a bunch of prior-probabilities, and then those probabilities change depending on the data that you got."
"Which makes it immensely complicated to quickly look over the summaries of what different people's experiments said about different worlds, and combine them together into a joint summary of what reality has told us about them all."
"It would, in fact, be possible to combine a lot of little experiments all of which suggested that - if you wrote the summary this way - the data was more likely if all-10s had 50% propensity to solve, versus less-than-50% propensity to solve, and get out a new update that the combined data was more likely if all-10s had less-than-50% to solve. If you multiplied enough 0.035 likelihoods from the 40%-propensity hypothesis, compared to the 0.031 likelihoods from the 50%-propensity hypothesis, then eventually the 40%-propensity hypothesis would come to dominate the predictions of its bucket, and then its bucket would start to dominate the other hypothesis."
"Which paradoxical-seeming combination, again, doesn't happen if you consider the 40%-propensity hypothesis separately, because then it's clear from the start that 40% propensity is gaining on 50% propensity in each experiment."
"Hence again the proverb: Different effect sizes are different hypotheses. Which argues again against thinking that 'all-14s are at least five times as likely to solve as all-10s' is a good way to split up the world for purposes of SCIENCE! Even though, in terms of 'truth-functional' scaffolds, it is a way the world can be. It could even be the metafact that is useful and that we're interested in. We should still ask the Science! question first, what are the exact real effectsizes, and then check the useful metafact afterwards."