The recent post about the Verducci Effect and Let's Make A Deal didn't elicit the response I was looking for. Reactions ranged from denial to sadness to even anger. I think the game show story hurt more than helped my case of why I think the Verducci Effect is an illusion. And as I said in the original article, I'm not completely certain. But now, after developing my thoughts a little better, I'm more certain than before.
Game shows aside, I'll explain my thought process with an notional example, a mental exercise actually. I'm most interested in the injury aspect of the Verducci Effect, so I'll concentrate on that in this post. The injury rates I'm going to use are created only for clarity, and they are not intended to match the true rates. I'm also going to make some simplifying assumptions to illustrate the broader point. Please keep in mind this example is only intended to demonstrate a concept.
The Example
Assume every MLB pitcher's career lasts exactly 5 years. Also assume that the league-wide injury rate is 1 out of 5 years, defined as however you like--say being on the DL. Also assume that one year out of each pitcher's career can be identified, after the fact, as a "career year," which by definition assures us of two things: no significant injury and an upswing of innings.
Take 200 pitchers and randomly assign a number from 1 through 5 to each of their 5 respective years completely independently. When a 1 comes up, call that an injury year. So far, we've got a 1 in 5 (20%) injury rate across the league. Some pitchers will have multiple injury years, some won't have any, and they are completely independent.
In my head I'm thinking of a table of cards, 200 x 5. For now, the cards are turned up so we can see the numbers 1 through 5. 20% of the cards are 1s--injuries.
Now, take all the "career years" for the pitchers off the table by removing one card from each row, which by definition cannot be an injury year. Let's choose the highest card and remove it. What percentage of cards will now be 1s (injuries)? Before, there were 200 out of 1000 (20%), and now there are the same number of injuries (200) but fewer cards remaining (800). 25% of the remaining years are 1s (injuries). If we now selected a card at random, we'd have a 1 in 4 chance at finding an injury.
Shrink the Sample
Let's do the same exercise but with a sub-sample of 50 pitchers instead of 200. There are still 20% injury years, and if we take each pitcher's known "career year" off the table, 25% of the remaining years will be injury years. Turn over the remaining cards, so you can't see the numbers 1-5. Turn one card face up at random--what are the chances of finding a 1-card (an injury)? It has to be 25%. We started with 250 cards, but there are now 200 cards remaining and 50 of them are injury cards.
Shrink It Again
Repeat the exercise with 10 pitchers. Does this change anything? No. Originally 20% of the cards were injuries, and after removing the career year cards, 25% of those remaining are injuries.
Down to One Player
Now consider a sub-sample of just a single pitcher. Again, turn over all the cards so we can't see the numbers 1-5. There are originally 5 cards on the table, with a probability of 1 in 5 being an injury card. Take away one card, which we know after the fact is not an injury card, leaving 4 cards. Turn one card over--the card dealt immediately following the "career" year. What is the chance it's an injury?
It would be 25%. We had a league-wide injury rate of 20%, but following career/high-inning years we would retrospectively observe a rate of 25%. Even though the injuries were distributed completely at random and completely independently, we'd see a false connection between high-inning years and injuries in subsequent years.
Even a small difference would appear statistically significant with a large data set, but it would be an illusion. The original probability of a year being an injury year was always 20%, but after looking back and removing a year in which we're virtually assured of no injury, we'd see a 25% injury rate.
Try It Yourself
If you don't believe me, you can play the game yourself. Shuffle a deck of cards and deal out 4 face up in row. Those 4 cards represent 4 years of a pitcher's career. Every time we see a diamond card, we'll call that an injury year. We'll say the highest non-diamond card is a career/high-inning year. It goes without saying that, on average, 1 in every 4 cards will be a diamond--an injury. That's our true baseline rate.
After dealing the 4 cards, remove the highest non-diamond card and set it aside. Look at the card immediately to the right of the one you removed. What is the probability it is a diamond? If you said 1 in 4 you'd be mistaken. It's 1 in 3. This is the same illusion.
Try it. I did, 78 times and got a diamond on 26 tries--exactly one third (p=0.04 for the sticklers out there). You have to re-shuffle each time for it to be completely random and independent. Also, if the high "career" card is the right-most card, you can either throw out that iteration or loop around to look at the first card. The effect is the same. In fact, just look at how many of the 3 remaining cards are diamonds, and you'll eventually see that it's 1 in 3.
I'm sure there is a name for this, but I don't know it. If anyone is familiar, fill me in. Otherwise, I'm sticking with the Monty Hall Effect. Also, the Verducci Effect may still be real, but it would have to be shown that the observed injury rates significantly exceed the rate predicted by the effect of the illusion.
Maybe I'm wrong, and that's ok. But every time I deal 4 cards and remove 1 non-diamond, I keep seeing diamonds 33% of the time. Just like the Monty Hall game, you probably won't believe it until you try it yourself.
[Edit: I am now completely certain the paradox/illusion exists as I described. However, after a good discussion with commenter Vince (see below), I'm no longer convinced that the way I set up the question applies to how the Verducci Effect is truly applied. In other words, the illusion is real if you set it up the way I did. It's just that strange, and subtle differences exist in how you look at the problem. For example, if in the pitcher example, you removed the first instance of a non-1, you would see a 1 in the following block 20% of the time, just like you'd expect. But if you select the highest number in the row, or if you remove the final non-1, you'd detect a 1 next 25% of the time.]