In a previous article I showed that the inference rules of propositional logic can be obtained from probability calculus. But actually, we can obtain much more, and even explain why most people don’t think the propositional logic way.
In this article, we will see that there is more to the traditional implication than it seems in the framework of propositional logic. Using probability calculus and bayesian inference, we will show why most people mistakenly use the A⇒B implication… and show that they are not completely mistaken after all.
Let’s take an example. Yesterday evening, my friend Bob was on his way to a party. He told me this: “If I can kiss Alice during the party, I will go to the cinema with her tomorrow evening”. I haven’t met Bob since the party, but a friend saw Alice and him at the cinema’s evening projection today.
Did you assume that Bob managed to get a kiss from Alice during the party? Logicians know this shortcut too well. In propositional logic, nothing tells us that Bob kissed Alice. Maybe they didn’t kiss and still went to the cinema today.
But using probability calculus, we can show that our probability estimate for their kiss increased when we learned they went to the cinema. This is what I will prove now.
Without further ado, let’s dive in.
A and B are propositions and I use the convention pe(⋅)=p(⋅∣e). If you need a cheatsheet about probability calculus or the notations I use, check this out.
If we let e = “A⇒B”, then pe(B∣A)=1 by definition of e. I already showed in a previous article that e is enough to derive the usual equivalent forms: A⇒B≡ˉB⇒ˉA≡ˉA+B using probability calculus. This article showed the following rules:
If A is true | then B is true |
If B is false | then A is false |
But actually, given this rule A⇒B, we can show that:
If B is true | then A is more likely |
If A is false | then B is less likely |
Part 1: when A is more likely
I will now show that we can rewind the arrow: given a rule such as A⇒B, we will show that the probability for A increases when we gain information about B. From the propositional logic vantage point, this is surprising because information about B doesn’t tell us anything about A. As we will see, in probability calculus, it’s a completely different story. This could explain why most people mistakenly use A⇒B as B⇒A, even though both are completely different in propositional logic.
Let e be an evidence such that pe(B)<pe(B∣A). For instance, the rule A⇒B is such evidence given that p(B)≠1. But so is the weaker rule: A⇒ more_plausible(B).
We will show that given evidence e, we also have: pe(A)<pe(A∣B) which means that evidence for B increases our belief in A.
We have: | pe(AB)=pe(A∣B)pe(B) and pe(AB)=pe(B∣A)pe(A)⇒pe(A∣B)/pe(A)=pe(B∣A)/pe(B) |
Hence: | pe(B)<pe(B∣A)⇒1<pe(B∣A)/pe(B)⇒1<pe(A∣B)/pe(A)⇒pe(A)<pe(A∣B) |
And that’s why given the rule A⇒B, the probability for A increases when we know that B is true, even though propositional logic don’t tell us anything about A.
Part 2: when B is less likely
Actually, we can even show that when A⇒B, the probability estimate for B decreases when we know that A is false!
I will now show that if pe(B)<pe(B∣A), then pe(B∣ˉA)<pe(B):
pe(B∣ˉA)=pe(ˉA∣B)pe(ˉA)⋅pe(B)=1−pe(A∣B)1−pe(A)⋅pe(B)But, we showed that pe(A)<pe(A∣B), so the fraction is less that 1. Thus the proof that:
pe(B∣ˉA)<pe(B)