Recientemente aprendí sobre un principio de razonamiento probabilístico llamado " explicar lejos ", y estoy tratando de captar una intuición para ello.
Déjame configurar un escenario. Sea el evento de que ocurra un terremoto. Que el evento sea el evento de que el gigante verde alegre pasee por la ciudad. Sea el evento de que el suelo tiembla. Deje . Como se ve, ya sea o pueden causar .
Uso el razonamiento "explicar", si ocurre , uno de o aumenta, pero el otro disminuye ya que no necesito razones alternativas para explicar por qué ocurrió Sin embargo, mi intuición me dice que la corriente que tanto y debería aumentar si se produce desde que ocurre hace que sea más probable que alguna de las causas de se produjo.
How do I reconcile my current intuition with the idea of explaining away? How do I use explaining away to justify that and are conditionally dependent on ?
fuente
Respuestas:
Clarification and notation
This isn't correct. You have (implicitly and reasonably) assumed that A is (marginally) independent of B and also that A and B are the only causes of C. This implies that A and B are indeed dependent conditional on C, their joint effect. These facts are consistent because explaining away is about P(A | C), which is not the same distribution as P(A). The conditioning bar notation is important here.
You are having the 'inference from semi-controlled demolition' (see below for details). To begin with, you already believe that C indicates that either A or B happened so you can't get any more certain that either A or B happened when you see C. But how about A and B given C? Well, this possible but less likely than either A and not B or B and not A. That is the 'explaining away' and what you want the intuition for.
Intuition
Let's move to a continuous model so we can visualise things more easily and think about correlation as a particular form of non-independence. Assume that reading scores (A) and math scores (B) are independently distributed in the general population. Now assume that a school will admit (C) a student with a combined reading and math score over some threshold. (It doesn't matter what that threshold is as long as it's at least a bit selective).
Here's a concrete example: Assume independent unit normally distributed reading and math scores and a sample of students, summarised below. When a student's reading and math score are together over the admission threshold (here 1.5) the student is shown as a red dot.
Because good math scores offset bad reading scores and vice versa, the population of admitted students will be such that reading and math are now dependent and negatively correlated (-0.65 here). This is also true in the non-admitted population (-0.19 here).
So, when you meet an randomly chosen student and you hear about her high math score then you should expect her to have gotten a lower reading score - the math score 'explains away' her admission. Of course she could also have a high reading score -- this certainly happens in the plot -- but it's less likely. And none of this affects our earlier assumption of no correlation, negative or positive, between math and reading scores in the general population.
Intuition check
Moving back to a discrete example closer to your original. Consider the best (and perhaps only) cartoon about 'explaining away'.
The government plot is A, the terrorist plot is B, and treat the general destruction as C, ignoring the fact there are two towers. If it is clear why the audience are being quite rational when they doubt the speaker's theory, then you understand 'explaining away'.
fuente
I think your intuition is ok but your understanding of "explain away" reasoning is wrong.
In the article you linked to
(emphasis added)
This is quite different from your:
You don't just needC to occur it also needs to have been explained away by confirmation of A or B before you reduce the probability of the other possible explanation
Think of it another way. The ground is shaking. You observeB , the giant is wandering around. This explains away C , so it seems unlikely that there is now an earthquake - you settle for the giant explanation. But observing the giant was key - until you had this as the likely explanation of the earthquake, nothing had been explained away. When all you had was C , in fact both P(A|C) and P(B|C) are > P(A) and P(B) respectively, as per @Glen_b's answer.
fuente
In the absence of specific additional information that changes the conditional probability ofA or B , Bayes rule tells you
IfP(C|A)P(C) and P(C|B)P(C) are both bigger than 1 (which you'd expect if the word 'explanation' is really to mean anything), then both A and B will be more conditionally more probable than they were before C was observed.
It will be of interest to see if one becomes relatively more likely after observingC compared to before.
That is, the relative probability of the two after observingC is the relative probability before (P(A)/P(B) ) times the ratio of the conditional probabilities of observing C given the two 'explanations'.
fuente
You're asking for intuition. What does it mean thatA and B are independent? It means that if I tell you that I've just seen the monster, your opinion about the occurrence or not of the earthquake doesn't change; and conversely. If you think that both P(C∣A) and P(C∣B) are high, and I tell you that the ground is shaking and there is no monster in the town, wouldn't that change your opinion about the occurrence of the earthquake, making it more probable?
fuente
From the linked abstract, it appears that "explaining away" is discussing a learning mechanism, a common way that humans reason, not a formal method of logic or probability. It's a human-like way of reasoning that's not formally correct, just as inductive reasoning is not formally correct (as opposed to deductive reasoning). So I think the formal logic and probability answers are very good, but not applicable. (Note that the abstract is in a Machine Intelligence context.)
Your giants example is very good for this. We believe that earthquakes or giants can cause the ground to shake. But we also believe that giants do not exist -- or are extremely unlikely to exist. The ground shakes. We will not investigate whether a giant is walking around, but rather we'll inquire as to whether an earthquake happened. Hearing that an earthquake did in fact happen, we are even more convinced that earthquakes are an adequate explanation of shaking ground and that giants are even more certain not to exist or are at least even more highly unlikely to exist.
We would only accept that a giant caused the ground to shake only if: 1) we actually witnessed the giant and were willing to believe that we were not being fooled and that our previous assumption that giants were highly-unlikely or impossible was wrong, or 2) we could totally eliminate the possibility of an earthquake and also eliminate all possibilities D, E, F, G, ... that we previously had not thought of but that now seem more likely than a giant.
In the giant case, it makes sense. This learning mechanism (an explanation we find likely becomes even more likely and causes other explanations to become less likely, each time that explanation works) is reasonable in general, but will burn us, too. For example, the ideas that the earth orbits the sun, or that ulcers are caused by bacteria had a hard time gaining traction because of "explaining away", which in this case we'd call confirmation bias.
The fact that the abstract is in a Machine Intelligence setting also makes me thing this is discussing a learning mechanism commonly used by humans (and other animals, I imagine) that could benefit learning systems even though it can also be highly flawed. The AI community tried formal systems for years without getting closer to human-like intelligence and I believe that pragmatics has won out over formalism and "explaining away" is something that we do and thus that AI needs to do.
fuente
I think an easier way to think of it is: If there is any variableC (0<P(C)<1) such that the occurrence of C increases the probability of both A and B , then A and B cannot be independent. In your example, you actually chose variables that you intuitively understand to be dependent, not independent. That is, the event that there is an earthquake and a giant stomping around aren't independent, since they both are more likely to occur when the floor is shaking. Here is another example: Let C be the event that it rains, and A be the event that you use an umbrella, and B the event that you wear rainboots. Clearly A and B are not independent because when C occurs, you are more likely to both wear galoshes and carry and umbrella. But if you lived in an area that never, ever rained, then A and B could potentially be independent--neither the umbrella nor galoshes are being used as rain gear, so perhaps you wear the galoshes in the garden and use the umbrella to catch fish. They are only able to be independent because they don't share a cause.
Here is a proof: SupposeA and B are independent and also conditionally independent given C .
It follows from 1 and 2 thatP(C)=P(C)2 hence P(C)=0 or P(C)=1 .
fuente