In essence probability is influenced by a person's information about the possible occurrence of an event. For example, let the event A {\displaystyle A} be 'I have a new phone'; event B {\displaystyle B} be 'I have a new watch'; and event C {\displaystyle C} be 'I am happy'; and suppose that having either a new phone or a new watch increases the probability of my being happy. Let us assume that the event C {\displaystyle C} has occurred – meaning 'I am happy'. Now if another person sees my new watch, he/she will reason that my likelihood of being happy was increased by my new watch, so there is less need to attribute my happiness to a new phone.
To make the example more numerically specific, suppose that there are four possible states Ω = { s 1 , s 2 , s 3 , s 4 } , {\displaystyle \Omega =\left\{s_{1},s_{2},s_{3},s_{4}\right\},} given in the middle four columns of the following table, in which the occurrence of event A {\displaystyle A} is signified by a 1 {\displaystyle 1} in row A {\displaystyle A} and its non-occurrence is signified by a 0 , {\displaystyle 0,} and likewise for B {\displaystyle B} and C . {\displaystyle C.} That is, A = { s 2 , s 4 } , B = { s 3 , s 4 } , {\displaystyle A=\left\{s_{2},s_{4}\right\},B=\left\{s_{3},s_{4}\right\},} and C = { s 2 , s 3 , s 4 } . {\displaystyle C=\left\{s_{2},s_{3},s_{4}\right\}.} The probability of s i {\displaystyle s_{i}} is 1 / 4 {\displaystyle 1/4} for every i . {\displaystyle i.}
and so
In this example, C {\displaystyle C} occurs if and only if at least one of A , B {\displaystyle A,B} occurs. Unconditionally (that is, without reference to C {\displaystyle C} ), A {\displaystyle A} and B {\displaystyle B} are independent of each other because P ( A ) {\displaystyle \operatorname {P} (A)} —the sum of the probabilities associated with a 1 {\displaystyle 1} in row A {\displaystyle A} —is 1 2 , {\displaystyle {\tfrac {1}{2}},} while P ( A ∣ B ) = P ( A and B ) / P ( B ) = 1 / 4 1 / 2 = 1 2 = P ( A ) . {\displaystyle \operatorname {P} (A\mid B)=\operatorname {P} (A{\text{ and }}B)/\operatorname {P} (B)={\tfrac {1/4}{1/2}}={\tfrac {1}{2}}=\operatorname {P} (A).} But conditional on C {\displaystyle C} having occurred (the last three columns in the table), we have P ( A ∣ C ) = P ( A and C ) / P ( C ) = 1 / 2 3 / 4 = 2 3 {\displaystyle \operatorname {P} (A\mid C)=\operatorname {P} (A{\text{ and }}C)/\operatorname {P} (C)={\tfrac {1/2}{3/4}}={\tfrac {2}{3}}} while P ( A ∣ C and B ) = P ( A and C and B ) / P ( C and B ) = 1 / 4 1 / 2 = 1 2 < P ( A ∣ C ) . {\displaystyle \operatorname {P} (A\mid C{\text{ and }}B)=\operatorname {P} (A{\text{ and }}C{\text{ and }}B)/\operatorname {P} (C{\text{ and }}B)={\tfrac {1/4}{1/2}}={\tfrac {1}{2}}<\operatorname {P} (A\mid C).} Since in the presence of C {\displaystyle C} the probability of A {\displaystyle A} is affected by the presence or absence of B , A {\displaystyle B,A} and B {\displaystyle B} are mutually dependent conditional on C . {\displaystyle C.}
Introduction to Artificial Intelligence by Sebastian Thrun and Peter Norvig, 2011 "Unit 3: Conditional Dependence"[permanent dead link] https://www.ai-class.com/course/video/videolecture/33 ↩
Introduction to learning Bayesian Networks from Data by Dirk Husmeier [1][permanent dead link] "Introduction to Learning Bayesian Networks from Data -Dirk Husmeier" http://www.bioss.sari.ac.uk/staff/dirk/papers/sbb_bnets.pdf ↩
Conditional Independence in Statistical theory "Conditional Independence in Statistical Theory", A. P. Dawid" Archived 2013-12-27 at the Wayback Machine http://edlab-www.cs.umass.edu/cs589/2010-lectures/conditional%20independence%20in%20statistical%20theory.pdf ↩
Probabilistic independence on Britannica "Probability->Applications of conditional probability->independence (equation 7) " http://www.britannica.com/EBchecked/topic/477530/probability-theory/32768/Applications-of-conditional-probability#toc32769 ↩
Introduction to Artificial Intelligence by Sebastian Thrun and Peter Norvig, 2011 "Unit 3: Explaining Away"[permanent dead link] https://www.ai-class.com/course/video/quizquestion/60 ↩
Bouckaert, Remco R. (1994). "11. Conditional dependence in probabilistic networks". In Cheeseman, P.; Oldford, R. W. (eds.). Selecting Models from Data, Artificial Intelligence and Statistics IV. Lecture Notes in Statistics. Vol. 89. Springer-Verlag. pp. 101–111, especially 104. ISBN 978-0-387-94281-0. 978-0-387-94281-0 ↩
Conditional Independence in Statistical theory "Conditional Independence in Statistical Theory", A. P. Dawid Archived 2013-12-27 at the Wayback Machine http://edlab-www.cs.umass.edu/cs589/2010-lectures/conditional%20independence%20in%20statistical%20theory.pdf ↩