Understanding Conditional Probability BasicsHelp understanding conditional probabilityConditional probability...
Prove that the total distance is minimised (when travelling across the longest path)
How does Dispel Magic work against Stoneskin?
Should we release the security issues we found in our product as CVE or we can just update those on weekly release notes?
My story is written in English, but is set in my home country. What language should I use for the dialogue?
Should QA ask requirements to developers?
Latest web browser compatible with Windows 98
What is the dot in “1.2.4."
Is it illegal in Germany to take sick leave if you caused your own illness with food?
Decoding assembly instructions in a Game Boy disassembler
What is the difference between "shut" and "close"?
what does the apostrophe mean in this notation?
Why don't MCU characters ever seem to have language issues?
Replacing Windows 7 security updates with anti-virus?
Time travel short story where dinosaur doesn't taste like chicken
Time dilation for a moving electronic clock
When were linguistics departments first established
Touchscreen-controlled dentist office snowman collector game
Ban on all campaign finance?
Who is our nearest neighbor
How could a female member of a species produce eggs unto death?
Does Linux have system calls to access all the features of the file systems it supports?
US to Europe trip with Canada layover- is 52 minutes enough?
Why would a jet engine that runs at temps excess of 2000°C burn when it crashes?
Rejected in 4th interview round citing insufficient years of experience
Understanding Conditional Probability Basics
Help understanding conditional probabilityConditional probability - problem with understandingProbability and understanding solveConditional probability with continuous random variablesIs the probability addition rule commutative?Difference between conditional probability and Bayes ruleHow do I understand “time” in conditional probability and chain ruleDifference between Conditional Probability and Bayes TheoremCorrect interpretation of a conditional probability constraintProbability - conditional probability on student passing
$begingroup$
In many online sources I've read a statement similar to:
Probability of B happening given A is equal to the probability of A and B both happening divided by B happening
or
$p(A | B) = p(A cap B) / p(B)$
But what is the difference between "B happening given A" and "A and B happening at the same time" ? If B happens given A, didn't A and B happen at the same time ? Why is that chance smaller than the intersection ?
probability bayesian
$endgroup$
add a comment |
$begingroup$
In many online sources I've read a statement similar to:
Probability of B happening given A is equal to the probability of A and B both happening divided by B happening
or
$p(A | B) = p(A cap B) / p(B)$
But what is the difference between "B happening given A" and "A and B happening at the same time" ? If B happens given A, didn't A and B happen at the same time ? Why is that chance smaller than the intersection ?
probability bayesian
$endgroup$
$begingroup$
The formula is right. But the statement is not right. It has to be: "Probability of $A$ happening given $B$ is equal to the probability of A and B both happening divided by B happening."
$endgroup$
– callculus
Jul 4 '15 at 6:42
1
$begingroup$
$P(A|B)$ isn´t smaller than $P(Acap B)$. $P(B)cdot P(A|B)=P(A cap B)$ and $P(B) leq 1$. Thus $P(A|B)geq P(A cap B)$
$endgroup$
– callculus
Jul 4 '15 at 7:02
$begingroup$
@calculus I am aware that the probability of B given some observations is going to reduce chances of A happening given B as a condition condition, but the problem is the play of words when stating the formula such as "both happening" and "this happening given that" is what I find confusing.
$endgroup$
– pwned
Jul 4 '15 at 7:23
$begingroup$
I understood your questions. I only make comments on you false statements. I hope it is clear now.
$endgroup$
– callculus
Jul 4 '15 at 7:27
$begingroup$
@pwned calculus's point is a good one. Your confusion seemed to be about the difference between "A and B" vs. "A given B". You noted that "A given B" sort of entails "A and B". calculus explained that your intuition is partly correct which is that "A given B" must be at least as probable as "A and B". The point is that, in a sense, "A given B" is a "weaker" condition than "A and B".
$endgroup$
– Jared
Jul 4 '15 at 7:37
add a comment |
$begingroup$
In many online sources I've read a statement similar to:
Probability of B happening given A is equal to the probability of A and B both happening divided by B happening
or
$p(A | B) = p(A cap B) / p(B)$
But what is the difference between "B happening given A" and "A and B happening at the same time" ? If B happens given A, didn't A and B happen at the same time ? Why is that chance smaller than the intersection ?
probability bayesian
$endgroup$
In many online sources I've read a statement similar to:
Probability of B happening given A is equal to the probability of A and B both happening divided by B happening
or
$p(A | B) = p(A cap B) / p(B)$
But what is the difference between "B happening given A" and "A and B happening at the same time" ? If B happens given A, didn't A and B happen at the same time ? Why is that chance smaller than the intersection ?
probability bayesian
probability bayesian
asked Jul 4 '15 at 6:31
pwnedpwned
34229
34229
$begingroup$
The formula is right. But the statement is not right. It has to be: "Probability of $A$ happening given $B$ is equal to the probability of A and B both happening divided by B happening."
$endgroup$
– callculus
Jul 4 '15 at 6:42
1
$begingroup$
$P(A|B)$ isn´t smaller than $P(Acap B)$. $P(B)cdot P(A|B)=P(A cap B)$ and $P(B) leq 1$. Thus $P(A|B)geq P(A cap B)$
$endgroup$
– callculus
Jul 4 '15 at 7:02
$begingroup$
@calculus I am aware that the probability of B given some observations is going to reduce chances of A happening given B as a condition condition, but the problem is the play of words when stating the formula such as "both happening" and "this happening given that" is what I find confusing.
$endgroup$
– pwned
Jul 4 '15 at 7:23
$begingroup$
I understood your questions. I only make comments on you false statements. I hope it is clear now.
$endgroup$
– callculus
Jul 4 '15 at 7:27
$begingroup$
@pwned calculus's point is a good one. Your confusion seemed to be about the difference between "A and B" vs. "A given B". You noted that "A given B" sort of entails "A and B". calculus explained that your intuition is partly correct which is that "A given B" must be at least as probable as "A and B". The point is that, in a sense, "A given B" is a "weaker" condition than "A and B".
$endgroup$
– Jared
Jul 4 '15 at 7:37
add a comment |
$begingroup$
The formula is right. But the statement is not right. It has to be: "Probability of $A$ happening given $B$ is equal to the probability of A and B both happening divided by B happening."
$endgroup$
– callculus
Jul 4 '15 at 6:42
1
$begingroup$
$P(A|B)$ isn´t smaller than $P(Acap B)$. $P(B)cdot P(A|B)=P(A cap B)$ and $P(B) leq 1$. Thus $P(A|B)geq P(A cap B)$
$endgroup$
– callculus
Jul 4 '15 at 7:02
$begingroup$
@calculus I am aware that the probability of B given some observations is going to reduce chances of A happening given B as a condition condition, but the problem is the play of words when stating the formula such as "both happening" and "this happening given that" is what I find confusing.
$endgroup$
– pwned
Jul 4 '15 at 7:23
$begingroup$
I understood your questions. I only make comments on you false statements. I hope it is clear now.
$endgroup$
– callculus
Jul 4 '15 at 7:27
$begingroup$
@pwned calculus's point is a good one. Your confusion seemed to be about the difference between "A and B" vs. "A given B". You noted that "A given B" sort of entails "A and B". calculus explained that your intuition is partly correct which is that "A given B" must be at least as probable as "A and B". The point is that, in a sense, "A given B" is a "weaker" condition than "A and B".
$endgroup$
– Jared
Jul 4 '15 at 7:37
$begingroup$
The formula is right. But the statement is not right. It has to be: "Probability of $A$ happening given $B$ is equal to the probability of A and B both happening divided by B happening."
$endgroup$
– callculus
Jul 4 '15 at 6:42
$begingroup$
The formula is right. But the statement is not right. It has to be: "Probability of $A$ happening given $B$ is equal to the probability of A and B both happening divided by B happening."
$endgroup$
– callculus
Jul 4 '15 at 6:42
1
1
$begingroup$
$P(A|B)$ isn´t smaller than $P(Acap B)$. $P(B)cdot P(A|B)=P(A cap B)$ and $P(B) leq 1$. Thus $P(A|B)geq P(A cap B)$
$endgroup$
– callculus
Jul 4 '15 at 7:02
$begingroup$
$P(A|B)$ isn´t smaller than $P(Acap B)$. $P(B)cdot P(A|B)=P(A cap B)$ and $P(B) leq 1$. Thus $P(A|B)geq P(A cap B)$
$endgroup$
– callculus
Jul 4 '15 at 7:02
$begingroup$
@calculus I am aware that the probability of B given some observations is going to reduce chances of A happening given B as a condition condition, but the problem is the play of words when stating the formula such as "both happening" and "this happening given that" is what I find confusing.
$endgroup$
– pwned
Jul 4 '15 at 7:23
$begingroup$
@calculus I am aware that the probability of B given some observations is going to reduce chances of A happening given B as a condition condition, but the problem is the play of words when stating the formula such as "both happening" and "this happening given that" is what I find confusing.
$endgroup$
– pwned
Jul 4 '15 at 7:23
$begingroup$
I understood your questions. I only make comments on you false statements. I hope it is clear now.
$endgroup$
– callculus
Jul 4 '15 at 7:27
$begingroup$
I understood your questions. I only make comments on you false statements. I hope it is clear now.
$endgroup$
– callculus
Jul 4 '15 at 7:27
$begingroup$
@pwned calculus's point is a good one. Your confusion seemed to be about the difference between "A and B" vs. "A given B". You noted that "A given B" sort of entails "A and B". calculus explained that your intuition is partly correct which is that "A given B" must be at least as probable as "A and B". The point is that, in a sense, "A given B" is a "weaker" condition than "A and B".
$endgroup$
– Jared
Jul 4 '15 at 7:37
$begingroup$
@pwned calculus's point is a good one. Your confusion seemed to be about the difference between "A and B" vs. "A given B". You noted that "A given B" sort of entails "A and B". calculus explained that your intuition is partly correct which is that "A given B" must be at least as probable as "A and B". The point is that, in a sense, "A given B" is a "weaker" condition than "A and B".
$endgroup$
– Jared
Jul 4 '15 at 7:37
add a comment |
3 Answers
3
active
oldest
votes
$begingroup$
"If B happens given A, didn't A and B happen at the same time?" Yes that is true (or it could be true--it could also be considering sequential events--not necessarily simultaneous ones). However, A and B happening together might be "unlikely" (really it just means they might not always happen together).
So consider that you have three outcomes and exactly two always occur together. That is there are three possible outcomes: A and B, A and C, or B and C. Now what is the probability that A occurs given that B occurs? Well the probability that they both occur is $frac{1}{3}$ (since A and B is only one possible outcome out of three). But what is the probability that given that B occurs that A occurs? Well there are only two events such that B occurs: either A and B or B and C. Therefore given that B occurs the probability that A also occurs is $frac{1}{2}$ (because you have excluded the possibility that A and C occur since you assumed B occurred).
This is indeed inline with Bayes' Theorem. The probability of B occurring is $frac{2}{3}$ and the probability of A and B occurring is $frac{1}{3}$ therefore the probability of A given B is: $frac{frac{1}{3}}{frac{2}{3}} = frac{1}{2}$. (this is probably a poor example of Bayes' Theorem because, in this case, if you chose any two events, you'd always get $p(x|y neq x) = frac{1}{2}$)
Second Example: Independent Events
Consider that you flip a (fair) coin twice, sequentially. What are the possible outcomes? You can have ${HH, TT, HT, TH}$ (since it's a fair coin, each of these possibilities are equally likely--that's important here although we could extend to situations where they are not equally likely--but then we start getting more into the Bernoulli random variable rather than conditional probabilities). What is the probability that you get a heads on the first flip? Clearly from the above we get $frac{2}{4} = frac{1}{2}$.
Now, what is the probability that you flip a heads given that the first flip is heads? There is only a single outcome that satisfies this (i.e. that you get two heads in a row). However there is only one other event, $HT$, such that heads was rolled first, therefore the probability that you get a heads given that the first flip was heads is $frac{1}{2}$. Again, this is the same as the probability of two heads (heads both times), $frac{1}{4}$, divided by the probability of getting heads the first time, $frac{2}{4}$: $frac{frac{1}{4}}{frac{2}{4}} = frac{1}{2}$.
Indeed we see by the above that the probability of getting heads given an initial flip of heads is the same as the probability of getting heads on the second roll--both are $frac{1}{2}$. This tells us that each flip is independent of the other since the probability of getting heads isn't affected by the previous flip.
$endgroup$
add a comment |
$begingroup$
The difference is that in one case, you already know B has happened. This is not subject to chance anymore, it's a fact. So you compute the probability of A happening, knowing for a fact that B happened
For the probability in the intersection, on the other hand, you don't know that B has happened and have to calculate the probability that both A and B occur
$endgroup$
add a comment |
$begingroup$
In summary, the essence of conditional probability is in every probability calculation, find the Whole
(i.e. universal set), find the Part
(intersection between the event you are trying to calculate the probability and the universal set) and divide the Part
by the Whole
to find the probability of given event (i.e the comparative relation between Part
and Whole
).
Let A
be any given event and U
be the set of all possible events or outcomes:
$ P(A)=frac{P(A)}{P(U)} = frac{P(A cap U)}{P(U)} = P(A|U) $
If you are struggling to see the full equivalence above, remember that in set theory $ A cap B = A $ when A is contained in B, which is always the case for any set when are trying to find the intersection with the Universal set (U
).
Why do we need to divide?
Proportions: Finding the likelihood or probability of some event can be interpreted as a proportion problem, which implies the existence of a Part
and a Whole
. The way that you find how much a Part
represent from the Whole
is by dividing the Part
by the Whole
: $ frac{Part}{Whole} $ (if you are still uncomfortable with the concept and for some reason can't visualize why division would provide you with such information, just take it as a rule for now, at some point the concept will click). So whenever you are thinking about the probability or likelihood of an event always think about the Part
and the Whole
, which leads to the next question: How do we find the Part
and the Whole
? We will need to understand how we represent events
to answer that question.
Sets, Subsets and Cardinalities: An event is a set of outcomes, the cardinality of a set is the measure of the size of that same set. Now, for the sake of our probabilistic theory discussion, you will see that a Part
is always a subset of the Whole
(you may be thinking of some edge cases that do not fit this rule, keep reading and you will see that this idea indeed holds true for all cases).
The first step to find a probability of any event is to find the set that represents the Whole
(i.e. the scope of all the possible outcomes, when you roll a die of 6 sides the Whole
set can be represented as the numbers from 1 through 6, since they form all the possible outcomes), the second step is to find the Part
(i.e. the set containing the subset of outcomes that you are trying to find the likelihood of occurrence).
Every probability calculation can be seen with the lens of a conditional probability (What!? Keep reading :) )
Let A be any given event and U be the set of all possible events or outcomes:
$ P(A)=frac{P(A)}{P(U)} = frac{P(A cap U)}{P(U)} = P(A|U) $
If you are struggling to see the full equivalence above, remember that in set theory A and B = A when A is contained in B, which is always the case for any set when are trying to find the intersection with the Universal set.
As an example, say we are interested in knowing the probability of getting an even number when rolling a die of 6 sides, you can always formulate the problem thinking about conditional probability, when we defined that we are going to be rolling a die of six sides we just implicitly defined the Whole set (i.e the possible outcomes are numbers from 1 to 6), the second piece of information is the event itself that we would like to obtain the probability (even number), that is the Part
. An important subtlety here, the Part you are interested here in the first place, is not ALL the even numbers but the ones that make part of your Whole set (i.e. subset of the Whole). How do you find that subset? The intersection of the two sets? precisely.
So the intersection gives you the Part
that you are interested in, and once you have the Whole
and Part
you divide the two and find the probability (frequency of occurrence, or the comparative relation between the Part
and the Whole
).
If the outcomes or event that you are trying to find the likelihood does not make any part of the Whole, in other words, if the intersection of both sets is empty then we know that the probability of such event is zero, the event is simply impossible to occur, there is no way you will get a 7 if you roll a die of 6 sides.
Always remember when you see P(A|B)
we are already explicitly saying that B
is the Whole
or universal set, all you have to do now is to find the comparative relation between the Part
and Whole
..
To end, let me clarify a few common misconceptions:
On the context of conditional probability, aside from the relationship between the
Whole
andPart
, the probability of what you are judging to be theWhole
set does not matter by itself! As an example, the probability of someone having a tumor is irrelevant, if what we are trying to find is the probability of someone having a benign tumor given the person is already diagnosed with a tumor.P(A|B)
is not the same asP(B|A)
: The probability of someone having a benign tumor given the person is already diagnosed with a tumor, is not the same as the probability of someone being diagnosed with a tumor given the fact the person has a benign tumor.Shouldn't
P(A|B)
be equal toP(A and B)
?
This is a common misinterpretation of what conditional probability means, we are not interested in the probability of A and B happening (i.e. $ frac{P(A cap B)}{P(U)} $), that's called joint probability and is a separate (and related) subject. Conditional probability means, now thatB
ALREADY HAPPENED, what is the probability ofA
also happening?
New contributor
$endgroup$
add a comment |
Your Answer
StackExchange.ifUsing("editor", function () {
return StackExchange.using("mathjaxEditing", function () {
StackExchange.MarkdownEditor.creationCallbacks.add(function (editor, postfix) {
StackExchange.mathjaxEditing.prepareWmdForMathJax(editor, postfix, [["$", "$"], ["\\(","\\)"]]);
});
});
}, "mathjax-editing");
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "69"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
noCode: true, onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f1348919%2funderstanding-conditional-probability-basics%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
3 Answers
3
active
oldest
votes
3 Answers
3
active
oldest
votes
active
oldest
votes
active
oldest
votes
$begingroup$
"If B happens given A, didn't A and B happen at the same time?" Yes that is true (or it could be true--it could also be considering sequential events--not necessarily simultaneous ones). However, A and B happening together might be "unlikely" (really it just means they might not always happen together).
So consider that you have three outcomes and exactly two always occur together. That is there are three possible outcomes: A and B, A and C, or B and C. Now what is the probability that A occurs given that B occurs? Well the probability that they both occur is $frac{1}{3}$ (since A and B is only one possible outcome out of three). But what is the probability that given that B occurs that A occurs? Well there are only two events such that B occurs: either A and B or B and C. Therefore given that B occurs the probability that A also occurs is $frac{1}{2}$ (because you have excluded the possibility that A and C occur since you assumed B occurred).
This is indeed inline with Bayes' Theorem. The probability of B occurring is $frac{2}{3}$ and the probability of A and B occurring is $frac{1}{3}$ therefore the probability of A given B is: $frac{frac{1}{3}}{frac{2}{3}} = frac{1}{2}$. (this is probably a poor example of Bayes' Theorem because, in this case, if you chose any two events, you'd always get $p(x|y neq x) = frac{1}{2}$)
Second Example: Independent Events
Consider that you flip a (fair) coin twice, sequentially. What are the possible outcomes? You can have ${HH, TT, HT, TH}$ (since it's a fair coin, each of these possibilities are equally likely--that's important here although we could extend to situations where they are not equally likely--but then we start getting more into the Bernoulli random variable rather than conditional probabilities). What is the probability that you get a heads on the first flip? Clearly from the above we get $frac{2}{4} = frac{1}{2}$.
Now, what is the probability that you flip a heads given that the first flip is heads? There is only a single outcome that satisfies this (i.e. that you get two heads in a row). However there is only one other event, $HT$, such that heads was rolled first, therefore the probability that you get a heads given that the first flip was heads is $frac{1}{2}$. Again, this is the same as the probability of two heads (heads both times), $frac{1}{4}$, divided by the probability of getting heads the first time, $frac{2}{4}$: $frac{frac{1}{4}}{frac{2}{4}} = frac{1}{2}$.
Indeed we see by the above that the probability of getting heads given an initial flip of heads is the same as the probability of getting heads on the second roll--both are $frac{1}{2}$. This tells us that each flip is independent of the other since the probability of getting heads isn't affected by the previous flip.
$endgroup$
add a comment |
$begingroup$
"If B happens given A, didn't A and B happen at the same time?" Yes that is true (or it could be true--it could also be considering sequential events--not necessarily simultaneous ones). However, A and B happening together might be "unlikely" (really it just means they might not always happen together).
So consider that you have three outcomes and exactly two always occur together. That is there are three possible outcomes: A and B, A and C, or B and C. Now what is the probability that A occurs given that B occurs? Well the probability that they both occur is $frac{1}{3}$ (since A and B is only one possible outcome out of three). But what is the probability that given that B occurs that A occurs? Well there are only two events such that B occurs: either A and B or B and C. Therefore given that B occurs the probability that A also occurs is $frac{1}{2}$ (because you have excluded the possibility that A and C occur since you assumed B occurred).
This is indeed inline with Bayes' Theorem. The probability of B occurring is $frac{2}{3}$ and the probability of A and B occurring is $frac{1}{3}$ therefore the probability of A given B is: $frac{frac{1}{3}}{frac{2}{3}} = frac{1}{2}$. (this is probably a poor example of Bayes' Theorem because, in this case, if you chose any two events, you'd always get $p(x|y neq x) = frac{1}{2}$)
Second Example: Independent Events
Consider that you flip a (fair) coin twice, sequentially. What are the possible outcomes? You can have ${HH, TT, HT, TH}$ (since it's a fair coin, each of these possibilities are equally likely--that's important here although we could extend to situations where they are not equally likely--but then we start getting more into the Bernoulli random variable rather than conditional probabilities). What is the probability that you get a heads on the first flip? Clearly from the above we get $frac{2}{4} = frac{1}{2}$.
Now, what is the probability that you flip a heads given that the first flip is heads? There is only a single outcome that satisfies this (i.e. that you get two heads in a row). However there is only one other event, $HT$, such that heads was rolled first, therefore the probability that you get a heads given that the first flip was heads is $frac{1}{2}$. Again, this is the same as the probability of two heads (heads both times), $frac{1}{4}$, divided by the probability of getting heads the first time, $frac{2}{4}$: $frac{frac{1}{4}}{frac{2}{4}} = frac{1}{2}$.
Indeed we see by the above that the probability of getting heads given an initial flip of heads is the same as the probability of getting heads on the second roll--both are $frac{1}{2}$. This tells us that each flip is independent of the other since the probability of getting heads isn't affected by the previous flip.
$endgroup$
add a comment |
$begingroup$
"If B happens given A, didn't A and B happen at the same time?" Yes that is true (or it could be true--it could also be considering sequential events--not necessarily simultaneous ones). However, A and B happening together might be "unlikely" (really it just means they might not always happen together).
So consider that you have three outcomes and exactly two always occur together. That is there are three possible outcomes: A and B, A and C, or B and C. Now what is the probability that A occurs given that B occurs? Well the probability that they both occur is $frac{1}{3}$ (since A and B is only one possible outcome out of three). But what is the probability that given that B occurs that A occurs? Well there are only two events such that B occurs: either A and B or B and C. Therefore given that B occurs the probability that A also occurs is $frac{1}{2}$ (because you have excluded the possibility that A and C occur since you assumed B occurred).
This is indeed inline with Bayes' Theorem. The probability of B occurring is $frac{2}{3}$ and the probability of A and B occurring is $frac{1}{3}$ therefore the probability of A given B is: $frac{frac{1}{3}}{frac{2}{3}} = frac{1}{2}$. (this is probably a poor example of Bayes' Theorem because, in this case, if you chose any two events, you'd always get $p(x|y neq x) = frac{1}{2}$)
Second Example: Independent Events
Consider that you flip a (fair) coin twice, sequentially. What are the possible outcomes? You can have ${HH, TT, HT, TH}$ (since it's a fair coin, each of these possibilities are equally likely--that's important here although we could extend to situations where they are not equally likely--but then we start getting more into the Bernoulli random variable rather than conditional probabilities). What is the probability that you get a heads on the first flip? Clearly from the above we get $frac{2}{4} = frac{1}{2}$.
Now, what is the probability that you flip a heads given that the first flip is heads? There is only a single outcome that satisfies this (i.e. that you get two heads in a row). However there is only one other event, $HT$, such that heads was rolled first, therefore the probability that you get a heads given that the first flip was heads is $frac{1}{2}$. Again, this is the same as the probability of two heads (heads both times), $frac{1}{4}$, divided by the probability of getting heads the first time, $frac{2}{4}$: $frac{frac{1}{4}}{frac{2}{4}} = frac{1}{2}$.
Indeed we see by the above that the probability of getting heads given an initial flip of heads is the same as the probability of getting heads on the second roll--both are $frac{1}{2}$. This tells us that each flip is independent of the other since the probability of getting heads isn't affected by the previous flip.
$endgroup$
"If B happens given A, didn't A and B happen at the same time?" Yes that is true (or it could be true--it could also be considering sequential events--not necessarily simultaneous ones). However, A and B happening together might be "unlikely" (really it just means they might not always happen together).
So consider that you have three outcomes and exactly two always occur together. That is there are three possible outcomes: A and B, A and C, or B and C. Now what is the probability that A occurs given that B occurs? Well the probability that they both occur is $frac{1}{3}$ (since A and B is only one possible outcome out of three). But what is the probability that given that B occurs that A occurs? Well there are only two events such that B occurs: either A and B or B and C. Therefore given that B occurs the probability that A also occurs is $frac{1}{2}$ (because you have excluded the possibility that A and C occur since you assumed B occurred).
This is indeed inline with Bayes' Theorem. The probability of B occurring is $frac{2}{3}$ and the probability of A and B occurring is $frac{1}{3}$ therefore the probability of A given B is: $frac{frac{1}{3}}{frac{2}{3}} = frac{1}{2}$. (this is probably a poor example of Bayes' Theorem because, in this case, if you chose any two events, you'd always get $p(x|y neq x) = frac{1}{2}$)
Second Example: Independent Events
Consider that you flip a (fair) coin twice, sequentially. What are the possible outcomes? You can have ${HH, TT, HT, TH}$ (since it's a fair coin, each of these possibilities are equally likely--that's important here although we could extend to situations where they are not equally likely--but then we start getting more into the Bernoulli random variable rather than conditional probabilities). What is the probability that you get a heads on the first flip? Clearly from the above we get $frac{2}{4} = frac{1}{2}$.
Now, what is the probability that you flip a heads given that the first flip is heads? There is only a single outcome that satisfies this (i.e. that you get two heads in a row). However there is only one other event, $HT$, such that heads was rolled first, therefore the probability that you get a heads given that the first flip was heads is $frac{1}{2}$. Again, this is the same as the probability of two heads (heads both times), $frac{1}{4}$, divided by the probability of getting heads the first time, $frac{2}{4}$: $frac{frac{1}{4}}{frac{2}{4}} = frac{1}{2}$.
Indeed we see by the above that the probability of getting heads given an initial flip of heads is the same as the probability of getting heads on the second roll--both are $frac{1}{2}$. This tells us that each flip is independent of the other since the probability of getting heads isn't affected by the previous flip.
edited Jul 4 '15 at 7:32
answered Jul 4 '15 at 6:47
JaredJared
5,20311116
5,20311116
add a comment |
add a comment |
$begingroup$
The difference is that in one case, you already know B has happened. This is not subject to chance anymore, it's a fact. So you compute the probability of A happening, knowing for a fact that B happened
For the probability in the intersection, on the other hand, you don't know that B has happened and have to calculate the probability that both A and B occur
$endgroup$
add a comment |
$begingroup$
The difference is that in one case, you already know B has happened. This is not subject to chance anymore, it's a fact. So you compute the probability of A happening, knowing for a fact that B happened
For the probability in the intersection, on the other hand, you don't know that B has happened and have to calculate the probability that both A and B occur
$endgroup$
add a comment |
$begingroup$
The difference is that in one case, you already know B has happened. This is not subject to chance anymore, it's a fact. So you compute the probability of A happening, knowing for a fact that B happened
For the probability in the intersection, on the other hand, you don't know that B has happened and have to calculate the probability that both A and B occur
$endgroup$
The difference is that in one case, you already know B has happened. This is not subject to chance anymore, it's a fact. So you compute the probability of A happening, knowing for a fact that B happened
For the probability in the intersection, on the other hand, you don't know that B has happened and have to calculate the probability that both A and B occur
answered Jul 4 '15 at 6:43
AntAnt
17.5k22974
17.5k22974
add a comment |
add a comment |
$begingroup$
In summary, the essence of conditional probability is in every probability calculation, find the Whole
(i.e. universal set), find the Part
(intersection between the event you are trying to calculate the probability and the universal set) and divide the Part
by the Whole
to find the probability of given event (i.e the comparative relation between Part
and Whole
).
Let A
be any given event and U
be the set of all possible events or outcomes:
$ P(A)=frac{P(A)}{P(U)} = frac{P(A cap U)}{P(U)} = P(A|U) $
If you are struggling to see the full equivalence above, remember that in set theory $ A cap B = A $ when A is contained in B, which is always the case for any set when are trying to find the intersection with the Universal set (U
).
Why do we need to divide?
Proportions: Finding the likelihood or probability of some event can be interpreted as a proportion problem, which implies the existence of a Part
and a Whole
. The way that you find how much a Part
represent from the Whole
is by dividing the Part
by the Whole
: $ frac{Part}{Whole} $ (if you are still uncomfortable with the concept and for some reason can't visualize why division would provide you with such information, just take it as a rule for now, at some point the concept will click). So whenever you are thinking about the probability or likelihood of an event always think about the Part
and the Whole
, which leads to the next question: How do we find the Part
and the Whole
? We will need to understand how we represent events
to answer that question.
Sets, Subsets and Cardinalities: An event is a set of outcomes, the cardinality of a set is the measure of the size of that same set. Now, for the sake of our probabilistic theory discussion, you will see that a Part
is always a subset of the Whole
(you may be thinking of some edge cases that do not fit this rule, keep reading and you will see that this idea indeed holds true for all cases).
The first step to find a probability of any event is to find the set that represents the Whole
(i.e. the scope of all the possible outcomes, when you roll a die of 6 sides the Whole
set can be represented as the numbers from 1 through 6, since they form all the possible outcomes), the second step is to find the Part
(i.e. the set containing the subset of outcomes that you are trying to find the likelihood of occurrence).
Every probability calculation can be seen with the lens of a conditional probability (What!? Keep reading :) )
Let A be any given event and U be the set of all possible events or outcomes:
$ P(A)=frac{P(A)}{P(U)} = frac{P(A cap U)}{P(U)} = P(A|U) $
If you are struggling to see the full equivalence above, remember that in set theory A and B = A when A is contained in B, which is always the case for any set when are trying to find the intersection with the Universal set.
As an example, say we are interested in knowing the probability of getting an even number when rolling a die of 6 sides, you can always formulate the problem thinking about conditional probability, when we defined that we are going to be rolling a die of six sides we just implicitly defined the Whole set (i.e the possible outcomes are numbers from 1 to 6), the second piece of information is the event itself that we would like to obtain the probability (even number), that is the Part
. An important subtlety here, the Part you are interested here in the first place, is not ALL the even numbers but the ones that make part of your Whole set (i.e. subset of the Whole). How do you find that subset? The intersection of the two sets? precisely.
So the intersection gives you the Part
that you are interested in, and once you have the Whole
and Part
you divide the two and find the probability (frequency of occurrence, or the comparative relation between the Part
and the Whole
).
If the outcomes or event that you are trying to find the likelihood does not make any part of the Whole, in other words, if the intersection of both sets is empty then we know that the probability of such event is zero, the event is simply impossible to occur, there is no way you will get a 7 if you roll a die of 6 sides.
Always remember when you see P(A|B)
we are already explicitly saying that B
is the Whole
or universal set, all you have to do now is to find the comparative relation between the Part
and Whole
..
To end, let me clarify a few common misconceptions:
On the context of conditional probability, aside from the relationship between the
Whole
andPart
, the probability of what you are judging to be theWhole
set does not matter by itself! As an example, the probability of someone having a tumor is irrelevant, if what we are trying to find is the probability of someone having a benign tumor given the person is already diagnosed with a tumor.P(A|B)
is not the same asP(B|A)
: The probability of someone having a benign tumor given the person is already diagnosed with a tumor, is not the same as the probability of someone being diagnosed with a tumor given the fact the person has a benign tumor.Shouldn't
P(A|B)
be equal toP(A and B)
?
This is a common misinterpretation of what conditional probability means, we are not interested in the probability of A and B happening (i.e. $ frac{P(A cap B)}{P(U)} $), that's called joint probability and is a separate (and related) subject. Conditional probability means, now thatB
ALREADY HAPPENED, what is the probability ofA
also happening?
New contributor
$endgroup$
add a comment |
$begingroup$
In summary, the essence of conditional probability is in every probability calculation, find the Whole
(i.e. universal set), find the Part
(intersection between the event you are trying to calculate the probability and the universal set) and divide the Part
by the Whole
to find the probability of given event (i.e the comparative relation between Part
and Whole
).
Let A
be any given event and U
be the set of all possible events or outcomes:
$ P(A)=frac{P(A)}{P(U)} = frac{P(A cap U)}{P(U)} = P(A|U) $
If you are struggling to see the full equivalence above, remember that in set theory $ A cap B = A $ when A is contained in B, which is always the case for any set when are trying to find the intersection with the Universal set (U
).
Why do we need to divide?
Proportions: Finding the likelihood or probability of some event can be interpreted as a proportion problem, which implies the existence of a Part
and a Whole
. The way that you find how much a Part
represent from the Whole
is by dividing the Part
by the Whole
: $ frac{Part}{Whole} $ (if you are still uncomfortable with the concept and for some reason can't visualize why division would provide you with such information, just take it as a rule for now, at some point the concept will click). So whenever you are thinking about the probability or likelihood of an event always think about the Part
and the Whole
, which leads to the next question: How do we find the Part
and the Whole
? We will need to understand how we represent events
to answer that question.
Sets, Subsets and Cardinalities: An event is a set of outcomes, the cardinality of a set is the measure of the size of that same set. Now, for the sake of our probabilistic theory discussion, you will see that a Part
is always a subset of the Whole
(you may be thinking of some edge cases that do not fit this rule, keep reading and you will see that this idea indeed holds true for all cases).
The first step to find a probability of any event is to find the set that represents the Whole
(i.e. the scope of all the possible outcomes, when you roll a die of 6 sides the Whole
set can be represented as the numbers from 1 through 6, since they form all the possible outcomes), the second step is to find the Part
(i.e. the set containing the subset of outcomes that you are trying to find the likelihood of occurrence).
Every probability calculation can be seen with the lens of a conditional probability (What!? Keep reading :) )
Let A be any given event and U be the set of all possible events or outcomes:
$ P(A)=frac{P(A)}{P(U)} = frac{P(A cap U)}{P(U)} = P(A|U) $
If you are struggling to see the full equivalence above, remember that in set theory A and B = A when A is contained in B, which is always the case for any set when are trying to find the intersection with the Universal set.
As an example, say we are interested in knowing the probability of getting an even number when rolling a die of 6 sides, you can always formulate the problem thinking about conditional probability, when we defined that we are going to be rolling a die of six sides we just implicitly defined the Whole set (i.e the possible outcomes are numbers from 1 to 6), the second piece of information is the event itself that we would like to obtain the probability (even number), that is the Part
. An important subtlety here, the Part you are interested here in the first place, is not ALL the even numbers but the ones that make part of your Whole set (i.e. subset of the Whole). How do you find that subset? The intersection of the two sets? precisely.
So the intersection gives you the Part
that you are interested in, and once you have the Whole
and Part
you divide the two and find the probability (frequency of occurrence, or the comparative relation between the Part
and the Whole
).
If the outcomes or event that you are trying to find the likelihood does not make any part of the Whole, in other words, if the intersection of both sets is empty then we know that the probability of such event is zero, the event is simply impossible to occur, there is no way you will get a 7 if you roll a die of 6 sides.
Always remember when you see P(A|B)
we are already explicitly saying that B
is the Whole
or universal set, all you have to do now is to find the comparative relation between the Part
and Whole
..
To end, let me clarify a few common misconceptions:
On the context of conditional probability, aside from the relationship between the
Whole
andPart
, the probability of what you are judging to be theWhole
set does not matter by itself! As an example, the probability of someone having a tumor is irrelevant, if what we are trying to find is the probability of someone having a benign tumor given the person is already diagnosed with a tumor.P(A|B)
is not the same asP(B|A)
: The probability of someone having a benign tumor given the person is already diagnosed with a tumor, is not the same as the probability of someone being diagnosed with a tumor given the fact the person has a benign tumor.Shouldn't
P(A|B)
be equal toP(A and B)
?
This is a common misinterpretation of what conditional probability means, we are not interested in the probability of A and B happening (i.e. $ frac{P(A cap B)}{P(U)} $), that's called joint probability and is a separate (and related) subject. Conditional probability means, now thatB
ALREADY HAPPENED, what is the probability ofA
also happening?
New contributor
$endgroup$
add a comment |
$begingroup$
In summary, the essence of conditional probability is in every probability calculation, find the Whole
(i.e. universal set), find the Part
(intersection between the event you are trying to calculate the probability and the universal set) and divide the Part
by the Whole
to find the probability of given event (i.e the comparative relation between Part
and Whole
).
Let A
be any given event and U
be the set of all possible events or outcomes:
$ P(A)=frac{P(A)}{P(U)} = frac{P(A cap U)}{P(U)} = P(A|U) $
If you are struggling to see the full equivalence above, remember that in set theory $ A cap B = A $ when A is contained in B, which is always the case for any set when are trying to find the intersection with the Universal set (U
).
Why do we need to divide?
Proportions: Finding the likelihood or probability of some event can be interpreted as a proportion problem, which implies the existence of a Part
and a Whole
. The way that you find how much a Part
represent from the Whole
is by dividing the Part
by the Whole
: $ frac{Part}{Whole} $ (if you are still uncomfortable with the concept and for some reason can't visualize why division would provide you with such information, just take it as a rule for now, at some point the concept will click). So whenever you are thinking about the probability or likelihood of an event always think about the Part
and the Whole
, which leads to the next question: How do we find the Part
and the Whole
? We will need to understand how we represent events
to answer that question.
Sets, Subsets and Cardinalities: An event is a set of outcomes, the cardinality of a set is the measure of the size of that same set. Now, for the sake of our probabilistic theory discussion, you will see that a Part
is always a subset of the Whole
(you may be thinking of some edge cases that do not fit this rule, keep reading and you will see that this idea indeed holds true for all cases).
The first step to find a probability of any event is to find the set that represents the Whole
(i.e. the scope of all the possible outcomes, when you roll a die of 6 sides the Whole
set can be represented as the numbers from 1 through 6, since they form all the possible outcomes), the second step is to find the Part
(i.e. the set containing the subset of outcomes that you are trying to find the likelihood of occurrence).
Every probability calculation can be seen with the lens of a conditional probability (What!? Keep reading :) )
Let A be any given event and U be the set of all possible events or outcomes:
$ P(A)=frac{P(A)}{P(U)} = frac{P(A cap U)}{P(U)} = P(A|U) $
If you are struggling to see the full equivalence above, remember that in set theory A and B = A when A is contained in B, which is always the case for any set when are trying to find the intersection with the Universal set.
As an example, say we are interested in knowing the probability of getting an even number when rolling a die of 6 sides, you can always formulate the problem thinking about conditional probability, when we defined that we are going to be rolling a die of six sides we just implicitly defined the Whole set (i.e the possible outcomes are numbers from 1 to 6), the second piece of information is the event itself that we would like to obtain the probability (even number), that is the Part
. An important subtlety here, the Part you are interested here in the first place, is not ALL the even numbers but the ones that make part of your Whole set (i.e. subset of the Whole). How do you find that subset? The intersection of the two sets? precisely.
So the intersection gives you the Part
that you are interested in, and once you have the Whole
and Part
you divide the two and find the probability (frequency of occurrence, or the comparative relation between the Part
and the Whole
).
If the outcomes or event that you are trying to find the likelihood does not make any part of the Whole, in other words, if the intersection of both sets is empty then we know that the probability of such event is zero, the event is simply impossible to occur, there is no way you will get a 7 if you roll a die of 6 sides.
Always remember when you see P(A|B)
we are already explicitly saying that B
is the Whole
or universal set, all you have to do now is to find the comparative relation between the Part
and Whole
..
To end, let me clarify a few common misconceptions:
On the context of conditional probability, aside from the relationship between the
Whole
andPart
, the probability of what you are judging to be theWhole
set does not matter by itself! As an example, the probability of someone having a tumor is irrelevant, if what we are trying to find is the probability of someone having a benign tumor given the person is already diagnosed with a tumor.P(A|B)
is not the same asP(B|A)
: The probability of someone having a benign tumor given the person is already diagnosed with a tumor, is not the same as the probability of someone being diagnosed with a tumor given the fact the person has a benign tumor.Shouldn't
P(A|B)
be equal toP(A and B)
?
This is a common misinterpretation of what conditional probability means, we are not interested in the probability of A and B happening (i.e. $ frac{P(A cap B)}{P(U)} $), that's called joint probability and is a separate (and related) subject. Conditional probability means, now thatB
ALREADY HAPPENED, what is the probability ofA
also happening?
New contributor
$endgroup$
In summary, the essence of conditional probability is in every probability calculation, find the Whole
(i.e. universal set), find the Part
(intersection between the event you are trying to calculate the probability and the universal set) and divide the Part
by the Whole
to find the probability of given event (i.e the comparative relation between Part
and Whole
).
Let A
be any given event and U
be the set of all possible events or outcomes:
$ P(A)=frac{P(A)}{P(U)} = frac{P(A cap U)}{P(U)} = P(A|U) $
If you are struggling to see the full equivalence above, remember that in set theory $ A cap B = A $ when A is contained in B, which is always the case for any set when are trying to find the intersection with the Universal set (U
).
Why do we need to divide?
Proportions: Finding the likelihood or probability of some event can be interpreted as a proportion problem, which implies the existence of a Part
and a Whole
. The way that you find how much a Part
represent from the Whole
is by dividing the Part
by the Whole
: $ frac{Part}{Whole} $ (if you are still uncomfortable with the concept and for some reason can't visualize why division would provide you with such information, just take it as a rule for now, at some point the concept will click). So whenever you are thinking about the probability or likelihood of an event always think about the Part
and the Whole
, which leads to the next question: How do we find the Part
and the Whole
? We will need to understand how we represent events
to answer that question.
Sets, Subsets and Cardinalities: An event is a set of outcomes, the cardinality of a set is the measure of the size of that same set. Now, for the sake of our probabilistic theory discussion, you will see that a Part
is always a subset of the Whole
(you may be thinking of some edge cases that do not fit this rule, keep reading and you will see that this idea indeed holds true for all cases).
The first step to find a probability of any event is to find the set that represents the Whole
(i.e. the scope of all the possible outcomes, when you roll a die of 6 sides the Whole
set can be represented as the numbers from 1 through 6, since they form all the possible outcomes), the second step is to find the Part
(i.e. the set containing the subset of outcomes that you are trying to find the likelihood of occurrence).
Every probability calculation can be seen with the lens of a conditional probability (What!? Keep reading :) )
Let A be any given event and U be the set of all possible events or outcomes:
$ P(A)=frac{P(A)}{P(U)} = frac{P(A cap U)}{P(U)} = P(A|U) $
If you are struggling to see the full equivalence above, remember that in set theory A and B = A when A is contained in B, which is always the case for any set when are trying to find the intersection with the Universal set.
As an example, say we are interested in knowing the probability of getting an even number when rolling a die of 6 sides, you can always formulate the problem thinking about conditional probability, when we defined that we are going to be rolling a die of six sides we just implicitly defined the Whole set (i.e the possible outcomes are numbers from 1 to 6), the second piece of information is the event itself that we would like to obtain the probability (even number), that is the Part
. An important subtlety here, the Part you are interested here in the first place, is not ALL the even numbers but the ones that make part of your Whole set (i.e. subset of the Whole). How do you find that subset? The intersection of the two sets? precisely.
So the intersection gives you the Part
that you are interested in, and once you have the Whole
and Part
you divide the two and find the probability (frequency of occurrence, or the comparative relation between the Part
and the Whole
).
If the outcomes or event that you are trying to find the likelihood does not make any part of the Whole, in other words, if the intersection of both sets is empty then we know that the probability of such event is zero, the event is simply impossible to occur, there is no way you will get a 7 if you roll a die of 6 sides.
Always remember when you see P(A|B)
we are already explicitly saying that B
is the Whole
or universal set, all you have to do now is to find the comparative relation between the Part
and Whole
..
To end, let me clarify a few common misconceptions:
On the context of conditional probability, aside from the relationship between the
Whole
andPart
, the probability of what you are judging to be theWhole
set does not matter by itself! As an example, the probability of someone having a tumor is irrelevant, if what we are trying to find is the probability of someone having a benign tumor given the person is already diagnosed with a tumor.P(A|B)
is not the same asP(B|A)
: The probability of someone having a benign tumor given the person is already diagnosed with a tumor, is not the same as the probability of someone being diagnosed with a tumor given the fact the person has a benign tumor.Shouldn't
P(A|B)
be equal toP(A and B)
?
This is a common misinterpretation of what conditional probability means, we are not interested in the probability of A and B happening (i.e. $ frac{P(A cap B)}{P(U)} $), that's called joint probability and is a separate (and related) subject. Conditional probability means, now thatB
ALREADY HAPPENED, what is the probability ofA
also happening?
New contributor
edited Mar 10 at 16:10
New contributor
answered Mar 10 at 2:30
bsdbsd
1011
1011
New contributor
New contributor
add a comment |
add a comment |
Thanks for contributing an answer to Mathematics Stack Exchange!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
Use MathJax to format equations. MathJax reference.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fmath.stackexchange.com%2fquestions%2f1348919%2funderstanding-conditional-probability-basics%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
$begingroup$
The formula is right. But the statement is not right. It has to be: "Probability of $A$ happening given $B$ is equal to the probability of A and B both happening divided by B happening."
$endgroup$
– callculus
Jul 4 '15 at 6:42
1
$begingroup$
$P(A|B)$ isn´t smaller than $P(Acap B)$. $P(B)cdot P(A|B)=P(A cap B)$ and $P(B) leq 1$. Thus $P(A|B)geq P(A cap B)$
$endgroup$
– callculus
Jul 4 '15 at 7:02
$begingroup$
@calculus I am aware that the probability of B given some observations is going to reduce chances of A happening given B as a condition condition, but the problem is the play of words when stating the formula such as "both happening" and "this happening given that" is what I find confusing.
$endgroup$
– pwned
Jul 4 '15 at 7:23
$begingroup$
I understood your questions. I only make comments on you false statements. I hope it is clear now.
$endgroup$
– callculus
Jul 4 '15 at 7:27
$begingroup$
@pwned calculus's point is a good one. Your confusion seemed to be about the difference between "A and B" vs. "A given B". You noted that "A given B" sort of entails "A and B". calculus explained that your intuition is partly correct which is that "A given B" must be at least as probable as "A and B". The point is that, in a sense, "A given B" is a "weaker" condition than "A and B".
$endgroup$
– Jared
Jul 4 '15 at 7:37