<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
		<id>http://eclr.humanities.manchester.ac.uk/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Admin</id>
		<title>ECLR - User contributions [en]</title>
		<link rel="self" type="application/atom+xml" href="http://eclr.humanities.manchester.ac.uk/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Admin"/>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php/Special:Contributions/Admin"/>
		<updated>2026-05-16T02:56:10Z</updated>
		<subtitle>User contributions</subtitle>
		<generator>MediaWiki 1.30.1</generator>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=MediaWiki:CustomBlockMainNav&amp;diff=3789</id>
		<title>MediaWiki:CustomBlockMainNav</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=MediaWiki:CustomBlockMainNav&amp;diff=3789"/>
				<updated>2015-07-07T13:06:05Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&amp;lt;!-- Vergiss nicht VoWi:Navigation anzupassen, wenn du diese Seite aenderst! --&amp;gt;&lt;br /&gt;
* [[Main_Page|Home]]&lt;br /&gt;
** [[MATLAB|MATLAB]]&lt;br /&gt;
** [[R|RStudio]]&lt;br /&gt;
** [[EVIEWS|EVIEWS]]&lt;br /&gt;
** [[Statistics|Statistics]]&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Conditional&amp;diff=2880</id>
		<title>Probability Conditional</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Conditional&amp;diff=2880"/>
				<updated>2013-08-09T10:59:22Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: /* Exercises */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Conditional Probability =&lt;br /&gt;
&lt;br /&gt;
An important consideration in the development of probability is that of &amp;#039;&amp;#039;conditional probability&amp;#039;&amp;#039;. This refers to the calculation of updating probabilities in the light of revealed information. For example, insurance companies nearly always set their home contents insurance premiums on the basis of the postcode in which the home is located. That is to say, insurance companies believe the risk depends upon the location; i.e., the probability of property crime is assessed conditional upon the location of the property. (A similar calculation is made to set car insurance premiums.) As a result, the premiums for two identical households located in different parts of the country can differ substantially.&lt;br /&gt;
&lt;br /&gt;
* In general, the probability of an event, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, occurring &amp;#039;&amp;#039;given&amp;#039;&amp;#039; that an event, &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;, has occurred is called the &amp;#039;&amp;#039;conditional probability&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and is denoted &amp;lt;math&amp;gt;\Pr (E|F)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
As another example, it has been well documented that the ability of a new born baby to survive is closely associated with its birth-weight. A birth-weight of less than 1500&amp;#039;&amp;#039;g&amp;#039;&amp;#039; is regarded as dangerously low. Consider &amp;lt;math&amp;gt;E=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;birth weight of a baby is less than&amp;#039;&amp;#039; 1500&amp;#039;&amp;#039;g&amp;#039;&amp;#039;, &amp;lt;math&amp;gt;F=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;mother smoked during pregnancy&amp;#039;&amp;#039;; then evidence as to whether &amp;lt;math&amp;gt;\Pr(E|F)&amp;gt;\Pr (E|\bar{F})&amp;lt;/math&amp;gt; is of considerable interest.&lt;br /&gt;
&lt;br /&gt;
As a preliminary to the main development, consider the simple experiment of rolling a fair die and observing the number of dots on the upturned face. Then &amp;lt;math&amp;gt;S=\left\{ 1,2,3,4,5,6\right\} &amp;lt;/math&amp;gt; and define events, &amp;lt;math&amp;gt;E=\left\{4\right\} &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F=\left\{ 4,5,6\right\} ;&amp;lt;/math&amp;gt; we are interested in &amp;lt;math&amp;gt;\Pr \left(E|F\right)&amp;lt;/math&amp;gt;. To work this out we take &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; as known. Given this knowledge the sample space becomes restricted to simply &amp;lt;math&amp;gt;\left\{ 4,5,6\right\} &amp;lt;/math&amp;gt; and, given no other information, each of these &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; outcome remains equally likely. So the required event, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;, is just one of three equally likely outcomes. It therefore seems reasonable that &amp;lt;math&amp;gt;\Pr (E|F)=\frac{1}{3}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
We shall now develop this idea more fully, using Venn Diagrams with the implied notion of area giving probability. Consider an abstract sample space, denoted by &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, with events &amp;lt;math&amp;gt;E\subset S,\,\,F\subset S&amp;lt;/math&amp;gt;. This is illustrated in the following Figure. Eventually we will want to construct the conditional probability, &amp;lt;math&amp;gt;\Pr \left( E|F\right)&amp;lt;/math&amp;gt;. Sticking with the above example that could be the probability that &amp;#039;&amp;#039;a child is underweight&amp;#039;&amp;#039;, given that &amp;#039;&amp;#039;the mother is a smoker&amp;#039;&amp;#039;. Two important areas used in the construction of this conditional probability are highlighted as &amp;lt;math&amp;gt;\mathbf{a}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{b}&amp;lt;/math&amp;gt;:&lt;br /&gt;
&lt;br /&gt;
[[File:ProbCond_venn1.jpg|frameless|400px]]&lt;br /&gt;
&lt;br /&gt;
In general, it is useful to think of &amp;lt;math&amp;gt;\Pr (E)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;\frac{area\left( E\right)}{area\left( S\right) }&amp;lt;/math&amp;gt;; and similarly for &amp;lt;math&amp;gt;\Pr (F)&amp;lt;/math&amp;gt;. The &amp;lt;math&amp;gt;\Pr (E\cap F)&amp;lt;/math&amp;gt; could equally be thought of as &amp;lt;math&amp;gt;\frac{area\left( a\right)}{area\left( S\right) }&amp;lt;/math&amp;gt;. With this in mind, consider what happens if we are now told that &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; has occurred. Incorporating this information implies that the effective sample space becomes restricted to &amp;lt;math&amp;gt;S^{*}=F&amp;lt;/math&amp;gt;, since &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; now defines what can happen. This now covers the sample area &amp;lt;math&amp;gt;a+b.&amp;lt;/math&amp;gt; On this new, restricted, sample space an outcome in &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; can only be observed if that outcome also belongs to &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;, the restricted sample space &amp;lt;math&amp;gt;S^*&amp;lt;/math&amp;gt;. And this only occurs in area &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt; which corresponds to the event &amp;lt;math&amp;gt;E\cap F&amp;lt;/math&amp;gt;. Thus the event of interest &amp;#039;&amp;#039;now&amp;#039;&amp;#039; is &amp;lt;math&amp;gt;E^{*}=E\cap F,&amp;lt;/math&amp;gt; as defined on the &amp;#039;&amp;#039;restricted &amp;#039;&amp;#039;sample space of &amp;lt;math&amp;gt;S^{*}=F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
In order to proceed with the construction of the conditional probability, &amp;lt;math&amp;gt;\Pr \left( E|F\right)&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;area(S)=z&amp;lt;/math&amp;gt;. Then, since the ratio of the area of the event of interest to that of the sample space gives probability, we have (on this restricted sample space):&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E|F) &amp;amp;=&amp;amp;\frac{area\left( E\cap F\right) }{area\left( F\right) } \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{a}{a+b} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{a/z}{\left( a+b\right) /z} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{\Pr \left( E\cap F\right) }{\Pr \left( F\right) },\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We have shown, for this example how a conditional probability can be expressed as a function of the joint probability &amp;lt;math&amp;gt;\Pr \left( E\cap F\right)&amp;lt;/math&amp;gt; and the marginal probability &amp;lt;math&amp;gt;\Pr \left( F\right)&amp;lt;/math&amp;gt;. This is a profound result and should be formulated in more general terms:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;The probability that &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; occurs, given that &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is known to have occurred, gives the &amp;#039;&amp;#039;&amp;#039;conditional probability&amp;#039;&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. This is denoted &amp;lt;math&amp;gt;Pr(E|F)&amp;lt;/math&amp;gt; and is calculated as&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (E|F)=\frac{\Pr (E\cap F)}{\Pr (F)}&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;and from the axioms of probability will generate a number lying between 0 and 1, since &amp;lt;math&amp;gt;\Pr (F)\geq \Pr (E\cap F)\geq 0.&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example: &amp;#039;&amp;#039;A Manufacturer of electrical components knows that the probability is 0.8 that an order will be ready for shipment on time and it is 0.6 that it will also be delivered on time. What is the probability that such an order will be delivered on time given that it was ready for shipment on time?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;R=&amp;lt;/math&amp;gt; READY, &amp;lt;math&amp;gt;D=&amp;lt;/math&amp;gt; DELIVERED ON TIME. &amp;lt;math&amp;gt;Pr(R)=0.8,Pr(R\cap D)=0.6.&amp;lt;/math&amp;gt; From this we need to calculate &amp;lt;math&amp;gt;Pr(D|R),&amp;lt;/math&amp;gt; using the above formula. This gives, &amp;lt;math&amp;gt;Pr(D|R)=Pr(R\cap D)/Pr(R)=6/8,\,\,&amp;lt;/math&amp;gt;or&amp;lt;math&amp;gt;\,\,75\%&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If we re-arrange the above formula for conditional probability, we obtain the so-called &amp;#039;&amp;#039;multiplication rule of probability &amp;#039;&amp;#039;for &amp;#039;&amp;#039;intersections&amp;#039;&amp;#039; of events:&lt;br /&gt;
&lt;br /&gt;
== Multiplication rule of probability ==&lt;br /&gt;
&lt;br /&gt;
The multiplication rule of probability can be stated as follows:&lt;br /&gt;
&lt;br /&gt;
* &amp;lt;math&amp;gt;\Pr (E\cap F)=\Pr (E|F)\times \Pr (F)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note that for any two events, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F,&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;(E\cap F)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;(E\cap \bar{F})&amp;lt;/math&amp;gt; are mutually exclusive with &amp;lt;math&amp;gt;E=(E\cap F)\cup (E\cap \bar{F})&amp;lt;/math&amp;gt;; this has been seen before. So the &amp;#039;&amp;#039;addition rule&amp;#039;&amp;#039; and &amp;#039;&amp;#039;multiplication rule&amp;#039;&amp;#039; of probability together give:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E) &amp;amp;=&amp;amp;\Pr (E\cap F)+\Pr (E\cap \bar{F}) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\Pr (E|F)\times \Pr (F)+\Pr (E|\bar{F})\times \Pr (\bar{F}).\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This is an extremely important and useful result, in practice, as we shall see shortly.&lt;br /&gt;
&lt;br /&gt;
=== Additional resources ===&lt;br /&gt;
&lt;br /&gt;
Khan Academy&lt;br /&gt;
&lt;br /&gt;
* Another application of this rule [https://www.khanacademy.org/math/probability/independent-dependent-probability/dependent_probability/v/dependent-probability-example-1]&lt;br /&gt;
&lt;br /&gt;
== Statistical Independence ==&lt;br /&gt;
&lt;br /&gt;
If the knowledge that &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; has occurred does NOT alter our probability assessment of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are said to be (statistically) &amp;#039;&amp;#039;independent&amp;#039;&amp;#039;. In this sense, &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; carries no information about &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Formally, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are &amp;#039;&amp;#039;&amp;#039;independent&amp;#039;&amp;#039;&amp;#039; events if and only if&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;Pr(E|F)=Pr(E)&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;which, in turn is true&amp;#039;&amp;#039; if and only if &amp;#039;&amp;#039;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;Pr(E\cap F)=Pr(E)\times Pr(F).&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This concept of independence is of enormous importance in practice. Consider the case of lung cancer and its connection to smoking (apologies to all smokers for being picked upon here). The first connection between smoking and lung cancer was made in the 1920s. However, for many decades after the tobacco industry spend a lot of money and effort to convince people that there was no connection between the two. In other words they claimed that the two events are &amp;#039;&amp;#039;independent&amp;#039;&amp;#039;, or &amp;lt;math&amp;gt;Pr(Cancer|Smoking)=Pr(Cancer|\cap{Smoking})=Pr(Cancer)&amp;lt;/math&amp;gt;. It was then the task of epidemiologists to show otherwise. This was famously and comprehensively achieved by the [http://en.wikipedia.org/wiki/British_Doctors_Study &amp;#039;&amp;#039;British Doctors Study&amp;#039;&amp;#039;].&lt;br /&gt;
&lt;br /&gt;
== Bayes’ Theorem ==&lt;br /&gt;
&lt;br /&gt;
One area where conditional probability is extremely important is that of clinical trials - testing the power of a diagnostic test to detect the presence of a particular disease. Suppose, then, that a new test is being developed and let &amp;lt;math&amp;gt;P=&amp;lt;/math&amp;gt; ‘&amp;#039;&amp;#039;test positive&amp;#039;&amp;#039;’ and &amp;lt;math&amp;gt;D=&amp;lt;/math&amp;gt; ‘&amp;#039;&amp;#039;presence of disease&amp;#039;&amp;#039;’, but where the results from applying the diagnostic test can never be wholly reliable. From the point of view of our previous discussion on conditional probability, we would of course require that &amp;lt;math&amp;gt;\Pr \left( P|D\right)&amp;lt;/math&amp;gt; to be large; i.e., the test should be effective at detecting the disease. However, if you think about, this is not necessarily the probability that we might be interested in from a diagnosis point of view. Rather, we should be more interested in &amp;lt;math&amp;gt;\Pr \left( D|P\right)&amp;lt;/math&amp;gt;, the probability of correct diagnosis, and require this to be large (with, presumably, &amp;lt;math&amp;gt;\Pr (D|\bar{P})&amp;lt;/math&amp;gt; being small). Here, what we are trying to attach a probability to is a possible ‘cause’. The observed outcome is a positive test result (&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;), but the presence or non-presence of the disease is what is of interest and this is uncertain. &amp;lt;math&amp;gt;\Pr (D|P)&amp;lt;/math&amp;gt; asks the question ‘&amp;#039;&amp;#039;what is the probability that it is the presence of the disease which caused the positive test result&amp;#039;&amp;#039;’? (Another recent news-worthy example would be the effect of exposure to depleted uranium on Gulf and Balkan war veterans. Given the presence of lymph, lung or brain cancer in such individuals (&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;), how likely is that the cause was exposure to depleted uranium weapons (&amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;)? Firstly, is &amp;lt;math&amp;gt;\Pr \left( D|P\right) &amp;lt;/math&amp;gt; high or low? Secondly, might there being something else (&amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;) which could offer a “better” explanation, such that &amp;lt;math&amp;gt;\Pr \left( F|P\right) &amp;gt;\Pr \left( D|F\right) &amp;lt;/math&amp;gt; ?)&lt;br /&gt;
&lt;br /&gt;
The situation is depicted in the following Figure, in which there are two possible ‘states’ in the population: &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt; (depicted by the lighter shaded area covering the left portion of the sample space) and &amp;lt;math&amp;gt;\bar{D}.&amp;lt;/math&amp;gt; It must be that &amp;lt;math&amp;gt;D\cup \bar{D}=S,&amp;lt;/math&amp;gt; since any individual in the population either has the disease or does not. The event of an observed positive test result is denoted by the closed loop, &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;. (Notice that the shading in the diagram is relatively darker where &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; intersects with &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&lt;br /&gt;
[[File:ProbCond_venn2.jpg|frameless|400px]]&lt;br /&gt;
&lt;br /&gt;
To investigate how we might construct the required probability, &amp;lt;math&amp;gt;\Pr \left(D|P\right)&amp;lt;/math&amp;gt;, proceed as follows:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr \left( D|P\right) &amp;amp;=&amp;amp;\frac{\Pr \left( D\cap P\right) }{\Pr (P)} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{\Pr (D\cap P)}{\Pr (P\cap D)+\Pr (P\cap \bar{D})},\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
since &amp;lt;math&amp;gt;P=(P\cap D)\cup (P\cap \bar{D}),&amp;lt;/math&amp;gt; and these are mutually exclusive. From the multiplication rule of probability, &amp;lt;math&amp;gt;\Pr \left( P\cap D\right) =\Pr(P|D)\times \Pr (D),&amp;lt;/math&amp;gt; and similarly for &amp;lt;math&amp;gt;\Pr \left( P\cap \bar{D}\right)&amp;lt;/math&amp;gt;. Thus&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\Pr \left( D|P\right) =\frac{\Pr \left( P|D\right) \times \Pr \left(D\right) }{\Pr \left( P|D\right) \times \Pr \left( D\right) +\Pr (P|\bar{D})\times \Pr \left( \bar{D}\right) },&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
which may be convenient to work with since &amp;lt;math&amp;gt;\Pr \left( P|D\right) &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr\left( P|\bar{D}\right) &amp;lt;/math&amp;gt; can be estimated from clinical trials and &amp;lt;math&amp;gt;\Pr\left( D\right) &amp;lt;/math&amp;gt; estimated from recent historical survey data.&lt;br /&gt;
&lt;br /&gt;
This sort of calculation (assigning probabilities to possible causes of observed events) is an example of &amp;#039;&amp;#039;Bayes’ Theorem&amp;#039;&amp;#039;. Of course, we may have to consider more than two possible causes, and the construction of the appropriate probabilities is as follows.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Consider a sample space, &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;A,B,C&amp;lt;/math&amp;gt; are three mutually exclusive events (possible causes), defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, such that &amp;lt;math&amp;gt;S=A\cup B\cup C&amp;lt;/math&amp;gt;. In such a situation, &amp;lt;math&amp;gt;A,B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; are said to form a &amp;#039;&amp;#039;&amp;#039;partition&amp;#039;&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;. &amp;#039;&amp;#039;&amp;#039;Bayes’ Theorem&amp;#039;&amp;#039;&amp;#039; states that:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (A|E)=\frac{\Pr (E|A)\times \Pr (A)}{\left\{ \Pr (E|A)\times \Pr(A)\right\} +\left\{ \Pr (E|B)\times \Pr (B)\right\} +\left\{ \Pr(E|C)\times \Pr (C)\right\} }.&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;And, more generally, consider a sample space, &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F_{1},F_{2},...,F_{k}&amp;lt;/math&amp;gt; are &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; mutually exclusive events (possible causes), which form a partition of &amp;lt;math&amp;gt;S:S=\bigcup_{j=1}^{k}F_{j}&amp;lt;/math&amp;gt;. &amp;#039;&amp;#039;&amp;#039;Bayes’ Theorem&amp;#039;&amp;#039;&amp;#039; then states that:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (F_{j}|E)=\frac{\Pr (E|F_{j})\times \Pr (F_{j})}{\sum_{s=1}^{k}\left\{\Pr (E|F_{s})\times \Pr (F_{s})\right\} }.&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
From the above formula, you should be able to satisfy yourself that &amp;lt;math&amp;gt;\sum_{j=1}^{k}\Pr \left( F_{j}|E\right) =1.&amp;lt;/math&amp;gt; If this is not at first clear, consider case (1) and show that &amp;lt;math&amp;gt;\Pr \left( A|E\right) +\Pr \left(&lt;br /&gt;
B|E\right) +\Pr \left( C|E\right) =1.&amp;lt;/math&amp;gt; The reason for this is that since &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; form a partition of &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; they must also form a partition of any event &amp;lt;math&amp;gt;E\subset S.&amp;lt;/math&amp;gt; In the above conditional probabilities, we are regarding &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; as the restricted sample space and therefore the probabilities assigned the mutually exclusive events &amp;lt;math&amp;gt;\left( A,B,C\right) &amp;lt;/math&amp;gt; which &amp;#039;&amp;#039;cover&amp;#039;&amp;#039; this (restricted) sample space, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, must sum to 1.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example&amp;#039;&amp;#039;: Box A contains 2 red balls. Box B contains 1 red and 1 white ball. Box A and Box B are identical. If a box is selected at random and one ball is withdrawn from it, what is the probability that the selected box was number 1 if the ball withdrawn from it turns out to be red?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; be the event of selecting Box A and &amp;lt;math&amp;gt;R&amp;lt;/math&amp;gt; the event of drawing a red ball. Require &amp;lt;math&amp;gt;Pr(A|R)&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;Pr(A|R)=Pr(A\cap R)/Pr(R);&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;Pr(A\cap R)=Pr(A)Pr(R|A)=(1/2)\times 1=1/2.&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
And,&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
Pr(R) &amp;amp;=&amp;amp;Pr(A\cap R)+Pr(\bar{A}\cap R) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\Pr (A)\times \Pr (R|A)\,\,\,\,+\,\,\,\,\Pr (\bar{A})\times \Pr (R|\bar{A}) \\&lt;br /&gt;
&amp;amp;=&amp;amp;(1/2)\,\,\,\,+\,\,\,\,(1/2)\times (1/2) \\&lt;br /&gt;
&amp;amp;=&amp;amp;3/4.\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Therefore, &amp;lt;math&amp;gt;\Pr (A|R)=(1/2)/(3/4)=2/3&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
= Additional resources =&lt;br /&gt;
&lt;br /&gt;
Khan Academy&lt;br /&gt;
&lt;br /&gt;
* A different example that intuitively leads to Bayes Formula [https://www.khanacademy.org/math/probability/independent-dependent-probability/dependent_probability/v/introduction-to-dependent-probability]&lt;br /&gt;
* Should you switch the door? A classic problem. [https://www.khanacademy.org/math/probability/independent-dependent-probability/dependent_probability/v/monty-hall-problem]&lt;br /&gt;
&lt;br /&gt;
= Exercises =&lt;br /&gt;
&lt;br /&gt;
You can find examples related to these topics here: [[Probability_Conditional_Exercises]].&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Conditional&amp;diff=2879</id>
		<title>Probability Conditional</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Conditional&amp;diff=2879"/>
				<updated>2013-08-09T10:58:12Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: /* Additional resources */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Conditional Probability =&lt;br /&gt;
&lt;br /&gt;
An important consideration in the development of probability is that of &amp;#039;&amp;#039;conditional probability&amp;#039;&amp;#039;. This refers to the calculation of updating probabilities in the light of revealed information. For example, insurance companies nearly always set their home contents insurance premiums on the basis of the postcode in which the home is located. That is to say, insurance companies believe the risk depends upon the location; i.e., the probability of property crime is assessed conditional upon the location of the property. (A similar calculation is made to set car insurance premiums.) As a result, the premiums for two identical households located in different parts of the country can differ substantially.&lt;br /&gt;
&lt;br /&gt;
* In general, the probability of an event, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, occurring &amp;#039;&amp;#039;given&amp;#039;&amp;#039; that an event, &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;, has occurred is called the &amp;#039;&amp;#039;conditional probability&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and is denoted &amp;lt;math&amp;gt;\Pr (E|F)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
As another example, it has been well documented that the ability of a new born baby to survive is closely associated with its birth-weight. A birth-weight of less than 1500&amp;#039;&amp;#039;g&amp;#039;&amp;#039; is regarded as dangerously low. Consider &amp;lt;math&amp;gt;E=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;birth weight of a baby is less than&amp;#039;&amp;#039; 1500&amp;#039;&amp;#039;g&amp;#039;&amp;#039;, &amp;lt;math&amp;gt;F=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;mother smoked during pregnancy&amp;#039;&amp;#039;; then evidence as to whether &amp;lt;math&amp;gt;\Pr(E|F)&amp;gt;\Pr (E|\bar{F})&amp;lt;/math&amp;gt; is of considerable interest.&lt;br /&gt;
&lt;br /&gt;
As a preliminary to the main development, consider the simple experiment of rolling a fair die and observing the number of dots on the upturned face. Then &amp;lt;math&amp;gt;S=\left\{ 1,2,3,4,5,6\right\} &amp;lt;/math&amp;gt; and define events, &amp;lt;math&amp;gt;E=\left\{4\right\} &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F=\left\{ 4,5,6\right\} ;&amp;lt;/math&amp;gt; we are interested in &amp;lt;math&amp;gt;\Pr \left(E|F\right)&amp;lt;/math&amp;gt;. To work this out we take &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; as known. Given this knowledge the sample space becomes restricted to simply &amp;lt;math&amp;gt;\left\{ 4,5,6\right\} &amp;lt;/math&amp;gt; and, given no other information, each of these &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; outcome remains equally likely. So the required event, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;, is just one of three equally likely outcomes. It therefore seems reasonable that &amp;lt;math&amp;gt;\Pr (E|F)=\frac{1}{3}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
We shall now develop this idea more fully, using Venn Diagrams with the implied notion of area giving probability. Consider an abstract sample space, denoted by &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, with events &amp;lt;math&amp;gt;E\subset S,\,\,F\subset S&amp;lt;/math&amp;gt;. This is illustrated in the following Figure. Eventually we will want to construct the conditional probability, &amp;lt;math&amp;gt;\Pr \left( E|F\right)&amp;lt;/math&amp;gt;. Sticking with the above example that could be the probability that &amp;#039;&amp;#039;a child is underweight&amp;#039;&amp;#039;, given that &amp;#039;&amp;#039;the mother is a smoker&amp;#039;&amp;#039;. Two important areas used in the construction of this conditional probability are highlighted as &amp;lt;math&amp;gt;\mathbf{a}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{b}&amp;lt;/math&amp;gt;:&lt;br /&gt;
&lt;br /&gt;
[[File:ProbCond_venn1.jpg|frameless|400px]]&lt;br /&gt;
&lt;br /&gt;
In general, it is useful to think of &amp;lt;math&amp;gt;\Pr (E)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;\frac{area\left( E\right)}{area\left( S\right) }&amp;lt;/math&amp;gt;; and similarly for &amp;lt;math&amp;gt;\Pr (F)&amp;lt;/math&amp;gt;. The &amp;lt;math&amp;gt;\Pr (E\cap F)&amp;lt;/math&amp;gt; could equally be thought of as &amp;lt;math&amp;gt;\frac{area\left( a\right)}{area\left( S\right) }&amp;lt;/math&amp;gt;. With this in mind, consider what happens if we are now told that &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; has occurred. Incorporating this information implies that the effective sample space becomes restricted to &amp;lt;math&amp;gt;S^{*}=F&amp;lt;/math&amp;gt;, since &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; now defines what can happen. This now covers the sample area &amp;lt;math&amp;gt;a+b.&amp;lt;/math&amp;gt; On this new, restricted, sample space an outcome in &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; can only be observed if that outcome also belongs to &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;, the restricted sample space &amp;lt;math&amp;gt;S^*&amp;lt;/math&amp;gt;. And this only occurs in area &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt; which corresponds to the event &amp;lt;math&amp;gt;E\cap F&amp;lt;/math&amp;gt;. Thus the event of interest &amp;#039;&amp;#039;now&amp;#039;&amp;#039; is &amp;lt;math&amp;gt;E^{*}=E\cap F,&amp;lt;/math&amp;gt; as defined on the &amp;#039;&amp;#039;restricted &amp;#039;&amp;#039;sample space of &amp;lt;math&amp;gt;S^{*}=F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
In order to proceed with the construction of the conditional probability, &amp;lt;math&amp;gt;\Pr \left( E|F\right)&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;area(S)=z&amp;lt;/math&amp;gt;. Then, since the ratio of the area of the event of interest to that of the sample space gives probability, we have (on this restricted sample space):&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E|F) &amp;amp;=&amp;amp;\frac{area\left( E\cap F\right) }{area\left( F\right) } \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{a}{a+b} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{a/z}{\left( a+b\right) /z} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{\Pr \left( E\cap F\right) }{\Pr \left( F\right) },\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We have shown, for this example how a conditional probability can be expressed as a function of the joint probability &amp;lt;math&amp;gt;\Pr \left( E\cap F\right)&amp;lt;/math&amp;gt; and the marginal probability &amp;lt;math&amp;gt;\Pr \left( F\right)&amp;lt;/math&amp;gt;. This is a profound result and should be formulated in more general terms:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;The probability that &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; occurs, given that &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is known to have occurred, gives the &amp;#039;&amp;#039;&amp;#039;conditional probability&amp;#039;&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. This is denoted &amp;lt;math&amp;gt;Pr(E|F)&amp;lt;/math&amp;gt; and is calculated as&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (E|F)=\frac{\Pr (E\cap F)}{\Pr (F)}&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;and from the axioms of probability will generate a number lying between 0 and 1, since &amp;lt;math&amp;gt;\Pr (F)\geq \Pr (E\cap F)\geq 0.&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example: &amp;#039;&amp;#039;A Manufacturer of electrical components knows that the probability is 0.8 that an order will be ready for shipment on time and it is 0.6 that it will also be delivered on time. What is the probability that such an order will be delivered on time given that it was ready for shipment on time?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;R=&amp;lt;/math&amp;gt; READY, &amp;lt;math&amp;gt;D=&amp;lt;/math&amp;gt; DELIVERED ON TIME. &amp;lt;math&amp;gt;Pr(R)=0.8,Pr(R\cap D)=0.6.&amp;lt;/math&amp;gt; From this we need to calculate &amp;lt;math&amp;gt;Pr(D|R),&amp;lt;/math&amp;gt; using the above formula. This gives, &amp;lt;math&amp;gt;Pr(D|R)=Pr(R\cap D)/Pr(R)=6/8,\,\,&amp;lt;/math&amp;gt;or&amp;lt;math&amp;gt;\,\,75\%&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If we re-arrange the above formula for conditional probability, we obtain the so-called &amp;#039;&amp;#039;multiplication rule of probability &amp;#039;&amp;#039;for &amp;#039;&amp;#039;intersections&amp;#039;&amp;#039; of events:&lt;br /&gt;
&lt;br /&gt;
== Multiplication rule of probability ==&lt;br /&gt;
&lt;br /&gt;
The multiplication rule of probability can be stated as follows:&lt;br /&gt;
&lt;br /&gt;
* &amp;lt;math&amp;gt;\Pr (E\cap F)=\Pr (E|F)\times \Pr (F)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note that for any two events, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F,&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;(E\cap F)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;(E\cap \bar{F})&amp;lt;/math&amp;gt; are mutually exclusive with &amp;lt;math&amp;gt;E=(E\cap F)\cup (E\cap \bar{F})&amp;lt;/math&amp;gt;; this has been seen before. So the &amp;#039;&amp;#039;addition rule&amp;#039;&amp;#039; and &amp;#039;&amp;#039;multiplication rule&amp;#039;&amp;#039; of probability together give:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E) &amp;amp;=&amp;amp;\Pr (E\cap F)+\Pr (E\cap \bar{F}) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\Pr (E|F)\times \Pr (F)+\Pr (E|\bar{F})\times \Pr (\bar{F}).\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This is an extremely important and useful result, in practice, as we shall see shortly.&lt;br /&gt;
&lt;br /&gt;
=== Additional resources ===&lt;br /&gt;
&lt;br /&gt;
Khan Academy&lt;br /&gt;
&lt;br /&gt;
* Another application of this rule [https://www.khanacademy.org/math/probability/independent-dependent-probability/dependent_probability/v/dependent-probability-example-1]&lt;br /&gt;
&lt;br /&gt;
== Statistical Independence ==&lt;br /&gt;
&lt;br /&gt;
If the knowledge that &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; has occurred does NOT alter our probability assessment of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are said to be (statistically) &amp;#039;&amp;#039;independent&amp;#039;&amp;#039;. In this sense, &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; carries no information about &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Formally, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are &amp;#039;&amp;#039;&amp;#039;independent&amp;#039;&amp;#039;&amp;#039; events if and only if&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;Pr(E|F)=Pr(E)&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;which, in turn is true&amp;#039;&amp;#039; if and only if &amp;#039;&amp;#039;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;Pr(E\cap F)=Pr(E)\times Pr(F).&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This concept of independence is of enormous importance in practice. Consider the case of lung cancer and its connection to smoking (apologies to all smokers for being picked upon here). The first connection between smoking and lung cancer was made in the 1920s. However, for many decades after the tobacco industry spend a lot of money and effort to convince people that there was no connection between the two. In other words they claimed that the two events are &amp;#039;&amp;#039;independent&amp;#039;&amp;#039;, or &amp;lt;math&amp;gt;Pr(Cancer|Smoking)=Pr(Cancer|\cap{Smoking})=Pr(Cancer)&amp;lt;/math&amp;gt;. It was then the task of epidemiologists to show otherwise. This was famously and comprehensively achieved by the [http://en.wikipedia.org/wiki/British_Doctors_Study &amp;#039;&amp;#039;British Doctors Study&amp;#039;&amp;#039;].&lt;br /&gt;
&lt;br /&gt;
== Bayes’ Theorem ==&lt;br /&gt;
&lt;br /&gt;
One area where conditional probability is extremely important is that of clinical trials - testing the power of a diagnostic test to detect the presence of a particular disease. Suppose, then, that a new test is being developed and let &amp;lt;math&amp;gt;P=&amp;lt;/math&amp;gt; ‘&amp;#039;&amp;#039;test positive&amp;#039;&amp;#039;’ and &amp;lt;math&amp;gt;D=&amp;lt;/math&amp;gt; ‘&amp;#039;&amp;#039;presence of disease&amp;#039;&amp;#039;’, but where the results from applying the diagnostic test can never be wholly reliable. From the point of view of our previous discussion on conditional probability, we would of course require that &amp;lt;math&amp;gt;\Pr \left( P|D\right)&amp;lt;/math&amp;gt; to be large; i.e., the test should be effective at detecting the disease. However, if you think about, this is not necessarily the probability that we might be interested in from a diagnosis point of view. Rather, we should be more interested in &amp;lt;math&amp;gt;\Pr \left( D|P\right)&amp;lt;/math&amp;gt;, the probability of correct diagnosis, and require this to be large (with, presumably, &amp;lt;math&amp;gt;\Pr (D|\bar{P})&amp;lt;/math&amp;gt; being small). Here, what we are trying to attach a probability to is a possible ‘cause’. The observed outcome is a positive test result (&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;), but the presence or non-presence of the disease is what is of interest and this is uncertain. &amp;lt;math&amp;gt;\Pr (D|P)&amp;lt;/math&amp;gt; asks the question ‘&amp;#039;&amp;#039;what is the probability that it is the presence of the disease which caused the positive test result&amp;#039;&amp;#039;’? (Another recent news-worthy example would be the effect of exposure to depleted uranium on Gulf and Balkan war veterans. Given the presence of lymph, lung or brain cancer in such individuals (&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;), how likely is that the cause was exposure to depleted uranium weapons (&amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;)? Firstly, is &amp;lt;math&amp;gt;\Pr \left( D|P\right) &amp;lt;/math&amp;gt; high or low? Secondly, might there being something else (&amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;) which could offer a “better” explanation, such that &amp;lt;math&amp;gt;\Pr \left( F|P\right) &amp;gt;\Pr \left( D|F\right) &amp;lt;/math&amp;gt; ?)&lt;br /&gt;
&lt;br /&gt;
The situation is depicted in the following Figure, in which there are two possible ‘states’ in the population: &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt; (depicted by the lighter shaded area covering the left portion of the sample space) and &amp;lt;math&amp;gt;\bar{D}.&amp;lt;/math&amp;gt; It must be that &amp;lt;math&amp;gt;D\cup \bar{D}=S,&amp;lt;/math&amp;gt; since any individual in the population either has the disease or does not. The event of an observed positive test result is denoted by the closed loop, &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;. (Notice that the shading in the diagram is relatively darker where &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; intersects with &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&lt;br /&gt;
[[File:ProbCond_venn2.jpg|frameless|400px]]&lt;br /&gt;
&lt;br /&gt;
To investigate how we might construct the required probability, &amp;lt;math&amp;gt;\Pr \left(D|P\right)&amp;lt;/math&amp;gt;, proceed as follows:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr \left( D|P\right) &amp;amp;=&amp;amp;\frac{\Pr \left( D\cap P\right) }{\Pr (P)} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{\Pr (D\cap P)}{\Pr (P\cap D)+\Pr (P\cap \bar{D})},\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
since &amp;lt;math&amp;gt;P=(P\cap D)\cup (P\cap \bar{D}),&amp;lt;/math&amp;gt; and these are mutually exclusive. From the multiplication rule of probability, &amp;lt;math&amp;gt;\Pr \left( P\cap D\right) =\Pr(P|D)\times \Pr (D),&amp;lt;/math&amp;gt; and similarly for &amp;lt;math&amp;gt;\Pr \left( P\cap \bar{D}\right)&amp;lt;/math&amp;gt;. Thus&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\Pr \left( D|P\right) =\frac{\Pr \left( P|D\right) \times \Pr \left(D\right) }{\Pr \left( P|D\right) \times \Pr \left( D\right) +\Pr (P|\bar{D})\times \Pr \left( \bar{D}\right) },&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
which may be convenient to work with since &amp;lt;math&amp;gt;\Pr \left( P|D\right) &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr\left( P|\bar{D}\right) &amp;lt;/math&amp;gt; can be estimated from clinical trials and &amp;lt;math&amp;gt;\Pr\left( D\right) &amp;lt;/math&amp;gt; estimated from recent historical survey data.&lt;br /&gt;
&lt;br /&gt;
This sort of calculation (assigning probabilities to possible causes of observed events) is an example of &amp;#039;&amp;#039;Bayes’ Theorem&amp;#039;&amp;#039;. Of course, we may have to consider more than two possible causes, and the construction of the appropriate probabilities is as follows.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Consider a sample space, &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;A,B,C&amp;lt;/math&amp;gt; are three mutually exclusive events (possible causes), defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, such that &amp;lt;math&amp;gt;S=A\cup B\cup C&amp;lt;/math&amp;gt;. In such a situation, &amp;lt;math&amp;gt;A,B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; are said to form a &amp;#039;&amp;#039;&amp;#039;partition&amp;#039;&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;. &amp;#039;&amp;#039;&amp;#039;Bayes’ Theorem&amp;#039;&amp;#039;&amp;#039; states that:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (A|E)=\frac{\Pr (E|A)\times \Pr (A)}{\left\{ \Pr (E|A)\times \Pr(A)\right\} +\left\{ \Pr (E|B)\times \Pr (B)\right\} +\left\{ \Pr(E|C)\times \Pr (C)\right\} }.&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;And, more generally, consider a sample space, &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F_{1},F_{2},...,F_{k}&amp;lt;/math&amp;gt; are &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; mutually exclusive events (possible causes), which form a partition of &amp;lt;math&amp;gt;S:S=\bigcup_{j=1}^{k}F_{j}&amp;lt;/math&amp;gt;. &amp;#039;&amp;#039;&amp;#039;Bayes’ Theorem&amp;#039;&amp;#039;&amp;#039; then states that:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (F_{j}|E)=\frac{\Pr (E|F_{j})\times \Pr (F_{j})}{\sum_{s=1}^{k}\left\{\Pr (E|F_{s})\times \Pr (F_{s})\right\} }.&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
From the above formula, you should be able to satisfy yourself that &amp;lt;math&amp;gt;\sum_{j=1}^{k}\Pr \left( F_{j}|E\right) =1.&amp;lt;/math&amp;gt; If this is not at first clear, consider case (1) and show that &amp;lt;math&amp;gt;\Pr \left( A|E\right) +\Pr \left(&lt;br /&gt;
B|E\right) +\Pr \left( C|E\right) =1.&amp;lt;/math&amp;gt; The reason for this is that since &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; form a partition of &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; they must also form a partition of any event &amp;lt;math&amp;gt;E\subset S.&amp;lt;/math&amp;gt; In the above conditional probabilities, we are regarding &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; as the restricted sample space and therefore the probabilities assigned the mutually exclusive events &amp;lt;math&amp;gt;\left( A,B,C\right) &amp;lt;/math&amp;gt; which &amp;#039;&amp;#039;cover&amp;#039;&amp;#039; this (restricted) sample space, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, must sum to 1.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example&amp;#039;&amp;#039;: Box A contains 2 red balls. Box B contains 1 red and 1 white ball. Box A and Box B are identical. If a box is selected at random and one ball is withdrawn from it, what is the probability that the selected box was number 1 if the ball withdrawn from it turns out to be red?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; be the event of selecting Box A and &amp;lt;math&amp;gt;R&amp;lt;/math&amp;gt; the event of drawing a red ball. Require &amp;lt;math&amp;gt;Pr(A|R)&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;Pr(A|R)=Pr(A\cap R)/Pr(R);&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;Pr(A\cap R)=Pr(A)Pr(R|A)=(1/2)\times 1=1/2.&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
And,&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
Pr(R) &amp;amp;=&amp;amp;Pr(A\cap R)+Pr(\bar{A}\cap R) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\Pr (A)\times \Pr (R|A)\,\,\,\,+\,\,\,\,\Pr (\bar{A})\times \Pr (R|\bar{A}) \\&lt;br /&gt;
&amp;amp;=&amp;amp;(1/2)\,\,\,\,+\,\,\,\,(1/2)\times (1/2) \\&lt;br /&gt;
&amp;amp;=&amp;amp;3/4.\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Therefore, &amp;lt;math&amp;gt;\Pr (A|R)=(1/2)/(3/4)=2/3&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
= Additional resources =&lt;br /&gt;
&lt;br /&gt;
Khan Academy&lt;br /&gt;
&lt;br /&gt;
* A different example that intuitively leads to Bayes Formula [https://www.khanacademy.org/math/probability/independent-dependent-probability/dependent_probability/v/introduction-to-dependent-probability]&lt;br /&gt;
* Should you switch the door? A classic problem. [https://www.khanacademy.org/math/probability/independent-dependent-probability/dependent_probability/v/monty-hall-problem]&lt;br /&gt;
&lt;br /&gt;
= Exercises =&lt;br /&gt;
&lt;br /&gt;
You can find examples related to these topics here: [[Probability&amp;lt;sub&amp;gt;C&amp;lt;/sub&amp;gt;onditional&amp;lt;sub&amp;gt;E&amp;lt;/sub&amp;gt;xercises]].&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Conditional&amp;diff=2878</id>
		<title>Probability Conditional</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Conditional&amp;diff=2878"/>
				<updated>2013-08-09T10:56:46Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Conditional Probability =&lt;br /&gt;
&lt;br /&gt;
An important consideration in the development of probability is that of &amp;#039;&amp;#039;conditional probability&amp;#039;&amp;#039;. This refers to the calculation of updating probabilities in the light of revealed information. For example, insurance companies nearly always set their home contents insurance premiums on the basis of the postcode in which the home is located. That is to say, insurance companies believe the risk depends upon the location; i.e., the probability of property crime is assessed conditional upon the location of the property. (A similar calculation is made to set car insurance premiums.) As a result, the premiums for two identical households located in different parts of the country can differ substantially.&lt;br /&gt;
&lt;br /&gt;
* In general, the probability of an event, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, occurring &amp;#039;&amp;#039;given&amp;#039;&amp;#039; that an event, &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;, has occurred is called the &amp;#039;&amp;#039;conditional probability&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and is denoted &amp;lt;math&amp;gt;\Pr (E|F)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
As another example, it has been well documented that the ability of a new born baby to survive is closely associated with its birth-weight. A birth-weight of less than 1500&amp;#039;&amp;#039;g&amp;#039;&amp;#039; is regarded as dangerously low. Consider &amp;lt;math&amp;gt;E=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;birth weight of a baby is less than&amp;#039;&amp;#039; 1500&amp;#039;&amp;#039;g&amp;#039;&amp;#039;, &amp;lt;math&amp;gt;F=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;mother smoked during pregnancy&amp;#039;&amp;#039;; then evidence as to whether &amp;lt;math&amp;gt;\Pr(E|F)&amp;gt;\Pr (E|\bar{F})&amp;lt;/math&amp;gt; is of considerable interest.&lt;br /&gt;
&lt;br /&gt;
As a preliminary to the main development, consider the simple experiment of rolling a fair die and observing the number of dots on the upturned face. Then &amp;lt;math&amp;gt;S=\left\{ 1,2,3,4,5,6\right\} &amp;lt;/math&amp;gt; and define events, &amp;lt;math&amp;gt;E=\left\{4\right\} &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F=\left\{ 4,5,6\right\} ;&amp;lt;/math&amp;gt; we are interested in &amp;lt;math&amp;gt;\Pr \left(E|F\right)&amp;lt;/math&amp;gt;. To work this out we take &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; as known. Given this knowledge the sample space becomes restricted to simply &amp;lt;math&amp;gt;\left\{ 4,5,6\right\} &amp;lt;/math&amp;gt; and, given no other information, each of these &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; outcome remains equally likely. So the required event, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;, is just one of three equally likely outcomes. It therefore seems reasonable that &amp;lt;math&amp;gt;\Pr (E|F)=\frac{1}{3}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
We shall now develop this idea more fully, using Venn Diagrams with the implied notion of area giving probability. Consider an abstract sample space, denoted by &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, with events &amp;lt;math&amp;gt;E\subset S,\,\,F\subset S&amp;lt;/math&amp;gt;. This is illustrated in the following Figure. Eventually we will want to construct the conditional probability, &amp;lt;math&amp;gt;\Pr \left( E|F\right)&amp;lt;/math&amp;gt;. Sticking with the above example that could be the probability that &amp;#039;&amp;#039;a child is underweight&amp;#039;&amp;#039;, given that &amp;#039;&amp;#039;the mother is a smoker&amp;#039;&amp;#039;. Two important areas used in the construction of this conditional probability are highlighted as &amp;lt;math&amp;gt;\mathbf{a}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{b}&amp;lt;/math&amp;gt;:&lt;br /&gt;
&lt;br /&gt;
[[File:ProbCond_venn1.jpg|frameless|400px]]&lt;br /&gt;
&lt;br /&gt;
In general, it is useful to think of &amp;lt;math&amp;gt;\Pr (E)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;\frac{area\left( E\right)}{area\left( S\right) }&amp;lt;/math&amp;gt;; and similarly for &amp;lt;math&amp;gt;\Pr (F)&amp;lt;/math&amp;gt;. The &amp;lt;math&amp;gt;\Pr (E\cap F)&amp;lt;/math&amp;gt; could equally be thought of as &amp;lt;math&amp;gt;\frac{area\left( a\right)}{area\left( S\right) }&amp;lt;/math&amp;gt;. With this in mind, consider what happens if we are now told that &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; has occurred. Incorporating this information implies that the effective sample space becomes restricted to &amp;lt;math&amp;gt;S^{*}=F&amp;lt;/math&amp;gt;, since &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; now defines what can happen. This now covers the sample area &amp;lt;math&amp;gt;a+b.&amp;lt;/math&amp;gt; On this new, restricted, sample space an outcome in &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; can only be observed if that outcome also belongs to &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;, the restricted sample space &amp;lt;math&amp;gt;S^*&amp;lt;/math&amp;gt;. And this only occurs in area &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt; which corresponds to the event &amp;lt;math&amp;gt;E\cap F&amp;lt;/math&amp;gt;. Thus the event of interest &amp;#039;&amp;#039;now&amp;#039;&amp;#039; is &amp;lt;math&amp;gt;E^{*}=E\cap F,&amp;lt;/math&amp;gt; as defined on the &amp;#039;&amp;#039;restricted &amp;#039;&amp;#039;sample space of &amp;lt;math&amp;gt;S^{*}=F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
In order to proceed with the construction of the conditional probability, &amp;lt;math&amp;gt;\Pr \left( E|F\right)&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;area(S)=z&amp;lt;/math&amp;gt;. Then, since the ratio of the area of the event of interest to that of the sample space gives probability, we have (on this restricted sample space):&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E|F) &amp;amp;=&amp;amp;\frac{area\left( E\cap F\right) }{area\left( F\right) } \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{a}{a+b} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{a/z}{\left( a+b\right) /z} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{\Pr \left( E\cap F\right) }{\Pr \left( F\right) },\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We have shown, for this example how a conditional probability can be expressed as a function of the joint probability &amp;lt;math&amp;gt;\Pr \left( E\cap F\right)&amp;lt;/math&amp;gt; and the marginal probability &amp;lt;math&amp;gt;\Pr \left( F\right)&amp;lt;/math&amp;gt;. This is a profound result and should be formulated in more general terms:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;The probability that &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; occurs, given that &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is known to have occurred, gives the &amp;#039;&amp;#039;&amp;#039;conditional probability&amp;#039;&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. This is denoted &amp;lt;math&amp;gt;Pr(E|F)&amp;lt;/math&amp;gt; and is calculated as&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (E|F)=\frac{\Pr (E\cap F)}{\Pr (F)}&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;and from the axioms of probability will generate a number lying between 0 and 1, since &amp;lt;math&amp;gt;\Pr (F)\geq \Pr (E\cap F)\geq 0.&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example: &amp;#039;&amp;#039;A Manufacturer of electrical components knows that the probability is 0.8 that an order will be ready for shipment on time and it is 0.6 that it will also be delivered on time. What is the probability that such an order will be delivered on time given that it was ready for shipment on time?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;R=&amp;lt;/math&amp;gt; READY, &amp;lt;math&amp;gt;D=&amp;lt;/math&amp;gt; DELIVERED ON TIME. &amp;lt;math&amp;gt;Pr(R)=0.8,Pr(R\cap D)=0.6.&amp;lt;/math&amp;gt; From this we need to calculate &amp;lt;math&amp;gt;Pr(D|R),&amp;lt;/math&amp;gt; using the above formula. This gives, &amp;lt;math&amp;gt;Pr(D|R)=Pr(R\cap D)/Pr(R)=6/8,\,\,&amp;lt;/math&amp;gt;or&amp;lt;math&amp;gt;\,\,75\%&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If we re-arrange the above formula for conditional probability, we obtain the so-called &amp;#039;&amp;#039;multiplication rule of probability &amp;#039;&amp;#039;for &amp;#039;&amp;#039;intersections&amp;#039;&amp;#039; of events:&lt;br /&gt;
&lt;br /&gt;
== Multiplication rule of probability ==&lt;br /&gt;
&lt;br /&gt;
The multiplication rule of probability can be stated as follows:&lt;br /&gt;
&lt;br /&gt;
* &amp;lt;math&amp;gt;\Pr (E\cap F)=\Pr (E|F)\times \Pr (F)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note that for any two events, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F,&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;(E\cap F)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;(E\cap \bar{F})&amp;lt;/math&amp;gt; are mutually exclusive with &amp;lt;math&amp;gt;E=(E\cap F)\cup (E\cap \bar{F})&amp;lt;/math&amp;gt;; this has been seen before. So the &amp;#039;&amp;#039;addition rule&amp;#039;&amp;#039; and &amp;#039;&amp;#039;multiplication rule&amp;#039;&amp;#039; of probability together give:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E) &amp;amp;=&amp;amp;\Pr (E\cap F)+\Pr (E\cap \bar{F}) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\Pr (E|F)\times \Pr (F)+\Pr (E|\bar{F})\times \Pr (\bar{F}).\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This is an extremely important and useful result, in practice, as we shall see shortly.&lt;br /&gt;
&lt;br /&gt;
=== Additional resources ===&lt;br /&gt;
&lt;br /&gt;
Khan Academy&lt;br /&gt;
&lt;br /&gt;
* Another application of this rule [https://www.khanacademy.org/math/probability/independent-dependent-probability/dependent_probability/v/dependent-probability-example-1]&lt;br /&gt;
&lt;br /&gt;
== Statistical Independence ==&lt;br /&gt;
&lt;br /&gt;
If the knowledge that &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; has occurred does NOT alter our probability assessment of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are said to be (statistically) &amp;#039;&amp;#039;independent&amp;#039;&amp;#039;. In this sense, &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; carries no information about &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Formally, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are &amp;#039;&amp;#039;&amp;#039;independent&amp;#039;&amp;#039;&amp;#039; events if and only if&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;Pr(E|F)=Pr(E)&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;which, in turn is true&amp;#039;&amp;#039; if and only if &amp;#039;&amp;#039;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;Pr(E\cap F)=Pr(E)\times Pr(F).&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This concept of independence is of enormous importance in practice. Consider the case of lung cancer and its connection to smoking (apologies to all smokers for being picked upon here). The first connection between smoking and lung cancer was made in the 1920s. However, for many decades after the tobacco industry spend a lot of money and effort to convince people that there was no connection between the two. In other words they claimed that the two events are &amp;#039;&amp;#039;independent&amp;#039;&amp;#039;, or &amp;lt;math&amp;gt;Pr(Cancer|Smoking)=Pr(Cancer|\cap{Smoking})=Pr(Cancer)&amp;lt;/math&amp;gt;. It was then the task of epidemiologists to show otherwise. This was famously and comprehensively achieved by the [http://en.wikipedia.org/wiki/British_Doctors_Study &amp;#039;&amp;#039;British Doctors Study&amp;#039;&amp;#039;].&lt;br /&gt;
&lt;br /&gt;
=== Additional resources ===&lt;br /&gt;
&lt;br /&gt;
Khan Academy&lt;br /&gt;
&lt;br /&gt;
* A different example that intuitively leads to Bayes Formula [https://www.khanacademy.org/math/probability/independent-dependent-probability/dependent_probability/v/introduction-to-dependent-probability]&lt;br /&gt;
&lt;br /&gt;
== Bayes’ Theorem ==&lt;br /&gt;
&lt;br /&gt;
One area where conditional probability is extremely important is that of clinical trials - testing the power of a diagnostic test to detect the presence of a particular disease. Suppose, then, that a new test is being developed and let &amp;lt;math&amp;gt;P=&amp;lt;/math&amp;gt; ‘&amp;#039;&amp;#039;test positive&amp;#039;&amp;#039;’ and &amp;lt;math&amp;gt;D=&amp;lt;/math&amp;gt; ‘&amp;#039;&amp;#039;presence of disease&amp;#039;&amp;#039;’, but where the results from applying the diagnostic test can never be wholly reliable. From the point of view of our previous discussion on conditional probability, we would of course require that &amp;lt;math&amp;gt;\Pr \left( P|D\right)&amp;lt;/math&amp;gt; to be large; i.e., the test should be effective at detecting the disease. However, if you think about, this is not necessarily the probability that we might be interested in from a diagnosis point of view. Rather, we should be more interested in &amp;lt;math&amp;gt;\Pr \left( D|P\right)&amp;lt;/math&amp;gt;, the probability of correct diagnosis, and require this to be large (with, presumably, &amp;lt;math&amp;gt;\Pr (D|\bar{P})&amp;lt;/math&amp;gt; being small). Here, what we are trying to attach a probability to is a possible ‘cause’. The observed outcome is a positive test result (&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;), but the presence or non-presence of the disease is what is of interest and this is uncertain. &amp;lt;math&amp;gt;\Pr (D|P)&amp;lt;/math&amp;gt; asks the question ‘&amp;#039;&amp;#039;what is the probability that it is the presence of the disease which caused the positive test result&amp;#039;&amp;#039;’? (Another recent news-worthy example would be the effect of exposure to depleted uranium on Gulf and Balkan war veterans. Given the presence of lymph, lung or brain cancer in such individuals (&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;), how likely is that the cause was exposure to depleted uranium weapons (&amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;)? Firstly, is &amp;lt;math&amp;gt;\Pr \left( D|P\right) &amp;lt;/math&amp;gt; high or low? Secondly, might there being something else (&amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;) which could offer a “better” explanation, such that &amp;lt;math&amp;gt;\Pr \left( F|P\right) &amp;gt;\Pr \left( D|F\right) &amp;lt;/math&amp;gt; ?)&lt;br /&gt;
&lt;br /&gt;
The situation is depicted in the following Figure, in which there are two possible ‘states’ in the population: &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt; (depicted by the lighter shaded area covering the left portion of the sample space) and &amp;lt;math&amp;gt;\bar{D}.&amp;lt;/math&amp;gt; It must be that &amp;lt;math&amp;gt;D\cup \bar{D}=S,&amp;lt;/math&amp;gt; since any individual in the population either has the disease or does not. The event of an observed positive test result is denoted by the closed loop, &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;. (Notice that the shading in the diagram is relatively darker where &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; intersects with &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&lt;br /&gt;
[[File:ProbCond_venn2.jpg|frameless|400px]]&lt;br /&gt;
&lt;br /&gt;
To investigate how we might construct the required probability, &amp;lt;math&amp;gt;\Pr \left(D|P\right)&amp;lt;/math&amp;gt;, proceed as follows:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr \left( D|P\right) &amp;amp;=&amp;amp;\frac{\Pr \left( D\cap P\right) }{\Pr (P)} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{\Pr (D\cap P)}{\Pr (P\cap D)+\Pr (P\cap \bar{D})},\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
since &amp;lt;math&amp;gt;P=(P\cap D)\cup (P\cap \bar{D}),&amp;lt;/math&amp;gt; and these are mutually exclusive. From the multiplication rule of probability, &amp;lt;math&amp;gt;\Pr \left( P\cap D\right) =\Pr(P|D)\times \Pr (D),&amp;lt;/math&amp;gt; and similarly for &amp;lt;math&amp;gt;\Pr \left( P\cap \bar{D}\right)&amp;lt;/math&amp;gt;. Thus&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\Pr \left( D|P\right) =\frac{\Pr \left( P|D\right) \times \Pr \left(D\right) }{\Pr \left( P|D\right) \times \Pr \left( D\right) +\Pr (P|\bar{D})\times \Pr \left( \bar{D}\right) },&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
which may be convenient to work with since &amp;lt;math&amp;gt;\Pr \left( P|D\right) &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr\left( P|\bar{D}\right) &amp;lt;/math&amp;gt; can be estimated from clinical trials and &amp;lt;math&amp;gt;\Pr\left( D\right) &amp;lt;/math&amp;gt; estimated from recent historical survey data.&lt;br /&gt;
&lt;br /&gt;
This sort of calculation (assigning probabilities to possible causes of observed events) is an example of &amp;#039;&amp;#039;Bayes’ Theorem&amp;#039;&amp;#039;. Of course, we may have to consider more than two possible causes, and the construction of the appropriate probabilities is as follows.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Consider a sample space, &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;A,B,C&amp;lt;/math&amp;gt; are three mutually exclusive events (possible causes), defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, such that &amp;lt;math&amp;gt;S=A\cup B\cup C&amp;lt;/math&amp;gt;. In such a situation, &amp;lt;math&amp;gt;A,B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; are said to form a &amp;#039;&amp;#039;&amp;#039;partition&amp;#039;&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;. &amp;#039;&amp;#039;&amp;#039;Bayes’ Theorem&amp;#039;&amp;#039;&amp;#039; states that:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (A|E)=\frac{\Pr (E|A)\times \Pr (A)}{\left\{ \Pr (E|A)\times \Pr(A)\right\} +\left\{ \Pr (E|B)\times \Pr (B)\right\} +\left\{ \Pr(E|C)\times \Pr (C)\right\} }.&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;And, more generally, consider a sample space, &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F_{1},F_{2},...,F_{k}&amp;lt;/math&amp;gt; are &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; mutually exclusive events (possible causes), which form a partition of &amp;lt;math&amp;gt;S:S=\bigcup_{j=1}^{k}F_{j}&amp;lt;/math&amp;gt;. &amp;#039;&amp;#039;&amp;#039;Bayes’ Theorem&amp;#039;&amp;#039;&amp;#039; then states that:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (F_{j}|E)=\frac{\Pr (E|F_{j})\times \Pr (F_{j})}{\sum_{s=1}^{k}\left\{\Pr (E|F_{s})\times \Pr (F_{s})\right\} }.&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
From the above formula, you should be able to satisfy yourself that &amp;lt;math&amp;gt;\sum_{j=1}^{k}\Pr \left( F_{j}|E\right) =1.&amp;lt;/math&amp;gt; If this is not at first clear, consider case (1) and show that &amp;lt;math&amp;gt;\Pr \left( A|E\right) +\Pr \left(&lt;br /&gt;
B|E\right) +\Pr \left( C|E\right) =1.&amp;lt;/math&amp;gt; The reason for this is that since &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; form a partition of &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; they must also form a partition of any event &amp;lt;math&amp;gt;E\subset S.&amp;lt;/math&amp;gt; In the above conditional probabilities, we are regarding &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; as the restricted sample space and therefore the probabilities assigned the mutually exclusive events &amp;lt;math&amp;gt;\left( A,B,C\right) &amp;lt;/math&amp;gt; which &amp;#039;&amp;#039;cover&amp;#039;&amp;#039; this (restricted) sample space, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, must sum to 1.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example&amp;#039;&amp;#039;: Box A contains 2 red balls. Box B contains 1 red and 1 white ball. Box A and Box B are identical. If a box is selected at random and one ball is withdrawn from it, what is the probability that the selected box was number 1 if the ball withdrawn from it turns out to be red?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; be the event of selecting Box A and &amp;lt;math&amp;gt;R&amp;lt;/math&amp;gt; the event of drawing a red ball. Require &amp;lt;math&amp;gt;Pr(A|R)&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;Pr(A|R)=Pr(A\cap R)/Pr(R);&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;Pr(A\cap R)=Pr(A)Pr(R|A)=(1/2)\times 1=1/2.&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
And,&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
Pr(R) &amp;amp;=&amp;amp;Pr(A\cap R)+Pr(\bar{A}\cap R) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\Pr (A)\times \Pr (R|A)\,\,\,\,+\,\,\,\,\Pr (\bar{A})\times \Pr (R|\bar{A}) \\&lt;br /&gt;
&amp;amp;=&amp;amp;(1/2)\,\,\,\,+\,\,\,\,(1/2)\times (1/2) \\&lt;br /&gt;
&amp;amp;=&amp;amp;3/4.\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Therefore, &amp;lt;math&amp;gt;\Pr (A|R)=(1/2)/(3/4)=2/3&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
= Additional resources =&lt;br /&gt;
&lt;br /&gt;
Khan Academy&lt;br /&gt;
&lt;br /&gt;
* A different example that intuitively leads to Bayes Formula [https://www.khanacademy.org/math/probability/independent-dependent-probability/dependent_probability/v/introduction-to-dependent-probability]&lt;br /&gt;
* Should you switch the door? A classic problem. [https://www.khanacademy.org/math/probability/independent-dependent-probability/dependent_probability/v/monty-hall-problem]&lt;br /&gt;
&lt;br /&gt;
= Exercises =&lt;br /&gt;
&lt;br /&gt;
You can find examples related to these topics here: [[Probability&amp;lt;sub&amp;gt;C&amp;lt;/sub&amp;gt;onditional&amp;lt;sub&amp;gt;E&amp;lt;/sub&amp;gt;xercises]].&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2877</id>
		<title>Probability Intro</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2877"/>
				<updated>2013-08-09T10:33:23Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Introducing Probability =&lt;br /&gt;
&lt;br /&gt;
So far we have been looking at ways of summarising samples of data drawn from an underlying population of interest. Although at times tedious, all such arithmetic calculations are fairly mechanical and straightforward to apply. To remind ourselves, one of the primary reasons for wishing to summarise data is so assist in the development of inferences about the population from which the data were taken. That is to say, we would like to elicit some information about the mechanism which generated the observed data.&lt;br /&gt;
&lt;br /&gt;
We now start on the process of developing mathematical ways of formulating inferences and this requires the use of &amp;#039;&amp;#039;probability&amp;#039;&amp;#039;. This becomes clear if we think back to one of the early questions posed in this course: &amp;#039;&amp;#039;prior to sampling is it possible to predict with absolute certainty what will be observed&amp;#039;&amp;#039;? The answer to this question is &amp;#039;&amp;#039;no&amp;#039;&amp;#039;; although it would be of interest to know how &amp;#039;&amp;#039;likely&amp;#039;&amp;#039; it is that certain values would be observed. Or, what is the &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; of observing certain values?&lt;br /&gt;
&lt;br /&gt;
Before proceeding, we need some more tools:&lt;br /&gt;
&lt;br /&gt;
= Venn diagrams =&lt;br /&gt;
&lt;br /&gt;
Venn diagrams (and diagrams in general) are of enormous help in trying to understand, and manipulate probability. We begin with some basic definitions, some of which we have encountered before.&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Experiment:&amp;#039;&amp;#039;&amp;#039; any process which, when applied, provides data or an outcome; e.g., rolling a die and observing the number of dots on the upturned face; recording the amount of rainfall in Manchester over a period of time.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Sample Space:&amp;#039;&amp;#039;&amp;#039; set of possible outcomes of an experiment; e.g., &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; (or &amp;lt;math&amp;gt;\Omega &amp;lt;/math&amp;gt;) &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{1,2,3,4,5,6\}&amp;lt;/math&amp;gt;, which is the sample space of rolling a dice. Or &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{x;x\geq 0\}&amp;lt;/math&amp;gt;, which is the sample space of an experiment where the outcomes can be any real non-negative number, or ‘&amp;#039;&amp;#039;the set of real non-negative real numbers&amp;#039;&amp;#039;’.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Simple Event&amp;#039;&amp;#039;&amp;#039;: just one of the possible outcomes on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Event:&amp;#039;&amp;#039;&amp;#039; a &amp;#039;&amp;#039;subset&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, denoted &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;; e.g., &amp;lt;math&amp;gt;E=\left\{ 2,4,6\right\}&amp;lt;/math&amp;gt; (i.e. any even number on a dice) or &amp;lt;math&amp;gt;E=\left\{ x;4&amp;lt;x\leq 10\right\} ,&amp;lt;/math&amp;gt; which means ‘&amp;#039;&amp;#039;the set of real numbers which are strictly bigger than&amp;#039;&amp;#039; &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;but less than or equal to &amp;#039;&amp;#039;&amp;lt;math&amp;gt;10&amp;lt;/math&amp;gt;’.&amp;lt;br /&amp;gt;Note that an event, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, is a collection of simple events.&lt;br /&gt;
&lt;br /&gt;
Such concepts can be represented by means of the following Venn Diagram:&lt;br /&gt;
&lt;br /&gt;
[[File:Venn_1.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
The sample space, &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; is depicted as a closed rectangle, and the event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; is a closed loop wholly contained within &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; and we write (in set notation) &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
In dealing with probability, and in particular the probability of an event (or events) occurring, we shall need to be familiar with &amp;#039;&amp;#039;&amp;#039;UNIONS, INTERSECTIONS&amp;#039;&amp;#039;&amp;#039; and &amp;#039;&amp;#039;&amp;#039;COMPLEMENTS&amp;#039;&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
To illustrate these concepts, consider the sample space &amp;lt;math&amp;gt;S=\{x;x\geq 0\},\,&amp;lt;/math&amp;gt; with the following events defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as depicted in Figure 3.2:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E=\{x;4&amp;lt;x\leq 10\},\,F=\{x;7&amp;lt;x\leq 17\},\,G=\{x;x&amp;gt;15\},\,H=\{x;9&amp;lt;x\leq&lt;br /&gt;
13\}. &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| (a) Event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: A closed loop&lt;br /&gt;
| (b) Union: &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2a.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2b.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (c) Intersection: &amp;lt;math&amp;gt;E\cap F&amp;lt;/math&amp;gt;&lt;br /&gt;
| (d) The Null set/event: &amp;lt;math&amp;gt;E\cap G=\emptyset &amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2c.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2d.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (e) Complement of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;&lt;br /&gt;
| (f) Subset of &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;H\subset F&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;H\cap F=H&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2e.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2f.jpg|frameless|300px]]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
        &lt;br /&gt;
&lt;br /&gt;
* The &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cup F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cup F=\{x;4&amp;lt;x\leq 17\};&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are either in &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or in &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) in both. This is illustrated on the Venn diagram by the dark shaded area in diagram (b).&lt;br /&gt;
* The&amp;#039;&amp;#039; intersection&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cap F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cap F=\left\{ x;7\leq x\leq 10\right\} ;&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are common to both &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; Again this is depicted by the dark shaded area in (c). If events have no elements in common (as, for example, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt;) then they are said to be &amp;#039;&amp;#039;mutually exclusive&amp;#039;&amp;#039;, and we can write &amp;lt;math&amp;gt;E\cap G=\emptyset ,&amp;lt;/math&amp;gt; meaning the &amp;#039;&amp;#039;null set&amp;#039;&amp;#039; which contains no elements. Such a situation is illustrated on the Venn Diagram by events (the two shaded closed loops in (d)) which do not overlap. Notice however that &amp;lt;math&amp;gt;G\cap F\neq \emptyset ,&amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have elements in common.&lt;br /&gt;
* The &amp;#039;&amp;#039;complement&amp;#039;&amp;#039; of an event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; say, is everything defined on the sample space which is not in &amp;lt;math&amp;gt;E.&amp;lt;/math&amp;gt; This event is denoted &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;, the dark shaded area in (e); here &amp;lt;math&amp;gt;\bar{E}=\left\{ x;x\leq 4\right\} \cup \left\{ x;x&amp;gt;10\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
* Finally note that &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; is a sub-set of &amp;lt;math&amp;gt;F;&amp;lt;/math&amp;gt; see (f). It is depicted as the dark closed loop wholly contained within &amp;lt;math&amp;gt;F,&amp;lt;/math&amp;gt; the lighter shaded area, so that &amp;lt;math&amp;gt;H\cap F=H;&amp;lt;/math&amp;gt; if an element in the sample space is a member of &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; then it must also be member of &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; (In mathematical logic, we employ this scenario to indicate that “&amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; implies &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;”, but not necessarily vice-versa.) Notice that &amp;lt;math&amp;gt;G\cap H=\emptyset &amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;H\cap E\neq \emptyset&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
= Probability =&lt;br /&gt;
&lt;br /&gt;
The term &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; (or some equivalent) is used in everyday conversation and so can not be unfamiliar to the reader. We talk of the probability, or chance, of rain; the likelihood of England winning the World Cup; or, perhaps more scientifically, the chance of getting a &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; when rolling a die. What we shall now do is develop a coherent theory of probability; a theory which allows us to combine and manipulate probabilities in a consistent and meaningful manner. We shall describe ways of dealing with, and describing, uncertainty. This will involve &amp;#039;&amp;#039;rules&amp;#039;&amp;#039; which govern our use of terms like probability.&lt;br /&gt;
&lt;br /&gt;
There have been a number of different approaches (interpretations) of probability. Most depend, at least to some extent, on the notion of relative frequency as now described:&lt;br /&gt;
&lt;br /&gt;
* Suppose an experiment has an outcome of interest &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;. The &amp;#039;&amp;#039;relative frequency interpretation&amp;#039;&amp;#039; of probability says that assuming the experiment can be repeated a large number of times then the relative frequency of observing the outcome &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; will settle down to a &amp;#039;&amp;#039;number&amp;#039;&amp;#039;, denoted &amp;lt;math&amp;gt;\Pr (E),&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;P(E)&amp;lt;/math&amp;gt; or Prob&amp;lt;math&amp;gt;(E),&amp;lt;/math&amp;gt; called the &amp;#039;&amp;#039;&amp;#039;probability&amp;#039;&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
This is illustrated in the next Figure where the proportion of heads obtained after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; flips of a fair coin is plotted against &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;, as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; increases; e.g., of the first 100 flips, 46 were heads (&amp;lt;math&amp;gt;46\%&amp;lt;/math&amp;gt;). Notice that the plot becomes less ‘wobbly’ after about &amp;lt;math&amp;gt;n=&amp;lt;/math&amp;gt;140 and appears to be settling down to the value of &amp;lt;math&amp;gt;\frac{1}{2}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
[[File:Prob_coin.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
Due to this interpretation of probability, we often use observed sample proportions to approximate underlying probabilities of interest; see, for example, Question 4 of Exercise 2. There are, of course, other interpretations of probability; e.g., the subjective interpretation which simply expresses the strength of one’s belief about an event of interest such as whether Manchester United will win the European Cup! Any one of these interpretations can be used in practical situations provided the implied notion of probability follows a simple set of &amp;#039;&amp;#039;axioms&amp;#039;&amp;#039; or &amp;#039;&amp;#039;rules&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
== The axioms of probability ==&lt;br /&gt;
&lt;br /&gt;
There are just &amp;#039;&amp;#039;three &amp;#039;&amp;#039;basic rules that must be obeyed when dealing with probabilities:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;For any event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; i.e., &amp;lt;math&amp;gt;E\subset S,\,\,\Pr (E)\geq 0&amp;lt;/math&amp;gt;; &amp;#039;&amp;#039;probabilities are non-negative&amp;#039;&amp;#039;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (S)=1;&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;having defined the sample space of outcomes, one of these outcomes must be observed&amp;#039;&amp;#039;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;If events &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are mutually exclusive defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, so that &amp;lt;math&amp;gt;E\cap F=\emptyset &amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\Pr \left( E\right)+\Pr \left( F\right) .&amp;lt;/math&amp;gt; In general, for any set of mutually exclusive events, &amp;lt;math&amp;gt;E_{1},E_{2},\ldots ,E_{k},&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S:&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (E_{1}\cup E_{2}\cup \ldots \cup E_{k})=\Pr (E_{1})+\Pr (E_{2})+\ldots\Pr (E_{k})&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;i.e., &amp;lt;math&amp;gt;\Pr \left( \bigcup_{j=1}^{k}E_{j}\right) =\sum_{j=1}^{k}\Pr (E_{j}).&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In terms of the Venn Diagram, one can (and should) usefully think of the area of &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; relative to that of &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as providing an indication of probability. (Note, from axiom 2, that the area of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is implicitly normalised to be unity).&lt;br /&gt;
&lt;br /&gt;
Also observe that, contrary to what you may have believed, it is not one of the rules that &amp;lt;math&amp;gt;\Pr (E)\leq 1&amp;lt;/math&amp;gt; for any event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;. Rather, this is an implication of the &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; rules given:&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;implications: &amp;#039;&amp;#039;&amp;#039;it must be that for any event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;E\cap \bar{E}=\emptyset &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;E\cup \bar{E}=S.&amp;lt;/math&amp;gt; By Axiom &amp;lt;math&amp;gt;1,&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr (E)\geq 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr \left( \bar{E}\right) \geq 0&amp;lt;/math&amp;gt; and by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr(E)+\Pr (\bar{E})=\Pr (S).&amp;lt;/math&amp;gt; So &amp;lt;math&amp;gt;\Pr \left( E\right) +\Pr \left( \bar{E}\right) =1,&amp;lt;/math&amp;gt; by Axiom &amp;lt;math&amp;gt;2.&amp;lt;/math&amp;gt; This implies that&lt;br /&gt;
&lt;br /&gt;
# &amp;lt;math&amp;gt;0\leq \Pr (E)\leq 1&amp;lt;/math&amp;gt;&lt;br /&gt;
# &amp;lt;math&amp;gt;\Pr (\bar{E})=1-\Pr (E)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The first of these is what we might have expected from probability (a number lying between &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;). The second implication is also very important; it says that the probability of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; not happening is ‘&amp;#039;&amp;#039;one minus the probability of it happening&amp;#039;&amp;#039;’. Thus when rolling a die, the probability of getting &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; is one minus the probability of getting either a &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;2&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;5.&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
These axioms imply how to calculate probabilities on a sample space of equally likely outcomes. For example, and as we have already noted, the experiment of rolling a fair die defines a sample space of six, mutually exclusive and equally likely outcomes (&amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; dots on the up-turned face). The axioms then say that each of the six probabilities are positive, add to 1 and are all the same. Thus, the probability of any one of the outcomes must be simply &amp;lt;math&amp;gt;\frac{1}{6};&amp;lt;/math&amp;gt; which may accord with your intuition. A similar sort of analysis reveals that the probability of drawing a club from a deck of &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; cards is &amp;lt;math&amp;gt;\frac{13}{52},&amp;lt;/math&amp;gt; since any one of the &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; cards has an equal chance of being drawn and &amp;lt;math&amp;gt;13&amp;lt;/math&amp;gt; of them are clubs; i.e., &amp;lt;math&amp;gt;13&amp;lt;/math&amp;gt; of the &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; are clubs, so the probability of drawing a club is &amp;lt;math&amp;gt;\frac{13}{52}.&amp;lt;/math&amp;gt; Notice the importance of the assumption of equally likely outcomes here.&lt;br /&gt;
&lt;br /&gt;
In this, and the next section of notes, we shall see how these axioms can be used. Firstly, consider the construction of a probability for the &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of two events; i.e., the probability that &amp;#039;&amp;#039;either &amp;#039;&amp;#039;&amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) &amp;#039;&amp;#039;both &amp;#039;&amp;#039;will occur. Such a probability is embodied in the &amp;#039;&amp;#039;addition rule of probability&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
== The addition rule of probability ==&lt;br /&gt;
&lt;br /&gt;
When rolling a fair die, let &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; denote the event of an “odd number of dots” and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; the event of the “number of dots being greater than, or equal, to &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;”&amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt; What is the probability of the event &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;? To calculate this we can collect together all the mutually exclusive (simple) events which comprise &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;, and then add up the probabilities (by axiom 3). These simple events are &amp;lt;math&amp;gt;1,3,4,5&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; dots. Each has a probability of &amp;lt;math&amp;gt;\frac{1}{6},&amp;lt;/math&amp;gt; so the required total probability is: &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\frac{5}{6}&amp;lt;/math&amp;gt;. Consider carefully how this probability is constructed and note, in particular, that &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) \neq \Pr \left( E\right) +\Pr \left( F\right) &amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have a simple event in common (namely &amp;lt;math&amp;gt;5&amp;lt;/math&amp;gt; dots).&lt;br /&gt;
&lt;br /&gt;
In general, we can calculate the probability of the union of events using the &amp;#039;&amp;#039;addition rule of probability&amp;#039;&amp;#039;, as follows.&lt;br /&gt;
&lt;br /&gt;
* For any events, &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F\subset S:\Pr (E\cup F)=\Pr (E)+\Pr (F)-\Pr (E\cap F).&amp;lt;/math&amp;gt; So, in general, &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) \leq \Pr (E)+\Pr (F).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This generalises to three events, &amp;lt;math&amp;gt;E_{1},E_{2}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;E_{3}&amp;lt;/math&amp;gt; as&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E_{1}\cup E_{2}\cup E_{3}) &amp;amp;=&amp;amp;\Pr (E_{1})+\Pr (E_{2})+\Pr (E_{3}) \\&lt;br /&gt;
&amp;amp;&amp;amp;-\Pr (E_{1}\cap E_{2})-\Pr (E_{1}\cap E_{3})-\Pr (E_{2}\cap E_{3}) \\&lt;br /&gt;
&amp;amp;&amp;amp;+\Pr (E_{1}\cap E_{2}\cap E_{3}).\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We can demonstrate this as follows.&lt;br /&gt;
&lt;br /&gt;
Note that&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E\cup F=\left( E\cap \bar{F}\right) \cup \left( E\cap F\right) \cup \left(\bar{E}\cap F\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
the union of &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; mutually exclusive events. These mutually exclusive events are depicted by the shaded areas &amp;lt;math&amp;gt;\mathbf{a,}&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\mathbf{b}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{c}&amp;lt;/math&amp;gt;, respectively, in the next Figure.&lt;br /&gt;
&lt;br /&gt;
[[File:Prob_add.jpg|frameless|500px]]&lt;br /&gt;
&lt;br /&gt;
Then by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, and from the fact that the three events &amp;lt;math&amp;gt;\left( E\cap\bar{F}\right) &amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\left( E\cap F\right) &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\left( \bar{E}\cap F\right)&amp;lt;/math&amp;gt; are mutually exclusive so that the “area” occupied by &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt; is simply &amp;lt;math&amp;gt;\mathbf{a+b+c,}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\Pr \left( E\cap \bar{F}\right) +\Pr \left( \bar{E}\cap F\right) +\Pr \left( E\cap F\right) .&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
But also by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, since &amp;lt;math&amp;gt;E=\left( E\cap \bar{F}\right) \cup \left(E\cap F\right) &amp;lt;/math&amp;gt;, it must be that &amp;lt;math&amp;gt;\Pr (E)=\Pr \left( E\cap \bar{F}\right)+\Pr (E\cap F);&amp;lt;/math&amp;gt; similarly, &amp;lt;math&amp;gt;\Pr \left( \bar{E}\cap F\right) =\Pr \left(F\right) -\Pr \left( E\cap F\right)&amp;lt;/math&amp;gt;. Putting all of this together gives&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\Pr (E\cup F)=\Pr (E)+\Pr (F)-\Pr (E\cap F).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
When &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are mutually exclusive, so that &amp;lt;math&amp;gt;E\cap F=\emptyset&amp;lt;/math&amp;gt;, this rule reveals Axiom 2: &amp;lt;math&amp;gt;\Pr (E\cup F)=\Pr (E)+\Pr (F)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example: &amp;#039;&amp;#039;What is the probability of drawing a Queen (&amp;lt;math&amp;gt;Q &amp;lt;/math&amp;gt;) or a Club (&amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;) in a single draw from a pack of cards? Now, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;52 &amp;lt;/math&amp;gt; cards are Queens, so &amp;lt;math&amp;gt;\Pr \left( Q\right) =\frac{4}{52},&amp;lt;/math&amp;gt; whilst &amp;lt;math&amp;gt;\Pr\left( C\right) =\frac{13}{52}.&amp;lt;/math&amp;gt; The probability of drawing the Queen of Clubs is simply &amp;lt;math&amp;gt;\frac{1}{52};&amp;lt;/math&amp;gt; i.e., &amp;lt;math&amp;gt;\Pr \left( Q\cap C\right) =\frac{1}{52}&amp;lt;/math&amp;gt;. What we require is a Club or a Queen, for which the probability is&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr \left( Q\cup C\right) &amp;amp;=&amp;amp;\Pr \left( Q\right) +\Pr \left( C\right) -\Pr\left( Q\cap C\right) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{4}{52}+\frac{13}{52}-\frac{1}{52} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{16}{52}=\frac{4}{13}.\end{aligned}&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example: &amp;#039;&amp;#039;Consider a car journey from Manchester to London via the M6 and M1. Let &amp;lt;math&amp;gt;E=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;heavy traffic somewhere on route&amp;#039;&amp;#039; and &amp;lt;math&amp;gt;F=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;roadworks somewhere on route&amp;#039;&amp;#039;. It is estimated that &amp;lt;math&amp;gt;\Pr (E)=0.8&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr (F)=0.4,&amp;lt;/math&amp;gt; whilst the probability of NOT encountering both is &amp;lt;math&amp;gt;\Pr (\overline{E\cap F})=0.6.&amp;lt;/math&amp;gt; What is the probability of encountering heavytraffic or roadworks?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;We require &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) .&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E\cup F) &amp;amp;=&amp;amp;\Pr (E)+\Pr (F)-\Pr (E\cap F) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\Pr (E)+\Pr (F)-(1-\Pr (\overline{E\cap F})) \\&lt;br /&gt;
&amp;amp;=&amp;amp;0.8+0.4-1+0.6 \\&lt;br /&gt;
&amp;amp;=&amp;amp;0.8=\Pr (E)\end{aligned}&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Notice that this implies, in this case, that &amp;lt;math&amp;gt;F\subset E&amp;lt;/math&amp;gt; (why?). This &amp;#039;&amp;#039;model &amp;#039;&amp;#039;then implies that when there are roadworks somewhere on route you are bound to encounter heavy traffic; on the other hand, you can encounter heavy traffic on route without ever passing through roadworks. (My own experience of this motorway inclines me towards this implication!)&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Similar concepts apply when manipulating proportions as follows:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example&amp;#039;&amp;#039;: A sample of 1000 undergraduates were asked whether they took either Mathematics, Physics or Chemistry at A-level. The following responses were obtained: 100 just took Mathematics; 70 just took Physics; 100 just took Chemistry; 150 took Mathematics and Physics, but not Chemistry; 40 took Mathematics and Chemistry, but not Physics; and, 240 took Physics and Chemistry, but not Mathematics. What proportion took all three?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;This can be addressed with the following diagram:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;[[File:Prob_Alevels.jpg|frameless|500px]]&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;The shaded area contains the number who took all three, which can be deduced from the above information (since the total of the numbers assigned to each part of the Venn diagram must be &amp;lt;math&amp;gt;1000&amp;lt;/math&amp;gt;). The answer is therefore &amp;lt;math&amp;gt;30\%&amp;lt;/math&amp;gt; (being &amp;lt;math&amp;gt;300&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;1000&amp;lt;/math&amp;gt;).&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Two further results on unions, intersections and complements which are of use (and which are fairly easy to demonstrate using Venn diagrams) are &amp;#039;&amp;#039;&amp;#039;de Morgan Laws&amp;#039;&amp;#039;&amp;#039;:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\left( \bar{A}\cap \bar{B}\right) =\left( \overline{A\cup B}\right) &amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\bar{A}\cup \bar{B}=\left( \overline{A\cap B}\right) &amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
= Additional resources =&lt;br /&gt;
&lt;br /&gt;
Khan Academy&lt;br /&gt;
&lt;br /&gt;
* Basic Probability and Venn Diagram [https://www.khanacademy.org/math/probability/independent-dependent-probability/addition_rule_probability/v/probability-with-playing-cards-and-venn-diagrams]&lt;br /&gt;
* Addition Rule [https://www.khanacademy.org/math/probability/independent-dependent-probability/addition_rule_probability/v/addition-rule-for-probability]&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Conditional&amp;diff=2876</id>
		<title>Probability Conditional</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Conditional&amp;diff=2876"/>
				<updated>2013-08-09T10:16:45Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: /* Exercises */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Conditional Probability =&lt;br /&gt;
&lt;br /&gt;
An important consideration in the development of probability is that of &amp;#039;&amp;#039;conditional probability&amp;#039;&amp;#039;. This refers to the calculation of updating probabilities in the light of revealed information. For example, insurance companies nearly always set their home contents insurance premiums on the basis of the postcode in which the home is located. That is to say, insurance companies believe the risk depends upon the location; i.e., the probability of property crime is assessed conditional upon the location of the property. (A similar calculation is made to set car insurance premiums.) As a result, the premiums for two identical households located in different parts of the country can differ substantially.&lt;br /&gt;
&lt;br /&gt;
* In general, the probability of an event, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, occurring &amp;#039;&amp;#039;given&amp;#039;&amp;#039; that an event, &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;, has occurred is called the &amp;#039;&amp;#039;conditional probability&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and is denoted &amp;lt;math&amp;gt;\Pr (E|F)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
As another example, it has been well documented that the ability of a new born baby to survive is closely associated with its birth-weight. A birth-weight of less than 1500&amp;#039;&amp;#039;g&amp;#039;&amp;#039; is regarded as dangerously low. Consider &amp;lt;math&amp;gt;E=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;birth weight of a baby is less than&amp;#039;&amp;#039; 1500&amp;#039;&amp;#039;g&amp;#039;&amp;#039;, &amp;lt;math&amp;gt;F=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;mother smoked during pregnancy&amp;#039;&amp;#039;; then evidence as to whether &amp;lt;math&amp;gt;\Pr(E|F)&amp;gt;\Pr (E|\bar{F})&amp;lt;/math&amp;gt; is of considerable interest.&lt;br /&gt;
&lt;br /&gt;
As a preliminary to the main development, consider the simple experiment of rolling a fair die and observing the number of dots on the upturned face. Then &amp;lt;math&amp;gt;S=\left\{ 1,2,3,4,5,6\right\} &amp;lt;/math&amp;gt; and define events, &amp;lt;math&amp;gt;E=\left\{4\right\} &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F=\left\{ 4,5,6\right\} ;&amp;lt;/math&amp;gt; we are interested in &amp;lt;math&amp;gt;\Pr \left(E|F\right)&amp;lt;/math&amp;gt;. To work this out we take &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; as known. Given this knowledge the sample space becomes restricted to simply &amp;lt;math&amp;gt;\left\{ 4,5,6\right\} &amp;lt;/math&amp;gt; and, given no other information, each of these &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; outcome remains equally likely. So the required event, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;, is just one of three equally likely outcomes. It therefore seems reasonable that &amp;lt;math&amp;gt;\Pr (E|F)=\frac{1}{3}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
We shall now develop this idea more fully, using Venn Diagrams with the implied notion of area giving probability. Consider an abstract sample space, denoted by &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, with events &amp;lt;math&amp;gt;E\subset S,\,\,F\subset S&amp;lt;/math&amp;gt;. This is illustrated in the following Figure. Eventually we will want to construct the conditional probability, &amp;lt;math&amp;gt;\Pr \left( E|F\right)&amp;lt;/math&amp;gt;. Sticking with the above example that could be the probability that &amp;#039;&amp;#039;a child is underweight&amp;#039;&amp;#039;, given that &amp;#039;&amp;#039;the mother is a smoker&amp;#039;&amp;#039;. Two important areas used in the construction of this conditional probability are highlighted as &amp;lt;math&amp;gt;\mathbf{a}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{b}&amp;lt;/math&amp;gt;:&lt;br /&gt;
&lt;br /&gt;
[[File:ProbCond_venn1.jpg|frameless|400px]]&lt;br /&gt;
&lt;br /&gt;
In general, it is useful to think of &amp;lt;math&amp;gt;\Pr (E)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;\frac{area\left( E\right)}{area\left( S\right) }&amp;lt;/math&amp;gt;; and similarly for &amp;lt;math&amp;gt;\Pr (F)&amp;lt;/math&amp;gt;. The &amp;lt;math&amp;gt;\Pr (E\cap F)&amp;lt;/math&amp;gt; could equally be thought of as &amp;lt;math&amp;gt;\frac{area\left( a\right)}{area\left( S\right) }&amp;lt;/math&amp;gt;. With this in mind, consider what happens if we are now told that &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; has occurred. Incorporating this information implies that the effective sample space becomes restricted to &amp;lt;math&amp;gt;S^{*}=F&amp;lt;/math&amp;gt;, since &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; now defines what can happen. This now covers the sample area &amp;lt;math&amp;gt;a+b.&amp;lt;/math&amp;gt; On this new, restricted, sample space an outcome in &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; can only be observed if that outcome also belongs to &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;, the restricted sample space &amp;lt;math&amp;gt;S^*&amp;lt;/math&amp;gt;. And this only occurs in area &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt; which corresponds to the event &amp;lt;math&amp;gt;E\cap F&amp;lt;/math&amp;gt;. Thus the event of interest &amp;#039;&amp;#039;now&amp;#039;&amp;#039; is &amp;lt;math&amp;gt;E^{*}=E\cap F,&amp;lt;/math&amp;gt; as defined on the &amp;#039;&amp;#039;restricted &amp;#039;&amp;#039;sample space of &amp;lt;math&amp;gt;S^{*}=F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
In order to proceed with the construction of the conditional probability, &amp;lt;math&amp;gt;\Pr \left( E|F\right)&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;area(S)=z&amp;lt;/math&amp;gt;. Then, since the ratio of the area of the event of interest to that of the sample space gives probability, we have (on this restricted sample space):&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E|F) &amp;amp;=&amp;amp;\frac{area\left( E\cap F\right) }{area\left( F\right) } \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{a}{a+b} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{a/z}{\left( a+b\right) /z} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{\Pr \left( E\cap F\right) }{\Pr \left( F\right) },\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We have shown, for this example how a conditional probability can be expressed as a function of the joint probability &amp;lt;math&amp;gt;\Pr \left( E\cap F\right)&amp;lt;/math&amp;gt; and the marginal probability &amp;lt;math&amp;gt;\Pr \left( F\right)&amp;lt;/math&amp;gt;. This is a profound result and should be formulated in more general terms:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;The probability that &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; occurs, given that &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is known to have occurred, gives the &amp;#039;&amp;#039;&amp;#039;conditional probability&amp;#039;&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. This is denoted &amp;lt;math&amp;gt;Pr(E|F)&amp;lt;/math&amp;gt; and is calculated as&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (E|F)=\frac{\Pr (E\cap F)}{\Pr (F)}&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;and from the axioms of probability will generate a number lying between 0 and 1, since &amp;lt;math&amp;gt;\Pr (F)\geq \Pr (E\cap F)\geq 0.&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example: &amp;#039;&amp;#039;A Manufacturer of electrical components knows that the probability is 0.8 that an order will be ready for shipment on time and it is 0.6 that it will also be delivered on time. What is the probability that such an order will be delivered on time given that it was ready for shipment on time?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;R=&amp;lt;/math&amp;gt; READY, &amp;lt;math&amp;gt;D=&amp;lt;/math&amp;gt; DELIVERED ON TIME. &amp;lt;math&amp;gt;Pr(R)=0.8,Pr(R\cap D)=0.6.&amp;lt;/math&amp;gt; From this we need to calculate &amp;lt;math&amp;gt;Pr(D|R),&amp;lt;/math&amp;gt; using the above formula. This gives, &amp;lt;math&amp;gt;Pr(D|R)=Pr(R\cap D)/Pr(R)=6/8,\,\,&amp;lt;/math&amp;gt;or&amp;lt;math&amp;gt;\,\,75\%&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If we re-arrange the above formula for conditional probability, we obtain the so-called &amp;#039;&amp;#039;multiplication rule of probability &amp;#039;&amp;#039;for &amp;#039;&amp;#039;intersections&amp;#039;&amp;#039; of events:&lt;br /&gt;
&lt;br /&gt;
== Multiplication rule of probability ==&lt;br /&gt;
&lt;br /&gt;
The multiplication rule of probability can be stated as follows:&lt;br /&gt;
&lt;br /&gt;
* &amp;lt;math&amp;gt;\Pr (E\cap F)=\Pr (E|F)\times \Pr (F)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note that for any two events, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F,&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;(E\cap F)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;(E\cap \bar{F})&amp;lt;/math&amp;gt; are mutually exclusive with &amp;lt;math&amp;gt;E=(E\cap F)\cup (E\cap \bar{F})&amp;lt;/math&amp;gt;; this has been seen before. So the &amp;#039;&amp;#039;addition rule&amp;#039;&amp;#039; and &amp;#039;&amp;#039;multiplication rule&amp;#039;&amp;#039; of probability together give:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E) &amp;amp;=&amp;amp;\Pr (E\cap F)+\Pr (E\cap \bar{F}) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\Pr (E|F)\times \Pr (F)+\Pr (E|\bar{F})\times \Pr (\bar{F}).\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This is an extremely important and useful result, in practice, as we shall see shortly.&lt;br /&gt;
&lt;br /&gt;
== Statistical Independence ==&lt;br /&gt;
&lt;br /&gt;
If the knowledge that &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; has occurred does NOT alter our probability assessment of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are said to be (statistically) &amp;#039;&amp;#039;independent&amp;#039;&amp;#039;. In this sense, &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; carries no information about &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Formally, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are &amp;#039;&amp;#039;&amp;#039;independent&amp;#039;&amp;#039;&amp;#039; events if and only if&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;Pr(E|F)=Pr(E)&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;which, in turn is true&amp;#039;&amp;#039; if and only if &amp;#039;&amp;#039;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;Pr(E\cap F)=Pr(E)\times Pr(F).&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This concept of independence is of enormous importance in practice. Consider the case of lung cancer and its connection to smoking (apologies to all smokers for being picked upon here). The first connection between smoking and lung cancer was made in the 1920s. However, for many decades after the tobacco industry spend a lot of money and effort to convince people that there was no connection between the two. In other words they claimed that the two events are &amp;#039;&amp;#039;independent&amp;#039;&amp;#039;, or &amp;lt;math&amp;gt;Pr(Cancer|Smoking)=Pr(Cancer|\cap{Smoking})=Pr(Cancer)&amp;lt;/math&amp;gt;. It was then the task of epidemiologists to show otherwise. This was famously and comprehensively achieved by the [http://en.wikipedia.org/wiki/British_Doctors_Study &amp;#039;&amp;#039;British Doctors Study&amp;#039;&amp;#039;].&lt;br /&gt;
&lt;br /&gt;
== Bayes’ Theorem ==&lt;br /&gt;
&lt;br /&gt;
One area where conditional probability is extremely important is that of clinical trials - testing the power of a diagnostic test to detect the presence of a particular disease. Suppose, then, that a new test is being developed and let &amp;lt;math&amp;gt;P=&amp;lt;/math&amp;gt; ‘&amp;#039;&amp;#039;test positive&amp;#039;&amp;#039;’ and &amp;lt;math&amp;gt;D=&amp;lt;/math&amp;gt; ‘&amp;#039;&amp;#039;presence of disease&amp;#039;&amp;#039;’, but where the results from applying the diagnostic test can never be wholly reliable. From the point of view of our previous discussion on conditional probability, we would of course require that &amp;lt;math&amp;gt;\Pr \left( P|D\right)&amp;lt;/math&amp;gt; to be large; i.e., the test should be effective at detecting the disease. However, if you think about, this is not necessarily the probability that we might be interested in from a diagnosis point of view. Rather, we should be more interested in &amp;lt;math&amp;gt;\Pr \left( D|P\right)&amp;lt;/math&amp;gt;, the probability of correct diagnosis, and require this to be large (with, presumably, &amp;lt;math&amp;gt;\Pr (D|\bar{P})&amp;lt;/math&amp;gt; being small). Here, what we are trying to attach a probability to is a possible ‘cause’. The observed outcome is a positive test result (&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;), but the presence or non-presence of the disease is what is of interest and this is uncertain. &amp;lt;math&amp;gt;\Pr (D|P)&amp;lt;/math&amp;gt; asks the question ‘&amp;#039;&amp;#039;what is the probability that it is the presence of the disease which caused the positive test result&amp;#039;&amp;#039;’? (Another recent news-worthy example would be the effect of exposure to depleted uranium on Gulf and Balkan war veterans. Given the presence of lymph, lung or brain cancer in such individuals (&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;), how likely is that the cause was exposure to depleted uranium weapons (&amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;)? Firstly, is &amp;lt;math&amp;gt;\Pr \left( D|P\right) &amp;lt;/math&amp;gt; high or low? Secondly, might there being something else (&amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;) which could offer a “better” explanation, such that &amp;lt;math&amp;gt;\Pr \left( F|P\right) &amp;gt;\Pr \left( D|F\right) &amp;lt;/math&amp;gt; ?)&lt;br /&gt;
&lt;br /&gt;
The situation is depicted in the following Figure, in which there are two possible ‘states’ in the population: &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt; (depicted by the lighter shaded area covering the left portion of the sample space) and &amp;lt;math&amp;gt;\bar{D}.&amp;lt;/math&amp;gt; It must be that &amp;lt;math&amp;gt;D\cup \bar{D}=S,&amp;lt;/math&amp;gt; since any individual in the population either has the disease or does not. The event of an observed positive test result is denoted by the closed loop, &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;. (Notice that the shading in the diagram is relatively darker where &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; intersects with &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&lt;br /&gt;
[[File:ProbCond_venn2.jpg|frameless|400px]]&lt;br /&gt;
&lt;br /&gt;
To investigate how we might construct the required probability, &amp;lt;math&amp;gt;\Pr \left(D|P\right)&amp;lt;/math&amp;gt;, proceed as follows:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr \left( D|P\right) &amp;amp;=&amp;amp;\frac{\Pr \left( D\cap P\right) }{\Pr (P)} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{\Pr (D\cap P)}{\Pr (P\cap D)+\Pr (P\cap \bar{D})},\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
since &amp;lt;math&amp;gt;P=(P\cap D)\cup (P\cap \bar{D}),&amp;lt;/math&amp;gt; and these are mutually exclusive. From the multiplication rule of probability, &amp;lt;math&amp;gt;\Pr \left( P\cap D\right) =\Pr(P|D)\times \Pr (D),&amp;lt;/math&amp;gt; and similarly for &amp;lt;math&amp;gt;\Pr \left( P\cap \bar{D}\right)&amp;lt;/math&amp;gt;. Thus&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\Pr \left( D|P\right) =\frac{\Pr \left( P|D\right) \times \Pr \left(D\right) }{\Pr \left( P|D\right) \times \Pr \left( D\right) +\Pr (P|\bar{D})\times \Pr \left( \bar{D}\right) },&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
which may be convenient to work with since &amp;lt;math&amp;gt;\Pr \left( P|D\right) &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr\left( P|\bar{D}\right) &amp;lt;/math&amp;gt; can be estimated from clinical trials and &amp;lt;math&amp;gt;\Pr\left( D\right) &amp;lt;/math&amp;gt; estimated from recent historical survey data.&lt;br /&gt;
&lt;br /&gt;
This sort of calculation (assigning probabilities to possible causes of observed events) is an example of &amp;#039;&amp;#039;Bayes’ Theorem&amp;#039;&amp;#039;. Of course, we may have to consider more than two possible causes, and the construction of the appropriate probabilities is as follows.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Consider a sample space, &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;A,B,C&amp;lt;/math&amp;gt; are three mutually exclusive events (possible causes), defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, such that &amp;lt;math&amp;gt;S=A\cup B\cup C&amp;lt;/math&amp;gt;. In such a situation, &amp;lt;math&amp;gt;A,B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; are said to form a &amp;#039;&amp;#039;&amp;#039;partition&amp;#039;&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;. &amp;#039;&amp;#039;&amp;#039;Bayes’ Theorem&amp;#039;&amp;#039;&amp;#039; states that:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (A|E)=\frac{\Pr (E|A)\times \Pr (A)}{\left\{ \Pr (E|A)\times \Pr(A)\right\} +\left\{ \Pr (E|B)\times \Pr (B)\right\} +\left\{ \Pr(E|C)\times \Pr (C)\right\} }.&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;And, more generally, consider a sample space, &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F_{1},F_{2},...,F_{k}&amp;lt;/math&amp;gt; are &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; mutually exclusive events (possible causes), which form a partition of &amp;lt;math&amp;gt;S:S=\bigcup_{j=1}^{k}F_{j}&amp;lt;/math&amp;gt;. &amp;#039;&amp;#039;&amp;#039;Bayes’ Theorem&amp;#039;&amp;#039;&amp;#039; then states that:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (F_{j}|E)=\frac{\Pr (E|F_{j})\times \Pr (F_{j})}{\sum_{s=1}^{k}\left\{\Pr (E|F_{s})\times \Pr (F_{s})\right\} }.&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
From the above formula, you should be able to satisfy yourself that &amp;lt;math&amp;gt;\sum_{j=1}^{k}\Pr \left( F_{j}|E\right) =1.&amp;lt;/math&amp;gt; If this is not at first clear, consider case (1) and show that &amp;lt;math&amp;gt;\Pr \left( A|E\right) +\Pr \left(&lt;br /&gt;
B|E\right) +\Pr \left( C|E\right) =1.&amp;lt;/math&amp;gt; The reason for this is that since &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; form a partition of &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; they must also form a partition of any event &amp;lt;math&amp;gt;E\subset S.&amp;lt;/math&amp;gt; In the above conditional probabilities, we are regarding &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; as the restricted sample space and therefore the probabilities assigned the mutually exclusive events &amp;lt;math&amp;gt;\left( A,B,C\right) &amp;lt;/math&amp;gt; which &amp;#039;&amp;#039;cover&amp;#039;&amp;#039; this (restricted) sample space, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, must sum to 1.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example&amp;#039;&amp;#039;: Box A contains 2 red balls. Box B contains 1 red and 1 white ball. Box A and Box B are identical. If a box is selected at random and one ball is withdrawn from it, what is the probability that the selected box was number 1 if the ball withdrawn from it turns out to be red?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; be the event of selecting Box A and &amp;lt;math&amp;gt;R&amp;lt;/math&amp;gt; the event of drawing a red ball. Require &amp;lt;math&amp;gt;Pr(A|R)&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;Pr(A|R)=Pr(A\cap R)/Pr(R);&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;Pr(A\cap R)=Pr(A)Pr(R|A)=(1/2)\times 1=1/2.&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
And,&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
Pr(R) &amp;amp;=&amp;amp;Pr(A\cap R)+Pr(\bar{A}\cap R) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\Pr (A)\times \Pr (R|A)\,\,\,\,+\,\,\,\,\Pr (\bar{A})\times \Pr (R|\bar{A}) \\&lt;br /&gt;
&amp;amp;=&amp;amp;(1/2)\,\,\,\,+\,\,\,\,(1/2)\times (1/2) \\&lt;br /&gt;
&amp;amp;=&amp;amp;3/4.\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Therefore, &amp;lt;math&amp;gt;\Pr (A|R)=(1/2)/(3/4)=2/3&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
= Exercises =&lt;br /&gt;
&lt;br /&gt;
You can find examples related to these topics here: [[Probability_Conditional_Exercises]].&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Conditional&amp;diff=2875</id>
		<title>Probability Conditional</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Conditional&amp;diff=2875"/>
				<updated>2013-08-09T10:16:13Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Conditional Probability =&lt;br /&gt;
&lt;br /&gt;
An important consideration in the development of probability is that of &amp;#039;&amp;#039;conditional probability&amp;#039;&amp;#039;. This refers to the calculation of updating probabilities in the light of revealed information. For example, insurance companies nearly always set their home contents insurance premiums on the basis of the postcode in which the home is located. That is to say, insurance companies believe the risk depends upon the location; i.e., the probability of property crime is assessed conditional upon the location of the property. (A similar calculation is made to set car insurance premiums.) As a result, the premiums for two identical households located in different parts of the country can differ substantially.&lt;br /&gt;
&lt;br /&gt;
* In general, the probability of an event, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, occurring &amp;#039;&amp;#039;given&amp;#039;&amp;#039; that an event, &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;, has occurred is called the &amp;#039;&amp;#039;conditional probability&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and is denoted &amp;lt;math&amp;gt;\Pr (E|F)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
As another example, it has been well documented that the ability of a new born baby to survive is closely associated with its birth-weight. A birth-weight of less than 1500&amp;#039;&amp;#039;g&amp;#039;&amp;#039; is regarded as dangerously low. Consider &amp;lt;math&amp;gt;E=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;birth weight of a baby is less than&amp;#039;&amp;#039; 1500&amp;#039;&amp;#039;g&amp;#039;&amp;#039;, &amp;lt;math&amp;gt;F=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;mother smoked during pregnancy&amp;#039;&amp;#039;; then evidence as to whether &amp;lt;math&amp;gt;\Pr(E|F)&amp;gt;\Pr (E|\bar{F})&amp;lt;/math&amp;gt; is of considerable interest.&lt;br /&gt;
&lt;br /&gt;
As a preliminary to the main development, consider the simple experiment of rolling a fair die and observing the number of dots on the upturned face. Then &amp;lt;math&amp;gt;S=\left\{ 1,2,3,4,5,6\right\} &amp;lt;/math&amp;gt; and define events, &amp;lt;math&amp;gt;E=\left\{4\right\} &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F=\left\{ 4,5,6\right\} ;&amp;lt;/math&amp;gt; we are interested in &amp;lt;math&amp;gt;\Pr \left(E|F\right)&amp;lt;/math&amp;gt;. To work this out we take &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; as known. Given this knowledge the sample space becomes restricted to simply &amp;lt;math&amp;gt;\left\{ 4,5,6\right\} &amp;lt;/math&amp;gt; and, given no other information, each of these &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; outcome remains equally likely. So the required event, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;, is just one of three equally likely outcomes. It therefore seems reasonable that &amp;lt;math&amp;gt;\Pr (E|F)=\frac{1}{3}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
We shall now develop this idea more fully, using Venn Diagrams with the implied notion of area giving probability. Consider an abstract sample space, denoted by &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, with events &amp;lt;math&amp;gt;E\subset S,\,\,F\subset S&amp;lt;/math&amp;gt;. This is illustrated in the following Figure. Eventually we will want to construct the conditional probability, &amp;lt;math&amp;gt;\Pr \left( E|F\right)&amp;lt;/math&amp;gt;. Sticking with the above example that could be the probability that &amp;#039;&amp;#039;a child is underweight&amp;#039;&amp;#039;, given that &amp;#039;&amp;#039;the mother is a smoker&amp;#039;&amp;#039;. Two important areas used in the construction of this conditional probability are highlighted as &amp;lt;math&amp;gt;\mathbf{a}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{b}&amp;lt;/math&amp;gt;:&lt;br /&gt;
&lt;br /&gt;
[[File:ProbCond_venn1.jpg|frameless|400px]]&lt;br /&gt;
&lt;br /&gt;
In general, it is useful to think of &amp;lt;math&amp;gt;\Pr (E)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;\frac{area\left( E\right)}{area\left( S\right) }&amp;lt;/math&amp;gt;; and similarly for &amp;lt;math&amp;gt;\Pr (F)&amp;lt;/math&amp;gt;. The &amp;lt;math&amp;gt;\Pr (E\cap F)&amp;lt;/math&amp;gt; could equally be thought of as &amp;lt;math&amp;gt;\frac{area\left( a\right)}{area\left( S\right) }&amp;lt;/math&amp;gt;. With this in mind, consider what happens if we are now told that &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; has occurred. Incorporating this information implies that the effective sample space becomes restricted to &amp;lt;math&amp;gt;S^{*}=F&amp;lt;/math&amp;gt;, since &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; now defines what can happen. This now covers the sample area &amp;lt;math&amp;gt;a+b.&amp;lt;/math&amp;gt; On this new, restricted, sample space an outcome in &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; can only be observed if that outcome also belongs to &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;, the restricted sample space &amp;lt;math&amp;gt;S^*&amp;lt;/math&amp;gt;. And this only occurs in area &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt; which corresponds to the event &amp;lt;math&amp;gt;E\cap F&amp;lt;/math&amp;gt;. Thus the event of interest &amp;#039;&amp;#039;now&amp;#039;&amp;#039; is &amp;lt;math&amp;gt;E^{*}=E\cap F,&amp;lt;/math&amp;gt; as defined on the &amp;#039;&amp;#039;restricted &amp;#039;&amp;#039;sample space of &amp;lt;math&amp;gt;S^{*}=F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
In order to proceed with the construction of the conditional probability, &amp;lt;math&amp;gt;\Pr \left( E|F\right)&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;area(S)=z&amp;lt;/math&amp;gt;. Then, since the ratio of the area of the event of interest to that of the sample space gives probability, we have (on this restricted sample space):&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E|F) &amp;amp;=&amp;amp;\frac{area\left( E\cap F\right) }{area\left( F\right) } \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{a}{a+b} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{a/z}{\left( a+b\right) /z} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{\Pr \left( E\cap F\right) }{\Pr \left( F\right) },\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We have shown, for this example how a conditional probability can be expressed as a function of the joint probability &amp;lt;math&amp;gt;\Pr \left( E\cap F\right)&amp;lt;/math&amp;gt; and the marginal probability &amp;lt;math&amp;gt;\Pr \left( F\right)&amp;lt;/math&amp;gt;. This is a profound result and should be formulated in more general terms:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;The probability that &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; occurs, given that &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is known to have occurred, gives the &amp;#039;&amp;#039;&amp;#039;conditional probability&amp;#039;&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. This is denoted &amp;lt;math&amp;gt;Pr(E|F)&amp;lt;/math&amp;gt; and is calculated as&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (E|F)=\frac{\Pr (E\cap F)}{\Pr (F)}&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;and from the axioms of probability will generate a number lying between 0 and 1, since &amp;lt;math&amp;gt;\Pr (F)\geq \Pr (E\cap F)\geq 0.&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example: &amp;#039;&amp;#039;A Manufacturer of electrical components knows that the probability is 0.8 that an order will be ready for shipment on time and it is 0.6 that it will also be delivered on time. What is the probability that such an order will be delivered on time given that it was ready for shipment on time?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;R=&amp;lt;/math&amp;gt; READY, &amp;lt;math&amp;gt;D=&amp;lt;/math&amp;gt; DELIVERED ON TIME. &amp;lt;math&amp;gt;Pr(R)=0.8,Pr(R\cap D)=0.6.&amp;lt;/math&amp;gt; From this we need to calculate &amp;lt;math&amp;gt;Pr(D|R),&amp;lt;/math&amp;gt; using the above formula. This gives, &amp;lt;math&amp;gt;Pr(D|R)=Pr(R\cap D)/Pr(R)=6/8,\,\,&amp;lt;/math&amp;gt;or&amp;lt;math&amp;gt;\,\,75\%&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If we re-arrange the above formula for conditional probability, we obtain the so-called &amp;#039;&amp;#039;multiplication rule of probability &amp;#039;&amp;#039;for &amp;#039;&amp;#039;intersections&amp;#039;&amp;#039; of events:&lt;br /&gt;
&lt;br /&gt;
== Multiplication rule of probability ==&lt;br /&gt;
&lt;br /&gt;
The multiplication rule of probability can be stated as follows:&lt;br /&gt;
&lt;br /&gt;
* &amp;lt;math&amp;gt;\Pr (E\cap F)=\Pr (E|F)\times \Pr (F)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note that for any two events, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F,&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;(E\cap F)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;(E\cap \bar{F})&amp;lt;/math&amp;gt; are mutually exclusive with &amp;lt;math&amp;gt;E=(E\cap F)\cup (E\cap \bar{F})&amp;lt;/math&amp;gt;; this has been seen before. So the &amp;#039;&amp;#039;addition rule&amp;#039;&amp;#039; and &amp;#039;&amp;#039;multiplication rule&amp;#039;&amp;#039; of probability together give:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E) &amp;amp;=&amp;amp;\Pr (E\cap F)+\Pr (E\cap \bar{F}) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\Pr (E|F)\times \Pr (F)+\Pr (E|\bar{F})\times \Pr (\bar{F}).\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This is an extremely important and useful result, in practice, as we shall see shortly.&lt;br /&gt;
&lt;br /&gt;
== Statistical Independence ==&lt;br /&gt;
&lt;br /&gt;
If the knowledge that &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; has occurred does NOT alter our probability assessment of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are said to be (statistically) &amp;#039;&amp;#039;independent&amp;#039;&amp;#039;. In this sense, &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; carries no information about &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Formally, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are &amp;#039;&amp;#039;&amp;#039;independent&amp;#039;&amp;#039;&amp;#039; events if and only if&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;Pr(E|F)=Pr(E)&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;which, in turn is true&amp;#039;&amp;#039; if and only if &amp;#039;&amp;#039;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;Pr(E\cap F)=Pr(E)\times Pr(F).&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This concept of independence is of enormous importance in practice. Consider the case of lung cancer and its connection to smoking (apologies to all smokers for being picked upon here). The first connection between smoking and lung cancer was made in the 1920s. However, for many decades after the tobacco industry spend a lot of money and effort to convince people that there was no connection between the two. In other words they claimed that the two events are &amp;#039;&amp;#039;independent&amp;#039;&amp;#039;, or &amp;lt;math&amp;gt;Pr(Cancer|Smoking)=Pr(Cancer|\cap{Smoking})=Pr(Cancer)&amp;lt;/math&amp;gt;. It was then the task of epidemiologists to show otherwise. This was famously and comprehensively achieved by the [http://en.wikipedia.org/wiki/British_Doctors_Study &amp;#039;&amp;#039;British Doctors Study&amp;#039;&amp;#039;].&lt;br /&gt;
&lt;br /&gt;
== Bayes’ Theorem ==&lt;br /&gt;
&lt;br /&gt;
One area where conditional probability is extremely important is that of clinical trials - testing the power of a diagnostic test to detect the presence of a particular disease. Suppose, then, that a new test is being developed and let &amp;lt;math&amp;gt;P=&amp;lt;/math&amp;gt; ‘&amp;#039;&amp;#039;test positive&amp;#039;&amp;#039;’ and &amp;lt;math&amp;gt;D=&amp;lt;/math&amp;gt; ‘&amp;#039;&amp;#039;presence of disease&amp;#039;&amp;#039;’, but where the results from applying the diagnostic test can never be wholly reliable. From the point of view of our previous discussion on conditional probability, we would of course require that &amp;lt;math&amp;gt;\Pr \left( P|D\right)&amp;lt;/math&amp;gt; to be large; i.e., the test should be effective at detecting the disease. However, if you think about, this is not necessarily the probability that we might be interested in from a diagnosis point of view. Rather, we should be more interested in &amp;lt;math&amp;gt;\Pr \left( D|P\right)&amp;lt;/math&amp;gt;, the probability of correct diagnosis, and require this to be large (with, presumably, &amp;lt;math&amp;gt;\Pr (D|\bar{P})&amp;lt;/math&amp;gt; being small). Here, what we are trying to attach a probability to is a possible ‘cause’. The observed outcome is a positive test result (&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;), but the presence or non-presence of the disease is what is of interest and this is uncertain. &amp;lt;math&amp;gt;\Pr (D|P)&amp;lt;/math&amp;gt; asks the question ‘&amp;#039;&amp;#039;what is the probability that it is the presence of the disease which caused the positive test result&amp;#039;&amp;#039;’? (Another recent news-worthy example would be the effect of exposure to depleted uranium on Gulf and Balkan war veterans. Given the presence of lymph, lung or brain cancer in such individuals (&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;), how likely is that the cause was exposure to depleted uranium weapons (&amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;)? Firstly, is &amp;lt;math&amp;gt;\Pr \left( D|P\right) &amp;lt;/math&amp;gt; high or low? Secondly, might there being something else (&amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;) which could offer a “better” explanation, such that &amp;lt;math&amp;gt;\Pr \left( F|P\right) &amp;gt;\Pr \left( D|F\right) &amp;lt;/math&amp;gt; ?)&lt;br /&gt;
&lt;br /&gt;
The situation is depicted in the following Figure, in which there are two possible ‘states’ in the population: &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt; (depicted by the lighter shaded area covering the left portion of the sample space) and &amp;lt;math&amp;gt;\bar{D}.&amp;lt;/math&amp;gt; It must be that &amp;lt;math&amp;gt;D\cup \bar{D}=S,&amp;lt;/math&amp;gt; since any individual in the population either has the disease or does not. The event of an observed positive test result is denoted by the closed loop, &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;. (Notice that the shading in the diagram is relatively darker where &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; intersects with &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&lt;br /&gt;
[[File:ProbCond_venn2.jpg|frameless|400px]]&lt;br /&gt;
&lt;br /&gt;
To investigate how we might construct the required probability, &amp;lt;math&amp;gt;\Pr \left(D|P\right)&amp;lt;/math&amp;gt;, proceed as follows:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr \left( D|P\right) &amp;amp;=&amp;amp;\frac{\Pr \left( D\cap P\right) }{\Pr (P)} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{\Pr (D\cap P)}{\Pr (P\cap D)+\Pr (P\cap \bar{D})},\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
since &amp;lt;math&amp;gt;P=(P\cap D)\cup (P\cap \bar{D}),&amp;lt;/math&amp;gt; and these are mutually exclusive. From the multiplication rule of probability, &amp;lt;math&amp;gt;\Pr \left( P\cap D\right) =\Pr(P|D)\times \Pr (D),&amp;lt;/math&amp;gt; and similarly for &amp;lt;math&amp;gt;\Pr \left( P\cap \bar{D}\right)&amp;lt;/math&amp;gt;. Thus&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\Pr \left( D|P\right) =\frac{\Pr \left( P|D\right) \times \Pr \left(D\right) }{\Pr \left( P|D\right) \times \Pr \left( D\right) +\Pr (P|\bar{D})\times \Pr \left( \bar{D}\right) },&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
which may be convenient to work with since &amp;lt;math&amp;gt;\Pr \left( P|D\right) &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr\left( P|\bar{D}\right) &amp;lt;/math&amp;gt; can be estimated from clinical trials and &amp;lt;math&amp;gt;\Pr\left( D\right) &amp;lt;/math&amp;gt; estimated from recent historical survey data.&lt;br /&gt;
&lt;br /&gt;
This sort of calculation (assigning probabilities to possible causes of observed events) is an example of &amp;#039;&amp;#039;Bayes’ Theorem&amp;#039;&amp;#039;. Of course, we may have to consider more than two possible causes, and the construction of the appropriate probabilities is as follows.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Consider a sample space, &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;A,B,C&amp;lt;/math&amp;gt; are three mutually exclusive events (possible causes), defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, such that &amp;lt;math&amp;gt;S=A\cup B\cup C&amp;lt;/math&amp;gt;. In such a situation, &amp;lt;math&amp;gt;A,B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; are said to form a &amp;#039;&amp;#039;&amp;#039;partition&amp;#039;&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;. &amp;#039;&amp;#039;&amp;#039;Bayes’ Theorem&amp;#039;&amp;#039;&amp;#039; states that:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (A|E)=\frac{\Pr (E|A)\times \Pr (A)}{\left\{ \Pr (E|A)\times \Pr(A)\right\} +\left\{ \Pr (E|B)\times \Pr (B)\right\} +\left\{ \Pr(E|C)\times \Pr (C)\right\} }.&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;And, more generally, consider a sample space, &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F_{1},F_{2},...,F_{k}&amp;lt;/math&amp;gt; are &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; mutually exclusive events (possible causes), which form a partition of &amp;lt;math&amp;gt;S:S=\bigcup_{j=1}^{k}F_{j}&amp;lt;/math&amp;gt;. &amp;#039;&amp;#039;&amp;#039;Bayes’ Theorem&amp;#039;&amp;#039;&amp;#039; then states that:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (F_{j}|E)=\frac{\Pr (E|F_{j})\times \Pr (F_{j})}{\sum_{s=1}^{k}\left\{\Pr (E|F_{s})\times \Pr (F_{s})\right\} }.&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
From the above formula, you should be able to satisfy yourself that &amp;lt;math&amp;gt;\sum_{j=1}^{k}\Pr \left( F_{j}|E\right) =1.&amp;lt;/math&amp;gt; If this is not at first clear, consider case (1) and show that &amp;lt;math&amp;gt;\Pr \left( A|E\right) +\Pr \left(&lt;br /&gt;
B|E\right) +\Pr \left( C|E\right) =1.&amp;lt;/math&amp;gt; The reason for this is that since &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; form a partition of &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; they must also form a partition of any event &amp;lt;math&amp;gt;E\subset S.&amp;lt;/math&amp;gt; In the above conditional probabilities, we are regarding &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; as the restricted sample space and therefore the probabilities assigned the mutually exclusive events &amp;lt;math&amp;gt;\left( A,B,C\right) &amp;lt;/math&amp;gt; which &amp;#039;&amp;#039;cover&amp;#039;&amp;#039; this (restricted) sample space, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, must sum to 1.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example&amp;#039;&amp;#039;: Box A contains 2 red balls. Box B contains 1 red and 1 white ball. Box A and Box B are identical. If a box is selected at random and one ball is withdrawn from it, what is the probability that the selected box was number 1 if the ball withdrawn from it turns out to be red?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; be the event of selecting Box A and &amp;lt;math&amp;gt;R&amp;lt;/math&amp;gt; the event of drawing a red ball. Require &amp;lt;math&amp;gt;Pr(A|R)&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;Pr(A|R)=Pr(A\cap R)/Pr(R);&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;Pr(A\cap R)=Pr(A)Pr(R|A)=(1/2)\times 1=1/2.&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
And,&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
Pr(R) &amp;amp;=&amp;amp;Pr(A\cap R)+Pr(\bar{A}\cap R) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\Pr (A)\times \Pr (R|A)\,\,\,\,+\,\,\,\,\Pr (\bar{A})\times \Pr (R|\bar{A}) \\&lt;br /&gt;
&amp;amp;=&amp;amp;(1/2)\,\,\,\,+\,\,\,\,(1/2)\times (1/2) \\&lt;br /&gt;
&amp;amp;=&amp;amp;3/4.\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Therefore, &amp;lt;math&amp;gt;\Pr (A|R)=(1/2)/(3/4)=2/3&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
= Exercises =&lt;br /&gt;
&lt;br /&gt;
You can find examples related to these topics here: [[Probability_Conditional_Examples]].&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Conditional_Exercises&amp;diff=2874</id>
		<title>Probability Conditional Exercises</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Conditional_Exercises&amp;diff=2874"/>
				<updated>2013-08-09T10:14:24Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: Created page with &amp;quot;  = Conditional Probabilities Exercises =  &amp;lt;ol&amp;gt; &amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; are events such that &amp;lt;math&amp;gt;\Pr (A)=0.4&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr (A\cup B)=0.75.&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Conditional Probabilities Exercises =&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; are events such that &amp;lt;math&amp;gt;\Pr (A)=0.4&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr (A\cup B)=0.75.&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Find &amp;lt;math&amp;gt;\Pr (B)&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; are mutually exclusive.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Find &amp;lt;math&amp;gt;\Pr (B)&amp;lt;/math&amp;gt; if &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; are independent.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Events &amp;lt;math&amp;gt;A,&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; are such that &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; are mutually exclusive and &amp;lt;math&amp;gt;\Pr (A)=2/3,&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr (A\cup B)=5/6&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr (B\cup C)=4/5.&amp;lt;/math&amp;gt; If &amp;lt;math&amp;gt;\Pr (B|A)=1/2&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr (C|A)=3/10,&amp;lt;/math&amp;gt; are &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; statistically independent?&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;A sample of 1000 undergraduates were asked whether they took either Mathematics, Physics or Chemistry at A-level. The following responses were obtained: 100 just took Mathematics; 70 just took Physics; 100 just took Chemistry; 150 took Mathematics and Physics, but not Chemistry; 40 took Mathematics and Chemistry, but not Physics; and, 240 took Physics and Chemistry, but not Mathematics. Calculate the following:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Of those who took Mathematics, what proportion also took Physics (but not Chemistry) and what proportion took both Physics and Chemistry?&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Of those who took Physics and Chemistry, what proportion also took Mathematics?&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;The &amp;#039;&amp;#039;Survey of British Births&amp;#039;&amp;#039;, undertaken in the 1970s, aimed to improve the survival rate and care of British babies at, and soon after, birth by collecting and analysing data on new-born babies. A sample was taken designed to be representative of the whole population of British births and consisted of all babies born alive (or dead) after the 24th week of gestation, between 0001 hours on Sunday 5 April and 2400 hours on Saturday 11 April 1970. The total number in the sample so obtained was &amp;lt;math&amp;gt;n=17,530.&amp;lt;/math&amp;gt; A large amount of information was obtained, but one particular area of interest was the effect of the smoking habits of the mothers on newly born babies. In particular, the ability of a newly born baby to survive is closely associated with its birth-weight and a birth-weight of less than 1500&amp;#039;&amp;#039;g&amp;#039;&amp;#039; is considered dangerously low. Some of the relevant data are summarised as follows.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;For all new born babies in the sample, the proportion of mothers who:(i) &amp;#039;&amp;#039;smoked before and during pregnancy was&amp;#039;&amp;#039; 0.433(ii) &amp;#039;&amp;#039;gave up smoking prior to pregnancy was&amp;#039;&amp;#039; 0.170(iii) &amp;#039;&amp;#039;who had never smoked was&amp;#039;&amp;#039; 0.397.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;However, by breaking down the sample into mothers who smoked, had given up, or who had never smoked, the following statistics were obtained:(iv) &amp;lt;math&amp;gt;1.6\%&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;of the mothers who smoked gave birth to babies whose weight was less than&amp;#039;&amp;#039; 1500&amp;#039;&amp;#039;g&amp;#039;&amp;#039;,(v) &amp;lt;math&amp;gt;0.9\%&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;of the mothers who had given up smoking prior to pregnancy gave birth to babies whose weight was less than&amp;#039;&amp;#039; 1500&amp;#039;&amp;#039;g&amp;#039;&amp;#039;,(vi) &amp;lt;math&amp;gt;0.8\%&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;of mothers who had never smoked gave birth to babies whose weight was less than&amp;#039;&amp;#039; 1500&amp;#039;&amp;#039;g&amp;#039;&amp;#039;.&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Given this information, how would you estimate the risk, for a smoking mother, of giving birth to a dangerously under-weight baby? What is the corresponding risk for a mother who has never smoked? What is the overall risk of giving birth to an under-weight baby?&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Of the babies born under &amp;lt;math&amp;gt;1500&amp;lt;/math&amp;gt;g&amp;lt;math&amp;gt;,&amp;lt;/math&amp;gt; estimate the proportion of these (a) born to mothers who smoked before and during pregnancy; (b) born to mothers who had never smoked.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;On the basis of the above information, how would you assess the evidence on smoking during pregnancy as a factor which could result in babies being born under weight?&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ol&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Metal fatigue in an aeroplane’s wing can be caused by any one of three (relatively minor) defects, labelled &amp;lt;math&amp;gt;A,\,\,B\,\,&amp;lt;/math&amp;gt;and &amp;lt;math&amp;gt;C,&amp;lt;/math&amp;gt; occurring during the manufacturing process. The probabilities are estimated as: &amp;lt;math&amp;gt;\Pr (A)=0.3,&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr (B)=0.1,&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr (C)=0.6.&amp;lt;/math&amp;gt; At the quality control stage of production, a test has been developed which is used to detect the presence of a defect. Let &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt; be the event that the test detects a manufacturing defect with the following probabilities: &amp;lt;math&amp;gt;\Pr (D|A)=0.6,&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr (D|B)=0.2,&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr (D|C)=0.7.&amp;lt;/math&amp;gt; If the test detects a defect, which of &amp;lt;math&amp;gt;A,&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; is the most likely cause? (&amp;#039;&amp;#039;Hint&amp;#039;&amp;#039;: you need to find, and compare, &amp;lt;math&amp;gt;\Pr\left( A|D\right)&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\Pr \left( B|D\right) &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr \left( C|D\right) &amp;lt;/math&amp;gt; using Bayes Theorem.)&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Statistics&amp;diff=2873</id>
		<title>Statistics</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Statistics&amp;diff=2873"/>
				<updated>2013-08-09T10:11:24Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: /* Probability */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This page is currently being build and should be completed for the start of Academic Year 2013/14.&lt;br /&gt;
&lt;br /&gt;
The purpose of these pages is to give prospective and current postgraduate students in Economics at The University of Manchester an opportunity to understand what Statistics knowledge they are expected to have and, of course, a source of revision information. The material presented here is largely based on material written by our colleagues Len Gill, Denise Osborn and Chris Orme at The University of Manchester.&lt;br /&gt;
&lt;br /&gt;
This page is explicitly designed with MSc Economics (and/or Econometrics) and MA Economics students at The University of Manchester in mind. Below you will be able to identify which topic is expected knowledge for the MSc/MA.&lt;br /&gt;
&lt;br /&gt;
== The Basics ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Topic&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| &lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| MA Economics&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| MSc Economics&lt;br /&gt;
|-&lt;br /&gt;
| [[StatPrelim|Preliminaries and Notation]]&lt;br /&gt;
|&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[DataTypes|Data types]]&lt;br /&gt;
|&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[GraphicRep|Graphical Representation]]&lt;br /&gt;
|&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Descriptive|Descriptive Statistics]]&lt;br /&gt;
| &lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Regression|Correlation &amp;amp; Regression]]&lt;br /&gt;
| [[Regression_Examples|Exercises]]&lt;br /&gt;
| Correlation only&lt;br /&gt;
| Yes&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Probability ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Topic&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| &lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| MA Economics&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| MSc Economics&lt;br /&gt;
|-&lt;br /&gt;
| [[Probability_Intro|Introduction to Probability]]&lt;br /&gt;
|&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Probability_Conditional|Conditional Probability]]&lt;br /&gt;
| [[Probability_Conditional_Exercises|Exercises]]&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Probability_DiscreteRV|Discrete Random Variables]]&lt;br /&gt;
|&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Probability_Norm|Normal Distribution]]&lt;br /&gt;
|&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Probability_MomenntsExp|Moments and Expectations]]&lt;br /&gt;
|&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Statistical Inference ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==&amp;lt;div id=&amp;quot;DataSets&amp;quot;&amp;gt;&amp;lt;/div&amp;gt;Example Data Sets==&lt;br /&gt;
&lt;br /&gt;
Throughout these pages a few example datasets will be used.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Description&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Organisation&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Data File&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Source&lt;br /&gt;
|-&lt;br /&gt;
| Data used to calculate a quality of Life Index&lt;br /&gt;
| OECD (collecting data from various sources)&lt;br /&gt;
| [[media:OECD_BetterLifeIndex.xls|BetterLifeIndex.xls]]&lt;br /&gt;
| [http://www.oecdbetterlifeindex.org/]&lt;br /&gt;
|-&lt;br /&gt;
| Exchange Rate USD/UKP&lt;br /&gt;
| Board of Governors of the Federal reserve System&lt;br /&gt;
| [[media:USDUKP.xlsx|USDUKP.xlsx]]&lt;br /&gt;
| [http://www.federalreserve.gov/releases/h10/hist/dat00_uk.htm]&lt;br /&gt;
|-&lt;br /&gt;
| UK Gross Domestic Product&lt;br /&gt;
| Office of national Statistics&lt;br /&gt;
| [[media:UK_GDP_INF.xlsx|UK_GDP_INF.xlsx]]&lt;br /&gt;
| [http://www.guardian.co.uk/]&lt;br /&gt;
|-&lt;br /&gt;
| Passengers on the Titanic&lt;br /&gt;
| &lt;br /&gt;
|[[media:Titanic.xlsx|Titanic.xlsx]]&lt;br /&gt;
|[http://www.statsci.org/data/general/titanic.html]&lt;br /&gt;
|-&lt;br /&gt;
| CO2 and GDP&lt;br /&gt;
| [http://Gapminder.com Gapminder.com]&lt;br /&gt;
|[[media:GDP_CO2.xlsx|GDP_CO2.xlsx]]&lt;br /&gt;
|[http://Gapminder.com]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Statistics&amp;diff=2872</id>
		<title>Statistics</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Statistics&amp;diff=2872"/>
				<updated>2013-08-09T10:10:28Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: /* The Basics */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This page is currently being build and should be completed for the start of Academic Year 2013/14.&lt;br /&gt;
&lt;br /&gt;
The purpose of these pages is to give prospective and current postgraduate students in Economics at The University of Manchester an opportunity to understand what Statistics knowledge they are expected to have and, of course, a source of revision information. The material presented here is largely based on material written by our colleagues Len Gill, Denise Osborn and Chris Orme at The University of Manchester.&lt;br /&gt;
&lt;br /&gt;
This page is explicitly designed with MSc Economics (and/or Econometrics) and MA Economics students at The University of Manchester in mind. Below you will be able to identify which topic is expected knowledge for the MSc/MA.&lt;br /&gt;
&lt;br /&gt;
== The Basics ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Topic&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| &lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| MA Economics&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| MSc Economics&lt;br /&gt;
|-&lt;br /&gt;
| [[StatPrelim|Preliminaries and Notation]]&lt;br /&gt;
|&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[DataTypes|Data types]]&lt;br /&gt;
|&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[GraphicRep|Graphical Representation]]&lt;br /&gt;
|&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Descriptive|Descriptive Statistics]]&lt;br /&gt;
| &lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Regression|Correlation &amp;amp; Regression]]&lt;br /&gt;
| [[Regression_Examples|Exercises]]&lt;br /&gt;
| Correlation only&lt;br /&gt;
| Yes&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Probability ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Topic&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| MA Economics&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| MSc Economics&lt;br /&gt;
|-&lt;br /&gt;
| [[Probability_Intro|Introduction to Probability]]&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Probability_Conditional|Conditional Probability]]&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Probability_DiscreteRV|Discrete Random Variables]]&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Probability_Norm|Normal Distribution]]&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Probability_MomenntsExp|Moments and Expectations]]&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Statistical Inference ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==&amp;lt;div id=&amp;quot;DataSets&amp;quot;&amp;gt;&amp;lt;/div&amp;gt;Example Data Sets==&lt;br /&gt;
&lt;br /&gt;
Throughout these pages a few example datasets will be used.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Description&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Organisation&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Data File&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Source&lt;br /&gt;
|-&lt;br /&gt;
| Data used to calculate a quality of Life Index&lt;br /&gt;
| OECD (collecting data from various sources)&lt;br /&gt;
| [[media:OECD_BetterLifeIndex.xls|BetterLifeIndex.xls]]&lt;br /&gt;
| [http://www.oecdbetterlifeindex.org/]&lt;br /&gt;
|-&lt;br /&gt;
| Exchange Rate USD/UKP&lt;br /&gt;
| Board of Governors of the Federal reserve System&lt;br /&gt;
| [[media:USDUKP.xlsx|USDUKP.xlsx]]&lt;br /&gt;
| [http://www.federalreserve.gov/releases/h10/hist/dat00_uk.htm]&lt;br /&gt;
|-&lt;br /&gt;
| UK Gross Domestic Product&lt;br /&gt;
| Office of national Statistics&lt;br /&gt;
| [[media:UK_GDP_INF.xlsx|UK_GDP_INF.xlsx]]&lt;br /&gt;
| [http://www.guardian.co.uk/]&lt;br /&gt;
|-&lt;br /&gt;
| Passengers on the Titanic&lt;br /&gt;
| &lt;br /&gt;
|[[media:Titanic.xlsx|Titanic.xlsx]]&lt;br /&gt;
|[http://www.statsci.org/data/general/titanic.html]&lt;br /&gt;
|-&lt;br /&gt;
| CO2 and GDP&lt;br /&gt;
| [http://Gapminder.com Gapminder.com]&lt;br /&gt;
|[[media:GDP_CO2.xlsx|GDP_CO2.xlsx]]&lt;br /&gt;
|[http://Gapminder.com]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Statistics&amp;diff=2871</id>
		<title>Statistics</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Statistics&amp;diff=2871"/>
				<updated>2013-08-09T10:09:51Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: /* The Basics */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This page is currently being build and should be completed for the start of Academic Year 2013/14.&lt;br /&gt;
&lt;br /&gt;
The purpose of these pages is to give prospective and current postgraduate students in Economics at The University of Manchester an opportunity to understand what Statistics knowledge they are expected to have and, of course, a source of revision information. The material presented here is largely based on material written by our colleagues Len Gill, Denise Osborn and Chris Orme at The University of Manchester.&lt;br /&gt;
&lt;br /&gt;
This page is explicitly designed with MSc Economics (and/or Econometrics) and MA Economics students at The University of Manchester in mind. Below you will be able to identify which topic is expected knowledge for the MSc/MA.&lt;br /&gt;
&lt;br /&gt;
== The Basics ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Topic&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| &lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| MA Economics&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| MSc Economics&lt;br /&gt;
|-&lt;br /&gt;
| [[StatPrelim|Preliminaries and Notation]]&lt;br /&gt;
|&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[DataTypes|Data types]]&lt;br /&gt;
|&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[GraphicRep|Graphical Representation]]&lt;br /&gt;
|&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Descriptive|Descriptive Statistics]]&lt;br /&gt;
| &lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Regression|Correlation &amp;amp; Regression]]&lt;br /&gt;
| Exercises&lt;br /&gt;
| Correlation only&lt;br /&gt;
| Yes&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Probability ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Topic&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| MA Economics&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| MSc Economics&lt;br /&gt;
|-&lt;br /&gt;
| [[Probability_Intro|Introduction to Probability]]&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Probability_Conditional|Conditional Probability]]&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Probability_DiscreteRV|Discrete Random Variables]]&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Probability_Norm|Normal Distribution]]&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Probability_MomenntsExp|Moments and Expectations]]&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Statistical Inference ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==&amp;lt;div id=&amp;quot;DataSets&amp;quot;&amp;gt;&amp;lt;/div&amp;gt;Example Data Sets==&lt;br /&gt;
&lt;br /&gt;
Throughout these pages a few example datasets will be used.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Description&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Organisation&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Data File&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Source&lt;br /&gt;
|-&lt;br /&gt;
| Data used to calculate a quality of Life Index&lt;br /&gt;
| OECD (collecting data from various sources)&lt;br /&gt;
| [[media:OECD_BetterLifeIndex.xls|BetterLifeIndex.xls]]&lt;br /&gt;
| [http://www.oecdbetterlifeindex.org/]&lt;br /&gt;
|-&lt;br /&gt;
| Exchange Rate USD/UKP&lt;br /&gt;
| Board of Governors of the Federal reserve System&lt;br /&gt;
| [[media:USDUKP.xlsx|USDUKP.xlsx]]&lt;br /&gt;
| [http://www.federalreserve.gov/releases/h10/hist/dat00_uk.htm]&lt;br /&gt;
|-&lt;br /&gt;
| UK Gross Domestic Product&lt;br /&gt;
| Office of national Statistics&lt;br /&gt;
| [[media:UK_GDP_INF.xlsx|UK_GDP_INF.xlsx]]&lt;br /&gt;
| [http://www.guardian.co.uk/]&lt;br /&gt;
|-&lt;br /&gt;
| Passengers on the Titanic&lt;br /&gt;
| &lt;br /&gt;
|[[media:Titanic.xlsx|Titanic.xlsx]]&lt;br /&gt;
|[http://www.statsci.org/data/general/titanic.html]&lt;br /&gt;
|-&lt;br /&gt;
| CO2 and GDP&lt;br /&gt;
| [http://Gapminder.com Gapminder.com]&lt;br /&gt;
|[[media:GDP_CO2.xlsx|GDP_CO2.xlsx]]&lt;br /&gt;
|[http://Gapminder.com]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Statistics&amp;diff=2870</id>
		<title>Statistics</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Statistics&amp;diff=2870"/>
				<updated>2013-08-09T10:09:14Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: /* The Basics */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This page is currently being build and should be completed for the start of Academic Year 2013/14.&lt;br /&gt;
&lt;br /&gt;
The purpose of these pages is to give prospective and current postgraduate students in Economics at The University of Manchester an opportunity to understand what Statistics knowledge they are expected to have and, of course, a source of revision information. The material presented here is largely based on material written by our colleagues Len Gill, Denise Osborn and Chris Orme at The University of Manchester.&lt;br /&gt;
&lt;br /&gt;
This page is explicitly designed with MSc Economics (and/or Econometrics) and MA Economics students at The University of Manchester in mind. Below you will be able to identify which topic is expected knowledge for the MSc/MA.&lt;br /&gt;
&lt;br /&gt;
== The Basics ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Topic&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| &lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| MA Economics&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| MSc Economics&lt;br /&gt;
|-&lt;br /&gt;
| [[StatPrelim|Preliminaries and Notation]]&lt;br /&gt;
|&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[DataTypes|Data types]]&lt;br /&gt;
|&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[GraphicRep|Graphical Representation]]&lt;br /&gt;
|&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Descriptive|Descriptive Statistics]]&lt;br /&gt;
| Exercises&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Regression|Correlation &amp;amp; Regression]]&lt;br /&gt;
| &lt;br /&gt;
| Correlation only&lt;br /&gt;
| Yes&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Probability ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Topic&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| MA Economics&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| MSc Economics&lt;br /&gt;
|-&lt;br /&gt;
| [[Probability_Intro|Introduction to Probability]]&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Probability_Conditional|Conditional Probability]]&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Probability_DiscreteRV|Discrete Random Variables]]&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Probability_Norm|Normal Distribution]]&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Probability_MomenntsExp|Moments and Expectations]]&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Statistical Inference ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==&amp;lt;div id=&amp;quot;DataSets&amp;quot;&amp;gt;&amp;lt;/div&amp;gt;Example Data Sets==&lt;br /&gt;
&lt;br /&gt;
Throughout these pages a few example datasets will be used.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Description&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Organisation&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Data File&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Source&lt;br /&gt;
|-&lt;br /&gt;
| Data used to calculate a quality of Life Index&lt;br /&gt;
| OECD (collecting data from various sources)&lt;br /&gt;
| [[media:OECD_BetterLifeIndex.xls|BetterLifeIndex.xls]]&lt;br /&gt;
| [http://www.oecdbetterlifeindex.org/]&lt;br /&gt;
|-&lt;br /&gt;
| Exchange Rate USD/UKP&lt;br /&gt;
| Board of Governors of the Federal reserve System&lt;br /&gt;
| [[media:USDUKP.xlsx|USDUKP.xlsx]]&lt;br /&gt;
| [http://www.federalreserve.gov/releases/h10/hist/dat00_uk.htm]&lt;br /&gt;
|-&lt;br /&gt;
| UK Gross Domestic Product&lt;br /&gt;
| Office of national Statistics&lt;br /&gt;
| [[media:UK_GDP_INF.xlsx|UK_GDP_INF.xlsx]]&lt;br /&gt;
| [http://www.guardian.co.uk/]&lt;br /&gt;
|-&lt;br /&gt;
| Passengers on the Titanic&lt;br /&gt;
| &lt;br /&gt;
|[[media:Titanic.xlsx|Titanic.xlsx]]&lt;br /&gt;
|[http://www.statsci.org/data/general/titanic.html]&lt;br /&gt;
|-&lt;br /&gt;
| CO2 and GDP&lt;br /&gt;
| [http://Gapminder.com Gapminder.com]&lt;br /&gt;
|[[media:GDP_CO2.xlsx|GDP_CO2.xlsx]]&lt;br /&gt;
|[http://Gapminder.com]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Statistics&amp;diff=2869</id>
		<title>Statistics</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Statistics&amp;diff=2869"/>
				<updated>2013-08-09T10:08:59Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: /* The Basics */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;This page is currently being build and should be completed for the start of Academic Year 2013/14.&lt;br /&gt;
&lt;br /&gt;
The purpose of these pages is to give prospective and current postgraduate students in Economics at The University of Manchester an opportunity to understand what Statistics knowledge they are expected to have and, of course, a source of revision information. The material presented here is largely based on material written by our colleagues Len Gill, Denise Osborn and Chris Orme at The University of Manchester.&lt;br /&gt;
&lt;br /&gt;
This page is explicitly designed with MSc Economics (and/or Econometrics) and MA Economics students at The University of Manchester in mind. Below you will be able to identify which topic is expected knowledge for the MSc/MA.&lt;br /&gt;
&lt;br /&gt;
== The Basics ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Topic&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| &lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| MA Economics&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| MSc Economics&lt;br /&gt;
|-&lt;br /&gt;
| [[StatPrelim|Preliminaries and Notation]]&lt;br /&gt;
|&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[DataTypes|Data types]]&lt;br /&gt;
|&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[GraphicRep|Graphical Representation]]&lt;br /&gt;
|&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Descriptive|Descriptive Statistics]]&lt;br /&gt;
| Exercises&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Regression|Correlation &amp;amp; Regression]]&lt;br /&gt;
| Correlation only&lt;br /&gt;
| Yes&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
== Probability ==&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Topic&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| MA Economics&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| MSc Economics&lt;br /&gt;
|-&lt;br /&gt;
| [[Probability_Intro|Introduction to Probability]]&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Probability_Conditional|Conditional Probability]]&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Probability_DiscreteRV|Discrete Random Variables]]&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Probability_Norm|Normal Distribution]]&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
|-&lt;br /&gt;
| [[Probability_MomenntsExp|Moments and Expectations]]&lt;br /&gt;
| Yes&lt;br /&gt;
| Yes&lt;br /&gt;
&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Statistical Inference ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==&amp;lt;div id=&amp;quot;DataSets&amp;quot;&amp;gt;&amp;lt;/div&amp;gt;Example Data Sets==&lt;br /&gt;
&lt;br /&gt;
Throughout these pages a few example datasets will be used.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Description&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Organisation&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Data File&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Source&lt;br /&gt;
|-&lt;br /&gt;
| Data used to calculate a quality of Life Index&lt;br /&gt;
| OECD (collecting data from various sources)&lt;br /&gt;
| [[media:OECD_BetterLifeIndex.xls|BetterLifeIndex.xls]]&lt;br /&gt;
| [http://www.oecdbetterlifeindex.org/]&lt;br /&gt;
|-&lt;br /&gt;
| Exchange Rate USD/UKP&lt;br /&gt;
| Board of Governors of the Federal reserve System&lt;br /&gt;
| [[media:USDUKP.xlsx|USDUKP.xlsx]]&lt;br /&gt;
| [http://www.federalreserve.gov/releases/h10/hist/dat00_uk.htm]&lt;br /&gt;
|-&lt;br /&gt;
| UK Gross Domestic Product&lt;br /&gt;
| Office of national Statistics&lt;br /&gt;
| [[media:UK_GDP_INF.xlsx|UK_GDP_INF.xlsx]]&lt;br /&gt;
| [http://www.guardian.co.uk/]&lt;br /&gt;
|-&lt;br /&gt;
| Passengers on the Titanic&lt;br /&gt;
| &lt;br /&gt;
|[[media:Titanic.xlsx|Titanic.xlsx]]&lt;br /&gt;
|[http://www.statsci.org/data/general/titanic.html]&lt;br /&gt;
|-&lt;br /&gt;
| CO2 and GDP&lt;br /&gt;
| [http://Gapminder.com Gapminder.com]&lt;br /&gt;
|[[media:GDP_CO2.xlsx|GDP_CO2.xlsx]]&lt;br /&gt;
|[http://Gapminder.com]&lt;br /&gt;
|}&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Conditional&amp;diff=2868</id>
		<title>Probability Conditional</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Conditional&amp;diff=2868"/>
				<updated>2013-08-09T09:58:51Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: /* Conditional Probability */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Conditional Probability =&lt;br /&gt;
&lt;br /&gt;
An important consideration in the development of probability is that of &amp;#039;&amp;#039;conditional probability&amp;#039;&amp;#039;. This refers to the calculation of updating probabilities in the light of revealed information. For example, insurance companies nearly always set their home contents insurance premiums on the basis of the postcode in which the home is located. That is to say, insurance companies believe the risk depends upon the location; i.e., the probability of property crime is assessed conditional upon the location of the property. (A similar calculation is made to set car insurance premiums.) As a result, the premiums for two identical households located in different parts of the country can differ substantially.&lt;br /&gt;
&lt;br /&gt;
* In general, the probability of an event, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, occurring &amp;#039;&amp;#039;given&amp;#039;&amp;#039; that an event, &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;, has occurred is called the &amp;#039;&amp;#039;conditional probability&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and is denoted &amp;lt;math&amp;gt;\Pr (E|F)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
As another example, it has been well documented that the ability of a new born baby to survive is closely associated with its birth-weight. A birth-weight of less than 1500&amp;#039;&amp;#039;g&amp;#039;&amp;#039; is regarded as dangerously low. Consider &amp;lt;math&amp;gt;E=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;birth weight of a baby is less than&amp;#039;&amp;#039; 1500&amp;#039;&amp;#039;g&amp;#039;&amp;#039;, &amp;lt;math&amp;gt;F=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;mother smoked during pregnancy&amp;#039;&amp;#039;; then evidence as to whether &amp;lt;math&amp;gt;\Pr(E|F)&amp;gt;\Pr (E|\bar{F})&amp;lt;/math&amp;gt; is of considerable interest.&lt;br /&gt;
&lt;br /&gt;
As a preliminary to the main development, consider the simple experiment of rolling a fair die and observing the number of dots on the upturned face. Then &amp;lt;math&amp;gt;S=\left\{ 1,2,3,4,5,6\right\} &amp;lt;/math&amp;gt; and define events, &amp;lt;math&amp;gt;E=\left\{4\right\} &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F=\left\{ 4,5,6\right\} ;&amp;lt;/math&amp;gt; we are interested in &amp;lt;math&amp;gt;\Pr \left(E|F\right)&amp;lt;/math&amp;gt;. To work this out we take &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; as known. Given this knowledge the sample space becomes restricted to simply &amp;lt;math&amp;gt;\left\{ 4,5,6\right\} &amp;lt;/math&amp;gt; and, given no other information, each of these &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; outcome remains equally likely. So the required event, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;, is just one of three equally likely outcomes. It therefore seems reasonable that &amp;lt;math&amp;gt;\Pr (E|F)=\frac{1}{3}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
We shall now develop this idea more fully, using Venn Diagrams with the implied notion of area giving probability. Consider an abstract sample space, denoted by &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, with events &amp;lt;math&amp;gt;E\subset S,\,\,F\subset S&amp;lt;/math&amp;gt;. This is illustrated in the following Figure. Eventually we will want to construct the conditional probability, &amp;lt;math&amp;gt;\Pr \left( E|F\right)&amp;lt;/math&amp;gt;. Sticking with the above example that could be the probability that &amp;#039;&amp;#039;a child is underweight&amp;#039;&amp;#039;, given that &amp;#039;&amp;#039;the mother is a smoker&amp;#039;&amp;#039;. Two important areas used in the construction of this conditional probability are highlighted as &amp;lt;math&amp;gt;\mathbf{a}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{b}&amp;lt;/math&amp;gt;:&lt;br /&gt;
&lt;br /&gt;
[[File:ProbCond_venn1.jpg|frameless|400px]]&lt;br /&gt;
&lt;br /&gt;
In general, it is useful to think of &amp;lt;math&amp;gt;\Pr (E)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;\frac{area\left( E\right)}{area\left( S\right) }&amp;lt;/math&amp;gt;; and similarly for &amp;lt;math&amp;gt;\Pr (F)&amp;lt;/math&amp;gt;. The &amp;lt;math&amp;gt;\Pr (E\cap F)&amp;lt;/math&amp;gt; could equally be thought of as &amp;lt;math&amp;gt;\frac{area\left( a\right)}{area\left( S\right) }&amp;lt;/math&amp;gt;. With this in mind, consider what happens if we are now told that &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; has occurred. Incorporating this information implies that the effective sample space becomes restricted to &amp;lt;math&amp;gt;S^{*}=F&amp;lt;/math&amp;gt;, since &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; now defines what can happen. This now covers the sample area &amp;lt;math&amp;gt;a+b.&amp;lt;/math&amp;gt; On this new, restricted, sample space an outcome in &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; can only be observed if that outcome also belongs to &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;, the restricted sample space &amp;lt;math&amp;gt;S^*&amp;lt;/math&amp;gt;. And this only occurs in area &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt; which corresponds to the event &amp;lt;math&amp;gt;E\cap F&amp;lt;/math&amp;gt;. Thus the event of interest &amp;#039;&amp;#039;now&amp;#039;&amp;#039; is &amp;lt;math&amp;gt;E^{*}=E\cap F,&amp;lt;/math&amp;gt; as defined on the &amp;#039;&amp;#039;restricted &amp;#039;&amp;#039;sample space of &amp;lt;math&amp;gt;S^{*}=F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
In order to proceed with the construction of the conditional probability, &amp;lt;math&amp;gt;\Pr \left( E|F\right)&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;area(S)=z&amp;lt;/math&amp;gt;. Then, since the ratio of the area of the event of interest to that of the sample space gives probability, we have (on this restricted sample space):&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E|F) &amp;amp;=&amp;amp;\frac{area\left( E\cap F\right) }{area\left( F\right) } \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{a}{a+b} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{a/z}{\left( a+b\right) /z} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{\Pr \left( E\cap F\right) }{\Pr \left( F\right) },\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We have shown, for this example how a conditional probability can be expressed as a function of the joint probability &amp;lt;math&amp;gt;\Pr \left( E\cap F\right)&amp;lt;/math&amp;gt; and the marginal probability &amp;lt;math&amp;gt;\Pr \left( F\right)&amp;lt;/math&amp;gt;. This is a profound result and should be formulated in more general terms:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;The probability that &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; occurs, given that &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is known to have occurred, gives the &amp;#039;&amp;#039;&amp;#039;conditional probability&amp;#039;&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. This is denoted &amp;lt;math&amp;gt;Pr(E|F)&amp;lt;/math&amp;gt; and is calculated as&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (E|F)=\frac{\Pr (E\cap F)}{\Pr (F)}&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;and from the axioms of probability will generate a number lying between 0 and 1, since &amp;lt;math&amp;gt;\Pr (F)\geq \Pr (E\cap F)\geq 0.&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example: &amp;#039;&amp;#039;A Manufacturer of electrical components knows that the probability is 0.8 that an order will be ready for shipment on time and it is 0.6 that it will also be delivered on time. What is the probability that such an order will be delivered on time given that it was ready for shipment on time?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;R=&amp;lt;/math&amp;gt; READY, &amp;lt;math&amp;gt;D=&amp;lt;/math&amp;gt; DELIVERED ON TIME. &amp;lt;math&amp;gt;Pr(R)=0.8,Pr(R\cap D)=0.6.&amp;lt;/math&amp;gt; From this we need to calculate &amp;lt;math&amp;gt;Pr(D|R),&amp;lt;/math&amp;gt; using the above formula. This gives, &amp;lt;math&amp;gt;Pr(D|R)=Pr(R\cap D)/Pr(R)=6/8,\,\,&amp;lt;/math&amp;gt;or&amp;lt;math&amp;gt;\,\,75\%&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If we re-arrange the above formula for conditional probability, we obtain the so-called &amp;#039;&amp;#039;multiplication rule of probability &amp;#039;&amp;#039;for &amp;#039;&amp;#039;intersections&amp;#039;&amp;#039; of events:&lt;br /&gt;
&lt;br /&gt;
== Multiplication rule of probability ==&lt;br /&gt;
&lt;br /&gt;
The multiplication rule of probability can be stated as follows:&lt;br /&gt;
&lt;br /&gt;
* &amp;lt;math&amp;gt;\Pr (E\cap F)=\Pr (E|F)\times \Pr (F)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note that for any two events, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F,&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;(E\cap F)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;(E\cap \bar{F})&amp;lt;/math&amp;gt; are mutually exclusive with &amp;lt;math&amp;gt;E=(E\cap F)\cup (E\cap \bar{F})&amp;lt;/math&amp;gt;; this has been seen before. So the &amp;#039;&amp;#039;addition rule&amp;#039;&amp;#039; and &amp;#039;&amp;#039;multiplication rule&amp;#039;&amp;#039; of probability together give:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E) &amp;amp;=&amp;amp;\Pr (E\cap F)+\Pr (E\cap \bar{F}) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\Pr (E|F)\times \Pr (F)+\Pr (E|\bar{F})\times \Pr (\bar{F}).\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This is an extremely important and useful result, in practice, as we shall see shortly.&lt;br /&gt;
&lt;br /&gt;
== Statistical Independence ==&lt;br /&gt;
&lt;br /&gt;
If the knowledge that &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; has occurred does NOT alter our probability assessment of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are said to be (statistically) &amp;#039;&amp;#039;independent&amp;#039;&amp;#039;. In this sense, &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; carries no information about &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Formally, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are &amp;#039;&amp;#039;&amp;#039;independent&amp;#039;&amp;#039;&amp;#039; events if and only if&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;Pr(E|F)=Pr(E)&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;which, in turn is true&amp;#039;&amp;#039; if and only if &amp;#039;&amp;#039;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;Pr(E\cap F)=Pr(E)\times Pr(F).&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This concept of independence is of enormous importance in practice. Consider the case of lung cancer and its connection to smoking (apologies to all smokers for being picked upon here). The first connection between smoking and lung cancer was made in the 1920s. However, for many decades after the tobacco industry spend a lot of money and effort to convince people that there was no connection between the two. In other words they claimed that the two events are &amp;#039;&amp;#039;independent&amp;#039;&amp;#039;, or &amp;lt;math&amp;gt;Pr(Cancer|Smoking)=Pr(Cancer|\cap{Smoking})=Pr(Cancer)&amp;lt;/math&amp;gt;. It was then the task of epidemiologists to show otherwise. This was famously and comprehensively achieved by the [http://en.wikipedia.org/wiki/British_Doctors_Study &amp;#039;&amp;#039;British Doctors Study&amp;#039;&amp;#039;].&lt;br /&gt;
&lt;br /&gt;
== Bayes’ Theorem ==&lt;br /&gt;
&lt;br /&gt;
One area where conditional probability is extremely important is that of clinical trials - testing the power of a diagnostic test to detect the presence of a particular disease. Suppose, then, that a new test is being developed and let &amp;lt;math&amp;gt;P=&amp;lt;/math&amp;gt; ‘&amp;#039;&amp;#039;test positive&amp;#039;&amp;#039;’ and &amp;lt;math&amp;gt;D=&amp;lt;/math&amp;gt; ‘&amp;#039;&amp;#039;presence of disease&amp;#039;&amp;#039;’, but where the results from applying the diagnostic test can never be wholly reliable. From the point of view of our previous discussion on conditional probability, we would of course require that &amp;lt;math&amp;gt;\Pr \left( P|D\right)&amp;lt;/math&amp;gt; to be large; i.e., the test should be effective at detecting the disease. However, if you think about, this is not necessarily the probability that we might be interested in from a diagnosis point of view. Rather, we should be more interested in &amp;lt;math&amp;gt;\Pr \left( D|P\right)&amp;lt;/math&amp;gt;, the probability of correct diagnosis, and require this to be large (with, presumably, &amp;lt;math&amp;gt;\Pr (D|\bar{P})&amp;lt;/math&amp;gt; being small). Here, what we are trying to attach a probability to is a possible ‘cause’. The observed outcome is a positive test result (&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;), but the presence or non-presence of the disease is what is of interest and this is uncertain. &amp;lt;math&amp;gt;\Pr (D|P)&amp;lt;/math&amp;gt; asks the question ‘&amp;#039;&amp;#039;what is the probability that it is the presence of the disease which caused the positive test result&amp;#039;&amp;#039;’? (Another recent news-worthy example would be the effect of exposure to depleted uranium on Gulf and Balkan war veterans. Given the presence of lymph, lung or brain cancer in such individuals (&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;), how likely is that the cause was exposure to depleted uranium weapons (&amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;)? Firstly, is &amp;lt;math&amp;gt;\Pr \left( D|P\right) &amp;lt;/math&amp;gt; high or low? Secondly, might there being something else (&amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;) which could offer a “better” explanation, such that &amp;lt;math&amp;gt;\Pr \left( F|P\right) &amp;gt;\Pr \left( D|F\right) &amp;lt;/math&amp;gt; ?)&lt;br /&gt;
&lt;br /&gt;
The situation is depicted in the following Figure, in which there are two possible ‘states’ in the population: &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt; (depicted by the lighter shaded area covering the left portion of the sample space) and &amp;lt;math&amp;gt;\bar{D}.&amp;lt;/math&amp;gt; It must be that &amp;lt;math&amp;gt;D\cup \bar{D}=S,&amp;lt;/math&amp;gt; since any individual in the population either has the disease or does not. The event of an observed positive test result is denoted by the closed loop, &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;. (Notice that the shading in the diagram is relatively darker where &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; intersects with &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&lt;br /&gt;
[[File:ProbCond_venn2.jpg|frameless|400px]]&lt;br /&gt;
&lt;br /&gt;
To investigate how we might construct the required probability, &amp;lt;math&amp;gt;\Pr \left(D|P\right)&amp;lt;/math&amp;gt;, proceed as follows:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr \left( D|P\right) &amp;amp;=&amp;amp;\frac{\Pr \left( D\cap P\right) }{\Pr (P)} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{\Pr (D\cap P)}{\Pr (P\cap D)+\Pr (P\cap \bar{D})},\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
since &amp;lt;math&amp;gt;P=(P\cap D)\cup (P\cap \bar{D}),&amp;lt;/math&amp;gt; and these are mutually exclusive. From the multiplication rule of probability, &amp;lt;math&amp;gt;\Pr \left( P\cap D\right) =\Pr(P|D)\times \Pr (D),&amp;lt;/math&amp;gt; and similarly for &amp;lt;math&amp;gt;\Pr \left( P\cap \bar{D}\right)&amp;lt;/math&amp;gt;. Thus&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\Pr \left( D|P\right) =\frac{\Pr \left( P|D\right) \times \Pr \left(D\right) }{\Pr \left( P|D\right) \times \Pr \left( D\right) +\Pr (P|\bar{D})\times \Pr \left( \bar{D}\right) },&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
which may be convenient to work with since &amp;lt;math&amp;gt;\Pr \left( P|D\right) &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr\left( P|\bar{D}\right) &amp;lt;/math&amp;gt; can be estimated from clinical trials and &amp;lt;math&amp;gt;\Pr\left( D\right) &amp;lt;/math&amp;gt; estimated from recent historical survey data.&lt;br /&gt;
&lt;br /&gt;
This sort of calculation (assigning probabilities to possible causes of observed events) is an example of &amp;#039;&amp;#039;Bayes’ Theorem&amp;#039;&amp;#039;. Of course, we may have to consider more than two possible causes, and the construction of the appropriate probabilities is as follows.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Consider a sample space, &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;A,B,C&amp;lt;/math&amp;gt; are three mutually exclusive events (possible causes), defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, such that &amp;lt;math&amp;gt;S=A\cup B\cup C&amp;lt;/math&amp;gt;. In such a situation, &amp;lt;math&amp;gt;A,B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; are said to form a &amp;#039;&amp;#039;&amp;#039;partition&amp;#039;&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;. &amp;#039;&amp;#039;&amp;#039;Bayes’ Theorem&amp;#039;&amp;#039;&amp;#039; states that:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (A|E)=\frac{\Pr (E|A)\times \Pr (A)}{\left\{ \Pr (E|A)\times \Pr(A)\right\} +\left\{ \Pr (E|B)\times \Pr (B)\right\} +\left\{ \Pr(E|C)\times \Pr (C)\right\} }.&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;And, more generally, consider a sample space, &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F_{1},F_{2},...,F_{k}&amp;lt;/math&amp;gt; are &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; mutually exclusive events (possible causes), which form a partition of &amp;lt;math&amp;gt;S:S=\bigcup_{j=1}^{k}F_{j}&amp;lt;/math&amp;gt;. &amp;#039;&amp;#039;&amp;#039;Bayes’ Theorem&amp;#039;&amp;#039;&amp;#039; then states that:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (F_{j}|E)=\frac{\Pr (E|F_{j})\times \Pr (F_{j})}{\sum_{s=1}^{k}\left\{\Pr (E|F_{s})\times \Pr (F_{s})\right\} }.&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
From the above formula, you should be able to satisfy yourself that &amp;lt;math&amp;gt;\sum_{j=1}^{k}\Pr \left( F_{j}|E\right) =1.&amp;lt;/math&amp;gt; If this is not at first clear, consider case (1) and show that &amp;lt;math&amp;gt;\Pr \left( A|E\right) +\Pr \left(&lt;br /&gt;
B|E\right) +\Pr \left( C|E\right) =1.&amp;lt;/math&amp;gt; The reason for this is that since &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; form a partition of &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; they must also form a partition of any event &amp;lt;math&amp;gt;E\subset S.&amp;lt;/math&amp;gt; In the above conditional probabilities, we are regarding &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; as the restricted sample space and therefore the probabilities assigned the mutually exclusive events &amp;lt;math&amp;gt;\left( A,B,C\right) &amp;lt;/math&amp;gt; which &amp;#039;&amp;#039;cover&amp;#039;&amp;#039; this (restricted) sample space, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, must sum to 1.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example&amp;#039;&amp;#039;: Box A contains 2 red balls. Box B contains 1 red and 1 white ball. Box A and Box B are identical. If a box is selected at random and one ball is withdrawn from it, what is the probability that the selected box was number 1 if the ball withdrawn from it turns out to be red?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; be the event of selecting Box A and &amp;lt;math&amp;gt;R&amp;lt;/math&amp;gt; the event of drawing a red ball. Require &amp;lt;math&amp;gt;Pr(A|R)&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;Pr(A|R)=Pr(A\cap R)/Pr(R);&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;Pr(A\cap R)=Pr(A)Pr(R|A)=(1/2)\times 1=1/2.&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
And,&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
Pr(R) &amp;amp;=&amp;amp;Pr(A\cap R)+Pr(\bar{A}\cap R) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\Pr (A)\times \Pr (R|A)\,\,\,\,+\,\,\,\,\Pr (\bar{A})\times \Pr (R|\bar{A}) \\&lt;br /&gt;
&amp;amp;=&amp;amp;(1/2)\,\,\,\,+\,\,\,\,(1/2)\times (1/2) \\&lt;br /&gt;
&amp;amp;=&amp;amp;3/4.\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Therefore, &amp;lt;math&amp;gt;\Pr (A|R)=(1/2)/(3/4)=2/3&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=File:ProbCond_venn2.jpg&amp;diff=2867</id>
		<title>File:ProbCond venn2.jpg</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=File:ProbCond_venn2.jpg&amp;diff=2867"/>
				<updated>2013-08-09T09:58:25Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=File:ProbCond_venn1.jpg&amp;diff=2866</id>
		<title>File:ProbCond venn1.jpg</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=File:ProbCond_venn1.jpg&amp;diff=2866"/>
				<updated>2013-08-09T09:58:11Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Conditional&amp;diff=2865</id>
		<title>Probability Conditional</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Conditional&amp;diff=2865"/>
				<updated>2013-08-09T09:57:26Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: Created page with &amp;quot;  = Conditional Probability =  An important consideration in the development of probability is that of &amp;#039;&amp;#039;conditional probability&amp;#039;&amp;#039;. This refers to the calculation of updating ...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Conditional Probability =&lt;br /&gt;
&lt;br /&gt;
An important consideration in the development of probability is that of &amp;#039;&amp;#039;conditional probability&amp;#039;&amp;#039;. This refers to the calculation of updating probabilities in the light of revealed information. For example, insurance companies nearly always set their home contents insurance premiums on the basis of the postcode in which the home is located. That is to say, insurance companies believe the risk depends upon the location; i.e., the probability of property crime is assessed conditional upon the location of the property. (A similar calculation is made to set car insurance premiums.) As a result, the premiums for two identical households located in different parts of the country can differ substantially.&lt;br /&gt;
&lt;br /&gt;
* In general, the probability of an event, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, occurring &amp;#039;&amp;#039;given&amp;#039;&amp;#039; that an event, &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;, has occurred is called the &amp;#039;&amp;#039;conditional probability&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; and is denoted &amp;lt;math&amp;gt;\Pr (E|F)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
As another example, it has been well documented that the ability of a new born baby to survive is closely associated with its birth-weight. A birth-weight of less than 1500&amp;#039;&amp;#039;g&amp;#039;&amp;#039; is regarded as dangerously low. Consider &amp;lt;math&amp;gt;E=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;birth weight of a baby is less than&amp;#039;&amp;#039; 1500&amp;#039;&amp;#039;g&amp;#039;&amp;#039;, &amp;lt;math&amp;gt;F=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;mother smoked during pregnancy&amp;#039;&amp;#039;; then evidence as to whether &amp;lt;math&amp;gt;\Pr(E|F)&amp;gt;\Pr (E|\bar{F})&amp;lt;/math&amp;gt; is of considerable interest.&lt;br /&gt;
&lt;br /&gt;
As a preliminary to the main development, consider the simple experiment of rolling a fair die and observing the number of dots on the upturned face. Then &amp;lt;math&amp;gt;S=\left\{ 1,2,3,4,5,6\right\} &amp;lt;/math&amp;gt; and define events, &amp;lt;math&amp;gt;E=\left\{4\right\} &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F=\left\{ 4,5,6\right\} ;&amp;lt;/math&amp;gt; we are interested in &amp;lt;math&amp;gt;\Pr \left(E|F\right)&amp;lt;/math&amp;gt;. To work this out we take &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; as known. Given this knowledge the sample space becomes restricted to simply &amp;lt;math&amp;gt;\left\{ 4,5,6\right\} &amp;lt;/math&amp;gt; and, given no other information, each of these &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; outcome remains equally likely. So the required event, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;, is just one of three equally likely outcomes. It therefore seems reasonable that &amp;lt;math&amp;gt;\Pr (E|F)=\frac{1}{3}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
We shall now develop this idea more fully, using Venn Diagrams with the implied notion of area giving probability. Consider an abstract sample space, denoted by &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, with events &amp;lt;math&amp;gt;E\subset S,\,\,F\subset S&amp;lt;/math&amp;gt;. This is illustrated in the following Figure. Eventually we will want to construct the conditional probability, &amp;lt;math&amp;gt;\Pr \left( E|F\right)&amp;lt;/math&amp;gt;. Sticking with the above example that could be the probability that &amp;#039;&amp;#039;a child is underweight&amp;#039;&amp;#039;, given that &amp;#039;&amp;#039;the mother is a smoker&amp;#039;&amp;#039;. Two important areas used in the construction of this conditional probability are highlighted as &amp;lt;math&amp;gt;\mathbf{a}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{b}&amp;lt;/math&amp;gt;:&lt;br /&gt;
&lt;br /&gt;
[[File:ProbCond_venn1.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
In general, it is useful to think of &amp;lt;math&amp;gt;\Pr (E)&amp;lt;/math&amp;gt; as &amp;lt;math&amp;gt;\frac{area\left( E\right)}{area\left( S\right) }&amp;lt;/math&amp;gt;; and similarly for &amp;lt;math&amp;gt;\Pr (F)&amp;lt;/math&amp;gt;. The &amp;lt;math&amp;gt;\Pr (E\cap F)&amp;lt;/math&amp;gt; could equally be thought of as &amp;lt;math&amp;gt;\frac{area\left( a\right)}{area\left( S\right) }&amp;lt;/math&amp;gt;. With this in mind, consider what happens if we are now told that &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; has occurred. Incorporating this information implies that the effective sample space becomes restricted to &amp;lt;math&amp;gt;S^{*}=F&amp;lt;/math&amp;gt;, since &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; now defines what can happen. This now covers the sample area &amp;lt;math&amp;gt;a+b.&amp;lt;/math&amp;gt; On this new, restricted, sample space an outcome in &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; can only be observed if that outcome also belongs to &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;, the restricted sample space &amp;lt;math&amp;gt;S^*&amp;lt;/math&amp;gt;. And this only occurs in area &amp;lt;math&amp;gt;a&amp;lt;/math&amp;gt; which corresponds to the event &amp;lt;math&amp;gt;E\cap F&amp;lt;/math&amp;gt;. Thus the event of interest &amp;#039;&amp;#039;now&amp;#039;&amp;#039; is &amp;lt;math&amp;gt;E^{*}=E\cap F,&amp;lt;/math&amp;gt; as defined on the &amp;#039;&amp;#039;restricted &amp;#039;&amp;#039;sample space of &amp;lt;math&amp;gt;S^{*}=F&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
In order to proceed with the construction of the conditional probability, &amp;lt;math&amp;gt;\Pr \left( E|F\right)&amp;lt;/math&amp;gt;, let &amp;lt;math&amp;gt;area(S)=z&amp;lt;/math&amp;gt;. Then, since the ratio of the area of the event of interest to that of the sample space gives probability, we have (on this restricted sample space):&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E|F) &amp;amp;=&amp;amp;\frac{area\left( E\cap F\right) }{area\left( F\right) } \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{a}{a+b} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{a/z}{\left( a+b\right) /z} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{\Pr \left( E\cap F\right) }{\Pr \left( F\right) },\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We have shown, for this example how a conditional probability can be expressed as a function of the joint probability &amp;lt;math&amp;gt;\Pr \left( E\cap F\right)&amp;lt;/math&amp;gt; and the marginal probability &amp;lt;math&amp;gt;\Pr \left( F\right)&amp;lt;/math&amp;gt;. This is a profound result and should be formulated in more general terms:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;The probability that &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; occurs, given that &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is known to have occurred, gives the &amp;#039;&amp;#039;&amp;#039;conditional probability&amp;#039;&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; given &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;. This is denoted &amp;lt;math&amp;gt;Pr(E|F)&amp;lt;/math&amp;gt; and is calculated as&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (E|F)=\frac{\Pr (E\cap F)}{\Pr (F)}&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;and from the axioms of probability will generate a number lying between 0 and 1, since &amp;lt;math&amp;gt;\Pr (F)\geq \Pr (E\cap F)\geq 0.&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example: &amp;#039;&amp;#039;A Manufacturer of electrical components knows that the probability is 0.8 that an order will be ready for shipment on time and it is 0.6 that it will also be delivered on time. What is the probability that such an order will be delivered on time given that it was ready for shipment on time?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;R=&amp;lt;/math&amp;gt; READY, &amp;lt;math&amp;gt;D=&amp;lt;/math&amp;gt; DELIVERED ON TIME. &amp;lt;math&amp;gt;Pr(R)=0.8,Pr(R\cap D)=0.6.&amp;lt;/math&amp;gt; From this we need to calculate &amp;lt;math&amp;gt;Pr(D|R),&amp;lt;/math&amp;gt; using the above formula. This gives, &amp;lt;math&amp;gt;Pr(D|R)=Pr(R\cap D)/Pr(R)=6/8,\,\,&amp;lt;/math&amp;gt;or&amp;lt;math&amp;gt;\,\,75\%&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
If we re-arrange the above formula for conditional probability, we obtain the so-called &amp;#039;&amp;#039;multiplication rule of probability &amp;#039;&amp;#039;for &amp;#039;&amp;#039;intersections&amp;#039;&amp;#039; of events:&lt;br /&gt;
&lt;br /&gt;
== Multiplication rule of probability ==&lt;br /&gt;
&lt;br /&gt;
The multiplication rule of probability can be stated as follows:&lt;br /&gt;
&lt;br /&gt;
* &amp;lt;math&amp;gt;\Pr (E\cap F)=\Pr (E|F)\times \Pr (F)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Note that for any two events, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F,&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;(E\cap F)&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;(E\cap \bar{F})&amp;lt;/math&amp;gt; are mutually exclusive with &amp;lt;math&amp;gt;E=(E\cap F)\cup (E\cap \bar{F})&amp;lt;/math&amp;gt;; this has been seen before. So the &amp;#039;&amp;#039;addition rule&amp;#039;&amp;#039; and &amp;#039;&amp;#039;multiplication rule&amp;#039;&amp;#039; of probability together give:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E) &amp;amp;=&amp;amp;\Pr (E\cap F)+\Pr (E\cap \bar{F}) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\Pr (E|F)\times \Pr (F)+\Pr (E|\bar{F})\times \Pr (\bar{F}).\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This is an extremely important and useful result, in practice, as we shall see shortly.&lt;br /&gt;
&lt;br /&gt;
== Statistical Independence ==&lt;br /&gt;
&lt;br /&gt;
If the knowledge that &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; has occurred does NOT alter our probability assessment of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are said to be (statistically) &amp;#039;&amp;#039;independent&amp;#039;&amp;#039;. In this sense, &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; carries no information about &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Formally, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are &amp;#039;&amp;#039;&amp;#039;independent&amp;#039;&amp;#039;&amp;#039; events if and only if&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;Pr(E|F)=Pr(E)&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;which, in turn is true&amp;#039;&amp;#039; if and only if &amp;#039;&amp;#039;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;Pr(E\cap F)=Pr(E)\times Pr(F).&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This concept of independence is of enormous importance in practice. Consider the case of lung cancer and its connection to smoking (apologies to all smokers for being picked upon here). The first connection between smoking and lung cancer was made in the 1920s. However, for many decades after the tobacco industry spend a lot of money and effort to convince people that there was no connection between the two. In other words they claimed that the two events are &amp;#039;&amp;#039;independent&amp;#039;&amp;#039;, or &amp;lt;math&amp;gt;Pr(Cancer|Smoking)=Pr(Cancer|\cap{Smoking})=Pr(Cancer)&amp;lt;/math&amp;gt;. It was then the task of epidemiologists to show otherwise. This was famously and comprehensively achieved by the [http://en.wikipedia.org/wiki/British_Doctors_Study &amp;#039;&amp;#039;British Doctors Study&amp;#039;&amp;#039;].&lt;br /&gt;
&lt;br /&gt;
== Bayes’ Theorem ==&lt;br /&gt;
&lt;br /&gt;
One area where conditional probability is extremely important is that of clinical trials - testing the power of a diagnostic test to detect the presence of a particular disease. Suppose, then, that a new test is being developed and let &amp;lt;math&amp;gt;P=&amp;lt;/math&amp;gt; ‘&amp;#039;&amp;#039;test positive&amp;#039;&amp;#039;’ and &amp;lt;math&amp;gt;D=&amp;lt;/math&amp;gt; ‘&amp;#039;&amp;#039;presence of disease&amp;#039;&amp;#039;’, but where the results from applying the diagnostic test can never be wholly reliable. From the point of view of our previous discussion on conditional probability, we would of course require that &amp;lt;math&amp;gt;\Pr \left( P|D\right)&amp;lt;/math&amp;gt; to be large; i.e., the test should be effective at detecting the disease. However, if you think about, this is not necessarily the probability that we might be interested in from a diagnosis point of view. Rather, we should be more interested in &amp;lt;math&amp;gt;\Pr \left( D|P\right)&amp;lt;/math&amp;gt;, the probability of correct diagnosis, and require this to be large (with, presumably, &amp;lt;math&amp;gt;\Pr (D|\bar{P})&amp;lt;/math&amp;gt; being small). Here, what we are trying to attach a probability to is a possible ‘cause’. The observed outcome is a positive test result (&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;), but the presence or non-presence of the disease is what is of interest and this is uncertain. &amp;lt;math&amp;gt;\Pr (D|P)&amp;lt;/math&amp;gt; asks the question ‘&amp;#039;&amp;#039;what is the probability that it is the presence of the disease which caused the positive test result&amp;#039;&amp;#039;’? (Another recent news-worthy example would be the effect of exposure to depleted uranium on Gulf and Balkan war veterans. Given the presence of lymph, lung or brain cancer in such individuals (&amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;), how likely is that the cause was exposure to depleted uranium weapons (&amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;)? Firstly, is &amp;lt;math&amp;gt;\Pr \left( D|P\right) &amp;lt;/math&amp;gt; high or low? Secondly, might there being something else (&amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;) which could offer a “better” explanation, such that &amp;lt;math&amp;gt;\Pr \left( F|P\right) &amp;gt;\Pr \left( D|F\right) &amp;lt;/math&amp;gt; ?)&lt;br /&gt;
&lt;br /&gt;
The situation is depicted in the following Figure, in which there are two possible ‘states’ in the population: &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt; (depicted by the lighter shaded area covering the left portion of the sample space) and &amp;lt;math&amp;gt;\bar{D}.&amp;lt;/math&amp;gt; It must be that &amp;lt;math&amp;gt;D\cup \bar{D}=S,&amp;lt;/math&amp;gt; since any individual in the population either has the disease or does not. The event of an observed positive test result is denoted by the closed loop, &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt;. (Notice that the shading in the diagram is relatively darker where &amp;lt;math&amp;gt;P&amp;lt;/math&amp;gt; intersects with &amp;lt;math&amp;gt;D&amp;lt;/math&amp;gt;.)&lt;br /&gt;
&lt;br /&gt;
[[File:ProbCond_venn2.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
To investigate how we might construct the required probability, &amp;lt;math&amp;gt;\Pr \left(D|P\right)&amp;lt;/math&amp;gt;, proceed as follows:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr \left( D|P\right) &amp;amp;=&amp;amp;\frac{\Pr \left( D\cap P\right) }{\Pr (P)} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{\Pr (D\cap P)}{\Pr (P\cap D)+\Pr (P\cap \bar{D})},\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
since &amp;lt;math&amp;gt;P=(P\cap D)\cup (P\cap \bar{D}),&amp;lt;/math&amp;gt; and these are mutually exclusive. From the multiplication rule of probability, &amp;lt;math&amp;gt;\Pr \left( P\cap D\right) =\Pr(P|D)\times \Pr (D),&amp;lt;/math&amp;gt; and similarly for &amp;lt;math&amp;gt;\Pr \left( P\cap \bar{D}\right)&amp;lt;/math&amp;gt;. Thus&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\Pr \left( D|P\right) =\frac{\Pr \left( P|D\right) \times \Pr \left(D\right) }{\Pr \left( P|D\right) \times \Pr \left( D\right) +\Pr (P|\bar{D})\times \Pr \left( \bar{D}\right) },&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
which may be convenient to work with since &amp;lt;math&amp;gt;\Pr \left( P|D\right) &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr\left( P|\bar{D}\right) &amp;lt;/math&amp;gt; can be estimated from clinical trials and &amp;lt;math&amp;gt;\Pr\left( D\right) &amp;lt;/math&amp;gt; estimated from recent historical survey data.&lt;br /&gt;
&lt;br /&gt;
This sort of calculation (assigning probabilities to possible causes of observed events) is an example of &amp;#039;&amp;#039;Bayes’ Theorem&amp;#039;&amp;#039;. Of course, we may have to consider more than two possible causes, and the construction of the appropriate probabilities is as follows.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Consider a sample space, &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;A,B,C&amp;lt;/math&amp;gt; are three mutually exclusive events (possible causes), defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, such that &amp;lt;math&amp;gt;S=A\cup B\cup C&amp;lt;/math&amp;gt;. In such a situation, &amp;lt;math&amp;gt;A,B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; are said to form a &amp;#039;&amp;#039;&amp;#039;partition&amp;#039;&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;. &amp;#039;&amp;#039;&amp;#039;Bayes’ Theorem&amp;#039;&amp;#039;&amp;#039; states that:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (A|E)=\frac{\Pr (E|A)\times \Pr (A)}{\left\{ \Pr (E|A)\times \Pr(A)\right\} +\left\{ \Pr (E|B)\times \Pr (B)\right\} +\left\{ \Pr(E|C)\times \Pr (C)\right\} }.&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;And, more generally, consider a sample space, &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, where &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F_{1},F_{2},...,F_{k}&amp;lt;/math&amp;gt; are &amp;lt;math&amp;gt;k&amp;lt;/math&amp;gt; mutually exclusive events (possible causes), which form a partition of &amp;lt;math&amp;gt;S:S=\bigcup_{j=1}^{k}F_{j}&amp;lt;/math&amp;gt;. &amp;#039;&amp;#039;&amp;#039;Bayes’ Theorem&amp;#039;&amp;#039;&amp;#039; then states that:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (F_{j}|E)=\frac{\Pr (E|F_{j})\times \Pr (F_{j})}{\sum_{s=1}^{k}\left\{\Pr (E|F_{s})\times \Pr (F_{s})\right\} }.&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
From the above formula, you should be able to satisfy yourself that &amp;lt;math&amp;gt;\sum_{j=1}^{k}\Pr \left( F_{j}|E\right) =1.&amp;lt;/math&amp;gt; If this is not at first clear, consider case (1) and show that &amp;lt;math&amp;gt;\Pr \left( A|E\right) +\Pr \left(&lt;br /&gt;
B|E\right) +\Pr \left( C|E\right) =1.&amp;lt;/math&amp;gt; The reason for this is that since &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;B&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt; form a partition of &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; they must also form a partition of any event &amp;lt;math&amp;gt;E\subset S.&amp;lt;/math&amp;gt; In the above conditional probabilities, we are regarding &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; as the restricted sample space and therefore the probabilities assigned the mutually exclusive events &amp;lt;math&amp;gt;\left( A,B,C\right) &amp;lt;/math&amp;gt; which &amp;#039;&amp;#039;cover&amp;#039;&amp;#039; this (restricted) sample space, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, must sum to 1.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example&amp;#039;&amp;#039;: Box A contains 2 red balls. Box B contains 1 red and 1 white ball. Box A and Box B are identical. If a box is selected at random and one ball is withdrawn from it, what is the probability that the selected box was number 1 if the ball withdrawn from it turns out to be red?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Let &amp;lt;math&amp;gt;A&amp;lt;/math&amp;gt; be the event of selecting Box A and &amp;lt;math&amp;gt;R&amp;lt;/math&amp;gt; the event of drawing a red ball. Require &amp;lt;math&amp;gt;Pr(A|R)&amp;lt;/math&amp;gt;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;Pr(A|R)=Pr(A\cap R)/Pr(R);&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;Pr(A\cap R)=Pr(A)Pr(R|A)=(1/2)\times 1=1/2.&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
And,&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
Pr(R) &amp;amp;=&amp;amp;Pr(A\cap R)+Pr(\bar{A}\cap R) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\Pr (A)\times \Pr (R|A)\,\,\,\,+\,\,\,\,\Pr (\bar{A})\times \Pr (R|\bar{A}) \\&lt;br /&gt;
&amp;amp;=&amp;amp;(1/2)\,\,\,\,+\,\,\,\,(1/2)\times (1/2) \\&lt;br /&gt;
&amp;amp;=&amp;amp;3/4.\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Therefore, &amp;lt;math&amp;gt;\Pr (A|R)=(1/2)/(3/4)=2/3&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2863</id>
		<title>Probability Intro</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2863"/>
				<updated>2013-08-07T21:29:31Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: /* Probability */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Introducing Probability =&lt;br /&gt;
&lt;br /&gt;
So far we have been looking at ways of summarising samples of data drawn from an underlying population of interest. Although at times tedious, all such arithmetic calculations are fairly mechanical and straightforward to apply. To remind ourselves, one of the primary reasons for wishing to summarise data is so assist in the development of inferences about the population from which the data were taken. That is to say, we would like to elicit some information about the mechanism which generated the observed data.&lt;br /&gt;
&lt;br /&gt;
We now start on the process of developing mathematical ways of formulating inferences and this requires the use of &amp;#039;&amp;#039;probability&amp;#039;&amp;#039;. This becomes clear if we think back to one of the early questions posed in this course: &amp;#039;&amp;#039;prior to sampling is it possible to predict with absolute certainty what will be observed&amp;#039;&amp;#039;? The answer to this question is &amp;#039;&amp;#039;no&amp;#039;&amp;#039;; although it would be of interest to know how &amp;#039;&amp;#039;likely&amp;#039;&amp;#039; it is that certain values would be observed. Or, what is the &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; of observing certain values?&lt;br /&gt;
&lt;br /&gt;
Before proceeding, we need some more tools:&lt;br /&gt;
&lt;br /&gt;
= Venn diagrams =&lt;br /&gt;
&lt;br /&gt;
Venn diagrams (and diagrams in general) are of enormous help in trying to understand, and manipulate probability. We begin with some basic definitions, some of which we have encountered before.&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Experiment:&amp;#039;&amp;#039;&amp;#039; any process which, when applied, provides data or an outcome; e.g., rolling a die and observing the number of dots on the upturned face; recording the amount of rainfall in Manchester over a period of time.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Sample Space:&amp;#039;&amp;#039;&amp;#039; set of possible outcomes of an experiment; e.g., &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; (or &amp;lt;math&amp;gt;\Omega &amp;lt;/math&amp;gt;) &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{1,2,3,4,5,6\}&amp;lt;/math&amp;gt;, which is the sample space of rolling a dice. Or &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{x;x\geq 0\}&amp;lt;/math&amp;gt;, which is the sample space of an experiment where the outcomes can be any real non-negative number, or ‘&amp;#039;&amp;#039;the set of real non-negative real numbers&amp;#039;&amp;#039;’.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Simple Event&amp;#039;&amp;#039;&amp;#039;: just one of the possible outcomes on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Event:&amp;#039;&amp;#039;&amp;#039; a &amp;#039;&amp;#039;subset&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, denoted &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;; e.g., &amp;lt;math&amp;gt;E=\left\{ 2,4,6\right\}&amp;lt;/math&amp;gt; (i.e. any even number on a dice) or &amp;lt;math&amp;gt;E=\left\{ x;4&amp;lt;x\leq 10\right\} ,&amp;lt;/math&amp;gt; which means ‘&amp;#039;&amp;#039;the set of real numbers which are strictly bigger than&amp;#039;&amp;#039; &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;but less than or equal to &amp;#039;&amp;#039;&amp;lt;math&amp;gt;10&amp;lt;/math&amp;gt;’.&amp;lt;br /&amp;gt;Note that an event, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, is a collection of simple events.&lt;br /&gt;
&lt;br /&gt;
Such concepts can be represented by means of the following Venn Diagram:&lt;br /&gt;
&lt;br /&gt;
[[File:Venn_1.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
The sample space, &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; is depicted as a closed rectangle, and the event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; is a closed loop wholly contained within &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; and we write (in set notation) &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
In dealing with probability, and in particular the probability of an event (or events) occurring, we shall need to be familiar with &amp;#039;&amp;#039;&amp;#039;UNIONS, INTERSECTIONS&amp;#039;&amp;#039;&amp;#039; and &amp;#039;&amp;#039;&amp;#039;COMPLEMENTS&amp;#039;&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
To illustrate these concepts, consider the sample space &amp;lt;math&amp;gt;S=\{x;x\geq 0\},\,&amp;lt;/math&amp;gt; with the following events defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as depicted in Figure 3.2:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E=\{x;4&amp;lt;x\leq 10\},\,F=\{x;7&amp;lt;x\leq 17\},\,G=\{x;x&amp;gt;15\},\,H=\{x;9&amp;lt;x\leq&lt;br /&gt;
13\}. &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| (a) Event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: A closed loop&lt;br /&gt;
| (b) Union: &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2a.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2b.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (c) Intersection: &amp;lt;math&amp;gt;E\cap F&amp;lt;/math&amp;gt;&lt;br /&gt;
| (d) The Null set/event: &amp;lt;math&amp;gt;E\cap G=\emptyset &amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2c.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2d.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (e) Complement of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;&lt;br /&gt;
| (f) Subset of &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;H\subset F&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;H\cap F=H&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2e.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2f.jpg|frameless|300px]]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
        &lt;br /&gt;
&lt;br /&gt;
* The &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cup F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cup F=\{x;4&amp;lt;x\leq 17\};&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are either in &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or in &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) in both. This is illustrated on the Venn diagram by the dark shaded area in diagram (b).&lt;br /&gt;
* The&amp;#039;&amp;#039; intersection&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cap F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cap F=\left\{ x;7\leq x\leq 10\right\} ;&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are common to both &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; Again this is depicted by the dark shaded area in (c). If events have no elements in common (as, for example, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt;) then they are said to be &amp;#039;&amp;#039;mutually exclusive&amp;#039;&amp;#039;, and we can write &amp;lt;math&amp;gt;E\cap G=\emptyset ,&amp;lt;/math&amp;gt; meaning the &amp;#039;&amp;#039;null set&amp;#039;&amp;#039; which contains no elements. Such a situation is illustrated on the Venn Diagram by events (the two shaded closed loops in (d)) which do not overlap. Notice however that &amp;lt;math&amp;gt;G\cap F\neq \emptyset ,&amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have elements in common.&lt;br /&gt;
* The &amp;#039;&amp;#039;complement&amp;#039;&amp;#039; of an event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; say, is everything defined on the sample space which is not in &amp;lt;math&amp;gt;E.&amp;lt;/math&amp;gt; This event is denoted &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;, the dark shaded area in (e); here &amp;lt;math&amp;gt;\bar{E}=\left\{ x;x\leq 4\right\} \cup \left\{ x;x&amp;gt;10\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
* Finally note that &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; is a sub-set of &amp;lt;math&amp;gt;F;&amp;lt;/math&amp;gt; see (f). It is depicted as the dark closed loop wholly contained within &amp;lt;math&amp;gt;F,&amp;lt;/math&amp;gt; the lighter shaded area, so that &amp;lt;math&amp;gt;H\cap F=H;&amp;lt;/math&amp;gt; if an element in the sample space is a member of &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; then it must also be member of &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; (In mathematical logic, we employ this scenario to indicate that “&amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; implies &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;”, but not necessarily vice-versa.) Notice that &amp;lt;math&amp;gt;G\cap H=\emptyset &amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;H\cap E\neq \emptyset&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
= Probability =&lt;br /&gt;
&lt;br /&gt;
The term &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; (or some equivalent) is used in everyday conversation and so can not be unfamiliar to the reader. We talk of the probability, or chance, of rain; the likelihood of England winning the World Cup; or, perhaps more scientifically, the chance of getting a &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; when rolling a die. What we shall now do is develop a coherent theory of probability; a theory which allows us to combine and manipulate probabilities in a consistent and meaningful manner. We shall describe ways of dealing with, and describing, uncertainty. This will involve &amp;#039;&amp;#039;rules&amp;#039;&amp;#039; which govern our use of terms like probability.&lt;br /&gt;
&lt;br /&gt;
There have been a number of different approaches (interpretations) of probability. Most depend, at least to some extent, on the notion of relative frequency as now described:&lt;br /&gt;
&lt;br /&gt;
* Suppose an experiment has an outcome of interest &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;. The &amp;#039;&amp;#039;relative frequency interpretation&amp;#039;&amp;#039; of probability says that assuming the experiment can be repeated a large number of times then the relative frequency of observing the outcome &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; will settle down to a &amp;#039;&amp;#039;number&amp;#039;&amp;#039;, denoted &amp;lt;math&amp;gt;\Pr (E),&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;P(E)&amp;lt;/math&amp;gt; or Prob&amp;lt;math&amp;gt;(E),&amp;lt;/math&amp;gt; called the &amp;#039;&amp;#039;&amp;#039;probability&amp;#039;&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
This is illustrated in the next Figure where the proportion of heads obtained after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; flips of a fair coin is plotted against &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;, as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; increases; e.g., of the first 100 flips, 46 were heads (&amp;lt;math&amp;gt;46\%&amp;lt;/math&amp;gt;). Notice that the plot becomes less ‘wobbly’ after about &amp;lt;math&amp;gt;n=&amp;lt;/math&amp;gt;140 and appears to be settling down to the value of &amp;lt;math&amp;gt;\frac{1}{2}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
[[File:Prob_coin.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
Due to this interpretation of probability, we often use observed sample proportions to approximate underlying probabilities of interest; see, for example, Question 4 of Exercise 2. There are, of course, other interpretations of probability; e.g., the subjective interpretation which simply expresses the strength of one’s belief about an event of interest such as whether Manchester United will win the European Cup! Any one of these interpretations can be used in practical situations provided the implied notion of probability follows a simple set of &amp;#039;&amp;#039;axioms&amp;#039;&amp;#039; or &amp;#039;&amp;#039;rules&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
== The axioms of probability ==&lt;br /&gt;
&lt;br /&gt;
There are just &amp;#039;&amp;#039;three &amp;#039;&amp;#039;basic rules that must be obeyed when dealing with probabilities:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;For any event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; i.e., &amp;lt;math&amp;gt;E\subset S,\,\,\Pr (E)\geq 0&amp;lt;/math&amp;gt;; &amp;#039;&amp;#039;probabilities are non-negative&amp;#039;&amp;#039;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (S)=1;&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;having defined the sample space of outcomes, one of these outcomes must be observed&amp;#039;&amp;#039;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;If events &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are mutually exclusive defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, so that &amp;lt;math&amp;gt;E\cap F=\emptyset &amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\Pr \left( E\right)+\Pr \left( F\right) .&amp;lt;/math&amp;gt; In general, for any set of mutually exclusive events, &amp;lt;math&amp;gt;E_{1},E_{2},\ldots ,E_{k},&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S:&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (E_{1}\cup E_{2}\cup \ldots \cup E_{k})=\Pr (E_{1})+\Pr (E_{2})+\ldots\Pr (E_{k})&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;i.e., &amp;lt;math&amp;gt;\Pr \left( \bigcup_{j=1}^{k}E_{j}\right) =\sum_{j=1}^{k}\Pr (E_{j}).&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In terms of the Venn Diagram, one can (and should) usefully think of the area of &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; relative to that of &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as providing an indication of probability. (Note, from axiom 2, that the area of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is implicitly normalised to be unity).&lt;br /&gt;
&lt;br /&gt;
Also observe that, contrary to what you may have believed, it is not one of the rules that &amp;lt;math&amp;gt;\Pr (E)\leq 1&amp;lt;/math&amp;gt; for any event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;. Rather, this is an implication of the &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; rules given:&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;implications: &amp;#039;&amp;#039;&amp;#039;it must be that for any event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;E\cap \bar{E}=\emptyset &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;E\cup \bar{E}=S.&amp;lt;/math&amp;gt; By Axiom &amp;lt;math&amp;gt;1,&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr (E)\geq 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr \left( \bar{E}\right) \geq 0&amp;lt;/math&amp;gt; and by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr(E)+\Pr (\bar{E})=\Pr (S).&amp;lt;/math&amp;gt; So &amp;lt;math&amp;gt;\Pr \left( E\right) +\Pr \left( \bar{E}\right) =1,&amp;lt;/math&amp;gt; by Axiom &amp;lt;math&amp;gt;2.&amp;lt;/math&amp;gt; This implies that&lt;br /&gt;
&lt;br /&gt;
# &amp;lt;math&amp;gt;0\leq \Pr (E)\leq 1&amp;lt;/math&amp;gt;&lt;br /&gt;
# &amp;lt;math&amp;gt;\Pr (\bar{E})=1-\Pr (E)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The first of these is what we might have expected from probability (a number lying between &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;). The second implication is also very important; it says that the probability of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; not happening is ‘&amp;#039;&amp;#039;one minus the probability of it happening&amp;#039;&amp;#039;’. Thus when rolling a die, the probability of getting &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; is one minus the probability of getting either a &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;2&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;5.&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
These axioms imply how to calculate probabilities on a sample space of equally likely outcomes. For example, and as we have already noted, the experiment of rolling a fair die defines a sample space of six, mutually exclusive and equally likely outcomes (&amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; dots on the up-turned face). The axioms then say that each of the six probabilities are positive, add to 1 and are all the same. Thus, the probability of any one of the outcomes must be simply &amp;lt;math&amp;gt;\frac{1}{6};&amp;lt;/math&amp;gt; which may accord with your intuition. A similar sort of analysis reveals that the probability of drawing a club from a deck of &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; cards is &amp;lt;math&amp;gt;\frac{13}{52},&amp;lt;/math&amp;gt; since any one of the &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; cards has an equal chance of being drawn and &amp;lt;math&amp;gt;13&amp;lt;/math&amp;gt; of them are clubs; i.e., &amp;lt;math&amp;gt;13&amp;lt;/math&amp;gt; of the &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; are clubs, so the probability of drawing a club is &amp;lt;math&amp;gt;\frac{13}{52}.&amp;lt;/math&amp;gt; Notice the importance of the assumption of equally likely outcomes here.&lt;br /&gt;
&lt;br /&gt;
In this, and the next section of notes, we shall see how these axioms can be used. Firstly, consider the construction of a probability for the &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of two events; i.e., the probability that &amp;#039;&amp;#039;either &amp;#039;&amp;#039;&amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) &amp;#039;&amp;#039;both &amp;#039;&amp;#039;will occur. Such a probability is embodied in the &amp;#039;&amp;#039;addition rule of probability&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
== The addition rule of probability ==&lt;br /&gt;
&lt;br /&gt;
When rolling a fair die, let &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; denote the event of an “odd number of dots” and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; the event of the “number of dots being greater than, or equal, to &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;”&amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt; What is the probability of the event &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;? To calculate this we can collect together all the mutually exclusive (simple) events which comprise &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;, and then add up the probabilities (by axiom 3). These simple events are &amp;lt;math&amp;gt;1,3,4,5&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; dots. Each has a probability of &amp;lt;math&amp;gt;\frac{1}{6},&amp;lt;/math&amp;gt; so the required total probability is: &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\frac{5}{6}&amp;lt;/math&amp;gt;. Consider carefully how this probability is constructed and note, in particular, that &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) \neq \Pr \left( E\right) +\Pr \left( F\right) &amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have a simple event in common (namely &amp;lt;math&amp;gt;5&amp;lt;/math&amp;gt; dots).&lt;br /&gt;
&lt;br /&gt;
In general, we can calculate the probability of the union of events using the &amp;#039;&amp;#039;addition rule of probability&amp;#039;&amp;#039;, as follows.&lt;br /&gt;
&lt;br /&gt;
* For any events, &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F\subset S:\Pr (E\cup F)=\Pr (E)+\Pr (F)-\Pr (E\cap F).&amp;lt;/math&amp;gt; So, in general, &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) \leq \Pr (E)+\Pr (F).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This generalises to three events, &amp;lt;math&amp;gt;E_{1},E_{2}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;E_{3}&amp;lt;/math&amp;gt; as&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E_{1}\cup E_{2}\cup E_{3}) &amp;amp;=&amp;amp;\Pr (E_{1})+\Pr (E_{2})+\Pr (E_{3}) \\&lt;br /&gt;
&amp;amp;&amp;amp;-\Pr (E_{1}\cap E_{2})-\Pr (E_{1}\cap E_{3})-\Pr (E_{2}\cap E_{3}) \\&lt;br /&gt;
&amp;amp;&amp;amp;+\Pr (E_{1}\cap E_{2}\cap E_{3}).\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We can demonstrate this as follows.&lt;br /&gt;
&lt;br /&gt;
Note that&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E\cup F=\left( E\cap \bar{F}\right) \cup \left( E\cap F\right) \cup \left(\bar{E}\cap F\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
the union of &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; mutually exclusive events. These mutually exclusive events are depicted by the shaded areas &amp;lt;math&amp;gt;\mathbf{a,}&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\mathbf{b}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{c}&amp;lt;/math&amp;gt;, respectively, in the next Figure.&lt;br /&gt;
&lt;br /&gt;
[[File:Prob_add.jpg|frameless|500px]]&lt;br /&gt;
&lt;br /&gt;
Then by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, and from the fact that the three events &amp;lt;math&amp;gt;\left( E\cap\bar{F}\right) &amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\left( E\cap F\right) &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\left( \bar{E}\cap F\right)&amp;lt;/math&amp;gt; are mutually exclusive so that the “area” occupied by &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt; is simply &amp;lt;math&amp;gt;\mathbf{a+b+c,}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\Pr \left( E\cap \bar{F}\right) +\Pr \left( \bar{E}\cap F\right) +\Pr \left( E\cap F\right) .&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
But also by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, since &amp;lt;math&amp;gt;E=\left( E\cap \bar{F}\right) \cup \left(E\cap F\right) &amp;lt;/math&amp;gt;, it must be that &amp;lt;math&amp;gt;\Pr (E)=\Pr \left( E\cap \bar{F}\right)+\Pr (E\cap F);&amp;lt;/math&amp;gt; similarly, &amp;lt;math&amp;gt;\Pr \left( \bar{E}\cap F\right) =\Pr \left(F\right) -\Pr \left( E\cap F\right)&amp;lt;/math&amp;gt;. Putting all of this together gives&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\Pr (E\cup F)=\Pr (E)+\Pr (F)-\Pr (E\cap F).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
When &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are mutually exclusive, so that &amp;lt;math&amp;gt;E\cap F=\emptyset&amp;lt;/math&amp;gt;, this rule reveals Axiom 2: &amp;lt;math&amp;gt;\Pr (E\cup F)=\Pr (E)+\Pr (F)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example: &amp;#039;&amp;#039;What is the probability of drawing a Queen (&amp;lt;math&amp;gt;Q &amp;lt;/math&amp;gt;) or a Club (&amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;) in a single draw from a pack of cards? Now, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;52 &amp;lt;/math&amp;gt; cards are Queens, so &amp;lt;math&amp;gt;\Pr \left( Q\right) =\frac{4}{52},&amp;lt;/math&amp;gt; whilst &amp;lt;math&amp;gt;\Pr\left( C\right) =\frac{13}{52}.&amp;lt;/math&amp;gt; The probability of drawing the Queen of Clubs is simply &amp;lt;math&amp;gt;\frac{1}{52};&amp;lt;/math&amp;gt; i.e., &amp;lt;math&amp;gt;\Pr \left( Q\cap C\right) =\frac{1}{52}&amp;lt;/math&amp;gt;. What we require is a Club or a Queen, for which the probability is&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr \left( Q\cup C\right) &amp;amp;=&amp;amp;\Pr \left( Q\right) +\Pr \left( C\right) -\Pr\left( Q\cap C\right) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{4}{52}+\frac{13}{52}-\frac{1}{52} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{16}{52}=\frac{4}{13}.\end{aligned}&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example: &amp;#039;&amp;#039;Consider a car journey from Manchester to London via the M6 and M1. Let &amp;lt;math&amp;gt;E=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;heavy traffic somewhere on route&amp;#039;&amp;#039; and &amp;lt;math&amp;gt;F=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;roadworks somewhere on route&amp;#039;&amp;#039;. It is estimated that &amp;lt;math&amp;gt;\Pr (E)=0.8&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr (F)=0.4,&amp;lt;/math&amp;gt; whilst the probability of NOT encountering both is &amp;lt;math&amp;gt;\Pr (\overline{E\cap F})=0.6.&amp;lt;/math&amp;gt; What is the probability of encountering heavytraffic or roadworks?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;We require &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) .&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E\cup F) &amp;amp;=&amp;amp;\Pr (E)+\Pr (F)-\Pr (E\cap F) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\Pr (E)+\Pr (F)-(1-\Pr (\overline{E\cap F})) \\&lt;br /&gt;
&amp;amp;=&amp;amp;0.8+0.4-1+0.6 \\&lt;br /&gt;
&amp;amp;=&amp;amp;0.8=\Pr (E)\end{aligned}&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Notice that this implies, in this case, that &amp;lt;math&amp;gt;F\subset E&amp;lt;/math&amp;gt; (why?). This &amp;#039;&amp;#039;model &amp;#039;&amp;#039;then implies that when there are roadworks somewhere on route you are bound to encounter heavy traffic; on the other hand, you can encounter heavy traffic on route without ever passing through roadworks. (My own experience of this motorway inclines me towards this implication!)&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Similar concepts apply when manipulating proportions as follows:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example&amp;#039;&amp;#039;: A sample of 1000 undergraduates were asked whether they took either Mathematics, Physics or Chemistry at A-level. The following responses were obtained: 100 just took Mathematics; 70 just took Physics; 100 just took Chemistry; 150 took Mathematics and Physics, but not Chemistry; 40 took Mathematics and Chemistry, but not Physics; and, 240 took Physics and Chemistry, but not Mathematics. What proportion took all three?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;This can be addressed with the following diagram:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;[[File:Prob_Alevels.jpg|frameless|500px]]&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;The shaded area contains the number who took all three, which can be deduced from the above information (since the total of the numbers assigned to each part of the Venn diagram must be &amp;lt;math&amp;gt;1000&amp;lt;/math&amp;gt;). The answer is therefore &amp;lt;math&amp;gt;30\%&amp;lt;/math&amp;gt; (being &amp;lt;math&amp;gt;300&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;1000&amp;lt;/math&amp;gt;).&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Two further results on unions, intersections and complements which are of use (and which are fairly easy to demonstrate using Venn diagrams) are &amp;#039;&amp;#039;&amp;#039;de Morgan Laws&amp;#039;&amp;#039;&amp;#039;:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\left( \bar{A}\cap \bar{B}\right) =\left( \overline{A\cup B}\right) &amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\bar{A}\cup \bar{B}=\left( \overline{A\cap B}\right) &amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2862</id>
		<title>Probability Intro</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2862"/>
				<updated>2013-08-07T21:27:45Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: /* Probability */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Introducing Probability =&lt;br /&gt;
&lt;br /&gt;
So far we have been looking at ways of summarising samples of data drawn from an underlying population of interest. Although at times tedious, all such arithmetic calculations are fairly mechanical and straightforward to apply. To remind ourselves, one of the primary reasons for wishing to summarise data is so assist in the development of inferences about the population from which the data were taken. That is to say, we would like to elicit some information about the mechanism which generated the observed data.&lt;br /&gt;
&lt;br /&gt;
We now start on the process of developing mathematical ways of formulating inferences and this requires the use of &amp;#039;&amp;#039;probability&amp;#039;&amp;#039;. This becomes clear if we think back to one of the early questions posed in this course: &amp;#039;&amp;#039;prior to sampling is it possible to predict with absolute certainty what will be observed&amp;#039;&amp;#039;? The answer to this question is &amp;#039;&amp;#039;no&amp;#039;&amp;#039;; although it would be of interest to know how &amp;#039;&amp;#039;likely&amp;#039;&amp;#039; it is that certain values would be observed. Or, what is the &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; of observing certain values?&lt;br /&gt;
&lt;br /&gt;
Before proceeding, we need some more tools:&lt;br /&gt;
&lt;br /&gt;
= Venn diagrams =&lt;br /&gt;
&lt;br /&gt;
Venn diagrams (and diagrams in general) are of enormous help in trying to understand, and manipulate probability. We begin with some basic definitions, some of which we have encountered before.&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Experiment:&amp;#039;&amp;#039;&amp;#039; any process which, when applied, provides data or an outcome; e.g., rolling a die and observing the number of dots on the upturned face; recording the amount of rainfall in Manchester over a period of time.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Sample Space:&amp;#039;&amp;#039;&amp;#039; set of possible outcomes of an experiment; e.g., &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; (or &amp;lt;math&amp;gt;\Omega &amp;lt;/math&amp;gt;) &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{1,2,3,4,5,6\}&amp;lt;/math&amp;gt;, which is the sample space of rolling a dice. Or &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{x;x\geq 0\}&amp;lt;/math&amp;gt;, which is the sample space of an experiment where the outcomes can be any real non-negative number, or ‘&amp;#039;&amp;#039;the set of real non-negative real numbers&amp;#039;&amp;#039;’.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Simple Event&amp;#039;&amp;#039;&amp;#039;: just one of the possible outcomes on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Event:&amp;#039;&amp;#039;&amp;#039; a &amp;#039;&amp;#039;subset&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, denoted &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;; e.g., &amp;lt;math&amp;gt;E=\left\{ 2,4,6\right\}&amp;lt;/math&amp;gt; (i.e. any even number on a dice) or &amp;lt;math&amp;gt;E=\left\{ x;4&amp;lt;x\leq 10\right\} ,&amp;lt;/math&amp;gt; which means ‘&amp;#039;&amp;#039;the set of real numbers which are strictly bigger than&amp;#039;&amp;#039; &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;but less than or equal to &amp;#039;&amp;#039;&amp;lt;math&amp;gt;10&amp;lt;/math&amp;gt;’.&amp;lt;br /&amp;gt;Note that an event, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, is a collection of simple events.&lt;br /&gt;
&lt;br /&gt;
Such concepts can be represented by means of the following Venn Diagram:&lt;br /&gt;
&lt;br /&gt;
[[File:Venn_1.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
The sample space, &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; is depicted as a closed rectangle, and the event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; is a closed loop wholly contained within &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; and we write (in set notation) &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
In dealing with probability, and in particular the probability of an event (or events) occurring, we shall need to be familiar with &amp;#039;&amp;#039;&amp;#039;UNIONS, INTERSECTIONS&amp;#039;&amp;#039;&amp;#039; and &amp;#039;&amp;#039;&amp;#039;COMPLEMENTS&amp;#039;&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
To illustrate these concepts, consider the sample space &amp;lt;math&amp;gt;S=\{x;x\geq 0\},\,&amp;lt;/math&amp;gt; with the following events defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as depicted in Figure 3.2:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E=\{x;4&amp;lt;x\leq 10\},\,F=\{x;7&amp;lt;x\leq 17\},\,G=\{x;x&amp;gt;15\},\,H=\{x;9&amp;lt;x\leq&lt;br /&gt;
13\}. &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| (a) Event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: A closed loop&lt;br /&gt;
| (b) Union: &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2a.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2b.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (c) Intersection: &amp;lt;math&amp;gt;E\cap F&amp;lt;/math&amp;gt;&lt;br /&gt;
| (d) The Null set/event: &amp;lt;math&amp;gt;E\cap G=\emptyset &amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2c.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2d.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (e) Complement of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;&lt;br /&gt;
| (f) Subset of &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;H\subset F&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;H\cap F=H&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2e.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2f.jpg|frameless|300px]]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
        &lt;br /&gt;
&lt;br /&gt;
* The &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cup F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cup F=\{x;4&amp;lt;x\leq 17\};&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are either in &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or in &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) in both. This is illustrated on the Venn diagram by the dark shaded area in diagram (b).&lt;br /&gt;
* The&amp;#039;&amp;#039; intersection&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cap F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cap F=\left\{ x;7\leq x\leq 10\right\} ;&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are common to both &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; Again this is depicted by the dark shaded area in (c). If events have no elements in common (as, for example, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt;) then they are said to be &amp;#039;&amp;#039;mutually exclusive&amp;#039;&amp;#039;, and we can write &amp;lt;math&amp;gt;E\cap G=\emptyset ,&amp;lt;/math&amp;gt; meaning the &amp;#039;&amp;#039;null set&amp;#039;&amp;#039; which contains no elements. Such a situation is illustrated on the Venn Diagram by events (the two shaded closed loops in (d)) which do not overlap. Notice however that &amp;lt;math&amp;gt;G\cap F\neq \emptyset ,&amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have elements in common.&lt;br /&gt;
* The &amp;#039;&amp;#039;complement&amp;#039;&amp;#039; of an event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; say, is everything defined on the sample space which is not in &amp;lt;math&amp;gt;E.&amp;lt;/math&amp;gt; This event is denoted &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;, the dark shaded area in (e); here &amp;lt;math&amp;gt;\bar{E}=\left\{ x;x\leq 4\right\} \cup \left\{ x;x&amp;gt;10\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
* Finally note that &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; is a sub-set of &amp;lt;math&amp;gt;F;&amp;lt;/math&amp;gt; see (f). It is depicted as the dark closed loop wholly contained within &amp;lt;math&amp;gt;F,&amp;lt;/math&amp;gt; the lighter shaded area, so that &amp;lt;math&amp;gt;H\cap F=H;&amp;lt;/math&amp;gt; if an element in the sample space is a member of &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; then it must also be member of &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; (In mathematical logic, we employ this scenario to indicate that “&amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; implies &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;”, but not necessarily vice-versa.) Notice that &amp;lt;math&amp;gt;G\cap H=\emptyset &amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;H\cap E\neq \emptyset&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
= Probability =&lt;br /&gt;
&lt;br /&gt;
The term &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; (or some equivalent) is used in everyday conversation and so can not be unfamiliar to the reader. We talk of the probability, or chance, of rain; the likelihood of England winning the World Cup; or, perhaps more scientifically, the chance of getting a &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; when rolling a die. What we shall now do is develop a coherent theory of probability; a theory which allows us to combine and manipulate probabilities in a consistent and meaningful manner. We shall describe ways of dealing with, and describing, uncertainty. This will involve &amp;#039;&amp;#039;rules&amp;#039;&amp;#039; which govern our use of terms like probability.&lt;br /&gt;
&lt;br /&gt;
There have been a number of different approaches (interpretations) of probability. Most depend, at least to some extent, on the notion of relative frequency as now described:&lt;br /&gt;
&lt;br /&gt;
* Suppose an experiment has an outcome of interest &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;. The &amp;#039;&amp;#039;relative frequency interpretation&amp;#039;&amp;#039; of probability says that assuming the experiment can be repeated a large number of times then the relative frequency of observing the outcome &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; will settle down to a &amp;#039;&amp;#039;number&amp;#039;&amp;#039;, denoted &amp;lt;math&amp;gt;\Pr (E),&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;P(E)&amp;lt;/math&amp;gt; or Prob&amp;lt;math&amp;gt;(E),&amp;lt;/math&amp;gt; called the &amp;#039;&amp;#039;&amp;#039;probability&amp;#039;&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
This is illustrated in the next Figure where the proportion of heads obtained after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; flips of a fair coin is plotted against &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;, as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; increases; e.g., of the first 100 flips, 46 were heads (&amp;lt;math&amp;gt;46\%&amp;lt;/math&amp;gt;). Notice that the plot becomes less ‘wobbly’ after about &amp;lt;math&amp;gt;n=&amp;lt;/math&amp;gt;220 and appears to be settling down to the value of &amp;lt;math&amp;gt;\frac{1}{2}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
[[File:Prob_coin.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
Due to this interpretation of probability, we often use observed sample proportions to approximate underlying probabilities of interest; see, for example, Question 4 of Exercise 2. There are, of course, other interpretations of probability; e.g., the subjective interpretation which simply expresses the strength of one’s belief about an event of interest such as whether Manchester United will win the European Cup! Any one of these interpretations can be used in practical situations provided the implied notion of probability follows a simple set of &amp;#039;&amp;#039;axioms&amp;#039;&amp;#039; or &amp;#039;&amp;#039;rules&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
== The axioms of probability ==&lt;br /&gt;
&lt;br /&gt;
There are just &amp;#039;&amp;#039;three &amp;#039;&amp;#039;basic rules that must be obeyed when dealing with probabilities:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;For any event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; i.e., &amp;lt;math&amp;gt;E\subset S,\,\,\Pr (E)\geq 0&amp;lt;/math&amp;gt;; &amp;#039;&amp;#039;probabilities are non-negative&amp;#039;&amp;#039;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (S)=1;&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;having defined the sample space of outcomes, one of these outcomes must be observed&amp;#039;&amp;#039;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;If events &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are mutually exclusive defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, so that &amp;lt;math&amp;gt;E\cap F=\emptyset &amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\Pr \left( E\right)+\Pr \left( F\right) .&amp;lt;/math&amp;gt; In general, for any set of mutually exclusive events, &amp;lt;math&amp;gt;E_{1},E_{2},\ldots ,E_{k},&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S:&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (E_{1}\cup E_{2}\cup \ldots \cup E_{k})=\Pr (E_{1})+\Pr (E_{2})+\ldots\Pr (E_{k})&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;i.e., &amp;lt;math&amp;gt;\Pr \left( \bigcup_{j=1}^{k}E_{j}\right) =\sum_{j=1}^{k}\Pr (E_{j}).&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In terms of the Venn Diagram, one can (and should) usefully think of the area of &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; relative to that of &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as providing an indication of probability. (Note, from axiom 2, that the area of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is implicitly normalised to be unity).&lt;br /&gt;
&lt;br /&gt;
Also observe that, contrary to what you may have believed, it is not one of the rules that &amp;lt;math&amp;gt;\Pr (E)\leq 1&amp;lt;/math&amp;gt; for any event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;. Rather, this is an implication of the &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; rules given:&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;implications: &amp;#039;&amp;#039;&amp;#039;it must be that for any event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;E\cap \bar{E}=\emptyset &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;E\cup \bar{E}=S.&amp;lt;/math&amp;gt; By Axiom &amp;lt;math&amp;gt;1,&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr (E)\geq 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr \left( \bar{E}\right) \geq 0&amp;lt;/math&amp;gt; and by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr(E)+\Pr (\bar{E})=\Pr (S).&amp;lt;/math&amp;gt; So &amp;lt;math&amp;gt;\Pr \left( E\right) +\Pr \left( \bar{E}\right) =1,&amp;lt;/math&amp;gt; by Axiom &amp;lt;math&amp;gt;2.&amp;lt;/math&amp;gt; This implies that&lt;br /&gt;
&lt;br /&gt;
# &amp;lt;math&amp;gt;0\leq \Pr (E)\leq 1&amp;lt;/math&amp;gt;&lt;br /&gt;
# &amp;lt;math&amp;gt;\Pr (\bar{E})=1-\Pr (E)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The first of these is what we might have expected from probability (a number lying between &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;). The second implication is also very important; it says that the probability of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; not happening is ‘&amp;#039;&amp;#039;one minus the probability of it happening&amp;#039;&amp;#039;’. Thus when rolling a die, the probability of getting &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; is one minus the probability of getting either a &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;2&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;5.&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
These axioms imply how to calculate probabilities on a sample space of equally likely outcomes. For example, and as we have already noted, the experiment of rolling a fair die defines a sample space of six, mutually exclusive and equally likely outcomes (&amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; dots on the up-turned face). The axioms then say that each of the six probabilities are positive, add to 1 and are all the same. Thus, the probability of any one of the outcomes must be simply &amp;lt;math&amp;gt;\frac{1}{6};&amp;lt;/math&amp;gt; which may accord with your intuition. A similar sort of analysis reveals that the probability of drawing a club from a deck of &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; cards is &amp;lt;math&amp;gt;\frac{13}{52},&amp;lt;/math&amp;gt; since any one of the &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; cards has an equal chance of being drawn and &amp;lt;math&amp;gt;13&amp;lt;/math&amp;gt; of them are clubs; i.e., &amp;lt;math&amp;gt;13&amp;lt;/math&amp;gt; of the &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; are clubs, so the probability of drawing a club is &amp;lt;math&amp;gt;\frac{13}{52}.&amp;lt;/math&amp;gt; Notice the importance of the assumption of equally likely outcomes here.&lt;br /&gt;
&lt;br /&gt;
In this, and the next section of notes, we shall see how these axioms can be used. Firstly, consider the construction of a probability for the &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of two events; i.e., the probability that &amp;#039;&amp;#039;either &amp;#039;&amp;#039;&amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) &amp;#039;&amp;#039;both &amp;#039;&amp;#039;will occur. Such a probability is embodied in the &amp;#039;&amp;#039;addition rule of probability&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
== The addition rule of probability ==&lt;br /&gt;
&lt;br /&gt;
When rolling a fair die, let &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; denote the event of an “odd number of dots” and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; the event of the “number of dots being greater than, or equal, to &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;”&amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt; What is the probability of the event &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;? To calculate this we can collect together all the mutually exclusive (simple) events which comprise &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;, and then add up the probabilities (by axiom 3). These simple events are &amp;lt;math&amp;gt;1,3,4,5&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; dots. Each has a probability of &amp;lt;math&amp;gt;\frac{1}{6},&amp;lt;/math&amp;gt; so the required total probability is: &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\frac{5}{6}&amp;lt;/math&amp;gt;. Consider carefully how this probability is constructed and note, in particular, that &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) \neq \Pr \left( E\right) +\Pr \left( F\right) &amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have a simple event in common (namely &amp;lt;math&amp;gt;5&amp;lt;/math&amp;gt; dots).&lt;br /&gt;
&lt;br /&gt;
In general, we can calculate the probability of the union of events using the &amp;#039;&amp;#039;addition rule of probability&amp;#039;&amp;#039;, as follows.&lt;br /&gt;
&lt;br /&gt;
* For any events, &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F\subset S:\Pr (E\cup F)=\Pr (E)+\Pr (F)-\Pr (E\cap F).&amp;lt;/math&amp;gt; So, in general, &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) \leq \Pr (E)+\Pr (F).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This generalises to three events, &amp;lt;math&amp;gt;E_{1},E_{2}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;E_{3}&amp;lt;/math&amp;gt; as&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E_{1}\cup E_{2}\cup E_{3}) &amp;amp;=&amp;amp;\Pr (E_{1})+\Pr (E_{2})+\Pr (E_{3}) \\&lt;br /&gt;
&amp;amp;&amp;amp;-\Pr (E_{1}\cap E_{2})-\Pr (E_{1}\cap E_{3})-\Pr (E_{2}\cap E_{3}) \\&lt;br /&gt;
&amp;amp;&amp;amp;+\Pr (E_{1}\cap E_{2}\cap E_{3}).\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We can demonstrate this as follows.&lt;br /&gt;
&lt;br /&gt;
Note that&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E\cup F=\left( E\cap \bar{F}\right) \cup \left( E\cap F\right) \cup \left(\bar{E}\cap F\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
the union of &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; mutually exclusive events. These mutually exclusive events are depicted by the shaded areas &amp;lt;math&amp;gt;\mathbf{a,}&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\mathbf{b}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{c}&amp;lt;/math&amp;gt;, respectively, in the next Figure.&lt;br /&gt;
&lt;br /&gt;
[[File:Prob_add.jpg|frameless|500px]]&lt;br /&gt;
&lt;br /&gt;
Then by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, and from the fact that the three events &amp;lt;math&amp;gt;\left( E\cap\bar{F}\right) &amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\left( E\cap F\right) &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\left( \bar{E}\cap F\right)&amp;lt;/math&amp;gt; are mutually exclusive so that the “area” occupied by &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt; is simply &amp;lt;math&amp;gt;\mathbf{a+b+c,}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\Pr \left( E\cap \bar{F}\right) +\Pr \left( \bar{E}\cap F\right) +\Pr \left( E\cap F\right) .&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
But also by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, since &amp;lt;math&amp;gt;E=\left( E\cap \bar{F}\right) \cup \left(E\cap F\right) &amp;lt;/math&amp;gt;, it must be that &amp;lt;math&amp;gt;\Pr (E)=\Pr \left( E\cap \bar{F}\right)+\Pr (E\cap F);&amp;lt;/math&amp;gt; similarly, &amp;lt;math&amp;gt;\Pr \left( \bar{E}\cap F\right) =\Pr \left(F\right) -\Pr \left( E\cap F\right)&amp;lt;/math&amp;gt;. Putting all of this together gives&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\Pr (E\cup F)=\Pr (E)+\Pr (F)-\Pr (E\cap F).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
When &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are mutually exclusive, so that &amp;lt;math&amp;gt;E\cap F=\emptyset&amp;lt;/math&amp;gt;, this rule reveals Axiom 2: &amp;lt;math&amp;gt;\Pr (E\cup F)=\Pr (E)+\Pr (F)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example: &amp;#039;&amp;#039;What is the probability of drawing a Queen (&amp;lt;math&amp;gt;Q &amp;lt;/math&amp;gt;) or a Club (&amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;) in a single draw from a pack of cards? Now, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;52 &amp;lt;/math&amp;gt; cards are Queens, so &amp;lt;math&amp;gt;\Pr \left( Q\right) =\frac{4}{52},&amp;lt;/math&amp;gt; whilst &amp;lt;math&amp;gt;\Pr\left( C\right) =\frac{13}{52}.&amp;lt;/math&amp;gt; The probability of drawing the Queen of Clubs is simply &amp;lt;math&amp;gt;\frac{1}{52};&amp;lt;/math&amp;gt; i.e., &amp;lt;math&amp;gt;\Pr \left( Q\cap C\right) =\frac{1}{52}&amp;lt;/math&amp;gt;. What we require is a Club or a Queen, for which the probability is&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr \left( Q\cup C\right) &amp;amp;=&amp;amp;\Pr \left( Q\right) +\Pr \left( C\right) -\Pr\left( Q\cap C\right) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{4}{52}+\frac{13}{52}-\frac{1}{52} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{16}{52}=\frac{4}{13}.\end{aligned}&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example: &amp;#039;&amp;#039;Consider a car journey from Manchester to London via the M6 and M1. Let &amp;lt;math&amp;gt;E=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;heavy traffic somewhere on route&amp;#039;&amp;#039; and &amp;lt;math&amp;gt;F=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;roadworks somewhere on route&amp;#039;&amp;#039;. It is estimated that &amp;lt;math&amp;gt;\Pr (E)=0.8&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr (F)=0.4,&amp;lt;/math&amp;gt; whilst the probability of NOT encountering both is &amp;lt;math&amp;gt;\Pr (\overline{E\cap F})=0.6.&amp;lt;/math&amp;gt; What is the probability of encountering heavytraffic or roadworks?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;We require &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) .&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E\cup F) &amp;amp;=&amp;amp;\Pr (E)+\Pr (F)-\Pr (E\cap F) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\Pr (E)+\Pr (F)-(1-\Pr (\overline{E\cap F})) \\&lt;br /&gt;
&amp;amp;=&amp;amp;0.8+0.4-1+0.6 \\&lt;br /&gt;
&amp;amp;=&amp;amp;0.8=\Pr (E)\end{aligned}&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Notice that this implies, in this case, that &amp;lt;math&amp;gt;F\subset E&amp;lt;/math&amp;gt; (why?). This &amp;#039;&amp;#039;model &amp;#039;&amp;#039;then implies that when there are roadworks somewhere on route you are bound to encounter heavy traffic; on the other hand, you can encounter heavy traffic on route without ever passing through roadworks. (My own experience of this motorway inclines me towards this implication!)&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Similar concepts apply when manipulating proportions as follows:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example&amp;#039;&amp;#039;: A sample of 1000 undergraduates were asked whether they took either Mathematics, Physics or Chemistry at A-level. The following responses were obtained: 100 just took Mathematics; 70 just took Physics; 100 just took Chemistry; 150 took Mathematics and Physics, but not Chemistry; 40 took Mathematics and Chemistry, but not Physics; and, 240 took Physics and Chemistry, but not Mathematics. What proportion took all three?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;This can be addressed with the following diagram:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;[[File:Prob_Alevels.jpg|frameless|500px]]&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;The shaded area contains the number who took all three, which can be deduced from the above information (since the total of the numbers assigned to each part of the Venn diagram must be &amp;lt;math&amp;gt;1000&amp;lt;/math&amp;gt;). The answer is therefore &amp;lt;math&amp;gt;30\%&amp;lt;/math&amp;gt; (being &amp;lt;math&amp;gt;300&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;1000&amp;lt;/math&amp;gt;).&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Two further results on unions, intersections and complements which are of use (and which are fairly easy to demonstrate using Venn diagrams) are &amp;#039;&amp;#039;&amp;#039;de Morgan Laws&amp;#039;&amp;#039;&amp;#039;:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\left( \bar{A}\cap \bar{B}\right) =\left( \overline{A\cup B}\right) &amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\bar{A}\cup \bar{B}=\left( \overline{A\cap B}\right) &amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2861</id>
		<title>Probability Intro</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2861"/>
				<updated>2013-08-07T21:27:32Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: /* Probability */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Introducing Probability =&lt;br /&gt;
&lt;br /&gt;
So far we have been looking at ways of summarising samples of data drawn from an underlying population of interest. Although at times tedious, all such arithmetic calculations are fairly mechanical and straightforward to apply. To remind ourselves, one of the primary reasons for wishing to summarise data is so assist in the development of inferences about the population from which the data were taken. That is to say, we would like to elicit some information about the mechanism which generated the observed data.&lt;br /&gt;
&lt;br /&gt;
We now start on the process of developing mathematical ways of formulating inferences and this requires the use of &amp;#039;&amp;#039;probability&amp;#039;&amp;#039;. This becomes clear if we think back to one of the early questions posed in this course: &amp;#039;&amp;#039;prior to sampling is it possible to predict with absolute certainty what will be observed&amp;#039;&amp;#039;? The answer to this question is &amp;#039;&amp;#039;no&amp;#039;&amp;#039;; although it would be of interest to know how &amp;#039;&amp;#039;likely&amp;#039;&amp;#039; it is that certain values would be observed. Or, what is the &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; of observing certain values?&lt;br /&gt;
&lt;br /&gt;
Before proceeding, we need some more tools:&lt;br /&gt;
&lt;br /&gt;
= Venn diagrams =&lt;br /&gt;
&lt;br /&gt;
Venn diagrams (and diagrams in general) are of enormous help in trying to understand, and manipulate probability. We begin with some basic definitions, some of which we have encountered before.&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Experiment:&amp;#039;&amp;#039;&amp;#039; any process which, when applied, provides data or an outcome; e.g., rolling a die and observing the number of dots on the upturned face; recording the amount of rainfall in Manchester over a period of time.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Sample Space:&amp;#039;&amp;#039;&amp;#039; set of possible outcomes of an experiment; e.g., &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; (or &amp;lt;math&amp;gt;\Omega &amp;lt;/math&amp;gt;) &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{1,2,3,4,5,6\}&amp;lt;/math&amp;gt;, which is the sample space of rolling a dice. Or &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{x;x\geq 0\}&amp;lt;/math&amp;gt;, which is the sample space of an experiment where the outcomes can be any real non-negative number, or ‘&amp;#039;&amp;#039;the set of real non-negative real numbers&amp;#039;&amp;#039;’.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Simple Event&amp;#039;&amp;#039;&amp;#039;: just one of the possible outcomes on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Event:&amp;#039;&amp;#039;&amp;#039; a &amp;#039;&amp;#039;subset&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, denoted &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;; e.g., &amp;lt;math&amp;gt;E=\left\{ 2,4,6\right\}&amp;lt;/math&amp;gt; (i.e. any even number on a dice) or &amp;lt;math&amp;gt;E=\left\{ x;4&amp;lt;x\leq 10\right\} ,&amp;lt;/math&amp;gt; which means ‘&amp;#039;&amp;#039;the set of real numbers which are strictly bigger than&amp;#039;&amp;#039; &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;but less than or equal to &amp;#039;&amp;#039;&amp;lt;math&amp;gt;10&amp;lt;/math&amp;gt;’.&amp;lt;br /&amp;gt;Note that an event, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, is a collection of simple events.&lt;br /&gt;
&lt;br /&gt;
Such concepts can be represented by means of the following Venn Diagram:&lt;br /&gt;
&lt;br /&gt;
[[File:Venn_1.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
The sample space, &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; is depicted as a closed rectangle, and the event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; is a closed loop wholly contained within &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; and we write (in set notation) &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
In dealing with probability, and in particular the probability of an event (or events) occurring, we shall need to be familiar with &amp;#039;&amp;#039;&amp;#039;UNIONS, INTERSECTIONS&amp;#039;&amp;#039;&amp;#039; and &amp;#039;&amp;#039;&amp;#039;COMPLEMENTS&amp;#039;&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
To illustrate these concepts, consider the sample space &amp;lt;math&amp;gt;S=\{x;x\geq 0\},\,&amp;lt;/math&amp;gt; with the following events defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as depicted in Figure 3.2:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E=\{x;4&amp;lt;x\leq 10\},\,F=\{x;7&amp;lt;x\leq 17\},\,G=\{x;x&amp;gt;15\},\,H=\{x;9&amp;lt;x\leq&lt;br /&gt;
13\}. &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| (a) Event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: A closed loop&lt;br /&gt;
| (b) Union: &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2a.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2b.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (c) Intersection: &amp;lt;math&amp;gt;E\cap F&amp;lt;/math&amp;gt;&lt;br /&gt;
| (d) The Null set/event: &amp;lt;math&amp;gt;E\cap G=\emptyset &amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2c.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2d.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (e) Complement of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;&lt;br /&gt;
| (f) Subset of &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;H\subset F&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;H\cap F=H&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2e.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2f.jpg|frameless|300px]]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
        &lt;br /&gt;
&lt;br /&gt;
* The &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cup F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cup F=\{x;4&amp;lt;x\leq 17\};&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are either in &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or in &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) in both. This is illustrated on the Venn diagram by the dark shaded area in diagram (b).&lt;br /&gt;
* The&amp;#039;&amp;#039; intersection&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cap F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cap F=\left\{ x;7\leq x\leq 10\right\} ;&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are common to both &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; Again this is depicted by the dark shaded area in (c). If events have no elements in common (as, for example, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt;) then they are said to be &amp;#039;&amp;#039;mutually exclusive&amp;#039;&amp;#039;, and we can write &amp;lt;math&amp;gt;E\cap G=\emptyset ,&amp;lt;/math&amp;gt; meaning the &amp;#039;&amp;#039;null set&amp;#039;&amp;#039; which contains no elements. Such a situation is illustrated on the Venn Diagram by events (the two shaded closed loops in (d)) which do not overlap. Notice however that &amp;lt;math&amp;gt;G\cap F\neq \emptyset ,&amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have elements in common.&lt;br /&gt;
* The &amp;#039;&amp;#039;complement&amp;#039;&amp;#039; of an event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; say, is everything defined on the sample space which is not in &amp;lt;math&amp;gt;E.&amp;lt;/math&amp;gt; This event is denoted &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;, the dark shaded area in (e); here &amp;lt;math&amp;gt;\bar{E}=\left\{ x;x\leq 4\right\} \cup \left\{ x;x&amp;gt;10\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
* Finally note that &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; is a sub-set of &amp;lt;math&amp;gt;F;&amp;lt;/math&amp;gt; see (f). It is depicted as the dark closed loop wholly contained within &amp;lt;math&amp;gt;F,&amp;lt;/math&amp;gt; the lighter shaded area, so that &amp;lt;math&amp;gt;H\cap F=H;&amp;lt;/math&amp;gt; if an element in the sample space is a member of &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; then it must also be member of &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; (In mathematical logic, we employ this scenario to indicate that “&amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; implies &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;”, but not necessarily vice-versa.) Notice that &amp;lt;math&amp;gt;G\cap H=\emptyset &amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;H\cap E\neq \emptyset&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
= Probability =&lt;br /&gt;
&lt;br /&gt;
The term &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; (or some equivalent) is used in everyday conversation and so can not be unfamiliar to the reader. We talk of the probability, or chance, of rain; the likelihood of England winning the World Cup; or, perhaps more scientifically, the chance of getting a &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; when rolling a die. What we shall now do is develop a coherent theory of probability; a theory which allows us to combine and manipulate probabilities in a consistent and meaningful manner. We shall describe ways of dealing with, and describing, uncertainty. This will involve &amp;#039;&amp;#039;rules&amp;#039;&amp;#039; which govern our use of terms like probability.&lt;br /&gt;
&lt;br /&gt;
There have been a number of different approaches (interpretations) of probability. Most depend, at least to some extent, on the notion of relative frequency as now described:&lt;br /&gt;
&lt;br /&gt;
* Suppose an experiment has an outcome of interest &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;. The &amp;#039;&amp;#039;relative frequency interpretation&amp;#039;&amp;#039; of probability says that assuming the experiment can be repeated a large number of times then the relative frequency of observing the outcome &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; will settle down to a &amp;#039;&amp;#039;number&amp;#039;&amp;#039;, denoted &amp;lt;math&amp;gt;\Pr (E),&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;P(E)&amp;lt;/math&amp;gt; or Prob&amp;lt;math&amp;gt;(E),&amp;lt;/math&amp;gt; called the &amp;#039;&amp;#039;&amp;#039;probability&amp;#039;&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
This is illustrated in the next Figure where the proportion of heads obtained after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; flips of a fair coin is plotted against &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;, as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; increases; e.g., of the first 100 flips, 46 were heads (&amp;lt;math&amp;gt;55\%&amp;lt;/math&amp;gt;). Notice that the plot becomes less ‘wobbly’ after about &amp;lt;math&amp;gt;n=&amp;lt;/math&amp;gt;220 and appears to be settling down to the value of &amp;lt;math&amp;gt;\frac{1}{2}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
[[File:Prob_coin.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
Due to this interpretation of probability, we often use observed sample proportions to approximate underlying probabilities of interest; see, for example, Question 4 of Exercise 2. There are, of course, other interpretations of probability; e.g., the subjective interpretation which simply expresses the strength of one’s belief about an event of interest such as whether Manchester United will win the European Cup! Any one of these interpretations can be used in practical situations provided the implied notion of probability follows a simple set of &amp;#039;&amp;#039;axioms&amp;#039;&amp;#039; or &amp;#039;&amp;#039;rules&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
== The axioms of probability ==&lt;br /&gt;
&lt;br /&gt;
There are just &amp;#039;&amp;#039;three &amp;#039;&amp;#039;basic rules that must be obeyed when dealing with probabilities:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;For any event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; i.e., &amp;lt;math&amp;gt;E\subset S,\,\,\Pr (E)\geq 0&amp;lt;/math&amp;gt;; &amp;#039;&amp;#039;probabilities are non-negative&amp;#039;&amp;#039;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (S)=1;&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;having defined the sample space of outcomes, one of these outcomes must be observed&amp;#039;&amp;#039;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;If events &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are mutually exclusive defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, so that &amp;lt;math&amp;gt;E\cap F=\emptyset &amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\Pr \left( E\right)+\Pr \left( F\right) .&amp;lt;/math&amp;gt; In general, for any set of mutually exclusive events, &amp;lt;math&amp;gt;E_{1},E_{2},\ldots ,E_{k},&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S:&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (E_{1}\cup E_{2}\cup \ldots \cup E_{k})=\Pr (E_{1})+\Pr (E_{2})+\ldots\Pr (E_{k})&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;i.e., &amp;lt;math&amp;gt;\Pr \left( \bigcup_{j=1}^{k}E_{j}\right) =\sum_{j=1}^{k}\Pr (E_{j}).&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In terms of the Venn Diagram, one can (and should) usefully think of the area of &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; relative to that of &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as providing an indication of probability. (Note, from axiom 2, that the area of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is implicitly normalised to be unity).&lt;br /&gt;
&lt;br /&gt;
Also observe that, contrary to what you may have believed, it is not one of the rules that &amp;lt;math&amp;gt;\Pr (E)\leq 1&amp;lt;/math&amp;gt; for any event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;. Rather, this is an implication of the &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; rules given:&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;implications: &amp;#039;&amp;#039;&amp;#039;it must be that for any event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;E\cap \bar{E}=\emptyset &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;E\cup \bar{E}=S.&amp;lt;/math&amp;gt; By Axiom &amp;lt;math&amp;gt;1,&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr (E)\geq 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr \left( \bar{E}\right) \geq 0&amp;lt;/math&amp;gt; and by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr(E)+\Pr (\bar{E})=\Pr (S).&amp;lt;/math&amp;gt; So &amp;lt;math&amp;gt;\Pr \left( E\right) +\Pr \left( \bar{E}\right) =1,&amp;lt;/math&amp;gt; by Axiom &amp;lt;math&amp;gt;2.&amp;lt;/math&amp;gt; This implies that&lt;br /&gt;
&lt;br /&gt;
# &amp;lt;math&amp;gt;0\leq \Pr (E)\leq 1&amp;lt;/math&amp;gt;&lt;br /&gt;
# &amp;lt;math&amp;gt;\Pr (\bar{E})=1-\Pr (E)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The first of these is what we might have expected from probability (a number lying between &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;). The second implication is also very important; it says that the probability of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; not happening is ‘&amp;#039;&amp;#039;one minus the probability of it happening&amp;#039;&amp;#039;’. Thus when rolling a die, the probability of getting &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; is one minus the probability of getting either a &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;2&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;5.&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
These axioms imply how to calculate probabilities on a sample space of equally likely outcomes. For example, and as we have already noted, the experiment of rolling a fair die defines a sample space of six, mutually exclusive and equally likely outcomes (&amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; dots on the up-turned face). The axioms then say that each of the six probabilities are positive, add to 1 and are all the same. Thus, the probability of any one of the outcomes must be simply &amp;lt;math&amp;gt;\frac{1}{6};&amp;lt;/math&amp;gt; which may accord with your intuition. A similar sort of analysis reveals that the probability of drawing a club from a deck of &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; cards is &amp;lt;math&amp;gt;\frac{13}{52},&amp;lt;/math&amp;gt; since any one of the &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; cards has an equal chance of being drawn and &amp;lt;math&amp;gt;13&amp;lt;/math&amp;gt; of them are clubs; i.e., &amp;lt;math&amp;gt;13&amp;lt;/math&amp;gt; of the &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; are clubs, so the probability of drawing a club is &amp;lt;math&amp;gt;\frac{13}{52}.&amp;lt;/math&amp;gt; Notice the importance of the assumption of equally likely outcomes here.&lt;br /&gt;
&lt;br /&gt;
In this, and the next section of notes, we shall see how these axioms can be used. Firstly, consider the construction of a probability for the &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of two events; i.e., the probability that &amp;#039;&amp;#039;either &amp;#039;&amp;#039;&amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) &amp;#039;&amp;#039;both &amp;#039;&amp;#039;will occur. Such a probability is embodied in the &amp;#039;&amp;#039;addition rule of probability&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
== The addition rule of probability ==&lt;br /&gt;
&lt;br /&gt;
When rolling a fair die, let &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; denote the event of an “odd number of dots” and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; the event of the “number of dots being greater than, or equal, to &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;”&amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt; What is the probability of the event &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;? To calculate this we can collect together all the mutually exclusive (simple) events which comprise &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;, and then add up the probabilities (by axiom 3). These simple events are &amp;lt;math&amp;gt;1,3,4,5&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; dots. Each has a probability of &amp;lt;math&amp;gt;\frac{1}{6},&amp;lt;/math&amp;gt; so the required total probability is: &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\frac{5}{6}&amp;lt;/math&amp;gt;. Consider carefully how this probability is constructed and note, in particular, that &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) \neq \Pr \left( E\right) +\Pr \left( F\right) &amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have a simple event in common (namely &amp;lt;math&amp;gt;5&amp;lt;/math&amp;gt; dots).&lt;br /&gt;
&lt;br /&gt;
In general, we can calculate the probability of the union of events using the &amp;#039;&amp;#039;addition rule of probability&amp;#039;&amp;#039;, as follows.&lt;br /&gt;
&lt;br /&gt;
* For any events, &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F\subset S:\Pr (E\cup F)=\Pr (E)+\Pr (F)-\Pr (E\cap F).&amp;lt;/math&amp;gt; So, in general, &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) \leq \Pr (E)+\Pr (F).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This generalises to three events, &amp;lt;math&amp;gt;E_{1},E_{2}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;E_{3}&amp;lt;/math&amp;gt; as&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E_{1}\cup E_{2}\cup E_{3}) &amp;amp;=&amp;amp;\Pr (E_{1})+\Pr (E_{2})+\Pr (E_{3}) \\&lt;br /&gt;
&amp;amp;&amp;amp;-\Pr (E_{1}\cap E_{2})-\Pr (E_{1}\cap E_{3})-\Pr (E_{2}\cap E_{3}) \\&lt;br /&gt;
&amp;amp;&amp;amp;+\Pr (E_{1}\cap E_{2}\cap E_{3}).\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We can demonstrate this as follows.&lt;br /&gt;
&lt;br /&gt;
Note that&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E\cup F=\left( E\cap \bar{F}\right) \cup \left( E\cap F\right) \cup \left(\bar{E}\cap F\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
the union of &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; mutually exclusive events. These mutually exclusive events are depicted by the shaded areas &amp;lt;math&amp;gt;\mathbf{a,}&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\mathbf{b}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{c}&amp;lt;/math&amp;gt;, respectively, in the next Figure.&lt;br /&gt;
&lt;br /&gt;
[[File:Prob_add.jpg|frameless|500px]]&lt;br /&gt;
&lt;br /&gt;
Then by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, and from the fact that the three events &amp;lt;math&amp;gt;\left( E\cap\bar{F}\right) &amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\left( E\cap F\right) &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\left( \bar{E}\cap F\right)&amp;lt;/math&amp;gt; are mutually exclusive so that the “area” occupied by &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt; is simply &amp;lt;math&amp;gt;\mathbf{a+b+c,}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\Pr \left( E\cap \bar{F}\right) +\Pr \left( \bar{E}\cap F\right) +\Pr \left( E\cap F\right) .&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
But also by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, since &amp;lt;math&amp;gt;E=\left( E\cap \bar{F}\right) \cup \left(E\cap F\right) &amp;lt;/math&amp;gt;, it must be that &amp;lt;math&amp;gt;\Pr (E)=\Pr \left( E\cap \bar{F}\right)+\Pr (E\cap F);&amp;lt;/math&amp;gt; similarly, &amp;lt;math&amp;gt;\Pr \left( \bar{E}\cap F\right) =\Pr \left(F\right) -\Pr \left( E\cap F\right)&amp;lt;/math&amp;gt;. Putting all of this together gives&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\Pr (E\cup F)=\Pr (E)+\Pr (F)-\Pr (E\cap F).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
When &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are mutually exclusive, so that &amp;lt;math&amp;gt;E\cap F=\emptyset&amp;lt;/math&amp;gt;, this rule reveals Axiom 2: &amp;lt;math&amp;gt;\Pr (E\cup F)=\Pr (E)+\Pr (F)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example: &amp;#039;&amp;#039;What is the probability of drawing a Queen (&amp;lt;math&amp;gt;Q &amp;lt;/math&amp;gt;) or a Club (&amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;) in a single draw from a pack of cards? Now, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;52 &amp;lt;/math&amp;gt; cards are Queens, so &amp;lt;math&amp;gt;\Pr \left( Q\right) =\frac{4}{52},&amp;lt;/math&amp;gt; whilst &amp;lt;math&amp;gt;\Pr\left( C\right) =\frac{13}{52}.&amp;lt;/math&amp;gt; The probability of drawing the Queen of Clubs is simply &amp;lt;math&amp;gt;\frac{1}{52};&amp;lt;/math&amp;gt; i.e., &amp;lt;math&amp;gt;\Pr \left( Q\cap C\right) =\frac{1}{52}&amp;lt;/math&amp;gt;. What we require is a Club or a Queen, for which the probability is&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr \left( Q\cup C\right) &amp;amp;=&amp;amp;\Pr \left( Q\right) +\Pr \left( C\right) -\Pr\left( Q\cap C\right) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{4}{52}+\frac{13}{52}-\frac{1}{52} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{16}{52}=\frac{4}{13}.\end{aligned}&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example: &amp;#039;&amp;#039;Consider a car journey from Manchester to London via the M6 and M1. Let &amp;lt;math&amp;gt;E=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;heavy traffic somewhere on route&amp;#039;&amp;#039; and &amp;lt;math&amp;gt;F=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;roadworks somewhere on route&amp;#039;&amp;#039;. It is estimated that &amp;lt;math&amp;gt;\Pr (E)=0.8&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr (F)=0.4,&amp;lt;/math&amp;gt; whilst the probability of NOT encountering both is &amp;lt;math&amp;gt;\Pr (\overline{E\cap F})=0.6.&amp;lt;/math&amp;gt; What is the probability of encountering heavytraffic or roadworks?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;We require &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) .&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E\cup F) &amp;amp;=&amp;amp;\Pr (E)+\Pr (F)-\Pr (E\cap F) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\Pr (E)+\Pr (F)-(1-\Pr (\overline{E\cap F})) \\&lt;br /&gt;
&amp;amp;=&amp;amp;0.8+0.4-1+0.6 \\&lt;br /&gt;
&amp;amp;=&amp;amp;0.8=\Pr (E)\end{aligned}&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Notice that this implies, in this case, that &amp;lt;math&amp;gt;F\subset E&amp;lt;/math&amp;gt; (why?). This &amp;#039;&amp;#039;model &amp;#039;&amp;#039;then implies that when there are roadworks somewhere on route you are bound to encounter heavy traffic; on the other hand, you can encounter heavy traffic on route without ever passing through roadworks. (My own experience of this motorway inclines me towards this implication!)&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Similar concepts apply when manipulating proportions as follows:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example&amp;#039;&amp;#039;: A sample of 1000 undergraduates were asked whether they took either Mathematics, Physics or Chemistry at A-level. The following responses were obtained: 100 just took Mathematics; 70 just took Physics; 100 just took Chemistry; 150 took Mathematics and Physics, but not Chemistry; 40 took Mathematics and Chemistry, but not Physics; and, 240 took Physics and Chemistry, but not Mathematics. What proportion took all three?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;This can be addressed with the following diagram:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;[[File:Prob_Alevels.jpg|frameless|500px]]&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;The shaded area contains the number who took all three, which can be deduced from the above information (since the total of the numbers assigned to each part of the Venn diagram must be &amp;lt;math&amp;gt;1000&amp;lt;/math&amp;gt;). The answer is therefore &amp;lt;math&amp;gt;30\%&amp;lt;/math&amp;gt; (being &amp;lt;math&amp;gt;300&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;1000&amp;lt;/math&amp;gt;).&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Two further results on unions, intersections and complements which are of use (and which are fairly easy to demonstrate using Venn diagrams) are &amp;#039;&amp;#039;&amp;#039;de Morgan Laws&amp;#039;&amp;#039;&amp;#039;:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\left( \bar{A}\cap \bar{B}\right) =\left( \overline{A\cup B}\right) &amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\bar{A}\cup \bar{B}=\left( \overline{A\cap B}\right) &amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=File:Prob_coin.jpg&amp;diff=2860</id>
		<title>File:Prob coin.jpg</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=File:Prob_coin.jpg&amp;diff=2860"/>
				<updated>2013-08-07T21:25:27Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: Admin uploaded a new version of &amp;amp;quot;File:Prob coin.jpg&amp;amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2859</id>
		<title>Probability Intro</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2859"/>
				<updated>2013-08-07T21:14:49Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: /* The addition rule of probability */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Introducing Probability =&lt;br /&gt;
&lt;br /&gt;
So far we have been looking at ways of summarising samples of data drawn from an underlying population of interest. Although at times tedious, all such arithmetic calculations are fairly mechanical and straightforward to apply. To remind ourselves, one of the primary reasons for wishing to summarise data is so assist in the development of inferences about the population from which the data were taken. That is to say, we would like to elicit some information about the mechanism which generated the observed data.&lt;br /&gt;
&lt;br /&gt;
We now start on the process of developing mathematical ways of formulating inferences and this requires the use of &amp;#039;&amp;#039;probability&amp;#039;&amp;#039;. This becomes clear if we think back to one of the early questions posed in this course: &amp;#039;&amp;#039;prior to sampling is it possible to predict with absolute certainty what will be observed&amp;#039;&amp;#039;? The answer to this question is &amp;#039;&amp;#039;no&amp;#039;&amp;#039;; although it would be of interest to know how &amp;#039;&amp;#039;likely&amp;#039;&amp;#039; it is that certain values would be observed. Or, what is the &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; of observing certain values?&lt;br /&gt;
&lt;br /&gt;
Before proceeding, we need some more tools:&lt;br /&gt;
&lt;br /&gt;
= Venn diagrams =&lt;br /&gt;
&lt;br /&gt;
Venn diagrams (and diagrams in general) are of enormous help in trying to understand, and manipulate probability. We begin with some basic definitions, some of which we have encountered before.&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Experiment:&amp;#039;&amp;#039;&amp;#039; any process which, when applied, provides data or an outcome; e.g., rolling a die and observing the number of dots on the upturned face; recording the amount of rainfall in Manchester over a period of time.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Sample Space:&amp;#039;&amp;#039;&amp;#039; set of possible outcomes of an experiment; e.g., &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; (or &amp;lt;math&amp;gt;\Omega &amp;lt;/math&amp;gt;) &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{1,2,3,4,5,6\}&amp;lt;/math&amp;gt;, which is the sample space of rolling a dice. Or &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{x;x\geq 0\}&amp;lt;/math&amp;gt;, which is the sample space of an experiment where the outcomes can be any real non-negative number, or ‘&amp;#039;&amp;#039;the set of real non-negative real numbers&amp;#039;&amp;#039;’.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Simple Event&amp;#039;&amp;#039;&amp;#039;: just one of the possible outcomes on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Event:&amp;#039;&amp;#039;&amp;#039; a &amp;#039;&amp;#039;subset&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, denoted &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;; e.g., &amp;lt;math&amp;gt;E=\left\{ 2,4,6\right\}&amp;lt;/math&amp;gt; (i.e. any even number on a dice) or &amp;lt;math&amp;gt;E=\left\{ x;4&amp;lt;x\leq 10\right\} ,&amp;lt;/math&amp;gt; which means ‘&amp;#039;&amp;#039;the set of real numbers which are strictly bigger than&amp;#039;&amp;#039; &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;but less than or equal to &amp;#039;&amp;#039;&amp;lt;math&amp;gt;10&amp;lt;/math&amp;gt;’.&amp;lt;br /&amp;gt;Note that an event, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, is a collection of simple events.&lt;br /&gt;
&lt;br /&gt;
Such concepts can be represented by means of the following Venn Diagram:&lt;br /&gt;
&lt;br /&gt;
[[File:Venn_1.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
The sample space, &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; is depicted as a closed rectangle, and the event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; is a closed loop wholly contained within &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; and we write (in set notation) &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
In dealing with probability, and in particular the probability of an event (or events) occurring, we shall need to be familiar with &amp;#039;&amp;#039;&amp;#039;UNIONS, INTERSECTIONS&amp;#039;&amp;#039;&amp;#039; and &amp;#039;&amp;#039;&amp;#039;COMPLEMENTS&amp;#039;&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
To illustrate these concepts, consider the sample space &amp;lt;math&amp;gt;S=\{x;x\geq 0\},\,&amp;lt;/math&amp;gt; with the following events defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as depicted in Figure 3.2:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E=\{x;4&amp;lt;x\leq 10\},\,F=\{x;7&amp;lt;x\leq 17\},\,G=\{x;x&amp;gt;15\},\,H=\{x;9&amp;lt;x\leq&lt;br /&gt;
13\}. &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| (a) Event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: A closed loop&lt;br /&gt;
| (b) Union: &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2a.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2b.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (c) Intersection: &amp;lt;math&amp;gt;E\cap F&amp;lt;/math&amp;gt;&lt;br /&gt;
| (d) The Null set/event: &amp;lt;math&amp;gt;E\cap G=\emptyset &amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2c.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2d.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (e) Complement of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;&lt;br /&gt;
| (f) Subset of &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;H\subset F&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;H\cap F=H&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2e.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2f.jpg|frameless|300px]]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
        &lt;br /&gt;
&lt;br /&gt;
* The &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cup F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cup F=\{x;4&amp;lt;x\leq 17\};&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are either in &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or in &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) in both. This is illustrated on the Venn diagram by the dark shaded area in diagram (b).&lt;br /&gt;
* The&amp;#039;&amp;#039; intersection&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cap F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cap F=\left\{ x;7\leq x\leq 10\right\} ;&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are common to both &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; Again this is depicted by the dark shaded area in (c). If events have no elements in common (as, for example, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt;) then they are said to be &amp;#039;&amp;#039;mutually exclusive&amp;#039;&amp;#039;, and we can write &amp;lt;math&amp;gt;E\cap G=\emptyset ,&amp;lt;/math&amp;gt; meaning the &amp;#039;&amp;#039;null set&amp;#039;&amp;#039; which contains no elements. Such a situation is illustrated on the Venn Diagram by events (the two shaded closed loops in (d)) which do not overlap. Notice however that &amp;lt;math&amp;gt;G\cap F\neq \emptyset ,&amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have elements in common.&lt;br /&gt;
* The &amp;#039;&amp;#039;complement&amp;#039;&amp;#039; of an event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; say, is everything defined on the sample space which is not in &amp;lt;math&amp;gt;E.&amp;lt;/math&amp;gt; This event is denoted &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;, the dark shaded area in (e); here &amp;lt;math&amp;gt;\bar{E}=\left\{ x;x\leq 4\right\} \cup \left\{ x;x&amp;gt;10\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
* Finally note that &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; is a sub-set of &amp;lt;math&amp;gt;F;&amp;lt;/math&amp;gt; see (f). It is depicted as the dark closed loop wholly contained within &amp;lt;math&amp;gt;F,&amp;lt;/math&amp;gt; the lighter shaded area, so that &amp;lt;math&amp;gt;H\cap F=H;&amp;lt;/math&amp;gt; if an element in the sample space is a member of &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; then it must also be member of &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; (In mathematical logic, we employ this scenario to indicate that “&amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; implies &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;”, but not necessarily vice-versa.) Notice that &amp;lt;math&amp;gt;G\cap H=\emptyset &amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;H\cap E\neq \emptyset&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
= Probability =&lt;br /&gt;
&lt;br /&gt;
The term &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; (or some equivalent) is used in everyday conversation and so can not be unfamiliar to the reader. We talk of the probability, or chance, of rain; the likelihood of England winning the World Cup; or, perhaps more scientifically, the chance of getting a &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; when rolling a die. What we shall now do is develop a coherent theory of probability; a theory which allows us to combine and manipulate probabilities in a consistent and meaningful manner. We shall describe ways of dealing with, and describing, uncertainty. This will involve &amp;#039;&amp;#039;rules&amp;#039;&amp;#039; which govern our use of terms like probability.&lt;br /&gt;
&lt;br /&gt;
There have been a number of different approaches (interpretations) of probability. Most depend, at least to some extent, on the notion of relative frequency as now described:&lt;br /&gt;
&lt;br /&gt;
* Suppose an experiment has an outcome of interest &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;. The &amp;#039;&amp;#039;relative frequency interpretation&amp;#039;&amp;#039; of probability says that assuming the experiment can be repeated a large number of times then the relative frequency of observing the outcome &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; will settle down to a &amp;#039;&amp;#039;number&amp;#039;&amp;#039;, denoted &amp;lt;math&amp;gt;\Pr (E),&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;P(E)&amp;lt;/math&amp;gt; or Prob&amp;lt;math&amp;gt;(E),&amp;lt;/math&amp;gt; called the &amp;#039;&amp;#039;&amp;#039;probability&amp;#039;&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
This is illustrated in the next Figure where the proportion of heads obtained after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; flips of a fair coin is plotted against &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;, as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; increases; e.g., of the first &amp;lt;math&amp;gt;100&amp;lt;/math&amp;gt; flips, &amp;lt;math&amp;gt;55&amp;lt;/math&amp;gt; were heads (&amp;lt;math&amp;gt;55\%&amp;lt;/math&amp;gt;). Notice that the plot becomes less ‘wobbly’ after about &amp;lt;math&amp;gt;n=220&amp;lt;/math&amp;gt; and appears to be settling down to the value of &amp;lt;math&amp;gt;\frac{1}{2}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
[[File:Prob_coin.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
Due to this interpretation of probability, we often use observed sample proportions to approximate underlying probabilities of interest; see, for example, Question 4 of Exercise 2. There are, of course, other interpretations of probability; e.g., the subjective interpretation which simply expresses the strength of one’s belief about an event of interest such as whether Manchester United will win the European Cup! Any one of these interpretations can be used in practical situations provided the implied notion of probability follows a simple set of &amp;#039;&amp;#039;axioms&amp;#039;&amp;#039; or &amp;#039;&amp;#039;rules&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
== The axioms of probability ==&lt;br /&gt;
&lt;br /&gt;
There are just &amp;#039;&amp;#039;three &amp;#039;&amp;#039;basic rules that must be obeyed when dealing with probabilities:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;For any event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; i.e., &amp;lt;math&amp;gt;E\subset S,\,\,\Pr (E)\geq 0&amp;lt;/math&amp;gt;; &amp;#039;&amp;#039;probabilities are non-negative&amp;#039;&amp;#039;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (S)=1;&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;having defined the sample space of outcomes, one of these outcomes must be observed&amp;#039;&amp;#039;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;If events &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are mutually exclusive defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, so that &amp;lt;math&amp;gt;E\cap F=\emptyset &amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\Pr \left( E\right)+\Pr \left( F\right) .&amp;lt;/math&amp;gt; In general, for any set of mutually exclusive events, &amp;lt;math&amp;gt;E_{1},E_{2},\ldots ,E_{k},&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S:&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (E_{1}\cup E_{2}\cup \ldots \cup E_{k})=\Pr (E_{1})+\Pr (E_{2})+\ldots\Pr (E_{k})&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;i.e., &amp;lt;math&amp;gt;\Pr \left( \bigcup_{j=1}^{k}E_{j}\right) =\sum_{j=1}^{k}\Pr (E_{j}).&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In terms of the Venn Diagram, one can (and should) usefully think of the area of &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; relative to that of &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as providing an indication of probability. (Note, from axiom 2, that the area of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is implicitly normalised to be unity).&lt;br /&gt;
&lt;br /&gt;
Also observe that, contrary to what you may have believed, it is not one of the rules that &amp;lt;math&amp;gt;\Pr (E)\leq 1&amp;lt;/math&amp;gt; for any event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;. Rather, this is an implication of the &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; rules given:&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;implications: &amp;#039;&amp;#039;&amp;#039;it must be that for any event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;E\cap \bar{E}=\emptyset &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;E\cup \bar{E}=S.&amp;lt;/math&amp;gt; By Axiom &amp;lt;math&amp;gt;1,&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr (E)\geq 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr \left( \bar{E}\right) \geq 0&amp;lt;/math&amp;gt; and by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr(E)+\Pr (\bar{E})=\Pr (S).&amp;lt;/math&amp;gt; So &amp;lt;math&amp;gt;\Pr \left( E\right) +\Pr \left( \bar{E}\right) =1,&amp;lt;/math&amp;gt; by Axiom &amp;lt;math&amp;gt;2.&amp;lt;/math&amp;gt; This implies that&lt;br /&gt;
&lt;br /&gt;
# &amp;lt;math&amp;gt;0\leq \Pr (E)\leq 1&amp;lt;/math&amp;gt;&lt;br /&gt;
# &amp;lt;math&amp;gt;\Pr (\bar{E})=1-\Pr (E)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The first of these is what we might have expected from probability (a number lying between &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;). The second implication is also very important; it says that the probability of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; not happening is ‘&amp;#039;&amp;#039;one minus the probability of it happening&amp;#039;&amp;#039;’. Thus when rolling a die, the probability of getting &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; is one minus the probability of getting either a &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;2&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;5.&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
These axioms imply how to calculate probabilities on a sample space of equally likely outcomes. For example, and as we have already noted, the experiment of rolling a fair die defines a sample space of six, mutually exclusive and equally likely outcomes (&amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; dots on the up-turned face). The axioms then say that each of the six probabilities are positive, add to 1 and are all the same. Thus, the probability of any one of the outcomes must be simply &amp;lt;math&amp;gt;\frac{1}{6};&amp;lt;/math&amp;gt; which may accord with your intuition. A similar sort of analysis reveals that the probability of drawing a club from a deck of &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; cards is &amp;lt;math&amp;gt;\frac{13}{52},&amp;lt;/math&amp;gt; since any one of the &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; cards has an equal chance of being drawn and &amp;lt;math&amp;gt;13&amp;lt;/math&amp;gt; of them are clubs; i.e., &amp;lt;math&amp;gt;13&amp;lt;/math&amp;gt; of the &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; are clubs, so the probability of drawing a club is &amp;lt;math&amp;gt;\frac{13}{52}.&amp;lt;/math&amp;gt; Notice the importance of the assumption of equally likely outcomes here.&lt;br /&gt;
&lt;br /&gt;
In this, and the next section of notes, we shall see how these axioms can be used. Firstly, consider the construction of a probability for the &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of two events; i.e., the probability that &amp;#039;&amp;#039;either &amp;#039;&amp;#039;&amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) &amp;#039;&amp;#039;both &amp;#039;&amp;#039;will occur. Such a probability is embodied in the &amp;#039;&amp;#039;addition rule of probability&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
== The addition rule of probability ==&lt;br /&gt;
&lt;br /&gt;
When rolling a fair die, let &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; denote the event of an “odd number of dots” and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; the event of the “number of dots being greater than, or equal, to &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;”&amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt; What is the probability of the event &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;? To calculate this we can collect together all the mutually exclusive (simple) events which comprise &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;, and then add up the probabilities (by axiom 3). These simple events are &amp;lt;math&amp;gt;1,3,4,5&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; dots. Each has a probability of &amp;lt;math&amp;gt;\frac{1}{6},&amp;lt;/math&amp;gt; so the required total probability is: &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\frac{5}{6}&amp;lt;/math&amp;gt;. Consider carefully how this probability is constructed and note, in particular, that &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) \neq \Pr \left( E\right) +\Pr \left( F\right) &amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have a simple event in common (namely &amp;lt;math&amp;gt;5&amp;lt;/math&amp;gt; dots).&lt;br /&gt;
&lt;br /&gt;
In general, we can calculate the probability of the union of events using the &amp;#039;&amp;#039;addition rule of probability&amp;#039;&amp;#039;, as follows.&lt;br /&gt;
&lt;br /&gt;
* For any events, &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F\subset S:\Pr (E\cup F)=\Pr (E)+\Pr (F)-\Pr (E\cap F).&amp;lt;/math&amp;gt; So, in general, &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) \leq \Pr (E)+\Pr (F).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This generalises to three events, &amp;lt;math&amp;gt;E_{1},E_{2}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;E_{3}&amp;lt;/math&amp;gt; as&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E_{1}\cup E_{2}\cup E_{3}) &amp;amp;=&amp;amp;\Pr (E_{1})+\Pr (E_{2})+\Pr (E_{3}) \\&lt;br /&gt;
&amp;amp;&amp;amp;-\Pr (E_{1}\cap E_{2})-\Pr (E_{1}\cap E_{3})-\Pr (E_{2}\cap E_{3}) \\&lt;br /&gt;
&amp;amp;&amp;amp;+\Pr (E_{1}\cap E_{2}\cap E_{3}).\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We can demonstrate this as follows.&lt;br /&gt;
&lt;br /&gt;
Note that&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E\cup F=\left( E\cap \bar{F}\right) \cup \left( E\cap F\right) \cup \left(\bar{E}\cap F\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
the union of &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; mutually exclusive events. These mutually exclusive events are depicted by the shaded areas &amp;lt;math&amp;gt;\mathbf{a,}&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\mathbf{b}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{c}&amp;lt;/math&amp;gt;, respectively, in the next Figure.&lt;br /&gt;
&lt;br /&gt;
[[File:Prob_add.jpg|frameless|500px]]&lt;br /&gt;
&lt;br /&gt;
Then by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, and from the fact that the three events &amp;lt;math&amp;gt;\left( E\cap\bar{F}\right) &amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\left( E\cap F\right) &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\left( \bar{E}\cap F\right)&amp;lt;/math&amp;gt; are mutually exclusive so that the “area” occupied by &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt; is simply &amp;lt;math&amp;gt;\mathbf{a+b+c,}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\Pr \left( E\cap \bar{F}\right) +\Pr \left( \bar{E}\cap F\right) +\Pr \left( E\cap F\right) .&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
But also by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, since &amp;lt;math&amp;gt;E=\left( E\cap \bar{F}\right) \cup \left(E\cap F\right) &amp;lt;/math&amp;gt;, it must be that &amp;lt;math&amp;gt;\Pr (E)=\Pr \left( E\cap \bar{F}\right)+\Pr (E\cap F);&amp;lt;/math&amp;gt; similarly, &amp;lt;math&amp;gt;\Pr \left( \bar{E}\cap F\right) =\Pr \left(F\right) -\Pr \left( E\cap F\right)&amp;lt;/math&amp;gt;. Putting all of this together gives&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\Pr (E\cup F)=\Pr (E)+\Pr (F)-\Pr (E\cap F).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
When &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are mutually exclusive, so that &amp;lt;math&amp;gt;E\cap F=\emptyset&amp;lt;/math&amp;gt;, this rule reveals Axiom 2: &amp;lt;math&amp;gt;\Pr (E\cup F)=\Pr (E)+\Pr (F)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example: &amp;#039;&amp;#039;What is the probability of drawing a Queen (&amp;lt;math&amp;gt;Q &amp;lt;/math&amp;gt;) or a Club (&amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;) in a single draw from a pack of cards? Now, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;52 &amp;lt;/math&amp;gt; cards are Queens, so &amp;lt;math&amp;gt;\Pr \left( Q\right) =\frac{4}{52},&amp;lt;/math&amp;gt; whilst &amp;lt;math&amp;gt;\Pr\left( C\right) =\frac{13}{52}.&amp;lt;/math&amp;gt; The probability of drawing the Queen of Clubs is simply &amp;lt;math&amp;gt;\frac{1}{52};&amp;lt;/math&amp;gt; i.e., &amp;lt;math&amp;gt;\Pr \left( Q\cap C\right) =\frac{1}{52}&amp;lt;/math&amp;gt;. What we require is a Club or a Queen, for which the probability is&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr \left( Q\cup C\right) &amp;amp;=&amp;amp;\Pr \left( Q\right) +\Pr \left( C\right) -\Pr\left( Q\cap C\right) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{4}{52}+\frac{13}{52}-\frac{1}{52} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{16}{52}=\frac{4}{13}.\end{aligned}&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example: &amp;#039;&amp;#039;Consider a car journey from Manchester to London via the M6 and M1. Let &amp;lt;math&amp;gt;E=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;heavy traffic somewhere on route&amp;#039;&amp;#039; and &amp;lt;math&amp;gt;F=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;roadworks somewhere on route&amp;#039;&amp;#039;. It is estimated that &amp;lt;math&amp;gt;\Pr (E)=0.8&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr (F)=0.4,&amp;lt;/math&amp;gt; whilst the probability of NOT encountering both is &amp;lt;math&amp;gt;\Pr (\overline{E\cap F})=0.6.&amp;lt;/math&amp;gt; What is the probability of encountering heavytraffic or roadworks?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;We require &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) .&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E\cup F) &amp;amp;=&amp;amp;\Pr (E)+\Pr (F)-\Pr (E\cap F) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\Pr (E)+\Pr (F)-(1-\Pr (\overline{E\cap F})) \\&lt;br /&gt;
&amp;amp;=&amp;amp;0.8+0.4-1+0.6 \\&lt;br /&gt;
&amp;amp;=&amp;amp;0.8=\Pr (E)\end{aligned}&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Notice that this implies, in this case, that &amp;lt;math&amp;gt;F\subset E&amp;lt;/math&amp;gt; (why?). This &amp;#039;&amp;#039;model &amp;#039;&amp;#039;then implies that when there are roadworks somewhere on route you are bound to encounter heavy traffic; on the other hand, you can encounter heavy traffic on route without ever passing through roadworks. (My own experience of this motorway inclines me towards this implication!)&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Similar concepts apply when manipulating proportions as follows:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example&amp;#039;&amp;#039;: A sample of 1000 undergraduates were asked whether they took either Mathematics, Physics or Chemistry at A-level. The following responses were obtained: 100 just took Mathematics; 70 just took Physics; 100 just took Chemistry; 150 took Mathematics and Physics, but not Chemistry; 40 took Mathematics and Chemistry, but not Physics; and, 240 took Physics and Chemistry, but not Mathematics. What proportion took all three?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;This can be addressed with the following diagram:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;[[File:Prob_Alevels.jpg|frameless|500px]]&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;The shaded area contains the number who took all three, which can be deduced from the above information (since the total of the numbers assigned to each part of the Venn diagram must be &amp;lt;math&amp;gt;1000&amp;lt;/math&amp;gt;). The answer is therefore &amp;lt;math&amp;gt;30\%&amp;lt;/math&amp;gt; (being &amp;lt;math&amp;gt;300&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;1000&amp;lt;/math&amp;gt;).&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Two further results on unions, intersections and complements which are of use (and which are fairly easy to demonstrate using Venn diagrams) are &amp;#039;&amp;#039;&amp;#039;de Morgan Laws&amp;#039;&amp;#039;&amp;#039;:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\left( \bar{A}\cap \bar{B}\right) =\left( \overline{A\cup B}\right) &amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\bar{A}\cup \bar{B}=\left( \overline{A\cap B}\right) &amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2858</id>
		<title>Probability Intro</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2858"/>
				<updated>2013-08-07T21:12:08Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: /* The addition rule of probability */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Introducing Probability =&lt;br /&gt;
&lt;br /&gt;
So far we have been looking at ways of summarising samples of data drawn from an underlying population of interest. Although at times tedious, all such arithmetic calculations are fairly mechanical and straightforward to apply. To remind ourselves, one of the primary reasons for wishing to summarise data is so assist in the development of inferences about the population from which the data were taken. That is to say, we would like to elicit some information about the mechanism which generated the observed data.&lt;br /&gt;
&lt;br /&gt;
We now start on the process of developing mathematical ways of formulating inferences and this requires the use of &amp;#039;&amp;#039;probability&amp;#039;&amp;#039;. This becomes clear if we think back to one of the early questions posed in this course: &amp;#039;&amp;#039;prior to sampling is it possible to predict with absolute certainty what will be observed&amp;#039;&amp;#039;? The answer to this question is &amp;#039;&amp;#039;no&amp;#039;&amp;#039;; although it would be of interest to know how &amp;#039;&amp;#039;likely&amp;#039;&amp;#039; it is that certain values would be observed. Or, what is the &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; of observing certain values?&lt;br /&gt;
&lt;br /&gt;
Before proceeding, we need some more tools:&lt;br /&gt;
&lt;br /&gt;
= Venn diagrams =&lt;br /&gt;
&lt;br /&gt;
Venn diagrams (and diagrams in general) are of enormous help in trying to understand, and manipulate probability. We begin with some basic definitions, some of which we have encountered before.&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Experiment:&amp;#039;&amp;#039;&amp;#039; any process which, when applied, provides data or an outcome; e.g., rolling a die and observing the number of dots on the upturned face; recording the amount of rainfall in Manchester over a period of time.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Sample Space:&amp;#039;&amp;#039;&amp;#039; set of possible outcomes of an experiment; e.g., &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; (or &amp;lt;math&amp;gt;\Omega &amp;lt;/math&amp;gt;) &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{1,2,3,4,5,6\}&amp;lt;/math&amp;gt;, which is the sample space of rolling a dice. Or &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{x;x\geq 0\}&amp;lt;/math&amp;gt;, which is the sample space of an experiment where the outcomes can be any real non-negative number, or ‘&amp;#039;&amp;#039;the set of real non-negative real numbers&amp;#039;&amp;#039;’.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Simple Event&amp;#039;&amp;#039;&amp;#039;: just one of the possible outcomes on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Event:&amp;#039;&amp;#039;&amp;#039; a &amp;#039;&amp;#039;subset&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, denoted &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;; e.g., &amp;lt;math&amp;gt;E=\left\{ 2,4,6\right\}&amp;lt;/math&amp;gt; (i.e. any even number on a dice) or &amp;lt;math&amp;gt;E=\left\{ x;4&amp;lt;x\leq 10\right\} ,&amp;lt;/math&amp;gt; which means ‘&amp;#039;&amp;#039;the set of real numbers which are strictly bigger than&amp;#039;&amp;#039; &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;but less than or equal to &amp;#039;&amp;#039;&amp;lt;math&amp;gt;10&amp;lt;/math&amp;gt;’.&amp;lt;br /&amp;gt;Note that an event, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, is a collection of simple events.&lt;br /&gt;
&lt;br /&gt;
Such concepts can be represented by means of the following Venn Diagram:&lt;br /&gt;
&lt;br /&gt;
[[File:Venn_1.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
The sample space, &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; is depicted as a closed rectangle, and the event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; is a closed loop wholly contained within &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; and we write (in set notation) &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
In dealing with probability, and in particular the probability of an event (or events) occurring, we shall need to be familiar with &amp;#039;&amp;#039;&amp;#039;UNIONS, INTERSECTIONS&amp;#039;&amp;#039;&amp;#039; and &amp;#039;&amp;#039;&amp;#039;COMPLEMENTS&amp;#039;&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
To illustrate these concepts, consider the sample space &amp;lt;math&amp;gt;S=\{x;x\geq 0\},\,&amp;lt;/math&amp;gt; with the following events defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as depicted in Figure 3.2:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E=\{x;4&amp;lt;x\leq 10\},\,F=\{x;7&amp;lt;x\leq 17\},\,G=\{x;x&amp;gt;15\},\,H=\{x;9&amp;lt;x\leq&lt;br /&gt;
13\}. &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| (a) Event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: A closed loop&lt;br /&gt;
| (b) Union: &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2a.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2b.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (c) Intersection: &amp;lt;math&amp;gt;E\cap F&amp;lt;/math&amp;gt;&lt;br /&gt;
| (d) The Null set/event: &amp;lt;math&amp;gt;E\cap G=\emptyset &amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2c.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2d.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (e) Complement of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;&lt;br /&gt;
| (f) Subset of &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;H\subset F&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;H\cap F=H&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2e.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2f.jpg|frameless|300px]]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
        &lt;br /&gt;
&lt;br /&gt;
* The &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cup F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cup F=\{x;4&amp;lt;x\leq 17\};&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are either in &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or in &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) in both. This is illustrated on the Venn diagram by the dark shaded area in diagram (b).&lt;br /&gt;
* The&amp;#039;&amp;#039; intersection&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cap F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cap F=\left\{ x;7\leq x\leq 10\right\} ;&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are common to both &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; Again this is depicted by the dark shaded area in (c). If events have no elements in common (as, for example, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt;) then they are said to be &amp;#039;&amp;#039;mutually exclusive&amp;#039;&amp;#039;, and we can write &amp;lt;math&amp;gt;E\cap G=\emptyset ,&amp;lt;/math&amp;gt; meaning the &amp;#039;&amp;#039;null set&amp;#039;&amp;#039; which contains no elements. Such a situation is illustrated on the Venn Diagram by events (the two shaded closed loops in (d)) which do not overlap. Notice however that &amp;lt;math&amp;gt;G\cap F\neq \emptyset ,&amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have elements in common.&lt;br /&gt;
* The &amp;#039;&amp;#039;complement&amp;#039;&amp;#039; of an event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; say, is everything defined on the sample space which is not in &amp;lt;math&amp;gt;E.&amp;lt;/math&amp;gt; This event is denoted &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;, the dark shaded area in (e); here &amp;lt;math&amp;gt;\bar{E}=\left\{ x;x\leq 4\right\} \cup \left\{ x;x&amp;gt;10\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
* Finally note that &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; is a sub-set of &amp;lt;math&amp;gt;F;&amp;lt;/math&amp;gt; see (f). It is depicted as the dark closed loop wholly contained within &amp;lt;math&amp;gt;F,&amp;lt;/math&amp;gt; the lighter shaded area, so that &amp;lt;math&amp;gt;H\cap F=H;&amp;lt;/math&amp;gt; if an element in the sample space is a member of &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; then it must also be member of &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; (In mathematical logic, we employ this scenario to indicate that “&amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; implies &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;”, but not necessarily vice-versa.) Notice that &amp;lt;math&amp;gt;G\cap H=\emptyset &amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;H\cap E\neq \emptyset&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
= Probability =&lt;br /&gt;
&lt;br /&gt;
The term &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; (or some equivalent) is used in everyday conversation and so can not be unfamiliar to the reader. We talk of the probability, or chance, of rain; the likelihood of England winning the World Cup; or, perhaps more scientifically, the chance of getting a &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; when rolling a die. What we shall now do is develop a coherent theory of probability; a theory which allows us to combine and manipulate probabilities in a consistent and meaningful manner. We shall describe ways of dealing with, and describing, uncertainty. This will involve &amp;#039;&amp;#039;rules&amp;#039;&amp;#039; which govern our use of terms like probability.&lt;br /&gt;
&lt;br /&gt;
There have been a number of different approaches (interpretations) of probability. Most depend, at least to some extent, on the notion of relative frequency as now described:&lt;br /&gt;
&lt;br /&gt;
* Suppose an experiment has an outcome of interest &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;. The &amp;#039;&amp;#039;relative frequency interpretation&amp;#039;&amp;#039; of probability says that assuming the experiment can be repeated a large number of times then the relative frequency of observing the outcome &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; will settle down to a &amp;#039;&amp;#039;number&amp;#039;&amp;#039;, denoted &amp;lt;math&amp;gt;\Pr (E),&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;P(E)&amp;lt;/math&amp;gt; or Prob&amp;lt;math&amp;gt;(E),&amp;lt;/math&amp;gt; called the &amp;#039;&amp;#039;&amp;#039;probability&amp;#039;&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
This is illustrated in the next Figure where the proportion of heads obtained after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; flips of a fair coin is plotted against &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;, as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; increases; e.g., of the first &amp;lt;math&amp;gt;100&amp;lt;/math&amp;gt; flips, &amp;lt;math&amp;gt;55&amp;lt;/math&amp;gt; were heads (&amp;lt;math&amp;gt;55\%&amp;lt;/math&amp;gt;). Notice that the plot becomes less ‘wobbly’ after about &amp;lt;math&amp;gt;n=220&amp;lt;/math&amp;gt; and appears to be settling down to the value of &amp;lt;math&amp;gt;\frac{1}{2}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
[[File:Prob_coin.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
Due to this interpretation of probability, we often use observed sample proportions to approximate underlying probabilities of interest; see, for example, Question 4 of Exercise 2. There are, of course, other interpretations of probability; e.g., the subjective interpretation which simply expresses the strength of one’s belief about an event of interest such as whether Manchester United will win the European Cup! Any one of these interpretations can be used in practical situations provided the implied notion of probability follows a simple set of &amp;#039;&amp;#039;axioms&amp;#039;&amp;#039; or &amp;#039;&amp;#039;rules&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
== The axioms of probability ==&lt;br /&gt;
&lt;br /&gt;
There are just &amp;#039;&amp;#039;three &amp;#039;&amp;#039;basic rules that must be obeyed when dealing with probabilities:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;For any event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; i.e., &amp;lt;math&amp;gt;E\subset S,\,\,\Pr (E)\geq 0&amp;lt;/math&amp;gt;; &amp;#039;&amp;#039;probabilities are non-negative&amp;#039;&amp;#039;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (S)=1;&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;having defined the sample space of outcomes, one of these outcomes must be observed&amp;#039;&amp;#039;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;If events &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are mutually exclusive defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, so that &amp;lt;math&amp;gt;E\cap F=\emptyset &amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\Pr \left( E\right)+\Pr \left( F\right) .&amp;lt;/math&amp;gt; In general, for any set of mutually exclusive events, &amp;lt;math&amp;gt;E_{1},E_{2},\ldots ,E_{k},&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S:&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (E_{1}\cup E_{2}\cup \ldots \cup E_{k})=\Pr (E_{1})+\Pr (E_{2})+\ldots\Pr (E_{k})&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;i.e., &amp;lt;math&amp;gt;\Pr \left( \bigcup_{j=1}^{k}E_{j}\right) =\sum_{j=1}^{k}\Pr (E_{j}).&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In terms of the Venn Diagram, one can (and should) usefully think of the area of &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; relative to that of &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as providing an indication of probability. (Note, from axiom 2, that the area of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is implicitly normalised to be unity).&lt;br /&gt;
&lt;br /&gt;
Also observe that, contrary to what you may have believed, it is not one of the rules that &amp;lt;math&amp;gt;\Pr (E)\leq 1&amp;lt;/math&amp;gt; for any event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;. Rather, this is an implication of the &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; rules given:&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;implications: &amp;#039;&amp;#039;&amp;#039;it must be that for any event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;E\cap \bar{E}=\emptyset &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;E\cup \bar{E}=S.&amp;lt;/math&amp;gt; By Axiom &amp;lt;math&amp;gt;1,&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr (E)\geq 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr \left( \bar{E}\right) \geq 0&amp;lt;/math&amp;gt; and by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr(E)+\Pr (\bar{E})=\Pr (S).&amp;lt;/math&amp;gt; So &amp;lt;math&amp;gt;\Pr \left( E\right) +\Pr \left( \bar{E}\right) =1,&amp;lt;/math&amp;gt; by Axiom &amp;lt;math&amp;gt;2.&amp;lt;/math&amp;gt; This implies that&lt;br /&gt;
&lt;br /&gt;
# &amp;lt;math&amp;gt;0\leq \Pr (E)\leq 1&amp;lt;/math&amp;gt;&lt;br /&gt;
# &amp;lt;math&amp;gt;\Pr (\bar{E})=1-\Pr (E)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The first of these is what we might have expected from probability (a number lying between &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;). The second implication is also very important; it says that the probability of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; not happening is ‘&amp;#039;&amp;#039;one minus the probability of it happening&amp;#039;&amp;#039;’. Thus when rolling a die, the probability of getting &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; is one minus the probability of getting either a &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;2&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;5.&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
These axioms imply how to calculate probabilities on a sample space of equally likely outcomes. For example, and as we have already noted, the experiment of rolling a fair die defines a sample space of six, mutually exclusive and equally likely outcomes (&amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; dots on the up-turned face). The axioms then say that each of the six probabilities are positive, add to 1 and are all the same. Thus, the probability of any one of the outcomes must be simply &amp;lt;math&amp;gt;\frac{1}{6};&amp;lt;/math&amp;gt; which may accord with your intuition. A similar sort of analysis reveals that the probability of drawing a club from a deck of &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; cards is &amp;lt;math&amp;gt;\frac{13}{52},&amp;lt;/math&amp;gt; since any one of the &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; cards has an equal chance of being drawn and &amp;lt;math&amp;gt;13&amp;lt;/math&amp;gt; of them are clubs; i.e., &amp;lt;math&amp;gt;13&amp;lt;/math&amp;gt; of the &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; are clubs, so the probability of drawing a club is &amp;lt;math&amp;gt;\frac{13}{52}.&amp;lt;/math&amp;gt; Notice the importance of the assumption of equally likely outcomes here.&lt;br /&gt;
&lt;br /&gt;
In this, and the next section of notes, we shall see how these axioms can be used. Firstly, consider the construction of a probability for the &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of two events; i.e., the probability that &amp;#039;&amp;#039;either &amp;#039;&amp;#039;&amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) &amp;#039;&amp;#039;both &amp;#039;&amp;#039;will occur. Such a probability is embodied in the &amp;#039;&amp;#039;addition rule of probability&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
== The addition rule of probability ==&lt;br /&gt;
&lt;br /&gt;
When rolling a fair die, let &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; denote the event of an “odd number of dots” and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; the event of the “number of dots being greater than, or equal, to &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;”&amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt; What is the probability of the event &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;? To calculate this we can collect together all the mutually exclusive (simple) events which comprise &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;, and then add up the probabilities (by axiom 3). These simple events are &amp;lt;math&amp;gt;1,3,4,5&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; dots. Each has a probability of &amp;lt;math&amp;gt;\frac{1}{6},&amp;lt;/math&amp;gt; so the required total probability is: &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\frac{5}{6}&amp;lt;/math&amp;gt;. Consider carefully how this probability is constructed and note, in particular, that &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) \neq \Pr \left( E\right) +\Pr \left( F\right) &amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have a simple event in common (namely &amp;lt;math&amp;gt;5&amp;lt;/math&amp;gt; dots).&lt;br /&gt;
&lt;br /&gt;
In general, we can calculate the probability of the union of events using the &amp;#039;&amp;#039;addition rule of probability&amp;#039;&amp;#039;, as follows.&lt;br /&gt;
&lt;br /&gt;
* For any events, &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F\subset S:\Pr (E\cup F)=\Pr (E)+\Pr (F)-\Pr (E\cap F).&amp;lt;/math&amp;gt; So, in general, &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) \leq \Pr (E)+\Pr (F).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This generalises to three events, &amp;lt;math&amp;gt;E_{1},E_{2}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;E_{3}&amp;lt;/math&amp;gt; as&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E_{1}\cup E_{2}\cup E_{3}) &amp;amp;=&amp;amp;\Pr (E_{1})+\Pr (E_{2})+\Pr (E_{3}) \\&lt;br /&gt;
&amp;amp;&amp;amp;-\Pr (E_{1}\cap E_{2})-\Pr (E_{1}\cap E_{3})-\Pr (E_{2}\cap E_{3}) \\&lt;br /&gt;
&amp;amp;&amp;amp;+\Pr (E_{1}\cap E_{2}\cap E_{3}).\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We can demonstrate this as follows.&lt;br /&gt;
&lt;br /&gt;
Note that&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E\cup F=\left( E\cap \bar{F}\right) \cup \left( E\cap F\right) \cup \left(\bar{E}\cap F\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
the union of &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; mutually exclusive events. These mutually exclusive events are depicted by the shaded areas &amp;lt;math&amp;gt;\mathbf{a,}&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\mathbf{b}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{c}&amp;lt;/math&amp;gt;, respectively, in the next Figure.&lt;br /&gt;
&lt;br /&gt;
[[File:Prob_add.jpg|frameless|500px]]&lt;br /&gt;
&lt;br /&gt;
Then by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, and from the fact that the three events &amp;lt;math&amp;gt;\left( E\cap\bar{F}\right) &amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\left( E\cap F\right) &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\left( \bar{E}\cap F\right)&amp;lt;/math&amp;gt; are mutually exclusive so that the “area” occupied by &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt; is simply &amp;lt;math&amp;gt;\mathbf{a+b+c,}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\Pr \left( E\cap \bar{F}\right) +\Pr \left( \bar{E}\cap F\right) +\Pr \left( E\cap F\right) .&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
But also by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, since &amp;lt;math&amp;gt;E=\left( E\cap \bar{F}\right) \cup \left(E\cap F\right) &amp;lt;/math&amp;gt;, it must be that &amp;lt;math&amp;gt;\Pr (E)=\Pr \left( E\cap \bar{F}\right)+\Pr (E\cap F);&amp;lt;/math&amp;gt; similarly, &amp;lt;math&amp;gt;\Pr \left( \bar{E}\cap F\right) =\Pr \left(F\right) -\Pr \left( E\cap F\right)&amp;lt;/math&amp;gt;. Putting all of this together gives&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\Pr (E\cup F)=\Pr (E)+\Pr (F)-\Pr (E\cap F).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
When &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are mutually exclusive, so that &amp;lt;math&amp;gt;E\cap F=\emptyset&amp;lt;/math&amp;gt;, this rule reveals Axiom 2: &amp;lt;math&amp;gt;\Pr (E\cup F)=\Pr (E)+\Pr (F)&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example: &amp;#039;&amp;#039;What is the probability of drawing a Queen (&amp;lt;math&amp;gt;Q &amp;lt;/math&amp;gt;) or a Club (&amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;) in a single draw from a pack of cards? Now, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;52 &amp;lt;/math&amp;gt; cards are Queens, so &amp;lt;math&amp;gt;\Pr \left( Q\right) =\frac{4}{52},&amp;lt;/math&amp;gt; whilst &amp;lt;math&amp;gt;\Pr\left( C\right) =\frac{13}{52}.&amp;lt;/math&amp;gt; The probability of drawing the Queen of Clubs is simply &amp;lt;math&amp;gt;\frac{1}{52};&amp;lt;/math&amp;gt; i.e., &amp;lt;math&amp;gt;\Pr \left( Q\cap C\right) =\frac{1}{52}&amp;lt;/math&amp;gt;. What we require is a Club or a Queen, for which the probability is&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr \left( Q\cup C\right) &amp;amp;=&amp;amp;\Pr \left( Q\right) +\Pr \left( C\right) -\Pr\left( Q\cap C\right) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{4}{52}+\frac{13}{52}-\frac{1}{52} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{16}{52}=\frac{4}{13}.\end{aligned}&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example: &amp;#039;&amp;#039;Consider a car journey from Manchester to London via the M6 and M1. Let &amp;lt;math&amp;gt;E=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;heavy traffic somewhere on route&amp;#039;&amp;#039; and &amp;lt;math&amp;gt;F=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;roadworks somewhere on route&amp;#039;&amp;#039;. It is estimated that &amp;lt;math&amp;gt;\Pr (E)=0.8&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr (F)=0.4,&amp;lt;/math&amp;gt; whilst the probability of NOT encountering both is &amp;lt;math&amp;gt;\Pr (\overline{E\cap F})=0.6.&amp;lt;/math&amp;gt; What is the probability of encountering heavytraffic or roadworks?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;We require &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) .&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E\cup F) &amp;amp;=&amp;amp;\Pr (E)+\Pr (F)-\Pr (E\cap F) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\Pr (E)+\Pr (F)-(1-\Pr (\overline{E\cap F})) \\&lt;br /&gt;
&amp;amp;=&amp;amp;0.8+0.4-1+0.6 \\&lt;br /&gt;
&amp;amp;=&amp;amp;0.8=\Pr (E)\end{aligned}&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Notice that this implies, in this case, that &amp;lt;math&amp;gt;F\subset E&amp;lt;/math&amp;gt; (why?). This &amp;#039;&amp;#039;model &amp;#039;&amp;#039;then implies that when there are roadworks somewhere on route you are bound to encounter heavy traffic; on the other hand, you can encounter heavy traffic on route without ever passing through roadworks. (My own experience of this motorway inclines me towards this implication!)&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Similar concepts apply when manipulating proportions as follows:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example&amp;#039;&amp;#039;&amp;lt;math&amp;gt;:&amp;lt;/math&amp;gt; A sample of &amp;lt;math&amp;gt;1000&amp;lt;/math&amp;gt; undergraduates were asked whether they took either Mathematics, Physics or Chemistry at A-level. The following responses were obtained: &amp;lt;math&amp;gt;100&amp;lt;/math&amp;gt; just took Mathematics; &amp;lt;math&amp;gt;70&amp;lt;/math&amp;gt; just took Physics; &amp;lt;math&amp;gt;100&amp;lt;/math&amp;gt; just took Chemistry; &amp;lt;math&amp;gt;150&amp;lt;/math&amp;gt; took Mathematics and Physics, but not Chemistry; &amp;lt;math&amp;gt;40&amp;lt;/math&amp;gt; took Mathematics and Chemistry, but not Physics; and, &amp;lt;math&amp;gt;240&amp;lt;/math&amp;gt; took Physics and Chemistry, but not Mathematics. What proportion took all three?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;This can be addressed with the following diagram:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;[[File:Prob_Alevels.jpg|frameless|500px]]&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;The shaded area contains the number who took all three, which can be deduced from the above information (since the total of the numbers assigned to each part of the Venn diagram must be &amp;lt;math&amp;gt;1000&amp;lt;/math&amp;gt;). The answer is therefore &amp;lt;math&amp;gt;30\%&amp;lt;/math&amp;gt; (being &amp;lt;math&amp;gt;300&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;1000&amp;lt;/math&amp;gt;).&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Two further results on unions, intersections and complements which are of use (and which are fairly easy to demonstrate using Venn diagrams) are &amp;#039;&amp;#039;&amp;#039;de Morgan Laws&amp;#039;&amp;#039;&amp;#039;:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\left( \bar{A}\cap \bar{B}\right) =\left( \overline{A\cup B}\right) &amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\bar{A}\cup \bar{B}=\left( \overline{A\cap B}\right) &amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2857</id>
		<title>Probability Intro</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2857"/>
				<updated>2013-08-07T21:11:46Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: /* The addition rule of probability */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Introducing Probability =&lt;br /&gt;
&lt;br /&gt;
So far we have been looking at ways of summarising samples of data drawn from an underlying population of interest. Although at times tedious, all such arithmetic calculations are fairly mechanical and straightforward to apply. To remind ourselves, one of the primary reasons for wishing to summarise data is so assist in the development of inferences about the population from which the data were taken. That is to say, we would like to elicit some information about the mechanism which generated the observed data.&lt;br /&gt;
&lt;br /&gt;
We now start on the process of developing mathematical ways of formulating inferences and this requires the use of &amp;#039;&amp;#039;probability&amp;#039;&amp;#039;. This becomes clear if we think back to one of the early questions posed in this course: &amp;#039;&amp;#039;prior to sampling is it possible to predict with absolute certainty what will be observed&amp;#039;&amp;#039;? The answer to this question is &amp;#039;&amp;#039;no&amp;#039;&amp;#039;; although it would be of interest to know how &amp;#039;&amp;#039;likely&amp;#039;&amp;#039; it is that certain values would be observed. Or, what is the &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; of observing certain values?&lt;br /&gt;
&lt;br /&gt;
Before proceeding, we need some more tools:&lt;br /&gt;
&lt;br /&gt;
= Venn diagrams =&lt;br /&gt;
&lt;br /&gt;
Venn diagrams (and diagrams in general) are of enormous help in trying to understand, and manipulate probability. We begin with some basic definitions, some of which we have encountered before.&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Experiment:&amp;#039;&amp;#039;&amp;#039; any process which, when applied, provides data or an outcome; e.g., rolling a die and observing the number of dots on the upturned face; recording the amount of rainfall in Manchester over a period of time.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Sample Space:&amp;#039;&amp;#039;&amp;#039; set of possible outcomes of an experiment; e.g., &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; (or &amp;lt;math&amp;gt;\Omega &amp;lt;/math&amp;gt;) &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{1,2,3,4,5,6\}&amp;lt;/math&amp;gt;, which is the sample space of rolling a dice. Or &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{x;x\geq 0\}&amp;lt;/math&amp;gt;, which is the sample space of an experiment where the outcomes can be any real non-negative number, or ‘&amp;#039;&amp;#039;the set of real non-negative real numbers&amp;#039;&amp;#039;’.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Simple Event&amp;#039;&amp;#039;&amp;#039;: just one of the possible outcomes on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Event:&amp;#039;&amp;#039;&amp;#039; a &amp;#039;&amp;#039;subset&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, denoted &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;; e.g., &amp;lt;math&amp;gt;E=\left\{ 2,4,6\right\}&amp;lt;/math&amp;gt; (i.e. any even number on a dice) or &amp;lt;math&amp;gt;E=\left\{ x;4&amp;lt;x\leq 10\right\} ,&amp;lt;/math&amp;gt; which means ‘&amp;#039;&amp;#039;the set of real numbers which are strictly bigger than&amp;#039;&amp;#039; &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;but less than or equal to &amp;#039;&amp;#039;&amp;lt;math&amp;gt;10&amp;lt;/math&amp;gt;’.&amp;lt;br /&amp;gt;Note that an event, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, is a collection of simple events.&lt;br /&gt;
&lt;br /&gt;
Such concepts can be represented by means of the following Venn Diagram:&lt;br /&gt;
&lt;br /&gt;
[[File:Venn_1.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
The sample space, &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; is depicted as a closed rectangle, and the event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; is a closed loop wholly contained within &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; and we write (in set notation) &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
In dealing with probability, and in particular the probability of an event (or events) occurring, we shall need to be familiar with &amp;#039;&amp;#039;&amp;#039;UNIONS, INTERSECTIONS&amp;#039;&amp;#039;&amp;#039; and &amp;#039;&amp;#039;&amp;#039;COMPLEMENTS&amp;#039;&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
To illustrate these concepts, consider the sample space &amp;lt;math&amp;gt;S=\{x;x\geq 0\},\,&amp;lt;/math&amp;gt; with the following events defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as depicted in Figure 3.2:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E=\{x;4&amp;lt;x\leq 10\},\,F=\{x;7&amp;lt;x\leq 17\},\,G=\{x;x&amp;gt;15\},\,H=\{x;9&amp;lt;x\leq&lt;br /&gt;
13\}. &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| (a) Event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: A closed loop&lt;br /&gt;
| (b) Union: &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2a.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2b.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (c) Intersection: &amp;lt;math&amp;gt;E\cap F&amp;lt;/math&amp;gt;&lt;br /&gt;
| (d) The Null set/event: &amp;lt;math&amp;gt;E\cap G=\emptyset &amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2c.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2d.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (e) Complement of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;&lt;br /&gt;
| (f) Subset of &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;H\subset F&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;H\cap F=H&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2e.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2f.jpg|frameless|300px]]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
        &lt;br /&gt;
&lt;br /&gt;
* The &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cup F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cup F=\{x;4&amp;lt;x\leq 17\};&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are either in &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or in &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) in both. This is illustrated on the Venn diagram by the dark shaded area in diagram (b).&lt;br /&gt;
* The&amp;#039;&amp;#039; intersection&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cap F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cap F=\left\{ x;7\leq x\leq 10\right\} ;&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are common to both &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; Again this is depicted by the dark shaded area in (c). If events have no elements in common (as, for example, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt;) then they are said to be &amp;#039;&amp;#039;mutually exclusive&amp;#039;&amp;#039;, and we can write &amp;lt;math&amp;gt;E\cap G=\emptyset ,&amp;lt;/math&amp;gt; meaning the &amp;#039;&amp;#039;null set&amp;#039;&amp;#039; which contains no elements. Such a situation is illustrated on the Venn Diagram by events (the two shaded closed loops in (d)) which do not overlap. Notice however that &amp;lt;math&amp;gt;G\cap F\neq \emptyset ,&amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have elements in common.&lt;br /&gt;
* The &amp;#039;&amp;#039;complement&amp;#039;&amp;#039; of an event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; say, is everything defined on the sample space which is not in &amp;lt;math&amp;gt;E.&amp;lt;/math&amp;gt; This event is denoted &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;, the dark shaded area in (e); here &amp;lt;math&amp;gt;\bar{E}=\left\{ x;x\leq 4\right\} \cup \left\{ x;x&amp;gt;10\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
* Finally note that &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; is a sub-set of &amp;lt;math&amp;gt;F;&amp;lt;/math&amp;gt; see (f). It is depicted as the dark closed loop wholly contained within &amp;lt;math&amp;gt;F,&amp;lt;/math&amp;gt; the lighter shaded area, so that &amp;lt;math&amp;gt;H\cap F=H;&amp;lt;/math&amp;gt; if an element in the sample space is a member of &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; then it must also be member of &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; (In mathematical logic, we employ this scenario to indicate that “&amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; implies &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;”, but not necessarily vice-versa.) Notice that &amp;lt;math&amp;gt;G\cap H=\emptyset &amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;H\cap E\neq \emptyset&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
= Probability =&lt;br /&gt;
&lt;br /&gt;
The term &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; (or some equivalent) is used in everyday conversation and so can not be unfamiliar to the reader. We talk of the probability, or chance, of rain; the likelihood of England winning the World Cup; or, perhaps more scientifically, the chance of getting a &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; when rolling a die. What we shall now do is develop a coherent theory of probability; a theory which allows us to combine and manipulate probabilities in a consistent and meaningful manner. We shall describe ways of dealing with, and describing, uncertainty. This will involve &amp;#039;&amp;#039;rules&amp;#039;&amp;#039; which govern our use of terms like probability.&lt;br /&gt;
&lt;br /&gt;
There have been a number of different approaches (interpretations) of probability. Most depend, at least to some extent, on the notion of relative frequency as now described:&lt;br /&gt;
&lt;br /&gt;
* Suppose an experiment has an outcome of interest &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;. The &amp;#039;&amp;#039;relative frequency interpretation&amp;#039;&amp;#039; of probability says that assuming the experiment can be repeated a large number of times then the relative frequency of observing the outcome &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; will settle down to a &amp;#039;&amp;#039;number&amp;#039;&amp;#039;, denoted &amp;lt;math&amp;gt;\Pr (E),&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;P(E)&amp;lt;/math&amp;gt; or Prob&amp;lt;math&amp;gt;(E),&amp;lt;/math&amp;gt; called the &amp;#039;&amp;#039;&amp;#039;probability&amp;#039;&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
This is illustrated in the next Figure where the proportion of heads obtained after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; flips of a fair coin is plotted against &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;, as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; increases; e.g., of the first &amp;lt;math&amp;gt;100&amp;lt;/math&amp;gt; flips, &amp;lt;math&amp;gt;55&amp;lt;/math&amp;gt; were heads (&amp;lt;math&amp;gt;55\%&amp;lt;/math&amp;gt;). Notice that the plot becomes less ‘wobbly’ after about &amp;lt;math&amp;gt;n=220&amp;lt;/math&amp;gt; and appears to be settling down to the value of &amp;lt;math&amp;gt;\frac{1}{2}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
[[File:Prob_coin.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
Due to this interpretation of probability, we often use observed sample proportions to approximate underlying probabilities of interest; see, for example, Question 4 of Exercise 2. There are, of course, other interpretations of probability; e.g., the subjective interpretation which simply expresses the strength of one’s belief about an event of interest such as whether Manchester United will win the European Cup! Any one of these interpretations can be used in practical situations provided the implied notion of probability follows a simple set of &amp;#039;&amp;#039;axioms&amp;#039;&amp;#039; or &amp;#039;&amp;#039;rules&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
== The axioms of probability ==&lt;br /&gt;
&lt;br /&gt;
There are just &amp;#039;&amp;#039;three &amp;#039;&amp;#039;basic rules that must be obeyed when dealing with probabilities:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;For any event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; i.e., &amp;lt;math&amp;gt;E\subset S,\,\,\Pr (E)\geq 0&amp;lt;/math&amp;gt;; &amp;#039;&amp;#039;probabilities are non-negative&amp;#039;&amp;#039;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (S)=1;&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;having defined the sample space of outcomes, one of these outcomes must be observed&amp;#039;&amp;#039;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;If events &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are mutually exclusive defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, so that &amp;lt;math&amp;gt;E\cap F=\emptyset &amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\Pr \left( E\right)+\Pr \left( F\right) .&amp;lt;/math&amp;gt; In general, for any set of mutually exclusive events, &amp;lt;math&amp;gt;E_{1},E_{2},\ldots ,E_{k},&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S:&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (E_{1}\cup E_{2}\cup \ldots \cup E_{k})=\Pr (E_{1})+\Pr (E_{2})+\ldots\Pr (E_{k})&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;i.e., &amp;lt;math&amp;gt;\Pr \left( \bigcup_{j=1}^{k}E_{j}\right) =\sum_{j=1}^{k}\Pr (E_{j}).&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In terms of the Venn Diagram, one can (and should) usefully think of the area of &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; relative to that of &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as providing an indication of probability. (Note, from axiom 2, that the area of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is implicitly normalised to be unity).&lt;br /&gt;
&lt;br /&gt;
Also observe that, contrary to what you may have believed, it is not one of the rules that &amp;lt;math&amp;gt;\Pr (E)\leq 1&amp;lt;/math&amp;gt; for any event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;. Rather, this is an implication of the &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; rules given:&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;implications: &amp;#039;&amp;#039;&amp;#039;it must be that for any event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;E\cap \bar{E}=\emptyset &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;E\cup \bar{E}=S.&amp;lt;/math&amp;gt; By Axiom &amp;lt;math&amp;gt;1,&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr (E)\geq 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr \left( \bar{E}\right) \geq 0&amp;lt;/math&amp;gt; and by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr(E)+\Pr (\bar{E})=\Pr (S).&amp;lt;/math&amp;gt; So &amp;lt;math&amp;gt;\Pr \left( E\right) +\Pr \left( \bar{E}\right) =1,&amp;lt;/math&amp;gt; by Axiom &amp;lt;math&amp;gt;2.&amp;lt;/math&amp;gt; This implies that&lt;br /&gt;
&lt;br /&gt;
# &amp;lt;math&amp;gt;0\leq \Pr (E)\leq 1&amp;lt;/math&amp;gt;&lt;br /&gt;
# &amp;lt;math&amp;gt;\Pr (\bar{E})=1-\Pr (E)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The first of these is what we might have expected from probability (a number lying between &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;). The second implication is also very important; it says that the probability of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; not happening is ‘&amp;#039;&amp;#039;one minus the probability of it happening&amp;#039;&amp;#039;’. Thus when rolling a die, the probability of getting &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; is one minus the probability of getting either a &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;2&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;5.&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
These axioms imply how to calculate probabilities on a sample space of equally likely outcomes. For example, and as we have already noted, the experiment of rolling a fair die defines a sample space of six, mutually exclusive and equally likely outcomes (&amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; dots on the up-turned face). The axioms then say that each of the six probabilities are positive, add to 1 and are all the same. Thus, the probability of any one of the outcomes must be simply &amp;lt;math&amp;gt;\frac{1}{6};&amp;lt;/math&amp;gt; which may accord with your intuition. A similar sort of analysis reveals that the probability of drawing a club from a deck of &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; cards is &amp;lt;math&amp;gt;\frac{13}{52},&amp;lt;/math&amp;gt; since any one of the &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; cards has an equal chance of being drawn and &amp;lt;math&amp;gt;13&amp;lt;/math&amp;gt; of them are clubs; i.e., &amp;lt;math&amp;gt;13&amp;lt;/math&amp;gt; of the &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; are clubs, so the probability of drawing a club is &amp;lt;math&amp;gt;\frac{13}{52}.&amp;lt;/math&amp;gt; Notice the importance of the assumption of equally likely outcomes here.&lt;br /&gt;
&lt;br /&gt;
In this, and the next section of notes, we shall see how these axioms can be used. Firstly, consider the construction of a probability for the &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of two events; i.e., the probability that &amp;#039;&amp;#039;either &amp;#039;&amp;#039;&amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) &amp;#039;&amp;#039;both &amp;#039;&amp;#039;will occur. Such a probability is embodied in the &amp;#039;&amp;#039;addition rule of probability&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
== The addition rule of probability ==&lt;br /&gt;
&lt;br /&gt;
When rolling a fair die, let &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; denote the event of an “odd number of dots” and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; the event of the “number of dots being greater than, or equal, to &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;”&amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt; What is the probability of the event &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;? To calculate this we can collect together all the mutually exclusive (simple) events which comprise &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;, and then add up the probabilities (by axiom 3). These simple events are &amp;lt;math&amp;gt;1,3,4,5&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; dots. Each has a probability of &amp;lt;math&amp;gt;\frac{1}{6},&amp;lt;/math&amp;gt; so the required total probability is: &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\frac{5}{6}&amp;lt;/math&amp;gt;. Consider carefully how this probability is constructed and note, in particular, that &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) \neq \Pr \left( E\right) +\Pr \left( F\right) &amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have a simple event in common (namely &amp;lt;math&amp;gt;5&amp;lt;/math&amp;gt; dots).&lt;br /&gt;
&lt;br /&gt;
In general, we can calculate the probability of the union of events using the &amp;#039;&amp;#039;addition rule of probability&amp;#039;&amp;#039;, as follows.&lt;br /&gt;
&lt;br /&gt;
* For any events, &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F\subset S:\Pr (E\cup F)=\Pr (E)+\Pr (F)-\Pr (E\cap F).&amp;lt;/math&amp;gt; So, in general, &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) \leq \Pr (E)+\Pr (F).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This generalises to three events, &amp;lt;math&amp;gt;E_{1},E_{2}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;E_{3}&amp;lt;/math&amp;gt; as&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E_{1}\cup E_{2}\cup E_{3}) &amp;amp;=&amp;amp;\Pr (E_{1})+\Pr (E_{2})+\Pr (E_{3}) \\&lt;br /&gt;
&amp;amp;&amp;amp;-\Pr (E_{1}\cap E_{2})-\Pr (E_{1}\cap E_{3})-\Pr (E_{2}\cap E_{3}) \\&lt;br /&gt;
&amp;amp;&amp;amp;+\Pr (E_{1}\cap E_{2}\cap E_{3}).\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We can demonstrate this as follows.&lt;br /&gt;
&lt;br /&gt;
Note that&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E\cup F=\left( E\cap \bar{F}\right) \cup \left( E\cap F\right) \cup \left(\bar{E}\cap F\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
the union of &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; mutually exclusive events. These mutually exclusive events are depicted by the shaded areas &amp;lt;math&amp;gt;\mathbf{a,}&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\mathbf{b}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{c}&amp;lt;/math&amp;gt;, respectively, in the next Figure.&lt;br /&gt;
&lt;br /&gt;
[[File:Prob_add.jpg|frameless|500px]]&lt;br /&gt;
&lt;br /&gt;
Then by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, and from the fact that the three events &amp;lt;math&amp;gt;\left( E\cap\bar{F}\right) &amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\left( E\cap F\right) &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\left( \bar{E}\cap F\right)&amp;lt;/math&amp;gt; are mutually exclusive so that the “area” occupied by &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt; is simply &amp;lt;math&amp;gt;\mathbf{a+b+c,}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\Pr \left( E\cap \bar{F}\right) +\Pr \left( \bar{E}\cap F\right) +\Pr \left( E\cap F\right) .&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
But also by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, since &amp;lt;math&amp;gt;E=\left( E\cap \bar{F}\right) \cup \left(E\cap F\right) &amp;lt;/math&amp;gt;, it must be that &amp;lt;math&amp;gt;\Pr (E)=\Pr \left( E\cap \bar{F}\right)+\Pr (E\cap F);&amp;lt;/math&amp;gt; similarly, &amp;lt;math&amp;gt;\Pr \left( \bar{E}\cap F\right) =\Pr \left(F\right) -\Pr \left( E\cap F\right)&amp;lt;/math&amp;gt;. Putting all of this together gives&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\Pr (E\cup F)=\Pr (E)+\Pr (F)-\Pr (E\cap F).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
When &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are mutually exclusive, so that &amp;lt;math&amp;gt;E\cap F=\emptyset&amp;lt;/math&amp;gt;, this rule reveals Axiom 2: &amp;lt;math&amp;gt;\Pr (E\cup F)=\Pr (E)+\Pr (F).\smallskip &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example: &amp;#039;&amp;#039;What is the probability of drawing a Queen (&amp;lt;math&amp;gt;Q &amp;lt;/math&amp;gt;) or a Club (&amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;) in a single draw from a pack of cards? Now, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;52 &amp;lt;/math&amp;gt; cards are Queens, so &amp;lt;math&amp;gt;\Pr \left( Q\right) =\frac{4}{52},&amp;lt;/math&amp;gt; whilst &amp;lt;math&amp;gt;\Pr\left( C\right) =\frac{13}{52}.&amp;lt;/math&amp;gt; The probability of drawing the Queen of Clubs is simply &amp;lt;math&amp;gt;\frac{1}{52};&amp;lt;/math&amp;gt; i.e., &amp;lt;math&amp;gt;\Pr \left( Q\cap C\right) =\frac{1}{52}&amp;lt;/math&amp;gt;. What we require is a Club or a Queen, for which the probability is&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr \left( Q\cup C\right) &amp;amp;=&amp;amp;\Pr \left( Q\right) +\Pr \left( C\right) -\Pr\left( Q\cap C\right) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{4}{52}+\frac{13}{52}-\frac{1}{52} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{16}{52}=\frac{4}{13}.\end{aligned}&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example: &amp;#039;&amp;#039;Consider a car journey from Manchester to London via the M6 and M1. Let &amp;lt;math&amp;gt;E=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;heavy traffic somewhere on route&amp;#039;&amp;#039; and &amp;lt;math&amp;gt;F=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;roadworks somewhere on route&amp;#039;&amp;#039;. It is estimated that &amp;lt;math&amp;gt;\Pr (E)=0.8&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr (F)=0.4,&amp;lt;/math&amp;gt; whilst the probability of NOT encountering both is &amp;lt;math&amp;gt;\Pr (\overline{E\cap F})=0.6.&amp;lt;/math&amp;gt; What is the probability of encountering heavytraffic or roadworks?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;We require &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) .&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E\cup F) &amp;amp;=&amp;amp;\Pr (E)+\Pr (F)-\Pr (E\cap F) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\Pr (E)+\Pr (F)-(1-\Pr (\overline{E\cap F})) \\&lt;br /&gt;
&amp;amp;=&amp;amp;0.8+0.4-1+0.6 \\&lt;br /&gt;
&amp;amp;=&amp;amp;0.8=\Pr (E)\end{aligned}&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Notice that this implies, in this case, that &amp;lt;math&amp;gt;F\subset E&amp;lt;/math&amp;gt; (why?). This &amp;#039;&amp;#039;model &amp;#039;&amp;#039;then implies that when there are roadworks somewhere on route you are bound to encounter heavy traffic; on the other hand, you can encounter heavy traffic on route without ever passing through roadworks. (My own experience of this motorway inclines me towards this implication!)&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Similar concepts apply when manipulating proportions as follows:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example&amp;#039;&amp;#039;&amp;lt;math&amp;gt;:&amp;lt;/math&amp;gt; A sample of &amp;lt;math&amp;gt;1000&amp;lt;/math&amp;gt; undergraduates were asked whether they took either Mathematics, Physics or Chemistry at A-level. The following responses were obtained: &amp;lt;math&amp;gt;100&amp;lt;/math&amp;gt; just took Mathematics; &amp;lt;math&amp;gt;70&amp;lt;/math&amp;gt; just took Physics; &amp;lt;math&amp;gt;100&amp;lt;/math&amp;gt; just took Chemistry; &amp;lt;math&amp;gt;150&amp;lt;/math&amp;gt; took Mathematics and Physics, but not Chemistry; &amp;lt;math&amp;gt;40&amp;lt;/math&amp;gt; took Mathematics and Chemistry, but not Physics; and, &amp;lt;math&amp;gt;240&amp;lt;/math&amp;gt; took Physics and Chemistry, but not Mathematics. What proportion took all three?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;This can be addressed with the following diagram:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;[[File:Prob_Alevels.jpg|frameless|500px]]&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;The shaded area contains the number who took all three, which can be deduced from the above information (since the total of the numbers assigned to each part of the Venn diagram must be &amp;lt;math&amp;gt;1000&amp;lt;/math&amp;gt;). The answer is therefore &amp;lt;math&amp;gt;30\%&amp;lt;/math&amp;gt; (being &amp;lt;math&amp;gt;300&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;1000&amp;lt;/math&amp;gt;).&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Two further results on unions, intersections and complements which are of use (and which are fairly easy to demonstrate using Venn diagrams) are &amp;#039;&amp;#039;&amp;#039;de Morgan Laws&amp;#039;&amp;#039;&amp;#039;:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\left( \bar{A}\cap \bar{B}\right) =\left( \overline{A\cup B}\right) &amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\bar{A}\cup \bar{B}=\left( \overline{A\cap B}\right) &amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2856</id>
		<title>Probability Intro</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2856"/>
				<updated>2013-08-07T21:11:24Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: /* The addition rule of probability */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Introducing Probability =&lt;br /&gt;
&lt;br /&gt;
So far we have been looking at ways of summarising samples of data drawn from an underlying population of interest. Although at times tedious, all such arithmetic calculations are fairly mechanical and straightforward to apply. To remind ourselves, one of the primary reasons for wishing to summarise data is so assist in the development of inferences about the population from which the data were taken. That is to say, we would like to elicit some information about the mechanism which generated the observed data.&lt;br /&gt;
&lt;br /&gt;
We now start on the process of developing mathematical ways of formulating inferences and this requires the use of &amp;#039;&amp;#039;probability&amp;#039;&amp;#039;. This becomes clear if we think back to one of the early questions posed in this course: &amp;#039;&amp;#039;prior to sampling is it possible to predict with absolute certainty what will be observed&amp;#039;&amp;#039;? The answer to this question is &amp;#039;&amp;#039;no&amp;#039;&amp;#039;; although it would be of interest to know how &amp;#039;&amp;#039;likely&amp;#039;&amp;#039; it is that certain values would be observed. Or, what is the &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; of observing certain values?&lt;br /&gt;
&lt;br /&gt;
Before proceeding, we need some more tools:&lt;br /&gt;
&lt;br /&gt;
= Venn diagrams =&lt;br /&gt;
&lt;br /&gt;
Venn diagrams (and diagrams in general) are of enormous help in trying to understand, and manipulate probability. We begin with some basic definitions, some of which we have encountered before.&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Experiment:&amp;#039;&amp;#039;&amp;#039; any process which, when applied, provides data or an outcome; e.g., rolling a die and observing the number of dots on the upturned face; recording the amount of rainfall in Manchester over a period of time.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Sample Space:&amp;#039;&amp;#039;&amp;#039; set of possible outcomes of an experiment; e.g., &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; (or &amp;lt;math&amp;gt;\Omega &amp;lt;/math&amp;gt;) &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{1,2,3,4,5,6\}&amp;lt;/math&amp;gt;, which is the sample space of rolling a dice. Or &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{x;x\geq 0\}&amp;lt;/math&amp;gt;, which is the sample space of an experiment where the outcomes can be any real non-negative number, or ‘&amp;#039;&amp;#039;the set of real non-negative real numbers&amp;#039;&amp;#039;’.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Simple Event&amp;#039;&amp;#039;&amp;#039;: just one of the possible outcomes on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Event:&amp;#039;&amp;#039;&amp;#039; a &amp;#039;&amp;#039;subset&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, denoted &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;; e.g., &amp;lt;math&amp;gt;E=\left\{ 2,4,6\right\}&amp;lt;/math&amp;gt; (i.e. any even number on a dice) or &amp;lt;math&amp;gt;E=\left\{ x;4&amp;lt;x\leq 10\right\} ,&amp;lt;/math&amp;gt; which means ‘&amp;#039;&amp;#039;the set of real numbers which are strictly bigger than&amp;#039;&amp;#039; &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;but less than or equal to &amp;#039;&amp;#039;&amp;lt;math&amp;gt;10&amp;lt;/math&amp;gt;’.&amp;lt;br /&amp;gt;Note that an event, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, is a collection of simple events.&lt;br /&gt;
&lt;br /&gt;
Such concepts can be represented by means of the following Venn Diagram:&lt;br /&gt;
&lt;br /&gt;
[[File:Venn_1.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
The sample space, &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; is depicted as a closed rectangle, and the event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; is a closed loop wholly contained within &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; and we write (in set notation) &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
In dealing with probability, and in particular the probability of an event (or events) occurring, we shall need to be familiar with &amp;#039;&amp;#039;&amp;#039;UNIONS, INTERSECTIONS&amp;#039;&amp;#039;&amp;#039; and &amp;#039;&amp;#039;&amp;#039;COMPLEMENTS&amp;#039;&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
To illustrate these concepts, consider the sample space &amp;lt;math&amp;gt;S=\{x;x\geq 0\},\,&amp;lt;/math&amp;gt; with the following events defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as depicted in Figure 3.2:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E=\{x;4&amp;lt;x\leq 10\},\,F=\{x;7&amp;lt;x\leq 17\},\,G=\{x;x&amp;gt;15\},\,H=\{x;9&amp;lt;x\leq&lt;br /&gt;
13\}. &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| (a) Event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: A closed loop&lt;br /&gt;
| (b) Union: &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2a.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2b.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (c) Intersection: &amp;lt;math&amp;gt;E\cap F&amp;lt;/math&amp;gt;&lt;br /&gt;
| (d) The Null set/event: &amp;lt;math&amp;gt;E\cap G=\emptyset &amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2c.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2d.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (e) Complement of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;&lt;br /&gt;
| (f) Subset of &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;H\subset F&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;H\cap F=H&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2e.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2f.jpg|frameless|300px]]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
        &lt;br /&gt;
&lt;br /&gt;
* The &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cup F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cup F=\{x;4&amp;lt;x\leq 17\};&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are either in &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or in &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) in both. This is illustrated on the Venn diagram by the dark shaded area in diagram (b).&lt;br /&gt;
* The&amp;#039;&amp;#039; intersection&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cap F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cap F=\left\{ x;7\leq x\leq 10\right\} ;&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are common to both &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; Again this is depicted by the dark shaded area in (c). If events have no elements in common (as, for example, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt;) then they are said to be &amp;#039;&amp;#039;mutually exclusive&amp;#039;&amp;#039;, and we can write &amp;lt;math&amp;gt;E\cap G=\emptyset ,&amp;lt;/math&amp;gt; meaning the &amp;#039;&amp;#039;null set&amp;#039;&amp;#039; which contains no elements. Such a situation is illustrated on the Venn Diagram by events (the two shaded closed loops in (d)) which do not overlap. Notice however that &amp;lt;math&amp;gt;G\cap F\neq \emptyset ,&amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have elements in common.&lt;br /&gt;
* The &amp;#039;&amp;#039;complement&amp;#039;&amp;#039; of an event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; say, is everything defined on the sample space which is not in &amp;lt;math&amp;gt;E.&amp;lt;/math&amp;gt; This event is denoted &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;, the dark shaded area in (e); here &amp;lt;math&amp;gt;\bar{E}=\left\{ x;x\leq 4\right\} \cup \left\{ x;x&amp;gt;10\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
* Finally note that &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; is a sub-set of &amp;lt;math&amp;gt;F;&amp;lt;/math&amp;gt; see (f). It is depicted as the dark closed loop wholly contained within &amp;lt;math&amp;gt;F,&amp;lt;/math&amp;gt; the lighter shaded area, so that &amp;lt;math&amp;gt;H\cap F=H;&amp;lt;/math&amp;gt; if an element in the sample space is a member of &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; then it must also be member of &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; (In mathematical logic, we employ this scenario to indicate that “&amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; implies &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;”, but not necessarily vice-versa.) Notice that &amp;lt;math&amp;gt;G\cap H=\emptyset &amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;H\cap E\neq \emptyset&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
= Probability =&lt;br /&gt;
&lt;br /&gt;
The term &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; (or some equivalent) is used in everyday conversation and so can not be unfamiliar to the reader. We talk of the probability, or chance, of rain; the likelihood of England winning the World Cup; or, perhaps more scientifically, the chance of getting a &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; when rolling a die. What we shall now do is develop a coherent theory of probability; a theory which allows us to combine and manipulate probabilities in a consistent and meaningful manner. We shall describe ways of dealing with, and describing, uncertainty. This will involve &amp;#039;&amp;#039;rules&amp;#039;&amp;#039; which govern our use of terms like probability.&lt;br /&gt;
&lt;br /&gt;
There have been a number of different approaches (interpretations) of probability. Most depend, at least to some extent, on the notion of relative frequency as now described:&lt;br /&gt;
&lt;br /&gt;
* Suppose an experiment has an outcome of interest &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;. The &amp;#039;&amp;#039;relative frequency interpretation&amp;#039;&amp;#039; of probability says that assuming the experiment can be repeated a large number of times then the relative frequency of observing the outcome &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; will settle down to a &amp;#039;&amp;#039;number&amp;#039;&amp;#039;, denoted &amp;lt;math&amp;gt;\Pr (E),&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;P(E)&amp;lt;/math&amp;gt; or Prob&amp;lt;math&amp;gt;(E),&amp;lt;/math&amp;gt; called the &amp;#039;&amp;#039;&amp;#039;probability&amp;#039;&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
This is illustrated in the next Figure where the proportion of heads obtained after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; flips of a fair coin is plotted against &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;, as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; increases; e.g., of the first &amp;lt;math&amp;gt;100&amp;lt;/math&amp;gt; flips, &amp;lt;math&amp;gt;55&amp;lt;/math&amp;gt; were heads (&amp;lt;math&amp;gt;55\%&amp;lt;/math&amp;gt;). Notice that the plot becomes less ‘wobbly’ after about &amp;lt;math&amp;gt;n=220&amp;lt;/math&amp;gt; and appears to be settling down to the value of &amp;lt;math&amp;gt;\frac{1}{2}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
[[File:Prob_coin.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
Due to this interpretation of probability, we often use observed sample proportions to approximate underlying probabilities of interest; see, for example, Question 4 of Exercise 2. There are, of course, other interpretations of probability; e.g., the subjective interpretation which simply expresses the strength of one’s belief about an event of interest such as whether Manchester United will win the European Cup! Any one of these interpretations can be used in practical situations provided the implied notion of probability follows a simple set of &amp;#039;&amp;#039;axioms&amp;#039;&amp;#039; or &amp;#039;&amp;#039;rules&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
== The axioms of probability ==&lt;br /&gt;
&lt;br /&gt;
There are just &amp;#039;&amp;#039;three &amp;#039;&amp;#039;basic rules that must be obeyed when dealing with probabilities:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;For any event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; i.e., &amp;lt;math&amp;gt;E\subset S,\,\,\Pr (E)\geq 0&amp;lt;/math&amp;gt;; &amp;#039;&amp;#039;probabilities are non-negative&amp;#039;&amp;#039;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (S)=1;&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;having defined the sample space of outcomes, one of these outcomes must be observed&amp;#039;&amp;#039;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;If events &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are mutually exclusive defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, so that &amp;lt;math&amp;gt;E\cap F=\emptyset &amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\Pr \left( E\right)+\Pr \left( F\right) .&amp;lt;/math&amp;gt; In general, for any set of mutually exclusive events, &amp;lt;math&amp;gt;E_{1},E_{2},\ldots ,E_{k},&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S:&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (E_{1}\cup E_{2}\cup \ldots \cup E_{k})=\Pr (E_{1})+\Pr (E_{2})+\ldots\Pr (E_{k})&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;i.e., &amp;lt;math&amp;gt;\Pr \left( \bigcup_{j=1}^{k}E_{j}\right) =\sum_{j=1}^{k}\Pr (E_{j}).&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In terms of the Venn Diagram, one can (and should) usefully think of the area of &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; relative to that of &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as providing an indication of probability. (Note, from axiom 2, that the area of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is implicitly normalised to be unity).&lt;br /&gt;
&lt;br /&gt;
Also observe that, contrary to what you may have believed, it is not one of the rules that &amp;lt;math&amp;gt;\Pr (E)\leq 1&amp;lt;/math&amp;gt; for any event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;. Rather, this is an implication of the &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; rules given:&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;implications: &amp;#039;&amp;#039;&amp;#039;it must be that for any event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;E\cap \bar{E}=\emptyset &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;E\cup \bar{E}=S.&amp;lt;/math&amp;gt; By Axiom &amp;lt;math&amp;gt;1,&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr (E)\geq 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr \left( \bar{E}\right) \geq 0&amp;lt;/math&amp;gt; and by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr(E)+\Pr (\bar{E})=\Pr (S).&amp;lt;/math&amp;gt; So &amp;lt;math&amp;gt;\Pr \left( E\right) +\Pr \left( \bar{E}\right) =1,&amp;lt;/math&amp;gt; by Axiom &amp;lt;math&amp;gt;2.&amp;lt;/math&amp;gt; This implies that&lt;br /&gt;
&lt;br /&gt;
# &amp;lt;math&amp;gt;0\leq \Pr (E)\leq 1&amp;lt;/math&amp;gt;&lt;br /&gt;
# &amp;lt;math&amp;gt;\Pr (\bar{E})=1-\Pr (E)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The first of these is what we might have expected from probability (a number lying between &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;). The second implication is also very important; it says that the probability of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; not happening is ‘&amp;#039;&amp;#039;one minus the probability of it happening&amp;#039;&amp;#039;’. Thus when rolling a die, the probability of getting &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; is one minus the probability of getting either a &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;2&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;5.&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
These axioms imply how to calculate probabilities on a sample space of equally likely outcomes. For example, and as we have already noted, the experiment of rolling a fair die defines a sample space of six, mutually exclusive and equally likely outcomes (&amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; dots on the up-turned face). The axioms then say that each of the six probabilities are positive, add to 1 and are all the same. Thus, the probability of any one of the outcomes must be simply &amp;lt;math&amp;gt;\frac{1}{6};&amp;lt;/math&amp;gt; which may accord with your intuition. A similar sort of analysis reveals that the probability of drawing a club from a deck of &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; cards is &amp;lt;math&amp;gt;\frac{13}{52},&amp;lt;/math&amp;gt; since any one of the &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; cards has an equal chance of being drawn and &amp;lt;math&amp;gt;13&amp;lt;/math&amp;gt; of them are clubs; i.e., &amp;lt;math&amp;gt;13&amp;lt;/math&amp;gt; of the &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; are clubs, so the probability of drawing a club is &amp;lt;math&amp;gt;\frac{13}{52}.&amp;lt;/math&amp;gt; Notice the importance of the assumption of equally likely outcomes here.&lt;br /&gt;
&lt;br /&gt;
In this, and the next section of notes, we shall see how these axioms can be used. Firstly, consider the construction of a probability for the &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of two events; i.e., the probability that &amp;#039;&amp;#039;either &amp;#039;&amp;#039;&amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) &amp;#039;&amp;#039;both &amp;#039;&amp;#039;will occur. Such a probability is embodied in the &amp;#039;&amp;#039;addition rule of probability&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
== The addition rule of probability ==&lt;br /&gt;
&lt;br /&gt;
When rolling a fair die, let &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; denote the event of an “odd number of dots” and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; the event of the “number of dots being greater than, or equal, to &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;”&amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt; What is the probability of the event &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;? To calculate this we can collect together all the mutually exclusive (simple) events which comprise &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;, and then add up the probabilities (by axiom 3). These simple events are &amp;lt;math&amp;gt;1,3,4,5&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; dots. Each has a probability of &amp;lt;math&amp;gt;\frac{1}{6},&amp;lt;/math&amp;gt; so the required total probability is: &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\frac{5}{6}&amp;lt;/math&amp;gt;. Consider carefully how this probability is constructed and note, in particular, that &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) \neq \Pr \left( E\right) +\Pr \left( F\right) &amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have a simple event in common (namely &amp;lt;math&amp;gt;5&amp;lt;/math&amp;gt; dots).&lt;br /&gt;
&lt;br /&gt;
In general, we can calculate the probability of the union of events using the &amp;#039;&amp;#039;addition rule of probability&amp;#039;&amp;#039;, as follows.&lt;br /&gt;
&lt;br /&gt;
* For any events, &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F\subset S:\Pr (E\cup F)=\Pr (E)+\Pr (F)-\Pr (E\cap F).&amp;lt;/math&amp;gt; So, in general, &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) \leq \Pr (E)+\Pr (F).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This generalises to three events, &amp;lt;math&amp;gt;E_{1},E_{2}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;E_{3}&amp;lt;/math&amp;gt; as&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E_{1}\cup E_{2}\cup E_{3}) &amp;amp;=&amp;amp;\Pr (E_{1})+\Pr (E_{2})+\Pr (E_{3}) \\&lt;br /&gt;
&amp;amp;&amp;amp;-\Pr (E_{1}\cap E_{2})-\Pr (E_{1}\cap E_{3})-\Pr (E_{2}\cap E_{3}) \\&lt;br /&gt;
&amp;amp;&amp;amp;+\Pr (E_{1}\cap E_{2}\cap E_{3}).\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We can demonstrate this as follows.&lt;br /&gt;
&lt;br /&gt;
Note that&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E\cup F=\left( E\cap \bar{F}\right) \cup \left( E\cap F\right) \cup \left(\bar{E}\cap F\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
the union of &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; mutually exclusive events. These mutually exclusive events are depicted by the shaded areas &amp;lt;math&amp;gt;\mathbf{a,}&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\mathbf{b}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{c}&amp;lt;/math&amp;gt;, respectively, in the next Figure.&lt;br /&gt;
&lt;br /&gt;
[[File:Prob_add.jpg|frameless|500px]]&lt;br /&gt;
&lt;br /&gt;
Then by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, and from the fact that the three events &amp;lt;math&amp;gt;\left( E\cap\bar{F}\right) &amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\left( E\cap F\right) &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\left( \bar{E}\cap F\right)&amp;lt;/math&amp;gt; are mutually exclusive so that the “area” occupied by &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt; is simply &amp;lt;math&amp;gt;\mathbf{a+b+c,}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\Pr \left( E\cap \bar{F}\right) +\Pr \left( \bar{E}\cap F\right) +\Pr \left( E\cap F\right) .&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
But also by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, since &amp;lt;math&amp;gt;E=\left( E\cap \bar{F}\right) \cup \left(E\cap F\right) &amp;lt;/math&amp;gt;, it must be that &amp;lt;math&amp;gt;\Pr (E)=\Pr \left( E\cap \bar{F}\right)+\Pr (E\cap F);&amp;lt;/math&amp;gt; similarly, &amp;lt;math&amp;gt;\Pr \left( \bar{E}\cap F\right) =\Pr \left(F\right) -\Pr \left( E\cap F\right)&amp;lt;/math&amp;gt;. Putting all of this together gives&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\Pr (E\cup F)=\Pr (E)+\Pr (F)-\Pr (E\cap F).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
When &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are mutually exclusive, so that &amp;lt;math&amp;gt;E\cap F=\emptyset&amp;lt;/math&amp;gt;, this rule reveals Axiom 2: &amp;lt;math&amp;gt;\Pr (E\cup F)=\Pr (E)+\Pr (F).\smallskip &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example: &amp;#039;&amp;#039;What is the probability of drawing a Queen (&amp;lt;math&amp;gt;Q &amp;lt;/math&amp;gt;) or a Club (&amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;) in a single draw from a pack of cards? Now, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;52 &amp;lt;/math&amp;gt; cards are Queens, so &amp;lt;math&amp;gt;\Pr \left( Q\right) =\frac{4}{52},&amp;lt;/math&amp;gt; whilst &amp;lt;math&amp;gt;\Pr\left( C\right) =\frac{13}{52}.&amp;lt;/math&amp;gt; The probability of drawing the Queen of Clubs is simply &amp;lt;math&amp;gt;\frac{1}{52};&amp;lt;/math&amp;gt; i.e., &amp;lt;math&amp;gt;\Pr \left( Q\cap C\right) =\frac{1}{52}&amp;lt;/math&amp;gt;. What we require is a Club or a Queen, for which the probability is&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr \left( Q\cup C\right) &amp;amp;=&amp;amp;\Pr \left( Q\right) +\Pr \left( C\right) -\Pr\left( Q\cap C\right) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{4}{52}+\frac{13}{52}-\frac{1}{52} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{16}{52}=\frac{4}{13}.\end{aligned}&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example: &amp;#039;&amp;#039;Consider a car journey from Manchester to London via the M6 and M1. Let &amp;lt;math&amp;gt;E=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;heavy traffic somewhere on route&amp;#039;&amp;#039; and &amp;lt;math&amp;gt;F=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;roadworks somewhere on route&amp;#039;&amp;#039;. It is estimated that &amp;lt;math&amp;gt;\Pr (E)=0.8&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr (F)=0.4,&amp;lt;/math&amp;gt; whilst the probability of NOT encountering both is &amp;lt;math&amp;gt;\Pr (\overline{E\cap F})=0.6.&amp;lt;/math&amp;gt; What is the probability of encountering heavytraffic or roadworks?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;We require &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) .&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E\cup F) &amp;amp;=&amp;amp;\Pr (E)+\Pr (F)-\Pr (E\cap F) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\Pr (E)+\Pr (F)-(1-\Pr (\overline{E\cap F})) \\&lt;br /&gt;
&amp;amp;=&amp;amp;0.8+0.4-1+0.6 \\&lt;br /&gt;
&amp;amp;=&amp;amp;0.8=\Pr (E)\end{aligned}&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Notice that this implies, in this case, that &amp;lt;math&amp;gt;F\subset E&amp;lt;/math&amp;gt; (why?). This &amp;#039;&amp;#039;model &amp;#039;&amp;#039;then implies that when there are roadworks somewhere on route you are bound to encounter heavy traffic; on the other hand, you can encounter heavy traffic on route without ever passing through roadworks. (My own experience of this motorway inclines me towards this implication!)&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Similar concepts apply when manipulating proportions as follows:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example&amp;#039;&amp;#039;&amp;lt;math&amp;gt;:&amp;lt;/math&amp;gt; A sample of &amp;lt;math&amp;gt;1000&amp;lt;/math&amp;gt; undergraduates were asked whether they took either Mathematics, Physics or Chemistry at A-level. The following responses were obtained: &amp;lt;math&amp;gt;100&amp;lt;/math&amp;gt; just took Mathematics; &amp;lt;math&amp;gt;70&amp;lt;/math&amp;gt; just took Physics; &amp;lt;math&amp;gt;100&amp;lt;/math&amp;gt; just took Chemistry; &amp;lt;math&amp;gt;150&amp;lt;/math&amp;gt; took Mathematics and Physics, but not Chemistry; &amp;lt;math&amp;gt;40&amp;lt;/math&amp;gt; took Mathematics and Chemistry, but not Physics; and, &amp;lt;math&amp;gt;240&amp;lt;/math&amp;gt; took Physics and Chemistry, but not Mathematics. What proportion took all three?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;This can be addressed with the following diagram:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;[[File:Prob_Alevels.jpg|frameless|600px]]&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;The shaded area contains the number who took all three, which can be deduced from the above information (since the total of the numbers assigned to each part of the Venn diagram must be &amp;lt;math&amp;gt;1000&amp;lt;/math&amp;gt;). The answer is therefore &amp;lt;math&amp;gt;30\%&amp;lt;/math&amp;gt; (being &amp;lt;math&amp;gt;300&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;1000&amp;lt;/math&amp;gt;).&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Two further results on unions, intersections and complements which are of use (and which are fairly easy to demonstrate using Venn diagrams) are &amp;#039;&amp;#039;&amp;#039;de Morgan Laws&amp;#039;&amp;#039;&amp;#039;:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\left( \bar{A}\cap \bar{B}\right) =\left( \overline{A\cup B}\right) &amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\bar{A}\cup \bar{B}=\left( \overline{A\cap B}\right) &amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=File:Prob_add.jpg&amp;diff=2855</id>
		<title>File:Prob add.jpg</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=File:Prob_add.jpg&amp;diff=2855"/>
				<updated>2013-08-07T21:10:25Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=File:Prob_Alevels.jpg&amp;diff=2854</id>
		<title>File:Prob Alevels.jpg</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=File:Prob_Alevels.jpg&amp;diff=2854"/>
				<updated>2013-08-07T21:09:58Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2853</id>
		<title>Probability Intro</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2853"/>
				<updated>2013-08-07T21:09:20Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Introducing Probability =&lt;br /&gt;
&lt;br /&gt;
So far we have been looking at ways of summarising samples of data drawn from an underlying population of interest. Although at times tedious, all such arithmetic calculations are fairly mechanical and straightforward to apply. To remind ourselves, one of the primary reasons for wishing to summarise data is so assist in the development of inferences about the population from which the data were taken. That is to say, we would like to elicit some information about the mechanism which generated the observed data.&lt;br /&gt;
&lt;br /&gt;
We now start on the process of developing mathematical ways of formulating inferences and this requires the use of &amp;#039;&amp;#039;probability&amp;#039;&amp;#039;. This becomes clear if we think back to one of the early questions posed in this course: &amp;#039;&amp;#039;prior to sampling is it possible to predict with absolute certainty what will be observed&amp;#039;&amp;#039;? The answer to this question is &amp;#039;&amp;#039;no&amp;#039;&amp;#039;; although it would be of interest to know how &amp;#039;&amp;#039;likely&amp;#039;&amp;#039; it is that certain values would be observed. Or, what is the &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; of observing certain values?&lt;br /&gt;
&lt;br /&gt;
Before proceeding, we need some more tools:&lt;br /&gt;
&lt;br /&gt;
= Venn diagrams =&lt;br /&gt;
&lt;br /&gt;
Venn diagrams (and diagrams in general) are of enormous help in trying to understand, and manipulate probability. We begin with some basic definitions, some of which we have encountered before.&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Experiment:&amp;#039;&amp;#039;&amp;#039; any process which, when applied, provides data or an outcome; e.g., rolling a die and observing the number of dots on the upturned face; recording the amount of rainfall in Manchester over a period of time.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Sample Space:&amp;#039;&amp;#039;&amp;#039; set of possible outcomes of an experiment; e.g., &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; (or &amp;lt;math&amp;gt;\Omega &amp;lt;/math&amp;gt;) &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{1,2,3,4,5,6\}&amp;lt;/math&amp;gt;, which is the sample space of rolling a dice. Or &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{x;x\geq 0\}&amp;lt;/math&amp;gt;, which is the sample space of an experiment where the outcomes can be any real non-negative number, or ‘&amp;#039;&amp;#039;the set of real non-negative real numbers&amp;#039;&amp;#039;’.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Simple Event&amp;#039;&amp;#039;&amp;#039;: just one of the possible outcomes on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Event:&amp;#039;&amp;#039;&amp;#039; a &amp;#039;&amp;#039;subset&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, denoted &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;; e.g., &amp;lt;math&amp;gt;E=\left\{ 2,4,6\right\}&amp;lt;/math&amp;gt; (i.e. any even number on a dice) or &amp;lt;math&amp;gt;E=\left\{ x;4&amp;lt;x\leq 10\right\} ,&amp;lt;/math&amp;gt; which means ‘&amp;#039;&amp;#039;the set of real numbers which are strictly bigger than&amp;#039;&amp;#039; &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;but less than or equal to &amp;#039;&amp;#039;&amp;lt;math&amp;gt;10&amp;lt;/math&amp;gt;’.&amp;lt;br /&amp;gt;Note that an event, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, is a collection of simple events.&lt;br /&gt;
&lt;br /&gt;
Such concepts can be represented by means of the following Venn Diagram:&lt;br /&gt;
&lt;br /&gt;
[[File:Venn_1.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
The sample space, &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; is depicted as a closed rectangle, and the event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; is a closed loop wholly contained within &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; and we write (in set notation) &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
In dealing with probability, and in particular the probability of an event (or events) occurring, we shall need to be familiar with &amp;#039;&amp;#039;&amp;#039;UNIONS, INTERSECTIONS&amp;#039;&amp;#039;&amp;#039; and &amp;#039;&amp;#039;&amp;#039;COMPLEMENTS&amp;#039;&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
To illustrate these concepts, consider the sample space &amp;lt;math&amp;gt;S=\{x;x\geq 0\},\,&amp;lt;/math&amp;gt; with the following events defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as depicted in Figure 3.2:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E=\{x;4&amp;lt;x\leq 10\},\,F=\{x;7&amp;lt;x\leq 17\},\,G=\{x;x&amp;gt;15\},\,H=\{x;9&amp;lt;x\leq&lt;br /&gt;
13\}. &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| (a) Event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: A closed loop&lt;br /&gt;
| (b) Union: &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2a.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2b.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (c) Intersection: &amp;lt;math&amp;gt;E\cap F&amp;lt;/math&amp;gt;&lt;br /&gt;
| (d) The Null set/event: &amp;lt;math&amp;gt;E\cap G=\emptyset &amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2c.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2d.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (e) Complement of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;&lt;br /&gt;
| (f) Subset of &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;H\subset F&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;H\cap F=H&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2e.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2f.jpg|frameless|300px]]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
        &lt;br /&gt;
&lt;br /&gt;
* The &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cup F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cup F=\{x;4&amp;lt;x\leq 17\};&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are either in &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or in &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) in both. This is illustrated on the Venn diagram by the dark shaded area in diagram (b).&lt;br /&gt;
* The&amp;#039;&amp;#039; intersection&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cap F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cap F=\left\{ x;7\leq x\leq 10\right\} ;&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are common to both &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; Again this is depicted by the dark shaded area in (c). If events have no elements in common (as, for example, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt;) then they are said to be &amp;#039;&amp;#039;mutually exclusive&amp;#039;&amp;#039;, and we can write &amp;lt;math&amp;gt;E\cap G=\emptyset ,&amp;lt;/math&amp;gt; meaning the &amp;#039;&amp;#039;null set&amp;#039;&amp;#039; which contains no elements. Such a situation is illustrated on the Venn Diagram by events (the two shaded closed loops in (d)) which do not overlap. Notice however that &amp;lt;math&amp;gt;G\cap F\neq \emptyset ,&amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have elements in common.&lt;br /&gt;
* The &amp;#039;&amp;#039;complement&amp;#039;&amp;#039; of an event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; say, is everything defined on the sample space which is not in &amp;lt;math&amp;gt;E.&amp;lt;/math&amp;gt; This event is denoted &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;, the dark shaded area in (e); here &amp;lt;math&amp;gt;\bar{E}=\left\{ x;x\leq 4\right\} \cup \left\{ x;x&amp;gt;10\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
* Finally note that &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; is a sub-set of &amp;lt;math&amp;gt;F;&amp;lt;/math&amp;gt; see (f). It is depicted as the dark closed loop wholly contained within &amp;lt;math&amp;gt;F,&amp;lt;/math&amp;gt; the lighter shaded area, so that &amp;lt;math&amp;gt;H\cap F=H;&amp;lt;/math&amp;gt; if an element in the sample space is a member of &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; then it must also be member of &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; (In mathematical logic, we employ this scenario to indicate that “&amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; implies &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;”, but not necessarily vice-versa.) Notice that &amp;lt;math&amp;gt;G\cap H=\emptyset &amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;H\cap E\neq \emptyset&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
= Probability =&lt;br /&gt;
&lt;br /&gt;
The term &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; (or some equivalent) is used in everyday conversation and so can not be unfamiliar to the reader. We talk of the probability, or chance, of rain; the likelihood of England winning the World Cup; or, perhaps more scientifically, the chance of getting a &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; when rolling a die. What we shall now do is develop a coherent theory of probability; a theory which allows us to combine and manipulate probabilities in a consistent and meaningful manner. We shall describe ways of dealing with, and describing, uncertainty. This will involve &amp;#039;&amp;#039;rules&amp;#039;&amp;#039; which govern our use of terms like probability.&lt;br /&gt;
&lt;br /&gt;
There have been a number of different approaches (interpretations) of probability. Most depend, at least to some extent, on the notion of relative frequency as now described:&lt;br /&gt;
&lt;br /&gt;
* Suppose an experiment has an outcome of interest &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;. The &amp;#039;&amp;#039;relative frequency interpretation&amp;#039;&amp;#039; of probability says that assuming the experiment can be repeated a large number of times then the relative frequency of observing the outcome &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; will settle down to a &amp;#039;&amp;#039;number&amp;#039;&amp;#039;, denoted &amp;lt;math&amp;gt;\Pr (E),&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;P(E)&amp;lt;/math&amp;gt; or Prob&amp;lt;math&amp;gt;(E),&amp;lt;/math&amp;gt; called the &amp;#039;&amp;#039;&amp;#039;probability&amp;#039;&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
This is illustrated in the next Figure where the proportion of heads obtained after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; flips of a fair coin is plotted against &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;, as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; increases; e.g., of the first &amp;lt;math&amp;gt;100&amp;lt;/math&amp;gt; flips, &amp;lt;math&amp;gt;55&amp;lt;/math&amp;gt; were heads (&amp;lt;math&amp;gt;55\%&amp;lt;/math&amp;gt;). Notice that the plot becomes less ‘wobbly’ after about &amp;lt;math&amp;gt;n=220&amp;lt;/math&amp;gt; and appears to be settling down to the value of &amp;lt;math&amp;gt;\frac{1}{2}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
[[File:Prob_coin.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
Due to this interpretation of probability, we often use observed sample proportions to approximate underlying probabilities of interest; see, for example, Question 4 of Exercise 2. There are, of course, other interpretations of probability; e.g., the subjective interpretation which simply expresses the strength of one’s belief about an event of interest such as whether Manchester United will win the European Cup! Any one of these interpretations can be used in practical situations provided the implied notion of probability follows a simple set of &amp;#039;&amp;#039;axioms&amp;#039;&amp;#039; or &amp;#039;&amp;#039;rules&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
== The axioms of probability ==&lt;br /&gt;
&lt;br /&gt;
There are just &amp;#039;&amp;#039;three &amp;#039;&amp;#039;basic rules that must be obeyed when dealing with probabilities:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;For any event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; i.e., &amp;lt;math&amp;gt;E\subset S,\,\,\Pr (E)\geq 0&amp;lt;/math&amp;gt;; &amp;#039;&amp;#039;probabilities are non-negative&amp;#039;&amp;#039;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (S)=1;&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;having defined the sample space of outcomes, one of these outcomes must be observed&amp;#039;&amp;#039;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;If events &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are mutually exclusive defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, so that &amp;lt;math&amp;gt;E\cap F=\emptyset &amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\Pr \left( E\right)+\Pr \left( F\right) .&amp;lt;/math&amp;gt; In general, for any set of mutually exclusive events, &amp;lt;math&amp;gt;E_{1},E_{2},\ldots ,E_{k},&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S:&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (E_{1}\cup E_{2}\cup \ldots \cup E_{k})=\Pr (E_{1})+\Pr (E_{2})+\ldots\Pr (E_{k})&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;i.e., &amp;lt;math&amp;gt;\Pr \left( \bigcup_{j=1}^{k}E_{j}\right) =\sum_{j=1}^{k}\Pr (E_{j}).&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In terms of the Venn Diagram, one can (and should) usefully think of the area of &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; relative to that of &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as providing an indication of probability. (Note, from axiom 2, that the area of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is implicitly normalised to be unity).&lt;br /&gt;
&lt;br /&gt;
Also observe that, contrary to what you may have believed, it is not one of the rules that &amp;lt;math&amp;gt;\Pr (E)\leq 1&amp;lt;/math&amp;gt; for any event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;. Rather, this is an implication of the &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; rules given:&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;implications: &amp;#039;&amp;#039;&amp;#039;it must be that for any event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;E\cap \bar{E}=\emptyset &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;E\cup \bar{E}=S.&amp;lt;/math&amp;gt; By Axiom &amp;lt;math&amp;gt;1,&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr (E)\geq 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr \left( \bar{E}\right) \geq 0&amp;lt;/math&amp;gt; and by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr(E)+\Pr (\bar{E})=\Pr (S).&amp;lt;/math&amp;gt; So &amp;lt;math&amp;gt;\Pr \left( E\right) +\Pr \left( \bar{E}\right) =1,&amp;lt;/math&amp;gt; by Axiom &amp;lt;math&amp;gt;2.&amp;lt;/math&amp;gt; This implies that&lt;br /&gt;
&lt;br /&gt;
# &amp;lt;math&amp;gt;0\leq \Pr (E)\leq 1&amp;lt;/math&amp;gt;&lt;br /&gt;
# &amp;lt;math&amp;gt;\Pr (\bar{E})=1-\Pr (E)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The first of these is what we might have expected from probability (a number lying between &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;). The second implication is also very important; it says that the probability of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; not happening is ‘&amp;#039;&amp;#039;one minus the probability of it happening&amp;#039;&amp;#039;’. Thus when rolling a die, the probability of getting &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; is one minus the probability of getting either a &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;2&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;5.&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
These axioms imply how to calculate probabilities on a sample space of equally likely outcomes. For example, and as we have already noted, the experiment of rolling a fair die defines a sample space of six, mutually exclusive and equally likely outcomes (&amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; dots on the up-turned face). The axioms then say that each of the six probabilities are positive, add to 1 and are all the same. Thus, the probability of any one of the outcomes must be simply &amp;lt;math&amp;gt;\frac{1}{6};&amp;lt;/math&amp;gt; which may accord with your intuition. A similar sort of analysis reveals that the probability of drawing a club from a deck of &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; cards is &amp;lt;math&amp;gt;\frac{13}{52},&amp;lt;/math&amp;gt; since any one of the &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; cards has an equal chance of being drawn and &amp;lt;math&amp;gt;13&amp;lt;/math&amp;gt; of them are clubs; i.e., &amp;lt;math&amp;gt;13&amp;lt;/math&amp;gt; of the &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; are clubs, so the probability of drawing a club is &amp;lt;math&amp;gt;\frac{13}{52}.&amp;lt;/math&amp;gt; Notice the importance of the assumption of equally likely outcomes here.&lt;br /&gt;
&lt;br /&gt;
In this, and the next section of notes, we shall see how these axioms can be used. Firstly, consider the construction of a probability for the &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of two events; i.e., the probability that &amp;#039;&amp;#039;either &amp;#039;&amp;#039;&amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) &amp;#039;&amp;#039;both &amp;#039;&amp;#039;will occur. Such a probability is embodied in the &amp;#039;&amp;#039;addition rule of probability&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
== The addition rule of probability ==&lt;br /&gt;
&lt;br /&gt;
When rolling a fair die, let &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; denote the event of an “odd number of dots” and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; the event of the “number of dots being greater than, or equal, to &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;”&amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt; What is the probability of the event &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;? To calculate this we can collect together all the mutually exclusive (simple) events which comprise &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;, and then add up the probabilities (by axiom 3). These simple events are &amp;lt;math&amp;gt;1,3,4,5&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; dots. Each has a probability of &amp;lt;math&amp;gt;\frac{1}{6},&amp;lt;/math&amp;gt; so the required total probability is: &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\frac{5}{6}&amp;lt;/math&amp;gt;. Consider carefully how this probability is constructed and note, in particular, that &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) \neq \Pr \left( E\right) +\Pr \left( F\right) &amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have a simple event in common (namely &amp;lt;math&amp;gt;5&amp;lt;/math&amp;gt; dots).&lt;br /&gt;
&lt;br /&gt;
In general, we can calculate the probability of the union of events using the &amp;#039;&amp;#039;addition rule of probability&amp;#039;&amp;#039;, as follows.&lt;br /&gt;
&lt;br /&gt;
* For any events, &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F\subset S:\Pr (E\cup F)=\Pr (E)+\Pr (F)-\Pr (E\cap F).&amp;lt;/math&amp;gt; So, in general, &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) \leq \Pr (E)+\Pr (F).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This generalises to three events, &amp;lt;math&amp;gt;E_{1},E_{2}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;E_{3}&amp;lt;/math&amp;gt; as&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E_{1}\cup E_{2}\cup E_{3}) &amp;amp;=&amp;amp;\Pr (E_{1})+\Pr (E_{2})+\Pr (E_{3}) \\&lt;br /&gt;
&amp;amp;&amp;amp;-\Pr (E_{1}\cap E_{2})-\Pr (E_{1}\cap E_{3})-\Pr (E_{2}\cap E_{3}) \\&lt;br /&gt;
&amp;amp;&amp;amp;+\Pr (E_{1}\cap E_{2}\cap E_{3}).\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
We can demonstrate this as follows.&lt;br /&gt;
&lt;br /&gt;
Note that&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E\cup F=\left( E\cap \bar{F}\right) \cup \left( E\cap F\right) \cup \left(\bar{E}\cap F\right)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
the union of &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; mutually exclusive events. These mutually exclusive events are depicted by the shaded areas &amp;lt;math&amp;gt;\mathbf{a,}&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\mathbf{b}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\mathbf{c}&amp;lt;/math&amp;gt;, respectively, in the next Figure.&lt;br /&gt;
&lt;br /&gt;
[[File:Prob_add.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
Then by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, and from the fact that the three events &amp;lt;math&amp;gt;\left( E\cap\bar{F}\right) &amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;\left( E\cap F\right) &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\left( \bar{E}\cap F\right)&amp;lt;/math&amp;gt; are mutually exclusive so that the “area” occupied by &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt; is simply &amp;lt;math&amp;gt;\mathbf{a+b+c,}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\Pr \left( E\cap \bar{F}\right) +\Pr \left( \bar{E}\cap F\right) +\Pr \left( E\cap F\right) .&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
But also by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, since &amp;lt;math&amp;gt;E=\left( E\cap \bar{F}\right) \cup \left(E\cap F\right) &amp;lt;/math&amp;gt;, it must be that &amp;lt;math&amp;gt;\Pr (E)=\Pr \left( E\cap \bar{F}\right)+\Pr (E\cap F);&amp;lt;/math&amp;gt; similarly, &amp;lt;math&amp;gt;\Pr \left( \bar{E}\cap F\right) =\Pr \left(F\right) -\Pr \left( E\cap F\right)&amp;lt;/math&amp;gt;. Putting all of this together gives&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\Pr (E\cup F)=\Pr (E)+\Pr (F)-\Pr (E\cap F).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
When &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are mutually exclusive, so that &amp;lt;math&amp;gt;E\cap F=\emptyset&amp;lt;/math&amp;gt;, this rule reveals Axiom 2: &amp;lt;math&amp;gt;\Pr (E\cup F)=\Pr (E)+\Pr (F).\smallskip &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example: &amp;#039;&amp;#039;What is the probability of drawing a Queen (&amp;lt;math&amp;gt;Q &amp;lt;/math&amp;gt;) or a Club (&amp;lt;math&amp;gt;C&amp;lt;/math&amp;gt;) in a single draw from a pack of cards? Now, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;52 &amp;lt;/math&amp;gt; cards are Queens, so &amp;lt;math&amp;gt;\Pr \left( Q\right) =\frac{4}{52},&amp;lt;/math&amp;gt; whilst &amp;lt;math&amp;gt;\Pr\left( C\right) =\frac{13}{52}.&amp;lt;/math&amp;gt; The probability of drawing the Queen of Clubs is simply &amp;lt;math&amp;gt;\frac{1}{52};&amp;lt;/math&amp;gt; i.e., &amp;lt;math&amp;gt;\Pr \left( Q\cap C\right) =\frac{1}{52}&amp;lt;/math&amp;gt;. What we require is a Club or a Queen, for which the probability is&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr \left( Q\cup C\right) &amp;amp;=&amp;amp;\Pr \left( Q\right) +\Pr \left( C\right) -\Pr\left( Q\cap C\right) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{4}{52}+\frac{13}{52}-\frac{1}{52} \\&lt;br /&gt;
&amp;amp;=&amp;amp;\frac{16}{52}=\frac{4}{13}.\end{aligned}&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example: &amp;#039;&amp;#039;Consider a car journey from Manchester to London via the M6 and M1. Let &amp;lt;math&amp;gt;E=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;heavy traffic somewhere on route&amp;#039;&amp;#039; and &amp;lt;math&amp;gt;F=&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;roadworks somewhere on route&amp;#039;&amp;#039;. It is estimated that &amp;lt;math&amp;gt;\Pr (E)=0.8&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr (F)=0.4,&amp;lt;/math&amp;gt; whilst the probability of NOT encountering both is &amp;lt;math&amp;gt;\Pr (\overline{E\cap F})=0.6.&amp;lt;/math&amp;gt; What is the probability of encountering heavytraffic or roadworks?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;We require &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) .&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E\cup F) &amp;amp;=&amp;amp;\Pr (E)+\Pr (F)-\Pr (E\cap F) \\&lt;br /&gt;
&amp;amp;=&amp;amp;\Pr (E)+\Pr (F)-(1-\Pr (\overline{E\cap F})) \\&lt;br /&gt;
&amp;amp;=&amp;amp;0.8+0.4-1+0.6 \\&lt;br /&gt;
&amp;amp;=&amp;amp;0.8=\Pr (E)\end{aligned}&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;Notice that this implies, in this case, that &amp;lt;math&amp;gt;F\subset E&amp;lt;/math&amp;gt; (why?). This &amp;#039;&amp;#039;model &amp;#039;&amp;#039;then implies that when there are roadworks somewhere on route you are bound to encounter heavy traffic; on the other hand, you can encounter heavy traffic on route without ever passing through roadworks. (My own experience of this motorway inclines me towards this implication!)&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Similar concepts apply when manipulating proportions as follows:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;#039;&amp;#039;Example&amp;#039;&amp;#039;&amp;lt;math&amp;gt;:&amp;lt;/math&amp;gt; A sample of &amp;lt;math&amp;gt;1000&amp;lt;/math&amp;gt; undergraduates were asked whether they took either Mathematics, Physics or Chemistry at A-level. The following responses were obtained: &amp;lt;math&amp;gt;100&amp;lt;/math&amp;gt; just took Mathematics; &amp;lt;math&amp;gt;70&amp;lt;/math&amp;gt; just took Physics; &amp;lt;math&amp;gt;100&amp;lt;/math&amp;gt; just took Chemistry; &amp;lt;math&amp;gt;150&amp;lt;/math&amp;gt; took Mathematics and Physics, but not Chemistry; &amp;lt;math&amp;gt;40&amp;lt;/math&amp;gt; took Mathematics and Chemistry, but not Physics; and, &amp;lt;math&amp;gt;240&amp;lt;/math&amp;gt; took Physics and Chemistry, but not Mathematics. What proportion took all three?&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;This can be addressed with the following diagram:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;[[File:Prob_Alevels.jpg|frameless|600px]]&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;The shaded area contains the number who took all three, which can be deduced from the above information (since the total of the numbers assigned to each part of the Venn diagram must be &amp;lt;math&amp;gt;1000&amp;lt;/math&amp;gt;). The answer is therefore &amp;lt;math&amp;gt;30\%&amp;lt;/math&amp;gt; (being &amp;lt;math&amp;gt;300&amp;lt;/math&amp;gt; out of &amp;lt;math&amp;gt;1000&amp;lt;/math&amp;gt;).&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;Two further results on unions, intersections and complements which are of use (and which are fairly easy to demonstrate using Venn diagrams) are &amp;#039;&amp;#039;&amp;#039;de Morgan Laws&amp;#039;&amp;#039;&amp;#039;:&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;ul&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\left( \bar{A}\cap \bar{B}\right) =\left( \overline{A\cup B}\right) &amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\bar{A}\cup \bar{B}=\left( \overline{A\cap B}\right) &amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&amp;lt;/li&amp;gt;&amp;lt;/ul&amp;gt;&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=File:Prob_coin.jpg&amp;diff=2852</id>
		<title>File:Prob coin.jpg</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=File:Prob_coin.jpg&amp;diff=2852"/>
				<updated>2013-08-07T20:59:35Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2851</id>
		<title>Probability Intro</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2851"/>
				<updated>2013-08-07T20:59:13Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Introducing Probability =&lt;br /&gt;
&lt;br /&gt;
So far we have been looking at ways of summarising samples of data drawn from an underlying population of interest. Although at times tedious, all such arithmetic calculations are fairly mechanical and straightforward to apply. To remind ourselves, one of the primary reasons for wishing to summarise data is so assist in the development of inferences about the population from which the data were taken. That is to say, we would like to elicit some information about the mechanism which generated the observed data.&lt;br /&gt;
&lt;br /&gt;
We now start on the process of developing mathematical ways of formulating inferences and this requires the use of &amp;#039;&amp;#039;probability&amp;#039;&amp;#039;. This becomes clear if we think back to one of the early questions posed in this course: &amp;#039;&amp;#039;prior to sampling is it possible to predict with absolute certainty what will be observed&amp;#039;&amp;#039;? The answer to this question is &amp;#039;&amp;#039;no&amp;#039;&amp;#039;; although it would be of interest to know how &amp;#039;&amp;#039;likely&amp;#039;&amp;#039; it is that certain values would be observed. Or, what is the &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; of observing certain values?&lt;br /&gt;
&lt;br /&gt;
Before proceeding, we need some more tools:&lt;br /&gt;
&lt;br /&gt;
= Venn diagrams =&lt;br /&gt;
&lt;br /&gt;
Venn diagrams (and diagrams in general) are of enormous help in trying to understand, and manipulate probability. We begin with some basic definitions, some of which we have encountered before.&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Experiment:&amp;#039;&amp;#039;&amp;#039; any process which, when applied, provides data or an outcome; e.g., rolling a die and observing the number of dots on the upturned face; recording the amount of rainfall in Manchester over a period of time.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Sample Space:&amp;#039;&amp;#039;&amp;#039; set of possible outcomes of an experiment; e.g., &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; (or &amp;lt;math&amp;gt;\Omega &amp;lt;/math&amp;gt;) &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{1,2,3,4,5,6\}&amp;lt;/math&amp;gt;, which is the sample space of rolling a dice. Or &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{x;x\geq 0\}&amp;lt;/math&amp;gt;, which is the sample space of an experiment where the outcomes can be any real non-negative number, or ‘&amp;#039;&amp;#039;the set of real non-negative real numbers&amp;#039;&amp;#039;’.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Simple Event&amp;#039;&amp;#039;&amp;#039;: just one of the possible outcomes on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Event:&amp;#039;&amp;#039;&amp;#039; a &amp;#039;&amp;#039;subset&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, denoted &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;; e.g., &amp;lt;math&amp;gt;E=\left\{ 2,4,6\right\}&amp;lt;/math&amp;gt; (i.e. any even number on a dice) or &amp;lt;math&amp;gt;E=\left\{ x;4&amp;lt;x\leq 10\right\} ,&amp;lt;/math&amp;gt; which means ‘&amp;#039;&amp;#039;the set of real numbers which are strictly bigger than&amp;#039;&amp;#039; &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;but less than or equal to &amp;#039;&amp;#039;&amp;lt;math&amp;gt;10&amp;lt;/math&amp;gt;’.&amp;lt;br /&amp;gt;Note that an event, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, is a collection of simple events.&lt;br /&gt;
&lt;br /&gt;
Such concepts can be represented by means of the following Venn Diagram:&lt;br /&gt;
&lt;br /&gt;
[[File:Venn_1.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
The sample space, &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; is depicted as a closed rectangle, and the event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; is a closed loop wholly contained within &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; and we write (in set notation) &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
In dealing with probability, and in particular the probability of an event (or events) occurring, we shall need to be familiar with &amp;#039;&amp;#039;&amp;#039;UNIONS, INTERSECTIONS&amp;#039;&amp;#039;&amp;#039; and &amp;#039;&amp;#039;&amp;#039;COMPLEMENTS&amp;#039;&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
To illustrate these concepts, consider the sample space &amp;lt;math&amp;gt;S=\{x;x\geq 0\},\,&amp;lt;/math&amp;gt; with the following events defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as depicted in Figure 3.2:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E=\{x;4&amp;lt;x\leq 10\},\,F=\{x;7&amp;lt;x\leq 17\},\,G=\{x;x&amp;gt;15\},\,H=\{x;9&amp;lt;x\leq&lt;br /&gt;
13\}. &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| (a) Event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: A closed loop&lt;br /&gt;
| (b) Union: &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2a.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2b.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (c) Intersection: &amp;lt;math&amp;gt;E\cap F&amp;lt;/math&amp;gt;&lt;br /&gt;
| (d) The Null set/event: &amp;lt;math&amp;gt;E\cap G=\emptyset &amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2c.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2d.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (e) Complement of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;&lt;br /&gt;
| (f) Subset of &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;H\subset F&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;H\cap F=H&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2e.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2f.jpg|frameless|300px]]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
        &lt;br /&gt;
&lt;br /&gt;
* The &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cup F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cup F=\{x;4&amp;lt;x\leq 17\};&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are either in &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or in &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) in both. This is illustrated on the Venn diagram by the dark shaded area in diagram (b).&lt;br /&gt;
* The&amp;#039;&amp;#039; intersection&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cap F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cap F=\left\{ x;7\leq x\leq 10\right\} ;&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are common to both &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; Again this is depicted by the dark shaded area in (c). If events have no elements in common (as, for example, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt;) then they are said to be &amp;#039;&amp;#039;mutually exclusive&amp;#039;&amp;#039;, and we can write &amp;lt;math&amp;gt;E\cap G=\emptyset ,&amp;lt;/math&amp;gt; meaning the &amp;#039;&amp;#039;null set&amp;#039;&amp;#039; which contains no elements. Such a situation is illustrated on the Venn Diagram by events (the two shaded closed loops in (d)) which do not overlap. Notice however that &amp;lt;math&amp;gt;G\cap F\neq \emptyset ,&amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have elements in common.&lt;br /&gt;
* The &amp;#039;&amp;#039;complement&amp;#039;&amp;#039; of an event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; say, is everything defined on the sample space which is not in &amp;lt;math&amp;gt;E.&amp;lt;/math&amp;gt; This event is denoted &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;, the dark shaded area in (e); here &amp;lt;math&amp;gt;\bar{E}=\left\{ x;x\leq 4\right\} \cup \left\{ x;x&amp;gt;10\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
* Finally note that &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; is a sub-set of &amp;lt;math&amp;gt;F;&amp;lt;/math&amp;gt; see (f). It is depicted as the dark closed loop wholly contained within &amp;lt;math&amp;gt;F,&amp;lt;/math&amp;gt; the lighter shaded area, so that &amp;lt;math&amp;gt;H\cap F=H;&amp;lt;/math&amp;gt; if an element in the sample space is a member of &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; then it must also be member of &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; (In mathematical logic, we employ this scenario to indicate that “&amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; implies &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;”, but not necessarily vice-versa.) Notice that &amp;lt;math&amp;gt;G\cap H=\emptyset &amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;H\cap E\neq \emptyset&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
= Probability =&lt;br /&gt;
&lt;br /&gt;
The term &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; (or some equivalent) is used in everyday conversation and so can not be unfamiliar to the reader. We talk of the probability, or chance, of rain; the likelihood of England winning the World Cup; or, perhaps more scientifically, the chance of getting a &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; when rolling a die. What we shall now do is develop a coherent theory of probability; a theory which allows us to combine and manipulate probabilities in a consistent and meaningful manner. We shall describe ways of dealing with, and describing, uncertainty. This will involve &amp;#039;&amp;#039;rules&amp;#039;&amp;#039; which govern our use of terms like probability.&lt;br /&gt;
&lt;br /&gt;
There have been a number of different approaches (interpretations) of probability. Most depend, at least to some extent, on the notion of relative frequency as now described:&lt;br /&gt;
&lt;br /&gt;
* Suppose an experiment has an outcome of interest &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;. The &amp;#039;&amp;#039;relative frequency interpretation&amp;#039;&amp;#039; of probability says that assuming the experiment can be repeated a large number of times then the relative frequency of observing the outcome &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; will settle down to a &amp;#039;&amp;#039;number&amp;#039;&amp;#039;, denoted &amp;lt;math&amp;gt;\Pr (E),&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;P(E)&amp;lt;/math&amp;gt; or Prob&amp;lt;math&amp;gt;(E),&amp;lt;/math&amp;gt; called the &amp;#039;&amp;#039;&amp;#039;probability&amp;#039;&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
This is illustrated in the next Figure where the proportion of heads obtained after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; flips of a fair coin is plotted against &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;, as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; increases; e.g., of the first &amp;lt;math&amp;gt;100&amp;lt;/math&amp;gt; flips, &amp;lt;math&amp;gt;55&amp;lt;/math&amp;gt; were heads (&amp;lt;math&amp;gt;55\%&amp;lt;/math&amp;gt;). Notice that the plot becomes less ‘wobbly’ after about &amp;lt;math&amp;gt;n=220&amp;lt;/math&amp;gt; and appears to be settling down to the value of &amp;lt;math&amp;gt;\frac{1}{2}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
[[File:Prob_coin.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
Due to this interpretation of probability, we often use observed sample proportions to approximate underlying probabilities of interest; see, for example, Question 4 of Exercise 2. There are, of course, other interpretations of probability; e.g., the subjective interpretation which simply expresses the strength of one’s belief about an event of interest such as whether Manchester United will win the European Cup! Any one of these interpretations can be used in practical situations provided the implied notion of probability follows a simple set of &amp;#039;&amp;#039;axioms&amp;#039;&amp;#039; or &amp;#039;&amp;#039;rules&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
== The axioms of probability ==&lt;br /&gt;
&lt;br /&gt;
There are just &amp;#039;&amp;#039;three &amp;#039;&amp;#039;basic rules that must be obeyed when dealing with probabilities:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;For any event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; i.e., &amp;lt;math&amp;gt;E\subset S,\,\,\Pr (E)\geq 0&amp;lt;/math&amp;gt;; &amp;#039;&amp;#039;probabilities are non-negative&amp;#039;&amp;#039;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (S)=1;&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;having defined the sample space of outcomes, one of these outcomes must be observed&amp;#039;&amp;#039;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;If events &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are mutually exclusive defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, so that &amp;lt;math&amp;gt;E\cap F=\emptyset &amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\Pr \left( E\right)+\Pr \left( F\right) .&amp;lt;/math&amp;gt; In general, for any set of mutually exclusive events, &amp;lt;math&amp;gt;E_{1},E_{2},\ldots ,E_{k},&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S:&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (E_{1}\cup E_{2}\cup \ldots \cup E_{k})=\Pr (E_{1})+\Pr (E_{2})+\ldots\Pr (E_{k})&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;i.e., &amp;lt;math&amp;gt;\Pr \left( \bigcup_{j=1}^{k}E_{j}\right) =\sum_{j=1}^{k}\Pr (E_{j}).&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In terms of the Venn Diagram, one can (and should) usefully think of the area of &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; relative to that of &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as providing an indication of probability. (Note, from axiom 2, that the area of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is implicitly normalised to be unity).&lt;br /&gt;
&lt;br /&gt;
Also observe that, contrary to what you may have believed, it is not one of the rules that &amp;lt;math&amp;gt;\Pr (E)\leq 1&amp;lt;/math&amp;gt; for any event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;. Rather, this is an implication of the &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; rules given:&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;implications: &amp;#039;&amp;#039;&amp;#039;it must be that for any event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;E\cap \bar{E}=\emptyset &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;E\cup \bar{E}=S.&amp;lt;/math&amp;gt; By Axiom &amp;lt;math&amp;gt;1,&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr (E)\geq 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr \left( \bar{E}\right) \geq 0&amp;lt;/math&amp;gt; and by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr(E)+\Pr (\bar{E})=\Pr (S).&amp;lt;/math&amp;gt; So &amp;lt;math&amp;gt;\Pr \left( E\right) +\Pr \left( \bar{E}\right) =1,&amp;lt;/math&amp;gt; by Axiom &amp;lt;math&amp;gt;2.&amp;lt;/math&amp;gt; This implies that&lt;br /&gt;
&lt;br /&gt;
# &amp;lt;math&amp;gt;0\leq \Pr (E)\leq 1&amp;lt;/math&amp;gt;&lt;br /&gt;
# &amp;lt;math&amp;gt;\Pr (\bar{E})=1-\Pr (E)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The first of these is what we might have expected from probability (a number lying between &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;). The second implication is also very important; it says that the probability of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; not happening is ‘&amp;#039;&amp;#039;one minus the probability of it happening&amp;#039;&amp;#039;’. Thus when rolling a die, the probability of getting &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; is one minus the probability of getting either a &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;2&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;5.&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
These axioms imply how to calculate probabilities on a sample space of equally likely outcomes. For example, and as we have already noted, the experiment of rolling a fair die defines a sample space of six, mutually exclusive and equally likely outcomes (&amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; dots on the up-turned face). The axioms then say that each of the six probabilities are positive, add to 1 and are all the same. Thus, the probability of any one of the outcomes must be simply &amp;lt;math&amp;gt;\frac{1}{6};&amp;lt;/math&amp;gt; which may accord with your intuition. A similar sort of analysis reveals that the probability of drawing a club from a deck of &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; cards is &amp;lt;math&amp;gt;\frac{13}{52},&amp;lt;/math&amp;gt; since any one of the &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; cards has an equal chance of being drawn and &amp;lt;math&amp;gt;13&amp;lt;/math&amp;gt; of them are clubs; i.e., &amp;lt;math&amp;gt;13&amp;lt;/math&amp;gt; of the &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; are clubs, so the probability of drawing a club is &amp;lt;math&amp;gt;\frac{13}{52}.&amp;lt;/math&amp;gt; Notice the importance of the assumption of equally likely outcomes here.&lt;br /&gt;
&lt;br /&gt;
In this, and the next section of notes, we shall see how these axioms can be used. Firstly, consider the construction of a probability for the &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of two events; i.e., the probability that &amp;#039;&amp;#039;either &amp;#039;&amp;#039;&amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) &amp;#039;&amp;#039;both &amp;#039;&amp;#039;will occur. Such a probability is embodied in the &amp;#039;&amp;#039;addition rule of probability&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
== The addition rule of probability ==&lt;br /&gt;
&lt;br /&gt;
When rolling a fair die, let &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; denote the event of an “odd number of dots” and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; the event of the “number of dots being greater than, or equal, to &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;”&amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt; What is the probability of the event &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;? To calculate this we can collect together all the mutually exclusive (simple) events which comprise &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;, and then add up the probabilities (by axiom 3). These simple events are &amp;lt;math&amp;gt;1,3,4,5&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; dots. Each has a probability of &amp;lt;math&amp;gt;\frac{1}{6},&amp;lt;/math&amp;gt; so the required total probability is: &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\frac{5}{6}&amp;lt;/math&amp;gt;. Consider carefully how this probability is constructed and note, in particular, that &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) \neq \Pr \left( E\right) +\Pr \left( F\right) &amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have a simple event in common (namely &amp;lt;math&amp;gt;5&amp;lt;/math&amp;gt; dots).&lt;br /&gt;
&lt;br /&gt;
In general, we can calculate the probability of the union of events using the &amp;#039;&amp;#039;addition rule of probability&amp;#039;&amp;#039;, as follows.&lt;br /&gt;
&lt;br /&gt;
* For any events, &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F\subset S:\Pr (E\cup F)=\Pr (E)+\Pr (F)-\Pr (E\cap F).&amp;lt;/math&amp;gt; So, in general, &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) \leq \Pr (E)+\Pr (F).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This generalises to three events, &amp;lt;math&amp;gt;E_{1},E_{2}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;E_{3}&amp;lt;/math&amp;gt; as&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E_{1}\cup E_{2}\cup E_{3}) &amp;amp;=&amp;amp;\Pr (E_{1})+\Pr (E_{2})+\Pr (E_{3}) \\&lt;br /&gt;
&amp;amp;&amp;amp;-\Pr (E_{1}\cap E_{2})-\Pr (E_{1}\cap E_{3})-\Pr (E_{2}\cap E_{3}) \\&lt;br /&gt;
&amp;amp;&amp;amp;+\Pr (E_{1}\cap E_{2}\cap E_{3}).\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2850</id>
		<title>Probability Intro</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2850"/>
				<updated>2013-08-07T20:57:49Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Introducing Probability =&lt;br /&gt;
&lt;br /&gt;
So far we have been looking at ways of summarising samples of data drawn from an underlying population of interest. Although at times tedious, all such arithmetic calculations are fairly mechanical and straightforward to apply. To remind ourselves, one of the primary reasons for wishing to summarise data is so assist in the development of inferences about the population from which the data were taken. That is to say, we would like to elicit some information about the mechanism which generated the observed data.&lt;br /&gt;
&lt;br /&gt;
We now start on the process of developing mathematical ways of formulating inferences and this requires the use of &amp;#039;&amp;#039;probability&amp;#039;&amp;#039;. This becomes clear if we think back to one of the early questions posed in this course: &amp;#039;&amp;#039;prior to sampling is it possible to predict with absolute certainty what will be observed&amp;#039;&amp;#039;? The answer to this question is &amp;#039;&amp;#039;no&amp;#039;&amp;#039;; although it would be of interest to know how &amp;#039;&amp;#039;likely&amp;#039;&amp;#039; it is that certain values would be observed. Or, what is the &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; of observing certain values?&lt;br /&gt;
&lt;br /&gt;
Before proceeding, we need some more tools:&lt;br /&gt;
&lt;br /&gt;
= Venn diagrams =&lt;br /&gt;
&lt;br /&gt;
Venn diagrams (and diagrams in general) are of enormous help in trying to understand, and manipulate probability. We begin with some basic definitions, some of which we have encountered before.&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Experiment:&amp;#039;&amp;#039;&amp;#039; any process which, when applied, provides data or an outcome; e.g., rolling a die and observing the number of dots on the upturned face; recording the amount of rainfall in Manchester over a period of time.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Sample Space:&amp;#039;&amp;#039;&amp;#039; set of possible outcomes of an experiment; e.g., &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; (or &amp;lt;math&amp;gt;\Omega &amp;lt;/math&amp;gt;) &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{1,2,3,4,5,6\}&amp;lt;/math&amp;gt;, which is the sample space of rolling a dice. Or &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{x;x\geq 0\}&amp;lt;/math&amp;gt;, which is the sample space of an experiment where the outcomes can be any real non-negative number, or ‘&amp;#039;&amp;#039;the set of real non-negative real numbers&amp;#039;&amp;#039;’.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Simple Event&amp;#039;&amp;#039;&amp;#039;: just one of the possible outcomes on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Event:&amp;#039;&amp;#039;&amp;#039; a &amp;#039;&amp;#039;subset&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, denoted &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;; e.g., &amp;lt;math&amp;gt;E=\left\{ 2,4,6\right\}&amp;lt;/math&amp;gt; (i.e. any even number on a dice) or &amp;lt;math&amp;gt;E=\left\{ x;4&amp;lt;x\leq 10\right\} ,&amp;lt;/math&amp;gt; which means ‘&amp;#039;&amp;#039;the set of real numbers which are strictly bigger than&amp;#039;&amp;#039; &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;but less than or equal to &amp;#039;&amp;#039;&amp;lt;math&amp;gt;10&amp;lt;/math&amp;gt;’.&amp;lt;br /&amp;gt;Note that an event, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, is a collection of simple events.&lt;br /&gt;
&lt;br /&gt;
Such concepts can be represented by means of the following Venn Diagram:&lt;br /&gt;
&lt;br /&gt;
[[File:Venn_1.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
The sample space, &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; is depicted as a closed rectangle, and the event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; is a closed loop wholly contained within &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; and we write (in set notation) &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
In dealing with probability, and in particular the probability of an event (or events) occurring, we shall need to be familiar with &amp;#039;&amp;#039;&amp;#039;UNIONS, INTERSECTIONS&amp;#039;&amp;#039;&amp;#039; and &amp;#039;&amp;#039;&amp;#039;COMPLEMENTS&amp;#039;&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
To illustrate these concepts, consider the sample space &amp;lt;math&amp;gt;S=\{x;x\geq 0\},\,&amp;lt;/math&amp;gt; with the following events defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as depicted in Figure 3.2:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E=\{x;4&amp;lt;x\leq 10\},\,F=\{x;7&amp;lt;x\leq 17\},\,G=\{x;x&amp;gt;15\},\,H=\{x;9&amp;lt;x\leq&lt;br /&gt;
13\}. &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| (a) Event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: A closed loop&lt;br /&gt;
| (b) Union: &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2a.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2b.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (c) Intersection: &amp;lt;math&amp;gt;E\cap F&amp;lt;/math&amp;gt;&lt;br /&gt;
| (d) The Null set/event: &amp;lt;math&amp;gt;E\cap G=\emptyset &amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2c.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2d.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (e) Complement of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;&lt;br /&gt;
| (f) Subset of &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;H\subset F&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;H\cap F=H&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2e.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2f.jpg|frameless|300px]]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
        &lt;br /&gt;
&lt;br /&gt;
* The &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cup F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cup F=\{x;4&amp;lt;x\leq 17\};&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are either in &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or in &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) in both. This is illustrated on the Venn diagram by the dark shaded area in diagram (b).&lt;br /&gt;
* The&amp;#039;&amp;#039; intersection&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cap F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cap F=\left\{ x;7\leq x\leq 10\right\} ;&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are common to both &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; Again this is depicted by the dark shaded area in (c). If events have no elements in common (as, for example, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt;) then they are said to be &amp;#039;&amp;#039;mutually exclusive&amp;#039;&amp;#039;, and we can write &amp;lt;math&amp;gt;E\cap G=\emptyset ,&amp;lt;/math&amp;gt; meaning the &amp;#039;&amp;#039;null set&amp;#039;&amp;#039; which contains no elements. Such a situation is illustrated on the Venn Diagram by events (the two shaded closed loops in (d)) which do not overlap. Notice however that &amp;lt;math&amp;gt;G\cap F\neq \emptyset ,&amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have elements in common.&lt;br /&gt;
* The &amp;#039;&amp;#039;complement&amp;#039;&amp;#039; of an event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; say, is everything defined on the sample space which is not in &amp;lt;math&amp;gt;E.&amp;lt;/math&amp;gt; This event is denoted &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;, the dark shaded area in (e); here &amp;lt;math&amp;gt;\bar{E}=\left\{ x;x\leq 4\right\} \cup \left\{ x;x&amp;gt;10\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
* Finally note that &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; is a sub-set of &amp;lt;math&amp;gt;F;&amp;lt;/math&amp;gt; see (f). It is depicted as the dark closed loop wholly contained within &amp;lt;math&amp;gt;F,&amp;lt;/math&amp;gt; the lighter shaded area, so that &amp;lt;math&amp;gt;H\cap F=H;&amp;lt;/math&amp;gt; if an element in the sample space is a member of &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; then it must also be member of &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; (In mathematical logic, we employ this scenario to indicate that “&amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; implies &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;”, but not necessarily vice-versa.) Notice that &amp;lt;math&amp;gt;G\cap H=\emptyset &amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;H\cap E\neq \emptyset&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
= Probability =&lt;br /&gt;
&lt;br /&gt;
The term &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; (or some equivalent) is used in everyday conversation and so can not be unfamiliar to the reader. We talk of the probability, or chance, of rain; the likelihood of England winning the World Cup; or, perhaps more scientifically, the chance of getting a &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; when rolling a die. What we shall now do is develop a coherent theory of probability; a theory which allows us to combine and manipulate probabilities in a consistent and meaningful manner. We shall describe ways of dealing with, and describing, uncertainty. This will involve &amp;#039;&amp;#039;rules&amp;#039;&amp;#039; which govern our use of terms like probability.&lt;br /&gt;
&lt;br /&gt;
There have been a number of different approaches (interpretations) of probability. Most depend, at least to some extent, on the notion of relative frequency as now described:&lt;br /&gt;
&lt;br /&gt;
* Suppose an experiment has an outcome of interest &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;. The &amp;#039;&amp;#039;relative frequency interpretation&amp;#039;&amp;#039; of probability says that assuming the experiment can be repeated a large number of times then the relative frequency of observing the outcome &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; will settle down to a &amp;#039;&amp;#039;number&amp;#039;&amp;#039;, denoted &amp;lt;math&amp;gt;\Pr (E),&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;P(E)&amp;lt;/math&amp;gt; or Prob&amp;lt;math&amp;gt;(E),&amp;lt;/math&amp;gt; called the &amp;#039;&amp;#039;&amp;#039;probability&amp;#039;&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
This is illustrated in the next Figure where the proportion of heads obtained after &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; flips of a fair coin is plotted against &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt;, as &amp;lt;math&amp;gt;n&amp;lt;/math&amp;gt; increases; e.g., of the first &amp;lt;math&amp;gt;100&amp;lt;/math&amp;gt; flips, &amp;lt;math&amp;gt;55&amp;lt;/math&amp;gt; were heads (&amp;lt;math&amp;gt;55\%&amp;lt;/math&amp;gt;). Notice that the plot becomes less ‘wobbly’ after about &amp;lt;math&amp;gt;n=220&amp;lt;/math&amp;gt; and appears to be settling down to the value of &amp;lt;math&amp;gt;\frac{1}{2}&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
[[File:Prob_coin2d.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
Due to this interpretation of probability, we often use observed sample proportions to approximate underlying probabilities of interest; see, for example, Question 4 of Exercise 2. There are, of course, other interpretations of probability; e.g., the subjective interpretation which simply expresses the strength of one’s belief about an event of interest such as whether Manchester United will win the European Cup! Any one of these interpretations can be used in practical situations provided the implied notion of probability follows a simple set of &amp;#039;&amp;#039;axioms&amp;#039;&amp;#039; or &amp;#039;&amp;#039;rules&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
== The axioms of probability ==&lt;br /&gt;
&lt;br /&gt;
There are just &amp;#039;&amp;#039;three &amp;#039;&amp;#039;basic rules that must be obeyed when dealing with probabilities:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;For any event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; i.e., &amp;lt;math&amp;gt;E\subset S,\,\,\Pr (E)\geq 0&amp;lt;/math&amp;gt;; &amp;#039;&amp;#039;probabilities are non-negative&amp;#039;&amp;#039;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (S)=1;&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;having defined the sample space of outcomes, one of these outcomes must be observed&amp;#039;&amp;#039;.&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;&amp;lt;p&amp;gt;If events &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; are mutually exclusive defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, so that &amp;lt;math&amp;gt;E\cap F=\emptyset &amp;lt;/math&amp;gt;, then &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\Pr \left( E\right)+\Pr \left( F\right) .&amp;lt;/math&amp;gt; In general, for any set of mutually exclusive events, &amp;lt;math&amp;gt;E_{1},E_{2},\ldots ,E_{k},&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S:&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;&amp;lt;math&amp;gt;\Pr (E_{1}\cup E_{2}\cup \ldots \cup E_{k})=\Pr (E_{1})+\Pr (E_{2})+\ldots\Pr (E_{k})&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&lt;br /&gt;
&amp;lt;p&amp;gt;i.e., &amp;lt;math&amp;gt;\Pr \left( \bigcup_{j=1}^{k}E_{j}\right) =\sum_{j=1}^{k}\Pr (E_{j}).&amp;lt;/math&amp;gt;&amp;lt;/p&amp;gt;&amp;lt;/li&amp;gt;&amp;lt;/ol&amp;gt;&lt;br /&gt;
&lt;br /&gt;
In terms of the Venn Diagram, one can (and should) usefully think of the area of &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; relative to that of &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as providing an indication of probability. (Note, from axiom 2, that the area of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; is implicitly normalised to be unity).&lt;br /&gt;
&lt;br /&gt;
Also observe that, contrary to what you may have believed, it is not one of the rules that &amp;lt;math&amp;gt;\Pr (E)\leq 1&amp;lt;/math&amp;gt; for any event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;. Rather, this is an implication of the &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; rules given:&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;implications: &amp;#039;&amp;#039;&amp;#039;it must be that for any event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; defined on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;E\cap \bar{E}=\emptyset &amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;E\cup \bar{E}=S.&amp;lt;/math&amp;gt; By Axiom &amp;lt;math&amp;gt;1,&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr (E)\geq 0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;\Pr \left( \bar{E}\right) \geq 0&amp;lt;/math&amp;gt; and by Axiom &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\Pr(E)+\Pr (\bar{E})=\Pr (S).&amp;lt;/math&amp;gt; So &amp;lt;math&amp;gt;\Pr \left( E\right) +\Pr \left( \bar{E}\right) =1,&amp;lt;/math&amp;gt; by Axiom &amp;lt;math&amp;gt;2.&amp;lt;/math&amp;gt; This implies that&lt;br /&gt;
&lt;br /&gt;
# &amp;lt;math&amp;gt;0\leq \Pr (E)\leq 1&amp;lt;/math&amp;gt;&lt;br /&gt;
# &amp;lt;math&amp;gt;\Pr (\bar{E})=1-\Pr (E)&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
The first of these is what we might have expected from probability (a number lying between &amp;lt;math&amp;gt;0&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;). The second implication is also very important; it says that the probability of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; not happening is ‘&amp;#039;&amp;#039;one minus the probability of it happening&amp;#039;&amp;#039;’. Thus when rolling a die, the probability of getting &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; is one minus the probability of getting either a &amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;2&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;3&amp;lt;/math&amp;gt;, &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;5.&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
These axioms imply how to calculate probabilities on a sample space of equally likely outcomes. For example, and as we have already noted, the experiment of rolling a fair die defines a sample space of six, mutually exclusive and equally likely outcomes (&amp;lt;math&amp;gt;1&amp;lt;/math&amp;gt; to &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; dots on the up-turned face). The axioms then say that each of the six probabilities are positive, add to 1 and are all the same. Thus, the probability of any one of the outcomes must be simply &amp;lt;math&amp;gt;\frac{1}{6};&amp;lt;/math&amp;gt; which may accord with your intuition. A similar sort of analysis reveals that the probability of drawing a club from a deck of &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; cards is &amp;lt;math&amp;gt;\frac{13}{52},&amp;lt;/math&amp;gt; since any one of the &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; cards has an equal chance of being drawn and &amp;lt;math&amp;gt;13&amp;lt;/math&amp;gt; of them are clubs; i.e., &amp;lt;math&amp;gt;13&amp;lt;/math&amp;gt; of the &amp;lt;math&amp;gt;52&amp;lt;/math&amp;gt; are clubs, so the probability of drawing a club is &amp;lt;math&amp;gt;\frac{13}{52}.&amp;lt;/math&amp;gt; Notice the importance of the assumption of equally likely outcomes here.&lt;br /&gt;
&lt;br /&gt;
In this, and the next section of notes, we shall see how these axioms can be used. Firstly, consider the construction of a probability for the &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of two events; i.e., the probability that &amp;#039;&amp;#039;either &amp;#039;&amp;#039;&amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) &amp;#039;&amp;#039;both &amp;#039;&amp;#039;will occur. Such a probability is embodied in the &amp;#039;&amp;#039;addition rule of probability&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
== The addition rule of probability ==&lt;br /&gt;
&lt;br /&gt;
When rolling a fair die, let &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; denote the event of an “odd number of dots” and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; the event of the “number of dots being greater than, or equal, to &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt;”&amp;lt;math&amp;gt;.&amp;lt;/math&amp;gt; What is the probability of the event &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;? To calculate this we can collect together all the mutually exclusive (simple) events which comprise &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;, and then add up the probabilities (by axiom 3). These simple events are &amp;lt;math&amp;gt;1,3,4,5&amp;lt;/math&amp;gt; or &amp;lt;math&amp;gt;6&amp;lt;/math&amp;gt; dots. Each has a probability of &amp;lt;math&amp;gt;\frac{1}{6},&amp;lt;/math&amp;gt; so the required total probability is: &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) =\frac{5}{6}&amp;lt;/math&amp;gt;. Consider carefully how this probability is constructed and note, in particular, that &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) \neq \Pr \left( E\right) +\Pr \left( F\right) &amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have a simple event in common (namely &amp;lt;math&amp;gt;5&amp;lt;/math&amp;gt; dots).&lt;br /&gt;
&lt;br /&gt;
In general, we can calculate the probability of the union of events using the &amp;#039;&amp;#039;addition rule of probability&amp;#039;&amp;#039;, as follows.&lt;br /&gt;
&lt;br /&gt;
* For any events, &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F\subset S:\Pr (E\cup F)=\Pr (E)+\Pr (F)-\Pr (E\cap F).&amp;lt;/math&amp;gt; So, in general, &amp;lt;math&amp;gt;\Pr \left( E\cup F\right) \leq \Pr (E)+\Pr (F).&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This generalises to three events, &amp;lt;math&amp;gt;E_{1},E_{2}&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;E_{3}&amp;lt;/math&amp;gt; as&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;\begin{aligned}&lt;br /&gt;
\Pr (E_{1}\cup E_{2}\cup E_{3}) &amp;amp;=&amp;amp;\Pr (E_{1})+\Pr (E_{2})+\Pr (E_{3}) \\&lt;br /&gt;
&amp;amp;&amp;amp;-\Pr (E_{1}\cap E_{2})-\Pr (E_{1}\cap E_{3})-\Pr (E_{2}\cap E_{3}) \\&lt;br /&gt;
&amp;amp;&amp;amp;+\Pr (E_{1}\cap E_{2}\cap E_{3}).\end{aligned}&amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2849</id>
		<title>Probability Intro</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2849"/>
				<updated>2013-08-07T20:38:43Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: /* Venn diagrams */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Introducing Probability =&lt;br /&gt;
&lt;br /&gt;
So far we have been looking at ways of summarising samples of data drawn from an underlying population of interest. Although at times tedious, all such arithmetic calculations are fairly mechanical and straightforward to apply. To remind ourselves, one of the primary reasons for wishing to summarise data is so assist in the development of inferences about the population from which the data were taken. That is to say, we would like to elicit some information about the mechanism which generated the observed data.&lt;br /&gt;
&lt;br /&gt;
We now start on the process of developing mathematical ways of formulating inferences and this requires the use of &amp;#039;&amp;#039;probability&amp;#039;&amp;#039;. This becomes clear if we think back to one of the early questions posed in this course: &amp;#039;&amp;#039;prior to sampling is it possible to predict with absolute certainty what will be observed&amp;#039;&amp;#039;? The answer to this question is &amp;#039;&amp;#039;no&amp;#039;&amp;#039;; although it would be of interest to know how &amp;#039;&amp;#039;likely&amp;#039;&amp;#039; it is that certain values would be observed. Or, what is the &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; of observing certain values?&lt;br /&gt;
&lt;br /&gt;
Before proceeding, we need some more tools:&lt;br /&gt;
&lt;br /&gt;
= Venn diagrams =&lt;br /&gt;
&lt;br /&gt;
Venn diagrams (and diagrams in general) are of enormous help in trying to understand, and manipulate probability. We begin with some basic definitions, some of which we have encountered before.&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Experiment:&amp;#039;&amp;#039;&amp;#039; any process which, when applied, provides data or an outcome; e.g., rolling a die and observing the number of dots on the upturned face; recording the amount of rainfall in Manchester over a period of time.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Sample Space:&amp;#039;&amp;#039;&amp;#039; set of possible outcomes of an experiment; e.g., &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; (or &amp;lt;math&amp;gt;\Omega &amp;lt;/math&amp;gt;) &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{1,2,3,4,5,6\}&amp;lt;/math&amp;gt;, which is the sample space of rolling a dice. Or &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{x;x\geq 0\}&amp;lt;/math&amp;gt;, which is the sample space of an experiment where the outcomes can be any real non-negative number, or ‘&amp;#039;&amp;#039;the set of real non-negative real numbers&amp;#039;&amp;#039;’.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Simple Event&amp;#039;&amp;#039;&amp;#039;: just one of the possible outcomes on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Event:&amp;#039;&amp;#039;&amp;#039; a &amp;#039;&amp;#039;subset&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, denoted &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;; e.g., &amp;lt;math&amp;gt;E=\left\{ 2,4,6\right\}&amp;lt;/math&amp;gt; (i.e. any even number on a dice) or &amp;lt;math&amp;gt;E=\left\{ x;4&amp;lt;x\leq 10\right\} ,&amp;lt;/math&amp;gt; which means ‘&amp;#039;&amp;#039;the set of real numbers which are strictly bigger than&amp;#039;&amp;#039; &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;but less than or equal to &amp;#039;&amp;#039;&amp;lt;math&amp;gt;10&amp;lt;/math&amp;gt;’.&amp;lt;br /&amp;gt;Note that an event, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, is a collection of simple events.&lt;br /&gt;
&lt;br /&gt;
Such concepts can be represented by means of the following Venn Diagram:&lt;br /&gt;
&lt;br /&gt;
[[File:Venn_1.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
The sample space, &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; is depicted as a closed rectangle, and the event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; is a closed loop wholly contained within &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; and we write (in set notation) &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
In dealing with probability, and in particular the probability of an event (or events) occurring, we shall need to be familiar with &amp;#039;&amp;#039;&amp;#039;UNIONS, INTERSECTIONS&amp;#039;&amp;#039;&amp;#039; and &amp;#039;&amp;#039;&amp;#039;COMPLEMENTS&amp;#039;&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
To illustrate these concepts, consider the sample space &amp;lt;math&amp;gt;S=\{x;x\geq 0\},\,&amp;lt;/math&amp;gt; with the following events defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as depicted in Figure 3.2:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E=\{x;4&amp;lt;x\leq 10\},\,F=\{x;7&amp;lt;x\leq 17\},\,G=\{x;x&amp;gt;15\},\,H=\{x;9&amp;lt;x\leq&lt;br /&gt;
13\}. &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| (a) Event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: A closed loop&lt;br /&gt;
| (b) Union: &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2a.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2b.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (c) Intersection: &amp;lt;math&amp;gt;E\cap F&amp;lt;/math&amp;gt;&lt;br /&gt;
| (d) The Null set/event: &amp;lt;math&amp;gt;E\cap G=\emptyset &amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2c.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2d.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (e) Complement of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;&lt;br /&gt;
| (f) Subset of &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;H\subset F&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;H\cap F=H&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2e.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2f.jpg|frameless|300px]]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
        &lt;br /&gt;
&lt;br /&gt;
* The &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cup F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cup F=\{x;4&amp;lt;x\leq 17\};&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are either in &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or in &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) in both. This is illustrated on the Venn diagram by the dark shaded area in diagram (b).&lt;br /&gt;
* The&amp;#039;&amp;#039; intersection&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cap F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cap F=\left\{ x;7\leq x\leq 10\right\} ;&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are common to both &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; Again this is depicted by the dark shaded area in (c). If events have no elements in common (as, for example, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt;) then they are said to be &amp;#039;&amp;#039;mutually exclusive&amp;#039;&amp;#039;, and we can write &amp;lt;math&amp;gt;E\cap G=\emptyset ,&amp;lt;/math&amp;gt; meaning the &amp;#039;&amp;#039;null set&amp;#039;&amp;#039; which contains no elements. Such a situation is illustrated on the Venn Diagram by events (the two shaded closed loops in (d)) which do not overlap. Notice however that &amp;lt;math&amp;gt;G\cap F\neq \emptyset ,&amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have elements in common.&lt;br /&gt;
* The &amp;#039;&amp;#039;complement&amp;#039;&amp;#039; of an event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; say, is everything defined on the sample space which is not in &amp;lt;math&amp;gt;E.&amp;lt;/math&amp;gt; This event is denoted &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;, the dark shaded area in (e); here &amp;lt;math&amp;gt;\bar{E}=\left\{ x;x\leq 4\right\} \cup \left\{ x;x&amp;gt;10\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
* Finally note that &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; is a sub-set of &amp;lt;math&amp;gt;F;&amp;lt;/math&amp;gt; see (f). It is depicted as the dark closed loop wholly contained within &amp;lt;math&amp;gt;F,&amp;lt;/math&amp;gt; the lighter shaded area, so that &amp;lt;math&amp;gt;H\cap F=H;&amp;lt;/math&amp;gt; if an element in the sample space is a member of &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; then it must also be member of &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; (In mathematical logic, we employ this scenario to indicate that “&amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; implies &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;”, but not necessarily vice-versa.) Notice that &amp;lt;math&amp;gt;G\cap H=\emptyset &amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;H\cap E\neq \emptyset&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2848</id>
		<title>Probability Intro</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2848"/>
				<updated>2013-08-07T20:38:24Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: /* Venn diagrams */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Introducing Probability =&lt;br /&gt;
&lt;br /&gt;
So far we have been looking at ways of summarising samples of data drawn from an underlying population of interest. Although at times tedious, all such arithmetic calculations are fairly mechanical and straightforward to apply. To remind ourselves, one of the primary reasons for wishing to summarise data is so assist in the development of inferences about the population from which the data were taken. That is to say, we would like to elicit some information about the mechanism which generated the observed data.&lt;br /&gt;
&lt;br /&gt;
We now start on the process of developing mathematical ways of formulating inferences and this requires the use of &amp;#039;&amp;#039;probability&amp;#039;&amp;#039;. This becomes clear if we think back to one of the early questions posed in this course: &amp;#039;&amp;#039;prior to sampling is it possible to predict with absolute certainty what will be observed&amp;#039;&amp;#039;? The answer to this question is &amp;#039;&amp;#039;no&amp;#039;&amp;#039;; although it would be of interest to know how &amp;#039;&amp;#039;likely&amp;#039;&amp;#039; it is that certain values would be observed. Or, what is the &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; of observing certain values?&lt;br /&gt;
&lt;br /&gt;
Before proceeding, we need some more tools:&lt;br /&gt;
&lt;br /&gt;
= Venn diagrams =&lt;br /&gt;
&lt;br /&gt;
Venn diagrams (and diagrams in general) are of enormous help in trying to understand, and manipulate probability. We begin with some basic definitions, some of which we have encountered before.&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Experiment:&amp;#039;&amp;#039;&amp;#039; any process which, when applied, provides data or an outcome; e.g., rolling a die and observing the number of dots on the upturned face; recording the amount of rainfall in Manchester over a period of time.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Sample Space:&amp;#039;&amp;#039;&amp;#039; set of possible outcomes of an experiment; e.g., &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; (or &amp;lt;math&amp;gt;\Omega &amp;lt;/math&amp;gt;) &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{1,2,3,4,5,6\}&amp;lt;/math&amp;gt;, which is the sample space of rolling a dice. Or &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{x;x\geq 0\}&amp;lt;/math&amp;gt;, which is the sample space of an experiment where the outcomes can be any real non-negative number, or ‘&amp;#039;&amp;#039;the set of real non-negative real numbers&amp;#039;&amp;#039;’.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Simple Event&amp;#039;&amp;#039;&amp;#039;: just one of the possible outcomes on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Event:&amp;#039;&amp;#039;&amp;#039; a &amp;#039;&amp;#039;subset&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, denoted &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;; e.g., &amp;lt;math&amp;gt;E=\left\{ 2,4,6\right\}&amp;lt;/math&amp;gt; (i.e. any even number on a dice) or &amp;lt;math&amp;gt;E=\left\{ x;4&amp;lt;x\leq 10\right\} ,&amp;lt;/math&amp;gt; which means ‘&amp;#039;&amp;#039;the set of real numbers which are strictly bigger than&amp;#039;&amp;#039; &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;but less than or equal to &amp;#039;&amp;#039;&amp;lt;math&amp;gt;10&amp;lt;/math&amp;gt;’.&amp;lt;br /&amp;gt;Note that an event, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, is a collection of simple events.&lt;br /&gt;
&lt;br /&gt;
Such concepts can be represented by means of the following Venn Diagram:&lt;br /&gt;
&lt;br /&gt;
[[File:Venn_1.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
The sample space, &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; is depicted as a closed rectangle, and the event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; is a closed loop wholly contained within &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; and we write (in set notation) &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
In dealing with probability, and in particular the probability of an event (or events) occurring, we shall need to be familiar with &amp;#039;&amp;#039;&amp;#039;UNIONS, INTERSECTIONS&amp;#039;&amp;#039;&amp;#039; and &amp;#039;&amp;#039;&amp;#039;COMPLEMENTS&amp;#039;&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
To illustrate these concepts, consider the sample space &amp;lt;math&amp;gt;S=\{x;x\geq 0\},\,&amp;lt;/math&amp;gt; with the following events defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as depicted in Figure 3.2:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E=\{x;4&amp;lt;x\leq 10\},\,F=\{x;7&amp;lt;x\leq 17\},\,G=\{x;x&amp;gt;15\},\,H=\{x;9&amp;lt;x\leq&lt;br /&gt;
13\}. &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|=&lt;br /&gt;
| (a) Event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: A closed loop&lt;br /&gt;
| (b) Union: &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2a.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2b.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (c) Intersection: &amp;lt;math&amp;gt;E\cap F&amp;lt;/math&amp;gt;&lt;br /&gt;
| (d) The Null set/event: &amp;lt;math&amp;gt;E\cap G=\emptyset &amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2c.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2d.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (e) Complement of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;&lt;br /&gt;
| (f) Subset of &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;H\subset F&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;H\cap F=H&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2e.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2f.jpg|frameless|300px]]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
        &lt;br /&gt;
&lt;br /&gt;
* The &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cup F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cup F=\{x;4&amp;lt;x\leq 17\};&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are either in &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or in &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) in both. This is illustrated on the Venn diagram by the dark shaded area in diagram (b).&lt;br /&gt;
* The&amp;#039;&amp;#039; intersection&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cap F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cap F=\left\{ x;7\leq x\leq 10\right\} ;&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are common to both &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; Again this is depicted by the dark shaded area in (c). If events have no elements in common (as, for example, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt;) then they are said to be &amp;#039;&amp;#039;mutually exclusive&amp;#039;&amp;#039;, and we can write &amp;lt;math&amp;gt;E\cap G=\emptyset ,&amp;lt;/math&amp;gt; meaning the &amp;#039;&amp;#039;null set&amp;#039;&amp;#039; which contains no elements. Such a situation is illustrated on the Venn Diagram by events (the two shaded closed loops in (d)) which do not overlap. Notice however that &amp;lt;math&amp;gt;G\cap F\neq \emptyset ,&amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have elements in common.&lt;br /&gt;
* The &amp;#039;&amp;#039;complement&amp;#039;&amp;#039; of an event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; say, is everything defined on the sample space which is not in &amp;lt;math&amp;gt;E.&amp;lt;/math&amp;gt; This event is denoted &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;, the dark shaded area in (e); here &amp;lt;math&amp;gt;\bar{E}=\left\{ x;x\leq 4\right\} \cup \left\{ x;x&amp;gt;10\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
* Finally note that &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; is a sub-set of &amp;lt;math&amp;gt;F;&amp;lt;/math&amp;gt; see (f). It is depicted as the dark closed loop wholly contained within &amp;lt;math&amp;gt;F,&amp;lt;/math&amp;gt; the lighter shaded area, so that &amp;lt;math&amp;gt;H\cap F=H;&amp;lt;/math&amp;gt; if an element in the sample space is a member of &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; then it must also be member of &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; (In mathematical logic, we employ this scenario to indicate that “&amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; implies &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;”, but not necessarily vice-versa.) Notice that &amp;lt;math&amp;gt;G\cap H=\emptyset &amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;H\cap E\neq \emptyset&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2847</id>
		<title>Probability Intro</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2847"/>
				<updated>2013-08-07T20:38:10Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: /* Venn diagrams */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Introducing Probability =&lt;br /&gt;
&lt;br /&gt;
So far we have been looking at ways of summarising samples of data drawn from an underlying population of interest. Although at times tedious, all such arithmetic calculations are fairly mechanical and straightforward to apply. To remind ourselves, one of the primary reasons for wishing to summarise data is so assist in the development of inferences about the population from which the data were taken. That is to say, we would like to elicit some information about the mechanism which generated the observed data.&lt;br /&gt;
&lt;br /&gt;
We now start on the process of developing mathematical ways of formulating inferences and this requires the use of &amp;#039;&amp;#039;probability&amp;#039;&amp;#039;. This becomes clear if we think back to one of the early questions posed in this course: &amp;#039;&amp;#039;prior to sampling is it possible to predict with absolute certainty what will be observed&amp;#039;&amp;#039;? The answer to this question is &amp;#039;&amp;#039;no&amp;#039;&amp;#039;; although it would be of interest to know how &amp;#039;&amp;#039;likely&amp;#039;&amp;#039; it is that certain values would be observed. Or, what is the &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; of observing certain values?&lt;br /&gt;
&lt;br /&gt;
Before proceeding, we need some more tools:&lt;br /&gt;
&lt;br /&gt;
= Venn diagrams =&lt;br /&gt;
&lt;br /&gt;
Venn diagrams (and diagrams in general) are of enormous help in trying to understand, and manipulate probability. We begin with some basic definitions, some of which we have encountered before.&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Experiment:&amp;#039;&amp;#039;&amp;#039; any process which, when applied, provides data or an outcome; e.g., rolling a die and observing the number of dots on the upturned face; recording the amount of rainfall in Manchester over a period of time.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Sample Space:&amp;#039;&amp;#039;&amp;#039; set of possible outcomes of an experiment; e.g., &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; (or &amp;lt;math&amp;gt;\Omega &amp;lt;/math&amp;gt;) &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{1,2,3,4,5,6\}&amp;lt;/math&amp;gt;, which is the sample space of rolling a dice. Or &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{x;x\geq 0\}&amp;lt;/math&amp;gt;, which is the sample space of an experiment where the outcomes can be any real non-negative number, or ‘&amp;#039;&amp;#039;the set of real non-negative real numbers&amp;#039;&amp;#039;’.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Simple Event&amp;#039;&amp;#039;&amp;#039;: just one of the possible outcomes on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Event:&amp;#039;&amp;#039;&amp;#039; a &amp;#039;&amp;#039;subset&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, denoted &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;; e.g., &amp;lt;math&amp;gt;E=\left\{ 2,4,6\right\}&amp;lt;/math&amp;gt; (i.e. any even number on a dice) or &amp;lt;math&amp;gt;E=\left\{ x;4&amp;lt;x\leq 10\right\} ,&amp;lt;/math&amp;gt; which means ‘&amp;#039;&amp;#039;the set of real numbers which are strictly bigger than&amp;#039;&amp;#039; &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;but less than or equal to &amp;#039;&amp;#039;&amp;lt;math&amp;gt;10&amp;lt;/math&amp;gt;’.&amp;lt;br /&amp;gt;Note that an event, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, is a collection of simple events.&lt;br /&gt;
&lt;br /&gt;
Such concepts can be represented by means of the following Venn Diagram:&lt;br /&gt;
&lt;br /&gt;
[[File:Venn_1.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
The sample space, &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; is depicted as a closed rectangle, and the event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; is a closed loop wholly contained within &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; and we write (in set notation) &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
In dealing with probability, and in particular the probability of an event (or events) occurring, we shall need to be familiar with &amp;#039;&amp;#039;&amp;#039;UNIONS, INTERSECTIONS&amp;#039;&amp;#039;&amp;#039; and &amp;#039;&amp;#039;&amp;#039;COMPLEMENTS&amp;#039;&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
To illustrate these concepts, consider the sample space &amp;lt;math&amp;gt;S=\{x;x\geq 0\},\,&amp;lt;/math&amp;gt; with the following events defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as depicted in Figure 3.2:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E=\{x;4&amp;lt;x\leq 10\},\,F=\{x;7&amp;lt;x\leq 17\},\,G=\{x;x&amp;gt;15\},\,H=\{x;9&amp;lt;x\leq&lt;br /&gt;
13\}. &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|--&lt;br /&gt;
| (a) Event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: A closed loop&lt;br /&gt;
| (b) Union: &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2a.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2b.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (c) Intersection: &amp;lt;math&amp;gt;E\cap F&amp;lt;/math&amp;gt;&lt;br /&gt;
| (d) The Null set/event: &amp;lt;math&amp;gt;E\cap G=\emptyset &amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2c.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2d.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (e) Complement of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;&lt;br /&gt;
| (f) Subset of &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;H\subset F&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;H\cap F=H&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2e.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2f.jpg|frameless|300px]]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
        &lt;br /&gt;
&lt;br /&gt;
* The &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cup F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cup F=\{x;4&amp;lt;x\leq 17\};&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are either in &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or in &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) in both. This is illustrated on the Venn diagram by the dark shaded area in diagram (b).&lt;br /&gt;
* The&amp;#039;&amp;#039; intersection&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cap F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cap F=\left\{ x;7\leq x\leq 10\right\} ;&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are common to both &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; Again this is depicted by the dark shaded area in (c). If events have no elements in common (as, for example, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt;) then they are said to be &amp;#039;&amp;#039;mutually exclusive&amp;#039;&amp;#039;, and we can write &amp;lt;math&amp;gt;E\cap G=\emptyset ,&amp;lt;/math&amp;gt; meaning the &amp;#039;&amp;#039;null set&amp;#039;&amp;#039; which contains no elements. Such a situation is illustrated on the Venn Diagram by events (the two shaded closed loops in (d)) which do not overlap. Notice however that &amp;lt;math&amp;gt;G\cap F\neq \emptyset ,&amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have elements in common.&lt;br /&gt;
* The &amp;#039;&amp;#039;complement&amp;#039;&amp;#039; of an event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; say, is everything defined on the sample space which is not in &amp;lt;math&amp;gt;E.&amp;lt;/math&amp;gt; This event is denoted &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;, the dark shaded area in (e); here &amp;lt;math&amp;gt;\bar{E}=\left\{ x;x\leq 4\right\} \cup \left\{ x;x&amp;gt;10\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
* Finally note that &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; is a sub-set of &amp;lt;math&amp;gt;F;&amp;lt;/math&amp;gt; see (f). It is depicted as the dark closed loop wholly contained within &amp;lt;math&amp;gt;F,&amp;lt;/math&amp;gt; the lighter shaded area, so that &amp;lt;math&amp;gt;H\cap F=H;&amp;lt;/math&amp;gt; if an element in the sample space is a member of &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; then it must also be member of &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; (In mathematical logic, we employ this scenario to indicate that “&amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; implies &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;”, but not necessarily vice-versa.) Notice that &amp;lt;math&amp;gt;G\cap H=\emptyset &amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;H\cap E\neq \emptyset&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=File:Venn_2f.jpg&amp;diff=2846</id>
		<title>File:Venn 2f.jpg</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=File:Venn_2f.jpg&amp;diff=2846"/>
				<updated>2013-08-07T20:36:51Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=File:Venn_2e.jpg&amp;diff=2845</id>
		<title>File:Venn 2e.jpg</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=File:Venn_2e.jpg&amp;diff=2845"/>
				<updated>2013-08-07T20:36:39Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=File:Venn_2d.jpg&amp;diff=2844</id>
		<title>File:Venn 2d.jpg</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=File:Venn_2d.jpg&amp;diff=2844"/>
				<updated>2013-08-07T20:36:27Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=File:Venn_2c.jpg&amp;diff=2843</id>
		<title>File:Venn 2c.jpg</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=File:Venn_2c.jpg&amp;diff=2843"/>
				<updated>2013-08-07T20:36:11Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=File:Venn_2b.jpg&amp;diff=2842</id>
		<title>File:Venn 2b.jpg</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=File:Venn_2b.jpg&amp;diff=2842"/>
				<updated>2013-08-07T20:35:59Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=File:Venn_2a.jpg&amp;diff=2841</id>
		<title>File:Venn 2a.jpg</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=File:Venn_2a.jpg&amp;diff=2841"/>
				<updated>2013-08-07T20:35:46Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2840</id>
		<title>Probability Intro</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2840"/>
				<updated>2013-08-07T20:35:30Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: /* Venn diagrams */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Introducing Probability =&lt;br /&gt;
&lt;br /&gt;
So far we have been looking at ways of summarising samples of data drawn from an underlying population of interest. Although at times tedious, all such arithmetic calculations are fairly mechanical and straightforward to apply. To remind ourselves, one of the primary reasons for wishing to summarise data is so assist in the development of inferences about the population from which the data were taken. That is to say, we would like to elicit some information about the mechanism which generated the observed data.&lt;br /&gt;
&lt;br /&gt;
We now start on the process of developing mathematical ways of formulating inferences and this requires the use of &amp;#039;&amp;#039;probability&amp;#039;&amp;#039;. This becomes clear if we think back to one of the early questions posed in this course: &amp;#039;&amp;#039;prior to sampling is it possible to predict with absolute certainty what will be observed&amp;#039;&amp;#039;? The answer to this question is &amp;#039;&amp;#039;no&amp;#039;&amp;#039;; although it would be of interest to know how &amp;#039;&amp;#039;likely&amp;#039;&amp;#039; it is that certain values would be observed. Or, what is the &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; of observing certain values?&lt;br /&gt;
&lt;br /&gt;
Before proceeding, we need some more tools:&lt;br /&gt;
&lt;br /&gt;
= Venn diagrams =&lt;br /&gt;
&lt;br /&gt;
Venn diagrams (and diagrams in general) are of enormous help in trying to understand, and manipulate probability. We begin with some basic definitions, some of which we have encountered before.&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Experiment:&amp;#039;&amp;#039;&amp;#039; any process which, when applied, provides data or an outcome; e.g., rolling a die and observing the number of dots on the upturned face; recording the amount of rainfall in Manchester over a period of time.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Sample Space:&amp;#039;&amp;#039;&amp;#039; set of possible outcomes of an experiment; e.g., &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; (or &amp;lt;math&amp;gt;\Omega &amp;lt;/math&amp;gt;) &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{1,2,3,4,5,6\}&amp;lt;/math&amp;gt;, which is the sample space of rolling a dice. Or &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{x;x\geq 0\}&amp;lt;/math&amp;gt;, which is the sample space of an experiment where the outcomes can be any real non-negative number, or ‘&amp;#039;&amp;#039;the set of real non-negative real numbers&amp;#039;&amp;#039;’.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Simple Event&amp;#039;&amp;#039;&amp;#039;: just one of the possible outcomes on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Event:&amp;#039;&amp;#039;&amp;#039; a &amp;#039;&amp;#039;subset&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, denoted &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;; e.g., &amp;lt;math&amp;gt;E=\left\{ 2,4,6\right\}&amp;lt;/math&amp;gt; (i.e. any even number on a dice) or &amp;lt;math&amp;gt;E=\left\{ x;4&amp;lt;x\leq 10\right\} ,&amp;lt;/math&amp;gt; which means ‘&amp;#039;&amp;#039;the set of real numbers which are strictly bigger than&amp;#039;&amp;#039; &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;but less than or equal to &amp;#039;&amp;#039;&amp;lt;math&amp;gt;10&amp;lt;/math&amp;gt;’.&amp;lt;br /&amp;gt;Note that an event, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, is a collection of simple events.&lt;br /&gt;
&lt;br /&gt;
Such concepts can be represented by means of the following Venn Diagram:&lt;br /&gt;
&lt;br /&gt;
[[File:Venn_1.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
The sample space, &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; is depicted as a closed rectangle, and the event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; is a closed loop wholly contained within &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; and we write (in set notation) &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
In dealing with probability, and in particular the probability of an event (or events) occurring, we shall need to be familiar with &amp;#039;&amp;#039;&amp;#039;UNIONS, INTERSECTIONS&amp;#039;&amp;#039;&amp;#039; and &amp;#039;&amp;#039;&amp;#039;COMPLEMENTS&amp;#039;&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
To illustrate these concepts, consider the sample space &amp;lt;math&amp;gt;S=\{x;x\geq 0\},\,&amp;lt;/math&amp;gt; with the following events defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as depicted in Figure 3.2:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E=\{x;4&amp;lt;x\leq 10\},\,F=\{x;7&amp;lt;x\leq 17\},\,G=\{x;x&amp;gt;15\},\,H=\{x;9&amp;lt;x\leq&lt;br /&gt;
13\}. &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
| (a) Event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: A closed loop&lt;br /&gt;
| (b) Union: &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2a.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2b.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (c) Intersection: &amp;lt;math&amp;gt;E\cap F&amp;lt;/math&amp;gt;&lt;br /&gt;
| (d) The Null set/event: &amp;lt;math&amp;gt;E\cap G=\emptyset &amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2c.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2d.jpg|frameless|300px]]&lt;br /&gt;
|-&lt;br /&gt;
| (e) Complement of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;&lt;br /&gt;
| (f) Subset of &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;H\subset F&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;H\cap F=H&amp;lt;/math&amp;gt;&lt;br /&gt;
|-&lt;br /&gt;
| [[File:Venn_2e.jpg|frameless|300px]] &lt;br /&gt;
| [[File:Venn_2f.jpg|frameless|300px]]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
        &lt;br /&gt;
&lt;br /&gt;
* The &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cup F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cup F=\{x;4&amp;lt;x\leq 17\};&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are either in &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or in &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) in both. This is illustrated on the Venn diagram by the dark shaded area in diagram (b).&lt;br /&gt;
* The&amp;#039;&amp;#039; intersection&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cap F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cap F=\left\{ x;7\leq x\leq 10\right\} ;&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are common to both &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; Again this is depicted by the dark shaded area in (c). If events have no elements in common (as, for example, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt;) then they are said to be &amp;#039;&amp;#039;mutually exclusive&amp;#039;&amp;#039;, and we can write &amp;lt;math&amp;gt;E\cap G=\emptyset ,&amp;lt;/math&amp;gt; meaning the &amp;#039;&amp;#039;null set&amp;#039;&amp;#039; which contains no elements. Such a situation is illustrated on the Venn Diagram by events (the two shaded closed loops in (d)) which do not overlap. Notice however that &amp;lt;math&amp;gt;G\cap F\neq \emptyset ,&amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have elements in common.&lt;br /&gt;
* The &amp;#039;&amp;#039;complement&amp;#039;&amp;#039; of an event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; say, is everything defined on the sample space which is not in &amp;lt;math&amp;gt;E.&amp;lt;/math&amp;gt; This event is denoted &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;, the dark shaded area in (e); here &amp;lt;math&amp;gt;\bar{E}=\left\{ x;x\leq 4\right\} \cup \left\{ x;x&amp;gt;10\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
* Finally note that &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; is a sub-set of &amp;lt;math&amp;gt;F;&amp;lt;/math&amp;gt; see (f). It is depicted as the dark closed loop wholly contained within &amp;lt;math&amp;gt;F,&amp;lt;/math&amp;gt; the lighter shaded area, so that &amp;lt;math&amp;gt;H\cap F=H;&amp;lt;/math&amp;gt; if an element in the sample space is a member of &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; then it must also be member of &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; (In mathematical logic, we employ this scenario to indicate that “&amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; implies &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;”, but not necessarily vice-versa.) Notice that &amp;lt;math&amp;gt;G\cap H=\emptyset &amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;H\cap E\neq \emptyset&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2839</id>
		<title>Probability Intro</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2839"/>
				<updated>2013-08-07T20:32:46Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: /* Venn diagrams */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Introducing Probability =&lt;br /&gt;
&lt;br /&gt;
So far we have been looking at ways of summarising samples of data drawn from an underlying population of interest. Although at times tedious, all such arithmetic calculations are fairly mechanical and straightforward to apply. To remind ourselves, one of the primary reasons for wishing to summarise data is so assist in the development of inferences about the population from which the data were taken. That is to say, we would like to elicit some information about the mechanism which generated the observed data.&lt;br /&gt;
&lt;br /&gt;
We now start on the process of developing mathematical ways of formulating inferences and this requires the use of &amp;#039;&amp;#039;probability&amp;#039;&amp;#039;. This becomes clear if we think back to one of the early questions posed in this course: &amp;#039;&amp;#039;prior to sampling is it possible to predict with absolute certainty what will be observed&amp;#039;&amp;#039;? The answer to this question is &amp;#039;&amp;#039;no&amp;#039;&amp;#039;; although it would be of interest to know how &amp;#039;&amp;#039;likely&amp;#039;&amp;#039; it is that certain values would be observed. Or, what is the &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; of observing certain values?&lt;br /&gt;
&lt;br /&gt;
Before proceeding, we need some more tools:&lt;br /&gt;
&lt;br /&gt;
= Venn diagrams =&lt;br /&gt;
&lt;br /&gt;
Venn diagrams (and diagrams in general) are of enormous help in trying to understand, and manipulate probability. We begin with some basic definitions, some of which we have encountered before.&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Experiment:&amp;#039;&amp;#039;&amp;#039; any process which, when applied, provides data or an outcome; e.g., rolling a die and observing the number of dots on the upturned face; recording the amount of rainfall in Manchester over a period of time.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Sample Space:&amp;#039;&amp;#039;&amp;#039; set of possible outcomes of an experiment; e.g., &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; (or &amp;lt;math&amp;gt;\Omega &amp;lt;/math&amp;gt;) &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{1,2,3,4,5,6\}&amp;lt;/math&amp;gt;, which is the sample space of rolling a dice. Or &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{x;x\geq 0\}&amp;lt;/math&amp;gt;, which is the sample space of an experiment where the outcomes can be any real non-negative number, or ‘&amp;#039;&amp;#039;the set of real non-negative real numbers&amp;#039;&amp;#039;’.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Simple Event&amp;#039;&amp;#039;&amp;#039;: just one of the possible outcomes on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Event:&amp;#039;&amp;#039;&amp;#039; a &amp;#039;&amp;#039;subset&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, denoted &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;; e.g., &amp;lt;math&amp;gt;E=\left\{ 2,4,6\right\}&amp;lt;/math&amp;gt; (i.e. any even number on a dice) or &amp;lt;math&amp;gt;E=\left\{ x;4&amp;lt;x\leq 10\right\} ,&amp;lt;/math&amp;gt; which means ‘&amp;#039;&amp;#039;the set of real numbers which are strictly bigger than&amp;#039;&amp;#039; &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;but less than or equal to &amp;#039;&amp;#039;&amp;lt;math&amp;gt;10&amp;lt;/math&amp;gt;’.&amp;lt;br /&amp;gt;Note that an event, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, is a collection of simple events.&lt;br /&gt;
&lt;br /&gt;
Such concepts can be represented by means of the following Venn Diagram:&lt;br /&gt;
&lt;br /&gt;
[[File:Venn_1.jpg|frameless|600px]]&lt;br /&gt;
&lt;br /&gt;
The sample space, &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; is depicted as a closed rectangle, and the event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; is a closed loop wholly contained within &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; and we write (in set notation) &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
In dealing with probability, and in particular the probability of an event (or events) occurring, we shall need to be familiar with &amp;#039;&amp;#039;&amp;#039;UNIONS, INTERSECTIONS&amp;#039;&amp;#039;&amp;#039; and &amp;#039;&amp;#039;&amp;#039;COMPLEMENTS&amp;#039;&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
To illustrate these concepts, consider the sample space &amp;lt;math&amp;gt;S=\{x;x\geq 0\},\,&amp;lt;/math&amp;gt; with the following events defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as depicted in Figure 3.2:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E=\{x;4&amp;lt;x\leq 10\},\,F=\{x;7&amp;lt;x\leq 17\},\,G=\{x;x&amp;gt;15\},\,H=\{x;9&amp;lt;x\leq&lt;br /&gt;
13\}. &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
PUT THIS IN WIKITABLE!&lt;br /&gt;
&lt;br /&gt;
(a) Event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: A closed loop (b) Union: &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt; [[File:Venn_2a.jpg|frameless|300px]] [[File:Venn_2b.jpg|frameless|300px]] (c) Intersection: &amp;lt;math&amp;gt;E\cap F&amp;lt;/math&amp;gt; (d) The Null set/event: &amp;lt;math&amp;gt;E\cap G=\emptyset &amp;lt;/math&amp;gt; [[File:Venn_2c.jpg|frameless|300px]] [[File:Venn_2d.jpg|frameless|300px]] (e) Complement of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt; (f) Subset of &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;H\subset F&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;H\cap F=H&amp;lt;/math&amp;gt; [[File:Venn_2e.jpg|frameless|300px]] [[File:Venn_2f.jpg|frameless|300px]]&lt;br /&gt;
&lt;br /&gt;
* The &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cup F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cup F=\{x;4&amp;lt;x\leq 17\};&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are either in &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or in &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) in both. This is illustrated on the Venn diagram by the dark shaded area in diagram (b).&lt;br /&gt;
* The&amp;#039;&amp;#039; intersection&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cap F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cap F=\left\{ x;7\leq x\leq 10\right\} ;&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are common to both &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; Again this is depicted by the dark shaded area in (c). If events have no elements in common (as, for example, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt;) then they are said to be &amp;#039;&amp;#039;mutually exclusive&amp;#039;&amp;#039;, and we can write &amp;lt;math&amp;gt;E\cap G=\emptyset ,&amp;lt;/math&amp;gt; meaning the &amp;#039;&amp;#039;null set&amp;#039;&amp;#039; which contains no elements. Such a situation is illustrated on the Venn Diagram by events (the two shaded closed loops in (d)) which do not overlap. Notice however that &amp;lt;math&amp;gt;G\cap F\neq \emptyset ,&amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have elements in common.&lt;br /&gt;
* The &amp;#039;&amp;#039;complement&amp;#039;&amp;#039; of an event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; say, is everything defined on the sample space which is not in &amp;lt;math&amp;gt;E.&amp;lt;/math&amp;gt; This event is denoted &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;, the dark shaded area in (e); here &amp;lt;math&amp;gt;\bar{E}=\left\{ x;x\leq 4\right\} \cup \left\{ x;x&amp;gt;10\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
* Finally note that &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; is a sub-set of &amp;lt;math&amp;gt;F;&amp;lt;/math&amp;gt; see (f). It is depicted as the dark closed loop wholly contained within &amp;lt;math&amp;gt;F,&amp;lt;/math&amp;gt; the lighter shaded area, so that &amp;lt;math&amp;gt;H\cap F=H;&amp;lt;/math&amp;gt; if an element in the sample space is a member of &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; then it must also be member of &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; (In mathematical logic, we employ this scenario to indicate that “&amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; implies &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;”, but not necessarily vice-versa.) Notice that &amp;lt;math&amp;gt;G\cap H=\emptyset &amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;H\cap E\neq \emptyset&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=File:Venn_1.jpg&amp;diff=2838</id>
		<title>File:Venn 1.jpg</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=File:Venn_1.jpg&amp;diff=2838"/>
				<updated>2013-08-07T20:31:41Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2837</id>
		<title>Probability Intro</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2837"/>
				<updated>2013-08-07T20:31:29Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: /* Venn diagrams */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Introducing Probability =&lt;br /&gt;
&lt;br /&gt;
So far we have been looking at ways of summarising samples of data drawn from an underlying population of interest. Although at times tedious, all such arithmetic calculations are fairly mechanical and straightforward to apply. To remind ourselves, one of the primary reasons for wishing to summarise data is so assist in the development of inferences about the population from which the data were taken. That is to say, we would like to elicit some information about the mechanism which generated the observed data.&lt;br /&gt;
&lt;br /&gt;
We now start on the process of developing mathematical ways of formulating inferences and this requires the use of &amp;#039;&amp;#039;probability&amp;#039;&amp;#039;. This becomes clear if we think back to one of the early questions posed in this course: &amp;#039;&amp;#039;prior to sampling is it possible to predict with absolute certainty what will be observed&amp;#039;&amp;#039;? The answer to this question is &amp;#039;&amp;#039;no&amp;#039;&amp;#039;; although it would be of interest to know how &amp;#039;&amp;#039;likely&amp;#039;&amp;#039; it is that certain values would be observed. Or, what is the &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; of observing certain values?&lt;br /&gt;
&lt;br /&gt;
Before proceeding, we need some more tools:&lt;br /&gt;
&lt;br /&gt;
= Venn diagrams =&lt;br /&gt;
&lt;br /&gt;
Venn diagrams (and diagrams in general) are of enormous help in trying to understand, and manipulate probability. We begin with some basic definitions, some of which we have encountered before.&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Experiment:&amp;#039;&amp;#039;&amp;#039; any process which, when applied, provides data or an outcome; e.g., rolling a die and observing the number of dots on the upturned face; recording the amount of rainfall in Manchester over a period of time.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Sample Space:&amp;#039;&amp;#039;&amp;#039; set of possible outcomes of an experiment; e.g., &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; (or &amp;lt;math&amp;gt;\Omega &amp;lt;/math&amp;gt;) &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{1,2,3,4,5,6\}&amp;lt;/math&amp;gt;, which is the sample space of rolling a dice. Or &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{x;x\geq 0\}&amp;lt;/math&amp;gt;, which is the sample space of an experiment where the outcomes can be any real non-negative number, or ‘&amp;#039;&amp;#039;the set of real non-negative real numbers&amp;#039;&amp;#039;’.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Simple Event&amp;#039;&amp;#039;&amp;#039;: just one of the possible outcomes on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Event:&amp;#039;&amp;#039;&amp;#039; a &amp;#039;&amp;#039;subset&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, denoted &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;; e.g., &amp;lt;math&amp;gt;E=\left\{ 2,4,6\right\}&amp;lt;/math&amp;gt; (i.e. any even number on a dice) or &amp;lt;math&amp;gt;E=\left\{ x;4&amp;lt;x\leq 10\right\} ,&amp;lt;/math&amp;gt; which means ‘&amp;#039;&amp;#039;the set of real numbers which are strictly bigger than&amp;#039;&amp;#039; &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;but less than or equal to &amp;#039;&amp;#039;&amp;lt;math&amp;gt;10&amp;lt;/math&amp;gt;’.&amp;lt;br /&amp;gt;Note that an event, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, is a collection of simple events.&lt;br /&gt;
&lt;br /&gt;
Such concepts can be represented by means of the following Venn Diagram:&lt;br /&gt;
&lt;br /&gt;
[[File:Venn_1.jpg|frameless|750px]]&lt;br /&gt;
&lt;br /&gt;
The sample space, &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; is depicted as a closed rectangle, and the event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; is a closed loop wholly contained within &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; and we write (in set notation) &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
In dealing with probability, and in particular the probability of an event (or events) occurring, we shall need to be familiar with &amp;#039;&amp;#039;&amp;#039;UNIONS, INTERSECTIONS&amp;#039;&amp;#039;&amp;#039; and &amp;#039;&amp;#039;&amp;#039;COMPLEMENTS&amp;#039;&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
To illustrate these concepts, consider the sample space &amp;lt;math&amp;gt;S=\{x;x\geq 0\},\,&amp;lt;/math&amp;gt; with the following events defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as depicted in Figure 3.2:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E=\{x;4&amp;lt;x\leq 10\},\,F=\{x;7&amp;lt;x\leq 17\},\,G=\{x;x&amp;gt;15\},\,H=\{x;9&amp;lt;x\leq&lt;br /&gt;
13\}. &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
PUT THIS IN WIKITABLE!&lt;br /&gt;
&lt;br /&gt;
(a) Event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: A closed loop (b) Union: &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt; [[File:Venn_2a.jpg|frameless|750px]] [[File:Venn_2b.jpg|frameless|750px]] (c) Intersection: &amp;lt;math&amp;gt;E\cap F&amp;lt;/math&amp;gt; (d) The Null set/event: &amp;lt;math&amp;gt;E\cap G=\emptyset &amp;lt;/math&amp;gt; [[File:Venn_2c.jpg|frameless|750px]] [[File:Venn_2d.jpg|frameless|750px]] (e) Complement of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt; (f) Subset of &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;H\subset F&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;H\cap F=H&amp;lt;/math&amp;gt; [[File:Venn_2e.jpg|frameless|750px]] [[File:Venn_2f.jpg|frameless|750px]]&lt;br /&gt;
&lt;br /&gt;
* The &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cup F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cup F=\{x;4&amp;lt;x\leq 17\};&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are either in &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or in &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) in both. This is illustrated on the Venn diagram by the dark shaded area in diagram (b).&lt;br /&gt;
* The&amp;#039;&amp;#039; intersection&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cap F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cap F=\left\{ x;7\leq x\leq 10\right\} ;&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are common to both &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; Again this is depicted by the dark shaded area in (c). If events have no elements in common (as, for example, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt;) then they are said to be &amp;#039;&amp;#039;mutually exclusive&amp;#039;&amp;#039;, and we can write &amp;lt;math&amp;gt;E\cap G=\emptyset ,&amp;lt;/math&amp;gt; meaning the &amp;#039;&amp;#039;null set&amp;#039;&amp;#039; which contains no elements. Such a situation is illustrated on the Venn Diagram by events (the two shaded closed loops in (d)) which do not overlap. Notice however that &amp;lt;math&amp;gt;G\cap F\neq \emptyset ,&amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have elements in common.&lt;br /&gt;
* The &amp;#039;&amp;#039;complement&amp;#039;&amp;#039; of an event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; say, is everything defined on the sample space which is not in &amp;lt;math&amp;gt;E.&amp;lt;/math&amp;gt; This event is denoted &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;, the dark shaded area in (e); here &amp;lt;math&amp;gt;\bar{E}=\left\{ x;x\leq 4\right\} \cup \left\{ x;x&amp;gt;10\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
* Finally note that &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; is a sub-set of &amp;lt;math&amp;gt;F;&amp;lt;/math&amp;gt; see (f). It is depicted as the dark closed loop wholly contained within &amp;lt;math&amp;gt;F,&amp;lt;/math&amp;gt; the lighter shaded area, so that &amp;lt;math&amp;gt;H\cap F=H;&amp;lt;/math&amp;gt; if an element in the sample space is a member of &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; then it must also be member of &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; (In mathematical logic, we employ this scenario to indicate that “&amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; implies &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;”, but not necessarily vice-versa.) Notice that &amp;lt;math&amp;gt;G\cap H=\emptyset &amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;H\cap E\neq \emptyset&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2836</id>
		<title>Probability Intro</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Probability_Intro&amp;diff=2836"/>
				<updated>2013-08-07T20:27:23Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: Created page with &amp;quot;  = Introducing Probability =  So far we have been looking at ways of summarising samples of data drawn from an underlying population of interest. Although at times tedious, a...&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
= Introducing Probability =&lt;br /&gt;
&lt;br /&gt;
So far we have been looking at ways of summarising samples of data drawn from an underlying population of interest. Although at times tedious, all such arithmetic calculations are fairly mechanical and straightforward to apply. To remind ourselves, one of the primary reasons for wishing to summarise data is so assist in the development of inferences about the population from which the data were taken. That is to say, we would like to elicit some information about the mechanism which generated the observed data.&lt;br /&gt;
&lt;br /&gt;
We now start on the process of developing mathematical ways of formulating inferences and this requires the use of &amp;#039;&amp;#039;probability&amp;#039;&amp;#039;. This becomes clear if we think back to one of the early questions posed in this course: &amp;#039;&amp;#039;prior to sampling is it possible to predict with absolute certainty what will be observed&amp;#039;&amp;#039;? The answer to this question is &amp;#039;&amp;#039;no&amp;#039;&amp;#039;; although it would be of interest to know how &amp;#039;&amp;#039;likely&amp;#039;&amp;#039; it is that certain values would be observed. Or, what is the &amp;#039;&amp;#039;probability&amp;#039;&amp;#039; of observing certain values?&lt;br /&gt;
&lt;br /&gt;
Before proceeding, we need some more tools:&lt;br /&gt;
&lt;br /&gt;
= Venn diagrams =&lt;br /&gt;
&lt;br /&gt;
Venn diagrams (and diagrams in general) are of enormous help in trying to understand, and manipulate probability. We begin with some basic definitions, some of which we have encountered before.&lt;br /&gt;
&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Experiment:&amp;#039;&amp;#039;&amp;#039; any process which, when applied, provides data or an outcome; e.g., rolling a die and observing the number of dots on the upturned face; recording the amount of rainfall in Manchester over a period of time.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Sample Space:&amp;#039;&amp;#039;&amp;#039; set of possible outcomes of an experiment; e.g., &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; (or &amp;lt;math&amp;gt;\Omega &amp;lt;/math&amp;gt;) &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{1,2,3,4,5,6\}&amp;lt;/math&amp;gt;, which is the sample space of rolling a dice. Or &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;=&amp;lt;/math&amp;gt; &amp;lt;math&amp;gt;\{x;x\geq 0\}&amp;lt;/math&amp;gt;, which is the sample space of an experiment where the outcomes can be any real non-negative number, or ‘&amp;#039;&amp;#039;the set of real non-negative real numbers&amp;#039;&amp;#039;’.&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Simple Event&amp;#039;&amp;#039;&amp;#039;: just one of the possible outcomes on &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;&lt;br /&gt;
* &amp;#039;&amp;#039;&amp;#039;Event:&amp;#039;&amp;#039;&amp;#039; a &amp;#039;&amp;#039;subset&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt;, denoted &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;; e.g., &amp;lt;math&amp;gt;E=\left\{ 2,4,6\right\}&amp;lt;/math&amp;gt; (i.e. any even number on a dice) or &amp;lt;math&amp;gt;E=\left\{ x;4&amp;lt;x\leq 10\right\} ,&amp;lt;/math&amp;gt; which means ‘&amp;#039;&amp;#039;the set of real numbers which are strictly bigger than&amp;#039;&amp;#039; &amp;lt;math&amp;gt;4&amp;lt;/math&amp;gt; &amp;#039;&amp;#039;but less than or equal to &amp;#039;&amp;#039;&amp;lt;math&amp;gt;10&amp;lt;/math&amp;gt;’.&amp;lt;br /&amp;gt;Note that an event, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;, is a collection of simple events.&lt;br /&gt;
&lt;br /&gt;
Such concepts can be represented by means of the following Venn Diagram:&lt;br /&gt;
&lt;br /&gt;
[[File:Venn_1.wmf|frameless|750px]]&lt;br /&gt;
&lt;br /&gt;
The sample space, &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; is depicted as a closed rectangle, and the event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; is a closed loop wholly contained within &amp;lt;math&amp;gt;S&amp;lt;/math&amp;gt; and we write (in set notation) &amp;lt;math&amp;gt;E\subset S&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
In dealing with probability, and in particular the probability of an event (or events) occurring, we shall need to be familiar with &amp;#039;&amp;#039;&amp;#039;UNIONS, INTERSECTIONS&amp;#039;&amp;#039;&amp;#039; and &amp;#039;&amp;#039;&amp;#039;COMPLEMENTS&amp;#039;&amp;#039;&amp;#039;.&lt;br /&gt;
&lt;br /&gt;
To illustrate these concepts, consider the sample space &amp;lt;math&amp;gt;S=\{x;x\geq 0\},\,&amp;lt;/math&amp;gt; with the following events defined on &amp;lt;math&amp;gt;S,&amp;lt;/math&amp;gt; as depicted in Figure 3.2:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;math&amp;gt;E=\{x;4&amp;lt;x\leq 10\},\,F=\{x;7&amp;lt;x\leq 17\},\,G=\{x;x&amp;gt;15\},\,H=\{x;9&amp;lt;x\leq&lt;br /&gt;
13\}. &amp;lt;/math&amp;gt;&lt;br /&gt;
&lt;br /&gt;
PUT THIS IN WIKITABLE!&lt;br /&gt;
&lt;br /&gt;
(a) Event &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: A closed loop (b) Union: &amp;lt;math&amp;gt;E\cup F&amp;lt;/math&amp;gt; [[File:Venn_2a.wmf|frameless|750px]] [[File:Venn_2b.wmf|frameless|750px]] (c) Intersection: &amp;lt;math&amp;gt;E\cap F&amp;lt;/math&amp;gt; (d) The Null set/event: &amp;lt;math&amp;gt;E\cap G=\emptyset &amp;lt;/math&amp;gt; [[File:Venn_2c.wmf|frameless|750px]] [[File:Venn_2d.wmf|frameless|750px]] (e) Complement of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt; (f) Subset of &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;: &amp;lt;math&amp;gt;H\subset F&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;H\cap F=H&amp;lt;/math&amp;gt; [[File:Venn_2e.wmf|frameless|750px]] [[File:Venn_2f.wmf|frameless|750px]]&lt;br /&gt;
&lt;br /&gt;
* The &amp;#039;&amp;#039;union&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cup F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cup F=\{x;4&amp;lt;x\leq 17\};&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are either in &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; or in &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; or (perhaps) in both. This is illustrated on the Venn diagram by the dark shaded area in diagram (b).&lt;br /&gt;
* The&amp;#039;&amp;#039; intersection&amp;#039;&amp;#039; of &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; is denoted &amp;lt;math&amp;gt;E\cap F,&amp;lt;/math&amp;gt; with &amp;lt;math&amp;gt;E\cap F=\left\{ x;7\leq x\leq 10\right\} ;&amp;lt;/math&amp;gt; i.e., it contains elements (simple events) which are common to both &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; Again this is depicted by the dark shaded area in (c). If events have no elements in common (as, for example, &amp;lt;math&amp;gt;E&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt;) then they are said to be &amp;#039;&amp;#039;mutually exclusive&amp;#039;&amp;#039;, and we can write &amp;lt;math&amp;gt;E\cap G=\emptyset ,&amp;lt;/math&amp;gt; meaning the &amp;#039;&amp;#039;null set&amp;#039;&amp;#039; which contains no elements. Such a situation is illustrated on the Venn Diagram by events (the two shaded closed loops in (d)) which do not overlap. Notice however that &amp;lt;math&amp;gt;G\cap F\neq \emptyset ,&amp;lt;/math&amp;gt; since &amp;lt;math&amp;gt;G&amp;lt;/math&amp;gt; and &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt; have elements in common.&lt;br /&gt;
* The &amp;#039;&amp;#039;complement&amp;#039;&amp;#039; of an event &amp;lt;math&amp;gt;E,&amp;lt;/math&amp;gt; say, is everything defined on the sample space which is not in &amp;lt;math&amp;gt;E.&amp;lt;/math&amp;gt; This event is denoted &amp;lt;math&amp;gt;\bar{E}&amp;lt;/math&amp;gt;, the dark shaded area in (e); here &amp;lt;math&amp;gt;\bar{E}=\left\{ x;x\leq 4\right\} \cup \left\{ x;x&amp;gt;10\right\}&amp;lt;/math&amp;gt;.&lt;br /&gt;
* Finally note that &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; is a sub-set of &amp;lt;math&amp;gt;F;&amp;lt;/math&amp;gt; see (f). It is depicted as the dark closed loop wholly contained within &amp;lt;math&amp;gt;F,&amp;lt;/math&amp;gt; the lighter shaded area, so that &amp;lt;math&amp;gt;H\cap F=H;&amp;lt;/math&amp;gt; if an element in the sample space is a member of &amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; then it must also be member of &amp;lt;math&amp;gt;F.&amp;lt;/math&amp;gt; (In mathematical logic, we employ this scenario to indicate that “&amp;lt;math&amp;gt;H&amp;lt;/math&amp;gt; implies &amp;lt;math&amp;gt;F&amp;lt;/math&amp;gt;”, but not necessarily vice-versa.) Notice that &amp;lt;math&amp;gt;G\cap H=\emptyset &amp;lt;/math&amp;gt; but &amp;lt;math&amp;gt;H\cap E\neq \emptyset&amp;lt;/math&amp;gt;.&lt;br /&gt;
&lt;br /&gt;
= Footnotes =&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=MediaWiki:CustomBlockMainNav&amp;diff=2835</id>
		<title>MediaWiki:CustomBlockMainNav</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=MediaWiki:CustomBlockMainNav&amp;diff=2835"/>
				<updated>2013-08-07T19:44:03Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: Created page with &amp;quot;&amp;lt;!-- Vergiss nicht VoWi:Navigation anzupassen, wenn du diese Seite aenderst! --&amp;gt; * Home ** MATLAB ** EVIEWS ** Statistics&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&amp;lt;!-- Vergiss nicht VoWi:Navigation anzupassen, wenn du diese Seite aenderst! --&amp;gt;&lt;br /&gt;
* [[Main_Page|Home]]&lt;br /&gt;
** [[MATLAB|MATLAB]]&lt;br /&gt;
** [[EVIEWS|EVIEWS]]&lt;br /&gt;
** [[Statistics|Statistics]]&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=MediaWiki:CustomNavBlocks&amp;diff=2834</id>
		<title>MediaWiki:CustomNavBlocks</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=MediaWiki:CustomNavBlocks&amp;diff=2834"/>
				<updated>2013-08-07T19:42:29Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;CustomBlockMainNav|ECLR&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=MediaWiki:CustomNavBlocks&amp;diff=2833</id>
		<title>MediaWiki:CustomNavBlocks</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=MediaWiki:CustomNavBlocks&amp;diff=2833"/>
				<updated>2013-08-07T19:42:12Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: Created page with &amp;quot;CustomBlockMainNav|ECLR Home&amp;quot;&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;CustomBlockMainNav|ECLR Home&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Main_Page&amp;diff=2829</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Main_Page&amp;diff=2829"/>
				<updated>2013-08-07T13:11:35Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
== Econometric Computing Learning Resource (ECLR) ==&lt;br /&gt;
&lt;br /&gt;
This is the home of the ECLR. It is the purpose of this resource to facilitate the application of Econometric techniques. There exist numerous software packages that can be used to solve Econometric problems. Some of them are menu driven (e.g. EVIEWS) and are great to tackle standard econometric problems. Others require more programming but therefore also allow the user to tackle non-standard problems. This resource provides support material for the use of [[EVIEWS]] and [[MATLAB]].&lt;br /&gt;
&lt;br /&gt;
This page will not really teach Econometrics, although it will provide sufficient econometric background to facilitate the implementation of the different econometric techniques. To do any serious econometrics you need to have some basic statistics knowledge. On this page [[Statistics]] you can review some of the required statistical background.&lt;br /&gt;
&lt;br /&gt;
{| class=&amp;quot;wikitable&amp;quot;&lt;br /&gt;
|-&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Programming&amp;lt;br&amp;gt; Languages&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| &lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Menu-driven&amp;lt;br&amp;gt;software&lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| &lt;br /&gt;
! scope=&amp;quot;col&amp;quot;| Statistics&amp;lt;br&amp;gt;Introduction&lt;br /&gt;
|-&lt;br /&gt;
| [[MATLAB|MATLAB]]&lt;br /&gt;
| Python (to come)&lt;br /&gt;
| [[EVIEWS|EVIEWS]]&lt;br /&gt;
| &lt;br /&gt;
| [[Statistics|Statistics Intro]]&lt;br /&gt;
|}&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Should you learn programming to do Econometrics? ==&lt;br /&gt;
First, the answer to this question is often no. If you need to apply a very standard econometric technique and don&amp;#039;t have to do that too often, then it may well be that a menu driven software package (like EVIEWS) is exactly the right way to go.&lt;br /&gt;
&lt;br /&gt;
However, if you want to achieve nonstandard things (and working on a PhD or even a MSc you are bound to want to do non-standard things) the picture is very different. A menu driven software package may do some related things but often not exactly what you want. Even if the software package has a &amp;quot;button&amp;quot; to do what you want to do, it is often badly documented what exactly happens underneath the hood. But most importantly, many nonstandard things just cannot be done in menu driven software. This is were MATLAB (or other software packages like GAUSS or OX) comes to the rescue. At the core it has nothing to do with econometrics. It is a matrix algebra maths programme in which you can do with data (stored in matrices) whatever you want. When we say whatever, then this is to be understood almost literally. And that is the point. &lt;br /&gt;
&lt;br /&gt;
Therefore, there are powerful reasons why you may want to learn to use a proper programming language to solve your econometric problem:&lt;br /&gt;
&lt;br /&gt;
# You can easily repeat the same analysis with many more dataset.&lt;br /&gt;
# For some analysis there are some very repetitive things you got to do. In some cases, writing a program will eliminate the need to do essentially the same thing many times. You instruct the software to do the repetitive work for you.&lt;br /&gt;
# By writing a MATLAB program (or script) you essentially also create a record of what you do. This is extremely useful when it comes to understanding previous work and to identify mistakes. This is possibly the most important advantage of writing your own programs.&lt;br /&gt;
# By having to instruct your computer what to actually do, you properly learn the underlying Econometrics. In other words, programming can be an awesome econometric learning tool.&lt;br /&gt;
&lt;br /&gt;
If you did any programming in any other programming language, such as Visual Basic, C, C++, GAUSS, etc. you will recognise a lot of common patterns. Learning how to use MATLAB to do econometrics is 100% a trial and error process and hence this wiki will provide guidance but you will have to practice yourself.&lt;br /&gt;
&lt;br /&gt;
== Authors, Maintenance and Contributions ==&lt;br /&gt;
&lt;br /&gt;
This wiki was created by [mailto:ralf.becker@manchester.ac.uk Ralf Becker] and [mailto:arthur.sinko@manchester.ac.uk Arthur Sinko] with the financial support of a University of Manchester Investing in Success grant. If you have any suggestions please contact us by email. Contributions to this wiki are encouraged. Please contact us for details.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Help for using WikiMedia ==&lt;br /&gt;
* [http://www.mediawiki.org/wiki/Manual:MediaWiki - FAQ]&lt;br /&gt;
* [http://www.mediawiki.org/wiki/Manual:MediaWiki - Configuration Settings]&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	<entry>
		<id>http://eclr.humanities.manchester.ac.uk/index.php?title=Main_Page&amp;diff=2828</id>
		<title>Main Page</title>
		<link rel="alternate" type="text/html" href="http://eclr.humanities.manchester.ac.uk/index.php?title=Main_Page&amp;diff=2828"/>
				<updated>2013-08-07T13:07:04Z</updated>
		
		<summary type="html">&lt;p&gt;Admin: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;&lt;br /&gt;
&lt;br /&gt;
== Econometric Computing Learning Resource (ECLR) ==&lt;br /&gt;
&lt;br /&gt;
This is the home of the ECLR. It is the purpose of this resource to facilitate the application of Econometric techniques. There exist numerous software packages that can be used to solve Econometric problems. Some of them are menu driven (e.g. EVIEWS) and are great to tackle standard econometric problems. Others require more programming but therefore also allow the user to tackle non-standard problems. This resource provides support material for the use of [[EVIEWS]] and [[MATLAB]].&lt;br /&gt;
&lt;br /&gt;
This page will not really teach Econometrics, although it will provide sufficient econometric background to facilitate the implementation of the different econometric techniques. To do any serious econometrics you need to have some basic statistics knowledge. On this page [[Statistics]] you can review some of the required statistical background.&lt;br /&gt;
&lt;br /&gt;
== Should you learn programming to do Econometrics? ==&lt;br /&gt;
First, the answer to this question is often no. If you need to apply a very standard econometric technique and don&amp;#039;t have to do that too often, then it may well be that a menu driven software package (like EVIEWS) is exactly the right way to go.&lt;br /&gt;
&lt;br /&gt;
However, if you want to achieve nonstandard things (and working on a PhD or even a MSc you are bound to want to do non-standard things) the picture is very different. A menu driven software package may do some related things but often not exactly what you want. Even if the software package has a &amp;quot;button&amp;quot; to do what you want to do, it is often badly documented what exactly happens underneath the hood. But most importantly, many nonstandard things just cannot be done in menu driven software. This is were MATLAB (or other software packages like GAUSS or OX) comes to the rescue. At the core it has nothing to do with econometrics. It is a matrix algebra maths programme in which you can do with data (stored in matrices) whatever you want. When we say whatever, then this is to be understood almost literally. And that is the point. &lt;br /&gt;
&lt;br /&gt;
Therefore, there are powerful reasons why you may want to learn to use a proper programming language to solve your econometric problem:&lt;br /&gt;
&lt;br /&gt;
# You can easily repeat the same analysis with many more dataset.&lt;br /&gt;
# For some analysis there are some very repetitive things you got to do. In some cases, writing a program will eliminate the need to do essentially the same thing many times. You instruct the software to do the repetitive work for you.&lt;br /&gt;
# By writing a MATLAB program (or script) you essentially also create a record of what you do. This is extremely useful when it comes to understanding previous work and to identify mistakes. This is possibly the most important advantage of writing your own programs.&lt;br /&gt;
# By having to instruct your computer what to actually do, you properly learn the underlying Econometrics. In other words, programming can be an awesome econometric learning tool.&lt;br /&gt;
&lt;br /&gt;
If you did any programming in any other programming language, such as Visual Basic, C, C++, GAUSS, etc. you will recognise a lot of common patterns. Learning how to use MATLAB to do econometrics is 100% a trial and error process and hence this wiki will provide guidance but you will have to practice yourself.&lt;br /&gt;
&lt;br /&gt;
== Authors, Maintenance and Contributions ==&lt;br /&gt;
&lt;br /&gt;
This wiki was created by [mailto:ralf.becker@manchester.ac.uk Ralf Becker] and [mailto:arthur.sinko@manchester.ac.uk Arthur Sinko] with the financial support of a University of Manchester Investing in Success grant. If you have any suggestions please contact us by email. Contributions to this wiki are encouraged. Please contact us for details.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Help for using WikiMedia ==&lt;br /&gt;
* [http://www.mediawiki.org/wiki/Manual:MediaWiki - FAQ]&lt;br /&gt;
* [http://www.mediawiki.org/wiki/Manual:MediaWiki - Configuration Settings]&lt;/div&gt;</summary>
		<author><name>Admin</name></author>	</entry>

	</feed>