Poker News and Poker Games Card Game Strategy Guide for Players |
Poker is a great game for players of all ages. Become a better player by learning about the history of the game, the players, and major tournaments. Interesting facts, trivia, and tidbits about the game. Practice your skills online in free games and use the interactive trainer to improve your game at Party Poker and play for great prizes, cash, tournament entries, or even just for the fun of it! It's free to sign up and start playing today!
Bayes' theoremBayes' theorem is a result in probability theory, which gives the conditional probability distribution of a random variable A given B in terms of the conditional probability distribution of variable B given A and the marginal probability distribution of A alone.In the context of Bayesian probability theory and of A alone is usually called the prior probability distribution or simply the prior. The conditional distribution of A given the "data" B is called the posterior probability distribution or just the posterior. As a mathematical theorem, Bayes' theorem is valid regardless of whether one adopts a frequentist or a Bayesian interpretation of probability. However, there is disagreement as to what kinds of variables can be substituted for A and B in the theorem; this topic is treated at greater length in the articles on Bayesian probability and frequentist probability. Historical remarksBayes' theorem is named after the Reverend Thomas Bayes (1702–61). Bayes worked on the problem of computing a distribution for the parameter of a binomial distribution (to use modern terminology); his work was edited and presented posthumously (1763) by his friend Richard Price, in An Essay towards solving a Problem in the Doctrine of Chances. Bayes' results were replicated and extended by Laplace in an essay of 1774, who apparently was not aware of Bayes' work. One of Bayes' results (Proposition 5) gives a simple description of conditional probability, and shows that it does not depend on the order in which things occur:
The main result (Proposition 9 in the essay) derived by Bayes is the following: assuming a uniform distribution for the prior distribution of the binomial parameter p, the probability that p is between two values a and b is where m is the number of observed successes and n the number of observed failures. His preliminary results, in particular Propositions 3, 4, and 5, imply the result now called Bayes' Theorem (as described below), but it does not appear that Bayes himself emphasized or focused on that result. What is "Bayesian" about Proposition 9 is that Bayes presented it as a probability for the parameter p. That is, not only can one compute probabilities for experimental outcomes, but also for the parameter which governs them, and the same algebra is used to make inferences of either kind. Interestingly, Bayes actually states his question in a way that might make the idea of assigning a probability distribution to a parameter palatable to a frequentist. He supposes that a billiard ball is thrown at random onto a billiard table, and that the probabilities p and q are the probabilities that subsequent billiard balls will fall above or below the first ball. By making the binomial parameter p depend on a random event, he cleverly escapes a philosophical quagmire that he most likely was not even aware was an issue. Statement of Bayes' theoremBayes' theorem is a relation among conditional and marginal probabilities. It can be viewed as a means of incorporating information, from an observation, for example, to produce a modified or updated probability distribution. To derive Bayes' theorem, note first from the definition of conditional probability that which is the theorem conventionally known as Bayes' theorem. Each term in Bayes' theorem has a conventional name. The term P(A) is called the prior probability of A. It is "prior" in the sense that it precedes any information about B. P(A) is also the marginal probability of A. The term P(A|B) is called the posterior probability of A, given B. It is "posterior" in the sense that it is derived from or entailed by the specified value of B. The term P(B|A), for a specific value of B, is called the likelihood function for A given B and can also be written as L(A|B). The term P(B) is the prior or marginal probability of B, and acts as the normalizing constant. With this terminology, the theorem may be paraphrased as Alternative forms of Bayes' theoremBayes' theorem is often embellished by noting that so the theorem can be restated as where AC is the complementary event of A. More generally, where {Ai} forms a partition of the event space, for any Ai in the partition. It can also be written neatly in terms of a likelihood ratio and odds as See also the law of total probability. Bayes' theorem for probability densitiesThere is also a version of Bayes' theorem for continuous distributions. It is somewhat harder to derive, since probability densities, strictly speaking, are not probabilities, so Bayes' theorem has to be established by a limit process; see Papoulis (citation below), Section 7.3 for an elementary derivation. Bayes' theorem for probability densities is formally similar to the theorem for probabilities: and there is an analogous statement of the law of total probability: As in the discrete case, the terms have standard names. f(x, y) is the joint distribution of X and Y, f(x|y) is the posterior distribution of X given Y=y, f(y|x) = L(x|y) is (as a function of x) the likelihood function of X given Y=y, and f(x) and f(y) are the marginal distributions of X and Y respectively, with f(x) being the prior distribution of X. Here we have indulged in a conventional abuse of notation, using f for each one of these terms, although each one is really a different function; the functions are distinguished by the names of their arguments. Extensions of Bayes' theoremTheorems analogous to Bayes' theorem hold in problems with more than two variables. These theorems are not given distinct names, as they may be mass-produced by applying the laws of probability. The general strategy is to work with a decomposition of the joint probability, and to marginalize (integrate) over the variables that are not of interest. Depending on the form of the decomposition, it may be possible to prove that some integrals must be 1, and thus they fall out of the decomposition; exploiting this property can reduce the computations very substantially. A Bayesian network is essentially a mechanism for automatically generating the extensions of Bayes' theorem that are appropriate for a given decomposition of the joint probability. ExampleTypical examples that use Bayes' theorem assume the philosophy underlying Bayesian probability that uncertainty and degrees of belief can be measured as probabilities. One such example follows. For additional worked out examples, please see the article on the examples of Bayesian inference. We wish to know about the proportion r of voters who will vote "yes" on a referendum. It is given that n=10 voters have been located at random, and m=7 say they will vote yes. From Bayes' theorem we have From this we see that once we have in hand the prior f(r) and the likelihood function f(n=10, m=7|r), we can compute the posterior f(r|n=10, m=7). The prior summarizes what we know about the distribution of r in the absence of any observation. We will assume in this case that the prior distribution of r is uniform over the interval [0, 1]. That is, f(r) = 1. That assumption should be considered provisional -- if some additional background information is found, we should modify the prior accordingly. Under the assumption of random sampling, choosing voters is just like choosing balls from an urn. The likelihood function for such a problem is just the probability of 7 successes in 10 trials for a binomial distribution. As with the prior, the likelihood is open to revision -- more complex assumptions will yield more complex likelihood functions. Maintaining the current assumptions, we compute the normalizing factor, and the posterior distribution for r is then for r between 0 and 1, inclusive. One may be interested in the probability that more than half the voters will vote "yes". The prior probability that more than half the voters will vote "yes" is 1/2, by the symmetry of the uniform distribution. In comparison, the posterior probability that more than half the voters will vote "yes", i.e., the conditional probability given the outcome of the opinion poll -- that seven of the 10 voters questioned will vote "yes" -- is which is about an "89% chance". ReferencesVersions of the essay
Commentaries
Additional material
See also
Directory of Poker ResourcesAll text is available under the terms of the GNU Free Documentation License. Source: Original text from the article in Wikipedia, The Free Encyclopedia |
Step One - The first step to learning how to play poker is easy... practice, practice, practice.
Step Two - Practice may not make you perfect, but taking poker lessons with Party Poker's poker school will improve your odds greatly. As you become more familiar with pot odds and good starting hands, you will see your game improve. Step Three - By setting up a PartyPoker account for free, it's safe, simple, and secure...play as little or as often as you like! Games are being played around the clock...
Step Four - Always have fun! Online playing features many great tournaments and prize package games. Build up bonues, play heads up or against a group, or join in the excitement of online tournament play!!
!
Free Poker Games Want to Find Free Catalogs? If you wish to find great magazines, then search the catalogs for clothes, recreation, and household items. Go to Free Catalogs Available Learn from the Pros
Looking for a great bargain? Search the millions of products available on eBay where you will find great prices! Discover for yourself how easy it is to get started! Join eBay today! |
Online Games Digest | Free Catalogs | Arkansas | Free Stuff | Affiliates | Findlings | Careers |
Poker News Archives
Learn how poker players got started, learn how poker tournaments are run, learn how pot odds work, and learn how online poker game freerolls work.