Professional Documents
Culture Documents
Gambling
- We assume that the gambler distributes all of his wealth across the horses.
- Let bi be the fraction of the gambler’s wealth invested in horse i, where bi ≥ 0 and ∑bi = 1.
- Then if horse i wins the race, the gambler will receive oi times the amount of wealth bet on
horse i. All the other bets are lost.
- Thus, at the end of the race, the gambler will have multiplied his wealth by a factor bioi if
horse i wins, and this will happen with probability pi .
- For notational convenience, we use b(i) and bi interchangeably
The horse race (cont)
- The wealth at the end of the race is a random variable, and the gambler wishes to
“maximize” the value of this random variable.
- Now since the gambler can reinvest his money, his wealth is the product of the gains for
each race. Let Sn be the gambler’s wealth after n races.
Then
𝑛
𝑆𝑛 = ∏ 𝑆(𝑋𝑖 )
𝑖=1
where S(X) = b(X)o(X) is the factor by which the gambler’s wealth is multiplied when
horse X wins.
Definitions
Definition The wealth relative S(X) = b(X)o(X) is the factor by which the gambler’s wealth
Definition The optimum doubling rate W*(p) is the maximum doubling rat over all
choices of the portfolio b:
𝑊 ∗ (𝑝) = max 𝑊(𝑏, 𝑝) = max 𝑝𝑖 log 𝑏𝑖 𝑜𝑖
𝑏 𝑏:𝑏𝑖 ≥0,∑𝑖 𝑏𝑖 =1
- We now consider a special case when the odds are fair with respect to some
1
distribution (∑ = 1)
𝑜𝑖
1
- We write 𝑟𝑖 = where ri can be interpreted as a probability mass function over the
𝑜𝑖
- The doubling rate is the difference between the distance of the bookie’s estimate from the
true distribution and the distance of the gambler’s estimate from the true distribution.
- Hence, the gambler can make money only if his estimate (as expressed by b) is better than
the bookie’s.
Gambling and side information
- Suppose the gambler has some information that is relevant to the outcome of the gamble.
For example, the gambler may have some information about the performance of the horses
in previous races.
- What is the value of this side information?
Gambling and side information (cont)
- Let horse X ∈ {1, 2, . . .,m} win the race with probability p(x) and pay odds of o(x) for 1.
side information Y, where b(x|y) is the proportion of wealth bet on horse x when y is
observed.
Theorem The increase ∆W in doubling rate due to side information Y for a horse race
X is
∆W = I(X; Y)
Gambling and side information (cont)
The entropy of English
Here are some examples of Markov approximations to English from Shannon’s original paper:
1. Zero-order approximation.
XFOML RXKHRJFFJUJ ZLPWCFWKCYJ
FFJEYVKCQSGXYD QPAAMKBZAACIBZLHJQD
2. First-order approximation.
OCRO HLI RGWR NMIELWIS EU LL NBNESEBYA TH EEI
ALHENHTTPA OOBTTVA NAH BRL
3. Second-order approximation.
ON IE ANTSOUTINYS ARE T INCTORE ST BE S DEAMY
ACHIN D ILONASIVE TUCOOWE AT TEASONARE FUSO
TIZIN ANDY TOBE SEACE CTISBE
4. Third-order approximation.
IN NO IST LAT WHEY CRATICT FROURE BERS GROCID
PONDENOME OF DEMONSTURES OF THE REPTAGIN IS
REGOACTIONA OF CRE
5. Fourth-order approximation.
THE GENERATED JOB PROVIDUAL BETTER TRAND THE DISPLAYED
CODE, ABOVERY UPONDULTS WELL THE CODERST IN THESTICAL
IT DO HOCK BOTHE MERG. (INSTATES CONS ERATION. NEVER
ANY OF PUBLE AND TO THEORY. EVENTIAL CALLEGAND TO ELAST
BENERATED IN WITH PIES AS IS WITH THE )
6. First-order word model . (The words are chosen independently but with frequencies as in
English.)
REPRESENTING AND SPEEDILY IS AN GOOD APT OR COME CAN
DIFFERENT NATURAL HERE HE THE A IN CAME THE TO OF TO
EXPERT GRAY COME TO FURNISHES THE LINE MESSAGE HAD BE
THESE.
7. Second-order word model. (The word transition probabilities match English text.)
THE HEAD AND IN FRONTAL ATTACK ON AN ENGLISH WRITER
THAT THE CHARACTER OF THIS POINT IS THEREFORE ANOTHER
METHOD FOR THE LETTERS THAT THE TIME OF WHO EVER TOLD
THE PROBLEM FOR AN UNEXPECTED
The entropy of English (cont)
o For example, the entropy of the zeroth-order model is log 27 = 4.76 bits per letter.
o The first order model gives an estimate of the entropy of 4.03 bits per letter, while the
fourth-order model gives an estimate of 2.8 bits per letter.
The entropy of English (cont)
(b) Show that if instead we put all of our money on horse 1, the most likely winner, we will eventually go
broke with probability 1.