Professional Documents
Culture Documents
Let RWM() denote the RWM algorithm described in class run with pa-
rameter > 0. Consider the version of the FPL algorithm FPL() defined
using the perturbation:
>
log( log(u1 )) log( log(uN ))
p1 = ,..., .
where, for j [1, N ], uj is drawn from the uniform distribution over [0, 1].
At round t [1, T ], wt is found via wt = M (x1:t1 + p1 ) = argminwW w
(x1:t1 +p1 ) using the notation adopted in the class lecture for FPL, with W
the set of coordinate vectors. Show that FPL() coincides with RWM().
where Pwe used the fact that ui is a uniform random variable. Fix t and let
Li = t1k=1 xik be the cumulative loss of coordinate i and Li = Li + p1i .
e
Since the set W consists of coordinate vectors , the minimization problem is
equivalent to finding i such that i = argmini Le i . Let
e i x) = 1 ee(xLi )
Fi (x) = Pr(L and
e(xLi )
fi (x) = e(xLi ) e
1
e i and let Gi (x) = 1 Fi (x). If pi = Pr(i = i), then
the random variable L
e i = min L
pi = Pr(L e j ) = Pr(L ei Le j j 6= i)
j
= E Pr(L ei Le j j 6= i|L
ei )
L
ei
hY i
=E Gj (L ei )
L
ei
i6=j
Z Y
= fi (x) Gj (x)dx.
i6=j
Using the definition of fi and Gj we see that the above expression is given
by
Z n Z
e(xLj ) eLj
Y x
Pn
e(xLi )
e = eLi ex ee j=1
j=1
eLi
= Pn Lj
.
j=1 e
B. Zero-sum games
For all the questions that follow, we consider a zero-sum game with payoffs
in [0, 1].
RT
max pRWM Mq min max p> Mq + .
qN pN qN T
Furthermore, we know that the regret of RWM is in O( log N / T ).
Therefore, after only O(log N/2 ) iterations of RWM we can obtain an
-approximation of the value of the game.
2
2. Use the proof given in class for von Neumanns theorem to show that
both players can come up with a strategy achieving and -approximation
of the value of the game (or Nash equilibrium) that are sparse: the
support of each mixed strategy is in O(log N/2 ). What fraction of the
payoff matrix does it suffice to consider to compute these strategies?
C. Bregman divergence
3
where F is a convex function. Furthermore, G(z, y) = BF (z||y). No-
tice that these properties can be easily verified for any function G. We
first prove the necessity of these conditions. The second bullet point
follows directly from the definition of Bregman divergence. To prove
the first bullet point, let G(x, y) = BF (x||y), then
G(x, y) + G(y, z)
= F (x) F (y) hF (y), x yi + F (y) F (z) hF (z), y zi
= F (x) F (z) hF (z), x zi + hF (z), x yi hF (y), x yi
= G(x, z) + hx y, F (z) F (y)i
= G(x, z) + hx y, z G(z, y)i.
In order for G to satisfy the first condition, the following equality must
then hold:
2. Using the results of the previous exercise, decide whether or not the
following functions are a Bregman divergence.
Solution:
4
By definition we have z i G(z, y) = log zi log yi + 1. Therefore, there
exists no function F such that z G(z, y) = F (z) F (y) and G
cannot be a Bregman divergence.
By definition we have z G(z, y) = ez + zez ey yey = F 0 (z) F 0 (y),
z
where F (z) = ze . Furthermore,