weipang142857's picture
Upload folder using huggingface_hub
f6cc031 verified
<Poster Width="1003" Height="1419">
<Panel left="27" right="94" width="462" height="117">
<Text>Contributions</Text>
<Text>An efficient algorithm, CGP-UCB, for the contextual GP bandit problem</Text>
<Text>Flexibly combining kernels over contexts and actions</Text>
<Text>Generic approach for deriving regret bounds for composite kernel functions</Text>
<Text>Evaluate CGP-UCB on automated vaccine design and sensor management</Text>
</Panel>
<Panel left="24" right="214" width="465" height="187">
<Text>Contextual Bandits [cf., Auer ’02; Langford & Zhang ’08]</Text>
<Text>Play a game for T rounds:</Text>
<Text>Receive context zt ∈ Z</Text>
<Text>Choose an action st ∈ S</Text>
<Text>Receive a payoff yt = f (st , zt ) + t (f unknown).</Text>
<Text>Cumulative regret for context specific action</Text>
<Text>Incur contextual regret rt = sups0∈S f (s0, zt ) − f (st , zt )</Text>
<Text>PTAfter T rounds, the cumulative contextual regret is RT =</Text>
<Text>t=1 rt .</Text>
<Text>Context-specific best action is a demanding benchmark.</Text>
</Panel>
<Panel left="24" right="406" width="464" height="190">
<Text>Gaussian Processes (GP)</Text>
<Text>Model payoff function using GPs: f ∼ GP(µ, k)</Text>
<Text>• observations yT = [y1 . . . yT ]T at inputs AT = {x1, . . . , xT }</Text>
<Text>yt = f (xt ) + t with i.i.d. Gaussian noise t ∼ N(0, σ 2)</Text>
<Text>Posterior distribution over f is a GP with</Text>
<Text>mean</Text>
<Text>covariance</Text>
<Text>variance</Text>
<Text>where kT (x) = [k(x1, x) . . . k(xT , x)]T and KT is the kernel matrix.</Text>
</Panel>
<Panel left="25" right="603" width="465" height="387">
<Text>GP-UCB [Srinivas, Krause, Kakade, Seeger ICML 2010]</Text>
<Figure left="75" right="636" width="143" height="116" no="1" OriWidth="0" OriHeight="0
" />
<Figure left="295" right="633" width="149" height="118" no="2" OriWidth="0" OriHeight="0
" />
<Text>Context free upper confidence bound algorithm (GP-UCB)</Text>
<Text>At round t, GP-UCB picks action st = xt such that</Text>
<Text>with appropriate βt . Trades exploration (high σ) and exploitation (high µ).</Text>
<Text>with appropriate βt . Trades exploration (high </Text>
<Text>Maximum information gain bounds regret</Text>
<Text>The (context-free) regret RT of GP-UCB is bounded by O∗( T βT γT ), where</Text>
<Text>γT is defined as the maximum information gain:</Text>
<Text>quantifies the reduction in uncertainty about f achieved by revealing yA.</Text>
<Text>Bounds for Kernels</Text>
<Text>Bounds on γT exist for linear, squared exponential and Mat´ern kernels.</Text>
</Panel>
<Panel left="26" right="992" width="463" height="128">
<Text>Contextual Upper Confidence Bound Algorithm (CGP-UCB)</Text>
<Text>where µt−1(·) and σt−1(·) are the posterior mean and standard deviation of the GP</Text>
<Text>over the joint set X = S × Z conditioned on the observations</Text>
<Text>(s1, z1, y1), . . . , (st−1, zt−1, yt−1).</Text>
</Panel>
<Panel left="25" right="1127" width="464" height="274">
<Text>Bounds on Contextual Regret</Text>
<Text>Then√for appropriate choices of βt , the contextual regret of CGP-UCB is bounded by</Text>
<Text>O∗( T γT βT ) w.h.p. Precisely,</Text>
<Text>onpLet δ ∈ (0, 1). Suppose one of the following assumptions holds</Text>
<Text>X is finite, f is sampled from a known GP prior with known noise variance σ 2,</Text>
<Text>dX is compact and convex, ⊆ [0, r ] , d ∈ N, r > 0. Suppose f is sampled from a</Text>
<Text>known GP prior with known noise variance σ 2, and that k(x, x0) has smooth</Text>
<Text>derivatives,</Text>
<Text>X is arbitrary; ||f ||k ≤ B. The noise variables t form an arbitrary martingale</Text>
<Text>difference sequence (meaning that E[εt | ε1, . . . , εt−1] = 0 for all t ∈ N),</Text>
<Text>uniformly bounded by σ.</Text>
<Text>where C1 = 8/ log(1 + σ −2).</Text>
</Panel>
<Panel left="513" right="94" width="464" height="404">
<Text>Composite Kernels</Text>
<Figure left="539" right="142" width="180" height="126" no="3" OriWidth="0.233526" OriHeight="0.12958
" />
<Figure left="773" right="130" width="176" height="135" no="4" OriWidth="0.231214" OriHeight="0.136282
" />
<Text>Product of squared exponential kernel</Text>
<Text>and linear kernel</Text>
<Text>Additive combination of payoff that</Text>
<Text>smoothly depends on context, and</Text>
<Text>exhibits clusters of actions.</Text>
<Text>Product kernel</Text>
<Text>• k = kS ⊗ kZ , where (kS ⊗ kZ )((s, z), (s0, z0)) = kZ (z, z0)kS (s, s0)</Text>
<Text>• Two context-action pairs are similar (large correlation) if the contexts are</Text>
<Text>similar and actions are similar</Text>
<Text>Additive kernel</Text>
<Text>• (kS ⊕ kZ )((s, z), (s0, z0)) = kZ (z, z0) + kS (s, s0)</Text>
<Text>• Generative model: first sample a function fS (s, z) that is constant along z, and</Text>
<Text>varies along s with regularity as expressed by ks; then sample a function fz(s, z),</Text>
<Text>which varies along z and is constant along s;</Text>
</Panel>
<Panel left="511" right="501" width="467" height="216">
<Text>Bounds for Composite Kernels</Text>
<Text>Maximum information gain for a GP with kernel k on set V</Text>
<Text>Product kernel</Text>
<Text>Let kZ be a kernel function on Z with rank at most d . Then</Text>
<Text>Additive kernel</Text>
<Text>Let kS and kZ be kernel functions on S and Z respectively. Then</Text>
</Panel>
<Panel left="513" right="723" width="464" height="336">
<Figure left="529" right="764" width="118" height="110" no="5" OriWidth="0.185549" OriHeight="0.127793
" />
<Figure left="672" right="764" width="150" height="121" no="6" OriWidth="0.198266" OriHeight="0.127793
" />
<Figure left="832" right="763" width="140" height="122" no="7" OriWidth="0.192486" OriHeight="0.125559
" />
<Text>Task Discover peptide sequences binding to MHC molecules</Text>
<Text>Context Features encoding the MHC alleles</Text>
<Text>Action Choose a stimulus (the vaccine) s ∈ S that maximizes an observed response</Text>
<Text>(binding affinity).</Text>
<Text>Kernels Use a finite inter-task covariance kernel KZ with rank mZ to model the</Text>
<Text>similarity of different experiments, and a Gaussian kernel kS (s, s0) to model the</Text>
<Text>experimental parameters.</Text>
</Panel>
<Panel left="510" right="1060" width="466" height="281">
<Text>Learning to Monitor Sensor Networks</Text>
<Figure left="520" right="1099" width="135" height="103" no="8" OriWidth="0.195376" OriHeight="0.119303
" />
<Figure left="670" right="1101" width="150" height="114" no="9" OriWidth="0.194798" OriHeight="0.117069
" />
<Figure left="826" right="1102" width="147" height="114" no="10" OriWidth="0.191329" OriHeight="0.116175
" />
<Text>Temperature data from a</Text>
<Text>network of 46 sensors at</Text>
<Text>Intel Research.</Text>
<Text>Time (h)</Text>
<Text>CGP-UCB using average</Text>
<Text>temperature</Text>
<Text>CGP-UCB using</Text>
<Text>minimum temperature</Text>
<Text>Task Given a sensor network, monitor maximum temperatures in building</Text>
<Text>Context Time of day</Text>
<Text>Action Pick 5 sensors to activate</Text>
<Text>Kernels Joint spatio-temporal covariance function using the Mat´ern kernel</Text>
</Panel>
</Poster>