coopkernelfc coopkernelfb arxiv:2110.15771v1 [cs.lg] 29

44
Collaborative Pure Exploration in Kernel Bandit YIHAN DU, IIIS, Tsinghua University, China WEI CHEN, Microsoft Research, China YUKO KUROKI, The University of Tokyo / RIKEN, Japan LONGBO HUANG, IIIS, Tsinghua University, China In this paper, we formulate a Collaborative Pure Exploration in Kernel Bandit problem (CoPE-KB), which provides a novel model for multi-agent multi-task decision making under limited communication and general reward functions, and is applicable to many online learning tasks, e.g., recommendation systems and network scheduling. We consider two settings of CoPE-KB, i.e., Fixed-Confidence (FC) and Fixed-Budget (FB), and design two optimal algorithms CoopKernelFC (for FC) and CoopKernelFB (for FB). Our algorithms are equipped with innovative and efficient kernelized estimators to simultaneously achieve computation and communication efficiency. Matching upper and lower bounds under both the statistical and communication metrics are established to demonstrate the optimality of our algorithms. The theoretical bounds successfully quantify the influences of task similarities on learning acceleration and only depend on the effective dimension of the kernelized feature space. Our analytical techniques, including data dimension decomposition, linear structured instance transformation and (communication) round-speedup induction, are novel and applicable to other bandit problems. Empirical evaluations are provided to validate our theoretical results and demonstrate the performance superiority of our algorithms. Additional Key Words and Phrases: collaborative pure exploration, kernel bandit, communication round, learning speedup ACM Reference Format: Yihan Du, Wei Chen, Yuko Kuroki, and Longbo Huang. 2021. Collaborative Pure Exploration in Kernel Bandit. In Woodstock ’21: ACM Symposium on Neural Gaze Detection, June 03–05, 2021, Woodstock, NY . ACM, New York, NY, USA, 44 pages. https://doi.org/10.1145/ 1122445.1122456 1 INTRODUCTION Pure exploration [3, 9, 12, 18, 21, 24] is a fundamental online learning problem in multi-armed bandits, where an agent sequentially chooses options (often called arms) and observes random feedback, with the objective of identifying the best option (arm). This problem finds various applications such as recommendation systems [31], online advertising [37] and neural architecture search [17]. However, the traditional single-agent pure exploration problem [3, 9, 12, 18, 21, 24] cannot be directly applied to many real-world distributed online learning platforms, which often face a large volume of user requests and need to coordinate multiple distributed computing devices to process the requests, e.g., geographically distributed data centers [30] and Web servers [44]. These computing devices communicate with each other to exchange information in order to attain globally optimal performance. To handle such distributed pure exploration problem, prior works [20, 22, 38] have developed the Collaborative Pure Exploration (CoPE) model, where there are multiple agents that communicate and cooperate in order to identify the best arm with learning speedup. Yet, existing results only investigate the classic multi-armed bandit (MAB) setting [3, 18, 21], Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]. © 2021 Association for Computing Machinery. Manuscript submitted to ACM 1 arXiv:2110.15771v1 [cs.LG] 29 Oct 2021

Upload: others

Post on 19-Feb-2022

3 views

Category:

Documents


0 download

TRANSCRIPT

Collaborative Pure Exploration in Kernel Bandit

YIHAN DU, IIIS, Tsinghua University, China

WEI CHEN,Microsoft Research, China

YUKO KUROKI, The University of Tokyo / RIKEN, Japan

LONGBO HUANG, IIIS, Tsinghua University, China

In this paper, we formulate a Collaborative Pure Exploration in Kernel Bandit problem (CoPE-KB), which provides a novel model for

multi-agent multi-task decision making under limited communication and general reward functions, and is applicable to many online

learning tasks, e.g., recommendation systems and network scheduling. We consider two settings of CoPE-KB, i.e., Fixed-Confidence

(FC) and Fixed-Budget (FB), and design two optimal algorithms CoopKernelFC (for FC) and CoopKernelFB (for FB). Our algorithms

are equipped with innovative and efficient kernelized estimators to simultaneously achieve computation and communication efficiency.

Matching upper and lower bounds under both the statistical and communication metrics are established to demonstrate the optimality

of our algorithms. The theoretical bounds successfully quantify the influences of task similarities on learning acceleration and only

depend on the effective dimension of the kernelized feature space. Our analytical techniques, including data dimension decomposition,

linear structured instance transformation and (communication) round-speedup induction, are novel and applicable to other bandit

problems. Empirical evaluations are provided to validate our theoretical results and demonstrate the performance superiority of our

algorithms.

Additional Key Words and Phrases: collaborative pure exploration, kernel bandit, communication round, learning speedup

ACM Reference Format:Yihan Du, Wei Chen, Yuko Kuroki, and Longbo Huang. 2021. Collaborative Pure Exploration in Kernel Bandit. In Woodstock ’21: ACM

Symposium on Neural Gaze Detection, June 03–05, 2021, Woodstock, NY. ACM, New York, NY, USA, 44 pages. https://doi.org/10.1145/

1122445.1122456

1 INTRODUCTION

Pure exploration [3, 9, 12, 18, 21, 24] is a fundamental online learning problem in multi-armed bandits, where an agent

sequentially chooses options (often called arms) and observes random feedback, with the objective of identifying the

best option (arm). This problem finds various applications such as recommendation systems [31], online advertising [37]

and neural architecture search [17]. However, the traditional single-agent pure exploration problem [3, 9, 12, 18, 21, 24]

cannot be directly applied to many real-world distributed online learning platforms, which often face a large volume of

user requests and need to coordinate multiple distributed computing devices to process the requests, e.g., geographically

distributed data centers [30] and Web servers [44]. These computing devices communicate with each other to exchange

information in order to attain globally optimal performance.

To handle such distributed pure exploration problem, prior works [20, 22, 38] have developed the Collaborative Pure

Exploration (CoPE) model, where there are multiple agents that communicate and cooperate in order to identify the best

arm with learning speedup. Yet, existing results only investigate the classic multi-armed bandit (MAB) setting [3, 18, 21],

Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not

made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components

of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to

redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

© 2021 Association for Computing Machinery.

Manuscript submitted to ACM

1

arX

iv:2

110.

1577

1v1

[cs

.LG

] 2

9 O

ct 2

021

Woodstock ’21, June 03–05, 2021, Woodstock, NY

and focus only on the fully-collaborative setting, i.e., the agents aim to solve a common task. However, in many

real-world applications such as recommendation systems [31], it is often the case that different computing devices

face different but correlated recommendation tasks. Moreover, there usually exists some structured dependency of user

utilities on the recommended items. In such applications, it is important to develop a more general CoPE model that

allows heterogeneous tasks and complex reward structures, and quantitatively investigate how task similarities impact

learning acceleration.

Motivated by the above facts, we propose a novel Collaborative Pure Exploration in Kernel Bandit (CoPE-KB) problem,

which generalizes traditional single-task CoPE problems [20, 22, 38] to the multi-task setting. It also generalizes the

classic MAB model to allow general (linear or nonlinear) reward structures via the powerful kernel representation.

Specifically, each agent is given a set of arms, and the expected reward of each arm is generated by a task-dependent

reward function with a low norm in a high-dimensional (possibly infinite-dimensional) Reproducing Kernel Hilbert

Space (RKHS) [33, 42], by which we can represent real-world nonlinear reward dependency as some linear function in

a high-dimensional space, and can go beyond linear rewards as commonly done in the literature, e.g., [10, 13, 14, 35, 40].

Each agent sequentially chooses arms to sample and observes noisy outcomes. The agents can broadcast and receive

messages to/from others in communication rounds, so that they can exploit the task similarity and collaborate to

expedite learning processes. The task of each agent is to find the best arm that maximizes the expected reward among

her arm set.

Our CoPE-KB formulation can handle different tasks in parallel and characterize the dependency of rewards on

options, which provides a more general and flexible model for real-world applications. For example, in distributed

recommendation systems [31], different computing devices can face different tasks, and it is inefficient to learn the

reward of each option individually. Instead, CoPE-KB enables us to directly learn the relationship between option

features and user utilities, and exploit the similarity of such relationship among different tasks to accelerate learning.

There are also many other applications, such as clinical trials [41], where we conduct multiple clinical trials in parallel

and utilize the common useful information to accelerate drug development, and neural architecture search [17], where

we simultaneously run different tests of neural architectures under different environmental setups to expedite search

processes.

We consider two important pure exploration settings under the CoPE-KB model, i.e., Fixed-Confidence (FC), where

agents aim to minimize the number of used samples under a given confidence, and Fixed-Budget (FB), where the goal is

to minimize the error probability under a given sample budget. Note that due to the high dimension (possibly infinite)

of the RKHS, it is highly non-trivial to simplify the burdensome computation and communication in the RKHS, and to

derive theoretical bounds only dependent on the effective dimension of the kernelized feature space. To tackle the above

challenges, we adopt efficient kernelized estimators and design novel algorithms CoopKernelFC and CoopKernelFB for

the FC and FB settings, respectively, which only cost Poly(𝑛𝑉 ) computation and communication complexity instead of

Poly(dim(H𝐾 )) as in [10, 43], where 𝑛 is the number of arms,𝑉 is the number of agents, andH𝐾 is the high-dimensional

RKHS. We also establish matching upper and lower bounds in terms of sampling and communication complexity to

demonstrate the optimality of our algorithms (within logarithmic factors).

Our work distinguishes itself from prior CoPE works, e.g., [20, 22, 38], in the following aspects: (i) Prior works [20, 22,

38] only consider the classic MAB setting, while we adopt a high-dimensional RKHS to allow more general real-world

reward dependency on option features. (ii) Unlike [20, 22, 38] which restrict tasks (given arm sets and rewards) among

agents to be the same, we allow different tasks for different agents, and explicitly quantify how task similarities impact

learning acceleration. (iii) In lower bound analysis, prior works [20, 38] mainly focus on a 2-armed case, whereas we

2

Collaborative Pure Exploration in Kernel Bandit Woodstock ’21, June 03–05, 2021, Woodstock, NY

derive a novel lower bound analysis for general multi-armed cases with high-dimensional linear reward structures.

Moreover, when reducing CoPE-KB to prior CoPE with classic MAB setting (all agents are solving the same classic

MAB task) [20, 38], our lower and upper bounds also match the existing state-of-the-art results in [38].

The contributions of this paper are summarized as follows:

• We formulate a novel Collaborative Pure Exploration in Kernel Bandit (CoPE-KB) problem, which models

distributed multi-task decision making problems with general reward functions, and finds applications in many

real-world online learning tasks, and study two settings of CoPE-KB, i.e., CoPE-KB with fixed-confidence (FC)

and CoPE-KB with fixed-budget (FB).

• For CoPE-KB with fixed-confidence (FC), we propose an algorithm CoopKernelFC, which adopts an effi-

cient kernelized estimator to significantly reduce computation and communication complexity from exist-

ing Poly(dim(H𝐾 )) to only Poly(𝑛𝑉 ). We derive matching upper and lower bounds of sample complexity

�̃� ( 𝜌∗

𝑉log𝛿−1) and communication rounds 𝑂 (logΔ−1

min). Here 𝜌∗ is the problem hardness (defined in Section 4.2),

and Δ−1

minis the minimum reward gap.

• For CoPE-KB with fixed-budget (FB), we design an efficient algorithm CoopKernelFB with error probability

�̃�

(exp

(−𝑇𝑉𝜌∗

)𝑛2𝑉

)and communication rounds 𝑂 (log(𝜔 ( ˜X))). A matching lower bound of communication

rounds is also established to validate the communication optimality of CoopKernelFB (within double-logarithmic

factors). Here𝑇 is the sample budget,˜X is the set of arms, and𝜔 ( ˜X) is the principle dimension of data projections

in˜X to RKHS (defined in Section 5.1.1).

• Our results explicitly quantify the impacts of task similarities on learning acceleration. Our novel analytical

techniques, including data dimension decomposition, linear structured instance transformation and round-

speedup induction, can be of independent interests and are applicable to other bandit problems.

Due to space limit, we defer all the proofs to Appendix.

2 RELATEDWORK

This work falls in the literature of multi-armed bandits [4, 8, 27, 28, 39]. Here we mainly review three most related lines

of research, i.e., collaborative pure exploration and kernel bandit.

Collaborative Pure Exploration (CoPE). The collaborative pure exploration literature is initiated by [20], which

considers the classic MAB and fully-collaborative settings, and designs fixed-confidence algorithms based on majority

vote with upper bound analysis. Tao et al. [38] further develop a fixed-budget algorithm by calling conventional

single-agent fixed-confidence algorithms, and completes the analysis of round-speedup lower bounds. Karpov et al.

[22] extend the formulation of [20, 38] to the best𝑚 arm identification problem, and designs fixed-confidence and

fixed-budget algorithms with tight round-speedup upper and lower bounds, which give a strong separation between

best arm identification and the extended best 𝑚 arm identification. Our CoPE-KB model encompasses the classic

MAB and fully-collaborative settings in the above works [20, 22, 38], but faces unique challenges on computation and

communication efficiency due to the high-dimensional reward structures.

Collaborative Regret Minimization. There are other works studying collaborative (distributed) bandit with the

regret minimization objective. Bistritz and Leshem [5], Liu and Zhao [29], Rosenski et al. [32] study the multi-player

bandit with collisions motivated by cognitive radio networks, where multiple players simultaneously choose arms from

the same set and receive no reward if more than one player choose the same arm (i.e., a collision happens). Bubeck and

Budzinski [6], Bubeck et al. [7] investigate a variant multi-player bandit problem where players cannot communicate but

3

Woodstock ’21, June 03–05, 2021, Woodstock, NY

have access to shared randomness, and they propose algorithms that achieve nearly optimal regrets without collisions.

Chakraborty et al. [11] introduce another distributed bandit problem, where each agent decides either to pull an arm

or to broadcast a message in order to maximize the total reward. Korda et al. [25], Szorenyi et al. [36] adapt bandit

algorithms to peer-to-peer networks, where the peers pick arms from the same set and can only communicate with a

few random others along network links. The above works consider different learning objectives and communication

protocols from ours, and do not involve the challenges of simultaneously handling multiple different tasks and analyzing

the relationship between communication rounds and learning speedup.

Kernel Bandit. There are a number of works for kernel bandit with the regret minimization objective. Srinivas et al.

[35] study the Gaussian process bandit problem with RKHS, which is the Bayesian version of kernel bandits, and

designs an Upper Confidence Bound (UCB) style algorithm. Chowdhury and Gopalan [13] further improve the regret

results of [35] by constructing tighter kernelized confidence intervals. Valko et al. [40] consider kernel bandit from

the frequentist perspective and provides an alternative regret analysis based on effective dimension. Deshmukh et al.

[14], Krause and Ong [26] study the multi-task kernel bandits, where the kernel function of RKHS is constituted by two

compositions from task similarities and arm features. Dubey et al. [16] investigate the multi-agent kernel bandit with a

local communication protocol, with the learning objective being to reduce the average regret suffered by per agent. For

kernel bandit with the pure exploration objective, there are only two works [10, 43] to our best knowledge. Camilleri

et al. [10] design a single-agent algorithm which uses a robust inverse propensity score estimator to reduce the sample

complexity incurred by rounding procedures. Zhu et al. [43] propose a variant of [10] which applies neural networks to

approximate nonlinear reward functions. All of these works consider either regret minimization or single-agent setting,

which largely differs from our problem, and they do not investigate the distributed decision making and (communication)

round-speedup trade-off. Thus, their algorithms and analysis cannot be applied to solve our CoPE-KB problem.

3 COLLABORATIVE PURE EXPLORATION IN KERNEL BANDIT (COPE-KB)

In this section, we present the formal formulation of the Collaborative Pure Exploration in Kernel Bandit (CoPE-KB), and

discuss the two important settings under CoPE-KB that will be investigated.

Agents and rewards. There are 𝑉 agents [𝑉 ] = {1, . . . ,𝑉 }, who collaborate to solve different but possibly related

instances (tasks) of the pure exploration in kernel bandit (PE-KB) problem. For each agent 𝑣 ∈ [𝑉 ], she is given a set of

𝑛 arms X𝑣 = {𝑥𝑣,1, . . . , 𝑥𝑣,𝑛} ⊆ R𝑑X , where 𝑑X is the dimension of arm feature vectors. The expected reward of each

arm 𝑥 ∈ X𝑣 is 𝑓𝑣 (𝑥), where 𝑓𝑣 : X𝑣 ↦→ R is an unknown reward function. Let X = ∪𝑣∈[𝑉 ]X𝑣 . Following the literature inkernel bandits [14, 26, 35, 40], we assume that for any 𝑣 ∈ [𝑉 ], 𝑓𝑣 has a bounded norm in a Reproducing Kernel Hilbert

Space (RKHS) specified by kernel 𝐾X : X × X ↦→ R (see below for more details). At each timestep 𝑡 , each agent 𝑣 pulls

an arm 𝑥𝑣,𝑡 ∈ X𝑣 and observes a random reward 𝑦𝑣,𝑡 = 𝑓 (𝑥𝑣,𝑡 ) + [𝑣,𝑡 , where [𝑣,𝑡 is an independent and zero-mean

1-sub-Gaussian noise (without loss of generality).1We assume that the best arms 𝑥𝑣,∗ = argmax𝑥 ∈X𝑣 𝑓𝑣 (𝑥) are unique

for all 𝑣 ∈ [𝑉 ], which is a common assumption in the pure exploration literature, e.g., [3, 12, 18, 24].

Multi-Task Kernel Composition.We assume that the functions 𝑓𝑣 are parametric functionals of a global function

𝐹 : X ×Z ↦→ R, which satisfies that, for each agent 𝑣 ∈ [𝑉 ], there exists a task feature vector 𝑧𝑣 ∈ Z such that

𝑓𝑣 (𝑥) = 𝐹 (𝑥, 𝑧𝑣), ∀𝑥 ∈ X𝑣 . (1)

Here X andZ denote the arm feature space and task feature space, respectively. Eq. (1) allows tasks to be different for

agents, whereas prior CoPE works [20, 22, 38] restrict the tasks (X𝑣 and 𝑓𝑣 ) to be the same for all agents 𝑣 ∈ [𝑉 ].1A random variable [ is called 1-sub-Gaussian if it satisfies that E[exp(_[ − _E[[ ]) ] ≤ exp(_2/2) for any _ ∈ R.

4

Collaborative Pure Exploration in Kernel Bandit Woodstock ’21, June 03–05, 2021, Woodstock, NY

Denote˜X = X ×Z and 𝑥 = (𝑥, 𝑧𝑣). As a standard assumption in kernel bandits [14, 16, 26, 35], we assume that 𝐹

has a bounded norm in a global RKHS H𝐾 specified by kernel 𝐾 :˜X × ˜X ↦→ R, and there exists a feature mapping

𝜙 :˜X ↦→ H𝐾 and an unknown parameter \∗ ∈ H𝐾 such that

𝐹 (𝑥) = 𝜙 (𝑥)⊤\∗, ∀𝑥 ∈ ˜X, and 𝐾 (𝑥, 𝑥 ′) = 𝜙 (𝑥)⊤𝜙 (𝑥 ′), ∀𝑥, 𝑥 ′ ∈ ˜X.

Here ∥\∗∥ :=√\∗⊤\∗ ≤ 𝐵 for some known constant 𝐵 > 0. 𝐾 :

˜X × ˜X ↦→ R is a product composite kernel, which

satisfies that for any 𝑧, 𝑧′ ∈ Z, 𝑥, 𝑥 ′ ∈ X,

𝐾 ((𝑥, 𝑧), (𝑥 ′, 𝑧′)) = 𝐾Z (𝑧, 𝑧′) · 𝐾X (𝑥, 𝑥 ′),

where 𝐾X is the arm feature kernel that depicts the feature structure of arms, and 𝐾Z is the task feature kernel that

measures the similarity of functions 𝑓𝑣 . For example, in the fully-collaborative setting, all agents solve a common task,

and we have that 𝑧𝑣 = 1 for all 𝑣 ∈ [𝑉 ], 𝐾Z (𝑧, 𝑧′) = 1 for all 𝑧, 𝑧′ ∈ Z, and 𝐾 = 𝐾X . On the contrary, if all tasks are

different, then rank(𝐾Z) = 𝑉 . 𝐾Z allows us to characterize the influences of task similarities (1 ≤ rank(𝐾Z) ≤ 𝑉 ) onlearning.

We give a simple 2-agent (2-task) illustrating example in Figure 1. Agent 1 is given Items 1,2 with the expected

rewards `1, `2, respectively, denoted by X1 = {𝑥1,1, 𝑥1,2}. Agent 2 is given Items 2,3 with the expected rewards

`2, `3, respectively, denoted by X2 = {𝑥2,1, 𝑥2,2}. Here 𝑥1,2 = 𝑥2,1 is the same item. In this case, 𝜙 (𝑥1,1) = [1, 0, 0]⊤,𝜙 (𝑥1,2) = 𝜙 (𝑥2,1) = [0, 1, 0]⊤, 𝜙 (𝑥2,2) = [0, 0, 1]⊤, and \∗ = [`1, `2, `3]⊤. The two agents can share the learned

information on the second dimension of \∗ to accelerate learning processes.

𝑥1,1

Item 1, reward 𝜇1

Task 1 Task 2

𝜙 𝑥 1,1 = 1,0,0 𝑇

𝜙 𝑥 1,2 = 𝜙 𝑥 2,1 = 0,1,0 𝑇

𝜙 𝑥 2,2 = 0,0,1 𝑇

θ∗ = 𝜇1, 𝜇2, 𝜇3𝑇

𝑥2,2

Item 3, reward 𝜇3

𝑥1,2 = 𝑥2,1

Item 2, reward 𝜇2

𝑥 1,1 = 𝑧1, 𝑥1,1

𝑥 1,2 = 𝑧1, 𝑥1,2 = 𝑥 2,1 = 𝑧2, 𝑥2,1

𝑥 2,2 = 𝑧2, 𝑥2,2

Fig. 1. Illustrating example.

Note that the RKHSH𝐾 can have infinite dimensions,

and any direct operation on H𝐾 , e.g., the calculation

of 𝜙 (𝑥) and explicit expression of the estimate of \∗, is

impracticable. In this paper, all our algorithms only query

the kernel function 𝐾 (·, ·) instead of directly operating

onH𝐾 , and 𝜙 (𝑥) and \∗ are only used in our theoretical

analysis, which is different from existing works, e.g., [10,

43].

Communication. Following the popular communica-

tion protocol in existing CoPE works [20, 22, 38], we

allow these 𝑉 agents to exchange information via com-

munication rounds, in which each agent can broadcast

and receive messages from others. While we do not re-

strict the exact length of a message, for practical implementation it should be bounded by 𝑂 (𝑛) bits. Here 𝑛 is the

number of arms for each agent, and we consider the number of bits for representing a real number as a constant.

In the CoPE-KB problem, our goal is to design computation and communication efficient algorithms to coordinate

different agents to simultaneously complete multiple tasks in collaboration and characterize how the take similarities

impact the learning speedup.

Fixed-Confidence and Fixed-Budget. We consider two versions of the CoPE-KB problem, one with fixed-confidence

(FC) and the other with fixed-budget (FB). Specifically, in the FC setting, given a confidence parameter 𝛿 ∈ (0, 1), theagents aim to identify 𝑥𝑣,∗ for all 𝑣 ∈ [𝑉 ] with probability at least 1 − 𝛿 and minimize the average number of samples

used by each agent. In the FB setting, on the other hand, the agents are given an overall 𝑇 ·𝑉 sample budget (𝑇 average

5

Woodstock ’21, June 03–05, 2021, Woodstock, NY

samples per agent), and aim to use at most𝑇 ·𝑉 samples to identify 𝑥𝑣,∗ for all 𝑣 ∈ [𝑉 ] and minimize the error probability.

In both FC and FB settings, agents are requested to minimize the number of communication rounds.

To evaluate the learning acceleration of our algorithms, following the CoPE literature, e.g., [20, 22, 38], we also define

the speedup metric of our algorithms. For a CoPE-KB instance I, let𝑇A𝑀 ,I denote the average number of samples used

by each agent in multi-agent algorithmA𝑀 to identify 𝑥𝑣,∗ for all 𝑣 ∈ [𝑉 ], and let𝑇A𝑆 ,I denote the average number of

samples used by each task for a single-agent algorithmA𝑆 to sequentially (without communication) identify 𝑥𝑣,∗ for all

𝑣 ∈ [𝑉 ]. Then, the speedup of A𝑀 on instance I is formally defined as

𝛽A𝑀 ,I = inf

A𝑆

𝑇A𝑆 ,I𝑇A𝑀 ,I

. (2)

It can be seen that 1 ≤ 𝛽A𝑀 ,I ≤ 𝑉 , where 𝛽A𝑀 ,I = 1 for the case where all tasks are different and 𝛽A𝑀 ,I can

approach 𝑉 for a fully-collaborative instance. By taking 𝑇A𝑀 ,I and 𝑇A𝑆 ,I as the smallest numbers of samples needed

to meet the confidence constraint, the definition of 𝛽A𝑀 ,I can be similarly defined for error probability results.

In particular, when all agents 𝑣 ∈ [𝑉 ] have the same arm set X𝑣 = X = {𝒆1, . . . , 𝒆𝑛} (i.e., standard bases in R𝑛) and

the same reward function 𝑓𝑣 (𝑥) = 𝑓 (𝑥) = 𝑥⊤\∗ for any 𝑥 ∈ X, all agents are solving a common classic MAB task, and

then the task featureZ = {1} and 𝐾Z (𝑧, 𝑧′) = 1 for any 𝑧, 𝑧′ ∈ Z. In this case, our CoPE-KB problem reduces to prior

CoPE with classic MAB setting [20, 38].

4 FIXED-CONFIDENCE COPE-KB

We start with the fixed-confidence (FC) setting and propose the CoopKernelFC algorithm. We explicitly quantify how

task similarities impact learning acceleration, and provide sample complexity and round-speedup lower bounds to

demonstrate the optimality of CoopKernelFC.

4.1 Algorithm CoopKernelFC

4.1.1 Algorithm. CoopKernelFC has three key components: (i) maintain alive arm sets for all agents, (ii) perform

sampling individually according to the globally optimal sample allocation, and (iii) exchange the distilled observa-

tion information to estimate reward gaps and eliminate sub-optimal arms, via efficient kernelized computation and

communication schemes.

The procedure of CoopKernelFC (Algorithm 1) for each agent 𝑣 is as follows. Agent 𝑣 maintains alive arm sets B (𝑡 )𝑣′

for all 𝑣 ′ ∈ [𝑉 ] by successively eliminating sub-optimal arms in each phase. In phase 𝑡 , she solves a global min-max

optimization, which takes into account the objectives and available arm sets of all agents, to obtain the optimal sample

allocation _∗𝑡 ∈ △ ˜X and optimal value 𝜌∗𝑡 (Line 4). Here △ ˜X is the collection of all distributions onX. b𝑡 is a regularizationparameter such that√︁

b𝑡 max

�̃�𝑖 ,�̃� 𝑗 ∈ ˜X𝑣 ,𝑣∈[𝑉 ]∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥(b𝑡 𝐼+∑𝑥∈X 1

𝑛𝑉𝜙 (𝑥)𝜙 (𝑥)⊤)−1 ≤ 1

(1 + Y)𝐵 · 2𝑡+1, (3)

which ensures the estimation bias for reward gap to be bounded by 2−(𝑡+1)

and can be efficiently computed by kernelized

transformation (specified in Section 4.1.2). Then, agent 𝑣 uses 𝜌∗𝑡 to compute the number of required samples 𝑁 (𝑡 ) , which

guarantees that the confidence radius of estimation for reward gaps is within 2−𝑡

(Line 5). In algorithm CoopKernelFC,

we use a rounding procedure ROUNDY (_, 𝑁 ) with approximation parameter 𝜖 from [2, 10], which rounds the sample

6

Collaborative Pure Exploration in Kernel Bandit Woodstock ’21, June 03–05, 2021, Woodstock, NY

Algorithm 1: Distributed Algorithm CoopKernelFC: for Agent 𝑣 (𝑣 ∈ [𝑉 ])Input: 𝛿 , ˜X1, . . . , ˜X𝑉 , 𝐾 (·, ·) :

˜X× ˜X ↦→ R, 𝐵, rounding procedure ROUNDY (·, ·) with approximation parameter Y.

1 Initialization: B (1)𝑣′ ← X𝑣′ for all 𝑣

′ ∈ [𝑉 ]. 𝑡 ← 1 ; // initialize alive arm sets B (1)𝑣′

2 while ∃𝑣 ′ ∈ [𝑉 ], |B (𝑡 )𝑣′ | > 1 do

3 𝛿𝑡 ← 𝛿2𝑡2

;

4 Let _∗𝑡 and 𝜌∗𝑡 be the optimal solution and optimal value of

min

_∈△ ˜Xmax

�̃�𝑖 ,�̃� 𝑗 ∈B (𝑡 )𝑣′ ,𝑣′∈[𝑉 ]

∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥2(b𝑡 𝐼+∑�̃�∈ ˜X _�̃�𝜙 (�̃�)𝜙 (�̃�)⊤)−1, where b𝑡 is a regularization parameter that

satisfies Eq. (3) ; // compute the optimal sample allocation

5 𝑁 (𝑡 ) ←⌈8(2𝑡 )2 (1 + Y)2𝜌∗𝑡 log

(2𝑛2𝑉 /𝛿𝑡

)⌉; // compute the number of required samples

6 (𝑠1, . . . , 𝑠𝑁 (𝑡 ) ) ← ROUNDY (_∗𝑡 , 𝑁 (𝑡 ) );7 Let �̃� (𝑡 )𝑣 be the sub-sequence of (𝑠1, . . . , 𝑠𝑁 (𝑡 ) ) which only contains the arms in

˜X𝑣 ; // generate the sample

sequence for agent 𝑣

8 Pull arms �̃� (𝑡 )𝑣 and observe random rewards 𝒚 (𝑡 )𝑣 ;

9 Broadcast {(𝑁 (𝑡 )𝑣,𝑖, 𝑦(𝑡 )𝑣,𝑖)}𝑖∈[𝑛] , where 𝑁

(𝑡 )𝑣,𝑖

is the number of samples and 𝑦(𝑡 )𝑣,𝑖

is the average observed reward

on arm 𝑥𝑣,𝑖 ;

10 Receive {(𝑁 (𝑡 )𝑣′,𝑖 , 𝑦

(𝑡 )𝑣′,𝑖 )}𝑖∈[𝑛] from all other agents 𝑣 ′ ∈ [𝑉 ] \ {𝑣};

11 For notational simplicity, we combine the subscripts 𝑣 ′, 𝑖 in 𝑥𝑣′,𝑖 , 𝑁(𝑡 )𝑣′,𝑖 , 𝑦

(𝑡 )𝑣′,𝑖 by using

𝑥 (𝑣′−1)𝑛+𝑖 , 𝑁(𝑡 )(𝑣′−1)𝑛+𝑖 , 𝑦

(𝑡 )(𝑣′−1)𝑛+𝑖 , respectively. Then, 𝑘𝑡 (𝑥) ← [

√︃𝑁(𝑡 )1𝐾 (𝑥, 𝑥1), . . . ,

√︃𝑁(𝑡 )𝑛𝑉𝐾 (𝑥, 𝑥𝑛𝑉 )]⊤ for

any 𝑥 ∈ ˜X. 𝐾 (𝑡 ) ← [√︃𝑁(𝑡 )𝑖𝑁(𝑡 )𝑗𝐾 (𝑥𝑖 , 𝑥 𝑗 )]𝑖, 𝑗 ∈[𝑛𝑉 ] . �̄� (𝑡 ) ← [

√︃𝑁(𝑡 )1𝑦(𝑡 )1, . . . ,

√︃𝑁(𝑡 )𝑛𝑉𝑦(𝑡 )𝑛𝑉]⊤ ; // organize

overall observation information

12 for all 𝑣 ′ ∈ [𝑉 ] do13 Δ̂𝑡 (𝑥𝑖 , 𝑥 𝑗 ) ← (𝑘𝑡 (𝑥𝑖 ) − 𝑘𝑡 (𝑥 𝑗 ))⊤ (𝐾 (𝑡 ) + 𝑁 (𝑡 )b𝑡 𝐼 )−1�̄� (𝑡 ) , ∀𝑥𝑖 , 𝑥 𝑗 ∈ B (𝑡 )𝑣′ ; // estimate the reward gap

between �̃�𝑖 and �̃� 𝑗

14 B (𝑡+1)𝑣′ ← B (𝑡 )

𝑣′ \ {𝑥 ∈ B(𝑡 )𝑣′ |∃𝑥

′ ∈ B (𝑡 )𝑣′ : Δ̂𝑡 (𝑥 ′, 𝑥) ≥ 2

−𝑡 } ; // discard sub-optimal arms

15 𝑡 ← 𝑡 + 1;

16 E return B (𝑡 )1, . . . ,B (𝑡 )

𝑉;

allocation _ ∈ △˜X into the integer numbers of samples ^ ∈ N | ˜X |

, such that

∑�̃� ∈ ˜X ^�̃� = 𝑁 and

max

�̃�𝑖 ,�̃� 𝑗 ∈B (𝑡 )𝑣′ ,𝑣′∈[𝑉 ]

∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥2(𝑁b𝑡 𝐼+∑�̃�∈ ˜X ^�̃�𝜙 (�̃�)𝜙 (�̃�)⊤)−1

≤ (1 + 𝜖) max

�̃�𝑖 ,�̃� 𝑗 ∈B (𝑡 )𝑣′ ,𝑣′∈[𝑉 ]

∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥2(𝑁b𝑡 𝐼+∑�̃�∈ ˜X 𝑁_�̃�𝜙 (�̃�)𝜙 (�̃�)⊤)−1.

By calling ROUNDY (_∗𝑡 , 𝑁 (𝑡 ) ), agent 𝑣 generates an overall sample sequence (𝑠1, . . . , 𝑠𝑁 (𝑡 ) ) according to _∗𝑡 , and extractsa sub-sequence �̃� (𝑡 )𝑣 that only contains the arms in

˜X𝑣 to sample (Lines 6-8). After sampling, she only communicates

the number of samples 𝑁(𝑡 )𝑣,𝑖

and average observed reward 𝑦(𝑡 )𝑣,𝑖

for each arm with other agents (Lines 9-10). With the

overall observation information, she estimates the reward gap Δ̂𝑡 (𝑥𝑖 , 𝑥 𝑗 ) between any arm pair 𝑥𝑖 , 𝑥 𝑗 ∈ B (𝑡 )𝑣′ for all

𝑣 ′ ∈ [𝑉 ] and discards sub-optimal arms (Lines 13-14).

4.1.2 Computation and Communication Efficiency. Here we explain the efficiency of CoopKernelFC. Note that in

CoPE-KB, due to its high-dimensional reward structures, directly using the empirical mean to estimate rewards will

7

Woodstock ’21, June 03–05, 2021, Woodstock, NY

cause loose sample complexity, and naively calculating and transmitting infinite-dimensional parameter \∗ will incur

huge computation and communication costs. As a result, we cannot directly compute and communicate scalar empirical

rewards as in prior CoPE with classic MAB works [20, 22, 38].

Computation Efficiency. CoopKernelFC uses three efficient kernelized operations, i.e., optimization solver (Line 4),

condition for regularization parameter b𝑡 (Eq. (3)) and estimator of reward gaps (Line 13). Unlike prior kernel bandit

algorithms [10, 43] which explicitly compute 𝜙 (𝑥) and maintain the estimate of \∗ on the infinite-dimensional RKHS,

CoopKernelFC only queries kernel function 𝐾 (·, ·) and significantly reduces the computation (memory) costs from

Poly(dim(H𝐾 )) to only Poly(𝑛𝑉 ).Below we give the formal expressions of these operations and defer their detailed derivation to Appendix A.1.

Kernelized Estimator. We first introduce the kernelized estimator of reward gaps (Line 13). Following the standard

estimation procedure in linear/kernel bandits [10, 19, 23, 43], we consider the following regularized least square estimator

of underlying reward parameter \∗

ˆ\𝑡 =©­«𝑁 (𝑡 )b𝑡 𝐼 +

𝑁 (𝑡 )∑︁𝑗=1

𝜙 (𝑠 𝑗 )𝜙 (𝑠 𝑗 )⊤ª®¬−1

𝑁 (𝑡 )∑︁𝑗=1

𝜙 (𝑠 𝑗 )𝑦 𝑗

Note that this form ofˆ\𝑡 has 𝑁

(𝑡 )terms in the summation, which are cumbersome to compute and communicate.

Since the samples (𝑠1, . . . , 𝑠𝑁 (𝑡 ) ) are composed by arms 𝑥1, . . . , 𝑥𝑛𝑉 , we merge repetitive computations for same arms

in the summation and obtain (for notational simplicity, we combine the subscripts 𝑣 ′, 𝑖 in 𝑥𝑣′,𝑖 , 𝑁(𝑡 )𝑣′,𝑖 , 𝑦

(𝑡 )𝑣′,𝑖 by using

𝑥 (𝑣′−1)𝑛+𝑖 , 𝑁(𝑡 )(𝑣′−1)𝑛+𝑖 , 𝑦

(𝑡 )(𝑣′−1)𝑛+𝑖 , respectively)

ˆ\𝑡(a)

=

(𝑁 (𝑡 )b𝑡 𝐼 +

𝑛𝑉∑︁𝑖=1

𝑁(𝑡 )𝑖𝜙 (𝑥𝑖 )𝜙 (𝑥𝑖 )⊤

)−1 𝑛𝑉∑︁𝑖=1

𝑁(𝑡 )𝑖𝜙 (𝑥𝑖 )𝑦 (𝑡 )𝑖

(b)

=Φ⊤𝑡(𝑁 (𝑡 )b𝑡 𝐼 + 𝐾 (𝑡 )

)−1

𝑦 (𝑡 ) . (4)

Here 𝑁(𝑡 )𝑖

is the number of samples and 𝑦(𝑡 )𝑖

is the average observed reward on arm 𝑥𝑖 for any 𝑖 ∈ [𝑛𝑉 ]. Φ𝑡 =

[√︃𝑁(𝑡 )1𝜙 (𝑥1)⊤; . . . ;

√︃𝑁(𝑡 )𝑛𝑉𝜙 (𝑥𝑛𝑉 )⊤] is the empiricallyweighted feature vector,𝐾 (𝑡 ) = Φ𝑡Φ

⊤𝑡 = [

√︃𝑁(𝑡 )𝑖𝑁(𝑡 )𝑗𝐾 (𝑥𝑖 , 𝑥 𝑗 )]𝑖, 𝑗 ∈[𝑛𝑉 ]

is the kernel matrix, and 𝑦 (𝑡 ) = [√︃𝑁(𝑡 )1𝑦(𝑡 )1, . . . ,

√︃𝑁(𝑡 )𝑛𝑉𝑦(𝑡 )𝑛𝑉]⊤ is the average observations. Equality (a) rearranges the

summation according to different chosen arms, and (b) follows from kernel transformation.

Then, by multiplying

(𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )

)⊤, we obtain the estimator of reward gaps 𝑓 (𝑥𝑖 ) − 𝑓 (𝑥 𝑗 ) as

Δ̂(𝑥𝑖 , 𝑥 𝑗 ) =(𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )

)⊤ˆ\𝑡 =

(𝑘𝑡 (𝑥𝑖 ) − 𝑘𝑡 (𝑥 𝑗 )

)⊤ (𝑁 (𝑡 )b𝑡 𝐼 + 𝐾 (𝑡 )

)−1

𝑦 (𝑡 ) , (5)

where 𝑘𝑡 (𝑥) = Φ𝑡𝜙 (𝑥) = [√︃𝑁(𝑡 )1𝐾 (𝑥, 𝑥1), . . . ,

√︃𝑁(𝑡 )𝑛𝑉𝐾 (𝑥, 𝑥𝑛𝑉 )]⊤ for any 𝑥 ∈ ˜X. This estimator not only transforms

heavy operations on the infinite-dimensional RKHS to efficient ones that only query the kernel function, but also

merges repetitive computations for same arms (equality (a)) and only requires calculations dependent on 𝑛𝑉 .

Kernelized Optimization Solver/Condition for Regularization Parameter. Now we introduce the optimization solver

(Line 4) and condition for regularization parameter b𝑡 (Eq. (3)).

For the kernelized optimization solver, we solve the min-max optimization in Line 4 by projected gradient descent,

which follows the procedure in [10]. Specifically, let𝐴(b, _) = b𝐼 +∑�̃� ∈ ˜X _�̃�𝜙 (𝑥)𝜙 (𝑥)

⊤for any b > 0, _ ∈ △

˜X . We define

8

Collaborative Pure Exploration in Kernel Bandit Woodstock ’21, June 03–05, 2021, Woodstock, NY

function ℎ(_) = max

�̃�𝑖 ,�̃� 𝑗 ∈B (𝑡 )𝑣 ,𝑣∈[𝑉 ]∥𝜙 (𝑥𝑖 ) −𝜙 (𝑥 𝑗 )∥2𝐴(b𝑡 ,_)−1

, and denote the optimal solution of ℎ(_) by 𝑥∗𝑖(_), 𝑥∗

𝑗(_). Then,

the gradient of ℎ(_) is given by

[∇_ℎ(_)]�̃� = −((𝜙 (𝑥∗𝑖 (_)) − 𝜙 (𝑥

∗𝑗 (_))

)⊤𝐴(b𝑡 , _)−1𝜙 (𝑥)

)2

, ∀𝑥 ∈ ˜X, (6)

which can be efficiently calculated by the following kernel transformation(𝜙 (𝑥∗𝑖 (_)) − 𝜙 (𝑥

∗𝑗 (_))

)⊤𝐴(b𝑡 , _)−1𝜙 (𝑥)

=b−1

𝑡

(𝐾 (𝑥∗𝑖 (_), 𝑥) − 𝐾 (𝑥

∗𝑗 (_), 𝑥) −

(𝑘_ (𝑥∗𝑖 (_)) − 𝑘_ (𝑥

∗𝑗 (_))

)⊤(b𝑡 𝐼 + 𝐾_)−1 𝑘_ (𝑥)

), (7)

where 𝑘_ (𝑥) = [ 1√_1

𝐾 (𝑥, 𝑥1), . . . , 1

_𝑛𝑉𝐾 (𝑥, 𝑥𝑛𝑉 )]⊤ and 𝐾_𝑢 = [ 1√

_𝑖_ 𝑗𝐾 (𝑥𝑖 , 𝑥 𝑗 )]𝑖, 𝑗 ∈[𝑛𝑉 ] .

For condition Eq. (3) on the regularization parameter b𝑡 , we can transform it to

max

�̃�𝑖 ,�̃� 𝑗 ∈ ˜X𝑣𝑣∈[𝑉 ]

√︂(𝐾 (𝑥𝑖 , 𝑥𝑖 ) + 𝐾 (𝑥 𝑗 , 𝑥 𝑗 ) − 2𝐾 (𝑥𝑖 , 𝑥 𝑗 )

)− ∥𝑘_𝑢 (𝑥𝑖 ) − 𝑘_𝑢 (𝑥 𝑗 )∥2(b𝑡 𝐼+𝐾_𝑢 )−1

≤ 1

(1 + Y)𝐵 · 2𝑡+1, (8)

where _𝑢 = [ 1

𝑛𝑉, . . . , 1

𝑛𝑉]⊤ is the uniform distribution on

˜X, 𝑘_𝑢 (𝑥) = [1√𝑛𝑉𝐾 (𝑥, 𝑥1), . . . , 1√

𝑛𝑉𝐾 (𝑥, 𝑥𝑛𝑉 )]⊤ and 𝐾_𝑢 =

[ 1

𝑛𝑉𝐾 (𝑥𝑖 , 𝑥 𝑗 )]𝑖, 𝑗 ∈[𝑛𝑉 ] .

Both the kernelized optimization solver and condition for b𝑡 avoid inefficient operations directly on infinite-

dimensional RKHS by querying the kernel function, and only cost Poly(𝑛𝑉 ) computation (memory) complexity

(Eqs. (7),(8) only contains scalar 𝐾 (𝑥𝑖 , 𝑥 𝑗 ), 𝑛𝑉 -dimensional vector 𝑘_ and 𝑛𝑉 × 𝑛𝑉 -dimensional matrix 𝐾_).

Communication Efficiency. By taking advantage of the kernelized estimator (Eq. (5)), CoopKernelFC merges repeti-

tive computations for the same arms and only transmits 𝑛𝑉 scalar tuples {(𝑁 (𝑡 )𝑣,𝑖, 𝑦(𝑡 )𝑣,𝑖)}𝑖∈[𝑛],𝑣∈[𝑉 ] among agents instead

of transmitting all 𝑁 (𝑡 ) samples as in [16]. This significantly reduces the communication cost from 𝑂 (𝑁 (𝑡 ) ) bits to𝑂 (𝑛𝑉 ) bits (Lines 9-10).

4.2 Theoretical performance of CoopKernelFC

Define the problem hardness of identifying the best arms 𝑥∗𝑣 for all 𝑣 ∈ [𝑉 ] as

𝜌∗ = min

_∈△ ˜Xmax

�̃� ∈X𝑣 ,𝑣∈[𝑉 ]

∥𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥)∥2(b∗𝐼+∑�̃�∈ ˜X _�̃�𝜙 (�̃�)𝜙 (�̃�)⊤)−1

(𝑓 (𝑥∗𝑣 ) − 𝑓 (𝑥))2, (9)

where b∗ = min𝑡 ≥1 b𝑡 . 𝜌∗is the information-theoretic lower bound of the CoPE-KB problem, which is adapted from

linear/kernel bandit pure exploration [10, 19, 23, 43]. Let 𝑆 denote the per-agent sample complexity, i.e., average number

of samples used by each agent in algorithm CoopKernelFC.

The sample complexity and number of communication rounds of CoopKernelFC are as follows.

Theorem 1 (Fixed-Confidence Upper Bound). With probability at least 1 − 𝛿 , algorithm CoopKernelFC returns the

correct answers 𝑥∗𝑣 for all 𝑣 ∈ [𝑉 ], with per-agent sample complexity

𝑆 = 𝑂

(𝜌∗

𝑉· logΔ−1

min

(log

(𝑛𝑉

𝛿

)+ log logΔ−1

min

))and communication rounds 𝑂 (logΔ−1

min).

9

Woodstock ’21, June 03–05, 2021, Woodstock, NY

Remark 1. 𝜌∗ is comprised of two sources of problem hardness, one due to handling different tasks and the other due to

distinguishing different arms (We will decompose the sample complexity into these two parts in Corollary 1(c)). We see

that the sample complexity ofCoopKernelFCmatches the lower bound (up to logarithmic factors). For fully-collaborative

instances where single-agent algorithms [19, 23] have �̃� (𝜌∗ log𝛿−1) sample complexity, our CoopKernelFC achieves

the maximum𝑉 -speedup (i.e., enjoys �̃� ( 𝜌∗

𝑉log𝛿−1) sample complexity) using only logarithmic communication rounds.

Interpretation. We further interpret Theorem 1 via standard expressive tools in kernel bandits [14, 35, 40], i.e.,

effective dimension and maximum information gain, to characterize the relationship between sample complexity and

data structures, and demonstrate how task similarity influences learning performance.

To this end, define the maximum information gain over all sample allocation _ ∈ △˜X as

Υ = max

_∈△ ˜Xlog det

(𝐼 + b∗−1𝐾_

).

Denote _∗ = argmax_∈△ ˜Xlog det

(𝐼 + b∗−1𝐾_

)and 𝛼1 ≥ · · · ≥ 𝛼𝑛𝑉 the eigenvalues of 𝐾_∗ , and define the effective

dimension of 𝐾_∗ as

𝑑eff

= min

{𝑗 : 𝑗b∗ log(𝑛𝑉 ) ≥

𝑛𝑉∑︁𝑖=𝑗+1

𝛼𝑖

}.

We then have the following corollary.

Corollary 1. The per-agent sample complexity of algorithm CoopKernelFC, denoted by 𝑆 , can also be bounded as

follows:

(a) 𝑆 = 𝑂

Δ2

min𝑉· 𝑔(Δmin, 𝛿)

), where Υ is the maximum information gain.

(b) 𝑆 = 𝑂

(𝑑eff

Δ2

min𝑉· log

(𝑛𝑉 ·

(1 + Trace(𝐾_∗ )

b∗𝑑eff

))𝑔(Δmin, 𝛿)

), where 𝑑

effis the effective dimension.

(c) 𝑆 = 𝑂

(rank(𝐾𝑧) · rank(𝐾𝑥 )

Δ2

min𝑉

· log

(Trace

(𝐼 + b−1

∗ 𝐾_∗)

rank(𝐾_∗ )

)𝑔(Δmin, 𝛿)

).

Here 𝑔(Δmin, 𝛿) = logΔ−1

min

(log

(𝑛𝑉𝛿

)+ log logΔ−1

min

).

Remark 2. Corollary 1(a) shows that, our sample complexity can be bounded by the maximum information gain of

any sample allocation on˜X, which extends conventional information-gain-based results in regret minimization kernel

bandits [13, 16, 35] to the pure exploration setting in the view of experimental (allocation) design.

In terms of dimension dependency, it is demonstrated in Corollary 1(b) that our result only depends on the effective

dimension of kernel representation, which is the number of principle directions that data projections in RKHS spread.

We also provide a fundamental decomposition of sample complexity into two compositions from task similarities and

arm features in Corollary 1(c), which shows that the more tasks are similar, the fewer samples we need for accomplishing

all tasks. For example, when tasks are the same (fully-collaborative), i.e., rank(𝐾𝑧) = 1, each agent only spends a1

𝑉

fraction of samples used by single-agent algorithms [10, 43]. Conversely, when the tasks are totally different, i.e.,

rank(𝐾𝑧) = 𝑉 , no advantage can be attained by multi-agent deployments, since the information from neighboring

agents is useless for solving local tasks.

10

Collaborative Pure Exploration in Kernel Bandit Woodstock ’21, June 03–05, 2021, Woodstock, NY

4.3 Lower Bound for Fixed-Confidence Setting

We now present lower bounds for the sample complexity and a round-speedup for fully-collaborative instances, using

a novel measure transformation techniques. The bounds validate the optimality of CoopKernelFC in both sampling

and communication. Specifically, Theorems 2 and 3 below formally present our bounds. In the theorems, we refer to a

distributed algorithm A for CoPE-KB as 𝛿-correct, if it returns the correct answers 𝑥∗𝑣 for all 𝑣 ∈ [𝑉 ] with probability

at least 1 − 𝛿 .

Theorem 2 (Fixed-Confidence Sample Complexity Lower Bound). Consider the fixed-confidence collaborative

pure exploration in kernel bandit problem with Gaussian noise [𝑣,𝑡 . Given any 𝛿 ∈ (0, 1), a 𝛿-correct distributed algorithmA must have per-agent sample complexity Ω( 𝜌

𝑉log𝛿−1).

Remark 3. Theorem 2 shows that even if the agents are allowed to share samples without limitation, each agent

still requires at least Ω̃( 𝜌∗

𝑉) samples on average. Together with Theorem 1, one sees that CoopKernelFC is within

logarithmic factors of the optimal sampling.

Theorem 3 (Fixed-Confidence Round-Speedup Lower Bound). There exists a fully-collaborative instance of the

fixed-confidence CoPE-KB problem with multi-armed and linear reward structures, for which given any 𝛿 ∈ (0, 1), a𝛿-correct and 𝛽-speedup distributed algorithm A must utilize

Ω©­«

logΔ−1

min

log(1 + 𝑉𝛽) + log logΔ−1

min

ª®¬communication rounds in expectation. In particular, when 𝛽 = 𝑉 , A must require Ω( logΔ−1

min

log logΔ−1

min

) communication rounds in

expectation.

Remark 4. Theorem 3 exhibits that logarithmic communication rounds are indispensable for achieving the full speedup,

which validates that CoopKernelFC is near-optimal in communication. Moreover, when CoPE-KB reduces to prior

CoPE with classic MAB setting [20, 38], i.e., all agents are solving the same classic MAB task, our upper and lower

bounds (Theorems 1 and 3) match the state-of-the-art results in [38].

Novel Analysis for Fixed-Confidence Round-Speedup Lower Bound.We highlight that our round-speedup lower

bound for the FC setting analysis has the following novel aspects. (i) Unlike prior CoPE work [38] which focuses

on a preliminary 2-armed case without considering reward structures, we investigate multi-armed instances with

high-dimensional linear reward structures. (ii) We develop a linear structured progress lemma (Lemma 3 in Appendix A.5),

which effectively handles the challenges due to different possible sample allocation on multiple arms and derives the

required communication rounds under linear reward structures. (iii) We propose multi-armed measure transformation

and linear structured instance transformation lemmas (Lemmas 4,5 in Appendix A.5), which bound the change of

probability measures in instance transformation with multiple arms and high-dimensional linear rewards, and serve as

basic analytical tools in our proof.

5 FIXED-BUDGET COPE-KB

We now turn to the fixed-budget (FB) setting and design an efficient algorithm CoopKernelFB. We also establish a

fixed-budget round-speedup lower bound to validate its communication optimality.

11

Woodstock ’21, June 03–05, 2021, Woodstock, NY

Algorithm 2: Distributed Algorithm CoopKernelFB: for Agent 𝑣 (𝑣 ∈ [𝑉 ])Input: Per-agent budget 𝑇 , ˜X1, . . . , ˜X𝑉 , 𝐾 (·, ·) :

˜X × ˜X ↦→ R, regularization parameter b∗, rounding procedure

ROUNDY (·, ·) with approximation parameter Y.

1 Initialization: 𝑅 ←⌈log

2(𝜔 ( ˜X))

⌉. 𝑁 ← ⌊𝑇𝑉 /𝑅⌋. B (1)

𝑣′ ← X𝑣′ for all 𝑣′ ∈ [𝑉 ]. 𝑡 ← 1 ; // pre-determine the

number of phases and the number of samples for each phase

2 while 𝑡 ≤ 𝑅 and ∃𝑣 ′ ∈ [𝑉 ], |B (𝑡 )𝑣′ | > 1 do

3 Let _∗𝑡 and 𝜌∗𝑡 be the optimal solution and optimal value of

min

_∈△ ˜Xmax

�̃�𝑖 ,�̃� 𝑗 ∈B (𝑡 )𝑣′ ,𝑣′∈[𝑉 ]

∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥2(b∗𝐼+∑�̃�∈ ˜X _�̃�𝜙 (�̃�)𝜙 (�̃�)⊤)−1; // compute the optimal sample allocation

4 (𝑠1, . . . , 𝑠𝑁 (𝑡 ) ) ← ROUNDY (_∗𝑡 , 𝑁 (𝑡 ) );5 Let �̃� (𝑡 )𝑣 be the sub-sequence of (𝑠1, . . . , 𝑠𝑁 ) which only contains the arms in

˜X𝑣 ; // generate the sample

sequence for agent 𝑣

6 Pull arms �̃� (𝑡 )𝑣 and observe random rewards 𝒚 (𝑡 )𝑣 ;

7 Broadcast {(𝑁 (𝑡 )𝑣,𝑖, 𝑦(𝑡 )𝑣,𝑖)}𝑖∈[𝑛] ;

8 Receive {(𝑁 (𝑡 )𝑣′,𝑖 , 𝑦

(𝑡 )𝑣′,𝑖 )}𝑖∈[𝑛] from all other agents 𝑣 ′ ∈ [𝑉 ] \ {𝑣};

9 𝑘𝑡 (𝑥) ← [√︃𝑁(𝑡 )1𝐾 (𝑥, 𝑥1), . . . ,

√︃𝑁(𝑡 )𝑛𝑉𝐾 (𝑥, 𝑥𝑛𝑉 )]⊤ for any 𝑥 ∈ ˜X. 𝐾 (𝑡 ) ← [

√︃𝑁(𝑡 )𝑖𝑁(𝑡 )𝑗𝐾 (𝑥𝑖 , 𝑥 𝑗 )]𝑖, 𝑗 ∈[𝑛𝑉 ] .

�̄� (𝑡 ) ← [√︃𝑁(𝑡 )1𝑦(𝑡 )1, . . . ,

√︃𝑁(𝑡 )𝑛𝑉𝑦(𝑡 )𝑛𝑉]⊤ ; // organize overall observation information

10 for all 𝑣 ′ ∈ [𝑉 ] do11 ˆ𝑓𝑡 (𝑥) ← 𝑘𝑡 (𝑥)⊤ (𝐾 (𝑡 ) + 𝑁 (𝑡 )b∗𝐼 )−1�̄� (𝑡 ) for all 𝑥 ∈ B (𝑡 )

𝑣′ ; // estimate the rewards of alive arms

12 Sort all 𝑥 ∈ B (𝑡 )𝑣′ by

ˆ𝑓𝑡 (𝑥) in decreasing order. Let 𝑥 (1) , . . . , 𝑥 ( |B (𝑡 )𝑣′ |)

denote the sorted arm sequence;

13 Let 𝑖𝑡+1 be the largest index such that 𝜔 ({𝑥 (1) , . . . , 𝑥 (𝑖𝑡+1) }) ≤ 𝜔 (B(𝑡 )𝑣′ )/2;

14 B (𝑡+1)𝑣′ ← {𝑥 (1) , . . . , 𝑥 (𝑖𝑡+1) } ; // cut down the alive arm set to half dimension

15 𝑡 ← 𝑡 + 1;

16 return B (𝑡 )1, . . . ,B (𝑡 )

𝑉;

5.1 Algorithm CoopKernelFB

5.1.1 Algorithm. CoopKernelFB consists of three key steps: (i) pre-determine the numbers of phases and samples

according to data dimension, (ii) maintain alive arm sets for all agents, plan a globally optimal sample allocation, (iii)

communicate observation information and cut down alive arms to a half in the dimension sense.

The procedure of CoopKernelFB is given in Algorithm 2. During initialization, we determine the number of phases

𝑅 and the number of samples for each phase 𝑁 according to the principle dimension 𝜔 ( ˜X) (Line 1), defined as:

𝜔 ( ˜S) = min

_∈△ ˜Xmax

�̃�𝑖 ,�̃� 𝑗 ∈ ˜S∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥2(b∗𝐼+∑�̃�∈ ˜X _�̃�𝜙 (�̃�)𝜙 (�̃�)⊤)−1

, ∀ ˜S ⊆ ˜X

i.e., the principle dimension of data projections in˜S to the RKHS. In each phase 𝑡 , each agent 𝑣 maintains alive arm sets

B (𝑡 )𝑣′ for all agents 𝑣 ′ ∈ [𝑉 ], and solves an integrated optimization to obtain a globally optimal sample allocation _∗𝑡

(Line 3). Then, she generates a sample sequence (𝑠 (𝑡 )1, . . . , 𝑠

(𝑡 )𝑁) according to _∗𝑡 , and selects the sub-sequence �̃� (𝑡 )𝑣 that

only contains her available arms to perform sampling (Lines 4-5). During communication, she only sends and receives

the number of samples 𝑁(𝑡 )𝑣,𝑖

and average observed reward 𝑦(𝑡 )𝑣,𝑖

for each arm to and from other agents (Lines 7-8). Using

12

Collaborative Pure Exploration in Kernel Bandit Woodstock ’21, June 03–05, 2021, Woodstock, NY

the shared information, she estimates rewards of alive arms and only selects the best half of them in the dimension

sense to enter the next phase (Lines 11-14).

5.1.2 Computation and Communication Efficiency. CoopKernelFB also adopts the efficient kernelized optimization

solver (Eqs. (6),(7)) to solve the min-max optimization in Line 3 and employs the kernelized estimator (Eq. (4)) to

estimate the rewards in Line 11. Moreover, CoopKernelFB only spends Poly(𝑛𝑉 ) computation time and 𝑂 (𝑛𝑉 )-bitcommunication costs.

5.2 Theoretical performance of CoopKernelFB

We present the error probability of CoopKernelFB in the following theorem, where _𝑢 = 1

𝑛𝑉1.

Theorem 4 (Fixed-Budget Upper Bound). Suppose 𝜔 ({𝑥∗𝑣 , 𝑥}) ≥ 1 for any 𝑥 ∈ X𝑣 \ {𝑥∗𝑣 }, 𝑣 ∈ [𝑉 ] and 𝑇 =

Ω(𝜌∗ log(𝜔 ( ˜X))), and the regularization parameter b∗ > 0 satisfies√︁b∗max

�̃�𝑖 ,�̃� 𝑗 ∈ ˜X𝑣 ,𝑣∈[𝑉 ] ∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥𝐴(b∗,_𝑢 )−1 ≤Δmin

2(1+Y)𝐵 . With at most 𝑇 samples per agent, CoopKernelFB returns the correct answers 𝑥∗𝑣 for all 𝑣 ∈ [𝑉 ], with error

probability

𝐸𝑟𝑟 = 𝑂

(𝑛2𝑉 log(𝜔 ( ˜X)) · exp

(− 𝑇𝑉

𝜌∗ log(𝜔 ( ˜X))

))and communication rounds 𝑂 (log(𝜔 ( ˜X))).

Remark 5.Theorem 4 implies that, to guarantee an error probability𝛿 ,CoopKernelFB only requires𝑂 ( 𝜌∗

log(𝜔 ( ˜X))𝑉

log( 𝑛2𝑉 log(𝜔 ( ˜X))

𝛿))

samples, which matches the sample complexity lower bound (Theorem 2) up to logarithmic factors. In addition,

CoopKernelFB attains the maximum 𝑉 -speedup for fully-collaborative instances with only logarithmic communication

rounds, which also matches the round-speedup lower bound (Theorem 5) within double logarithmic factors.

Technical Novelty in Error Probability Analysis. Our analysis extends prior single-agent analysis [23] to the

multi-agent setting. The single-agent analysis in [23] only uses a single universal Gaussian-process concentration

bound. Instead, we establish novel estimate concentrations and high probability events for each arm pair and each

agent to handle the distributed environment, and build a connection between the principle dimension 𝜔 (B (𝑡 )𝑣 ) andproblem hardness 𝜌∗ via elimination rules (Lines 13-14) to guarantee the identification correctness.

Interpretation. Similar to Corollary 1, we can also interpret the error probability result with the standard tools of

maximum information gain and effective dimension in kernel bandits [14, 35, 40], and decompose the error probability

into two compositions from task similarities and arm features.

Corollary 2. The error probability of algorithm CoopKernelFC, denoted by 𝐸𝑟𝑟 , can also be bounded as follows:

(a)𝐸𝑟𝑟 = 𝑂

(exp

(−

𝑇𝑉Δ2

min

Υ log(𝜔 ( ˜X))

)· 𝑛2𝑉 log(𝜔 ( ˜X))

), where Υ is the maximum information gain.

(b)𝐸𝑟𝑟 = 𝑂©­­«exp

©­­«−𝑇𝑉Δ2

min

𝑑eff

log

(𝑛𝑉 ·

(1 + Trace(𝐾_∗ )

b∗𝑑eff

))log(𝜔 ( ˜X))

ª®®¬ · 𝑛2𝑉 log(𝜔 ( ˜X))ª®®¬ , where 𝑑eff is the

effective dimension.

(c)𝐸𝑟𝑟 = 𝑂©­­«exp

©­­«−𝑇𝑉Δ2

min

rank(𝐾𝑧) · rank(𝐾𝑥 ) log

(Trace(𝐼+b−1

∗ 𝐾_∗ )rank(𝐾_∗ )

)log(𝜔 ( ˜X))

ª®®¬ · 𝑛2𝑉 log(𝜔 ( ˜X))ª®®¬ .

13

Woodstock ’21, June 03–05, 2021, Woodstock, NY

Remark 6. Corollaries 2(a), 2(b) bound the error probability by maximum information gain and effective dimension,

respectively, which capture essential structures of tasks and arm features and only depend on the effective dimension of

the feature space of kernel representation. Furthermore, we exhibit how task similarities influence the error probability

performance in Corollary 2(c). For example, in the fully-collaborative case where rank(𝐾𝑧) = 1, the error probability

enjoys an exponential decay factor of 𝑉 compared to conventional single-agent results [23] (achieves a 𝑉 -speedup).

Conversely, when the tasks are totally different with rank(𝐾𝑧) = 𝑉 , the error probability degenerates to conventional

single-agent results [23], since in this case information sharing brings no benefit.

5.3 Lower Bound for Fixed-Budget Setting

In this subsection, we establish a round-speedup lower bound for the FB setting.

Theorem 5 (Fixed-Budget Round-Speedup Lower Bound). There exists a fully-collaborative instance of the fixed-

budget CoPE-KB problem with multi-armed and linear reward structures, for which given any 𝛽 ∈ [ 𝑉

log(𝜔 ( ˜X)),𝑉 ], a

𝛽-speedup distributed algorithm A must utilize

Ω©­« log(𝜔 ( ˜X))

log(𝑉𝛽) + log log(𝜔 ( ˜X))

ª®¬communication rounds in expectation. In particular, when 𝛽 = 𝑉 , A must use Ω( log(𝜔 ( ˜X))

log log(𝜔 ( ˜X))) communication rounds in

expectation.

Remark 7. Theorem 5 shows that under the FB setting, to achieve the full speedup, agents require at least logarithmic

communication rounds with respect to the principle dimension 𝜔 ( ˜X), which validates the communication optimality

of CoopKernelFB. In the degenerated case when all agents solve the same non-structured pure exploration problem,

same as in prior classic MAB setting [20, 38], both our upper (Theorem 4) and lower (Theorem 5) bounds match the

state-of-the-art results in [38].

Novel Analysis for Fixed-Budget Round-Speedup Lower Bound. Different from the FC setting, here we borrow

the proof idea of prior limited adaptivity work [1] to establish a non-trivial lower bound analysis under Bayesian

environments, and perform instance transformation by changing data dimension instead of tuning reward gaps. In

our analysis, we employ novel techniques to calculate the information entropy and support size of posterior reward

distributions in order to build induction among different rounds and derive the required communication rounds.

6 EXPERIMENTS

In this section, we conduct experiments to validate the empirical performance of our algorithms. In our experiments, we

set𝑉 = 5, 𝑑 = 4, 𝑛 = 6, 𝛿 = 0.005 and 𝜙 (𝑥) = 𝐼𝑥 for any 𝑥 ∈ ˜X. The entries of \∗ form an arithmetic sequence that starts

from 0.1 and has the common difference Δmin, i.e., \∗ = [0.1, 0.1 + Δmin, . . . , 0.1 + (𝑑 − 1)Δmin]⊤. For the FC setting, we

vary the gap Δmin ∈ [0.1, 0.8] to generate different instances (points), and run 50 independent simulations to plot the

average sample complexity with 95% confidence intervals. For the FB setting, we change the budget 𝑇 ∈ [7000, 300000]to obtain different instances, and perform 100 independent runs to show the error probability across runs. The specific

values of gap Δmin and budget 𝑇 can be seen in X-axis of the figures.

Fixed-Confidence. In the FC setting (Figures 2(a)-2(c)), we compareCoopKernelFCwith five baselines:CoopKernel-IndAlloc

is an ablation variant ofCoopKernelFCwhich individually calculates sample allocations for different agents. IndRAGE [19],

IndALBA [34] and IndPolyALBA [15] are single-agent algorithms, which use 𝑉 copies of single-agent RAGE [19],

14

Collaborative Pure Exploration in Kernel Bandit Woodstock ’21, June 03–05, 2021, Woodstock, NY

0.10 0.15 0.20 0.25 0.30 0.35 0.40 0.45 0.50Gap

103

104

105

106

107Sa

mpl

e Co

mpl

exity

CoopKernel (ours)CoopKernel-IndAllocIndRAGEIndRAGE-VIndALBAIndPolyALBA

(a) FC, 𝐾Z = 1 (fully-collaborative)

0.10 0.15 0.20 0.25 0.30Gap

103

104

105

106

Sam

ple

Com

plex

ity

CoopKernel (ours)CoopKernel-IndAllocIndRAGEIndRAGE-VIndALBAIndPolyALBA

(b) FC, 1 < 𝐾Z < 𝑉

0.30 0.35 0.40 0.45 0.50 0.55 0.60 0.65 0.70Gap

103

104

105

106

Sam

ple

Com

plex

ity

CoopKernel (ours)CoopKernel-IndAllocIndRAGEIndRAGE-VIndALBAIndPolyALBA

(c) FC, 𝐾Z = 𝑉

5000 5500 6000 6500 7000 7500 8000 8500 9000Budget

0.00.10.20.30.40.50.60.70.80.91.0

Erro

r Pro

babi

lity

CoopKernelFB (ours)CoopKernelFB-IndAllocIndRAGE-FBIndUniformFB

(d) FB, 𝐾Z = 1 (fully-collaborative)

5000 5500 6000 6500 7000 7500 8000 8500 9000Budget

0.00.10.20.30.40.50.60.70.80.91.0

Erro

r Pro

babi

lity

CoopKernelFB (ours)CoopKernelFB-IndAllocIndRAGE-FBIndUniformFB

(e) FB, 1 < 𝐾Z < 𝑉

6000 6500 7000 7500 8000 8500 9000 950010000Budget

0.0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1.0

Erro

r Pro

babi

lity

CoopKernelFB (ours)CoopKernelFB-IndAllocIndRAGE-FBIndUniformFB

(f) FB, 𝐾Z = 𝑉

Fig. 2. Experimental results for FC and FB settings.

ALBA [34] and PolyALBA [15] policies to solve the 𝑉 tasks independently. IndRAGE/𝑉 is a 𝑉 -speedup baseline, which

divides the sample complexity of the best single-agent algorithm IndRAGE by 𝑉 . One can see that CoopKernelFC

achieves the best sample complexity in Figures 2(a),2(b), which demonstrates the effectiveness of our sample allocation

and cooperation scheme. Moreover, the empirical results also reflect the impacts of task similarities on learning speedup,

and keep consistent with our theoretical analysis. Specifically, in the fully-collaborative case (Figure 2(a)),CoopKernelFC

matches IndRAGE-𝑉 since it attains the 𝑉 speedup; in the intermediate (1 < 𝐾Z < 𝑉 ) case (Figure 2(b)), the curve

of CoopKernelFC lies between IndRAGE/𝑉 and IndRAGE, since it only achieves smaller than 𝑉 speedup due to the

decrease of task similarity; in the totally-different-task (𝐾Z = 𝑉 ) case (Figure 2(c)), CoopKernelFC performs similar to

the single-agent algorithm IndRAGE, since information sharing among agents brings no advantage in this case.

Fixed-Budget. In the FB setting (Figures 2(d)-2(f)), we compareCoopKernelFBwith three baselines:CoopKernelFB-IndAlloc

is an ablation variant of CoopKernelFB where agents calculate and use different sample allocations. IndPeaceFB [23]

and IndUniformFB solve the 𝑉 tasks independently by calling 𝑉 copies of single-agent PeaceFB [23] and uniform

sampling policies, respectively. As shown in Figures 2(d),2(e), our CoopKernelFB enjoys a lower error probability than

all other algorithms. In addition, these empirical results also validate the influences of task similarities on learning

performance, and match our theoretical analysis. Specifically, as the task similarity decreases in Figures 2(d) to 2(f), the

error probability of CoopKernelFB gets closer to that of single-agent IndRAGE-FB, due to the slow-down of its learning

speedup.

7 CONCLUSION

In this paper, we propose a novel Collaborative Pure Exploration in Kernel Bandit (CoPE-KB) problem with Fixed-

Confidence (FC) and Fixed-Budget (FB) settings. CoPE-KB aims to coordinate multiple agents to identify best arms with

15

Woodstock ’21, June 03–05, 2021, Woodstock, NY

general reward functions. We design two computation and communication efficient algorithms CoopKernelFC and

CoopKernelFB based on novel kernelized estimators. Matching upper and lower bounds are established to demonstrate

the statistical and communication optimality of our algorithms. Our theoretical results explicitly characterize the

impacts of task similarities on learning speedup and avoid heavy dependency on the high dimension of the kernelized

feature space. In our analysis, we also develop novel analytical techniques, including data dimension decomposition,

linear structured instance transformation and (communication) round-speedup induction, which are applicable to other

bandit problems and can be of independent interests.

REFERENCES[1] Arpit Agarwal, Shivani Agarwal, Sepehr Assadi, and Sanjeev Khanna. 2017. Learning with limited rounds of adaptivity: Coin tossing, multi-armed

bandits, and ranking from pairwise comparisons. In Conference on Learning Theory. PMLR, 39–75.

[2] Zeyuan Allen-Zhu, Yuanzhi Li, Aarti Singh, and Yining Wang. 2021. Near-optimal discrete optimization for experimental design: A regret

minimization approach. Mathematical Programming 186, 1 (2021), 439–478.

[3] Jean-Yves Audibert, Sébastien Bubeck, and Rémi Munos. 2010. Best arm identification in multi-armed bandits.. In Conference on Learning Theory.Citeseer, 41–53.

[4] Peter Auer, Nicolo Cesa-Bianchi, and Paul Fischer. 2002. Finite-time analysis of the multiarmed bandit problem. Machine Learning 47, 2-3 (2002),

235–256.

[5] Ilai Bistritz and Amir Leshem. 2018. Distributed Multi-Player Bandits-a Game of Thrones Approach.. In Advances in Neural Information ProcessingSystems. 7222–7232.

[6] Sébastien Bubeck and Thomas Budzinski. 2020. Coordination without communication: optimal regret in two players multi-armed bandits. In

Conference on Learning Theory. PMLR, 916–939.

[7] Sébastien Bubeck, Thomas Budzinski, and Mark Sellke. 2021. Cooperative and stochastic multi-player multi-armed bandit: Optimal regret with

neither communication nor collisions. In Conference on Learning Theory. PMLR, 821–822.

[8] Sébastien Bubeck and Nicolo Cesa-Bianchi. 2012. Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems. Machine Learning5, 1 (2012), 1–122.

[9] Séebastian Bubeck, Tengyao Wang, and Nitin Viswanathan. 2013. Multiple identifications in multi-armed bandits. In International Conference onMachine Learning. PMLR, 258–265.

[10] Romain Camilleri, Kevin Jamieson, and Julian Katz-Samuels. 2021. High-Dimensional Experimental Design and Kernel Bandits. In InternationalConference on Machine Learning. PMLR, 1227–1237.

[11] Mithun Chakraborty, Kai Yee Phoebe Chua, Sanmay Das, and Brendan Juba. 2017. Coordinated Versus Decentralized Exploration In Multi-Agent

Multi-Armed Bandits.. In Proceedings of the International Joint Conference on Artifical Intelligence. 164–170.[12] Lijie Chen, Jian Li, and Mingda Qiao. 2017. Towards instance optimal bounds for best arm identification. In Conference on Learning Theory. PMLR,

535–592.

[13] Sayak Ray Chowdhury and Aditya Gopalan. 2017. On kernelized multi-armed bandits. In International Conference on Machine Learning. PMLR,

844–853.

[14] Aniket Anand Deshmukh, Urun Dogan, and Clayton Scott. 2017. Multi-task learning for contextual bandits. In Advances in Neural InformationProcessing Systems. 4851–4859.

[15] Yihan Du, Yuko Kuroki, and Wei Chen. 2021. Combinatorial pure exploration with full-bandit or partial linear feedback. In Proceedings of the AAAIConference on Artificial Intelligence, Vol. 35. 7262–7270.

[16] Abhimanyu Dubey et al. 2020. Kernel methods for cooperative multi-agent contextual bandits. In International Conference on Machine Learning.2740–2750.

[17] Thomas Elsken, Jan Hendrik Metzen, and Frank Hutter. 2019. Neural architecture search: A survey. The Journal of Machine Learning Research 20, 1

(2019), 1997–2017.

[18] Eyal Even-Dar, Shie Mannor, Yishay Mansour, and Sridhar Mahadevan. 2006. Action Elimination and Stopping Conditions for the Multi-Armed

Bandit and Reinforcement Learning Problems. Journal of Machine Learning Research 7, 6 (2006).

[19] Tanner Fiez, Lalit Jain, Kevin G Jamieson, and Lillian Ratliff. 2019. Sequential experimental design for transductive linear bandits. Advances inneural information processing systems 32 (2019), 10667–10677.

[20] Eshcar Hillel, Zohar S Karnin, Tomer Koren, Ronny Lempel, and Oren Somekh. 2013. Distributed Exploration in Multi-Armed Bandits. In Advancesin Neural Information Processing Systems, Vol. 26. 854–862.

[21] Shivaram Kalyanakrishnan, Ambuj Tewari, Peter Auer, and Peter Stone. 2012. PAC subset selection in stochastic multi-armed bandits. In InternationalConference on Machine Learning, Vol. 12. 655–662.

[22] Nikolai Karpov, Qin Zhang, and Yuan Zhou. 2020. Collaborative top distribution identifications with limited interaction. In IEEE 61st AnnualSymposium on Foundations of Computer Science. 160–171.

16

Collaborative Pure Exploration in Kernel Bandit Woodstock ’21, June 03–05, 2021, Woodstock, NY

[23] Julian Katz-Samuels, Lalit Jain, Kevin G Jamieson, et al. 2020. An Empirical Process Approach to the Union Bound: Practical Algorithms for

Combinatorial and Linear Bandits. In Advances in Neural Information Processing Systems, Vol. 33.[24] Emilie Kaufmann, Olivier Cappé, and Aurélien Garivier. 2016. On the complexity of best-arm identification in multi-armed bandit models. Journal

of Machine Learning Research 17, 1 (2016), 1–42.

[25] Nathan Korda, Balazs Szorenyi, and Shuai Li. 2016. Distributed clustering of linear bandits in peer to peer networks. In International Conference onMachine Learning. PMLR, 1301–1309.

[26] Andreas Krause and Cheng Soon Ong. 2011. Contextual Gaussian Process Bandit Optimization.. In Advances in Neural Information ProcessingSystems. 2447–2455.

[27] Tze Leung Lai and Herbert Robbins. 1985. Asymptotically efficient adaptive allocation rules. Advances in Applied Mathematics 6, 1 (1985), 4–22.[28] Tor Lattimore and Csaba Szepesvári. 2020. Bandit algorithms. Cambridge University Press.

[29] Keqin Liu and Qing Zhao. 2010. Distributed learning in multi-armed bandit with multiple players. IEEE Transactions on Signal Processing 58, 11

(2010), 5667–5681.

[30] Zhenhua Liu, Minghong Lin, Adam Wierman, Steven H Low, and Lachlan LH Andrew. 2011. Geographical load balancing with renewables. 39, 3

(2011), 62–66.

[31] Trung T. Nguyen. 2021. On the Edge and Cloud: Recommendation Systems with Distributed Machine Learning. In 2021 International Conference onInformation Technology (ICIT). 929–934. https://doi.org/10.1109/ICIT52682.2021.9491121

[32] Jonathan Rosenski, Ohad Shamir, and Liran Szlak. 2016. Multi-player bandits–a musical chairs approach. In International Conference on MachineLearning. PMLR, 155–163.

[33] Bernhard Schölkopf, Alexander J Smola, Francis Bach, et al. 2002. Learning with kernels: support vector machines, regularization, optimization, andbeyond. MIT press.

[34] Marta Soare, Alessandro Lazaric, and Rémi Munos. 2014. Best-arm identification in linear bandits. In Advances in Neural Information ProcessingSystems, Vol. 27. 828–836.

[35] Niranjan Srinivas, Andreas Krause, Sham M Kakade, and Matthias Seeger. 2010. Gaussian process optimization in the bandit setting: No regret and

experimental design. In International Conference on Machine Learning.[36] Balazs Szorenyi, Róbert Busa-Fekete, István Hegedus, Róbert Ormándi, Márk Jelasity, and Balázs Kégl. 2013. Gossip-based distributed stochastic

bandit algorithms. In International Conference on Machine Learning. PMLR, 19–27.

[37] Liang Tang, Romer Rosales, Ajit Singh, and Deepak Agarwal. 2013. Automatic ad format selection via contextual bandits. In Proceedings of the 22ndACM International Conference on Information & Knowledge Management. 1587–1594.

[38] Chao Tao, Qin Zhang, and Yuan Zhou. 2019. Collaborative learning with limited interaction: Tight bounds for distributed exploration in multi-armed

bandits. In IEEE 60th Annual Symposium on Foundations of Computer Science. 126–146.[39] William R Thompson. 1933. On the likelihood that one unknown probability exceeds another in view of the evidence of two samples. Biometrika 25,

3/4 (1933), 285–294.

[40] Michal Valko, Nathan Korda, Rémi Munos, Ilias Flaounas, and Nello Cristianini. 2013. Finite-Time Analysis of Kernelised Contextual Bandits. In

Uncertainty in Artificial Intelligence.[41] Sofía S Villar, Jack Bowden, and James Wason. 2015. Multi-armed bandit models for the optimal design of clinical trials: benefits and challenges.

Statistical Science: A Review Journal of the Institute of Mathematical Statistics 30, 2 (2015), 199.[42] Grace Wahba. 1990. Spline models for observational data. SIAM.

[43] Yinglun Zhu, Dongruo Zhou, Ruoxi Jiang, Quanquan Gu, Rebecca Willett, and Robert Nowak. 2021. Pure Exploration in Kernel and Neural Bandits.

arXiv preprint arXiv:2106.12034 (2021).[44] Ling Zhuo, Cho-Li Wang, and Francis C.M. Lau. 2003. Document replication and distribution in extensible geographically distributed web servers. J.

Parallel and Distrib. Comput. 63, 10 (2003), 927–944.

17

Woodstock ’21, June 03–05, 2021, Woodstock, NY

APPENDIX

A PROOFS FOR THE FIXED-CONFIDENCE SETTING

A.1 Kernelized Computations in Algorithm CoopKernelFC

Kernelized Condition for Regularization Parameter b𝑡 .We first introduce how to compute the condition Eq. (3)

for regularization parameter b𝑡 in Line 4 of Algorithm 1.

Let Φ_ = [√_1𝜙 (𝑥1)⊤; . . . ;

√︁_𝑛𝑉𝜙 (𝑥𝑛𝑉 )⊤] and 𝐾_ = Φ_Φ

⊤_= [

√︁_𝑖_ 𝑗𝐾 (𝑥𝑖 , 𝑥 𝑗 )]𝑖, 𝑗 ∈[𝑛𝑉 ] for any _ ∈ △ ˜X . Let 𝑘_ (𝑥) =

Φ_𝜙 (𝑥) = [√_1𝐾 (𝑥, 𝑥1), . . . ,

√︁_𝑛𝑉𝐾 (𝑥, 𝑥𝑛𝑉 )]⊤ for any _ ∈ △

˜X , 𝑥 ∈ ˜X. Since(b𝑡 𝐼 + Φ⊤_ Φ_

)𝜙 (𝑥) = b𝑡𝜙 (𝑥) + Φ⊤_ 𝑘_ (𝑥)

for any 𝑥 ∈ ˜X, we have

𝜙 (𝑥) =b𝑡(b𝑡 𝐼 + Φ⊤_ Φ_

)−1

𝜙 (𝑥) +(b𝑡 𝐼 + Φ⊤_ Φ_

)−1

Φ⊤_𝑘_ (𝑥)

=b𝑡

(b𝑡 𝐼 + Φ⊤_ Φ_

)−1

𝜙 (𝑥) + Φ⊤_(b𝑡 𝐼 + 𝐾_)−1 𝑘_ (𝑥)

Thus,

𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 ) = b𝑡(b𝑡 𝐼 + Φ⊤_ Φ_

)−1 (𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )

)+ Φ⊤

_(b𝑡 𝐼 + 𝐾_)−1

(𝑘_ (𝑥𝑖 ) − 𝑘_ (𝑥 𝑗 )

)Multiplying

(𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )

)⊤on both sides, we have(

𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 ))⊤ (

𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 ))

=b𝑡(𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )

)⊤ (b𝑡 𝐼 + Φ⊤_ Φ_

)−1 (𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )

)+

(𝑘_ (𝑥𝑖 ) − 𝑘_ (𝑥 𝑗 )

)⊤ (b𝑡 𝐼 + 𝐾_)−1(𝑘_ (𝑥𝑖 ) − 𝑘_ (𝑥 𝑗 )

)Thus,

∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥2(b𝑡 𝐼+Φ⊤_Φ_

)−1

=(𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )

)⊤ (b𝑡 𝐼 + Φ⊤_ Φ_

)−1 (𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )

)=b−1

𝑡

(𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )

)⊤ (𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )

)− b−1

𝑡

(𝑘_ (𝑥𝑖 ) − 𝑘_ (𝑥 𝑗 )

)⊤ (b𝑡 𝐼 + 𝐾_)−1(𝑘_ (𝑥𝑖 ) − 𝑘_ (𝑥 𝑗 )

)=b−1

𝑡

(𝐾 (𝑥𝑖 , 𝑥𝑖 ) + 𝐾 (𝑥 𝑗 , 𝑥 𝑗 ) − 2𝐾 (𝑥𝑖 , 𝑥 𝑗 )

)− b−1

𝑡

(𝑘_ (𝑥𝑖 ) − 𝑘_ (𝑥 𝑗 )

)⊤ (b𝑡 𝐼 + 𝐾_)−1(𝑘_ (𝑥𝑖 ) − 𝑘_ (𝑥 𝑗 )

)=b−1

𝑡

(𝐾 (𝑥𝑖 , 𝑥𝑖 ) + 𝐾 (𝑥 𝑗 , 𝑥 𝑗 ) − 2𝐾 (𝑥𝑖 , 𝑥 𝑗 )

)− b−1

𝑡 ∥𝑘_ (𝑥𝑖 ) − 𝑘_ (𝑥 𝑗 )∥2(b𝑡 𝐼+𝐾_)−1

Let _𝑢 = 1

𝑛𝑉1 be the uniform distribution on

˜X. Then, the condition Eq. (3) for regularization parameter b𝑡√︁b𝑡 max

�̃�𝑖 ,�̃� 𝑗 ∈ ˜X𝑣 ,𝑣∈[𝑉 ]∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥(b𝑡 𝐼+∑𝑥∈X 1

𝑛𝑉𝜙 (𝑥)𝜙 (𝑥)⊤)−1 ≤ 1

(1 + Y)𝐵 · 2𝑡+1

is equivalent to the following efficient kernelized statement

max

�̃�𝑖 ,�̃� 𝑗 ∈ ˜X𝑣 ,𝑣∈[𝑉 ]

√︂(𝐾 (𝑥𝑖 , 𝑥𝑖 ) + 𝐾 (𝑥 𝑗 , 𝑥 𝑗 ) − 2𝐾 (𝑥𝑖 , 𝑥 𝑗 )

)− ∥𝑘_𝑢 (𝑥𝑖 ) − 𝑘_𝑢 (𝑥 𝑗 )∥2(b𝑡 𝐼+𝐾_𝑢 )−1

≤ 1

(1 + Y)𝐵 · 2𝑡+1.

Kernelized Optimization Solver. Now we present the efficient kernelized optimization solver for the following

convex optimization in Line 4 of Algorithm 1:

min

_∈△ ˜Xmax

�̃�𝑖 ,�̃� 𝑗 ∈B (𝑡 )𝑣 ,𝑣∈[𝑉 ]∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥2𝐴(b𝑡 ,_)−1

, (10)

18

Collaborative Pure Exploration in Kernel Bandit Woodstock ’21, June 03–05, 2021, Woodstock, NY

where 𝐴(b, _) = b𝐼 +∑�̃� ∈ ˜X _�̃�𝜙 (𝑥)𝜙 (𝑥)

⊤for any b > 0, _ ∈ △

˜X .

Define function ℎ(_) = max

�̃�𝑖 ,�̃� 𝑗 ∈B (𝑡 )𝑣 ,𝑣∈[𝑉 ]∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥2𝐴(b𝑡 ,_)−1

, and define 𝑥∗𝑖(_), 𝑥∗

𝑗(_) as the optimal solution of

ℎ(_). Then, the gradient of ℎ(_) with respect to _ is

[∇_ℎ(_)]�̃� = −((𝜙 (𝑥∗𝑖 (_)) − 𝜙 (𝑥

∗𝑗 (_))

)⊤𝐴(b𝑡 , _)−1𝜙 (𝑥)

)2

,∀𝑥 ∈ ˜X. (11)

Next, we show how to efficiently compute gradient [∇_ℎ(_)]�̃� with kernel function 𝐾 (·, ·).Since

(b𝑡 𝐼 + Φ⊤_ Φ_

)𝜙 (𝑥) = b𝑡𝜙 (𝑥) + Φ⊤_ 𝑘_ (𝑥) for any 𝑥 ∈ ˜X, we have

𝜙 (𝑥) =b𝑡(b𝑡 𝐼 + Φ⊤_ Φ_

)−1

𝜙 (𝑥) +(b𝑡 𝐼 + Φ⊤_ Φ_

)−1

Φ⊤_𝑘_ (𝑥)

=b𝑡

(b𝑡 𝐼 + Φ⊤_ Φ_

)−1

𝜙 (𝑥) + Φ⊤_(b𝑡 𝐼 + 𝐾_)−1 𝑘_ (𝑥)

Multiplying

(𝜙 (𝑥∗

𝑖(_)) − 𝜙 (𝑥∗

𝑗(_))

)⊤on both sides, we have(

𝜙 (𝑥∗𝑖 (_)) − 𝜙 (𝑥∗𝑗 (_))

)⊤𝜙 (𝑥)

=b𝑡

(𝜙 (𝑥∗𝑖 (_)) − 𝜙 (𝑥

∗𝑗 (_))

)⊤ (b𝑡 𝐼 + Φ⊤_ Φ_

)−1

𝜙 (𝑥) +(𝑘_ (𝑥∗𝑖 (_)) − 𝑘_ (𝑥

∗𝑗 (_))

)⊤(b𝑡 𝐼 + 𝐾_)−1 𝑘_ (𝑥)

Then, (𝜙 (𝑥∗𝑖 (_)) − 𝜙 (𝑥

∗𝑗 (_))

)⊤ (b𝑡 𝐼 + Φ⊤_ Φ_

)−1

𝜙 (𝑥)

=b−1

𝑡

(𝜙 (𝑥∗𝑖 (_)) − 𝜙 (𝑥

∗𝑗 (_))

)⊤𝜙 (𝑥) − b−1

𝑡

(𝑘_ (𝑥∗𝑖 (_)) − 𝑘_ (𝑥

∗𝑗 (_))

)⊤(b𝑡 𝐼 + 𝐾_)−1 𝑘_ (𝑥)

=b−1

𝑡

(𝐾 (𝑥∗𝑖 (_), 𝑥) − 𝐾 (𝑥

∗𝑗 (_), 𝑥) −

(𝑘_ (𝑥∗𝑖 (_)) − 𝑘_ (𝑥

∗𝑗 (_))

)⊤(b𝑡 𝐼 + 𝐾_)−1 𝑘_ (𝑥)

)(12)

Therefore, we can compute gradient ∇_ℎ(_) (Eq. (11)) using the equivalent kernelized expression Eq. (12), and then the

optimization (Eq. (10)) can be efficiently solved by projected gradient descent.

Innovative Kernelized Estimator. Finally, we explicate the innovative kernelized estimator of reward gaps in Line 13

of Algorithm 1, which plays an important role in boosting the computation and communication efficiency.

Letˆ\𝑡 denote the minimizer of the following regularized least square loss function:

L(\ ) = 𝑁 (𝑡 )b𝑡 ∥\ ∥2 +𝑁 (𝑡 )∑︁𝑗=1

(𝑦 𝑗 − 𝜙 (𝑠 𝑗 )⊤\ )2 .

Letting the derivative of L(\ ) equal to zero, we have

𝑁 (𝑡 )b𝑡 ˆ\𝑡 +𝑁 (𝑡 )∑︁𝑗=1

𝜙 (𝑥 𝑗 )𝜙 (𝑥 𝑗 )⊤ ˆ\𝑡 =

𝑁 (𝑡 )∑︁𝑗=1

𝜙 (𝑥 𝑗 )𝑦 𝑗 .

Rearranging the summation, we can obtain

𝑁 (𝑡 )b𝑡 ˆ\𝑡 +(𝑛𝑉∑︁𝑖=1

𝑁(𝑡 )𝑖𝜙 (𝑥𝑖 )𝜙 (𝑥𝑖 )⊤

)ˆ\𝑡 =

𝑛𝑉∑︁𝑖=1

𝑁(𝑡 )𝑖𝜙 (𝑥𝑖 )𝑦 (𝑡 )𝑖 , (13)

19

Woodstock ’21, June 03–05, 2021, Woodstock, NY

where 𝑁(𝑡 )𝑖

is the number of samples and 𝑦(𝑡 )𝑖

is the average observation on arm 𝑥𝑖 for any 𝑖 ∈ [𝑛𝑉 ]. Let Φ𝑡 =

[√︃𝑁(𝑡 )1𝜙 (𝑥1)⊤; . . . ;

√︃𝑁(𝑡 )𝑛𝑉𝜙 (𝑥𝑛𝑉 )⊤],𝐾 (𝑡 ) = Φ𝑡Φ

⊤𝑡 = [

√︃𝑁(𝑡 )𝑖𝑁(𝑡 )𝑗𝐾 (𝑥𝑖 , 𝑥 𝑗 )]𝑖, 𝑗 ∈[𝑛𝑉 ] and𝑦 (𝑡 ) = [

√︃𝑁(𝑡 )1𝑦(𝑡 )1, . . . ,

√︃𝑁(𝑡 )𝑛𝑉𝑦(𝑡 )𝑛𝑉]⊤.

Then, we can write Eq. (13) as (𝑁 (𝑡 )b𝑡 𝐼 + Φ⊤𝑡 Φ𝑡

)ˆ\𝑡 =Φ

⊤𝑡 𝑦(𝑡 ) .

Since

(𝑁 (𝑡 )b𝑡 𝐼 + Φ⊤𝑡 Φ𝑡

)≻ 0 and

(𝑁 (𝑡 )b𝑡 𝐼 + Φ𝑡Φ⊤𝑡

)≻ 0,

ˆ\𝑡 =

(𝑁 (𝑡 )b𝑡 𝐼 + Φ⊤𝑡 Φ𝑡

)−1

Φ⊤𝑡 𝑦(𝑡 )

= Φ⊤𝑡(𝑁 (𝑡 )b𝑡 𝐼 + Φ𝑡Φ⊤𝑡

)−1

𝑦 (𝑡 )

= Φ⊤𝑡(𝑁 (𝑡 )b𝑡 𝐼 + 𝐾 (𝑡 )

)−1

𝑦 (𝑡 ) .

Let 𝑘𝑡 (𝑥) = Φ𝑡𝜙 (𝑥) = [√︃𝑁(𝑡 )1𝐾 (𝑥, 𝑥1), . . . ,

√︃𝑁(𝑡 )𝑛𝑉𝐾 (𝑥, 𝑥𝑛𝑉 )]⊤ for any 𝑥 ∈ X. Then, we obtain the efficient kernelized

estimators of 𝑓 (𝑥𝑖 ) and 𝑓 (𝑥𝑖 ) − 𝑓 (𝑥 𝑗 ) as

ˆ𝑓 (𝑥𝑖 ) = 𝜙 (𝑥𝑖 )⊤ ˆ\𝑡 = 𝑘𝑡 (𝑥𝑖 )⊤(𝑁 (𝑡 )b𝑡 𝐼 + 𝐾 (𝑡 )

)−1

𝑦 (𝑡 ) ,

Δ̂(𝑥𝑖 , 𝑥 𝑗 ) =(𝑘𝑡 (𝑥𝑖 ) − 𝑘𝑡 (𝑥 𝑗 )

)⊤ (𝑁 (𝑡 )b𝑡 𝐼 + 𝐾 (𝑡 )

)−1

𝑦 (𝑡 ) .

A.2 Proof of Theorem 1

Our proof of Theorem 1 adapts the analysis procedure of [19, 23] to the multi-agent setting.

For any _ ∈ △˜X , let Φ_ = [

√_1𝜙 (𝑥∗𝑣 )⊤; . . . ;

√︁_𝑛𝑉𝜙 (𝑥𝑛𝑉 )⊤] and Φ⊤

_Φ_ =

∑�̃� ∈ ˜X _�̃�𝜙 (𝑥)𝜙 (𝑥)

⊤. In order to prove

Theorem 1, we first introduce the following lemmas.

Lemma 1 (Concentration). Defining event

G =

{ ���( ˆ𝑓𝑡 (𝑥𝑖 ) − ˆ𝑓𝑡 (𝑥 𝑗 ))−

(𝑓 (𝑥𝑖 ) − 𝑓 (𝑥 𝑗 )

) ��� < (1 + Y) · ∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥ (b𝑡 𝐼+Φ⊤_∗𝑡Φ_∗𝑡)−1 ·

©­«√︄

2 log

(2𝑛2𝑉 /𝛿𝑡

)𝑁𝑡

+√︁b𝑡 · ∥\∗∥2

ª®¬ ≤ 2−𝑡 , ∀𝑥𝑖 , 𝑥 𝑗 ∈ B (𝑡 )𝑣 , ∀𝑣 ∈ [𝑉 ], ∀𝑡 ≥ 1

},

we have

Pr [G] ≥ 1 − 𝛿.

Proof of Lemma 1. Letˆ\𝑡 be the regularized least square estimator of \∗ with samples 𝑥∗𝑣 , . . . , 𝑥𝑁𝑡 and 𝛾𝑡 = 𝑁𝑡 b𝑡 .

Recall that Φ𝑡 = [√︃𝑁(𝑡 )1𝜙 (𝑥∗𝑣 )⊤; . . . ;

√︃𝑁(𝑡 )𝑛𝑉𝜙 (𝑥𝑛𝑉 )⊤] and Φ⊤𝑡 Φ𝑡 =

∑𝑛𝑉𝑖=1

𝑁(𝑡 )𝑖𝜙 (𝑥𝑖 )𝜙 (𝑥𝑖 )⊤.

In addition, 𝑦 (𝑡 ) = [√︃𝑁(𝑡 )1𝑦(𝑡 )1, . . . ,

√︃𝑁(𝑡 )𝑛𝑉𝑦(𝑡 )𝑛𝑉]⊤ and

ˆ\𝑡 =(𝛾𝑡 𝐼 + Φ⊤𝑡 Φ𝑡

)−1 Φ⊤𝑡 𝑦(𝑡 )

.

Let [̄ (𝑡 ) = [√︃𝑁(𝑡 )1[̄(𝑡 )1, . . . ,

√︃𝑁(𝑡 )𝑛𝑉[̄(𝑡 )𝑛𝑉]⊤, where [̄ (𝑡 )

𝑖= 𝑦(𝑡 )𝑖− 𝜙 (𝑥𝑖 )⊤\∗ denote the average noise of the 𝑁 (𝑡 )𝑖 pulls

on arm 𝑥𝑖 for any 𝑖 ∈ [𝑛𝑉 ].Then, (

ˆ𝑓𝑡 (𝑥𝑖 ) − ˆ𝑓𝑡 (𝑥 𝑗 ))−

(𝑓 (𝑥𝑖 ) − 𝑓 (𝑥 𝑗 )

)20

Collaborative Pure Exploration in Kernel Bandit Woodstock ’21, June 03–05, 2021, Woodstock, NY

=(𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )

)⊤ (ˆ\𝑡 − \∗

)=

(𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )

)⊤ ( (𝛾𝑡 𝐼 + Φ⊤𝑡 Φ𝑡

)−1

Φ⊤𝑡 𝑦(𝑡 ) − \∗

)=

(𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )

)⊤ ( (𝛾𝑡 𝐼 + Φ⊤𝑡 Φ𝑡

)−1

Φ⊤𝑡(Φ𝑡\∗ + [̄ (𝑡 )

)− \∗

)=

(𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )

)⊤ ( (𝛾𝑡 𝐼 + Φ⊤𝑡 Φ𝑡

)−1

Φ⊤𝑡 Φ𝑡\∗ +

(𝛾𝑡 𝐼 + Φ⊤𝑡 Φ𝑡

)−1

Φ⊤𝑡 [̄(𝑡 ) − \∗

)=

(𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )

)⊤ ( (𝛾𝑡 𝐼 + Φ⊤𝑡 Φ𝑡

)−1(Φ⊤𝑡 Φ𝑡 + 𝛾𝑡 𝐼

)\∗ +

(𝛾𝑡 𝐼 + Φ⊤𝑡 Φ𝑡

)−1

Φ⊤𝑡 [̄(𝑡 )

(14)

− \∗ − 𝛾𝑡(𝛾𝑡 𝐼 + Φ⊤𝑡 Φ𝑡

)−1

\∗)

=(𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )

)⊤ (𝛾𝑡 𝐼 + Φ⊤𝑡 Φ𝑡

)−1

Φ⊤𝑡 [̄(𝑡 ) − 𝛾𝑡

(𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )

)⊤ (𝛾𝑡 𝐼 + Φ⊤𝑡 Φ𝑡

)−1

\∗ (15)

Since the mean of the first term is zero and its variance is bounded by(𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )

)⊤ (𝛾𝑡 𝐼 + Φ⊤𝑡 Φ𝑡

)−1

Φ⊤𝑡 Φ𝑡(𝛾𝑡 𝐼 + Φ⊤𝑡 Φ𝑡

)−1(𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )

)≤

(𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )

)⊤ (𝛾𝑡 𝐼 + Φ⊤𝑡 Φ𝑡

)−1(𝛾𝑡 𝐼 + Φ⊤𝑡 Φ𝑡

) (𝛾𝑡 𝐼 + Φ⊤𝑡 Φ𝑡

)−1(𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )

)=

(𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )

)⊤ (𝛾𝑡 𝐼 + Φ⊤𝑡 Φ𝑡

)−1(𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )

)=∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥2(𝛾𝑡 𝐼+Φ⊤𝑡 Φ𝑡 )−1

,

using the Hoeffding inequality, we have that with probability at least 1 − 𝛿𝑡𝑛2𝑉

,��� (𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 ))⊤ (𝛾𝑡 𝐼 + Φ⊤𝑡 Φ𝑡

)−1

Φ⊤𝑡 [(𝑡 )𝑣

��� < ∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥(𝛾𝑡 𝐼+Φ⊤𝑡 Φ𝑡 )−1

√︄2 log

(2𝑛2𝑉

𝛿𝑡

)Thus, with probability at least 1 − 𝛿𝑡

𝑛2𝑉,���( ˆ𝑓𝑡 (𝑥𝑖 ) − ˆ𝑓𝑡 (𝑥 𝑗 )

)−

(𝑓 (𝑥𝑖 ) − 𝑓 (𝑥 𝑗 )

) ���<∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥(𝛾𝑡 𝐼+Φ⊤𝑡 Φ𝑡 )−1

√︄2 log

(2𝑛2𝑉

𝛿𝑡

)+ 𝛾𝑡 ∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥(𝛾𝑡 𝐼+Φ⊤𝑡 Φ𝑡 )−1 ∥\∗∥(𝛾𝑡 𝐼+Φ⊤𝑡 Φ𝑡 )−1

≤∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥(𝛾𝑡 𝐼+Φ⊤𝑡 Φ𝑡 )−1

√︄2 log

(2𝑛2𝑉

𝛿𝑡

)+ √𝛾𝑡 · ∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥(𝛾𝑡 𝐼+Φ⊤𝑡 Φ𝑡 )−1 ∥\∗∥2

(a)

≤(1 + Y) · ∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥ (b𝑡 𝐼+Φ⊤

_∗𝑡Φ_∗𝑡)−1

√𝑁𝑡

√︄2 log

(2𝑛2𝑉

𝛿𝑡

)+

√︁b𝑡𝑁𝑡 ·

(1 + Y) · ∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥ (b𝑡 𝐼+Φ⊤_∗𝑡Φ_∗𝑡)−1

√𝑁𝑡

· ∥\∗∥2

=(1 + Y) · ∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥ (b𝑡 𝐼+Φ⊤_∗𝑡Φ_∗𝑡)−1

√︄2 log

(2𝑛2𝑉 /𝛿𝑡

)𝑁𝑡

+ (1 + Y)√︁b𝑡 · ∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥ (b𝑡 𝐼+Φ⊤

_∗𝑡Φ_∗𝑡)−1 ∥\∗∥2

≤(1 + Y) max

�̃�𝑖 ,�̃� 𝑗 ∈ ˜B (𝑡 )𝑣 ,𝑣∈[𝑉 ]∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥ (b𝑡 𝐼+Φ⊤

_∗𝑡Φ_∗𝑡)−1

√︄2 log

(2𝑛2𝑉 /𝛿𝑡

)𝑁𝑡

21

Woodstock ’21, June 03–05, 2021, Woodstock, NY

+ (1 + Y)√︁b𝑡 max

�̃�𝑖 ,�̃� 𝑗 ∈ ˜B (𝑡 )𝑣 ,𝑣∈[𝑉 ]∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥ (b𝑡 𝐼+Φ⊤

_∗𝑡Φ_∗𝑡)−1 ∥\∗∥2,

where (a) is due to the rounding procedure.

According to the choice of b𝑡 , it holds that

(1 + Y)√︁b𝑡 max

�̃�𝑖 ,�̃� 𝑗 ∈ ˜B (𝑡 )𝑣 ,𝑣∈[𝑉 ]∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥ (

b𝑡 𝐼+Φ⊤_∗𝑡Φ_∗𝑡

)−1 ∥\∗∥2

≤(1 + Y)√︁b𝑡 max

�̃�𝑖 ,�̃� 𝑗 ∈ ˜B (𝑡 )𝑣 ,𝑣∈[𝑉 ]∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥ (

b𝑡 𝐼+Φ⊤_𝑢Φ_𝑢)−1 ∥\∗∥2

≤(1 + Y)√︁b𝑡 max

�̃�𝑖 ,�̃� 𝑗 ∈ ˜X𝑣 ,𝑣∈[𝑉 ]∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥ (

b𝑡 𝐼+Φ⊤_𝑢Φ_𝑢)−1 · 𝐵

≤ 1

2𝑡+1 .

Thus, with probability at least 1 − 𝛿𝑡𝑛2𝑉

,���( ˆ𝑓𝑡 (𝑥𝑖 ) − ˆ𝑓𝑡 (𝑥 𝑗 ))−

(𝑓 (𝑥𝑖 ) − 𝑓 (𝑥 𝑗 )

) ���<(1 + Y) max

�̃�𝑖 ,�̃� 𝑗 ∈ ˜B (𝑡 )𝑣 ,𝑣∈[𝑉 ]∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥ (b𝑡 𝐼+Φ⊤

_∗𝑡Φ_∗𝑡)−1

√︄2 log

(2𝑛2𝑉 /𝛿𝑡

)𝑁𝑡

+ 1

2𝑡+1

=

√︄2(1 + Y)2𝜌∗𝑡 log

(2𝑛2𝑉 /𝛿𝑡

)𝑁𝑡

+ 1

2𝑡+1

≤ 1

2𝑡+1 +

1

2𝑡+1

=1

2𝑡

By a union bound over arms 𝑥𝑖 , 𝑥 𝑗 , agent 𝑣 and phase 𝑡 , we have that

Pr [G] ≥ 1 − 𝛿.

For any 𝑡 > 1 and 𝑣 ∈ [𝑉 ], let S (𝑡 )𝑣 = {𝑥 ∈ ˜X𝑣 : 𝑓 (𝑥∗𝑣 ) − 𝑓 (𝑥) ≤ 2−𝑡+2}.

Lemma 2. Assume that event G occurs. Then, for any phase 𝑡 > 1 and agent 𝑣 ∈ [𝑉 ], we have that 𝑥∗𝑣 ∈ B(𝑡 )𝑣 and

B (𝑡 )𝑣 ⊆ S (𝑡 )𝑣 .

Proof of Lemma 2. We prove the first statement by induction.

To begin, for any 𝑣 ∈ [𝑉 ], 𝑥∗𝑣 ∈ B(1)𝑣 trivially holds.

Suppose that 𝑥∗𝑣 ∈ B(𝑡 )𝑣 holds for any 𝑣 ∈ [𝑉 ], and there exists some 𝑣 ′ ∈ [𝑉 ] such that 𝑥∗

𝑣′ ∉ B(𝑡+1)𝑣′ . According to

the elimination rule of algorithm CoopKernelFC, we have that these exists some 𝑥 ′ ∈ B (𝑡 )𝑣 such that

ˆ𝑓𝑡 (𝑥 ′) − ˆ𝑓𝑡 (𝑥∗𝑣′) ≥ 2−𝑡 .

Using Lemma 1, we have

𝑓 (𝑥 ′) − 𝑓 (𝑥∗𝑣′) > ˆ𝑓𝑡 (𝑥 ′) − ˆ𝑓𝑡 (𝑥∗𝑣′) − 2−𝑡 ≥ 0,

22

Collaborative Pure Exploration in Kernel Bandit Woodstock ’21, June 03–05, 2021, Woodstock, NY

which contradicts the definition of 𝑥∗𝑣′ . Thus, we have that for any 𝑣 ∈ [𝑉 ], 𝑥

∗𝑣 ∈ B

(𝑡+1)𝑣 , which completes the proof of

the first statement.

Now, we prove the second statement also by induction.

To begin, we prove that for any 𝑣 ∈ [𝑉 ], B (2)𝑣 ⊆ S (2)𝑣 . Suppose that there exists some 𝑣 ′ ∈ [𝑉 ] such that B (2)𝑣′ ⊊ S (2)

𝑣′ .

Then, there exists some 𝑥 ′ ∈ B (2)𝑣′ such that 𝑓 (𝑥∗

𝑣′) − 𝑓 (𝑥′) > 2

−2+2 = 1. Using Lemma 1, we have that at the phase

𝑡 = 1,

ˆ𝑓𝑡 (𝑥∗𝑣′) − ˆ𝑓𝑡 (𝑥 ′) ≥ 𝑓 (𝑥∗𝑣′) − 𝑓 (𝑥′) − 2

−1 > 1 − 2−1 = 2

−1,

which implies that 𝑥 ′ should have been eliminated in phase 𝑡 = 1 and gives a contradiction.

Suppose that B (𝑡 )𝑣 ⊆ S (𝑡 )𝑣 (𝑡 > 1) holds for any 𝑣 ∈ [𝑉 ], and there exists some 𝑣 ′ ∈ [𝑉 ] such that B (𝑡+1)𝑣′ ⊊ S (𝑡+1)

𝑣′ .

Then, there exists some 𝑥 ′ ∈ B (𝑡+1)𝑣′ such that 𝑓 (𝑥∗

𝑣′) − 𝑓 (𝑥′) > 2

−(𝑡+1)+2 = 2 · 2−𝑡 . Using Lemma 1, we have that at the

phase 𝑡 ,

ˆ𝑓𝑡 (𝑥∗𝑣′) − ˆ𝑓𝑡 (𝑥 ′) ≥ 𝑓 (𝑥∗𝑣′) − 𝑓 (𝑥′) − 2

−𝑡 > 2 · 2−𝑡 − 2−𝑡 = 2

−𝑡 ,

which implies that 𝑥 ′ should have been eliminated in phase 𝑡 and gives a contradiction. Thus, we complete the proof of

Lemma 2. □

Now we prove Theorem 1.

Proof of Theorem 1. We first prove the correctness.

Let 𝑡∗ =⌈log

2Δ−1

min

⌉+ 1 be the index of the last phase of algorithm CoopKernelFC. According to Lemma 2, when

𝑡 = 𝑡∗, B (𝑡 )𝑣 = {𝑥∗𝑣 } holds for any 𝑣 ∈ [𝑉 ], and thus algorithm CoopKernelFC returns the correct answer 𝑥∗𝑣 for all

𝑣 ∈ [𝑉 ].Next, we prove the sample complexity.

In algorithm CoopKernelFC, the computation of _∗𝑡 , 𝜌∗𝑡 and 𝑁𝑡 is the same for all agents, and each agent 𝑣 just

generates partial samples that belong to his arm set˜X𝑣 from the total 𝑁𝑡 samples. Hence, to bound the overall sample

complexity, it suffices to bound

∑𝑡∗𝑡=1

𝑁𝑡 , and then we can obtain the per-agent sample complexity by dividing 𝑉 . Let

Y = 0.1. We have

𝑡∗∑︁𝑡=1

𝑁𝑡

=

𝑡∗∑︁𝑡=1

(8(2𝑡 )2 (1 + Y)2𝜌∗𝑡 log

(2𝑛2𝑉

𝛿𝑡

)+ 1

)

=

𝑡∗∑︁𝑡=2

8(2𝑡 )2(2−𝑡+2

)2

(1 + Y)2min_∈△X max

�̃�𝑖 ,�̃� 𝑗 ∈B (𝑡 )𝑣 ,𝑣∈[𝑉 ] ∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥2(b𝑡 𝐼+Φ⊤_Φ_

)−1(2−𝑡+2)2

log

(4𝑉𝑛2𝑡2

𝛿

)+ 𝑁1 + 𝑡∗

≤𝑡∗∑︁𝑡=2

©­­­«128(1 + Y)2min_∈△X max

�̃� ∈B (𝑡 )𝑣 ,𝑣∈[𝑉 ] ∥𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥)∥2(

b𝑡 𝐼+Φ⊤_Φ_)−1(

2−𝑡+2)2

log

(4𝑉𝑛2 (𝑡∗)2

𝛿

)ª®®®¬ + 𝑁1 + 𝑡∗

23

Woodstock ’21, June 03–05, 2021, Woodstock, NY

≤𝑡∗∑︁𝑡=2

©­­­«128(1 + Y)2 min

_∈△Xmax

�̃� ∈B (𝑡 )𝑣 ,𝑣∈[𝑉 ]

∥𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥)∥2(b𝑡 𝐼+Φ⊤_Φ_

)−1

(𝑓 (𝑥∗𝑣 ) − 𝑓 (𝑥))2log

(4𝑉𝑛2 (𝑡∗)2

𝛿

)ª®®®¬ + 𝑁1 + 𝑡∗

≤𝑡∗∑︁𝑡=2

©­­­«128(1 + Y)2 min

_∈△Xmax

�̃� ∈ ˜X𝑣 ,𝑣∈[𝑉 ]

∥𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥)∥2(b∗𝐼+Φ⊤_Φ_

)−1

(𝑓 (𝑥∗𝑣 ) − 𝑓 (𝑥))2log

(4𝑉𝑛2 (𝑡∗)2

𝛿

)ª®®®¬ + 𝑁1 + 𝑡∗

≤𝑡∗ ·(128(1 + Y)2𝜌∗ log

(4𝑉𝑛2 (𝑡∗)2

𝛿

))+ 𝑁1 + 𝑡∗

=𝑂

(𝜌∗ · logΔ−1

min

(log

(𝑉𝑛

𝛿

)+ log logΔ−1

min

))Thus, the per-agent sample complexity is bounded by

𝑂

(𝜌∗

𝑉· logΔ−1

min

(log

(𝑉𝑛

𝛿

)+ log logΔ−1

min

)).

Since algorithm CoopKernelFC has at most 𝑡∗ =⌈log

2Δ−1

min

⌉+ 1 phases, the number of communication rounds is

bounded by 𝑂 (logΔ−1

min). □

A.3 Proof of Corollary 1

Proof of Corollary 1. Recall that𝐾_ = Φ_Φ⊤_and _∗ = argmax_∈△ ˜X

log det

(𝐼 + b∗−1𝐾_

).We have that log det

(𝐼 + b∗−1𝐾_

)=

log det

(𝐼 + b∗−1Φ⊤

_Φ_

)= log det

(𝐼 + b∗−1

∑�̃� ′∈ ˜X _�̃� ′𝜙 (𝑥

′)𝜙 (𝑥 ′)⊤). Then,

𝜌∗ = min

_∈△Xmax

�̃� ∈ ˜X𝑣 ,𝑣∈[𝑉 ]

∥𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥)∥2(b∗𝐼+Φ⊤_Φ_

)−1

(𝑓 (𝑥∗𝑣 ) − 𝑓 (𝑥))2

≤ min

_∈△Xmax

�̃� ∈ ˜X𝑣 ,𝑣∈[𝑉 ]

∥𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥)∥2(b∗𝐼+Φ⊤_Φ_

)−1

Δ2

min

=1

Δ2

min

· min

_∈△Xmax

�̃� ∈ ˜X𝑣 ,𝑣∈[𝑉 ]∥𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥)∥2(

b∗𝐼+Φ⊤_Φ_)−1

≤ 1

Δ2

min

· min

_∈△X

(2 max

�̃� ∈ ˜X∥𝜙 (𝑥)∥ (

b∗𝐼+Φ⊤_Φ_)−1

)2

=4

Δ2

min

· min

_∈△Xmax

�̃� ∈ ˜X∥𝜙 (𝑥)∥2(

b∗𝐼+∑�̃�′∈ ˜X _�̃�′𝜙 (�̃� ′)𝜙 (�̃� ′)⊤

)−1

=4

Δ2

min

·max

�̃� ∈ ˜X∥𝜙 (𝑥)∥2(

b∗𝐼+∑�̃�′∈ ˜X _

∗�̃�′𝜙 (�̃�

′)𝜙 (�̃� ′)⊤)−1

(b)

=4

Δ2

min

·∑︁�̃� ∈ ˜X

_∗�̃�∥𝜙 (𝑥)∥2(

b∗𝐼+∑�̃�′∈ ˜X _

∗�̃�′𝜙 (�̃�

′)𝜙 (�̃� ′)⊤)−1,

where (b) is due to Lemma 9.

24

Collaborative Pure Exploration in Kernel Bandit Woodstock ’21, June 03–05, 2021, Woodstock, NY

Since _∗�̃�∥𝜙 (𝑥)∥2(

b∗𝐼+∑�̃�′∈ ˜X _

∗�̃�′𝜙 (�̃�

′)𝜙 (�̃� ′)⊤)−1≤ 1 for any 𝑥 ∈ X,

∑︁�̃� ∈ ˜X

_∗�̃�∥𝜙 (𝑥)∥2(

b∗𝐼+∑�̃�′∈ ˜X _

∗�̃�′𝜙 (�̃�

′)𝜙 (�̃� ′)⊤)−1

≤2

∑︁�̃� ∈ ˜X

log

©­«1 + _∗�̃�∥𝜙 (𝑥)∥2(

b∗𝐼+∑�̃�′∈ ˜X _

∗�̃�′𝜙 (�̃�

′)𝜙 (�̃� ′)⊤)−1

ª®¬(c)

≤2 log

det

(b∗𝐼 +

∑�̃� ∈ ˜X _

∗�̃�𝜙 (𝑥)𝜙 (𝑥)⊤

)det (b∗𝐼 )

=2 log det

©­«𝐼 + b−1

∗∑︁�̃� ∈ ˜X

_∗�̃�𝜙 (𝑥)𝜙 (𝑥)⊤ª®¬

=2 log det

(𝐼 + b−1

∗ 𝐾_∗),

where (c) comes from Lemma 10.

Thus, we have

𝜌∗ = min

_∈△Xmax

�̃� ∈ ˜X𝑣 ,𝑣∈[𝑉 ]

∥𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥)∥2(b∗𝐼+Φ⊤_Φ_

)−1

(𝑓 (𝑥∗𝑣 ) − 𝑓 (𝑥))2

≤ 4

Δ2

min

·∑︁�̃� ∈ ˜X

_∗�̃�∥𝜙 (𝑥)∥2(

b∗𝐼+∑�̃�′∈ ˜X _

∗�̃�′𝜙 (�̃�

′)𝜙 (�̃� ′)⊤)−1

≤ 8

Δ2

min

· log det

(𝐼 + b−1

∗ 𝐾_∗)

(16)

In the following, we interpret the term log det

(𝐼 + b−1

∗ 𝐾_∗)using two standard expressive tools, i.e., maximum

information gain and effective dimension, respectively.

Maximum Information Gain. Recall that the maximum information gain over all sample allocation _ ∈ △˜X is defined

as

Υ = max

_∈△ ˜Xlog det

(𝐼 + b∗−1𝐾_

).

Then, using Eq. (16) and the definitions of _∗, the per-agent sample complexity is bounded by

𝑂

(𝜌∗

𝑉· logΔ−1

min

(log

(𝑉𝑛

𝛿

)+ log logΔ−1

min

))=𝑂

©­­«log det

(𝐼 + b∗−1𝐾_∗

)Δ2

min𝑉

· logΔ−1

min

(log

(𝑉𝑛

𝛿

)+ log logΔ−1

min

)ª®®¬=𝑂

Δ2

min𝑉· logΔ−1

min

(log

(𝑉𝑛

𝛿

)+ log logΔ−1

min

))

25

Woodstock ’21, June 03–05, 2021, Woodstock, NY

Effective Dimension. Recall that 𝛼1 ≥ · · · ≥ 𝛼𝑛𝑉 denote the eigenvalues of 𝐾_∗ in decreasing order. The effective

dimension of 𝐾_∗ is defined as

𝑑eff

= min

{𝑗 : 𝑗b∗ log(𝑛𝑉 ) ≥

𝑛𝑉∑︁𝑖=𝑗+1

𝛼𝑖

},

and it holds that 𝑑effb∗ log(𝑛𝑉 ) ≥ ∑𝑛𝑉

𝑖=𝑑eff+1 𝛼𝑖 .

Let Y = 𝑑effb∗ log(𝑛𝑉 )−∑𝑛𝑉

𝑖=𝑑eff+1 𝛼𝑖 , and thus Y ≤ 𝑑effb∗ log(𝑛𝑉 ). Then, we have∑𝑑eff𝑖=1

𝛼𝑖 = Trace(𝐾_∗ )−∑𝑛𝑉𝑖=𝑑eff+1 𝛼𝑖 =

Trace(𝐾_∗ ) − 𝑑effb∗ log(𝑛𝑉 ) + Y and ∑𝑛𝑉𝑖=𝑑eff+1 𝛼𝑖 = 𝑑effb∗ log(𝑛𝑉 ) − Y.

log det

(𝐼 + b−1

∗ 𝐾_∗)

= log

(Π𝑛𝑉𝑖=1

(1 + b−1

∗ 𝛼𝑖))

= log

(Π𝑑eff𝑖=1

(1 + b−1

∗ 𝛼𝑖)· Π𝑛𝑉

𝑖=𝑑eff+1

(1 + b−1

∗ 𝛼𝑖))

≤ log

((1 + b−1

∗ ·Trace(𝐾_∗ ) − 𝑑effb∗ log(𝑛𝑉 ) + Y

𝑑eff

)𝑑eff (1 + b−1

∗ ·𝑑effb∗ log(𝑛𝑉 ) − Y𝑛𝑉 − 𝑑

eff

)𝑛𝑉−𝑑eff )≤𝑑

efflog

(1 + b−1

∗ ·Trace(𝐾_∗ ) − 𝑑effb∗ log(𝑛𝑉 ) + Y

𝑑eff

)+ log

(1 + 𝑑eff log(𝑛𝑉 )

𝑛𝑉 − 𝑑eff

)𝑛𝑉−𝑑eff=𝑑

efflog

(1 + b−1

∗ ·Trace(𝐾_∗ ) − 𝑑effb∗ log(𝑛𝑉 ) + Y

𝑑eff

)+ log

(1 + 𝑑eff log(𝑛𝑉 − 𝑑

eff+ 𝑑

eff)

𝑛𝑉 − 𝑑eff

)𝑛𝑉−𝑑eff(d)

≤𝑑eff

log

(1 + b−1

∗ ·Trace(𝐾_∗ ) − 𝑑effb∗ log(𝑛𝑉 ) + Y

𝑑eff

)+ log

(1 + 𝑑eff log(𝑛𝑉 + 𝑑

eff)

𝑛𝑉

)𝑛𝑉=𝑑

efflog

(1 + b−1

∗ ·Trace(𝐾_∗ ) − 𝑑effb∗ log(𝑛𝑉 ) + Y

𝑑eff

)+ 𝑛𝑉 log

(1 + 𝑑eff log(𝑛𝑉 + 𝑑

eff)

𝑛𝑉

)≤𝑑

efflog

(1 + Trace(𝐾_∗ )

b∗𝑑eff

)+ 𝑑

efflog(𝑛𝑉 + 𝑑

eff)

≤𝑑eff

log

(2𝑛𝑉 ·

(1 + Trace(𝐾_∗ )

b∗𝑑eff

)),

where inequality (d) is due to that

(1 + 𝑑eff log(𝑥+𝑑eff)

𝑥

)𝑥is monotonically increasing with respect to 𝑥 ≥ 1.

Then, using Eq. (16), the per-agent sample complexity is bounded by

𝑂

(𝜌∗

𝑉· logΔ−1

min

(log

(𝑉𝑛

𝛿

)+ log logΔ−1

min

))=𝑂

©­­«log det

(𝐼 + b∗−1𝐾_∗

)Δ2

min𝑉

· logΔ−1

min

(log

(𝑉𝑛

𝛿

)+ log logΔ−1

min

)ª®®¬=𝑂

(𝑑eff

Δ2

min𝑉· log

(𝑛𝑉 ·

(1 + Trace(𝐾_∗ )

b∗𝑑eff

))· logΔ−1

min

(log

(𝑉𝑛

𝛿

)+ log logΔ−1

min

))Decomposition. Let 𝐾 = [𝐾 (𝑥𝑖 , 𝑥 𝑗 )]𝑖, 𝑗 ∈[𝑛𝑉 ] , 𝐾𝑧 = [𝐾𝑧 (𝑧𝑣, 𝑧𝑣′)]𝑣,𝑣′∈[𝑉 ] and 𝐾𝑥 = [𝐾𝑥 (𝑥𝑖 , 𝑥 𝑗 )]𝑖, 𝑗 ∈[𝑛𝑉 ] . Since kernelfunction 𝐾 (¤,¤) is a Hadamard composition of 𝐾𝑧 (¤,¤) and 𝐾𝑥 (¤,¤), it holds that rank(𝐾_∗ ) = rank(𝐾) ≤ rank(𝐾𝑧) · rank(𝐾𝑥 ).

26

Collaborative Pure Exploration in Kernel Bandit Woodstock ’21, June 03–05, 2021, Woodstock, NY

log det

(𝐼 + b−1

∗ 𝐾_∗)

= log

(Π𝑛𝑉𝑖=1

(1 + b−1

∗ 𝛼𝑖))

= log

(Πrank(𝐾_∗ )𝑖=1

(1 + b−1

∗ 𝛼𝑖))

≤ log

©­«∑rank(𝐾_∗ )𝑖=1

(1 + b−1

∗ 𝛼𝑖)

rank(𝐾_∗ )ª®¬rank(𝐾_∗ )

=rank(𝐾_∗ ) log

©­«∑rank(𝐾_∗ )𝑖=1

(1 + b−1

∗ 𝛼𝑖)

rank(𝐾_∗ )ª®¬

≤rank(𝐾𝑧) · rank(𝐾𝑥 ) log

(Trace

(𝐼 + b−1

∗ 𝐾_∗)

rank(𝐾_∗ )

)Then, using Eq. (16), the per-agent sample complexity is bounded by

𝑂

(𝜌∗

𝑉· logΔ−1

min

(log

(𝑉𝑛

𝛿

)+ log logΔ−1

min

))=𝑂

©­­«log det

(𝐼 + b∗−1𝐾_∗

)Δ2

min𝑉

· logΔ−1

min

(log

(𝑉𝑛

𝛿

)+ log logΔ−1

min

)ª®®¬=𝑂

(rank(𝐾𝑧) · rank(𝐾𝑥 )

Δ2

min𝑉

· log

(Trace

(𝐼 + b−1

∗ 𝐾_∗)

rank(𝐾_∗ )

)· logΔ−1

min

(log

(𝑉𝑛

𝛿

)+ log logΔ−1

min

))□

A.4 Proof of Theorem 2

Proof of Theorem 2. Our proof of Theorem 2 adapts the analysis procedure of [19] to the multi-agent setting.

Suppose thatA is a 𝛿-correct algorithm for CoPE-KB. For any 𝑖 ∈ [𝑛𝑉 ], let a\ ∗,𝑖 = N(𝜙 (𝑥𝑖 )⊤\∗, 1) denote the rewarddistribution of arm 𝑥𝑖 , and𝑇𝑖 denote the number of times arm 𝑥𝑖 is pulled by algorithmA. Let Θ = {\ ∈ H𝐾 : ∃𝑣, ∃𝑥 ∈˜X𝑣 \ {𝑥∗𝑣 },

(𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥)

)⊤\ < 0}.

Since A is 𝛿-correct, according to the “Change of Distribution” lemma (Lemma A.3) in [24], we have that for any

\ ∈ Θ,𝑛𝑉∑︁𝑖=1

E[𝑇𝑖 ] · KL(a\ ∗,𝑖 , a\,𝑖 ) ≥ log(1/2.4𝛿) .

Thus, we have

min

\ ∈Θ

𝑛𝑉∑︁𝑖=1

E[𝑇𝑖 ] · KL(a\ ∗,𝑖 , a\,𝑖 ) ≥ log(1/2.4𝛿) .

Let 𝒕∗ be the optimal solution of the following optimization problem

min

𝑛𝑉∑︁𝑖=1

𝑡𝑖

27

Woodstock ’21, June 03–05, 2021, Woodstock, NY

𝑠 .𝑡 . min

\ ∈H𝐾

𝑛𝑉∑︁𝑖=1

𝑡𝑖 · KL(a\ ∗,𝑖 , a\,𝑖 ) ≥ log(1/2.4𝛿) .

Then, we have

min

\ ∈H𝐾

𝑛𝑉∑︁𝑖=1

𝑡∗𝑖∑𝑛𝑉𝑗=1

𝑡∗𝑗

· KL(a\ ∗,𝑖 , a\,𝑖 ) ≥log(1/2.4𝛿)∑𝑛𝑉

𝑗=1𝑡∗𝑗

≥ log(1/2.4𝛿)∑𝑛𝑉𝑗=1E[𝑇𝑗 ]

.

Since

∑𝑛𝑉𝑖=1

𝑡∗𝑖∑𝑛𝑉𝑗=1𝑡∗𝑗

= 1,

max

_∈△Xmin

\ ∈Θ

𝑛𝑉∑︁𝑖=1

_𝑖 · KL(a\ ∗,𝑖 , a\,𝑖 ) ≥log(1/2.4𝛿)∑𝑛𝑉𝑖=1E[𝑇𝑖 ]

.

Thus, we have

𝑛𝑉∑︁𝑖=1

E[𝑇𝑖 ] ≥log(1/2.4𝛿)

max_∈△X min\ ∈Θ∑𝑛𝑉𝑖=1

_𝑖 · KL(a\ ∗,𝑖 , a\,𝑖 )

= log(1/2.4𝛿) min

_∈△Xmax

\ ∈Θ1∑𝑛𝑉

𝑖=1_𝑖 · KL(a\ ∗,𝑖 , a\,𝑖 )

(17)

For _ ∈ △X , let𝐴(b∗, _) = b∗𝐼 +∑𝑛𝑉𝑖=1

_𝑖𝜙 (𝑥𝑖 )𝜙 (𝑥𝑖 )⊤. For any _ ∈ △X , 𝑣 ∈ [𝑉 ], 𝑗 ∈ [𝑛𝑉 ] such that 𝑥 𝑗 ∈ ˜X𝑣 \ {𝑥∗𝑣 }, define

\ 𝑗 (_) = \∗ −(2(𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥 𝑗 ))⊤\∗) · 𝐴(b∗, _)−1 (𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥 𝑗 ))(𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥 𝑗 ))⊤𝐴(b∗, _)−1 (𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥 𝑗 ))

Here (𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥 𝑗 ))⊤\ 𝑗 (_) = −(𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥 𝑗 ))⊤\∗ < 0, and thus \ 𝑗 (_) ∈ Θ.The KL-divergence between a\ ∗,𝑖 and a\ 𝑗 (_),𝑖 is

KL(a\ ∗,𝑖 , a\ 𝑗 (_),𝑖 )

=1

2

(𝜙 (𝑥𝑖 )⊤ (\∗ − \ 𝑗 (_))

)2

=2(𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥 𝑗 ))⊤\∗)2 (𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥 𝑗 ))⊤ · 𝐴(b∗, _)−1𝜙 (𝑥𝑖 )𝜙 (𝑥𝑖 )⊤𝐴(b∗, _)−1 (𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥 𝑗 ))(

(𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥 𝑗 ))⊤𝐴(b∗, _)−1 (𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥 𝑗 )))2

,

and thus,

𝑛𝑉∑︁𝑖=1

_𝑖 · KL(a\ ∗,𝑖 , a\ 𝑗 (_),𝑖 )

=2(𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥 𝑗 ))⊤\∗)2 (𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥 𝑗 ))⊤ · 𝐴(b∗, _)−1 (∑𝑛𝑉𝑖=1

_𝑖𝜙 (𝑥𝑖 )𝜙 (𝑥𝑖 )⊤)𝐴(b∗, _)−1 (𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥 𝑗 ))((𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥 𝑗 ))⊤𝐴(b∗, _)−1 (𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥 𝑗 ))

)2

≤2(𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥 𝑗 ))⊤\∗)2

(𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥 𝑗 ))⊤𝐴(b∗, _)−1 (𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥 𝑗 ))

=2(𝑓 (𝑥∗𝑣 ) − 𝑓 (𝑥 𝑗 ))2

(𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥 𝑗 ))⊤𝐴(b∗, _)−1 (𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥 𝑗 ))(18)

28

Collaborative Pure Exploration in Kernel Bandit Woodstock ’21, June 03–05, 2021, Woodstock, NY

Let J = { 𝑗 ∈ [𝑛𝑉 ] : 𝑥 𝑗 ≠ 𝑥∗𝑣 ,∀𝑣 ∈ [𝑉 ]} denote the set of indices of all sub-optimal arms. Then, plugging the above

Eq. (18) into Eq. (17), we have

𝑛𝑉∑︁𝑖=1

E[𝑇𝑖 ] ≥ log(1/2.4𝛿) min

_∈△Xmax

\ ∈Θ1∑𝑛𝑉

𝑖=1_𝑖 · KL(a\ ∗,𝑖 , a\,𝑖 )

≥ log(1/2.4𝛿) min

_∈△Xmax

𝑗 ∈J1∑𝑛𝑉

𝑖=1_𝑖 · KL(a\ ∗,𝑖 , a\ 𝑗 (_),𝑖 )

≥ log(1/2.4𝛿) min

_∈△Xmax

�̃� 𝑗 ∈ ˜X𝑣\{𝑥∗𝑣 },𝑣∈[𝑉 ]

(𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥 𝑗 ))⊤𝐴(b∗, _)−1 (𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥 𝑗 ))2(𝑓 (𝑥∗𝑣 ) − 𝑓 (𝑥 𝑗 ))2

=1

2

log(1/2.4𝛿)𝜌∗,

which completes the proof of Theorem 2.

A.5 Proof of Theorem 3

Our proof of Theorem 3 generalizes the 2-armed lower bound analysis in [38] to the multi-armed case with linear

reward structures. We first introduce some notations and definitions.

Consider a fully-collaborative instance I(X, \∗) of the CoPE-KB problem, where˜X = X = X𝑣, 𝑓 = 𝑓𝑣 for all 𝑣 ∈ [𝑉 ]

and 𝜙 (𝑥𝑖 )⊤\∗ is equal for all 𝑥𝑖 ≠ 𝑥∗. Let Δ = Δmin = (𝜙 (𝑥∗) − 𝜙 (𝑥𝑖 ))⊤\∗ and 𝑐 =𝜙 (𝑥𝑖 )⊤\ ∗

Δ for any 𝑥𝑖 ≠ 𝑥∗, where

𝑐 > 0 is a constant. Then, we have that𝜙 (𝑥∗)⊤\ ∗

Δ =𝜙 (𝑥𝑖 )⊤\ ∗+Δ

Δ = 1 + 𝑐 .For any integer 𝛼 ≥ 0, let E(𝛼,𝑇 ) be the event that A uses at least 𝛼 communication rounds and at most 𝑇

samples before the end of the 𝛼-th round, and let E+1 (𝛼,𝑇 ) be the event that A uses at least 𝛼 + 1 communication

rounds and at most 𝑇 samples before the end of the 𝛼-th round. Let 𝑇A and 𝑇A,𝑥𝑖 denote the expected number of

samples used by A, and the expected number of samples used on arm 𝑥𝑖 by A, respectively. Let _ be the sample

allocation of A, i.e., _𝑖 =𝑇A,𝑥𝑖𝑇A

. Let 𝜌 (X, \∗) = min_∈△X max𝑥 ∈X\{𝑥∗ }∥𝜙 (𝑥∗)−𝜙 (𝑥) ∥2(b∗𝐼+∑𝑥∈X _𝑥𝜙 (𝑥 )𝜙 (𝑥 )⊤)−1

(𝑓 (𝑥∗)−𝑓 (𝑥))2 and 𝑑 (X) =

min_∈△X max𝑥 ∈X\{𝑥∗ } ∥𝜙 (𝑥∗) − 𝜙 (𝑥)∥2(b∗𝐼+∑𝑥∈X _𝑥𝜙 (𝑥)𝜙 (𝑥)⊤)−1. Then, we have 𝜌 (X, \∗) = 𝑑 (X)

Δ2.

In order to prove Theorem 3, we first prove the following lemmas.

Lemma 3 (Linear Structured Progress Lemma). For any integer 𝛼 ≥ 0 and any 𝑞 ≥ 1, we have

Pr

I(X,\ ∗)

[E+1

(𝛼,𝜌 (X, \∗)𝑉𝑞

)]≥ Pr

I(X,\ ∗)

[E

(𝛼,𝜌 (X, \∗)𝑉𝑞

)]− 2𝛿 − 1

√𝑞.

Proof of Lemma 3. Let F be the event that A uses exactly 𝛼 communication rounds and at most𝜌 (X,\ ∗)𝑉𝑞

samples

before the end of the 𝛼-th round. Then, we have

Pr

I(X,\ ∗)

[E+1

(𝛼,𝜌 (X, \∗)𝑉𝑞

)]≥ Pr

I(X,\ ∗)

[E

(𝛼,𝜌 (X, \∗)𝑉𝑞

)]− Pr

I(X,\ ∗)[F ] .

Thus, to prove Lemma 3, it suffices to prove

Pr

I(X,\ ∗)[F ] ≤ 2𝛿 + 1

√𝑞. (19)

We can decompose F as

Pr

I(X,\ ∗)[F ] = Pr

I(X,\ ∗)[F ,A returns 𝑥∗] + Pr

I(X,\ ∗)[F ,A does not return 𝑥∗]

29

Woodstock ’21, June 03–05, 2021, Woodstock, NY

= Pr

I(X,\ ∗)[F ,A returns 𝑥∗] + 𝛿 (20)

Let𝑦𝑖 = 𝜙 (𝑥∗)−𝜙 (𝑥𝑖 ) for any 𝑖 ∈ [𝑛]. Let\ (b∗, _) = \∗−2(𝑦⊤𝑗 \ ∗)𝐴(b∗,_)−1𝑦 𝑗

𝑦⊤𝑗𝐴(b∗,_)−1𝑦 𝑗

, where𝐴(b∗, _) = b∗𝐼+∑𝑛𝑖=1

_𝑖𝜙 (𝑥𝑖 )𝜙 (𝑥𝑖 )⊤

and 𝑗 = argmax𝑖∈[𝑛]𝑦⊤𝑖 𝐴(b∗,_)−1𝑦𝑖

(𝑦⊤𝑖\ ∗)2 . Let I(X, \ (_)) denote the instance where the underlying parameter is \ (_). Under

I(X, \ (_)), it holds that 𝑦⊤𝑗\ (_) = −𝑦⊤

𝑗\∗ < 0, and thus 𝑥∗ is sub-optimal. Let DI denote the product distribution of

instance I with at most𝜌 (X,\ ∗)

𝑞 samples over all agents.

Using the Pinsker’s inequality (Lemma 11) and Gaussian KL-divergence computation, we have

∥DI(X,\ ∗) − DI(X,\ (_)) ∥TV

≤√︂

1

2

KL

(DI(X,\ ∗) ∥DI(X,\ (_))

)≤√√

1

4

∑︁𝑖∈[𝑛]

(𝜙 (𝑥𝑖 )⊤ (\∗ − \ (_)))2 · _𝑖𝑇A

=

√√√√1

4

∑︁𝑖∈[𝑛]

4(𝑦⊤𝑗\∗)2 · 𝑦⊤

𝑗𝐴(b∗, _)−1𝜙 (𝑥𝑖 )𝜙 (𝑥𝑖 )⊤𝐴(b∗, _)−1𝑦 𝑗

(𝑦⊤𝑗𝐴(b∗, _)−1𝑦 𝑗 )2

· _𝑖𝑇A

=

√√√𝑇A(𝑦⊤𝑗\∗)2 · 𝑦⊤

𝑗𝐴(b∗, _)−1 (∑𝑖∈[𝑛] _𝑖𝜙 (𝑥𝑖 )𝜙 (𝑥𝑖 )⊤)𝐴(b∗, _)−1𝑦 𝑗

(𝑦⊤𝑗𝐴(b∗, _)−1𝑦 𝑗 )2

√√√𝑇A(𝑦⊤𝑗\∗)2 · 𝑦⊤

𝑗𝐴(b∗, _)−1 (b∗𝐼 +

∑𝑖∈[𝑛] _𝑖𝜙 (𝑥𝑖 )𝜙 (𝑥𝑖 )⊤)𝐴(b∗, _)−1𝑦 𝑗

(𝑦⊤𝑗𝐴(b∗, _)−1𝑦 𝑗 )2

=

√√√𝑇A ·

(𝑦⊤𝑗\∗)2

𝑦⊤𝑗𝐴(b∗, _)−1𝑦 𝑗

≤√√√√𝜌 (X, \∗)

𝑞· 1

𝑦⊤𝑗𝐴(b∗,_)−1𝑦 𝑗

(𝑦⊤𝑗\ ∗)2

≤√√√√𝜌 (X, \∗)

𝑞· 1

min_∈△X𝑦⊤𝑗𝐴(b∗,_)−1𝑦 𝑗

(𝑦⊤𝑗\ ∗)2

=1

√𝑞

(21)

Since 𝑥∗ is sub-optimal under I(X, \ (_)), using the measure change technique, we have

Pr

I(X,\ ∗)[F ,A returns 𝑥∗] ≤ Pr

I(X,\ (_))[F ,A returns 𝑥∗]

+ ∥DI(X,\ ∗) − DI(X,\ (_)) ∥TV

≤𝛿 + 1

√𝑞

30

Collaborative Pure Exploration in Kernel Bandit Woodstock ’21, June 03–05, 2021, Woodstock, NY

Plugging the above equality into Eq. (20), we have

Pr

I(X,\ ∗)[F ] ≤ 2𝛿 + 1

√𝑞,

which completes the proof of Lemma 3. □

Let I(X, \/^) denote the instance where the underlying parameter is \/^. Under I(X, \/^), the reward gap is

(𝜙 (𝑥∗) −𝜙 (𝑥𝑖 ))⊤\/^ = 1

^ Δ and the sample complexity is 𝜌 (X, \/^) = ^2𝑑 (X)Δ2

. LetDI(X,\ ∗) andDI(X,\/^) denote theproduct distributions of instances I(X, \∗) and I(X, \/^) with 𝑇A samples, respectively.

Lemma 4 (Multi-armed Measure Transformation Lemma). Suppose that algorithm A uses 𝑇A =𝜌 (X,\ ∗)

Zsamples

over all agents on instance I(X, \∗), where Z ≥ 100. Then, for any event 𝐸 on I(X, \∗) and any 𝑄 ≥ Z , we have

Pr

DI(X,\/^ )[𝐸] ≤ Pr

DI(X,\∗ )[𝐸] · exp

((1 + 𝑐)2𝑑 (X)𝑛

√︄log(𝑛𝑄)

Z

)+ 1

𝑛𝑄2.

Proof. For any 𝑖 ∈ [𝑛], let 𝑍𝑖,1, . . . , 𝑍𝑖,𝑇A,𝑥𝑖 denote the observed 𝑇A,𝑥𝑖 samples on arm 𝑥𝑖 , and define event

𝐿𝑖 =

𝑇A,𝑥𝑖∑︁𝑡=1

𝑍𝑖,𝑡 ≥ 𝑇A,𝑥𝑖 · 𝜙 (𝑥𝑖 )⊤\/^ + 𝑧

𝜙 (𝑥𝑖 )⊤\/^ − 𝜙 (𝑥𝑖 )⊤\∗

,where 𝑧 ≥ 0 is a parameter specified later. We also define event 𝐿 = ∩𝑖∈[𝑛]𝐿𝑖 . Then,

Pr

DI(X,\/^ )[𝐸] ≤ Pr

DI(X,\/^ )[𝐸, 𝐿] + Pr

DI(X,\/^ )[¬𝐿]

Using the measure change technique, we bound the term PrDI(X,\/^ ) [𝐸, 𝐿] as

Pr

DI(X,\/^ )[𝐸, 𝐿]

= Pr

DI(X,\∗ )[𝐸, 𝐿] · exp

©­«−1

2

∑︁𝑖∈[𝑛]

𝑇A,𝑥𝑖∑︁𝑡=1

((𝑍𝑖,𝑡 − 𝜙 (𝑥𝑖 )⊤\/^)2 − (𝑍𝑖,𝑡 − 𝜙 (𝑥𝑖 )⊤\∗)2

)ª®¬= Pr

DI(X,\∗ )[𝐸, 𝐿] · exp

©­«−1

2

∑︁𝑖∈[𝑛]

𝑇A,𝑥𝑖∑︁𝑡=1

((𝜙 (𝑥𝑖 )⊤\/^)2 − (𝜙 (𝑥𝑖 )⊤\∗)2 − 2𝑍𝑖,𝑡 (𝜙 (𝑥𝑖 )⊤\/^ − 𝜙 (𝑥𝑖 )⊤\∗)

)ª®¬= Pr

DI(X,\∗ )[𝐸, 𝐿] · exp

(− 1

2

∑︁𝑖∈[𝑛]

( ((𝜙 (𝑥𝑖 )⊤\/^)2 − (𝜙 (𝑥𝑖 )⊤\∗)2

)·𝑇A,𝑥𝑖

− 2

(𝜙 (𝑥𝑖 )⊤\/^ − 𝜙 (𝑥𝑖 )⊤\∗

)·𝑇A,𝑥𝑖∑︁𝑡=1

𝑍𝑖,𝑡

))≤ Pr

DI(X,\∗ )[𝐸, 𝐿] · exp

(− 1

2

∑︁𝑖∈[𝑛]

( ((𝜙 (𝑥𝑖 )⊤\/^)2 − (𝜙 (𝑥𝑖 )⊤\∗)2

)·𝑇A,𝑥𝑖

− 2

(𝜙 (𝑥𝑖 )⊤\/^ − 𝜙 (𝑥𝑖 )⊤\∗

)·(𝑇A,𝑥𝑖 · 𝜙 (𝑥𝑖 )

⊤\/^ + 𝑧

𝜙 (𝑥𝑖 )⊤\/^ − 𝜙 (𝑥𝑖 )⊤\∗

) ))= Pr

DI(X,\∗ )[𝐸, 𝐿] · exp

(− 1

2

∑︁𝑖∈[𝑛]

( ((𝜙 (𝑥𝑖 )⊤\/^)2 − (𝜙 (𝑥𝑖 )⊤\∗)2

)·𝑇A,𝑥𝑖

31

Woodstock ’21, June 03–05, 2021, Woodstock, NY

− 2

(𝜙 (𝑥𝑖 )⊤\/^ − 𝜙 (𝑥𝑖 )⊤\∗

)·𝑇A,𝑥𝑖 · 𝜙 (𝑥𝑖 )

⊤\/^ − 2𝑧

))= Pr

DI(X,\∗ )[𝐸, 𝐿] · exp

(− 1

2

∑︁𝑖∈[𝑛]

( (𝜙 (𝑥𝑖 )⊤\/^ − 𝜙 (𝑥𝑖 )⊤\∗

) (𝜙 (𝑥𝑖 )⊤\/^ ·𝑇A,𝑥𝑖 + 𝜙 (𝑥𝑖 )

⊤\∗ ·𝑇A,𝑥𝑖

− 2 · 𝜙 (𝑥𝑖 )⊤\/^ ·𝑇A,𝑥𝑖)− 2𝑧

))= Pr

DI(X,\∗ )[𝐸, 𝐿] · exp

(− 1

2

∑︁𝑖∈[𝑛]

(−

(𝜙 (𝑥𝑖 )⊤\/^ − 𝜙 (𝑥𝑖 )⊤\∗

)2 ·𝑇A,𝑥𝑖 − 2𝑧

))

= Pr

DI(X,\∗ )[𝐸, 𝐿] · exp

©­«∑︁𝑖∈[𝑛]

(1

2

(𝜙 (𝑥𝑖 )⊤\/^ − 𝜙 (𝑥𝑖 )⊤\∗

)2 ·𝑇A,𝑥𝑖 + 𝑧

)ª®¬= Pr

DI(X,\∗ )[𝐸, 𝐿] · exp

©­«∑︁𝑖∈[𝑛]

(1

2

(1 − 1

^)2 (𝜙 (𝑥𝑖 )⊤\∗)2 ·

𝑑 (X)Δ2Z

· _𝑖 + 𝑧)ª®¬

≤ Pr

DI(X,\∗ )[𝐸, 𝐿] · exp

((1 + 𝑐)2

2

· 𝑑 (X)Z+ 𝑛𝑧

)Next, using the Chernoff-Hoeffding inequality, we bound the second term as

Pr

DI(X,\/^ )[¬𝐿] ≤

∑︁𝑖∈[𝑛]

Pr

DI(X,\/^ )[¬𝐿𝑖 ]

≤∑︁𝑖∈[𝑛]

exp

(−2

𝑧2

(𝜙 (𝑥𝑖 )⊤\/^ − 𝜙 (𝑥𝑖 )⊤\∗)2· Z

_𝑖𝜌 (X, \∗)

)=

∑︁𝑖∈[𝑛]

exp

(−2

𝑧2

(1 − 1

^ )2 (𝜙 (𝑥𝑖 )⊤\∗)2· Δ2Z

_𝑖𝑑 (X)

)≤

∑︁𝑖∈[𝑛]

exp

(− 2

(1 + 𝑐)2· 𝑧

2Z

𝑑 (X)

)=𝑛 · exp

(− 2

(1 + 𝑐)2· 𝑧

2Z

𝑑 (X)

)Thus,

Pr

DI(X,\/^ )[𝐸] ≤ Pr

DI(X,\/^ )[𝐸, 𝐿] + Pr

DI(X,\/^ )[¬𝐿]

≤ Pr

DI(X,\∗ )[𝐸, 𝐿] · exp

((1 + 𝑐)2

2

· 𝑑 (X)Z+ 𝑛𝑧

)+ 𝑛 · exp

(− 2

(1 + 𝑐)2· 𝑧

2Z

𝑑 (X)

)Let 𝑧 =

√︂(1+𝑐)2𝑑 (X) log(𝑛𝑄)

Z. Then, we have

Pr

DI(X,\/^ )[𝐸] ≤ Pr

DI(X,\∗ )[𝐸, 𝐿] · exp

©­« (1 + 𝑐)2

2

· 𝑑 (X)Z+ 𝑛

√︄(1 + 𝑐)2𝑑 (X) log(𝑛𝑄)

Z

ª®¬+ 𝑛 · exp (−2 log(𝑛𝑄))

32

Collaborative Pure Exploration in Kernel Bandit Woodstock ’21, June 03–05, 2021, Woodstock, NY

≤ Pr

DI(X,\∗ )[𝐸] · exp

((1 + 𝑐)2𝑑 (X)𝑛

√︄log(𝑛𝑄)

Z

)+ 1

𝑛𝑄2

Lemma 5 (Linear Structured Instance Transformation Lemma). For any integer 𝛼 ≥ 0, 𝑞 ≥ 100 and ^ ≥ 1, we

have

Pr

I(X,\/^)

[E

(𝛼 + 1,

𝜌 (X, \∗)𝑉𝑞

+ 𝜌 (X, \∗)

𝛽

)]≥ Pr

I(X,\ ∗)

[E+1

(𝛼,𝜌 (X, \∗)𝑉𝑞

)]− 𝛿 −

√︄(1 + 𝑐)2

4

· 𝑑 (X)𝑞

−(exp

((1 + 𝑐)2𝑑 (X)𝑛

√︄log(𝑛𝑉 )

𝛽

)− 1

)− 1

𝑛𝑉

Proof. Let ℓ = {𝑍𝑖,1, . . . , 𝑍𝑖,𝑇A,𝑥𝑖 }𝑖∈[𝑛] denote the 𝑇A samples of algorithm A on instance I(X, \∗). Let S denote

the set of all possible ℓ , conditioned on which E+1 (𝛼, 𝜌 (X,\∗)

𝑉𝑞) holds. Then, we have∑︁

𝑠∈SPr

I(X,\ ∗)[ℓ = 𝑠] =

∑︁𝑠∈S

Pr

I(X,\ ∗)

[E+1

(𝛼,𝜌 (X, \∗)𝑉𝑞

)]For any agent 𝑣 ∈ [𝑉 ], let K𝑣 be the event that agent 𝑣 uses more than

𝜌 (X,\ ∗)𝛽

samples during the (𝛼 + 1)-st round.Conditioned on 𝑠 ∈ S, K𝑣 only depends on the samples of agent 𝑣 during the (𝛼 + 1)-st round, and is independent of

other agents.

Using the facts that A is 𝛿-correct and 𝛽-speedup and all agents have the same performance on fully-collaborative

instances, we have

𝛿 ≥ Pr

I(X,\ ∗)

[A uses more than

𝜌 (X, \∗)𝛽

]≥

∑︁𝑠∈S

Pr

I(X,\ ∗)[ℓ = 𝑠] · Pr

I(X,\ ∗)[K1 ∨ · · · ∨ K𝑉 |ℓ = 𝑠]

=∑︁𝑠∈S

Pr

I(X,\ ∗)[ℓ = 𝑠] · ©­«1 −

∏𝑣∈[𝑉 ]

(1 − Pr

I(X,\ ∗)[K𝑖 |ℓ = 𝑠]

)ª®¬=E+1

(𝛼,𝜌 (X, \∗)𝑉𝑞

)−

∑︁𝑠∈S

Pr

I(X,\ ∗)[ℓ = 𝑠] ·

∏𝑣∈[𝑉 ]

(1 − Pr

I(X,\ ∗)[K𝑖 |ℓ = 𝑠]

)Rearranging the above inequality, we have∑︁

𝑠∈SPr

I(X,\ ∗)[ℓ = 𝑠] ·

∏𝑣∈[𝑉 ]

(1 − Pr

I(X,\ ∗)[K𝑖 |ℓ = 𝑠]

)≥E+1

(𝛼,𝜌 (X, \∗)𝑉𝑞

)− 𝛿 (22)

Using Lemma 4 with Z = 𝛽 and 𝑄 = 𝑉 , we have

Pr

I(X,\/^)

[E

(𝛼 + 1,

𝜌 (X, \∗)𝑉𝑞

+ 𝜌 (X, \∗)

𝛽

)]≥

∑︁𝑠∈S

Pr

I(X,\/^)[ℓ = 𝑠] · Pr

I(X,\/^)[¬K1 ∧ · · · ∧ ¬K𝑉 |ℓ = 𝑠]

=∑︁𝑠∈S

Pr

I(X,\/^)[ℓ = 𝑠] ·

∏𝑣∈[𝑉 ]

(1 − Pr

I(X,\/^)[K𝑖 |ℓ = 𝑠]

)33

Woodstock ’21, June 03–05, 2021, Woodstock, NY

≥∑︁𝑠∈S

Pr

I(X,\/^)[ℓ = 𝑠] ·

∏𝑣∈[𝑉 ]

max

{1 − Pr

I(X,\ ∗)[K𝑖 |ℓ = 𝑠] · exp

((1 + 𝑐)2𝑑 (X)𝑛

√︄log(𝑛𝑉 )

𝛽

)− 1

𝑛𝑉 2, 0

}≥

∑︁𝑠∈S

Pr

I(X,\/^)[ℓ = 𝑠] ·

©­«∏𝑣∈[𝑉 ]

max

{1 − Pr

I(X,\ ∗)[K𝑖 |ℓ = 𝑠] · exp

((1 + 𝑐)2𝑑 (X)𝑛

√︄log(𝑛𝑉 )

𝛽

), 0

}− 1

𝑛𝑉

ª®¬=

∑︁𝑠∈S

Pr

I(X,\/^)[ℓ = 𝑠] ·

( ∏𝑣∈[𝑉 ]

max

(1 − Pr

I(X,\ ∗)[K𝑖 |ℓ = 𝑠]

− Pr

I(X,\ ∗)[K𝑖 |ℓ = 𝑠] ·

(exp

((1 + 𝑐)2𝑑 (X)𝑛

√︄log(𝑛𝑉 )

𝛽

)− 1

), 0

)− 1

𝑛𝑉

)(e)

≥∑︁𝑠∈S

Pr

I(X,\/^)[ℓ = 𝑠] ·

©­«∏𝑣∈[𝑉 ]

(1 − Pr

I(X,\ ∗)[K𝑖 |ℓ = 𝑠]

)−

(exp

((1 + 𝑐)2𝑑 (X)𝑛

√︄log(𝑛𝑉 )

𝛽

)− 1

)− 1

𝑛𝑉

ª®¬≥

∑︁𝑠∈S

Pr

I(X,\/^)[ℓ = 𝑠] ·

∏𝑣∈[𝑉 ]

(1 − Pr

I(X,\ ∗)[K𝑖 |ℓ = 𝑠]

)−

(exp

((1 + 𝑐)2𝑑 (X)𝑛

√︄log(𝑛𝑉 )

𝛽

)− 1

)− 1

𝑛𝑉

≥∑︁𝑠∈S

Pr

I(X,\ ∗)[ℓ = 𝑠] ·

∏𝑣∈[𝑉 ]

(1 − Pr

I(X,\ ∗)[K𝑖 |ℓ = 𝑠]

)−

∑︁𝑠∈S

���� Pr

I(X,\/^)[ℓ = 𝑠] − Pr

I(X,\ ∗)[ℓ = 𝑠]

���� − (exp

((1 + 𝑐)2𝑑 (X)𝑛

√︄log(𝑛𝑉 )

𝛽

)− 1

)− 1

𝑛𝑉

where (e) comes from Lemma 12.

Using the Pinsker’s inequality (Lemma 11), we have∑︁𝑠∈S

���� Pr

I(X,\/^)[ℓ = 𝑠] − Pr

I(X,\ ∗)[ℓ = 𝑠]

���� (23)

≤√︂

1

2

KL(DI(X,\/^) ,DI(X,\ ∗) )

≤√√

1

4

∑︁𝑖∈[𝑛]

(𝜙 (𝑥𝑖 )⊤\/^ − 𝜙 (𝑥𝑖 )⊤\∗)2 · _𝑖𝑇A

=

√√1

4

∑︁𝑖∈[𝑛]

((1 − 1

^)2 (𝜙 (𝑥𝑖 )⊤\∗)2 ·

𝑑 (X)Δ2𝑞

_𝑖

)

34

Collaborative Pure Exploration in Kernel Bandit Woodstock ’21, June 03–05, 2021, Woodstock, NY

√︄(1 + 𝑐)2

4

· 𝑑 (X)𝑞

(24)

Thus, using Eqs. (22),(24), we have

Pr

I(X,\/^)

[E

(𝛼 + 1,

𝜌 (X, \∗)𝑉𝑞

+ 𝜌 (X, \∗)

𝛽

)]≥

∑︁𝑠∈S

Pr

I(X,\ ∗)[ℓ = 𝑠] ·

∏𝑣∈[𝑉 ]

(1 − Pr

I(X,\ ∗)[K𝑖 |ℓ = 𝑠]

)−

√︄(1 + 𝑐)2

4

· 𝑑 (X)𝑞

−(exp

((1 + 𝑐)2𝑑 (X)𝑛

√︄log(𝑛𝑉 )

𝛽

)− 1

)− 1

𝑛𝑉

≥ Pr

I(X,\ ∗)

[E+1

(𝛼,𝜌 (X, \∗)𝑉𝑞

)]− 𝛿 −

√︄(1 + 𝑐)2

4

· 𝑑 (X)𝑞

−(exp

((1 + 𝑐)2𝑑 (X)𝑛

√︄log(𝑛𝑉 )

𝛽

)− 1

)− 1

𝑛𝑉

Now we prove Theorem 3.

Proof of Theorem 3. Combining Lemmas 3,5, we have

Pr

I(X,\/^)

[E

(𝛼 + 1,

𝜌 (X, \∗)𝑉𝑞

+ 𝜌 (X, \∗)

𝛽

)]≥ Pr

I(X,\ ∗)

[E+1

(𝛼,𝜌 (X, \∗)𝑉𝑞

)]− 𝛿 −

√︄(1 + 𝑐)2

4

· 𝑑 (X)𝑞

−(exp

((1 + 𝑐)2𝑑 (X)𝑛

√︄log(𝑛𝑉 )

𝛽

)− 1

)− 1

𝑛𝑉

≥ Pr

I(X,\ ∗)

[E

(𝛼,𝜌 (X, \∗)𝑉𝑞

)]− 3𝛿 −

√︄(1 + 𝑐)2𝑑 (X)

𝑞

−(exp

((1 + 𝑐)2𝑑 (X)𝑛

√︄log(𝑛𝑉 )

𝛽

)− 1

)− 1

𝑛𝑉

Let ^ =

√︃1 + 𝑉𝑞

𝛽. Then, we have

Pr

I(X,\/^)

[E

(𝛼 + 1,

𝜌 (X, \/^)𝑉𝑞

)]= Pr

I(X,\/^)

[E

(𝛼 + 1,

^2 · 𝜌 (X, \∗)𝑉𝑞

)]= Pr

I(X,\/^)

[E

(𝛼 + 1,

𝜌 (X, \∗)𝑉𝑞

+ 𝜌 (X, \∗)

𝛽

)]≥ Pr

I(X,\ ∗)

[E

(𝛼,𝜌 (X, \∗)𝑉𝑞

)]− 3𝛿 −

√︄(1 + 𝑐)2𝑑 (X)

𝑞

35

Woodstock ’21, June 03–05, 2021, Woodstock, NY

−(exp

((1 + 𝑐)2𝑑 (X)𝑛

√︄log(𝑛𝑉 )

𝛽

)− 1

)− 1

𝑛𝑉(25)

Let I(X, \0) be the basic instance of induction, where the reward gap is Δ0 = (𝜙 (𝑥∗) −𝜙 (𝑥𝑖 ))⊤\0 = 1 for any 𝑖 ∈ [𝑛].Let 𝑡0 be the largest integer such that

Δ ·(1 + 𝑉𝑞

𝛽

) 𝑡02

≤ 1,

where 𝑞 = 1000𝑡20. Then, we have

𝑡0 = Ω©­«

log( 1

Δ )log(1 + 𝑉

𝛽) + log log( 1

Δ )ª®¬

Starting from I(X, \0), we repeatedly apply Eq. (25) for 𝑡0 times to switch to I(X, \∗) where the reward gap is Δ.

Since PrI(X,\0)[E

(0,𝜌 (X,\0)𝑉𝑞

)]= 1, by induction, we have

Pr

I(X,\ ∗)

[E

(𝑡0,

𝜌 (X, \∗)𝑉𝑞

)]≥1 − ©­«3𝛿 +

√︄(1 + 𝑐)2𝑑 (X)

𝑞+

(exp

((1 + 𝑐)2𝑑 (X)𝑛

√︄log(𝑛𝑉 )

𝛽

)− 1

)+ 1

𝑛𝑉

ª®¬ · 𝑡0When 𝛿 = 𝑂 ( 1

𝑛𝑉) and 𝑞,𝑉 , 𝛽 are large enough, we can have

©­«3𝛿 +

√︄(1 + 𝑐)2𝑑 (X)

𝑞+

(exp

((1 + 𝑐)2𝑑 (X)𝑛

√︄log(𝑛𝑉 )

𝛽

)− 1

)+ 1

𝑛𝑉

ª®¬ · 𝑡0 ≤ 1

2

and then

Pr

I(X,\ ∗)

[E

(𝑡0,

𝜌 (X, \∗)𝑉𝑞

)]≥ 1

2

.

Thus,

Pr

I(X,\ ∗)

A uses Ω©­«

log( 1

Δ )log(1 + 𝑉

𝛽) + log log( 1

Δ )ª®¬ communication rounds

≥ Pr

I(X,\ ∗)

[E

(𝑡0,

𝜌 (X, \∗)𝑉𝑞

)]≥ 1

2

,

which completes the proof of Theorem 3. □

B PROOFS FOR THE FIXED-BUDGET SETTING

B.1 Proof of Theorem 4

Proof of Theorem 4. Our proof of Theorem 4 adapts the error probability analysis in [23] to the multi-agent setting.

Since the number of samples used over all agents in each phase is 𝑁 = ⌊𝑇𝑉 /𝑅⌋, the total number of samples used by

algorithm CoopKernelFB is at most 𝑇𝑉 and the total number of samples used per agent is at most 𝑇 .

Now we prove the error probability upper bound.

Recall that for any _ ∈ △˜X , Φ_ = [

√_1𝜙 (𝑥∗𝑣 )⊤; . . . ;

√︁_𝑛𝑉𝜙 (𝑥𝑛𝑉 )⊤] and Φ⊤

_Φ_ =

∑�̃� ∈ ˜X _�̃�𝜙 (𝑥)𝜙 (𝑥)

⊤. Let 𝛾∗ = b∗𝑁 .

36

Collaborative Pure Exploration in Kernel Bandit Woodstock ’21, June 03–05, 2021, Woodstock, NY

For any 𝑥𝑖 , 𝑥 𝑗 ∈ B (𝑡 )𝑣 , 𝑣 ∈ [𝑉 ], 𝑡 ∈ [𝑅], define

Δ𝑡,�̃�𝑖 ,�̃� 𝑗 = inf

Δ>0

∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥2(b∗𝐼+Φ⊤

_∗𝑡Φ_∗𝑡)−1

Δ2≤ 2𝜌∗

and event

J𝑡,�̃�𝑖 ,�̃� 𝑗 ={���( ˆ𝑓𝑡 (𝑥𝑖 ) − ˆ𝑓𝑡 (𝑥 𝑗 )

)−

(𝑓 (𝑥𝑖 ) − 𝑓 (𝑥 𝑗 )

) ��� < Δ𝑡,�̃�𝑖 ,�̃� 𝑗

}.

In the following, we prove Pr

[¬J𝑡,�̃�𝑖 ,�̃� 𝑗

]≤ 2 exp

(− 𝑁

2(1+Y)𝜌∗).

Similar to the analysis procedure of Lemma 1, we have(ˆ𝑓𝑡 (𝑥𝑖 ) − ˆ𝑓𝑡 (𝑥 𝑗 )

)−

(𝑓 (𝑥𝑖 ) − 𝑓 (𝑥 𝑗 )

)=

(𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )

)⊤ (𝛾∗𝐼 + Φ⊤𝑡 Φ𝑡

)−1

Φ⊤𝑡 [̄(𝑡 ) − 𝛾∗

(𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )

)⊤ (𝛾∗𝐼 + Φ⊤𝑡 Φ𝑡

)−1

\∗,

where the mean of the first term is zero and its variance is bounded by

∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥2(𝛾∗𝐼+Φ⊤𝑡 Φ𝑡 )−1≤(1 + Y) · ∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥ (b∗𝐼+Φ⊤

_∗𝑡Φ_∗𝑡)−1

𝑁.

Using the Hoeffding inequality, we have

Pr

[���(𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 ))⊤ (𝛾∗𝐼 + Φ⊤𝑡 Φ𝑡

)−1

Φ⊤𝑡 [(𝑡 )𝑣

��� ≥ 1

2

Δ𝑡,�̃�𝑖 ,�̃� 𝑗

]≤2 exp

©­­­«−2

1

4Δ2

𝑡,�̃�𝑖 ,�̃� 𝑗

(1+Y) · ∥𝜙 (�̃�𝑖 )−𝜙 (�̃� 𝑗 ) ∥ (b∗𝐼+Φ⊤_∗𝑡

Φ_∗𝑡)−1

𝑁

ª®®®¬≤2 exp

©­­­­­«−1

2

𝑁

(1+Y) · ∥𝜙 (�̃�𝑖 )−𝜙 (�̃� 𝑗 ) ∥ (b∗𝐼+Φ⊤_∗𝑡

Φ_∗𝑡)−1

Δ2

𝑡,�̃�𝑖 ,�̃� 𝑗

ª®®®®®¬≤2 exp

(− 𝑁

2(1 + Y)𝜌∗

)Thus, with probability at least 1 − 2 exp

(− 𝑁

2(1+Y)𝜌∗), we have��� (𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 ))⊤ (

𝛾∗𝐼 + Φ⊤𝑡 Φ𝑡)−1

Φ⊤𝑡 [(𝑡 )𝑣

��� < 1

2

Δ𝑡,�̃�𝑖 ,�̃� 𝑗 .

Recall that b∗ satisfies (1 + Y)√︁b∗max

�̃�𝑖 ,�̃� 𝑗 ∈ ˜X𝑣 ,𝑣∈[𝑉 ] ∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥ (b∗𝐼+Φ⊤_𝑢Φ_𝑢 )−1𝐵 ≤ 1

2Δ𝑡,�̃�𝑖 ,�̃� 𝑗 . Then, we bound the

regularization (bias) term. ���𝛾∗ (𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 ))⊤ (𝛾∗𝐼 + Φ⊤𝑡 Φ𝑡

)−1

\∗���

≤𝛾∗∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥(𝛾∗𝐼+Φ⊤𝑡 Φ𝑡 )−1 ∥\∗∥(𝛾∗𝐼+Φ⊤𝑡 Φ𝑡 )−1

≤√𝛾∗ · ∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥(𝛾∗𝐼+Φ⊤𝑡 Φ𝑡 )−1 ∥\∗∥2

37

Woodstock ’21, June 03–05, 2021, Woodstock, NY

≤√︁b∗𝑁 ·

(1 + Y) · ∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥ (b∗𝐼+Φ⊤_∗𝑡Φ_∗𝑡)−1

√𝑁

· ∥\∗∥2

≤(1 + Y)√︁b∗ max

�̃�𝑖 ,�̃� 𝑗 ∈ ˜B (𝑡 )𝑣 ,𝑣∈[𝑉 ]∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥ (

b∗𝐼+Φ⊤_∗𝑡Φ_∗𝑡

)−1 ∥\∗∥2

≤(1 + Y)√︁b∗ max

�̃�𝑖 ,�̃� 𝑗 ∈ ˜B (𝑡 )𝑣 ,𝑣∈[𝑉 ]∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥ (

b∗𝐼+Φ⊤_𝑢Φ_𝑢)−1 ∥\∗∥2

≤(1 + Y)√︁b∗ max

�̃�𝑖 ,�̃� 𝑗 ∈ ˜X𝑣 ,𝑣∈[𝑉 ]∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥ (

b∗𝐼+Φ⊤_𝑢Φ_𝑢)−1 · 𝐵

≤ 1

2

Δ𝑡,�̃�𝑖 ,�̃� 𝑗

Thus, with probability at least 1 − 2 exp

(− 𝑁

2(1+Y)𝜌∗), we have���( ˆ𝑓𝑡 (𝑥𝑖 ) − ˆ𝑓𝑡 (𝑥 𝑗 )

)−

(𝑓 (𝑥𝑖 ) − 𝑓 (𝑥 𝑗 )

) ���≤

��� (𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 ))⊤ (𝛾∗𝐼 + Φ⊤𝑡 Φ𝑡

)−1

Φ⊤𝑡 [(𝑡 )𝑣

��� + ���𝛾∗ (𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 ))⊤ (𝛾∗𝐼 + Φ⊤𝑡 Φ𝑡

)−1

\∗���

<Δ𝑡,�̃�𝑖 ,�̃� 𝑗 ,

which completes the proof of Pr

[¬J𝑡,�̃�𝑖 ,�̃� 𝑗

]≤ 2 exp

(− 𝑁

2(1+Y)𝜌∗).

Define event

J =

{���( ˆ𝑓𝑡 (𝑥𝑖 ) − ˆ𝑓𝑡 (𝑥 𝑗 ))−

(𝑓 (𝑥𝑖 ) − 𝑓 (𝑥 𝑗 )

) ��� < Δ𝑡,�̃�𝑖 ,�̃� 𝑗 , ∀𝑥𝑖 , 𝑥 𝑗 ∈ B(𝑡 )𝑣 ,∀𝑣 ∈ [𝑉 ],∀𝑡 ∈ [𝑅]

},

By a union bound over 𝑥𝑖 , 𝑥 𝑗 ∈ B (𝑡 )𝑣 , 𝑣 ∈ [𝑉 ] and 𝑡 ∈ [𝑅], we have

Pr [¬J] ≤2𝑛2𝑉 log(𝜔 ( ˜X)) · exp

(− 𝑁

2(1 + Y)𝜌∗

)=𝑂

(𝑛2𝑉 log(𝜔 ( ˜X)) · exp

(− 𝑇𝑉

𝜌∗ log(𝜔 ( ˜X))

))In order to prove Theorem 4, it suffices to prove that conditioning on J , algorithm CoopKernelFB returns the correct

answers 𝑥∗𝑣 for all 𝑣 ∈ [𝑉 ].Suppose that there exist 𝑣 ∈ [𝑉 ] and 𝑡 ∈ [𝑅] such that 𝑥∗𝑣 is eliminated in phase 𝑡 . Define

B′ (𝑡 )𝑣 = {𝑥 ∈ B (𝑡 )𝑣 :ˆ𝑓𝑡 (𝑥) > ˆ𝑓𝑡 (𝑥∗𝑣 )},

which denotes the subset of arms that are ranked before 𝑥∗𝑣 by the estimated rewards in B (𝑡 )𝑣 . According to the

elimination rule, we have

𝜔 (B′ (𝑡 )𝑣 ∪ {𝑥∗𝑣 }) >1

2

𝜔 (B (𝑡 )𝑣 ) =1

2

max

�̃�𝑖 ,�̃� 𝑗 ∈B (𝑡 )𝑣∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥ (b∗𝐼+Φ⊤

_∗𝑡Φ_∗𝑡)−1 (26)

Define 𝑥0 = argmax�̃� ∈B′ (𝑡 )𝑣

Δ�̃� . We have

1

2

∥𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥0)∥ (b∗𝐼+Φ⊤_∗𝑡Φ_∗𝑡)−1

Δ2

�̃�0

38

Collaborative Pure Exploration in Kernel Bandit Woodstock ’21, June 03–05, 2021, Woodstock, NY

≤ 1

2

max

�̃�𝑖 ,�̃� 𝑗 ∈B (𝑡 )𝑣

∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥ (b∗𝐼+Φ⊤_∗𝑡Φ_∗𝑡)−1

Δ2

�̃�0

(f)

<𝜔 (B′ (𝑡 )𝑣 ∪ {𝑥∗𝑣 })

Δ2

�̃�0

(g)

= min

_∈△Xmax

�̃�𝑖 ,�̃� 𝑗 ∈B′ (𝑡 )𝑣 ∪{�̃�∗𝑣 }

∥𝜙 (𝑥𝑖 ) − 𝜙 (𝑥 𝑗 )∥ (b∗𝐼+Φ⊤_Φ_

)−1

Δ2

�̃�0

(h)

≤ min

_∈△Xmax

�̃� ∈B′ (𝑡 )𝑣

∥𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥)∥ (b∗𝐼+Φ⊤_Φ_

)−1

Δ2

�̃�

≤ min

_∈△Xmax

�̃� ∈X𝑣\{�̃�∗𝑣 },𝑣∈[𝑉 ]

∥𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥)∥ (b∗𝐼+Φ⊤_Φ_

)−1

Δ2

�̃�

=𝜌∗,

where (f) can be obtained by dividing Eq. (26) by Δ2

�̃�0

, (g) comes from the definition of 𝜔 (·), and (h) is due to the

definition of 𝑥0.

According to the definition

Δ𝑡,�̃�∗𝑣 ,�̃�0= inf

Δ>0

∥𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥0)∥2(b∗𝐼+Φ⊤

_∗𝑡Φ_∗𝑡)−1

Δ2≤ 2𝜌∗

,we have Δ𝑡,�̃�∗𝑣 ,�̃�0

≤ Δ�̃�0.

Conditioning on J , we have

���( ˆ𝑓𝑡 (𝑥∗𝑣 ) − ˆ𝑓𝑡 (𝑥0))−

(𝑓 (𝑥∗𝑣 ) − 𝑓 (𝑥0)

) ��� < Δ𝑡,�̃�∗𝑣 ,�̃�0≤ Δ�̃�0

. Then, we have

ˆ𝑓𝑡 (𝑥∗𝑣 ) − ˆ𝑓𝑡 (𝑥0) >(𝑓 (𝑥∗𝑣 ) − 𝑓 (𝑥0)

)− Δ�̃�0

= 0,

which contradicts the definition of 𝑥0 that satisfiesˆ𝑓𝑡 (𝑥0) > ˆ𝑓𝑡 (𝑥∗𝑣 ). Thus, for any 𝑣 ∈ [𝑉 ], 𝑥∗𝑣 will never be eliminated.

Since 𝜔 ({𝑥∗𝑣 , 𝑥}) ≥ 1 for any 𝑥 ∈ X𝑣 \ {𝑥∗𝑣 }, 𝑣 ∈ [𝑉 ] and 𝑅 =⌈log

2(𝜔 ( ˜X))

⌉≥

⌈log

2(𝜔 (X𝑣))

⌉for any 𝑣 ∈ [𝑉 ], we

have that conditioning on J , algorithm CoopKernelFB returns the correct answers 𝑥∗𝑣 for all 𝑣 ∈ [𝑉 ].For communication rounds, since algorithm CoopKernelFB has at most 𝑅 =

⌈log

2(𝜔 ( ˜X))

⌉phases, the number of

communication rounds is bounded by 𝑂 (log(𝜔 ( ˜X))). □

B.2 Proof of Corollary 2

Proof of Corollary 2. Following the analysis procedure of Corollary 1, we have

𝜌∗ = min

_∈△Xmax

�̃� ∈ ˜X𝑣 ,𝑣∈[𝑉 ]

∥𝜙 (𝑥∗𝑣 ) − 𝜙 (𝑥)∥2(b∗𝐼+Φ⊤_Φ_

)−1

(𝑓 (𝑥∗𝑣 ) − 𝑓 (𝑥))2

≤ 8

Δ2

min

· log det

(𝐼 + b−1

∗ 𝐾_∗)

39

Woodstock ’21, June 03–05, 2021, Woodstock, NY

Maximum Information Gain. Recall that the maximum information gain over all sample allocation _ ∈ △˜X is defined

as

Υ = max

_∈△ ˜Xlog det

(𝐼 + b∗−1𝐾_

).

Then, the error probability is bounded by

𝑂

(𝑛2𝑉 log(𝜔 ( ˜X)) · exp

(− 𝑇𝑉

𝜌∗ log(𝜔 ( ˜X))

))=𝑂

(𝑛2𝑉 log(𝜔 ( ˜X)) · exp

(−

𝑇𝑉Δ2

min

Υ log(𝜔 ( ˜X))

))Effective Dimension. Recall that 𝛼1 ≥ · · · ≥ 𝛼𝑛𝑉 denote the eigenvalues of 𝐾_∗ in decreasing order. The effective

dimension of 𝐾_∗ is defined as

𝑑eff

= min

{𝑗 : 𝑗b∗ log(𝑛𝑉 ) ≥

𝑛𝑉∑︁𝑖=𝑗+1

𝛼𝑖

},

and we have

log det

(𝐼 + b−1

∗ 𝐾_∗)≤ 𝑑

efflog

(2𝑛𝑉 ·

(1 + Trace(𝐾_∗ )

b∗𝑑eff

)).

Then, the error probability is bounded by

𝑂

(𝑛2𝑉 log(𝜔 ( ˜X)) · exp

(− 𝑇𝑉

𝜌∗ log(𝜔 ( ˜X))

))

=𝑂©­­«𝑛2𝑉 log(𝜔 ( ˜X)) · exp

©­­«−𝑇𝑉Δ2

min

𝑑eff

log

(𝑛𝑉 ·

(1 + Trace(𝐾_∗ )

b∗𝑑eff

))log(𝜔 ( ˜X))

ª®®¬ª®®¬

Decomposition. Recall that 𝐾 = [𝐾 (𝑥𝑖 , 𝑥 𝑗 )]𝑖, 𝑗 ∈[𝑛𝑉 ] , 𝐾𝑧 = [𝐾𝑧 (𝑧𝑣, 𝑧𝑣′)]𝑣,𝑣′∈[𝑉 ] , 𝐾𝑥 = [𝐾𝑥 (𝑥𝑖 , 𝑥 𝑗 )]𝑖, 𝑗 ∈[𝑛𝑉 ] , andrank(𝐾_∗ ) = rank(𝐾) ≤ rank(𝐾𝑧) · rank(𝐾𝑥 ). We have

log det

(𝐼 + b−1

∗ 𝐾_∗)≤ rank(𝐾𝑧) · rank(𝐾𝑥 ) log

(Trace

(𝐼 + b−1

∗ 𝐾_∗)

rank(𝐾_∗ )

)Then, the error probability is bounded by

𝑂

(𝑛2𝑉 log(𝜔 ( ˜X)) · exp

(− 𝑇𝑉

𝜌∗ log(𝜔 ( ˜X))

))

=𝑂©­­«𝑛2𝑉 log(𝜔 ( ˜X)) · exp

©­­«−𝑇𝑉Δ2

min

rank(𝐾𝑧) · rank(𝐾𝑥 ) log

(Trace(𝐼+b−1

∗ 𝐾_∗ )rank(𝐾_∗ )

)log(𝜔 ( ˜X))

ª®®¬ª®®¬

Therefore, we complete the proof of Corollary 2. □

B.3 Proof of Theorem 5

Our proof of Theorem 5 follows the analysis procedure in [1].

40

Collaborative Pure Exploration in Kernel Bandit Woodstock ’21, June 03–05, 2021, Woodstock, NY

We first introduce some definitions in information theory. For any random variable 𝐴, let 𝐻 (𝐴) denote the Shannonentropy of 𝐴. If 𝐴 is uniformly distributed on its support, 𝐻 (𝐴) = log |𝐴|. For any 𝑝 ∈ (0, 1), let 𝐻2 (𝑝) = −𝑝 log

2𝑝 −

(1 − 𝑝) log2(1 − 𝑝) denote the binary entropy, and 𝐻2 (𝑝) = 𝐻 (𝐴) for 𝐴 ∼ Bernoulli(𝑝). For any random variables 𝐴, 𝐵,

let 𝐻 (𝐴;𝐵) = 𝐻 (𝐴) − 𝐻 (𝐴|𝐵) = 𝐻 (𝐵) − 𝐻 (𝐵 |𝐴) denote the mutual information of 𝐴 and 𝐵.

Consider the following fully-collaborative instance DΔ,𝑝𝑑

: for a uniformly drawn index 𝑖∗ ∈ [𝑑], \∗𝑖∗ = 𝑝 + Δ and

\∗𝑗= 𝑝 for all 𝑗 ≠ 𝑖∗. The arm set X = {𝑥 ∈ {0, 1}𝑑 :

∑𝑑𝑖=1

𝑥 (𝑖) = 1}, and the feature mapping 𝜙 (𝑥) = 𝐼𝑥 for all 𝑥 ∈ X.Under instance DΔ,𝑝

𝑑, we have 𝜔 ( ˜X) = 𝜔 (X) = 𝑑 .

Let 𝛿∗ > 0 be a small constant that we specify later. There exists a single-agent algorithm A𝑆 (e.g., [23]) that uses at

most 𝑇 samples and guarantees 𝑂 (log𝑑 · exp(− Δ2𝑇𝑑 log𝑑

)) error probability on instance DΔ,𝑝𝑑

for any 𝑑 > 1. Restricting

the error probability to the constant 𝛿∗, we have that for any 𝑑 > 1,A𝑆 uses at most𝑇 = 𝑂 ( 𝑑Δ2

log𝑑 · log log𝑑) samples

to guarantee 𝛿∗ error probability on instance DΔ,𝑝𝑑

.

Let 𝛼 = 𝑉𝛽and 1 ≤ 𝛼 ≤ log𝑑 . According to the definition of speedup, a

𝑉𝛼 -speedup distributed algorithm A must

satisfy that for any 𝑑 > 1, A uses at most 𝑇 = 𝑂 (𝑛

Δ2

𝑉𝛼

· 𝑉 log𝑑 · log log𝑑) = 𝑂 ( 𝛼𝑑Δ2

log𝑑 · log log𝑑) samples over all

agents and guarantees 𝛿∗ error probability on instance DΔ,𝑝𝑑

.

Main Proof. Now, in order to prove Theorem 5, we first prove the following Lemma 6, which relaxes the sample budget

within logarithmic factors.

Lemma 6. For any 𝑑 > 1, any distributed algorithm A that can use at most 𝑂 ( 𝛼𝑑 (log𝛼+log log𝑑)2Δ2 (log𝑑)2 ) samples over all

agents and guarantee 𝛿∗ error probability on instance DΔ,𝑝𝑑

needs Ω( log𝑑

log𝛼+log log𝑑) communication rounds.

Below we prove Lemma 6 by induction (Lemmas 7,8).

Lemma 7 (Basic Step). For any 𝑑 > 1 and 1 ≤ 𝛼 ≤ log𝑑 , there is no 1-round algorithmA1 that can use𝑂 ( 𝛼𝑑Δ2) samples

and guarantee 𝛿1 error probability for some constant 𝛿1 ∈ (0, 1) on instance DΔ,𝑝𝑑

.

Proof of Lemma 7. Let 𝐼 denote the random variable of index 𝑖∗. Since 𝐼 is uniformly distributed on [𝑑],𝐻 (𝐼 ) = log𝑑 .

Let 𝑆 denote the sample profile of A1 on instance DΔ,𝑝𝑑

. According to the definitions of X and \∗, for instance DΔ,𝑝𝑑

,

identifying the best arm is equivalent to identifying 𝐼 . Suppose that A1 returns the best arm with error probability 𝛿1.

Using the Fano’s inequality (Lemma 13), we have

𝐻 (𝐼 |𝑆) ≤ 𝐻2 (𝛿1) + 𝛿1 log𝑑 (27)

Using Lemma 14, we have

𝐻 (𝐼 |𝑆) =𝐻 (𝐼 ) − 𝐻 (𝐼 ; 𝑆)

= log𝑑 −𝑂 (𝛼𝑑Δ2· Δ

2

𝑑)

= log𝑑 −𝑂 (𝛼)

Then, for some small enough constant 𝛿1 ∈ (0, 1), Eq. (27) cannot hold. Thus, for any 𝑑 > 1 and 1 ≤ 𝛼 ≤ log𝑑 , there is

no 1-round algorithm A1 that can use 𝑂 ( 𝛼𝑑Δ2) samples and guarantee 𝛿1 error probability on instance DΔ,𝑝

𝑑. □

Lemma 8 (Induction Step). Suppose that 1 ≤ 𝛼 ≤ log𝑑 and 𝛿 ∈ (0, 1). If for any 𝑑 > 1, there is no (𝑟 − 1)-roundalgorithmA𝑟−1 that can use𝑂 ( 𝛼𝑑

Δ2 (𝑟−1)2 ) samples and guarantee 𝛿 error probability on instance DΔ,𝑝𝑑

, then for any 𝑑 > 1,41

Woodstock ’21, June 03–05, 2021, Woodstock, NY

there is no 𝑟 -round algorithm A𝑟 that can use 𝑂 ( 𝛼𝑑Δ2𝑟 2) samples and guarantee 𝛿 −𝑂 ( 1

𝑟 2) error probability on instance

DΔ,𝑝𝑑

.

Proof of Lemma 8. We prove this lemma by contradiction. Suppose that for some 𝑑 > 1, there exists an 𝑟 -round

algorithm A𝑟 that can use 𝑂 ( 𝛼𝑑Δ2𝑟 2) samples and guarantee 𝛿 error probability on instance DΔ,𝑝

𝑑. In the following, we

show how to construct an (𝑟 − 1)-round algorithm A𝑟−1 that can use 𝑂 ( 𝛼𝑑Δ2 (𝑟−1)2 ) samples and guarantee at most

𝛿 +𝑂 ( 1

𝑟 2) error probability on instance DΔ,𝑝

˜𝑑for some

˜𝑑 .

Let 𝑆1 denote the sample profile of A𝑟 in the first round on instance DΔ,𝑝𝑑

. Since the number of samples of A𝑟 isbounded by 𝑂 ( 𝛼𝑑

Δ2𝑟 2), using Lemma 14, we have

𝐻 (𝐼 |𝑆1) =𝐻 (𝐼 ) − 𝐻 (𝐼 ; 𝑆1)

= log𝑑 −𝑂 ( 𝛼𝑑Δ2𝑟2

· Δ2

𝑑)

= log𝑑 −𝑂 ( 𝛼𝑟2)

Let DΔ,𝑝𝑑 |𝑆1

denote the posterior of DΔ,𝑝𝑑

after observing the sample profile 𝑆1.

Using Lemma 15 on random variable 𝐼 |𝑆1 with parameters 𝛾 = 𝑂 ( 𝛼𝑟 2) and 𝜖 = 𝑜 ( 1

𝑟 2), we can write DΔ,𝑝

𝑑 |𝑆1

as a convex

combination of distributions Q0,Q1, . . . ,Qℓ , i.e., DΔ,𝑝𝑑 |𝑆1

=∑ℓ𝑗=0

𝑞 𝑗Q 𝑗 such that 𝑞0 = 𝑜 ( 1

𝑟 2), and for any 𝑗 ≥ 1,

|supp(Q 𝑗 ) | ≥ exp

(log𝑑 − 𝛾

𝜖

)= exp (log𝑑 − 𝑜 (𝛼))

=𝑑

𝑒𝑜 (𝛼),

and

∥Q 𝑗 −U𝑗 ∥TV = 𝑜 ( 1

𝑟2),

whereU𝑗 is the uniform distribution on supp(Q 𝑗 ).Since

Pr[A𝑟 has an error|DΔ,𝑝𝑑 |𝑆1

] =ℓ∑︁𝑗=0

𝑞 𝑗 Pr[A𝑟 has an error|Q 𝑗 ] ≤ 𝛿,

using an average argument, there exists a distribution Q 𝑗 for some 𝑗 ≥ 1 such that

Pr[A𝑟 has an error|Q 𝑗 ] ≤ 𝛿.

Let˜𝑑 = |supp(Q 𝑗 ) | ≥ 𝑑

𝑒𝑜 (𝛼 ). Since ∥Q 𝑗 −U𝑗 ∥TV = 𝑜 ( 1

𝑟 2) andU𝑗 is equivalent to DΔ,𝑝

˜𝑑, we have

Pr[A𝑟 has an error|DΔ,𝑝˜𝑑] ≤ 𝛿 + 𝑜 ( 1

𝑟2) .

Under instance DΔ,𝑝˜𝑑

, the sample budget for (𝑟 − 1)-round algorithms is 𝑂 ( 𝛼 ˜𝑑Δ2 (𝑟−1)2 ), and it holds that

𝑂 ( 𝛼 ˜𝑑

Δ2 (𝑟 − 1)2) ≥ 𝑂 (

𝛼 𝑑

𝑒𝑜 (𝛼 )

Δ2 (𝑟 − 1)2) ≥ 𝑂 ( 𝛼𝑑

1−𝑜 (1)

Δ2 (𝑟 − 1)2) ≥ 𝑂 ( 𝛼𝑑

Δ2𝑟2)

42

Collaborative Pure Exploration in Kernel Bandit Woodstock ’21, June 03–05, 2021, Woodstock, NY

Then, we can construct an (𝑟 −1)-round algorithmA𝑟−1 using algorithmA𝑟 from the second round. The constructed

A𝑟−1 uses at most 𝑂 ( 𝛼𝑑Δ2𝑟 2) ≤ 𝑂 ( 𝛼 ˜𝑑

Δ2 (𝑟−1)2 ) samples and guarantees 𝛿 + 𝑜 ( 1

𝑟 2) error probability.

The specific procedure ofA𝑟−1 is as follows: ifA𝑟 samples some dimension 𝑖 ∈ supp(Q 𝑗 ), thenA𝑟−1 also samples 𝑖;

otherwise, if A𝑟 samples some dimension 𝑖 ∈ [𝑑] \ supp(Q 𝑗 ), then A𝑟−1 samples Bernoulli(𝑝) and feeds the outcome

to A𝑟 . Finally, if A𝑟 returns some dimension 𝑖 ∈ supp(Q 𝑗 ), then A𝑟−1 also returns 𝑖; otherwise, if A𝑟 returns some

dimension 𝑖 ∈ [𝑑] \ supp(Q 𝑗 ), then A𝑟−1 returns an arbitrary dimension in [𝑑] \ supp(Q 𝑗 ) (the error case). □

Proof of Lemma 6. Let 𝑟∗ =log𝑑

log𝛼+log log𝑑. Combining Lemmas 7,8, we obtain that there is no 𝑟∗-round algorithm

A that can use 𝑂 ( 𝛼𝑑Δ2𝑟 2

∗) samples and guarantee 𝛿1 −

∑𝑟∗𝑟=2

𝑜 ( 1

𝑟 2) error probability on instance DΔ,𝑝

𝑑for any 𝑑 > 1.

Let 𝛿∗ < 𝛿1 −∑𝑟∗𝑟=2

𝑜 ( 1

𝑟 2). Thus, for any 𝑑 > 1 and 1 ≤ 𝛼 ≤ log𝑑 , any distributed algorithm A that can use 𝑂 ( 𝛼𝑑

Δ2𝑟 2

∗)

samples and guarantee 𝛿∗ error probability on instance DΔ,𝑝𝑑

must cost 𝑟∗ = Ω

(log𝑑

log( 𝑉𝛽)+log log𝑑

)communication

rounds. □

Therefore, for any𝑑 > 1 and𝑉

log𝑑≤ 𝛽 ≤ 𝑉 , a 𝛽-speedup distributed algorithmA needs at leastΩ

(log𝜔 ( ˜X)

log( 𝑉𝛽)+log log𝜔 ( ˜X)

)communication rounds under instance DΔ,𝑝

𝑑, which completes the proof of Theorem 5.

C TECHNICAL TOOLS

Lemma 9 (Lemma 15 in [10]). For _∗ = argmax_∈△ ˜Xlog det

(𝐼 + b∗−1

∑�̃� ′∈ ˜X _�̃� ′𝜙 (𝑥

′)𝜙 (𝑥 ′)⊤), we have

max

�̃� ∈ ˜X∥𝜙 (𝑥)∥2(

b∗𝐼+∑�̃�′∈ ˜X _

∗�̃�′𝜙 (�̃�

′)𝜙 (�̃� ′)⊤)−1

=∑︁�̃� ∈ ˜X

_∗�̃�∥𝜙 (𝑥)∥2(

b∗𝐼+∑�̃�′∈ ˜X _

∗�̃�′𝜙 (�̃�

′)𝜙 (�̃� ′)⊤)−1

Lemma 10. For _∗ = argmax_∈△ ˜Xlog det

(𝐼 + b∗−1

∑�̃� ′∈ ˜X _�̃� ′𝜙 (𝑥

′)𝜙 (𝑥 ′)⊤), we have

∑︁�̃� ∈ ˜X

log

©­«1 + _∗�̃�∥𝜙 (𝑥)∥2(

b∗𝐼+∑�̃�′∈ ˜X _

∗�̃�′𝜙 (�̃�

′)𝜙 (�̃� ′)⊤)−1

ª®¬ ≤ log

det

(b∗𝐼 +

∑�̃� ∈ ˜X _

∗�̃�𝜙 (𝑥)𝜙 (𝑥)⊤

)det (b∗𝐼 )

Proof of Lemma 10. For any 𝑗 ∈ [𝑛𝑉 ], let𝑀𝑗 = det

(b∗𝐼 +

∑𝑖∈[ 𝑗 ] _

∗𝑖𝜙 (𝑥𝑖 )𝜙 (𝑥𝑖 )⊤

).

det

©­«b∗𝐼 +∑︁�̃� ∈ ˜X

_∗�̃�𝜙 (𝑥)𝜙 (𝑥)⊤ª®¬

= det

©­«b∗𝐼 +∑︁

𝑖∈[𝑛𝑉−1]_∗𝑖 𝜙 (𝑥𝑖 )𝜙 (𝑥𝑖 )

⊤ + _∗𝑛𝑉𝜙 (𝑥𝑛𝑉 )𝜙 (𝑥𝑛𝑉 )⊤ª®¬

= det (𝑀𝑛𝑉−1) det

(𝐼 + _∗𝑛𝑉 ·𝑀

− 1

2

𝑛𝑉−1𝜙 (𝑥𝑛𝑉 )

(𝑀− 1

2

𝑛𝑉−1𝜙 (𝑥𝑛𝑉 )

)⊤)= det (𝑀𝑛𝑉−1) det

(𝐼 + _∗𝑛𝑉 · 𝜙 (𝑥𝑛𝑉 )

⊤𝑀−1

𝑛𝑉−1𝜙 (𝑥𝑛𝑉 )

)= det (𝑀𝑛𝑉−1)

(1 + _∗𝑛𝑉 ∥𝜙 (𝑥𝑛𝑉 )∥

2

𝑀−1

𝑛𝑉−1

)= det (b∗𝐼 )

𝑛𝑉∏𝑖=1

(1 + _∗𝑖 ∥𝜙 (𝑥𝑖 )∥

2

𝑀−1

𝑖−1

)43

Woodstock ’21, June 03–05, 2021, Woodstock, NY

Thus,

det

(b∗𝐼 +

∑�̃� ∈ ˜X _

∗�̃�𝜙 (𝑥)𝜙 (𝑥)⊤

)det (b∗𝐼 )

=

𝑛𝑉∏𝑖=1

(1 + _∗𝑖 ∥𝜙 (𝑥𝑖 )∥

2

𝑀−1

𝑖−1

)Taking logarithm on both sides, we have

log

det

(b∗𝐼 +

∑�̃� ∈ ˜X _

∗�̃�𝜙 (𝑥)𝜙 (𝑥)⊤

)det (b∗𝐼 )

=

𝑛𝑉∑︁𝑖=1

log

(1 + _∗𝑖 ∥𝜙 (𝑥𝑖 )∥

2

𝑀−1

𝑖−1

)≥𝑛𝑉∑︁𝑖=1

log

(1 + _∗𝑖 ∥𝜙 (𝑥𝑖 )∥

2

𝑀−1

𝑛𝑉

),

which completes the proof of Lemma 10. □

Lemma 11 (Pinsker’s ineqality). If 𝑃 and 𝑄 are two probability distributions on a measurable space (𝑋, Σ), then for

any measurable event 𝐴 ∈ Σ, it holds that

|𝑃 (𝐴) −𝑄 (𝐴) | ≤√︂

1

2

KL(𝑃 ∥𝑄) .

Lemma 12 (Lemma 29 in [38]). For any 𝛾1, . . . , 𝛾𝐾 ∈ [0, 1] and 𝑥 ≥ 0, it holds that

𝐾∏𝑖=1

max{1 − 𝛾𝑖 − 𝛾𝑖𝑥, 0} ≥𝐾∏𝑖=1

(1 − 𝛾𝑖 ) − 𝑥

Lemma 13 (Fano’s Ineqality). Let 𝐴, 𝐵 be random variables and 𝑓 be a function that given 𝐴 predicts a value for 𝐵.

If Pr(𝑓 (𝐴) ≠ 𝐵) ≤ 𝛿 , then 𝐻 (𝐵 |𝐴) ≤ 𝐻2 (𝛿) + 𝛿 · log |𝐵 |.

Lemma 14. For the instance DΔ,𝑝𝑑

with sample profile S, we have 𝐻 (𝐼 ; 𝑆) = 𝑂 ( |𝑆 | · Δ2

𝑑).

The proof of Lemma 14 follows the same analysis procedure of Lemma 7 in [1].

Lemma 15 (Lemma 8 in [1]). Let 𝐴 ∼ D be a random variable on [𝑑] with 𝐻 (𝐴) ≥ log𝑑 − 𝛾 for some 𝛾 ≥ 1. For any

Y > exp(−𝛾), there exists ℓ + 1 distributions𝜓0,𝜓1, ...,𝜓ℓ on [𝑑] along with ℓ + 1 probabilities 𝑝0, 𝑝1, . . . , 𝑝ℓ (∑ℓ𝑖=0

𝑝𝑖 = 1)

for some ℓ = 𝑂 (𝛾/Y3) such that D =∑ℓ𝑖=1

𝑝𝑖𝜓𝑖 , 𝑝0 = 𝑂 (Y), and for any 𝑖 ≥ 1,

1. log |supp(𝜓𝑖 ) | ≥ log𝑑 − 𝛾/Y.2. ∥𝜓𝑖 −U𝑖 ∥TV = 𝑂 (Y) whereU𝑖 denotes the uniform distribution on supp(𝜓𝑖 ).

44