# provably_efficient_interactivegrounded_learning_with_personalized_reward__d1339c6e.pdf Provably Efficient Interaction-Grounded Learning with Personalized Reward Mengxiao Zhang University of Iowa mengxiao-zhang@uiowa.edu Yuheng Zhang University of Illinois Urbana-Champaign yuhengz2@illinois.edu Haipeng Luo University of Southern California haipengl@usc.edu Paul Mineiro Microsoft Research pmineiro@microsoft.com Interaction-Grounded Learning (IGL) [Xie et al., 2021] is a powerful framework in which a learner aims at maximizing unobservable rewards through interacting with an environment and observing reward-dependent feedback on the taken actions. To deal with personalized rewards that are ubiquitous in applications such as recommendation systems, Maghakian et al. [2022] study a version of IGL with context-dependent feedback, but their algorithm does not come with theoretical guarantees. In this work, we consider the same problem and provide the first provably efficient algorithms with sublinear regret under realizability. Our analysis reveals that the step-function estimator of prior work can deviate uncontrollably due to finite-sample effects. Our solution is a novel Lipschitz reward estimator which underestimates the true reward and enjoys favorable generalization performances. Building on this estimator, we propose two algorithms, one based on explore-thenexploit and the other based on inverse-gap weighting. We apply IGL to learning from image feedback and learning from text feedback, which are reward-free settings that arise in practice. Experimental results showcase the importance of using our Lipschitz reward estimator and the overall effectiveness of our algorithms. 1 Introduction Traditional bandit problems [Auer et al., 2002] or reinforcement learning problems [Sutton and Barto, 2018] assume that the learner has access to the reward, which is her learning goal. However, such explicit reward information is usually difficult to obtain in many real-world scenarios, including human-computer interface applications [Pantic and Rothkrantz, 2003, Freeman et al., 2017] and recommender systems [Yi et al., 2014, Wu et al., 2017]. Recently, Xie et al. [2021] study a new setting called Interaction-Grounded Learning (IGL), where the learner interacts with the environment and receives some feedback on her actions instead of explicit reward signals. The learner needs to discover the implicit information about the reward in the feedback and maximizes the reward. From an information-theoretic perspective, IGL is intractable unless further assumptions are made. Xie et al. [2021] introduce a conditional independence assumption which states that the feedback is conditionally independent of the action and context given the latent reward. However, this assumption is unrealistic in many scenarios. For example, different users interact with recommender systems in different ways even under the same latent reward [Maghakian et al., 2022]. This inspires us to study IGL with personalized rewards, a setting where the feedback depends on the context. Although Equal contribution. 38th Conference on Neural Information Processing Systems (Neur IPS 2024). Maghakian et al. [2022] study this for recommender systems, they only provide empirical results of their approach, leaving the following question open: Can we design provably efficient algorithms for interaction-grounded learning when the feedback depends on the context given the latent reward? Contributions. In this paper, we answer the question in the positive and provide the first provably efficient algorithms with sublinear regret guarantee for IGL with personalized reward. To achieve this, in Section 3.1, we first propose a novel reward estimator via inverse kinematics. Specifically, using the samples collected by applying a uniform policy, we construct a Lipschitz reward, which underestimates the reward for all the policies but approximates the reward for the optimal policy well. With this reward estimator, we propose two algorithms, one based on Explore-then-Exploit (Section 3.2) and the other based on inverse-gap weighting (Section 3.3). Both algorithms achieve O(T 2 3 ) regret. To the best of our knowledge, we are the first to propose provable regret guarantees for IGL with personalized reward. In Section 4, we implement both algorithms and apply them to both an image classification dataset and a conversational dataset. The empirical performance showcases the effectiveness of our algorithm and the importance of using the Lipschitz reward estimator. 1.1 Related Work Interaction-Grounded Learning (IGL). Xie et al. [2021] is the first work studying IGL and assumes that the feedback is independent of the context and action conditioned on the latent reward. Xie et al. [2022] further relaxes the assumption to an action-inclusive version, where the feedback is independent of context conditioned on both the action and the reward.2 Hu et al. [2024] follows the same setting as Xie et al. [2021] but proposes an information-theoretic approach to enforce the conditional independence assumption. However, as we mentioned before, context-dependent feedback is ubiquitous in real-word applications, so such conditional independence assumptions reduce the generality of the IGL framework. Maghakian et al. [2022] is the closest work to ours which also considers personalized rewards. Nonetheless, their work focuses on the empirical side and does not provide any theoretical guarantees. Contextual online learning with partial feedback. Our work is closely related to the recent trend of designing efficient contextual learning algorithms with partial feedback, including contextual bandits [Langford and Zhang, 2007, Agarwal et al., 2012, 2014, Foster and Krishnamurthy, 2018, Foster et al., 2018, Foster and Rakhlin, 2020, Simchi-Levi and Xu, 2021], where the learner receives explicit reward signal of her taken action; contextual bandits with feedback graphs [Zhang et al., 2024a,b], where the learner s observation of the reward is determined based on a feedback graph; and contextual partial monitoring [Bartók and Szepesvári, 2012, Kirschner et al., 2020], where the learner s observation is defined by a signal matrix or a linear observation operator. We adopt and generalize the ideas for designing contextual bandits algorithms [Foster and Rakhlin, 2020] to design algorithms for IGL. 2 Preliminary Problem setup. Throughout the paper, we use [N] to denote the set {1, 2, . . . , N} for a positive integer N. The problem of online Interative-Grounded Learning (IGL) with personalized reward we consider is defined as follows. The interaction between the learner and the environment lasts for T rounds. At each round t [T], the environment reveals a stochastic context xt X i.i.d. drawn from an unknown distribution D, and the learner decides an action at [K] from a finite action set of size K based on this context. Then, different from the classic contextual bandits in which the learner observes the binary reward of her chosen action r(xt, at) {0, 1}, she receives feedback yt Y. 2We point out that the idea behind Xie et al. [2022] cannot be used in our setting where feedback can be dependent on the context. Specifically, Xie et al. [2022] propose an algorithm which learns the value function and the reward decoder separately for each action. Generalizing their idea to our setting would then mean learning the value function and the reward decoder for each context, which is infeasible since there could be infinitely many contexts. Feedback dependence assumption. We follow the assumption proposed in [Maghakian et al., 2022] that the feedback is conditionally independent of the action given the context and the realized reward. Assumption 1. For arbitrary (x, a, r, y) tuple where the reward r and the feedback y are generated based on the context x and action a, we assume y is conditionally independent of action a given context x and reward r. In other words, we assume that y a | r, x. As mentioned, compared to prior work [Xie et al., 2021, 2022], this better captures many real-world applications such as recommender systems where different users interact with them in different ways [Beel et al., 2013, Shin, 2020, Maghakian et al., 2022]. Realizability. We assume that the learner has access to two function classes F (X [K] 7 [0, 1]) and Φ (X Y 7 {0, 1}), where F characterizes the mean of the reward for a given contextaction pair, and Φ characterizes the realized reward given the context and the received feedback. A policy π : X (K) specifies the action probability conditioned on the context. For each f F, we use πf to denote the induced policy which takes action greedily according to f, that is, πf(a|x) = 1{a = argmaxa [K] f(x, a )}, a [K]. We also use the shorthand π for the optimal policy πf . We then make the following realizability assumption following previous contextual bandits literature [Agarwal et al., 2012, Foster et al., 2018, Foster and Rakhlin, 2020, Simchi-Levi and Xu, 2021]. Assumption 2 (Realizability). There exists a regression function f F such that E[r(xt, a)|xt] = f (xt, a) for all a [K] and t [T]. Furthermore, there exists a feedback decoder ϕ Φ such that ϕ (xt, yt) = r(xt, at) for all t [T]. For simplicity, we also assume that F and Φ are finite with cardinality |F| and |Φ|. Our results can be further generalized to broader function classes, which will be discussed in later sections. Identifiability. As mentioned in [Xie et al., 2021, 2022, Maghakian et al., 2022], it is impossible to learn if we do not break the symmetry between reward being 1 and being 0. Following [Maghakian et al., 2022], we make the following assumption: Assumption 3 (Identifiability). For any x X, f defined in Assumption 2 satisfies that: 1) PK a=1 f (x, a) α for some 0 < α < K 2 ; 2) maxa [K] f (x, a) θ where θ > α K α. The first condition says that the sum of the expected reward over actions is less than K 2 given any context x. That is to say, the reward vector is sparse if f (x, a) {0, 1}. The second condition says that for each context x X, there exists an action that achieves a large enough expected reward. These two conditions are indeed satisfied by many real-world applications, including the s-multi-label classification problem (with s < K/2) where α = s and θ = 1. Regret. The learner s performance is measured via the notion of regret, which is defined as the expected difference between the learner s total reward and the one received by the optimal policy: t=1 f (xt, π (xt)) t=1 f (xt, at) Other notations. We denote the (K 1)-dimensional simplex by K. Let 1 be the all-one vector in an appropriate dimension and πUnif = 1 K 1 K. For a d-dimensional vector v Rd, we denote its i-th entry by vi. 1{ } is the indicator function and ei is the i-th standard basis vector in an appropriate dimension. 3 Methodology In this section, we discuss our methodology. In Section 3.1, we start from introducing how we construct a Lipschitz reward estimator based on uniform samples, which serves as the key for our algorithm construction. We prove that this estimator is an underestimator of the reward, and more importantly matches the reward for the optimal policy. Based on this estimator, we design two algorithms for this problem in Section 3.2 and Section 3.3, with the first one based on explore-thenexploit and the second one based on inverse-gap weighting (IGW). 3.1 Reward Estimator Construction via Uniform Exploration We first show how we construct a reward estimator based on feedback collected from a uniform policy, which serves as the most important component in our two algorithms. 3.1.1 Inverse Kinematics When the reward for each context-action pair is deterministic and binary, Maghakian et al. [2022] show that if the learner uniformly samples an action for any context and is able to accurately predict the posterior probability of her chosen action given the context and the feedback (inverse kinematics), then she is able to infer that the reward is 1 if that posterior probability is above certain threshold. Here, we generalize this thresholding reward estimator to the randomized binary reward case, and further prove that this estimator correctly models the reward for the optimal policy. To see this, we first prove the following lemma showing the exact posterior distribution over actions if the learner selects an action uniformly randomly. This is also proven in [Maghakian et al., 2022, Eq.(2)] as well, and we include it here for completeness. Lemma 1. For any context x X, suppose that the learner picks a uniformly random action a [K]. Let r and y be its realized reward and the corresponding feedback. Then, under Assumption 1 and Assumption 2, the posterior distribution of a given the context x and feedback y equals to Pr[a|x, y] = f (x, a) ϕ (x, y) PK a =1 f (x, a ) + (1 f (x, a))(1 ϕ (x, y)) K PK a =1 f (x, a ) , (1) where f and ϕ are the true expected reward and feedback decoder defined in Assumption 2. Now we show how to infer the true reward from this inverse kinematics. Specifically, we show: Lemma 2. For any context x X and action a [K], let r(x, a) be the realized reward and y be the feedback. Let h (x, y) K be such that for each a [K], h a(x, y) f (x, a) ϕ (x, y) PK a =1 f (x, a ) + (1 f (x, a))(1 ϕ (x, y)) K PK a =1 f (x, a ) . (2) Then we have r(x, a) = ϕ (x, y) 1{h a(x, y) θ α} for all a [K]; r(x, a) = ϕ (x, y) = 1{h a(x, y) θ α} for a = π (x). Proof. Note that since ϕ (x, y) {0, 1}, only one term can be non-zero in Eq. (2). If h a(x, y) θ α, where θ and α are defined in Assumption 3, then the reward ϕ (x, y) has to be 1 since otherwise, we have ϕ (x, y) = 0 and θ α h a(x, y) = 1 f (x, a) K PK a =1 f (x, a ) 1 K α < θ where the second and the third inequality are due to Assumption 3. This leads to a contradiction. Therefore, we know that the realized reward is 1 if h a(x, y) is no less than θ α. Therefore, 1{h a(x, y) θ α} can be viewed as an underestimator of r(x, a). To prove the second property, consider the case in which a = π (x). Then, we know that when ϕ (x, y) = 1, we must also have h a(x, y) θ h a(x, y) = f (x, π (x)) PK i=1 f (x, a) f (x, π (x)) where the first inequality uses the first property in Assumption 3 and the second inequality uses the second property in Assumption 3. Lemma 2 shows that the function h (x, y) defined in Eq. (2) satisfies two important properties. First, 1{h a(x, y) θ α} serves as a reward underestimator for all the policies. Second, it matches the reward of the optimal policy. Note that this is different from [Maghakian et al., 2022, Eq.(3)], since they only consider the deterministic binary reward for each context-action pair, and they do not show that the constructed reward estimator matches the reward of the optimal policy. Based on these two properties, we know that if we have access to h , then the policy π that maximizes the surrogate reward E h 1{h π(x)(x, y)} θ α i also maximizes the true expected reward. 3.1.2 Learning the Posterior Distribution via ERM Next, we show how to learn this h via uniformly collected samples. Define the function class H as: H = h : X Y 7 K, ha(x, y) = f(x, a)ϕ(x, y) PK i=1 f(x, i) + (1 f(x, a))(1 ϕ(x, y)) K PK i=1 f(x, i) f F, ϕ Φ, a [K] . Note that |H| = |F| |Φ|. Since h models the posterior distribution over actions when the learner selects her action uniformly randomly, we collect N tuples of (xt, at, yt) by sampling at uniformly from [K] and find the empirical risk minimizer (ERM) bh H over these samples using squared loss. In the following lemma, we show that bh enjoys O log |H| N excess risk with high probability. Lemma 3. Let {(xi, ai, yi)}N i=1 be N i.i.d. samples where xi D, a πUnif, and yi is the corresponding feedback. Let bh be the ERM with respect to the squared loss defined as follows: bh = argmin h H i=1 h(xi, yi) eai 2 2 Then, with probability at least 1 δ, we have Ex D,a πUnif,y|x,a h bh(x, y) ea 2 2 h (x, y) ea 2 2 i O Ex D,a πUnif,y|x,a h bh(x, y) h (x, y) 2 i O The full proof is deferred to Appendix A. Different from the classic one-dimensional squared loss, here we consider the ℓ2-loss between two vectors in K. Directly applying the generalization bound for each entry leads to a K-factor worse bound. Instead, our proof is based on the observation that the loss function h(x, y) ea 2 2, when seen as a function of h, satisfies the so-called strong 1-central condition [Van Erven et al., 2015, Definition 7]. Moreover, these results can be extended to function classes with infinite size and bounded covering number based on Theorem 7.7 of [Van Erven et al., 2015]. 3.1.3 Constructing Lipschitz Reward Estimators Based on bh Now we show how to construct a reward estimator based on the ERM bh. According to Section 3.1.1, an intuitive form of the reward (under)estimator is 1{bha(x, y) θ α}. However, since the indicator function is not Lipschitz, 1{bha(x, y) θ α} can be very different from 1{h a(x, y) θ α} even with the generalization bound proven in Lemma 3. To resolve this issue, we propose a Lipschitz variant of the indicator function (defined in Eq. (6)) and show that it also satisfies the two properties shown in Lemma 2. Lemma 4. Define G(v, β, σ) as G(v, β, σ) = 1 σ (v β)1{β v < β + σ} + 1{v β + σ}. (6) Then, for any context x X, action a [K], and feedback y Y generated via context x and the realized reward r(x, a), we have the following two properties with σ 1 2 θ α 1 K α > 0. Algorithm 1 Off-Policy IGL Input: number of exploration samples N, parameters α, θ and σ = 1 2 θ α 1 K α . for t = 1, 2, , 2N do Receive context xt, sample at πUnif, and observe signal yt. Calculate the empirical risk minimizer bh as in Eq. (3). Construct the reward decoder brσ(x, y, a) = G(bha(x, y), θ α σ, σ) where G is defined in Eq. (6). Calculate bπ = argmaxπ Π n P2N i=N+1 π(ai|xi) brσ(xi, yi, ai) o where Π = {πf : f F}. for t = 2N + 1, . . . , T do Execute policy bπ. r(x, a) = ϕ (x, y) G(h a(x, y), θ r(x, a) = ϕ (x, y) = G(h a(x, y), θ α σ, σ) if a = π (x), The proof shares a similar spirit to Lemma 2 and is deferred to Appendix A. Notably, the function G(v, β, σ) is 1 σ-Lipschitz in v, which is important in order to show concentration between the reward estimator with respect to the true posterior distribution G(h a(x, y), θ α σ, σ) and that constructed via the ERM function bh: G(bha(x, y), θ α σ, σ). In the following, we will show how to use the reward estimator G(bha(x, y), θ α σ, σ) to design algorithms with provable guarantees. 3.2 Off-Policy Algorithm Built on the reward estimator in Section 3.1, we present our off-policy algorithm, summarized in Algorithm 1. Our algorithm follows the explore-then-exploit idea and consists of two phases. In the exploration phase, we perform uniform exploration and collect the dataset {xi, ai, yi}2N i=1. The first N samples are used to learn the reward estimator br and the rest N samples are used to learn the policy bπ with br. For the exploitation phase, we employ the learned policy bπ in the remaining T 2N iterations. We present the regret bound of Algorithm 1 in the following theorem. Theorem 1. Under Assumptions 1-3, Algorithm 1 with N = T 2/3K2/3σ 2/3 log1/3(|H|T) guaran- tees that Reg CB O T 2/3K2/3σ 2/3 log1/3(|H|T) . We remark that this is the first provably efficient algorithm for IGL with personalized reward. The key of the proof is to show that the learned policy bπ is near-optimal under the true reward. This is achieved by using the properties of function G in Lemma 4 and proving that the reward decoder brσ is close to the ground-truth. The full proof is deferred to Appendix B. Besides the dependence on T, K and log |H|, our regret bound also depends on σ 1, which measures how sparse the reward vector is and characterizes the difficulty of the problem. For example, σ 1 = O(1) in the multi-class classification problem and σ 1 O(s2) in the s-multi-label classification problem. 3.3 On-Policy Algorithm Since on-policy algorithms are more favorable in practice, we further introduce an on-policy algorithm based on the inverse-gap weighting strategy. Following [Foster and Rakhlin, 2020], we assume access to an online regression oracle Alg Sq: at each round t [T], the oracle Alg Sq produces an estimator bft in the convex hull of F, then receives a context-action-reward tuple (xt, at, rt). The squared loss of the oracle for this round is defined as ( bft(xt, at) rt)2, which is on average assumed to be close to that of the best predictor in F. Assumption 4 (Bounded squared loss regret). For any sequence {(xt, at, rt)}T t=1, the regression oracle Alg Sq guarantees the following regret bound for some Reg Sq that depends on T, K, and F: bft(xt, at) rt 2 inf f F t=1 (f(xt, at) rt)2 Reg Sq. Algorithm 2 On-policy IGL Input: online regression oracle Alg Sq, number of exploration samples N, parameters α, θ, γ and σ = 1 2 θ α 1 K α . for t = 1, 2, , N do Receive context xt, sample at πUnif, and observe signal yt. Calculate the empirical risk minimizer bh as in Eq. (3). for t = N + 1, . . . , T do Receive context xt. Obtain an estimator bft from the oracle Alg Sq. Calculate the action distribution pt as ( 1 K+γ( b ft(xt,ba) b ft(xt,a)), a = ba, a =ba pa , a = ba , (7) where ba = argmaxa [K] bft(xt, a). Sample at from pt and receive feedback yt. Feed (xt, at, G(bhat(xt, yt), θ α σ, σ)) to the oracle Alg Sq where G is defined in Eq. (6). Based on the ground-truth inverse kinematics function h , we define a function f (x, a) := Ey|x,a G(h a(x, y), θ α σ, σ) , which is always a lower bound on f (x, a) according to Lemma 4. Since we only feed the surrogate reward to the oracle, we make another mild assumption that our regression function class F also realizes f . Assumption 5 (Lower Bound Realizability). We also assume that f F, where f is defined as f (x, a) = Ey|x,a G(h a(x, y), θ We now summarize our algorithm in Algorithm 2. After obtaining the inverse kinematics predictor bh in the same manner as Algorithm 1, instead of uniform exploring, we use the estimated reward from the oracle and an inverse-gap weighting strategy [Abe and Long, 1999, Foster and Rakhlin, 2020] defined in Eq. (7). Different from the contextual bandit problem where the true reward is given and fed to the oracle, we feed the predicted reward G(bhat(xt, yt), θ α σ, σ) to the oracle. One might wonder how such misspecification in rewards would affect the regret bound. Our key observation is that since we use the uniform policy to collect data used to train bh, the generalization error of bh is small for any a due to good coverage of the dataset on each action. Based on this observation, we prove the following theorem for Algorithm 2. Theorem 2. Under Assumptions 1-5, Algorithm 2 with certain choice of N and γ guarantees that Reg CB = O p KTReg Sq + σ 2/3(KT)2/3 log1/3(|H|T) . The proof mainly relies on the generalization bounds in Lemma 3 and the property of G in Lemma 4, and is deferred to Appendix C. We observe that Algorithm 2 enjoys the same dependence on T, K, σ 1 as Algorithm 1. For finite F, we can use Vovk s aggregation algorithm [Vovk, 1995] as the regression oracle and achieve Reg Sq = O (log |F|), making the second term in the regret bound negligible.3 While in theory our on-policy algorithm does not seem to be more favorable than the off-policy algorithm (mostly because they both need to uniformly explore for a certain period in order to build bh), it can still perform better in practice, as shown in our experiments. 4 Experiments In this section, we apply IGL to learning from image feedback and learning from text feedback. Specifically, we conduct experiments on the MNIST dataset and a conversational dataset to verify the effectiveness of our algorithms and the Lipschitz reward estimator constructed by Eq. (6). 3We refer readers to Foster and Rakhlin [2020] for more examples of regression oracles when F is not finite. Table 1: Performance of Algorithm 1 and Algorithm 2 on the MNIST dataset. Algorithm Reward Estimator Average Progressive Reward Test Accuracy Off-policy Algorithm 1 Binary 0.614 (0.012) 72.6% (1.5%) Lipschitz 0.638 (0.042) 75.9% (4.9%) On-policy Algorithm 2 Binary 0.718 (0.006) 89.4% (4.0%) Lipschitz 0.740 (0.022) 90.3% (3.7%) 4.1 Experiments on Learning from Image Feedback Experiment Setup We first conduct experiments on MNIST dataset and implement both the offpolicy algorithm Algorithm 1 and the on-policy algorithm Algorithm 2. The setup is as follows: at each step t, the learner receives an image xt with ground-truth label lt and picks action at from {0, , 9} as the predicted label. If the prediction at is correct, the learner receives as feedback an image yt with digit (lt + 1) mod 10; otherwise, the learner receives an image with digit (lt 1) mod 10. Therefore, different from the experimental setup in [Xie et al., 2021], our feedback yt does depend on both the context and the reward. Both function classes H and F are implemented as two-layer convolutional neural networks. We use Py Torch framework [Paszke et al., 2019] and parameter-free optimizers [Orabona and Tommasi, 2017] to learn the reward estimator and the policy. For both algorithms, we set the number of exploration samples N = 5000 and pick the parameter σ {0, 0.05, 0.1, 0.15, 0.2, 0.25, 0.3} and θ α { 1 2 }. For the on-policy algorithm, we use a time-varying exploration parameter γt = Kt as suggested by Foster and Krishnamurthy [2021]. We run the experiments on one NVIDIA Ge Force RTX 2080 Ti. The interaction lasts for T = 60000 rounds. We use two metrics to evaluate the performance of the algorithm, including the average progressive reward during the interaction process and the test accuracy on a held-out test set containing 10000 samples. When evaluating the on-policy algorithm on the test set, we take actions greedily according to bf T . Since we use the uniform policy to collect data for learning the reward estimator, the progressive reward at that phase is counted as 1 Results We run the experiments with 4 different random seeds and report the mean value and standard deviation in Table 1. The running averaged reward over the entire T rounds are shown in Figure 1. We can see that both algorithms achieve good performance, despite never observing any true labels. While the theoretical regret guarantees for both algorithms are of the same order, empirically, the on-policy Algorithm 2 performs better than Algorithm 1 with over 90% test accuracy. This is because the on-policy algorithm uses the inverse-gap weighting strategy to achieve a better trade-off between exploration and exploitation, while the off-policy algorithm learns the policy from uniform exploration. On the other hand, to demonstrate the effectiveness of the Lipschitz reward estimator, we compare the performances of the Lipschitz estimator Eq. (6) with a binary reward estimator brbinary(x, y, a) = 1{bha(x, y) θ α}, where the parameter θ α is searched over the same space. The results in Table 1 show that the Lipschitz reward estimator improves over the binary one by a clear margin for both algorithms. This matches our theoretical analysis that highlights the vital role of the Lipschitzness of the reward estimator in obtaining good regret guarantees. We also plot the performance of Algorithm 2 under both true reward and constructed reward in the left figure of Figure 2.4 The figure shows that the constructed reward is indeed a lower bound of the true reward, and the policy can learn from the constructed reward effectively. 4.2 Experiments on Learning From Text Feedback We further consider an application of learning with text feedback. The IGL framework fits this problem well since this is a natural reward-free setting involving learning from user s text feedback, which is idiosyncratically related to the quality of the actions taken and the question asked. Specifically, given the input of a question, the learner needs to select one of the K possible answers. Instead of 4Both the true reward and the constructed reward values are averaged over the rounds after the first N rounds of uniform exploration. Figure 1: Running averaged reward of Algorithm 1 and Algorithm 2 on MNIST. Note that Algorithm 1 uniformly explores in the first 2N = 10000 rounds, and thus its averaged reward at t = 10000 is about 1/K = 0.1. receiving whether the selected answer is correct, the learner only receives user s text feedback to the chosen answer. The goal of the learner is to choose the best answer only based on such text feedback. Dataset Construction Our dataset is constructed as follows. Specifically, we construct our question set S = {qi}i [20000] from Chatbot Arena datasets [Zheng et al., 2024]. Then, for each question qi S we ask a larger language model G with a high ELO score on the chatsys leaderboard [Chiang et al., 2024] to generate a good answer gi,0 with ri,0 = 1; and ask a (much) smaller language model B with a (much) lower ELO score to generate 4 bad answers bi,j with reward ri,j = 0, j {1, 2, 3, 4}. Specifically, we pick G to be Qwen1.5-32B-Chat with ELO score 1134 and B to be Qwen1.5-0.5B-Chat with ELO score5 less than 804 [Bai et al., 2023].6 For each questionanswer tuple (qi, gi,0, bi,1, bi,2, bi,3, bi,4), we ask another large language model R to simulate a user response fi,j, j {0, 1, 2, 3, 4} to the good (bad) answers under the instruction that the user is satisfied (unsatisfied) with the answer. We pick R to be Qwen1.5-32B-Chat as well. This forms our final dataset SConv = {(qi, (gi,0, fi,0, ri,0), {(bi,j, fi,j, ri,j)}j [4])}i [20000]. Again, the true reward is never revealed to the learner, and we only use this reward to measure the performance of our algorithm. The prompt we use is deferred to Appendix D.1. We generate our dataset using one A100 GPU for two weeks. 4.2.1 Algorithm Configurations and Results Given the superior performance of Algorithm 2 over Algorithm 1 on the MNIST dataset shown in Section 4.1, we only test Algorithm 2 on the conversational dataset. We use the first N = 10000 data points to learn bh and the remaining |S| N = 10000 data points to learn the optimal policy based on the reward function constructed via bh. We use the same parameter-free optimizer as the one in Section 4.1. Next, we introduce the construction of bh and the regression oracle: Inverse kinematic model. The model class H we consider is the pretrained language model Llama3-8B-Instruct [AI@Meta, 2024] with an additional multi-class classification head.7 The language model is prompted with a question-answer-feedback tuple (x, a, y); see Appendix D.2 for the prompt. To learn the inverse kinematic model, we use parameter efficient fine-tuning [Mangrulkar et al., 2024] with a rank-1 LORA adapter [Hu et al., 2021] and binary cross entropy loss, which is with respect to the indicator of whether-or-not the predicted action corresponds to the action selected in the tuple. 5The bad model B is not displayed on the leaderboard; 804 is the lowest displayed score. 6 Qwen1.5-32B-Chat is available at https://huggingface.co/Qwen/Qwen1.5-32B-Chat and Qwen1.5-0.5B-Chat is available at https://huggingface.co/Qwen/Qwen1.5-0.5B-Chat. 7 Llama-3-8B-Instruct : https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct. Figure 2: Performance of Algorithm 2 under true (unobserved) rewards and constructed rewards. Left figure: Results on MNIST dataset after the first N uniform exploration rounds. Right figure: Results on our conversational dataset. We successfully learn bh using one A100 GPU within 6 hours. After obtaining bh, we construct the reward estimator G(bha(x, y), θ α σ, σ) with σ = 0.1, α = K 2, and θ = 1. Regression oracle. Similar to the inverse kinematic model, the reward prediction function class F is again based on the pretrained Llama-3-8B-Instruct model but with an additional regression head. Specifically, the language model is prompted only with a question-answer pair (x, a) using the prompt deferred to Appendix D.3 and predicts a score in [0, 1] for this question-answer pair. The regression oracle again applies parameter efficient fine-tuning with a different rank-1 LORA adapter on the regression loss, which measures the error of predicting the output of the reward predictor G(bha(x, y), θ α σ, σ). This process is done on one A100 GPU within 3 hours. Results. The empirical results on the conversational dataset are shown in Figure 2. We show the running averaged true reward and the running average constructed reward received by Algorithm 2 after the first N = 10000 rounds of learning bh. The x-axis is the time horizon and y-axis is the value of average reward. Similar to our experiment results in Section 4.1, the right figure in Figure 2 shows that our constructed reward estimator is a lower bound on the true reward, matching our theoretical results in Lemma 4, and Algorithm 2 is able to learn the reward effectively through the text feedback with the constructed reward estimator. Acknowledgments and Disclosure of Funding HL and MZ were supported by NSF Award IIS-1943607. Naoki Abe and Philip M Long. Associative reinforcement learning using linear probabilistic concepts. In ICML, pages 3 11. Citeseer, 1999. Alekh Agarwal, Miroslav Dudík, Satyen Kale, John Langford, and Robert Schapire. Contextual bandit learning with predictable rewards. In Artificial Intelligence and Statistics, pages 19 26. PMLR, 2012. Alekh Agarwal, Daniel Hsu, Satyen Kale, John Langford, Lihong Li, and Robert Schapire. Taming the monster: A fast and simple algorithm for contextual bandits. In International Conference on Machine Learning, pages 1638 1646. PMLR, 2014. AI@Meta. Llama 3 model card. 2024. URL https://github.com/meta-llama/llama3/blob/ main/MODEL_CARD.md. Peter Auer, Nicolo Cesa-Bianchi, and Paul Fischer. Finite-time analysis of the multiarmed bandit problem. Machine learning, 47:235 256, 2002. Jinze Bai, Shuai Bai, Yunfei Chu, Zeyu Cui, Kai Dang, Xiaodong Deng, Yang Fan, Wenbin Ge, Yu Han, Fei Huang, Binyuan Hui, Luo Ji, Mei Li, Junyang Lin, Runji Lin, Dayiheng Liu, Gao Liu, Chengqiang Lu, Keming Lu, Jianxin Ma, Rui Men, Xingzhang Ren, Xuancheng Ren, Chuanqi Tan, Sinan Tan, Jianhong Tu, Peng Wang, Shijie Wang, Wei Wang, Shengguang Wu, Benfeng Xu, Jin Xu, An Yang, Hao Yang, Jian Yang, Shusheng Yang, Yang Yao, Bowen Yu, Hongyi Yuan, Zheng Yuan, Jianwei Zhang, Xingxuan Zhang, Yichang Zhang, Zhenru Zhang, Chang Zhou, Jingren Zhou, Xiaohuan Zhou, and Tianhang Zhu. Qwen technical report. ar Xiv preprint ar Xiv:2309.16609, 2023. Gábor Bartók and Csaba Szepesvári. Partial monitoring with side information. In International Conference on Algorithmic Learning Theory, pages 305 319. Springer, 2012. Joeran Beel, Stefan Langer, Andreas Nürnberger, and Marcel Genzmehr. The impact of demographics (age and gender) and other user-characteristics on evaluating recommender systems. In Research and Advanced Technology for Digital Libraries: International Conference on Theory and Practice of Digital Libraries, TPDL 2013, Valletta, Malta, September 22-26, 2013. Proceedings 3, pages 396 400. Springer, 2013. Wei-Lin Chiang, Lianmin Zheng, Ying Sheng, Anastasios Nikolas Angelopoulos, Tianle Li, Dacheng Li, Hao Zhang, Banghua Zhu, Michael Jordan, Joseph E Gonzalez, et al. Chatbot arena: An open platform for evaluating llms by human preference. ar Xiv preprint ar Xiv:2403.04132, 2024. Dylan Foster and Alexander Rakhlin. Beyond ucb: Optimal and efficient contextual bandits with regression oracles. In International Conference on Machine Learning, pages 3199 3210. PMLR, 2020. Dylan Foster, Alekh Agarwal, Miroslav Dudik, Haipeng Luo, and Robert Schapire. Practical contextual bandits with regression oracles. In International Conference on Machine Learning, pages 1539 1548. PMLR, 2018. Dylan J Foster and Akshay Krishnamurthy. Contextual bandits with surrogate losses: Margin bounds and efficient algorithms. Advances in Neural Information Processing Systems, 31, 2018. Dylan J Foster and Akshay Krishnamurthy. Efficient first-order contextual bandits: Prediction, allocation, and triangular discrimination. Advances in Neural Information Processing Systems, 34: 18907 18919, 2021. Euan Freeman, Graham Wilson, Dong-Bach Vo, Alex Ng, Ioannis Politis, and Stephen Brewster. Multimodal feedback in hci: haptics, non-speech audio, and their applications. In The Handbook of Multimodal-Multisensor Interfaces: Foundations, User Modeling, and Common Modality Combinations-Volume 1, pages 277 317. 2017. Edward J Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, and Weizhu Chen. Lora: Low-rank adaptation of large language models. ar Xiv preprint ar Xiv:2106.09685, 2021. Xiaoyan Hu, Farzan Farnia, and Ho-fung Leung. An information theoretic approach to interactiongrounded learning. ar Xiv preprint ar Xiv:2401.05015, 2024. Johannes Kirschner, Tor Lattimore, and Andreas Krause. Information directed sampling for linear partial monitoring. In Proceedings of Thirty Third Conference on Learning Theory, pages 2328 2369, 2020. John Langford and Tong Zhang. The epoch-greedy algorithm for multi-armed bandits with side information. Advances in neural information processing systems, 20, 2007. Jessica Maghakian, Paul Mineiro, Kishan Panaganti, Mark Rucker, Akanksha Saran, and Cheng Tan. Personalized reward learning with interaction-grounded learning (igl). ar Xiv preprint ar Xiv:2211.15823, 2022. S Mangrulkar, S. Gugger, L. Debut, Y. Belkada, and Paul S. Peft: State-of-the-art parameter-efficient fine-tuning methods. https://github.com/huggingface/peft, 2024. Accessed: 2024-05-14. Francesco Orabona and Tatiana Tommasi. Training deep networks without learning rates through coin betting. Advances in Neural Information Processing Systems, 30, 2017. URL https: //github.com/bremen79/parameterfree. Maja Pantic and Leon JM Rothkrantz. Toward an affect-sensitive multimodal human-computer interaction. Proceedings of the IEEE, 91(9):1370 1390, 2003. Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, et al. Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems, 32, 2019. Donghee Shin. How do users interact with algorithm recommender systems? the interaction of users, algorithms, and performance. Computers in human behavior, 109:106344, 2020. David Simchi-Levi and Yunzong Xu. Bypassing the monster: A faster and simpler optimal algorithm for contextual bandits under realizability. Mathematics of Operations Research, 2021. Richard S Sutton and Andrew G Barto. Reinforcement learning: An introduction. MIT press, 2018. Tim Van Erven, Peter Grunwald, Nishant A Mehta, Mark Reid, Robert Williamson, et al. Fast rates in statistical and online learning. Journal of Machine Learning Research, 54(6), 2015. Vladimir G Vovk. A game of prediction with expert advice. In Proceedings of the eighth annual conference on Computational learning theory, pages 51 60, 1995. Qingyun Wu, Hongning Wang, Liangjie Hong, and Yue Shi. Returning is believing: Optimizing longterm user engagement in recommender systems. In Proceedings of the 2017 ACM on Conference on Information and Knowledge Management, pages 1927 1936, 2017. Tengyang Xie, John Langford, Paul Mineiro, and Ida Momennejad. Interaction-grounded learning. In International Conference on Machine Learning, pages 11414 11423. PMLR, 2021. Tengyang Xie, Akanksha Saran, Dylan J Foster, Lekan Molu, Ida Momennejad, Nan Jiang, Paul Mineiro, and John Langford. Interaction-grounded learning with action-inclusive feedback. Advances in Neural Information Processing Systems, 35:12529 12541, 2022. Xing Yi, Liangjie Hong, Erheng Zhong, Nanthan Nan Liu, and Suju Rajan. Beyond clicks: dwell time for personalization. In Proceedings of the 8th ACM Conference on Recommender systems, pages 113 120, 2014. Mengxiao Zhang, Yuheng Zhang, Haipeng Luo, and Paul Mineiro. Efficient contextual bandits with uninformed feedback graphs. International Conference on Machine Learning, 2024a. Mengxiao Zhang, Yuheng Zhang, Olga Vrousgou, Haipeng Luo, and Paul Mineiro. Practical contextual bandits with feedback graphs. Advances in Neural Information Processing Systems, 36, 2024b. Lianmin Zheng, Wei-Lin Chiang, Ying Sheng, Siyuan Zhuang, Zhanghao Wu, Yonghao Zhuang, Zi Lin, Zhuohan Li, Dacheng Li, Eric Xing, et al. Judging llm-as-a-judge with mt-bench and chatbot arena. Advances in Neural Information Processing Systems, 36, 2024. A Proofs in Section 3.1 Lemma 1. For any context x X, suppose that the learner picks a uniformly random action a [K]. Let r and y be its realized reward and the corresponding feedback. Then, under Assumption 1 and Assumption 2, the posterior distribution of a given the context x and feedback y equals to Pr[a|x, y] = f (x, a) ϕ (x, y) PK a =1 f (x, a ) + (1 f (x, a))(1 ϕ (x, y)) K PK a =1 f (x, a ) , (1) where f and ϕ are the true expected reward and feedback decoder defined in Assumption 2. Pr[a|x, y] = Pr[a|x] Pr[y|x, a] = Pr[a|x]Pr[r = 1|x, a] Pr[y|x, a, r = 1] + Pr[r = 0|x, a] Pr[y|x, a, r = 0] = Pr[a|x]f (x, a) Pr[y|x, r = 1] + (1 f (x, a)) Pr[y|x, r = 0] = Pr[a|x]f (x, a) Pr[r = 1|x, y] Pr[r = 1|x] + Pr[a|x](1 f (x, a)) Pr[r = 0|x, y] Pr[r = 0|x] = Pr[a|x] f (x, a) Pr[r = 1|x, y] PK a =1 Pr[a |x]Pr[r = 1|x, a ] + Pr[a|x](1 f (x, a)) Pr[r = 0|x, y] PK a =1 Pr[a |x]Pr[r = 0|x, a ] = Pr[a|x] f (x, a) Pr[r = 1|x, y] PK a =1 Pr[a |x]f (x, a ) + Pr[a|x](1 f (x, a)) Pr[r = 0|x, y] PK a =1 Pr[a |x](1 f (x, a ) . Recall that πUnif = 1 K 1 is the policy which uniformly samples an action regardless of the context. Under πUnif, we know that Pr[a|x, y] = f (x, a) ϕ (x, y) PK a =1 f (x, a ) + (1 f (x, a))(1 ϕ (x, y)) K PK a =1 f (x, a ) . Lemma 3. Let {(xi, ai, yi)}N i=1 be N i.i.d. samples where xi D, a πUnif, and yi is the corresponding feedback. Let bh be the ERM with respect to the squared loss defined as follows: bh = argmin h H i=1 h(xi, yi) eai 2 2 Then, with probability at least 1 δ, we have Ex D,a πUnif,y|x,a h bh(x, y) ea 2 2 h (x, y) ea 2 2 i O Ex D,a πUnif,y|x,a h bh(x, y) h (x, y) 2 i O Proof. For notational convenience, we denote (x, y, a) by Z and define ℓh(Z) = h(x, y) ea 2 2. Now we aim to show that ℓh(Z) satisfies the strong η-central condition for some η > 0 [Van Erven et al., 2015]. Specifically, we aim to show that EZ [exp( η(ℓh(Z) ℓh (Z)))] 1. (8) To show this, direct calculation shows that EZ [exp( η(ℓh(Z) ℓh (Z)))] = Ex,y exp( η h(x, y) h (x, y) 2 2) Ea|x,y exp 2η(h(x, y) h (x, y)) (h (x, y) ea) . Since Ea|x,y[ea] = h (x, y), the random variable (h(x, y) h (x, y)) (h (x, y) ea) given x and y is zero-mean and is within the range [ 2 h(x, y) h (x, y) 2, 2 h(x, y) h (x, y) 2]. Therefore, we know that (h(x, y) h (x, y)) (h (x, y) ea) is h(x, y) h (x, y) 2 2-sub-Gaussian and we have EZ [exp( η(ℓh(Z) ℓh (Z)))] exp( η h(x, y) h (x, y) 2 2) exp 4η2 h(x, y) h (x, y) 2 2 2 = Ex,y exp 2η2 η h(x, y) h (x, y) 2 2 . Picking η = 1 2 proves Eq. (8). Noticing the fact that |ℓh(Z)| 4 for all h H and applying Theorem 7.6 in [Van Erven et al., 2015] proves Eq. (4). Now we prove Eq. (5). Noticing the fact that Ea|x,y[ea] = h (x, y) and applying this to Eq. (4), we know that with probability at least 1 δ, Ex,y h bh(x, y) h (x, y) 2 2 i O This means that Ex DEa πUnif Ey|x,a h bh(x, y) h (x, y) 2 2 i O Further applying Jensen s inequality shows that Ex DEa πUnif Ey|x,a h bh(x, y) h (x, y) 2 i = Ex DEa πUnif Ey|x,a q bh(x, y) h (x, y) 2 2 Ex DEa πUnif Ey|x,a h bh(x, y) h (x, y) 2 2 i (Jensen s inequality) Lemma 4. Define G(v, β, σ) as G(v, β, σ) = 1 σ (v β)1{β v < β + σ} + 1{v β + σ}. (6) Then, for any context x X, action a [K], and feedback y Y generated via context x and the realized reward r(x, a), we have the following two properties with σ 1 2 θ α 1 K α > 0. r(x, a) = ϕ (x, y) G(h a(x, y), θ r(x, a) = ϕ (x, y) = G(h a(x, y), θ α σ, σ) if a = π (x), Proof. To prove the first property, we show ϕ (x, y) = 1 if h a(x, y) θ α σ. Specifically, if h a(x, y) θ α σ and ϕ (x, y) = 0, then we know that θ α σ h a(x, y) = 1 f (x, a) K PK i=1 f (x, i) 1 K α, leading to a contradiction according to the definition of σ. Therefore, when h a(x, y) θ α σ, we have ϕ (x, y) = 1, which is surely an upper bound of G(h a(x, y), θ α σ, σ) since G(h a(x, y), θ α σ, σ) 1. When h a(x, y) < θ α σ, by definition, we know that G(h a(x, y), θ α σ, σ) = 0 ϕ (x, y). Combining the two cases finishes the proof for the first property. To prove the second property, note that when ϕ (x, y) = 1 and a = π (x), we have h π (x)(x, y) = f (x, π (x)) PK i=1 f (x, i) θ where the last inequality is by Assumption 3. This means that G(h a(x, y), θ α σ, σ) = 1 = ϕ (x, y). When ϕ (x, y) = 0, we have h π (x)(x, y) = 1 f (x, π (x)) K PK i=1 f (x, i) 1 K α θ meaning that G(h a(x, y), θ α σ, σ) = 0. This finishes the proof. B Proofs in Section 3.2 For any policy π and reward estimator rσ, we define the value function V (π, rσ) and the empirical value function b V (π, rσ) as follows: b V (π, rσ) = 1 i=N+1 K π(ai|xi) rσ(xi, yi, ai), (10) V (π, rσ) = Ex D,a πUnif,y|x,a [K π(a|x) rσ(x, y, a)] , (11) where {(xi, ai, yi)}2N i=N+1 is collected with the uniform policy πUnif. The true value function V (π) is defined as: V (π) = Ex D a=1 π(a|x) f (x, a) Using Lemma 4, we can establish the equivalence between V (π ) and V (π , r σ) where r σ(x, y, a) = G(h a(x, y), θ V (π ) = Ex D,a πUnif [K π (a|x) f (x, a)] = Ex D,a πUnif,y|x,a [K π (a|x) ϕ (x, y)] (Assumption 2) = Ex D,a πUnif,y|x,a K π (a|x) G h a(x, y), θ α σ, σ (Lemma 4) = V (π , r σ). (12) In the following lemma, we prove that bπ is near optimal under the true reward. Lemma 5. Under Assumption 2 and Assumption 3, the following holds with probability at least 1 3δ: V (π ) V (bπ) O where bπ is defined in Algorithm 1. Proof. We first introduce some high probability events that the following analysis is based on. First, according to Hoeffding s inequality together with a union bound over π Π (note that |Π| = |F| |H|), with probability at least 1 δ, we have for any π Π V (π, r σ) b V (π, r σ) O In addition, according to Hoeffding s inequality with a union bound over h H, with probability at least 1 δ, we have for any h H, the following inequality holds: 1 N i=N+1 h(xi, yi) h (xi, yi) 2 Ex D,a πUnif,y|x,a [ h(x, y) h (x, y) 2] Combining Eq. (14) with Eq. (5) in Lemma 3, we know that with probability at least 1 2δ, bhai(xi, yi) h ai(xi, yi) 1 i=N+1 bh(xi, yi) h (xi, yi) 2 The following analysis is based on the condition that Eq. (13) and Eq. (15) hold. Bounding b V (π, brσ) b V (π, r σ) . We first show that for any policy π Π, the prediction from brσ is close to r σ in terms of the empirical value function defined in Eq. (10), where brσ(x, y, a) G(bha(x, y), θ b V (π, brσ) b V (π, r σ) 1 i=N+1 K π(ai|xi) |brσ(xi, yi, ai) r σ(xi, yi, ai)| i=N+1 |brσ(xi, yi, ai) r σ(xi, yi, ai)| G(bhai(xi, yi), θ α σ, σ) G(h ai(xi, yi), θ bhai(xi, yi) h ai(xi, yi) i=N+1 bh(xi, yi) h (xi, yi) 2 The third inequality is because G(v, β, σ) is 1 σ-Lipschitz in v and the last inequality is from Eq. (15). Lower bound V (π) by b V (π, brσ). Then, we show that for any policy π Π, V (π) is lower bounded by b V (π, brσ): V (π) = Ex D a=1 π(a|x)f (x, a) = Ex D,a πUnif [Kπ(a|x)f (x, a)] = Ex D,a πUnif Ey|x,a [Kπ(a|x)ϕ (x, y)] (Assumption 2) Ex D,a πUnif Ey|x,a Kπ(a|x)G(h a(x, y), θ = V (π, r σ) (Eq. (11)) = V (π, r σ) b V (π, r σ) + b V (π, r σ) b V (π, brσ) + b V (π, brσ) b V (π, brσ) O The first inequality is from Lemma 4 and the second inequality is from Eq. (13) and Eq. (16). Recall that bπ = maxπ Π b V (π, brσ), we then know that with probability at least 1 3δ, V (π ) V (bπ) V (π ) b V (bπ, brσ) + O V (π ) b V (π , brσ) + O (optimality of bπ on b V (π, brσ)) V (π ) b V (π , r σ) + O V (π ) V (π , r σ) + O , (Eq. (12)) which finishes the proof. Next, we prove Theorem 1. Theorem 1. Under Assumptions 1-3, Algorithm 1 with N = T 2/3K2/3σ 2/3 log1/3(|H|T) guaran- tees that Reg CB O T 2/3K2/3σ 2/3 log1/3(|H|T) . Proof. We bound Reg CB as follows: t=1 f (xt, π (xt)) t=1 f (xt, at) t=2N+1 f (xt, π (xt)) t=2N+1 f (xt, at) t=2N+1 f (xt, π (xt)) t=2N+1 f (xt, bπ(xt)) 2N + T E [V (π ) V (bπ)] The last step is from Lemma 5 with δ = 1 T . Picking N = T 2/3K2/3σ 2/3 log1/3(|H|T) finishes the proof. C Proofs in Section 3.3 Theorem 2. Under Assumptions 1-5, Algorithm 2 with certain choice of N and γ guarantees that Reg CB = O p KTReg Sq + σ 2/3(KT)2/3 log1/3(|H|T) . Proof. According to Eq. (9), we know that with probability at least 1 δ, a=1 Ex DEy|x,a bha(x, y) h a(x, y) 2 K Ex D,a πUnif,y|x,a bha(x, y) h a(x, y) 2 K Ex D,a πUnif,y|x,a h bh(x, y) h (x, y) 2 2 i where the last inequality uses Eq. (9). Let a t = argmaxa f (xt, a) and recall that f (x, a) := Ey|x,a G(h a(x, y), θ α σ, σ) F. We have t=1 f (xt, a t ) t=1 f (xt, at) t=1 Ey|xt,a t [ϕ (xt, y)] t=1 Ey |xt,at[ϕ (xt, y )] (From Assumption 2) t=1 f (xt, a t ) t=1 f (xt, at) (From Lemma 4) t=1 ( bft(xt, at) f (xt, at))2 # The last step is from Foster and Rakhlin [2020, Lemma 3]. Recall that brσ(xt, yt, at) = G(bhat(xt, yt), θ α σ, σ). Direct calculation shows that t=1 ( bft(xt, at) brσ(xt, yt, at))2 t=1 (f (xt, at) brσ(xt, yt, at))2 # t=1 ( bft(at) f (xt, at))( bft(xt, at) + f (xt, at) 2brσ(xt, yt, at)) t=1 ( bft(xt, at) f (xt, at))2 # t=1 ( bft(xt, at) f (xt, at))(f (xt, at) Eyt[brσ(xt, yt, at)]) t=1 ( bft(xt, at) f (xt, at))2 # t=1 (f (xt, at) Eyt[brσ(xt, yt, at)])2 # where the last step is by AM-GM inequality. For the second term, we know that f (xt, at) Eyt[brσ(xt, yt, at)] 2 # G h at(xt, yt), θ α σ, σ G bhat(xt, yt), θ t=1 Eyt|xt,at " G h at(xt, yt), θ α σ, σ G bhat(xt, yt), θ a=1 Ey|xt,a " G h a(xt, y), θ α σ, σ G bha(xt, y), θ a=1 Ey|xt,a h a(xt, y) bha(xt, y) 2 # O TK log(|H|T) where the first inequality is from Jensen s inequality; the third inequality is from that G(v, β, σ) is 1 σ-Lipschitz in v; and the last inequality is by Eq. (18) with δ = 1 T . Combining Eqs. (19)-(21), we obtain that Reg CB O γReg Sq + TKγ log(|H|T) Picking N = 1 TKγ log(|H|T) and γ = min nq KT/Reg Sq, σ 2/3(KT)2/3 log 1/3(|H|T) o , we obtain that KTReg Sq + σ 2/3(KT)2/3 log1/3(|H|T) , which finishes the proof. D Omitted Details in Section 4 D.1 Prompt Used in Generating Dataset The prompt we use in answer generation basically the question itself shown as follows. We replace question by x in our experiment. The prompt we use in feedback generation as follows. We replace question and answer by x and a respectively in our experiments. We replace mood by satisfied ( not satisfied ) when the answer is generated by Qwen1.5-32B-Chat (Qwen1.5-0.5B-Chat). System: You are a user simulator. A user has been presented a Question and an Answer. Simulate the user s next statement. The user is {mood} with the Answer to the Question. Question: {question} Answer:{answer} D.2 Prompt Used in Learning Inverse Kinematic Model The prompt we use in learning the inverse kinematic model is as follows. We replace question , answer , and feedback by x, a, and y respectively in our experiments. System: You are a conversation evaluating agent. Given a User s Question, an Answer, and the User s Feedback: determine if the User s Feedback is consistent with Answer. Respond with Yes or No only. User s Question: {question} Answer:{answer} User s Feedback:{feedback} Respond with Yes or No only. D.3 Prompt Used in Learning Policy The prompt we use in learning the reward predictor is as follows. We replace question and answer by x and a respectively in our experiments. System: You are an Answer evaluating agent. Given a User s Question and an Answer: assess if the Answer is good. Respond with Yes or No only. User s Question:{question} Answer:{answer} Respond with Yes or No only. Neur IPS Paper Checklist Question: Do the main claims made in the abstract and introduction accurately reflect the paper s contributions and scope? Answer: [Yes] Justification: See abstract and Section 1. Guidelines: The answer NA means that the abstract and introduction do not include the claims made in the paper. The abstract and/or introduction should clearly state the claims made, including the contributions made in the paper and important assumptions and limitations. A No or NA answer to this question will not be perceived well by the reviewers. The claims made should match theoretical and experimental results, and reflect how much the results can be expected to generalize to other settings. It is fine to include aspirational goals as motivation as long as it is clear that these goals are not attained by the paper. 2. Limitations Question: Does the paper discuss the limitations of the work performed by the authors? Answer: [Yes] Justification: See Section 1 and Section 2. Guidelines: The answer NA means that the paper has no limitation while the answer No means that the paper has limitations, but those are not discussed in the paper. The authors are encouraged to create a separate "Limitations" section in their paper. The paper should point out any strong assumptions and how robust the results are to violations of these assumptions (e.g., independence assumptions, noiseless settings, model well-specification, asymptotic approximations only holding locally). The authors should reflect on how these assumptions might be violated in practice and what the implications would be. The authors should reflect on the scope of the claims made, e.g., if the approach was only tested on a few datasets or with a few runs. In general, empirical results often depend on implicit assumptions, which should be articulated. The authors should reflect on the factors that influence the performance of the approach. For example, a facial recognition algorithm may perform poorly when image resolution is low or images are taken in low lighting. Or a speech-to-text system might not be used reliably to provide closed captions for online lectures because it fails to handle technical jargon. The authors should discuss the computational efficiency of the proposed algorithms and how they scale with dataset size. If applicable, the authors should discuss possible limitations of their approach to address problems of privacy and fairness. While the authors might fear that complete honesty about limitations might be used by reviewers as grounds for rejection, a worse outcome might be that reviewers discover limitations that aren t acknowledged in the paper. The authors should use their best judgment and recognize that individual actions in favor of transparency play an important role in developing norms that preserve the integrity of the community. Reviewers will be specifically instructed to not penalize honesty concerning limitations. 3. Theory Assumptions and Proofs Question: For each theoretical result, does the paper provide the full set of assumptions and a complete (and correct) proof? Answer: [Yes] Justification: See Assumption 1, Assumption 2, Assumption 3, Assumption 4, Assumption 5, and the appendix. Guidelines: The answer NA means that the paper does not include theoretical results. All the theorems, formulas, and proofs in the paper should be numbered and crossreferenced. All assumptions should be clearly stated or referenced in the statement of any theorems. The proofs can either appear in the main paper or the supplemental material, but if they appear in the supplemental material, the authors are encouraged to provide a short proof sketch to provide intuition. Inversely, any informal proof provided in the core of the paper should be complemented by formal proofs provided in appendix or supplemental material. Theorems and Lemmas that the proof relies upon should be properly referenced. 4. Experimental Result Reproducibility Question: Does the paper fully disclose all the information needed to reproduce the main experimental results of the paper to the extent that it affects the main claims and/or conclusions of the paper (regardless of whether the code and data are provided or not)? Answer: [Yes] Justification: See Algorithm 1, Algorithm 2, and detailed descriptions in Section 4 and Appendix D. Guidelines: The answer NA means that the paper does not include experiments. If the paper includes experiments, a No answer to this question will not be perceived well by the reviewers: Making the paper reproducible is important, regardless of whether the code and data are provided or not. If the contribution is a dataset and/or model, the authors should describe the steps taken to make their results reproducible or verifiable. Depending on the contribution, reproducibility can be accomplished in various ways. For example, if the contribution is a novel architecture, describing the architecture fully might suffice, or if the contribution is a specific model and empirical evaluation, it may be necessary to either make it possible for others to replicate the model with the same dataset, or provide access to the model. In general. releasing code and data is often one good way to accomplish this, but reproducibility can also be provided via detailed instructions for how to replicate the results, access to a hosted model (e.g., in the case of a large language model), releasing of a model checkpoint, or other means that are appropriate to the research performed. While Neur IPS does not require releasing code, the conference does require all submissions to provide some reasonable avenue for reproducibility, which may depend on the nature of the contribution. For example (a) If the contribution is primarily a new algorithm, the paper should make it clear how to reproduce that algorithm. (b) If the contribution is primarily a new model architecture, the paper should describe the architecture clearly and fully. (c) If the contribution is a new model (e.g., a large language model), then there should either be a way to access this model for reproducing the results or a way to reproduce the model (e.g., with an open-source dataset or instructions for how to construct the dataset). (d) We recognize that reproducibility may be tricky in some cases, in which case authors are welcome to describe the particular way they provide for reproducibility. In the case of closed-source models, it may be that access to the model is limited in some way (e.g., to registered users), but it should be possible for other researchers to have some path to reproducing or verifying the results. 5. Open access to data and code Question: Does the paper provide open access to the data and code, with sufficient instructions to faithfully reproduce the main experimental results, as described in supplemental material? Answer: [Yes] Justification: See supplementary materials. Guidelines: The answer NA means that paper does not include experiments requiring code. Please see the Neur IPS code and data submission guidelines (https://nips.cc/ public/guides/Code Submission Policy) for more details. While we encourage the release of code and data, we understand that this might not be possible, so No is an acceptable answer. Papers cannot be rejected simply for not including code, unless this is central to the contribution (e.g., for a new open-source benchmark). The instructions should contain the exact command and environment needed to run to reproduce the results. See the Neur IPS code and data submission guidelines (https: //nips.cc/public/guides/Code Submission Policy) for more details. The authors should provide instructions on data access and preparation, including how to access the raw data, preprocessed data, intermediate data, and generated data, etc. The authors should provide scripts to reproduce all experimental results for the new proposed method and baselines. If only a subset of experiments are reproducible, they should state which ones are omitted from the script and why. At submission time, to preserve anonymity, the authors should release anonymized versions (if applicable). Providing as much information as possible in supplemental material (appended to the paper) is recommended, but including URLs to data and code is permitted. 6. Experimental Setting/Details Question: Does the paper specify all the training and test details (e.g., data splits, hyperparameters, how they were chosen, type of optimizer, etc.) necessary to understand the results? Answer: [Yes] Justification: See Section 4. Guidelines: The answer NA means that the paper does not include experiments. The experimental setting should be presented in the core of the paper to a level of detail that is necessary to appreciate the results and make sense of them. The full details can be provided either with the code, in appendix, or as supplemental material. 7. Experiment Statistical Significance Question: Does the paper report error bars suitably and correctly defined or other appropriate information about the statistical significance of the experiments? Answer: [Yes] Justification: See Section 4. Guidelines: The answer NA means that the paper does not include experiments. The authors should answer "Yes" if the results are accompanied by error bars, confidence intervals, or statistical significance tests, at least for the experiments that support the main claims of the paper. The factors of variability that the error bars are capturing should be clearly stated (for example, train/test split, initialization, random drawing of some parameter, or overall run with given experimental conditions). The method for calculating the error bars should be explained (closed form formula, call to a library function, bootstrap, etc.) The assumptions made should be given (e.g., Normally distributed errors). It should be clear whether the error bar is the standard deviation or the standard error of the mean. It is OK to report 1-sigma error bars, but one should state it. The authors should preferably report a 2-sigma error bar than state that they have a 96% CI, if the hypothesis of Normality of errors is not verified. For asymmetric distributions, the authors should be careful not to show in tables or figures symmetric error bars that would yield results that are out of range (e.g. negative error rates). If error bars are reported in tables or plots, The authors should explain in the text how they were calculated and reference the corresponding figures or tables in the text. 8. Experiments Compute Resources Question: For each experiment, does the paper provide sufficient information on the computer resources (type of compute workers, memory, time of execution) needed to reproduce the experiments? Answer: [Yes] Justification: See Section 4. Guidelines: The answer NA means that the paper does not include experiments. The paper should indicate the type of compute workers CPU or GPU, internal cluster, or cloud provider, including relevant memory and storage. The paper should provide the amount of compute required for each of the individual experimental runs as well as estimate the total compute. The paper should disclose whether the full research project required more compute than the experiments reported in the paper (e.g., preliminary or failed experiments that didn t make it into the paper). 9. Code Of Ethics Question: Does the research conducted in the paper conform, in every respect, with the Neur IPS Code of Ethics https://neurips.cc/public/Ethics Guidelines? Answer: [Yes] Justification: The authors have reviewed the Neur IPS Code of Ethics. The research conducted in this paper conforms with it in every respect. Guidelines: The answer NA means that the authors have not reviewed the Neur IPS Code of Ethics. If the authors answer No, they should explain the special circumstances that require a deviation from the Code of Ethics. The authors should make sure to preserve anonymity (e.g., if there is a special consideration due to laws or regulations in their jurisdiction). 10. Broader Impacts Question: Does the paper discuss both potential positive societal impacts and negative societal impacts of the work performed? Answer: [NA] Justification: This work is mostly theoretical, and we do not foresee any negative ethical or societal outcomes. Guidelines: The answer NA means that there is no societal impact of the work performed. If the authors answer NA or No, they should explain why their work has no societal impact or why the paper does not address societal impact. Examples of negative societal impacts include potential malicious or unintended uses (e.g., disinformation, generating fake profiles, surveillance), fairness considerations (e.g., deployment of technologies that could make decisions that unfairly impact specific groups), privacy considerations, and security considerations. The conference expects that many papers will be foundational research and not tied to particular applications, let alone deployments. However, if there is a direct path to any negative applications, the authors should point it out. For example, it is legitimate to point out that an improvement in the quality of generative models could be used to generate deepfakes for disinformation. On the other hand, it is not needed to point out that a generic algorithm for optimizing neural networks could enable people to train models that generate Deepfakes faster. The authors should consider possible harms that could arise when the technology is being used as intended and functioning correctly, harms that could arise when the technology is being used as intended but gives incorrect results, and harms following from (intentional or unintentional) misuse of the technology. If there are negative societal impacts, the authors could also discuss possible mitigation strategies (e.g., gated release of models, providing defenses in addition to attacks, mechanisms for monitoring misuse, mechanisms to monitor how a system learns from feedback over time, improving the efficiency and accessibility of ML). 11. Safeguards Question: Does the paper describe safeguards that have been put in place for responsible release of data or models that have a high risk for misuse (e.g., pretrained language models, image generators, or scraped datasets)? Answer: [NA] Justification: This paper poses no such risks. Guidelines: The answer NA means that the paper poses no such risks. Released models that have a high risk for misuse or dual-use should be released with necessary safeguards to allow for controlled use of the model, for example by requiring that users adhere to usage guidelines or restrictions to access the model or implementing safety filters. Datasets that have been scraped from the Internet could pose safety risks. The authors should describe how they avoided releasing unsafe images. We recognize that providing effective safeguards is challenging, and many papers do not require this, but we encourage authors to take this into account and make a best faith effort. 12. Licenses for existing assets Question: Are the creators or original owners of assets (e.g., code, data, models), used in the paper, properly credited and are the license and terms of use explicitly mentioned and properly respected? Answer: [Yes] Justification: We properly cite the used dataset and models and respect the licenses. Guidelines: The answer NA means that the paper does not use existing assets. The authors should cite the original paper that produced the code package or dataset. The authors should state which version of the asset is used and, if possible, include a URL. The name of the license (e.g., CC-BY 4.0) should be included for each asset. For scraped data from a particular source (e.g., website), the copyright and terms of service of that source should be provided. If assets are released, the license, copyright information, and terms of use in the package should be provided. For popular datasets, paperswithcode.com/datasets has curated licenses for some datasets. Their licensing guide can help determine the license of a dataset. For existing datasets that are re-packaged, both the original license and the license of the derived asset (if it has changed) should be provided. If this information is not available online, the authors are encouraged to reach out to the asset s creators. 13. New Assets Question: Are new assets introduced in the paper well documented and is the documentation provided alongside the assets? Answer: [NA] Justification: The paper does not release new assets. Guidelines: The answer NA means that the paper does not release new assets. Researchers should communicate the details of the dataset/code/model as part of their submissions via structured templates. This includes details about training, license, limitations, etc. The paper should discuss whether and how consent was obtained from people whose asset is used. At submission time, remember to anonymize your assets (if applicable). You can either create an anonymized URL or include an anonymized zip file. 14. Crowdsourcing and Research with Human Subjects Question: For crowdsourcing experiments and research with human subjects, does the paper include the full text of instructions given to participants and screenshots, if applicable, as well as details about compensation (if any)? Answer: [NA] Justification: The paper does not involve crowdsourcing nor research with human subjects. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Including this information in the supplemental material is fine, but if the main contribution of the paper involves human subjects, then as much detail as possible should be included in the main paper. According to the Neur IPS Code of Ethics, workers involved in data collection, curation, or other labor should be paid at least the minimum wage in the country of the data collector. 15. Institutional Review Board (IRB) Approvals or Equivalent for Research with Human Subjects Question: Does the paper describe potential risks incurred by study participants, whether such risks were disclosed to the subjects, and whether Institutional Review Board (IRB) approvals (or an equivalent approval/review based on the requirements of your country or institution) were obtained? Answer: [NA] Justification: The paper does not involve crowdsourcing nor research with human subjects. Guidelines: The answer NA means that the paper does not involve crowdsourcing nor research with human subjects. Depending on the country in which research is conducted, IRB approval (or equivalent) may be required for any human subjects research. If you obtained IRB approval, you should clearly state this in the paper. We recognize that the procedures for this may vary significantly between institutions and locations, and we expect authors to adhere to the Neur IPS Code of Ethics and the guidelines for their institution. For initial submissions, do not include any information that would break anonymity (if applicable), such as the institution conducting the review.