# satisficing_regret_minimization_in_bandits__8ba6b6b9.pdf Published as a conference paper at ICLR 2025 SATISFICING REGRET MINIMIZATION IN BANDITS Qing Feng Cornell University qf48@cornell.edu Tianyi Ma Shanghai Jiao Tong University sean ma@sjtu.edu.cn Ruihao Zhu Cornell University ruihao.zhu@cornell.edu Motivated by the concept of satisficing in decision-making, we consider the problem of satisficing regret minimization in bandit optimization. In this setting, the learner aims at selecting satisficing arms (arms with mean reward exceeding a certain threshold value) as frequently as possible. The performance is measured by satisficing regret, which is the cumulative deficit of the chosen arm s mean reward compared to the threshold. We propose SELECT, a general algorithmic template for Satisficing REgret Minimization via Samp Ling and Low Er Confidence bound Testing, that attains constant satisficing regret for a wide variety of bandit optimization problems in the realizable case (i.e., a satisficing arm exists). Specifically, given a class of bandit optimization problems and a corresponding learning oracle with sub-linear (standard) regret upper bound, SELECT iteratively makes use of the oracle to identify a potential satisficing arm with low regret. Then, it collects data samples from this arm, and continuously compares the LCB of the identified arm s mean reward against the threshold value to determine if it is a satisficing arm. As a complement, SELECT also enjoys the same (standard) regret guarantee as the oracle in the non-realizable case. Finally, we conduct numerical experiments to validate the performance of SELECT for several popular bandit optimization settings. 1 INTRODUCTION Multi-armed bandit (MAB) is a classic framework for decision-making under uncertainty. In MAB, a learner is given a set of arms with initially unknown mean reward. She repeatedly picks an arm to collect its noisy reward and her goal is to maximize the cumulative reward. The performance is usually measured by the notion of regret, which is the difference between the maximum possible total mean reward and the total mean reward collected by the learner. With its elegant form, MAB has found numerous applications in recommendation systems, ads display, and beyond, and many (near) optimal algorithms have been developed for various bandit optimization settings (Lattimore & Szepesv ari, 2020). While optimal decision is desired in certain high-stakes situations, it is perhaps not surprising that what we often need is just a good enough or satisficing alternative (Lufkin, 2021). Coined by Nobel laureate Herbert Simon (Simon, 1956), the term satisficing refers to the decision-making strategy that settles for an acceptable solution rather than exhaustively hunting for the best possible. The concept of satisficing finds its place in many real world decision-making under uncertainty settings (Caplin et al., 2011). For example, in inventory management, studies show that product stockouts can result in customer disengagement and impair a firm s profitability (Fitzsimons, 2000; Jing & Lewis, 2011). As such, retailers usually replenish their products periodically to improve service level (i.e., probability of no stockout for any product). Notably, due to various constraints (e.g., capacity and inventory cost), instead of blindly maximizing service level, retailers typically set a target service level (e.g., the ideal service level in retailing is 90% 95% (Levy et al., 2023)). Consequently, in face of the unknown demand distributions, one can cast the problem into a MAB setting with a different objective. In this setting, a retailer starts without complete knowledge of the demand distributions and proceeds by trying various stock level combinations (i.e., arms) of the products over time. Here, the objective is to hit the target service level as frequently as possible rather than maximizing the service level every time. As another example, in fashion product design, designers usually need to design a sequence of products for a group of customers with initially Published as a conference paper at ICLR 2025 unknown preferences. Instead of demanding for the best design every time, designers only need to pick a design that can hopefully meet the expectation of a certain portion of the customers for most of the time (Mabey, 2022; Whitenton, 2014). The above examples give rise to the study of satisficing bandits (Tamatsukuri & Takahashi, 2019; Michel et al., 2023) where the objective is to hit a satisficing arm (i.e., an arm whose mean reward exceeds a certain threshold value) as frequently as possible. The corresponding performance measure is satisficing regret, which is the cumulative deficit of the chosen arm s mean reward compared to the threshold. In particular, Michel et al. (2023) develops SAT-UCB, an algorithm for finitearmed satisficing bandits: In the realizable case (i.e., a satisficing arm exists), the algorithm attains a constant satisficing regret, i.e., independent of the length of the time horizon; Otherwise, in the non-realizable case (i.e., the threshold value is out of reach), it preserves the optimal logarithmic (standard) regret. Despite these, existing algorithms usually impose that there is a clear separation between the threshold value and the largest mean reward among all non-satisficing arms, which is referred to as the satisficing gap. Moreover, the satisficing regret bounds usually scale inverse proportionally to the satisficing gap. On the other hand, it is evident that in many bandit optimization problems (e.g., Lipschitz bandits and concave bandits), the satisficing gap can simply be 0, making the bounds vacuous. Therefore, it is not immediately clear if SAT-UCB (Michel et al., 2023) or other prior results can go beyond finite-armed bandits and if constant satisficing regret is possible for bandit optimization in general. Main Contributions: In this work, we propose SELECT, a novel algorithmic template for Satisficing REgret Minimization via Samp Ling and Low Er Confidence Bound Testing. More specifically: We describe the design of SELECT in Section 3. For a given bandit optimization problem class and an associated learning oracle with sub-linear (but not necessarily optimal) standard regret guarantee, SELECT runs in rounds with the help of the oracle. At the beginning of a round, SELECT first runs the oracle for a number of time steps and randomly samples an arm from its trajectory as a candidate satisficing arm. Then, it conducts forced sampling by pulling this arm for a pre-specified number of times. Finally, it continues to pull this arm and starts to monitor the resulted lower confidence bound (LCB) of its mean reward. SELECT terminates the current round and starts the next once the LCB falls below the threshold value; In Section 4, we establish that in the realizable case (i.e., when a satisficing arm exists), SELECT is able to achieve a constant satisficing regret. Notably, the satisficing regret bound has no dependence on the satisficing gap (see the forthcoming Remark 2 for a detailed discussion). Instead, it scales inverse proportionally to the exceeding gap, which measures the difference between the optimal reward and the threshold value, and is positive in general; Otherwise (i.e., when no satisficing arm exists), SELECT enjoys the same standard regret bound as the learning oracle. In Section 5, we instantiate SELECT to finite-armed bandits, concave bandits, and Lipschitz bandits, and demonstrate the corresponding satisficing and standard regret bounds. We also provide some discussions on the respective lower bounds; Finally, in Section 6 we conduct numerical experiments to demonstrate the performance of SELECT in fnite-armed bandits, concave bandits, and Lipschitz bandits. We use SAT-UCB, SATUCB+ (a heuristic with no regret guarantee proposed by Michel et al. (2023)), and the respective learning oracle as the benchmarks. Our results reveal that in the realizable case, SELECT does achieve constant satisficing regret. Moreover, both its satisficing regrets (in the realizable cases) and standard regrets (in the non-realizable cases) either significantly outperform the benchmarks or are close to the best-performing ones . 1.1 RELATED WORKS Besides Michel et al. (2023), the most relevant works to ours are Bubeck et al. (2013); Garivier et al. (2019); Tamatsukuri & Takahashi (2019), all of which focus on finite-armed bandits. Bubeck et al. (2013) and Garivier et al. (2019) introduce constant regret algorithms with knowledge of the optimal reward. This can be viewed as a special case of satisficing where the threshold value is exactly the optimal mean reward. H uy uk & Tekin (2021) further extend the algorithm in Garivier et al. (2019) Published as a conference paper at ICLR 2025 to multi-objective multi-armed bandit settings. Tamatsukuri & Takahashi (2019) also establishes constant regret for finite-armed satisficing bandits when there exists exactly one satisficing arm. As a follow-up paper of Michel et al. (2023), Hajiabolhassan & Ortner (2023) considers satisficing in finite-state MDP. Among others, Russo & Van Roy (2022) is one of the first to introduce the notion of satisficing regret, but they focus on a time-discounted setting. Furthermore, in Russo & Van Roy (2022), the learner is assumed to know the difference between the values of the threshold and the optimal reward. More broadly, the concept of satisficing has also been studied by Reverdy & Leonard (2014); Reverdy et al. (2017); Abernethy et al. (2016) in bandit learning. Both Reverdy & Leonard (2014) and Abernethy et al. (2016) focus on maximizing the number of times that the selected arm s realized reward exceeds a threshold value. This is equivalent to a MAB problem that sets each arm s mean reward as its probability of generating a value that exceeds the threshold. In view of this, the problems investigated in Reverdy & Leonard (2014) and Abernethy et al. (2016) are closer to conventional MABs. They are also different from Reverdy et al. (2017); Michel et al. (2023) and ours, which use the mean reward of the selected arms to compute the satisficing regret. We note that in Reverdy et al. (2017), a more general Bayesian setting is studied. In this setting, the satisficing regret also takes the learner s belief that some arm is satisficing into account. The notion of satisficing is also studied in the pure exploration setting (Locatelli et al., 2016; Mukherjee et al., 2017; Kano et al., 2019) as thresholding bandits. Their objective is to identify all arms with mean reward above a certain satisficing level with limited exploration budget. Unlike the satisficing regret we consider, the performance of their algorithm is measured by simple regret, i.e., the expected number of misidentification made by the final output. Note that any pure exploration algorithm (Audibert & Bubeck, 2010) is able to identify the optimal arm or a satisficing arm with high probability after a number of steps. However, due to a small but positive error probability, subsequent exploitation of the identified arm will always incur linear satisficing regret, thus a simple explore-then-exploit approach is unable to obtain a constant satisficing regret bound. Similar observations have also been made in Michel et al. (2023). 2 PROBLEM FORMULATION In this section, we present the setup of our problem. Consider a class of bandit optimization problem (X, R), where X is the arm set and R {r : X R} is the class of admissible reward functions. The learner is given a set of feasible arms X and a class of admissible reward functions R, pulling arm X X in a time step brings a random reward with mean r(X). The learner knows the underlying reward function r( ) belongs to the function class R, but the exact r( ) is initially unknown to the learner. In each time step t = 1, 2, . . . , T, the learner pulls an arm Xt X, and then observes and receives a noisy reward Yt = r(Xt) + ϵt, where ϵt is a conditionally 1-subgaussian random variable, i.e., E[ϵt|X1, ϵ1 . . . , Xt 1, ϵt 1, Xt] = 0 and E[eλϵt|X1, ϵ1 . . . , Xt 1, ϵt 1, Xt] eλ2/2 holds for all λ R. Remark 1. The notion of problem class (X, R) captures most of the bandit optimization problems in the stationary setting. For example, if X = [K] and R = {r : [K] R}, then the problem class (X, R) corresponds to finite-armed bandits with K arms; if X is any convex set in Rd and R = {r : X R, r is concave and 1-Lipschitz}, then the problem class (X, R) corresponds to concave bandits. In this work, we define two notions of regret. First, with the presence of satisficing level S, we consider the satisficing regret introduced in Reverdy et al. (2017); Michel et al. (2023), which is the expected cumulative deficit of the chosen arm s mean reward when compared to S, i.e. Regret S = E t=1 max{S r(Xt), 0} Because it is possible that no arm achieves the satisficing level, we also follow the conventional bandit literature to define the (standard) regret, which measures the expected cumulative difference of Published as a conference paper at ICLR 2025 the mean reward between the optimal arm and the chosen arm s, i.e., let X = arg max X X r(X), the regret is defined as Regret = T r(X ) E We say that the problem is realizable if r(X ) S; Otherwise, it is non-realizable. We recall that an arm X is satisficing if r(X) S; Otherwise, it is non-satisficing. Existing works on similar settings (e.g., Michel et al. 2023) usually derive satisficing regret bound that scales inverse proportional to the satisficing gap S = min{S r(X) : r(X) < S}. However, for bandit optimization problems with large and even infinitely many arms, S can approach 0 quickly. To address this issue, we define the notion of exceeding gap that captures the difference between the optimal reward and S, i.e., S = r(X ) S. 3 SELECT: SATISFICING REGRET MINIMIZATION VIA SAMPLING AND LOWER CONFIDENCE BOUND TESTING In this section, we present our algorithmic template SELECT for general bandit problems. When the arm space X is not too large, one can collect enough data samples for every arm to estimate its mean reward and gradually abandon any non-satisficing arms (see, e.g., Garivier et al. 2019; Michel et al. 2023). In general, however, the arm space can be large and may even contain infinitely many arms (e.g., continuous arm space). Therefore, it becomes impossible to identify all non-satisficing arms. We thus take an alternative approach: It repeatedly locates a potential satisficing arm with low (satisficing) regret. Then, it tests if this candidate arm is truly a satisficing arm or not. If not, it kicks off the next round of search. To ensure fast detection of candidate satisficing arms with low regret for a given bandit problem class (X, R), SELECT makes use of a bandit algorithm for (standard) regret minimization on (X, R). To this end, we assume access to a blackbox learning oracle that achieves sub-linear standard regret for the bandit problem class (X, R). Condition 1. For problem class (X, R), there exists a sequence of learning algorithms ALG such that for any reward function r R and any time horizon t 2, the regret of ALG(t) is bounded by C1tα log(t)β. where C1 1, 1/2 α < 1, β 0 are constants only dependent on the problem class and independent from the time horizon t. We remark that Condition 1 only asks for sub-linearity in standard regret upper bounds rather than optimality. Therefore, it is an extremely mild condition and can be easily satisfied by a wide range of bandit optimization problems and the corresponding bandit learning oracles, e.g., finite-armed bandits, Lipschitz bandits, etc. We will demonstrate this in the forthcoming Section 5 and Appendix D of the full version (Feng et al., 2025). With this, for a problem class (X, R) and its sub-linear regret learning algorithm ALG, SELECT runs in rounds. In round i, it takes the following steps (see Fig. 1 for a illustration): Step 1. Identify a Potential Satisficing Arm: Let γi = 2 i(1 α)/α, SELECT follows ALG(ti) for the first ti = γ 1/(1 α) i = 2i/α steps of round i and records its arm selections. Then, it samples an arm ˆXi uniformly random from the trajectory of ALG(ti) in this round. By virtue of ALG, we find an arm whose mean reward is at most O(tα 1 i ) = O(γi) below r(X ) in expectation, and only O(tα i ) satisficing regret is incurred in this step in the realizable case. We note that in the realizable case, as i increases, O(γi) will gradually become smaller than S, meaning that the sampled arm ˆXi is more likely to be a satisficing arm and enjoys no satisficing regret; Published as a conference paper at ICLR 2025 Figure 1: Illustration of a single round of SELECT Step 2. Forced Sampling on the Identified Arm: To validate if ˆXi is a satisficing arm or not, we need to collect enough data from pulling ˆXi before performing any statistical test (see the forthcoming Remark 2 for the reasons). However, when the arm set X is large, ALG(ti) might only have pulled ˆXi for limited number of times. To circumvent this, we perform a forced sampling on ˆXi by pulling it for Ti = γ 2 i times; Step 3. Lower Confidence Bound Tests: In this step, SELECT kicks off the LCB test as more data is collected from pulling ˆXi. We use k (initialized to 0) to denote the number of additional pieces of noisy reward data generated by pulling ˆXi in the current time step and ˆrtot i to denote the running total reward collected from Step 2 and the current step so far. At the same time, it keeps comparing the LCB of ˆXi s mean reward, i.e., ˆrtot i /(Ti + k) p 4 log(Ti + k)/(Ti + k), against the threshold value S after acquiring every piece of new data. SELECT terminates this round and enters the next whenever the LCB is less than S. The pseudo-code of SELECT is presented in Algorithm 1. Remark 2. (Key Novelties of SELECT) Our algorithm brings together several key ingredients, including sampling from oracle trajectories (Step 1), forced sampling to collect data (Step 2), and the LCB test (Step 3). In what follows, we further comment on the subtleties of each step. Step 1. Prior works on satisficing (Garivier et al. 2019; Michel et al. 2023) mainly use uniform exploration to find candidates of satisficing arms. This approach could incur major satisficing regret when the arm set is large or even possibly infinite, as conducting uniform exploration on a large arm set may result in pulling a large number of non-satisficing arms. To bypass this challenge, in Step 1, we search for candidates using a bandit learning oracle with sub-linear standard regret. Hence, we are able to find candidates of satisficing arms without incurring large regret. Step 2. One challenge from using the LCB test in Step 3 is that it is a more conservative design compared to prior works. In fact, if we directly enter Step 3 without Step 2, the LCB of r( ˆ Xi) can easily fall below S even if r( ˆXi) S. This is because we may not have pulled ˆXi for enough times in Step 1 and the corresponding confidence interval might be large. As a result, the lower confidence bound of r( ˆXi), which is the difference between empirical mean and confidence radius, can be significantly below S. This indicates, without the forced sampling in Step 2, major satisficing regret can be incurred due to frequent re-start of a new round. With the forced sampled data collected from Step 2, SELECT ensures that the width of the confidence interval is of order O(T 1/2 i ) = O(γi) (i.e., same as E[r(X ) r( ˆXi)]). Consequently, whenever γi shrinks to well below S, SELECT gradually becomes less likely to terminate a round (see the forthcoming Proposition 2); Step 3. In Step 3, SELECT compares the LCB of ˆ Xi s mean reward against S to determine if Xi is a satisficing arm. This deviates from prior works that use UCB (Garivier et al., 2019) or empirical mean (Michel et al., 2023). However, it turns out that this is essential in achieving a Published as a conference paper at ICLR 2025 constant satisficing regret that does not scale with 1/ S, which can quickly explode in many cases. As we will see, once entering Step 3, SELECT can terminate a round within 1 additional time step in expectation when facing a non-satisficing arm. In contrast, if it follows prior works to use UCB or empirical mean instead, the number of time steps required (and hence, the satisficing regret) will unavoidably scale with 1/ S (see e.g., Theorem 9 of Garivier et al. 2019 or Theorem 1 of Michel et al. 2023). Algorithm 1 Satisficing Regret Minimization via Sampling and Lower Confidence Bound Testing (SELECT) Input: time horizon T, satisficing level S γi 2 i(1 α)/α for all i = 1, 2, 3 . . . for round i = 1, 2, . . . do Set ti γ 1/(1 α) i Run ALG(ti), let { Xs}s [ti] be the trajectory of arm pulls Sample R uniformly at random from {1, 2, . . . , ti} and set ˆXi XR Set Ti γ 2 i , k 0 Let t be the current time step, pull arm ˆXi for the next Ti time steps, and set ˆrtot i Pt+Ti 1 s=t Ys while LCB( ˆXi) S do Set k k + 1 Pull arm ˆXi and observe Yt, where t is the current time step Update ˆrtot i ˆrtot i + Yt Set LCB( ˆXi) ˆrtot i Ti + k 4 log(Ti + k) end while end for 4 REGRET ANALYSIS In this section, we analyze the regret bounds of SELECT. We begin by providing an upper bound of satisficing regret in the realizable case, i.e., r(X ) S. Theorem 1. If r(X ) S, then the satisficing regret of SELECT is bounded by Regret S min α 1 α polylog C1 , C1T α polylog(T) We remark that, from the first term on the RHS of equation 1, SELECT can achieve a constant (w.r.t. T) satisficing regret in the realizable case. Moreover, even when T is relatively small and the constant satisficing regret guarantee is loose compared to the oracle s regret bound, it is able to adapt to the oracle s performance. Proof Sketch. A complete proof for Theorem 1 is provided in Appendix A of the full version (Feng et al., 2025). The proof relies on two critical results. In the first one, we show an upper bound on the satisficing regret of round i. Proposition 1. If r(X ) S, then the satisficing regret incurred in round i is bounded by 6C1 (1 α)β γ α/(1 α) i log(2/γi)β. The proof of this proposition is provided in Appendix A.1 of the full version (Feng et al., 2025). We also show that once SELECT runs for enough rounds so that γi = O( S), it is unlikely to start a new round. Published as a conference paper at ICLR 2025 Proposition 2. If r(X ) S, then for every i that satisfies 32C1 (1 α)β γi log(2/γi)max{β,1/2} S , the probability that round i ends within the time horizon T (conditioned on the event that round i is started within the time horizon) is bounded by 1/4. The proof of this proposition is provided in Appendix A.2 of the full version (Feng et al., 2025). Since each round of SELECT runs independently, Proposition 2 indicates that the total number of rounds for SELECT can be upper-bounded by a shifted geometric random variable with success rate 3/4. Combining the results of Proposition 1 and Proposition 2 we are able to establish the statement in Theorem 1. We also show that, in the non-realizable case, SELECT enjoys the same regret bound as the oracle in Condition 1. The proof is provided in Appendix B of the full version (Feng et al., 2025). Theorem 2. If r(X ) < S, the standard regret of SELECT is bounded by Regret C1T α polylog(T). With the above results, we establish that under a very general condition, SELECT does achieve constant satisficing regret in the realizable case without compromising the standard regret guarantee in the non-realizable case. Altogether, these empower SELECT the potential to become a general algorithm for decision-making under uncertainty. In what follows, we instantiate SELECT to several popular bandit optimization settings, including finite-armed bandits, concave bandits, and Lipschitz bandits. Along the way, we showcase how our results enable constant satisficing regret for bandits with large and even infinite arm space, which was not achieved by existing algorithms. We remark that the examples given here are non-exhaustive and similar results can be derived for other bandit optimization settings such as linear bandits. In all three examples of this section, we assume the mean reward of each arm is in [0, 1]. 1. Finite-Armed Bandits: Consider a finite-armed bandit problem with K arms, i.e., X = [K]. In this case, both the UCB algorithm (see, e.g., Bubeck & Cesa-Bianchi 2012) and Thompson sampling (see, e.g., Agrawal & Goyal 2017) achieve a regret bound of O( p KT log(T)). Combining them with Theorems 1 and 2, we have the following corollary. At the end of this section, we also provide some discussions on the lower bounds of the satisficing regrets. Corollary 1. By using either the UCB algorithm or Thompson sampling as ALG, if r(X ) S, the satisficing regret of SELECT is bounded by Regret S min K S polylog K KT polylog(T) ; If r(X ) < S, the standard regret of SELECT is bounded by Regret KT polylog(T). Remark 3 (Comparison with Garivier et al. 2019; Michel et al. 2023). Garivier et al. (2019); Michel et al. (2023) also provide algorithms for satisficing regret for finite-armed bandits. While our satisficing regret bound is incomparable to Garivier et al. (2019), which attains a O(K/ S) satisficing regret, we provide major improvement over Michel et al. (2023), which achieves O(K/ S + K/( S)2) satisficing regret. Compared to Michel et al. (2023), our regret bound remove the dependence on S and the additional 1/ S factor. We also point out that if we want to achieve the same satisficing regret as Garivier et al. (2019), we can simply change the LCB test in Step 3 to a UCB test. 2. Concave Bandits: Consider a concave bandit problem, i.e., X Rd is a bounded convex set, and r(X) is a concave and 1-Lipschitz continuous function defined on X. Agarwal et al. (2011) gives an algorithm that enjoys poly(d) T polylog(T) regret. Together with Theorems 1 and 2, we have the following corollary. Published as a conference paper at ICLR 2025 Corollary 2. By using the algorithm in Agarwal et al. (2011) as ALG, if r(X ) S, then the satisficing regret of SELECT is bounded by Regret S min poly(d) S polylog d T polylog(T) ; If r(X ) < S, the standard regret of SELECT is bounded by Regret poly(d) T polylog(T). 3. Lipschitz Bandits: Consider a Lipschitz bandit problem in d dimensions, i.e., X = [0, 1]d, and r(X) is an L-Lipschitz function (here Lipschitz functions are defined in the sense of - norm). Bubeck et al. (2011) introduces a uniformly discretized UCB algorithm to achieve a O(Ld/(d+2)T (d+1)/(d+2)p log(T)) regret upper bound bound. Together with Theorem 1 and Theorem 2, we have the following corollary. Corollary 3. By using the UCB algorithm with uniform discretization in Bubeck et al. (2011) as ALG, if r(X ) S, then the satisficing regret of SELECT is bounded by Regret S min Ld ( S/2)d+1 polylog L , L d d+2 T d+1 d+2 polylog(T) ; If r(X ) < S, the standard regret of SELECT is bounded by Regret L d d+2 T d+1 d+2 polylog(T). Remark 4. Recall that S = min{S r(X) : r(X) < S}, one can easily verify that in the realizable case, S = 0 for both concave bandits and the Lipschitz bandits. As such, one cannot directly apply the results in Garivier et al. (2019); Michel et al. (2023) to acquire a constant satisficing regret. With the notion of exceeding gap S and SELECT, we establish constant satisficing regret bounds for these two settings. 4. Lower Bounds: To complement our main results, we also present the satisficing regret lower bounds for finite-armed bandits and concave bandits. The first one is a satisficing regret lower bound for the finite-armed bandits. Theorem 3 (Finite-Armed Bandits). For every non-anticipatory learning algorithm π, every > 0 and T 1/ 2, there exists an instance of two-armed bandit such that r(X ) S = , and the satisficing regret incurred by π is at least Ω(1/ ). Remark 5. In Michel et al. (2023), the authors adapt the results from Bubeck et al. (2013) (for MAB with known optimal mean reward) to establish a Ω(1/ S) satisficing regret lower bound for finite-armed bandits. This is different than the one in Theorem3, which is Ω(1/ ) with being the exceeding gap (i.e., S). This difference originates from the lower bound instances. The lower bound instance in Bubeck et al. (2013) (when adapted to satisficing bandits) sets the threshold value to the optimal mean reward, i.e., S = r(X ) , which makes both the satisficing regret and standard regret lower bounds to be Ω(1/ S). In our lower bound proof, we allow S to be smaller than r(X ), which leads to a lower bound that depends on the difference between these two quantities. Next, we provide a satisficing regret lower bound for bandits with concave reward. Theorem 4 (Bandits with Concave reward). For every non-anticipatory learning algorithm π, every > 0 and every T 1/ 2, there exists an instance of 1-dimensional bandit with concave reward such that r(X ) S = , and the satisficing regret incurred by π is at least Ω(1/ ). 6 NUMERICAL EXPERIMENTS In this section, we conduct numerical experiments to test the performance of SELECT on finitearmed bandits, concave bandits, and Lipschitz bandits. All experiments are run on a PC with 4-core CPU and 16GB of RAM. 6.1 FINITE-ARMED BANDITS Setup: In this case, we consider an instance of 4 arms, and the expected reward of all arms are {0.6, 0.7, 0.8, 1}. The length of the time horizon is set to be 500 to 5000 with a stepsize of 500. We Published as a conference paper at ICLR 2025 conduct experiments for both the realizable case and the non-realizable case. For the realizable case, we set the satisficing level S = 0.93; for the non-realizable case, we set the satisficing level S = 1.5. In both cases, we compare SELECT against Thompson sampling in Agrawal & Goyal (2017) as well as SAT-UCB and SAT-UCB+ in Michel et al. (2023). For SELECT, we use Thompson Sampling as the learning oracle. The experiment is repeated for 1000 times and we report the average results. Results: The results of the realizable case are presented in Figure 2(a), and the results of the non-realizable case are presented in Figure 2(b). From Figure 2(a) one can see that SELECT, SAT-UCB and SAT-UCB+ exhibit constant satisficing regret, and incur less satisficing regret compared to Thompson sampling. Furthermore, SELECT incurs less regret compared to SAT-UCB and achieves comparable performance with SAT-UCB+ in the realizable case. From Figure 2(b) one can see that SELECT incurs roughly the same standard regret as Thompson sampling. Furthermore, SELECT also incurs less standard regret in the non-realizable case compared to either SAT-UCB or SAT-UCB+. (a) Satisficing Regret in the Realizable Case (b) Standard Regret in the Non-Realizable Case Figure 2: Comparison of SELECT , Thompson sampling, SAT-UCB and SAT-UCB+ 6.2 CONCAVE BANDITS Setup: In this case, we consider an instance with arm set [0, 1] and concave reward function r(x) = 1 16(x 0.25)2. The length of the time horizon is set to be 500 to 5000 with a stepsize of 500. We consider both the realizable case and the non-realizable case. For the realizable case, we set the satisficing level S = 0.3; for the non-realizable case, we set the satisficing level S = 1.5. For both cases, we compare our algorithm SELECT with the convex bandit algorithm introduced in Agarwal et al. (2011). For SELECT, we use the algorithm in Agarwal et al. (2011) as the learning oracle. We also use SAT-UCB and SAT-UCB+ as heuristics by viewing the problem as Lipschitz bandits. That is, we first discretize the arm space uniformly with stepsize L 1/3T 1/3, where L is the Lipschitz constant of the reward function, and then run SAT-UCB and SAT-UCB+ over the discretized arm space. The experiment is repeated for 1000 times and we report the average results. Results: The results of realizable case are provided in Figure 3(a), and the results of non-realizable case are provided in Figure 3(b). From Figure 3(a), we can see that SELECT converges to satisficing arms faster and incurs smaller satisficing regret than the algorithm in Agarwal et al. (2011), SAT-UCB and SAT-UCB+. The reason why SAT-UCB and SAT-UCB+ has not converged is that they usually repeatedly pull non-satisficing arms which are close to satisficing level but provide empirically highest UCB. As shown in Figure 3(b), SELECT can also reach good enough distribution compared to the algorithm in Agarwal et al. (2011) , SAT-UCB, and SAT-UCB+ in the non-realizable case. 6.3 LIPSCHITZ BANDITS Setup: In this case, we consider a two-dimensional Lipschitz bandit in domain (x, y) [0, 1]2 with Lipschitz reward f(x, y) = min{1, 3e 100((x 0.5)2+(y 0.7)2)}. The length of the time horizon is set to be 500 to 5000 with a stepsize of 500. We consider both the realizable case and the non-realizable case. For the realizable case, we set the satisficing level S = 0.5; for the non-realizable case, we set the satisficing level S = 1.5. For both cases, we compare our algorithm SELECT with the uniformly discretized UCB introduced in Bubeck et al. (2011) (hereafter referred to as the Uniform UCB ). Published as a conference paper at ICLR 2025 (a) Satisficing Regret in the Realizable Case (b) Standard Regret in the Non-Realizable Case Figure 3: Comparison of SELECT , Stochastic Convex, SAT-UCB and SAT-UCB+ for concave bandit For SELECT, we use the Uniform UCB as the learning oracle. We again use SAT-UCB and SATUCB+ as heuristics by uniformly discretizing the arm space with stepsize L 1/4T 1/4, where L is the Lipschitz constant of the reward function, and run the algorithms over the discretized arms. The experiment is repeated for 1000 times and we report the average results. Results: The results of realizable case are provided in Figure 4(a), and the results of non-realizable case are provided in Figure 4(b). From Figure 4(a), we can see that SELECT converges to satisficing arms faster and incurs smaller satisficing regret than Uniform UCB, SAT-UCB and SAT-UCB+. From 4(b), we can see that SELECT still works the best while Uniform UCB, SAT-UCB and SATUCB+ have similar performance. (a) Satisficing Regret in the Realizable Case (b) Standard Regret in the Non-Realizable Case Figure 4: Comparison of SELECT , Uniform UCB, SAT-UCB and SAT-UCB+ for Lipschitz bandit 7 CONCLUSION In this paper, we propose SELECT, a general algorithmic template, for satisficing regret minimization in bandit optimization. For a given bandit optimization problem and a sub-linear regret learning oracle, we show that SELECT attains constant satisficing regret in the realizable case and the same regret as the learning oracle in the non-realizable case. We instantiate SELECT to finite-armed bandits, concave bandits, and Lipschitz bandits, and also make a discussion on the corresponding lower bounds. Finally, numerical experiments show that SELECT attains constant regret in several popular bandit problems and demonstrates excellent performances when compared to other benchmarks in both the realizable and the non-realizable cases. For future studies, we suggest further exploration in the following directions. First, our work mainly focuses on the stationary settings, and further research could explore satisficing in the nonstationary online learning settings (Wei & Luo, 2021; Cheung et al., 2022). Also, the notion of satisficing regret indeed discourages exploration. Therefore, another direction is to study the role of satisficing regret in settings with limited adaptivity (Gao et al., 2019; Feng et al., 2023). Published as a conference paper at ICLR 2025 Jacob Abernethy, Kareem Amin, and Ruihao Zhu. Threshold bandits, with and without censored feedback. Advances in Neural Information Processing Systems, 29, 2016. Alekh Agarwal, Dean P Foster, Daniel J Hsu, Sham M Kakade, and Alexander Rakhlin. Stochastic convex optimization with bandit feedback. Advances in Neural Information Processing Systems, 24, 2011. Shipra Agrawal and Navin Goyal. Near-optimal regret bounds for thompson sampling. Journal of the ACM (JACM), 64(5):1 24, 2017. Jean-Yves Audibert and S ebastien Bubeck. Best arm identification in multi-armed bandits. In COLT-23th Conference on learning theory-2010, pp. 13 p, 2010. S ebastien Bubeck and Nicolo Cesa-Bianchi. Regret analysis of stochastic and nonstochastic multiarmed bandit problems. Foundations and Trends in Machine Learning, 5(1):1 122, 2012. S ebastien Bubeck, Gilles Stoltz, and Jia Yuan Yu. Lipschitz bandits without the lipschitz constant. In Algorithmic Learning Theory: 22nd International Conference, ALT 2011, Espoo, Finland, October 5-7, 2011. Proceedings 22, pp. 144 158. Springer, 2011. S ebastien Bubeck, Vianney Perchet, and Philippe Rigollet. Bounded regret in stochastic multi-armed bandits. In Conference on Learning Theory, pp. 122 134. PMLR, 2013. Andrew Caplin, Mark Dean, and Daniel Martin. Search and satisficing. American Economic Review, 101(7):2899 2922, 2011. Wang Chi Cheung, David Simchi-Levi, and Ruihao Zhu. Hedging the drift: Learning to optimize under nonstationarity. Management Science, 68(3):1696 1713, 2022. Qing Feng, Ruihao Zhu, and Stefanus Jasin. Temporal fairness in learning and earning: Price protection guarantee and phase transitions. Proceedings of the 24th ACM Conference on Economics and Computation, 2023. Qing Feng, Tianyi Ma, and Ruihao Zhu. Satisficing regret minimization in bandits, 2025. URL https://arxiv.org/abs/2406.06802. Gavan J. Fitzsimons. Consumer response to stockouts. Journal of Consumer Research, 27(2):249 266, 2000. Zijun Gao, Yanjun Han, Zhimei Ren, and Zhengqing Zhou. Batched multi-armed bandits problem. Advances in Neural Information Processing Systems, 32, 2019. Aur elien Garivier, Pierre M enard, and Gilles Stoltz. Explore first, exploit next: The true shape of regret in bandit problems. Mathematics of Operations Research, 44(2):377 399, 2019. Hossein Hajiabolhassan and Ronald Ortner. Online regret bounds for satisficing in mdps. In Sixteenth European Workshop on Reinforcement Learning, 2023. Alihan H uy uk and Cem Tekin. Multi-objective multi-armed bandit with lexicographically ordered and satisficing objectives. Machine Learning, 110(6):1233 1266, 2021. Xiaoqing Jing and Michael Lewis. Stockouts in online retailing. Journal of Marketing Research, 48 (2):342 354, 2011. Hideaki Kano, Junya Honda, Kentaro Sakamaki, Kentaro Matsuura, Atsuyoshi Nakamura, and Masashi Sugiyama. Good arm identification via bandit feedback. Machine Learning, 108:721 745, 2019. Tor Lattimore and Csaba Szepesv ari. Bandit algorithms. Cambridge University Press, 2020. Michael Levy, Barton Weitz, and Dhruv Grewal. Retailing Management. Mc Graw Hill, 2023. Published as a conference paper at ICLR 2025 Andrea Locatelli, Maurilio Gutzeit, and Alexandra Carpentier. An optimal algorithm for the thresholding bandit problem. In International Conference on Machine Learning, pp. 1690 1698. PMLR, 2016. Bryan Lufkin. Do maximisers or satisficers make better decisions? BBC, 2021. Chris Mabey. Optimizing vs satisficing: Tips for product design and designing your life. The BYU Design Review, 2022. Thomas Michel, Hossein Hajiabolhassan, and Ronald Ortner. Regret bounds for satisficing in multiarmed bandit problems. Transactions on Machine Learning Research, 2023. Subhojyoti Mukherjee, Kolar Purushothama Naveen, Nandan Sudarsanam, and Balaraman Ravindran. Thresholding bandits with augmented ucb. ar Xiv preprint ar Xiv:1704.02281, 2017. Paul Reverdy and Naomi E Leonard. Satisficing in gaussian bandit problems. In 53rd IEEE Conference on Decision and Control, pp. 5718 5723. IEEE, 2014. Paul Reverdy, Vaibhav Srivastava, and Naomi Ehrich Leonard. Satisficing in multi-armed bandit problems. IEEE Transactions on Automatic Control, 62(8):3788 3803, 2017. Daniel Russo and Benjamin Van Roy. Satisficing in time-sensitive bandit learning. Mathematics of Operations Research, 47(4):2815 2839, 2022. Herbert A Simon. Rational choice and the structure of the environment. Psychological Review, 1956. Akihiro Tamatsukuri and Tatsuji Takahashi. Guaranteed satisficing and finite regret: Analysis of a cognitive satisficing value function. Biosystems, 180:46 53, 2019. Chen-Yu Wei and Haipeng Luo. Non-stationary reinforcement learning without prior knowledge: an optimal black-box approach. Proceedings of Thirty Fourth Conference on Learning Theory, 134:4300 4354, 2021. Kathryn Whitenton. Satisficing: Quickly meet users main needs. Nielsen Norman Group, 2014.