skip to main content
research-article
Open Access

Learning Equilibria in Matching Markets with Bandit Feedback

Published:24 May 2023Publication History

Skip Abstract Section

Abstract

Large-scale, two-sided matching platforms must find market outcomes that align with user preferences while simultaneously learning these preferences from data. Classical notions of stability (Gale and Shapley, 1962; Shapley and Shubik, 1971) are, unfortunately, of limited value in the learning setting, given that preferences are inherently uncertain and destabilizing while they are being learned. To bridge this gap, we develop a framework and algorithms for learning stable market outcomes under uncertainty. Our primary setting is matching with transferable utilities, where the platform both matches agents and sets monetary transfers between them. We design an incentive-aware learning objective that captures the distance of a market outcome from equilibrium. Using this objective, we analyze the complexity of learning as a function of preference structure, casting learning as a stochastic multi-armed bandit problem. Algorithmically, we show that “optimism in the face of uncertainty,” the principle underlying many bandit algorithms, applies to a primal-dual formulation of matching with transfers and leads to near-optimal regret bounds. Our work takes a first step toward elucidating when and how stable matchings arise in large, data-driven marketplaces.

Skip 1INTRODUCTION Section

1 INTRODUCTION

Data-driven marketplaces face the simultaneous challenges of learning agent preferences and aligning market outcomes with the incentives induced by these preferences. Consider, for instance, online platforms that match two sides of a market to each other (e.g., Lyft, TaskRabbit, and Airbnb). On these platforms, customers are matched to service providers and pay for the service they receive. If agents on either side are not offered desirable matches at fair prices, then they would have an incentive to leave the platform and switch to a competing platform. Agent preferences, however, are often unknown to the platform and must be learned. When faced with uncertainty about agent preferences (and thus incentives), when can a marketplace efficiently explore and learn market outcomes that align with agent incentives?

We center our investigation around a model called matching with transferable utilities, proposed by Shapley and Shubik [42]. In this model, there is a two-sided market of customers and service providers. Each customer has a utility that they derive from being matched to a given provider and vice versa. The platform selects a matching between the two sides and assigns a monetary transfer between each pair of matched agents. Transfers are a salient feature of most real-world matching markets: riders pay drivers on Lyft, clients pay freelancers on TaskRabbit, and guests pay hosts on Airbnb. An agent’s net utility is their value for being matched to their partner plus the value of their transfer (either of which can be negative in the cases of costs and payments). In matching markets, the notion of stability captures alignment of a market outcome with agent incentives. Informally, a market outcome is stable if no pair of agents would rather match with each other than abide by the market outcome, and stable matchings can be computed when preferences are fully known.

In the context of large-scale matching platforms, however, the assumption that preferences are known breaks down. Platforms usually cannot have users report their complete preference profiles. Moreover, users may not even be aware of what their own preferences are. For example, a freelancer may not exactly know what types of projects they prefer until actually trying out specific ones. In reality, a data-driven platform is more likely to learn information about preferences from repeated feedback1 over time. Two questions now emerge: In such marketplaces, how can stable matchings be learned? And what underlying structural assumptions are necessary for efficient learning to be possible?

To address these questions, we propose and investigate a model for learning stable matchings from noisy feedback. We model the platform’s learning problem using stochastic multi-armed bandits, which lets us leverage the extensive body of work in the bandit literature to analyze the data efficiency of learning (see Lattimore and Szepesvári [30] for a textbook treatment). More specifically, our three main contributions are: (i) We develop an incentive-aware learning objective—Subset Instability—that captures the distance of a market outcome from equilibrium. (ii) Using Subset Instability as a measure of regret, we show that any “UCB-based” algorithm from the classical bandit literature can be adapted to this incentive-aware setting. (iii) We instantiate this idea for several families of preference structures to design efficient algorithms for incentive-aware learning. This helps elucidate how preference structure affects the complexity of learning stable matchings.

Designing the learning objective. Since mistakes are inevitable while exploring and learning, achieving exact stability at every timestep is an unattainable goal. To address this issue, we lean on approximation, focusing on learning market outcomes that are approximately stable. Thus, we need a metric that captures the distance of a market outcome from equilibrium.2

We introduce a notion for approximate stability that we call Subset Instability. Specifically, we define the Subset Instability of a market outcome to be the maximum difference, over all subsets \(\mathcal {S}\) of agents, between the total utility of the maximum weight matching on \(\mathcal {S}\) and the total utility of \(\mathcal {S}\) under the market outcome.3 We show that Subset Instability can be interpreted as the amount the platform would have to subsidize participants to keep them on the platform and make the resulting matching stable. We can also interpret Subset Instability as the platform’s cost of learning when facing competing platforms with greater knowledge of user preferences. Finally, we show that Subset Instability is the maximum gain in utility that a coalition of agents could have derived from an alternate matching such that no agent in the coalition is worse off.

Subset Instability also satisfies the following properties, which make it suitable for learning: (i) Subset Instability is equal to zero if and only if the market outcome is (exactly) stable; (ii) Subset Instability is robust to small perturbations to the utility functions of individual agents, which is essential for learning with noisy feedback; and (iii) Subset Instability upper bounds the utility difference of a market outcome from the socially optimal market outcome.

Designing algorithms for learning a stable matching. Using Subset Instability, we investigate the problem of learning a stable market outcome from noisy user feedback using the stochastic contextual bandit model (see, e.g., Lattimore and Szepesvári [30]). In each round, the platform selects a market outcome (i.e., a matching along with transfers) with the goal of minimizing cumulative instability.

We develop a general approach for designing bandit algorithms within our framework. Our approach is based on a primal-dual formulation of matching with transfers [42], in which the primal variables correspond to the matching and the dual variables can be used to set the transfers. We find that “optimism in the face of uncertainty,” the principle underlying many UCB-style bandit algorithms [5, 30], can be adapted to this primal-dual setting. The resulting algorithm is simple: Maintain upper confidence bounds on the agent utilities and compute, in each round, an optimal primal-dual pair in terms of these upper confidence bounds. The crux of the analysis is the following lemma, which bounds instability by the gap between the upper confidence bound and true utilities:

Lemma 1.1 (Informal, see 5.4 for a Formal Statement).

Given confidence sets for each utility value such that each confidence set contains the true utility, let \((X, \tau)\) be a stable matching with transfers with respect to the utility functions given by the upper confidence bounds. The instability of \((X, \tau)\) is upper bounded by the sum of the sizes of the confidence sets of pairs in X.

We can thus analyze our algorithms by combining Lemma 1.1 with the analyses of existing UCB-style algorithms. In particular, we can essentially inherit the bounds on the size of the confidence bounds from traditional analyses of multi-arm bandits.

Complexity of learning a stable matching. Our main technical result is a collection of regret bounds for different structural assumptions on agent preferences. These bounds resemble the classical stochastic multi-armed bandits bounds when rewards have related structural assumptions. We summarize these regret bounds in Table 1 and elaborate on them in more detail below.

Table 1.
Regret bound
Unstructured preferences\(\widetilde{O}\bigl (N \sqrt {nT}\bigr)\)
Typed preferences\(\widetilde{O}\bigl (|\mathcal {C}| \sqrt {nT} \bigr)\)
Separable linear preferences\(\widetilde{O}\bigl (d\sqrt {N} \sqrt {nT}\bigr)\)

Table 1. Regret Bounds for Varying Preference Structures (for N Agents and n Agents Per Round)

Theorem 1.2 (Unstructured Preferences, Informal).

For unstructured preferences, there exists a UCB-style algorithm that incurs \(\widetilde{O}(N \sqrt {nT})\) regret according to Subset Instability after T rounds, where N is the number of agents on the platform and n is the number of agents that arrive in any round. (This bound is optimal up to logarithmic factors.)

Theorem 1.3 (Typed Preferences, Informal).

Consider preferences such that each agent a has a type \(c_a \in \mathcal {C}\) and the utility of a when matched to another agent \(a^{\prime }\) is given by a function of the types \(c_a\) and \(c_{a^{\prime }}\). There exists a UCB-style algorithm that incurs \(\widetilde{O}(|\mathcal {C}| \sqrt {n T})\) regret according to Subset Instability after T rounds, where n is the maximum number of agents that arrive to the platform in any round.

Theorem 1.4 (Separable Linear Preferences, Informal).

Consider preferences such that the utility of an agent a when matched to another agent \(a^{\prime }\) is \(\langle \phi (a), c_{a^{\prime }}\rangle\), where \(\phi (a) \in \mathbb {R}^d\) is unknown and \(c_{a^{\prime }} \in \mathbb {R}^d\) is known. There exists a UCB-style algorithm that incurs \(\widetilde{O}(d \sqrt {N} \sqrt {n T})\) regret according to Subset Instability after T rounds, where N is the number of agents on the platform and n is the maximum number of agents that arrive in any round.

These results elucidate the role of preference structure on the complexity of learning a stable matching. Our regret bounds scale with \(N \sqrt {n T}\) for unstructured preferences (Theorem 1.2), \(|\mathcal {C}| \sqrt {n T}\) for typed preferences (Theorem 1.3), and \(d \sqrt {N} \sqrt {n T}\) for linear preferences (Theorem 1.4). To illustrate these differences in a simple setting, let us consider the case where all of the agents show up every round, so \(n = N\). In this case, our regret bound for unstructured preferences is superlinear in N; in fact, this dependence on N is necessary as we demonstrate via a lower bound (see Lemma 5.5). However, the complexity of learning a stable matching changes substantially with preference structure assumptions. In particular, our regret bounds are sublinear/linear in N for typed preferences and separable linear preferences. This means that in large markets, a centralized platform can efficiently learn a stable matching with these preference structure assumptions.

Connections and extensions. Key to our results and extensions is the primal-dual characterization of equilibria in matching markets with transfers. Specifically, equilibria are described by a linear program whose primal form maximizes total utility over matchings and whose dual variables correspond to transfers. This linear program inspires our definition of Subset Instability, connects Subset Instability to platform profit (see Section 6.2), and relates learning with Subset Instability to regret minimization in combinatorial bandits (see Section 5.4). We adapt ideas from combinatorial bandits to additionally obtain \(O(\log T)\) instance-dependent regret bounds (see Section 6.1).

Our approach also offers a new perspective on learning stable matchings in markets with non-transferable utilities [17, 32]. Although this setting does not admit a linear program formulation, we show Subset Instability can be extended to what we call NTU Subset Instability (see Section 6.3), which turns out to have several advantages over the instability measures studied in previous work. Our algorithmic principles extend to NTU Subset Instability: We prove regret bounds commensurate with those for markets with transferable utilities.

1.1 Related Work

In the machine learning literature, starting with Das and Kamenica [17] and Liu et al. [32], several works [9, 12, 17, 32, 33, 40] study learning stable matchings from bandit feedback in the Gale-Shapley stable marriage model [23]. A major difference between this setting and ours is the absence of monetary transfers between agents. These works focus on the utility difference rather than the instability measure that we consider. Cen and Shah [12] extend this bandits model to incorporate fixed, predetermined cost/transfer rules. However, they do not allow the platform to set arbitrary transfers between agents. Moreover, they also consider a weaker notion of stability that does not consider agents negotiating arbitrary transfers: Defecting agents must set their transfers according to a fixed, predetermined structure. In contrast, we follow the classical definition of stability [42].

Outside of the machine learning literature, several papers also consider the complexity of finding stable matchings in other feedback and cost models, e.g., communication complexity [4, 24, 43] and query complexity [4, 20]. Of these works, Shi [43], which studies the communication complexity of finding approximately stable matchings with transferable utilities, is perhaps most similar to ours. This work assumes agents know their preferences and focuses on the communication bottleneck, whereas we study the costs associated with learning preferences. Moreover, the approximate stability notion in Shi [43] is the maximum unhappiness of any pair of agents, whereas Subset Instability is equivalent to the maximum unhappiness over any subset of agents. For learning stable matchings, Subset Instability has the advantages of being more fine-grained and having a primal view that motivates a clean UCB-based algorithm.

Our notion of instability connects to historical works in coalitional game theory: Related are the concepts of the strong-\(\varepsilon\) core of Shapley and Shubik [41] and the indirect function of Martínez-Legaz [37], although each was introduced in a very different context than ours. Nonetheless, they reinforce the fact that our instability notion is a very natural one to consider.

A complementary line of work in economics [2, 11, 34, 35] considers stable matchings under incomplete information. These works focus on defining stability when the agents have incomplete information about their own preferences, whereas we focus on the platform’s problem of learning stable matchings from noisy feedback. As a result, these works relax the definition of stability to account for uncertainty in the preferences of agents, rather than the uncertainty experienced by the platform from noisy feedback.

Multi-armed bandits have also been applied to learning in other economic contexts. For example, learning a socially optimal matching (without learning transfers) is a standard application of combinatorial bandits [13, 15, 16, 22, 29]. Other applications at the interface of bandit methodology and economics include dynamic pricing [8, 27, 38], incentivizing exploration [21, 36], learning under competition [3], and learning in matching markets without incentives [26].

Finally, primal-dual methods have also been applied to other problems in the bandits literature (e.g., References [25, 31, 44]).

Skip 2PRELIMINARIES Section

2 PRELIMINARIES

The foundation of our framework is the matching with transfers model of Shapley and Shubik [42]. In this section, we introduce this model along with the concept of stable matching.

2.1 Matching with Transferable Utilities

Consider a two-sided market that consists of a finite set \(\mathcal {I}\) of customers on one side and a finite set \(\mathcal {J}\) of providers on the other. Let \(\mathcal {A}:=\mathcal {I}\cup \mathcal {J}\) be the set of all agents. A matching \(X\subseteq \mathcal {I}\times \mathcal {J}\) is a set of pairs \((i, j)\) that are pairwise disjoint, representing the pairs of agents that are matched. Let \(\mathscr{X}_\mathcal {A}\) denote the set of all matchings on \(\mathcal {A}\). For notational convenience, we define for each matching \(X\in \mathscr{X}_\mathcal {A}\) an equivalent functional representation \(\mu _{X}:\mathcal {A}\rightarrow \mathcal {A}\), where \(\mu _{X}(i) = j\) and \(\mu _{X}(j) = i\) for all matched pairs \((i, j)\in X\), and \(\mu _{X}(a) = a\) if \(a\in \mathcal {A}\) is unmatched.

When a pair of agents \((i, j)\in \mathcal {I}\times \mathcal {J}\) matches, each experiences a utility gain. We denote these utilities by a global utility function \(u: \mathcal {A}\times \mathcal {A}\rightarrow \mathbb {R}\), where \(u(a, a^{\prime })\) denotes the utility that agent a gains from being matched to agent \(a^{\prime }\). (If a and \(a^{\prime }\) are on the same side of the market, then we take \(u(a, a^{\prime })\) to be zero by default.) We allow these utilities to be negative if matching results in a net cost (e.g., if an agent is providing a service). We assume each agent \(a\in \mathcal {A}\) receives zero utility if unmatched, i.e., \(u(a, a) = 0\). When we wish to emphasize the role of an individual agent’s utility function, we will use the equivalent notation \(u_a(a^{\prime }):= u(a, a^{\prime })\).

A market outcome consists of a matching \(X\in \mathscr{X}_\mathcal {A}\) along with a vector \(\tau \in \mathbb {R}^{\mathcal {A}}\) of transfers, where \(\tau _a\) is the amount of money transferred from the platform to agent a for each \(a\in \mathcal {A}\). These monetary transfers are a salient feature of most real-world matching markets: Riders pay drivers on Lyft, clients pay freelancers on TaskRabbit, and guests pay hosts on Airbnb. Shapley and Shubik [42] capture this aspect of matching markets by augmenting the classical two-sided matching model with transfers of utility between agents. Transfers are typically required to be zero-sum, meaning that \(\tau _i+ \tau _j= 0\) for all matched pairs \((i, j)\in X\) and \(\tau _a= 0\) if a is unmatched. Here, X represents how agents are matched and \(\tau _a\) represents the transfer that agent a receives (or pays). The net utility that an agent a derives from a matching with transfers \((X, \tau)\) is therefore \(u(a, \mu _{X}(a)) + \tau _a\).

Stable matchings. In matching theory, stability captures when a market outcome aligns with individual agents’ preferences. Roughly speaking, a market outcome \((X, \tau)\) is stable if: (i) no individual agent a would rather be unmatched and (ii) no pair of agents \((i, j)\) can agree on a transfer such that both would rather match with each other than abide by \((X, \tau)\). Formally:

Definition 2.1.

A market outcome \((X, \tau)\) is stable if: (i) it is individually rational, i.e., (1) \(\begin{equation} u_a(\mu _{X}(a)) + \tau _a\ge 0 \end{equation}\) for all agents \(a\in \mathcal {A}\), and (ii) it has no blocking pairs, i.e., (2) \(\begin{equation} \bigl (u_i(\mu _{X}(i)) + \tau _i\bigr) + \bigl (u_j(\mu _{X}(j)) + \tau _j\bigr)\ge u_i(j) + u_j(i) \end{equation}\) for all pairs of agents \((i,j)\in \mathcal {I}\times \mathcal {J}\).4

A fundamental property of the matching with transfers model is that if \((X, \tau)\) is stable, then X is a maximum weight matching, i.e., X maximizes \(\sum _{a\in \mathcal {A}} u_a(\mu _{X}(a))\) over all matchings \(X\in \mathscr{X}_{\mathcal {A}}\) [42]. The same work shows that stable market outcomes coincide with Walrasian equilibria. (For completeness, we recapitulate the basic properties of this model in Appendix A.)

To make the matching with transfers model concrete, we use the simple market depicted in the center panel of Figure 1 as a running example throughout the article. This market consists of a customer Charlene and two providers Percy and Quinn, which we denote by \(\mathcal {I}= \lbrace C\rbrace\) and \(\mathcal {J}= \lbrace P, Q\rbrace\). If the agents’ utilities are as given in Figure 1, then Charlene would prefer Quinn, but Quinn’s cost of providing the service is much higher. Thus, matching Charlene and Percy is necessary for a stable outcome. This matching is stable for any transfer from Charlene to Percy in the interval \([5, 7]\).

Fig. 1.

Fig. 1. The left panel depicts a schematic of a matching (blue) with transfers (green). The center panel depicts a matching market with three agents and a stable matching with transfers for that market. (If the transfer 6 is replaced with any value between 5 and 7, then the outcome remains stable.) The right panel depicts the same market, but with utilities replaced by uncertainty sets; note that no matching with transfers is stable for all realizations of utilities.

Skip 3LEARNING PROBLEM AND FEEDBACK MODEL Section

3 LEARNING PROBLEM AND FEEDBACK MODEL

We instantiate the platform’s learning problem in a stochastic contextual bandits framework. Matching takes place over the course of T rounds. We denote the set of all customers by \(\mathcal {I}^*\), the set of all providers by \(\mathcal {J}^*\), and the set of all agents on the platform by \(\mathcal {A}^*= \mathcal {I}^*\cup \mathcal {J}^*\). Each agent \(a\in \mathcal {A}^*\) has an associated context \(c_a\in \mathcal {C}\), where \(\mathcal {C}\) is the set of all possible contexts. This context represents the side information available to the platform about the agent, e.g., demographic, location, or platform usage information. Each round, a set of agents arrives to each side of the market. The platform then selects a market outcome and incurs a regret equal to the instability of the market outcome (which we introduce formally in Section 4). Finally, the platform receives noisy feedback about the utilities of each matched pair \((i, j)\).

To interpret the noisy feedback, note that platforms in practice often receive feedback both explicitly (e.g., riders rating drivers after a Lyft ride) and implicitly (e.g., engagement metrics on an app). In either instance, feedback is likely to be sparse and noisy. For simplicity, we do not account for agents strategically manipulating their feedback to the platform and focus on the problem of learning preferences from unbiased reports.

We now describe this model more formally. In the t-th round:

  • A set \(\mathcal {I}^t\subseteq \mathcal {I}^*\) of customers and a set \(\mathcal {J}^t\subseteq \mathcal {J}^*\) of providers arrive to the market. Write \(\mathcal {I}^t\cup \mathcal {J}^t=: \mathcal {A}^t\). The platform observes the identity a and the context \(c_a\in \mathcal {C}\) of each agent \(a\in \mathcal {A}^t\).

  • The platform selects a matching with zero-sum transfers \((X^t, \tau ^t)\) between \(\mathcal {I}^t\) and \(\mathcal {J}^t\).

  • The platform observes noisy utilities \(u_a(\mu _{X^t}(a)) + \varepsilon _{a,t}\) for each agent \(a\in \mathcal {I}^t\cup \mathcal {J}^t\), where the \(\varepsilon _{a,t}\) are independent, 1-subgaussian random variables.5

  • The platform incurs regret equal to the instability of the selected market outcome \((X^t, \tau ^t)\). (We define instability formally in Section 4.)

The platform’s total regret \(R_T\) is thus the cumulative instability incurred up through round T.

3.1 Preference Structure

In this bandits framework, we can impose varying degrees of structure on agent preferences. We encode these preference structures via the functional form of agents’ utility functions and their relation to agent contexts. More formally, let \(\hspace{1.111pt}\mathcal {U}\) be the set of functions \(u:\mathcal {A}^*\times \mathcal {A}^*\rightarrow \mathbb {R}\), i.e., \(\hspace{1.111pt}\mathcal {U}\) is the set of all possible (global) utility functions. We now introduce several classes of preference structures as subsets of \(\hspace{1.111pt}\mathcal {U}\).

Unstructured preferences. The simplest setting we consider is one where the preferences are unstructured. Specifically, we consider the class of utility functions \(\begin{equation*} \hspace{1.111pt}\mathcal {U}_{\mathrm{unstructured}}= \lbrace u \in \hspace{1.111pt}\mathcal {U}\mid u(a, a^{\prime }) \in [-1, 1]\rbrace . \end{equation*}\) (Here, one can think of the context as being uninformative, i.e., \(\mathcal {C}\) is the singleton set.) In this setup, the platform must learn each agent’s utility function \(u_a(\cdot) = u(a, \cdot)\).

Typed preferences. We next consider a market where each agent comes in one of finitely many types, with agents of the same type having identical preferences. Assuming typed preference structures is standard in theoretical models of markets (see, e.g., Azevedo and Hatfield [7], Debreu and Scarf [18], Echenique et al. [19]). We can embed types into our framework by having each agent’s context represent their type, with \(|\mathcal {C}| \lt \infty\). The global utility function is then fully specified by agents’ contexts: \(\begin{equation*} \hspace{1.111pt}\mathcal {U}_{\mathrm{typed}}= \left\lbrace u \in \hspace{1.111pt}\mathcal {U}\mid u(a, a^{\prime }) = f(c_a, c_{a^{\prime }}) \text{ for some $f:\mathcal {C}\times \mathcal {C}\rightarrow [-1,1]$}\right\rbrace . \end{equation*}\)

Separable linear preferences. We next consider markets where each agent is associated with known information given by their context as well as hidden information that must be learned by the platform. (This differs from unstructured preferences, where all information was hidden, and typed preferences, where each agent’s context encapsulated their full preferences.) We explore this setting under the assumption that agents’ contexts and hidden information interact linearly.

We assume that all contexts belong to \(\mathcal {B}^d\) (i.e., \(\mathcal {C}=\mathcal {B}^d\)) where \(\mathcal {B}^d\) is the \(\ell _2\) unit ball in \(\mathbb {R}^d\). We also assume that there exists a function \(\phi :\mathcal {A}^*\rightarrow \mathcal {B}^d\) mapping each agent to the hidden information associated to that agent. The preference class \(\hspace{1.111pt}\mathcal {U}_{\mathrm{linear}}^{d}\) can then be defined as \(\begin{equation*} \hspace{1.111pt}\mathcal {U}_{\mathrm{linear}}^d = \lbrace u\in \mathcal {U} \bigm \vert u(a, a^{\prime }) = \langle c_{a^{\prime }}, \phi (a) \rangle \text{ for some $\phi :\mathcal {A}^*\rightarrow \mathcal {B}^d$}\rbrace . \end{equation*}\)

Skip 4MEASURING APPROXIMATE STABILITY Section

4 MEASURING APPROXIMATE STABILITY

When learning stable matchings, we must settle for guarantees of approximate stability, since exact stability—a binary notion—is unattainable when preferences are uncertain. To see this, we return to the example from Figure 1. Suppose that the platform has uncertainty sets given by the right panel. Recall that for the true utilities, all stable outcomes match Charlene with Percy. If the true utilities were instead the upper bounds of each uncertainty set, then all stable outcomes would match Charlene and Quinn. Given only the uncertainty sets, it is impossible for the platform to find an (exactly) stable matching, so it is necessary to introduce a measure of approximate stability as a relaxed benchmark for the platform; we turn to this now.

Given the insights of Shapley and Shubik [42]—that all stable outcomes maximize the sum of agents’ utilities—it might seem natural to measure distance from stability simply in terms of the utility difference. To define this formally, let \(\mathcal {A}\) be the set of agents participating in the market. (This corresponds to \(\mathcal {A}^t\) at timestep t in the bandits model.) The utility difference6 of a market outcome \((X, \tau)\) is given by: (3) \(\begin{equation} \left(\max _{X^{\prime }\in \mathscr{X}_\mathcal {A}} \sum _{a\in \mathcal {A}} u_a(\mu _{X^{\prime }}(a)) \right) - \left(\sum _{a \in \mathcal {A}} u_a(\mu _{X}(a)) + \tau _a)\right). \end{equation}\) The first term \(\max _{X^{\prime }\in \mathscr{X}_\mathcal {A}} \sum _{a\in \mathcal {A}} u_a(\mu _{X^{\prime }}(a))\) is the maximum total utility of any matching, and the second term \(\sum _{a \in \mathcal {A}} (u_a(\mu _{X}(a)) + \tau _a)\) is the total utility of market outcome \((X, \tau)\). Since transfers are zero-sum, (3) can be equivalently written as \(\begin{equation*} \left(\max _{X^{\prime }\in \mathscr{X}_\mathcal {A}} \sum _{a\in \mathcal {A}} u_a(\mu _{X^{\prime }}(a)) \right) - \sum _{a \in \mathcal {A}} u_a(\mu _{X}(a)). \end{equation*}\) But this shows that utility difference actually ignores the transfers \(\tau\) entirely! In fact, the utility difference can be zero even when the transfers lead to a market outcome that is far from stable (see Section B.1). Utility difference is therefore not incentive-aware, making it unsuitable as an objective for learning stable matchings with transfers.

In the remainder of this section, we propose a measure of instability—Subset Instability—which we will show serves as a suitable objective for learning stable matchings with transfers. Specifically, we show that Subset Instability captures the distance of a market outcome from equilibrium while reflecting both the platform’s objective and the users’ incentives. We additionally show that Subset Instability satisfies several structural properties that make it useful for learning.

4.1 Subset Instability

Subset Instability is based on utility difference, but rather than only looking at the market in aggregate, it takes a maximum ranging over all subsets of agents.

Definition 4.1.

Given utilities u, the Subset Instability \(I(X, \tau ; u, \mathcal {A})\) of a matching with transfers \((X, \tau)\) is \(\begin{equation*} \max _{\mathcal {S}\subseteq \mathcal {A}} \left[ \left(\max _{X^{\prime }\in \mathscr{X}_\mathcal {S}} \sum _{a\in \mathcal {S}} u_a(\mu _{X^{\prime }}(a)) \right) - \left(\sum _{a\in \mathcal {S}} u_a(\mu _{X}(a)) + \tau _a\right) \right]. \qquad \qquad \text{($\ast $)} \end{equation*}\) (The first term \(\max _{X^{\prime } \in \mathscr{X}_\mathcal {S}} \sum _{a\in \mathcal {S}} u_a(\mu _{X^{\prime }}(a))\) is the maximum total utility of any matching over \(\mathcal {S}\), and the second term \(\sum _{a \in \mathcal {A}} (u_a(\mu _{X}(a)) + \tau _a)\) is the total utility of the agents in \(\mathcal {S}\) under market outcome \((X, \tau)\).)

Intuitively, Subset Instability captures stability because it checks whether any subset of agents would prefer an alternate outcome. We provide a more extensive economic interpretation below; but before doing so, we first illustrate Definition 4.1 in the context of the example in Figure 1.

Consider the matching \(X= \lbrace (C, Q)\rbrace\) with transfers \(\tau _C = -11\) and \(\tau _Q = 11\). (This market outcome is stable for the upper bounds of the uncertainty sets of the platform in Figure 1, but not stable for the true utilities.) It is not hard to see that the subset \(\mathcal {S}\) that maximizes Subset Instability is \(\mathcal {S}= \left\lbrace C, P \right\rbrace\), in which case \(\max _{X^{\prime }\in \mathscr{X}_\mathcal {S}} \sum _{a\in \mathcal {S}} u_a(\mu _{X^{\prime }}(a))= 4\) and \(\sum _{a\in \mathcal {S}} \left(u_a(\mu _{X}(a)) + \tau _a\right) = 1\). Thus, the Subset Instability of \((X, \tau)\) is \(I(X, \tau ; u, \mathcal {A}) = 4 - 1 = 3\). In contrast, the utility difference of \((X, \tau)\) is 2.

We now discuss several interpretations of Subset Instability, which provide further insight into why Subset Instability serves as a meaningful notion of approximate stability in online marketplaces. In particular, Subset Instability can be interpreted as the minimum stabilizing subsidy, as the platform’s cost of learning, as a measure of user unhappiness, and as a distance from equilibrium.

Subset Instability as the platform’s minimum stabilizing subsidy. Subset Instability can be interpreted in terms of monetary subsidies from the platform to the agents. Specifically, the Subset Instability of a market outcome equals the minimum amount the platform could subsidize agents so the subsidized market outcome is individually rational and has no blocking pairs.

More formally, let \(s\in \mathbb {R}^\mathcal {A}_{\ge 0}\) denote subsidies made by the platform, where the variable \(s_a\ge 0\) represents the subsidy provided to agent a.7 For a market outcome \((X, \tau)\), the minimum stabilizing subsidy is (4) \(\begin{equation} \min _{s\in \mathbb {R}_{\ge 0}^{\mathcal {A}}} \Biggl \lbrace \sum _{a\in \mathcal {A}} s_{a} \biggm \vert \text{$(X, \tau + s)$ is stable}\Biggr \rbrace , \end{equation}\) where we define stability in analogy to Definition 2.1. Specifically, we say that a market outcome \((X, \tau)\) with subsidies s is stable if it is individually rational, i.e., \(u_a(\mu _{X}(a)) + \tau _a+ s_a\ge 0\) for all agents \(a\in \mathcal {A}\), and has no blocking pairs, i.e., \((u_i(\mu _{X}(i)) + \tau _i+ s_i) + (u_j(\mu _{X}(j)) + \tau _j+ s_j)\ge u_i(j) + u_j(i)\) for all pairs of agents \((i, j)\in \mathcal {I}\times \mathcal {J}\).

Given this setup, we show the following equivalence:

Proposition 4.2.

Minimum stabilizing subsidy equals Subset Instability for any market outcome.

The proof boils down to showing that the two definitions are “dual” to each other. To formalize this, we rewrite the minimum stabilizing subsidy as the solution to the following linear program:8 (5) \(\begin{align} \min _{s\in \mathbb {R}^{|\mathcal {A}|}} & \sum _{a\in \mathcal {A}} s_{a} & \\ \text{s.t.}\ \ & \bigl (u_i(\mu _{X}(i)) + \tau _i+ s_i\bigr) + \bigl (u_j(\mu _{X}(j)) + \tau _j+ s_j\bigr)\ge u_i(j) + u_j(i)\qquad &&\forall (i,j)\in \mathcal {I}\times \mathcal {J}\nonumber \nonumber\\ & u_a(\mu _{X}(a)) + \tau _a+ s_a\ge 0\quad &&\forall a\in \mathcal {A}\nonumber \nonumber\\ &s_a\ge 0 \quad &&\forall a\in \mathcal {A}. \nonumber \nonumber \end{align}\) The crux of our argument is that the dual linear program to (5) maximizes the combinatorial objective (\(\ast\)). The equivalence of (\(\ast\)) and (5) then follows from strong duality.

With this alternate formulation of Subset Instability in mind, we revisit the example in Figure 1. Again, consider the matching \(X= \lbrace (C, Q)\rbrace\) with transfers \(\tau _C = -11\) and \(\tau _Q = 11\). (This is stable for the upper bounds of the uncertainty sets of the platform in Figure 1, but not stable for the true utilities.) We have already shown above that the Subset Instability of this market outcome is 3. To see this via the subsidy formulation, note that the optimal subsidy s gives C and P a total of 3. (E.g., we give C a subsidy of \(s_C = 2\) and P a subsidy of \(s_P = 1\).) Indeed, if \(s_C + s_P = 3\), then \(\begin{equation*} \bigl (u_C(\mu _{X}(C)) + \tau _C + s_C\bigr) + \bigl (u_P(\mu _{X}(P)) + \tau _P + s_P\bigr) \ge u_C(P) + u_P(C) \end{equation*}\) holds (with equality), so the pair \((C, P)\) could no longer gain by matching with each other.

The subsidy perspective turns out to be useful when designing learning algorithms. In particular, while the formulation in Definition 4.1 involves a maximization over the \(2^{|\mathcal {A}|}\) subsets of \(\mathcal {A}\), the linear programming formulation (5) only involves \(O(|\mathcal {A}|)\) variables and \(O(|\mathcal {A}|^2)\) constraints.

Subset Instability as the platform’s cost of learning. We next connect minimum stabilizing subsidies to the platform’s cost of learning—how much the platform would have to pay to keep users on the platform in the presence of a worst-case (but budget-balanced) competitor with perfect knowledge of agent utilities.

Observe that (4) is the minimum amount the platform could subsidize agents so that no budget-balanced competitor could convince agents to leave. The way that we formalize “convincing agents to leave” is that: (a) an agent will leave the original platform if they prefer to be unmatched over being on the platform or (b) a pair of agents who are matched on the competitor’s platform will leave the original platform if they both prefer the new market outcome over their original market outcomes. Thus, if we imagine the platform as actually paying the subsidies, then the cumulative instability (i.e., our regret) can be realized as a “cost of learning”: It is how much the platform pays the agents to learn a stable outcome while ensuring that no agent has the incentive to leave during the learning process. Later on, we will see that our algorithmic approach can be extended to efficiently compute feasible subsidies for (5) that are within a constant factor of our regret bound, meaning that subsidies can be implemented using only the information that the platform has. Moreover, in Section 6.2, we show that cost of learning can also be explicitly connected to the platform’s revenue.

Subset Instability as a measure of user unhappiness. While the above interpretations focus on Subset Instability from the platform’s perspective, we show that Subset Instability can also be interpreted as a measure of user unhappiness. Given a subset \(\mathcal {S}\subseteq \mathcal {A}\) of agents, which we call a coalition, we define the unhappiness of \(\mathcal {S}\) with respect to a market outcome \((X, \tau)\) to be the maximum gain (relative to \((X, \tau)\)) in total utility that the members of coalition \(\mathcal {S}\) could achieve by matching only among themselves, such that no member is worse off than they were in \((X, \tau)\). (See Section B.3 for a formal definition.) The condition that no member is worse off ensures that all agents would actually want to participate in the coalition (i.e. they prefer it to the original market outcome).

User unhappiness differs from the original definition of Subset Instability in (\(\ast\)), becaus (\(\ast\)) does not require individuals to be better off in any alternative matching. However, we show that this difference is inconsequential:

Proposition 4.3.

The maximum unhappiness of any coalition \(\mathcal {S}\subseteq \mathcal {A}\) with respect to \((X, \tau)\) equals the Subset Instability \(I(X, \tau ; u, \mathcal {A})\).

See Section B.3 for a full proof. In the proof, we relate the maximum unhappiness of any coalition to the dual linear program to (5). To show this relation, we leverage the fact that optimal solutions to the dual program correspond to blocking pairs of agents as well as individual rationality violations.

The main takeaway from Proposition 4.3 is that Subset Instability not only measures costs to the platform, but also costs to users, in terms of the maximum amount they “leave on the table” by not negotiating an alternate arrangement amongst themselves.

Subset Instability as a distance from equilibrium. Finally, we connect Subset Instability to solution concepts for coalitional games, a general concept in game theory that includes matching with transfers as a special case. Coalitional games (also known as cooperative games) capture competition and cooperation amongst a group of agents. The core is the set of outcomes in a cooperative game such that no subset \(\mathcal {S}\) of agents can achieve higher total utility among themselves than according to the given outcome. In games where the core is empty, a natural relaxation is the strong \(\varepsilon\)-core [41], which is the set of outcomes in a cooperative game such that no subset \(\mathcal {S}\) of agents can achieve total utility among themselves that is at least \(\varepsilon\) greater than according to the given outcome.

Subset Instability can be seen as transporting the strong \(\varepsilon\)-core notion to a slightly different context. In particular, in the context of matching with transferable utilities, the core is exactly the set of stable matchings; since a stable matching always exists, the core is always nonempty. Even though the core is nonempty, we can nonetheless use the strong \(\varepsilon\)-core to measure distance from the core. More specifically, it is natural to consider the smallest \(\varepsilon\) such that \((X, \tau)\) is in the strong \(\varepsilon\)-core. This definition exactly aligns with Subset Instability, thus providing an alternate interpretation of Subset Instability within the context of coalitional game theory.

4.2 Properties of Subset Instability

We now describe additional properties of our instability measure that are important for learning. We show that Subset Instability is: (i) zero if and only if the matching with transfers is stable, (ii) Lipschitz in the true utility functions, and (iii) lower bounded by the utility difference.

Proposition 4.4.

Subset Instability satisfies the following properties:

  • Subset Instability is always nonnegative and is zero if and only if \((X, \tau)\) is stable.

  • Subset Instability is Lipschitz continuous with respect to agent utilities. That is, for any possible market outcome \((X, \tau)\) and any pair of utility functions u and \(\tilde{u}\) it holds that: \(\begin{equation*} |I(X, \tau ; u, \mathcal {A}) - I(X, \tau ; \tilde{u}, \mathcal {A})| \le 2\sum _{a\in \mathcal {A}} {\Vert }u_a- \tilde{u}_a{\Vert }_\infty . \end{equation*}\)

  • Subset Instability is always at least the utility difference.

We defer the proof to Section B.4.

These three properties show that Subset Instability is useful as a regret measure for learning stable matchings. The first property establishes that Subset Instability satisfies the basic desideratum of having zero instability coincide with exact stability. The second property shows that Subset Instability is robust to small perturbations to the utility functions of individual agents. The third property ensures that, when learning using Subset Instability as a loss function, the platform learns a socially optimal matching.

Note that the second property already implies the existence of an explore-then-commit algorithm that achieves \(\widetilde{O}(N^{4/3} T^{2/3})\) regret in the simple setting where \(\mathcal {A}^t= \mathcal {A}\) for some \(\mathcal {A}\) of size N for all t.9 In the next section, we will explore algorithms that improve the dependence on the number of rounds T to \(\sqrt {T}\) and also work in more general settings.

Skip 5REGRET BOUNDS Section

5 REGRET BOUNDS

In this section, we develop a general approach for designing algorithms that achieve near-optimal regret within our framework. To be precise, the platform’s regret is defined to be \(\begin{equation*} R_T= \sum _{t= 1}^TI(X^t, \tau ^t; u, \mathcal {A}^t). \end{equation*}\) While our framework bears some resemblance to the (incentive-free) combinatorial bandit problem of learning a maximum weight matching, two crucial differences differentiate our setting: (i) In each round, the platform must choose transfers in addition to a matching, and (ii) loss is measured with respect to instability rather than the utility difference. Nonetheless, we show that a suitable interpretation of “optimism in the face of uncertainty” can still apply.

Regret bounds for different preference structures. By instantiating this optimism-based approach, we derive regret bounds for the preference structures introduced in Section 3. We start with the simplest case of unstructured preferences, where we assume no structure on the utilities.

Theorem 5.1.

For preference class \(\hspace{1.111pt}\mathcal {U}_{\mathrm{unstructured}}\) (see Section 3), MatchUCB (defined in Section 5.3) incurs expected regret \(\mathbb {E}(R_T) = O(|\mathcal {A}| \sqrt {n T\log (|\mathcal {A}| T)})\), where \(n = \max _t|\mathcal {A}_t|\).

In Section 5.4, we additionally give a matching (up to logarithmic factors) lower bound showing for \(n = |\mathcal {A}|\) that such scaling in \(|\mathcal {A}|\) is indeed necessary. This demonstrates that the regret scales with \(|\mathcal {A}| \sqrt {n}\), which is superlinear in the size of the market. Roughly speaking, this bound means that the platform is required to learn a superconstant amount of information per agent in the marketplace. These results suggest that without preference structure, it is unlikely that a platform can efficiently learn a stable matching in large markets.

The next two bounds demonstrate that, with preference structure, efficient learning of a stable matching becomes possible. First, we consider typed preferences, which are purely specified by a function f mapping finitely many pairs of contexts to utilities.

Theorem 5.2.

For preference class \(\hspace{1.111pt}\mathcal {U}_{\mathrm{typed}}\) (see Section 3), MatchTypedUCB (defined in Section 5.3) incurs expected regret \(\mathbb {E}(R_T) = O({|\mathcal {C}|} \sqrt {n T\log (|\mathcal {A}|T)})\), where \(n = \max _t |\mathcal {A}_t|\).

For a fixed type space \(\mathcal {C}\), the regret bound in Theorem 5.2 scales sublinearly with the market size (captured by \(|\mathcal {A}|\) and n). This demonstrates that the platform can efficiently learn a stable matching when preferences are determined by types. In fact, the regret bound only depends on the number of agents who arrive on the platform in any round; notably, it does not depend on the total number of agents on the platform (beyond logarithmic factors).

Finally, we consider separable linear preferences, where the platform needs to learn hidden information associated with each agent.

Theorem 5.3.

For preference class \(\hspace{1.111pt}\mathcal {U}_{\mathrm{linear}}\) (see Section 3), MatchLinUCB (defined in Section 5.3) incurs expected regret \(\mathbb {E}(R_T) = O (d \sqrt {|\mathcal {A}|}\sqrt {n T\log (|\mathcal {A}| T)})\), where \(n = \max _t |\mathcal {A}_t|\).

When n is comparable to \(|\mathcal {A}|\), the regret bound in Theorem 5.3 scales linearly with the market size (captured by \(|\mathcal {A}|\)) and linearly with the dimension d. Roughly speaking, this means that the platform learns (at most) a constant amount of information per agent in the marketplace. We interpret this as indicating that the platform can efficiently learn a stable matching in large markets for separable linear preferences, although learning in this setting is more demanding than for typed preferences.

5.1 Algorithm

Following the principle of optimism, our algorithm selects at each round a stable market outcome using upper confidence bounds as if they were the true agent utilities. To design and analyze this algorithm, we leverage the fact that, in the full-information setting, stable market outcomes are optimal solutions to a pair of primal-dual linear programs whose coefficients depend on agents’ utility functions. This primal-dual perspective lets us compute a market outcome each round. A particular consequence is that any UCB-based algorithm for learning matchings in a semi-bandit setting can be transformed into an algorithm for learning both the matching and the prices.

Stable market outcomes via linear programming duality. Before proceeding with the details of our algorithm, we review how the primal-dual framework can be used to select a stable market outcome in the full information setting. Shapley and Shubik [42] show that stable market outcomes \((X, \tau)\) correspond to optimal primal-dual solutions to the following pair of primal and dual linear programs (where we omit the round index t and consider matchings over \(\mathcal {A}= \mathcal {I}\cup \mathcal {J}\)):

The primal program (P) is a linear programming formulation of the maximum weight matching problem: The Birkhoff-von Neumann theorem states that its extreme points are exactly the indicator vectors for matchings between \(\mathcal {I}\) and \(\mathcal {J}\). Each dual variable \(p_a\) in (D) can be interpreted as a price that roughly corresponds to agent a’s net utility. Specifically, given any optimal primal-dual pair \((Z, p)\), one can recover a matching \(\mu _{X}\) from the nonzero entries of Z and set transfers \(\tau _a= p_a- u_a(\mu _{X}(a))\) to obtain a stable outcome \((X, \tau)\). Moreover, any stable outcome induces an optimal primal-dual pair \((Z, p)\).

Algorithm overview. Leveraging the above primal-dual formulation of stability, we introduce a meta-algorithm MetaMatchUCB for learning stable outcomes (Algorithm 1). In each round, we compute a matching with transfers by solving the primal-dual linear programs for our upper confidence bounds: Suppose we have a collection \(\mathscr{C}\) of confidence sets \(C_{i,j}, C_{j,i}\subseteq \mathbb {R}\) such that \(u_i(j)\in C_{i,j}\) and \(u_j(i)\in C_{j,i}\) for all \((i, j) \in \mathcal {I}\times \mathcal {J}\). Our algorithm uses \(\mathscr{C}\) to get an upper confidence bound for each agent’s utility function and then computes a stable matching with transfers as if these upper confidence bounds were the true utilities (see ComputeMatch in Algorithm 2). This can be implemented efficiently if we use, e.g., the Hungarian algorithm [28] to solve (P) and (D).

5.2 Main Lemma

The key fact we need to analyze our algorithms is that Subset Instability is upper bounded by the sum of the sizes of the relevant confidence sets, assuming that the confidence sets over the utilities contain the true utilities. (In the following, we again omit the round index t.)

Lemma 5.4.

Suppose a collection of confidence sets \(\mathscr{C} = \lbrace C_{i,j}, C_{j,i} : (i, j)\in \mathcal {I}\times \mathcal {J}\rbrace\) is such that \(u_{i}(j) \in C_{i,j}\) and \(u_{j}(i) \in C_{j,i}\) for all \((i, j)\). Then, the instability of \((X^\mathrm{UCB}, \tau ^{\mathrm{UCB}}):={\rm C}{\rm\small{OMPUTE}}{\rm M}{\rm\small{ATCH}}(\mathscr{C})\) satisfies (6) \(\begin{equation} I(X^\mathrm{UCB}, \tau ^\mathrm{UCB}; u, \mathcal {A}^t) \le \sum _{a\in \mathcal {A}} \Bigl (\max \bigl (C_{a, \mu _{X^\mathrm{UCB}}(a)}\bigr) - \min \bigl (C_{a, \mu _{X^\mathrm{UCB}}(a)}\bigr)\Bigr). \end{equation}\)

Proof.

Since \((X^\mathrm{UCB}, \tau ^\mathrm{UCB})\) is stable with respect to \(u^{\mathrm{UCB}}\), we have \(I(X^\mathrm{UCB}, \tau ^\mathrm{UCB}; u^{\mathrm{UCB}}, \mathcal {A}^t) = 0\). Thus, it is equivalent to bound the difference \(I(X^\mathrm{UCB}, \tau ^\mathrm{UCB}; u, \mathcal {A}^t) - I(X^\mathrm{UCB}, \tau ^\mathrm{UCB}; u^{\mathrm{UCB}}, \mathcal {A}^t)\).

At this stage, it might be tempting to bound this difference using the Lipschitz continuity of Subset Instability (see Proposition 4.4). However, this would only allow us to obtain an upper bound of the form \(\sum _{a\in \mathcal {A}} \max _{a^{\prime } \in \mathcal {A}} (\max (C_{a, a^{\prime }}) - \min (C_{a, a^{\prime }}))\). The problem with this bound is that it depends on the sizes of the confidence sets for all pairs of agents, including those that are not matched in \(X^\mathrm{UCB}\), making it too weak to prove regret bounds for UCB-style algorithms.10 Thus, we proceed with a more fine-grained analysis.

Define the function \(\begin{equation*} f(\mathcal {S}, X, \tau ; u) = \left(\max _{X^{\prime }\in \mathscr{X}_\mathcal {S}} \sum _{a\in \mathcal {S}} u_a(\mu _{X^{\prime }}(a)) \right) - \left(\sum _{a\in \mathcal {S}} u_a(\mu _{X}(a)) + \tau _a\right). \end{equation*}\) By definition, \(I(X, \tau ; u, \mathcal {A}) = \max _{\mathcal {S}\subseteq \mathcal {A}} f(\mathcal {S}, X, \tau ; u)\). It follows that \(\begin{multline*} I(X^\mathrm{UCB}, \tau ^\mathrm{UCB}; u, \mathcal {A}^t) - I(X^\mathrm{UCB}, \tau ^\mathrm{UCB}; u^{\mathrm{UCB}}, \mathcal {A}^t) \\ \le \max _{\mathcal {S}\subseteq \mathcal {A}} (f(\mathcal {S}, X^\mathrm{UCB}, \tau ^\mathrm{UCB}; u) - f(\mathcal {S}, X^\mathrm{UCB}, \tau ^\mathrm{UCB}; u^{\mathrm{UCB}})). \end{multline*}\) To finish, we upper bound \(f(\mathcal {S}, X^\mathrm{UCB}, \tau ^\mathrm{UCB}; u) - f(\mathcal {S}, X^\mathrm{UCB}, \tau ^\mathrm{UCB}; u^{\mathrm{UCB}})\) for each \(\mathcal {S}\subseteq \mathcal {A}\). We decompose this expression into two terms: \(\begin{align*} f(\mathcal {S}, X^\mathrm{UCB}, \tau ^\mathrm{UCB}; &u) - f(\mathcal {S}, X^\mathrm{UCB}, \tau ^\mathrm{UCB}; u^\mathrm{UCB}) \\ &= \underbrace{\left(\max _{X^{\prime }\in \mathscr{X}_\mathcal {S}} \sum _{a\in \mathcal {S}} u_a(\mu _{X^{\prime }}(a)) - \max _{X^{\prime }\in \mathscr{X}_\mathcal {S}} \sum _{a\in \mathcal {S}} u^{\mathrm{UCB}}_a(\mu _{X^{\prime }}(a)) \right)}_{(\text{A})} \\ &\qquad \qquad + \underbrace{\left(\sum _{a\in \mathcal {S}} \bigl (u^{\mathrm{UCB}}_a(\mu _{X^\mathrm{UCB}}(a)) + \tau ^\mathrm{UCB}_a\bigr) - \sum _{a\in \mathcal {S}} \bigl (u_a(\mu _{X^\mathrm{UCB}}(a)) + \tau ^\mathrm{UCB}_a\bigr) \right)}_{(\text{B})}. \end{align*}\) To see that (A) is nonpositive, observe that the maximum weight matching of \(\mathcal {S}\) with respect to u is no larger than the maximum weight matching of \(\mathcal {S}\) with respect to \(u^{\mathrm{UCB}}\), since \(u^{\mathrm{UCB}}\) pointwise upper bounds u. To upper bound (B), observe that the transfers cancel out, so the expression is equivalent to

5.3 Instantiations of the Meta-algorithm

As formalized in MetaMatchUCB, the regret bound of Lemma 5.4 suggests a simple approach: At each round, select the matching with transfers returned by ComputeMatch and update confidence sets accordingly. To instantiate MetaMatchUCB, it remains to construct confidence intervals that contain the true utilities with high probability. This last step naturally depends on the assumptions made about the utilities and the noise.

Unstructured preferences. For this setting, we construct confidence intervals following the classical UCB approach: For each utility value involving the pair \((i, j)\in \mathcal {I}\times \mathcal {J}\), we take a confidence interval of length \(O(\sqrt {\log (|\mathcal {A}|T) / \smash{n_{ij}}})\) centered at the empirical mean, where \(n_{ij}\) is the number of times the pair has been matched thus far. We describe this construction precisely in Algorithm 3 (MatchUCB).

To analyze MatchUCB, recall that Lemma 5.4 bounds the regret at each step by the lengths of the confidence intervals of each pair in the selected matching. Bounding the lengths of the confidence intervals parallels the analysis of UCB for classical stochastic multi-armed bandits. We give the full proof of Theorem 5.1 in Section C.1.

Typed Preferences. For this setting, we construct our confidence intervals as follows: For each pair of types \(c_1\) and \(c_2\), we take a length \(O(\sqrt {\log (|\mathcal {A}|T) / \smash{n_{c_1c_2}}})\) confidence interval centered around the empirical mean, where \(n_{c_1c_2}\) is the number of times that an agent with type \(c_1\) has been matched with an agent with type \(c_2\). We describe this construction precisely in Algorithm 4 (MatchTypedUCB). We give the full proof of Theorem 5.2 in Section C.2.

Separable Linear Preferences. To build the confidence sets, we use a key idea from the design of LinUCB [30, 39]. The idea is to compute a confidence set for each hidden vector \(\phi (a)\) using the least squares estimate and use that to construct confidence sets \(C_{a,a^{\prime }}\) for the utilities.

More formally, let \(\mathcal {T}_{a}\) be the set of rounds where agent a is matched on the platform thus far, and for \(t^{\prime } \in \mathcal {T}_a\), let \(\mathcal {R}_{a,t^{\prime }}\) be the observed utility at time \(t^{\prime }\) for agent a. The center of the confidence set will be given by the least squares estimate \(\begin{equation*} \phi ^{\mathrm{LS}}(a) = \text{arg min}_{v \in \mathcal {B}^d} \left(\sum _{t^{\prime } \in \mathcal {T}_{a}} (\langle v, c_{\mu _{X_{t^{\prime }}}(a)} \rangle - \mathcal {R}_{a,t^{\prime }} \right). \end{equation*}\) The confidence set for \(\phi (a)\) is given by \(\begin{equation*} C_{\phi (a)} := \left\lbrace v \Biggm \vert \sum _{t^{\prime } \in \mathcal {T}_{a,t}} \Big \langle v - \phi ^{\mathrm{LS}}(a), c_{\mu _{X_{t^{\prime }}}(a)} \Big \rangle ^2 \le \beta \text{ and } \Vert v\Vert _2\le 1 \right\rbrace , \end{equation*}\) where \(\beta = O(D \log T + \frac{n_{a} \sqrt {\ln (n_{a} /\delta)}}{T^2})\) and \(n_{a}\) counts the number of times that a has appeared in selected matchings. The confidence set for \(u(a, a^{\prime })\) is given by \(\begin{equation*} C_{a, a^{\prime }} := \left\lbrace \langle c_{a^{\prime }}, v \rangle \mid v \in C_{\phi (a)} \right\rbrace \cap [-1, 1]. \end{equation*}\) We describe this construction precisely in Algorithm 5 (MatchLinUCB). We give the full proof of Theorem 5.3 in Section C.3.

5.4 Matching Lower Bound

For the case of unstructured preferences, we now show that MatchUCB achieves optimal regret (up to logarithmic factors) by showing a lower bound that (nearly) matches the upper bound in Theorem 5.1.

Lemma 5.5.

For any algorithm that learns a stable matching with respect to unstructured preferences, there exists an instance on which it has expected regret \(\widetilde{\Omega }(|A|^{3/2} \sqrt {T})\) (where regret is given by Subset Instability).

The idea behind this lemma is to show a lower bound for the easier problem of learning a maximum weight matching using utility difference as regret. By Proposition 4.4, this immediately implies a lower bound for learning a stable matching with regret measured by Subset Instability.

This lower bound illustrates the close connection between our setting and that of learning a maximum weight matching. Indeed, by applying MatchUCB and simply disregarding the transfers every round, we recover the classical UCB-based algorithm for learning the maximum weight matching [15, 22, 29]. From this perspective, the contribution of MatchUCB is an approach to set the dual variables while asymptotically maintaining the same regret as the primal-only problem.

Skip 6EXTENSIONS Section

6 EXTENSIONS

In this section, we discuss several extensions of our results: instance-dependent regret bounds, connections between subset instability and platform revenue, and non-transferable utilities. These extensions illustrate the generality of our framework and also suggest several avenues for future research.

In Section 6.1, we derive instance-dependent regret bounds for Subset Instability, which allow us to improve the \(O(\sqrt T)\) convergence from Section 5 to \(O(\log T)\) for any fixed instance. Achieving this logarithmic bound involves choosing “robust” dual solutions when setting transfers (rather than choosing an arbitrary optimal primal-dual pair as in ComputeMatch): We want our selected primal-dual pair to lead to stable outcomes even under perturbations of the transfers.

In Section 6.2, we connect the subsidy perspective of Subset Instability to platform revenue. We relate regret to platform revenue and show that, when there are search frictions, the platform can achieve substantial long-run profit despite starting with no knowledge of agent preferences.

In Section 6.3, we adapt our framework to matching with non-transferable utilities (where agents do not transfer money to other agents on the platform). We define an analogue of Subset Instability using the subsidy formulation and give an \(\widetilde{O}(\sqrt {T})\) regret algorithm for learning stable matchings.

6.1 Instance-dependent Regret Bounds

While our analyses in Section 5.1 focused on bounds that hold uniformly for all problem instances, we now explore instance-dependent regret bounds. Instance-dependent bounds capture a different facet of bandit algorithms: How does the number of mistakes made by the algorithm scale on each instance with respect to T? Bounds of this nature have been explored in previous works [9, 12, 32, 33, 40] on learning stable matchings in the non-transferable utilities setting, and we show that they can be obtained within our framework as well.

Our instance-dependent regret bound depends on a gap \(\Delta \gt 0\) determined by the true utility function u. We focus on the setting where agent utilities are unstructured (i.e., \(u\in \hspace{1.111pt}\mathcal {U}_{\mathrm{unstructured}}\)) and where the same set of agents \(\mathcal {A}\) arrives in each round. As is common in analyses of combinatorial bandit problems (e.g., References [15, 29]), the gap \(\Delta\) in the bound is global to the matching. Letting \(X^{\text{opt}}\) be a maximum weight matching with respect to u, we define the gap to \(\Delta\) be the difference in utility between the optimal and second-best matchings:11 \(\begin{equation*} \Delta := \inf _{X\ne X^{\text{opt}}} \Biggl \lbrace \sum _{a\in \mathcal {A}} u_a(\mu _{X^{\text{opt}}}(a)) - \sum _{a\in \mathcal {A}} u_a(\mu _{X}(a)) \Biggr \rbrace . \end{equation*}\) We prove the following regret bound:

Theorem 6.1 (Instance-Dependent Regret).

Suppose that \(\mathcal {A}_t= \mathcal {A}\) for all t. Let \(u\in \hspace{1.111pt}\mathcal {U}_{\mathrm{unstructured}}\) be any utility function and put \(\begin{equation*} \Delta := \inf _{X\ne X^*} \Biggl \lbrace \sum _{a\in \mathcal {A}} u_a(\mu _{X^*}(a)) - \sum _{a\in \mathcal {A}} u_a(\mu _{X}(a)) \Biggr \rbrace . \end{equation*}\) Then, MatchUCB\(^{\prime }\) incurs expected regret \(\mathbb {E}(R_T) = O(|\mathcal {A}|^5 \cdot \log (|\mathcal {A}|T)/\Delta ^2)\).

MatchUCB\(^{\prime }\) is MatchUCB with a slight adjustment to ComputeMatch needed to prove Theorem 6.1. MatchUCB\(^{\prime }\), like MatchUCB, does not depend on the gap \(\Delta\) and achieves the instance-independent regret bound in Theorem 5.1.12 That is, MatchUCB\(^{\prime }\) achieves both our instance-independent and instance-dependent regret bounds.

Our starting point for proving Theorem 6.1 is to upper bound the number of “mistakes” that a platform makes while exploring and learning, i.e., the number of rounds where the chosen matching is suboptimal. That is, we bound the number of rounds where the chosen market outcome is not stable with respect to the true utilities u. This is similar in spirit to the analysis of the combinatorial bandits problem of learning a maximum weight matching in Chen et al. [15]. However, a crucial difference is that a mistake can be incurred even when the selected matching is optimal if the selected transfers do not result in a stable market outcome. Ensuring that the selected transfers result in a stable market outcome when the utility estimates are sufficiently accurate is the main technical hurdle in our analysis.

To make this argument work, we need to specify more precisely how the primal-dual solution is chosen in line 5 of ComputeMatch (which we previously did not specify). In particular, poor choices of the primal-dual solution can lead to many rounds where the chosen outcome is unstable, because the transfers violate the stability constraints. To see this, consider a market with a single customer C and a single provider P such that \(u_C(P) = 2\) and \(u_P(C) = -1\), and suppose we have nearly tight upper bounds \(u^{\mathrm{UCB}}_C(P) = 2 + \varepsilon\) and \(u^{\mathrm{UCB}}_P(C) = -1 + \varepsilon\) on the utilities. Then, the market outcome with matching \(\lbrace (C, P)\rbrace\) with \(\tau _C = -2 - \varepsilon\) and \(\tau _P = -\tau _C\) could be selected by ComputeMatch, since it corresponds to an optimal primal-dual pair for \(u^{\mathrm{UCB}}\). However, it is not stable with respect to the true utilities u (as individual rationality is violated for C), regardless of how small \(\varepsilon\) is. Thus, without assuming more about how the optimal primal-dual pair is chosen in ComputeMatch, we cannot hope to bound the number of unstable market outcomes selected.

We show that, by carefully selecting an optimal primal-dual pair each round, we can bound the number of mistakes. In particular, we design an algorithm ComputeMatch\(^{\prime }\) to find primal-dual pairs that satisfy the following property: If the confidence sets are small enough, then the selected matching will be stable with respect to the true utilities.

Lemma 6.2.

Suppose ComputeMatch\(^{\prime }\) is run on a collection \(\mathscr{C}\) of confidence sets \(C_{i,j}\) and \(C_{j,i}\) over the agent utilities that satisfy \(\begin{equation*} \max (C_{i,j}) - \min (C_{i,j}) \le 0.05{} \frac{\Delta }{|\mathcal {A}|} \quad \text{and}\quad \max (C_{j, i}) - \min (C_{j, i}) \le 0.05{} \frac{\Delta }{|\mathcal {A}|} \end{equation*}\) for all \((i, j)\) in the matching returned by ComputeMatch\(^{\prime }\). Suppose also that the confidence sets \(\mathscr{C}\) contain the true utilities for all pairs of agents. Then, the market outcome returned by ComputeMatch\(^{\prime }\) is stable with respect to the true utilities u.

Lemma 6.2 does not hold for ComputeMatch; its proof relies on the particular specification of the optimal primal-dual pair in ComputeMatch\(^{\prime }\).

Using Lemma 6.2, we intuitively can bound the number of mistakes made by the algorithm by the number of samples needed to sufficiently reduce the size of the confidence sets. In Appendix D, we describe how we choose optimal primal-dual pairs in ComputeMatch\(^{\prime }\), prove Lemma 6.2, and provide a full proof of Theorem 6.1.

Theorem 6.1 opens the door to further exploring algorithmic properties of learning stable matchings. First, this result establishes fine-grained regret bounds, demonstrating the typical \(O(\log T)\) regret bounds from the combinatorial bandits literature [15] are achievable in our setting as well. Second, Theorem 6.1 provides insight into the number of mistakes made by the platform. In particular, we show within the proof of Theorem 6.1 that the platform fails to choose a matching that is stable with respect to u in at most \(O(|\mathcal {A}|^4 \cdot \log (|\mathcal {A}|T)/\Delta ^2)\) rounds.13 This means that the platform selects a stable matching in at least \(T - O(|\mathcal {A}|^5 \cdot \log (|\mathcal {A}|T)/\Delta ^2) = T - O(\log T)\) of the rounds.

As we described, our bounds in Theorem 6.1 rely on choosing an appropriate primal-dual solution. An interesting direction for future work would be to provide further insight into how different methods for finding optimal primal-dual pairs affect both regret bounds and the trajectory of the selected market outcomes over time.

6.2 Search Frictions and Platform Revenue

Next, we further ground Subset Instability by explicitly connecting it to the platform’s revenue under a stylized economic model of search frictions. A major motivation for this is that it helps explain when an online platform can earn a profit in competitive settings, even when they start out with no information about agent preferences.

More specifically, we incorporate search frictions where an agent must lose utility \(\varepsilon\) to find an alternative to the given match (e.g., from the time spent finding an alternate partner or from a cancellation fee). These search frictions weaken the requirements for stability: The platform now only needs matchings to be \(\varepsilon\)-stable: \(\begin{equation*} u_i(j) + u_j(i) - 2\varepsilon \le u_i(\mu _{X}(i)) + \tau _i+ u_j(\mu _{X}(j)) + \tau _j \end{equation*}\) for all \((i, j) \in \mathcal {I}\times \mathcal {J}\) and \(u_a(\mu _{X}(a)) + \tau _a\ge -\varepsilon\) for all \(a\in \mathcal {A}\).14

To model revenue, we take the subsidy perspective on Subset Instability. Specifically, recall that Subset Instability is equal to the minimum subsidy needed to maintain stability (see Proposition 4.2). With search frictions, that subsidy can potentially be negative, thus allowing the platform to generate revenue. We are interested in analyzing the maximum revenue (minimum subsidy) the platform can generate while ensuring stability with high probability over all rounds. For realism, we also want this subsidy to be computed online using only information that the platform has access to, but it turns out we can do this with minimal modifications to our algorithm.

More formally, in this modified model, the platform must select an \(\varepsilon\)-stable matching in each round with high probability by choosing appropriate subsidies. That is, in round t, the platform selects a matching with transfers \((X^t, \tau ^t)\) with the modification that the transfers need not be zero-sum. The transfers thus incorporate the amount that platform is subsidizing or charging agents for participation on the platform. The net profit of the platform is then \(-\sum _{t=1}^{T} \sum _{a\in \mathcal {A}} \tau _a^t\). We impose the stability requirement that \(\begin{equation*} \mathbb {P}[(X^t, \tau ^t) \text{ is $\varepsilon $-stable for all } 1 \le t\le T] \ge 0.99. \end{equation*}\) Given this setup, we show the following:

Theorem 6.3.

For preference class \(\hspace{1.111pt}\mathcal {U}_{\mathrm{unstructured}}\) (see Section 3), there exists an algorithm giving the platform \(\begin{equation*} \varepsilon T \sum _{t=1}^T |\mathcal {A}_t| - O\Bigl (|\mathcal {A}| \sqrt {n T} \sqrt {\log (|\mathcal {A}| |T|)}\Bigr) \end{equation*}\) revenue in the presence of search frictions while maintaining stability with high probability.

In particular, if \(\mathcal {A}_t = \mathcal {A}\) in every round, then the platform will starting making a profit within \(O(|\mathcal {A}|/\varepsilon ^2\cdot \log (|\mathcal {A}|/\varepsilon ^2))\) rounds.

We defer the proof of Theorem 6.3 to Appendix E.

Qualitatively, Theorem 6.3 captures that if the platform “pays to learn” in initial rounds, then the information that it obtains will help it achieve a profit in the long run. We note that both the revenue objective and the model for search frictions that we consider in these preliminary results are stylized. An interesting direction for future work would be to integrate more realistic platform objectives and models for search frictions into the framework.

6.3 Matching with Non-transferable Utilities

While we have focused on matching with transferable utilities, utilities are not always transferable in practice, as in the cases of dating markets and college admissions (i.e., most people are not willing to date an undesirable partner in exchange for money, and a typical college admission slot is not sold for money). We can extend our findings to this setting following the model of matching with non-transferable utilities (NTU) [23], which has also been studied in previous work [12, 17, 32, 40]. The definition of Subset Instability extends naturally and has advantages over the “utility difference” metric that is commonly used in prior work. Our algorithmic meta-approach also sheds new light on the convergence properties of the centralized UCB algorithm of Liu et al. [32].

The starting point of our instability measure is slightly different than in Section 4. Since stable matchings in the NTU model need not maximize total utility, we cannot define instability based on a maximum over all subsets of agents of the utility difference for that subset. However, the subsidy formulation of Subset Instability (see (4)) translates well to this setting. Our instability measure will correspond to the minimum amount the platform could subsidize agents so individual rationality holds and no blocking pairs remain. For matching with NTU, we formalize this notion as follows:

Definition 6.4

(NTU Subset Instability).

For utilities u and agents \(\mathcal {A}\), the NTU Subset Instability \(I(X; u, \mathcal {A})\) of a matching X is \(\begin{align*} \min _{s\in \mathbb {R}^{|\mathcal {A}|}} & \sum _{a\in \mathcal {A}} s_{a} & {(\dagger)} \\ \text{s.t.}\ \ & \min \bigl (u_i(j) - u_i(\mu _{X}(i)) - s_i, u_j(i) - u_j(\mu _{X}(j)) - s_j\bigr)\le 0 \qquad &&\forall (i,j)\in \mathcal {I}\times \mathcal {J}\\ & u_a(\mu _{X}(a)) + s_a\ge 0\quad &&\forall a\in \mathcal {A}\\ &s_a\ge 0 \quad &&\forall a\in \mathcal {A}. \end{align*}\)

NTU Subsidy Instability inherits some of the same appealing properties as Subsidy Instability.

Proposition 6.5 (Informal).

NTU Subset Instability satisfies the following properties:

(1)

NTU Subset Instability is always nonnegative and is zero if and only if \((X, \tau)\) is stable.

(2)

NTU Subset Instability is Lipschitz continuous with respect to agent utilities. That is, for any matching X and any pair of utility functions u and \(\tilde{u}\), it holds that: \(\begin{equation*} |I(X; u, \mathcal {A}) - I(X; \tilde{u}, \mathcal {A})| \le 2\sum _{a\in \mathcal {A}} {\Vert }u_a- \tilde{u}_a{\Vert }_\infty . \end{equation*}\)

The proofs of this and subsequent results are deferred to Appendix F. Together, the preceding properties mean that NTU Subsidy Instability is useful as a regret measure for learning stable matchings.

As in the transferable utilities setting, Property 2 implies the existence of an explore-then-commit algorithm with \(\widetilde{O}(|A|^{4/3} T^{2/3})\) regret. We show that this can be improved to a \(\sqrt {T}\) dependence by adapting our approach from Section 5:

Theorem 6.6.

For matchings with non-transferable utilities, there exists an algorithm that for any utility function u incurs regret \(R_T= O({|\mathcal {A}|}^{3/2} \sqrt {T} \sqrt {\log (|\mathcal {A}| T)})\).

While Theorem 6.6 illustrates that our approach easily generalizes to the NTU setting, we highlight two crucial differences between these settings. First, learning a stable matching is incomparable to learning a maximum weight matching, because stable matchings do not maximize the sum of agents’ utilities in the NTU setting. Next, the instability measure is not equivalent to the cumulative unhappiness of agents, unlike in the setting with transferable utilities. Intuitively, these definitions cease to be equivalent, because non-transferable utilities render the problem more “discontinuous” and thus obstruct the duality results we applied earlier.

These results provide a preliminary application of our framework to the setting of matching with non-transferable utilities; an interesting direction for future inquiry would be to more thoroughly investigate notions of approximate stability and regret in this setting.

6.3.1 Comparison to the Utility Difference Measure.

It turns out that the algorithm underlying Theorem 6.6 is equivalent to the centralized UCB algorithm from previous work [12, 32], albeit derived from a different angle. However, an important difference is that Theorem 6.6 guarantees low regret relative to the incentive-aware NTU Subset Instability, as opposed to the incentive-unaware “utility difference” measure in prior work. In this section, we outline several properties that make our instability measure more suitable especially in the NTU setting. In particular, we show for utility difference that:

(a)

There is no canonical formalization of utility difference when multiple stable matchings exist.

(b)

The utility difference of a matching can be positive even if the matching is stable and negative even if the matching is unstable.

(c)

Even when restricting to markets with unique stable matchings, the utility difference of a matching can be discontinuous in the true agent utilities. As a result, it does not allow for instance-independent regret bounds that are sublinear in T.

For (a), the utility difference requires specifying a stable matching to serve as a benchmark against which to measure relative utility. However, when multiple stable matchings exist, some ambiguity arises as to which one should be chosen as the benchmark. Because of this, previous works [12, 17, 32, 40] study two different benchmarks. In particular, they assume providers’ preferences are known and benchmark with respect to the customer-optimal and customer-pessimal stable matchings. For (b), notice that the utility difference for the maximum weight matching is negative, even though it is typically not stable in the NTU setting. Moreover, because of the ambiguity in the benchmark from (a), the utility difference may not be zero even when the matching is stable. For (c), to see that utility difference is not continuous as a function of the underlying agent utilities, consider the following example:

Example 6.7.

Consider a market where there is a single customer i and two providers \(j_1\) and \(j_2\). Suppose their utility functions are given by \(u_{i}(j_1) = \varepsilon\), \(u_{i}(j_2) = 2\varepsilon\), \(u_{j_1}(i) = 1\), and \(u_{j_2}(i) = 0.5\). Then, the unique stable matching \(\lbrace (i, j_2)\rbrace\) has total utility \(0.5 + 2\varepsilon\). Now, consider the perturbed utility function \(\tilde{u}\) such that \(\tilde{u}_{i}(j_1) = 2\varepsilon\), \(\tilde{u}_{i}(j_2) = \varepsilon\), \(\tilde{u}_{j_1}(i) = 1\), and \(\tilde{u}_{j_2}(i) = 0.5\). For this perturbed utility function, the unique stable matching is \(\lbrace (i, j_1)\rbrace\), which has total utility \(1 + 2\varepsilon\). The utility difference (either optimal or pessimal) for matching \(\lbrace (i, j_2)\rbrace\) is 0 for u and \(0.5 + \varepsilon\) for \(\tilde{u}\). Since this holds for any \(\varepsilon \gt 0\), taking \(\varepsilon \rightarrow 0\) shows that utility difference is not continuous in the utility function.

That utility difference is discontinuous in agent utilities rules out the existence of bandit algorithms that achieve sublinear instance-independent regret when using utility difference as the regret measure. In particular, the analyses in previous work [12, 32, 33, 40] focus entirely on instance-dependent regret bounds. They show that centralized UCB achieves logarithmic instance-dependent regret with respect to the utility difference relative to the customer-pessimal stable matching (but does not achieve sublinear regret with respect to the customer-optimal stable matching). Our insight here is that a new measure of instability can present a more appealing evaluation metric and paint a clearer picture of an algorithm’s convergence to the set of stable matchings as a whole.

Skip 7IN WHAT SETTINGS ARE EQUILIBRIA LEARNABLE? Section

7 IN WHAT SETTINGS ARE EQUILIBRIA LEARNABLE?

A core insight of our work is that, in a stochastic environment, “optimism in the face of uncertainty” can be effectively leveraged for the problem of learning stable matchings. This motivates us to ask: In what other settings, and with what other algorithmic methods, can equilibria be learned?

One interesting open direction is to understand when equilibria can be learned in adversarial environments where the utility functions can change between rounds. From an economic perspective, adversarial environments could capture evolving market conditions. In the adversarial bandit setting, most work relies on gradient-based algorithms instead of UCB-based algorithms to attain optimal regret bounds (see, e.g., References [1, 6]). Can these gradient-based algorithms similarly be adapted to Subset Instability?

Another interesting open direction is to consider more general market settings, even in stochastic environments. For example, within the context of matching markets, each agent might match to more than one agent on the other side of the market; and outside of matching markets, a buyer might purchase multiple units of multiple goods. In markets with transferable utilities, incentive-aligned outcomes can be captured by Walrasian equilibria (see, e.g., Bichler et al. [10]). Can Subset Instability and our UCB-based algorithms be adapted to learning Walrasian equilibria in general?

Addressing these questions would provide a richer understanding of when and how large-scale, data-driven marketplaces can efficiently learn market equilibria.

APPENDICES

Skip ACLASSICAL RESULTS FOR MATCHING WITH TRANSFERABLE UTILITIES Section

A CLASSICAL RESULTS FOR MATCHING WITH TRANSFERABLE UTILITIES

To be self-contained, we briefly state and prove the key results from Shapley and Shubik [42] we need.

First, we explicitly relate the primal-dual formulation in Section 5.1 to stable matchings.

Theorem A.1 ([42]).

If \((X, \tau)\) is stable, then \((Z, p)\) is an optimal primal-dual pair to (P) and (D), where \(p_a= \tau _a+ u_a(X(a))\) and Z is the indicator matrix in \(\mathbb {R}^{\mathcal {I}\times \mathcal {J}}\) corresponding to X.

Moreover, if \((Z, p)\) is an optimal primal-dual pair to (P) and (D) such that Z lies at an extreme point of the feasible set, then \((X, \tau)\) is stable where \(\tau _a= p_a- u_a(X(a))\) and X is the matching corresponding to the nonzero entries of Z.

Proof.

Both statements follow from the complementary slackness conditions and the definition of stability in Definition 2.1. The complementary slackness conditions are:

  • If \(Z_{i, j} \gt 0\), then \(p_i+ p_j= u_i(j) + u_j(i)\).

  • If \(p_i\gt 0\), then \(\sum _jZ_{i, j} = 1\).

  • If \(p_j\gt 0\), then \(\sum _iZ_{i, j} = 1\).

Suppose that \((X, \tau)\) is stable. Let us first show that \((Z, p)\) is feasible. We see that Z is primal feasible by definition. For dual feasibility, since there are no blocking pairs, we know that \(\begin{equation*} \bigl (u_i(\mu _{X}(i)) + \tau _i\bigr) + \bigl (u_j(\mu _{X}(j)) + \tau _j\bigr)\ge u_i(j) + u_j(i), \end{equation*}\) which implies \(\begin{equation*} p_i+ p_j\ge u_i(j) + u_j(i). \end{equation*}\) The individual rationality condition \(u_a(\mu _{X}(a)) + \tau _a\ge 0\) tells us \(p_a\ge 0\). Hence, p is dual feasible. Next, we show that \((Z, p)\) is an optimal primal-dual pair by checking the Karush–Kuhn–Tucker conditions. We have already shown primal and dual feasibility, so it suffices to show complementary slackness. The first condition follows from zero-sum transfers. To see the second and third conditions, we show the contrapositive: If \(i\in \mathcal {I}\) is such that \(\sum _jZ_{i,j} \lt 1\), then \(\sum _jZ_{i,j} = 0\) by our assumption on Z. Hence, i is unmatched (i.e., \(u_i(\mu _{X}(i)) = 0\) and \(\tau _i= 0\)), which implies \(p_i= 0\). The analogous argument applies for \(j\in \mathcal {J}\).

We now prove the second part of the theorem. Suppose \((Z, p)\) is an optimal solution to (P) and (D) such that Z is at a vertex. By the Birkhoff-von Neumann theorem, since Z is a vertex, it corresponds to a matching. We wish to show that \((X, \tau)\) has no blocking pairs, is individually rational, and has zero-sum transfers. Dual feasibility tells us that: \(\begin{equation*} p_i+ p_j\ge u_i(j) + u_j(i), \end{equation*}\) which means that: \(\begin{equation*} (u_i(\mu _{X}(i)) + \tau _i) + (u_j(\mu _{X}(j)) + \tau _j)\ge u_i(j) + u_j(i), \end{equation*}\) so there are no blocking pairs. Dual feasibility also tells us that \(p_a\ge 0\), which means that \(u_a(\mu _{X}(a)) + \tau _a\ge 0\), so individual rationality is satisfied. To show that there are zero-sum transfers, we use complementary slackness. The first complementary slackness condition tells us that if \(Z_{i, j} \gt 0\), then \(p_i+ p_j= u_i(j) + u_j(i)\). Using the fact that Z corresponds to a matching, this in particular means that if \((i, j) \in X\), then we know \(\tau _i+ \tau _j= 0\). To show that agents who are unnmatched receive 0 transfers, let us use the second and third complementary slackness conditions. The contrapositive tells us that if a is unmatched, then \(p_a= 0\), which implies \(\tau _a= 0\).□

Since (P) is exactly the maximum weight matching linear program, Theorem A.1 immediately tells us that if \((X, p)\) is stable, then X is a maximum weight matching. This means that stable matchings with transferable utilities maximize social welfare.

Skip BProofs for Section 4 Section

B Proofs for Section 4

This section contains further exposition (including proofs) for Section 4.

B.1 Limitations of Utility Difference as an Instability Measure

To illustrate why utility difference fails to be a good measure of instability, we describe a matching with transfers that (i) is far from stable, and (ii) has zero utility difference (but large Subset Instability).

Example B.1.

Consider the following market with two agents: \(\mathcal {I}= \lbrace i\rbrace\) and \(\mathcal {J}= \lbrace j\rbrace\). Suppose that \(u_i(j) = 2\) and \(u_j(i) = -1\). Consider the matching \(X= \lbrace (i, j)\rbrace\) with transfers \(\tau _i= -\xi\) and \(\tau _j= \xi\) for some \(\xi \gt 0\). We will show that this matching with transfers will have the properties stated above when \(\xi\) is large.

This matching with transfers has a utility difference equal to zero (for any \(\xi\)), since it maximizes the sum of utilities. Indeed, it is stable for any \(\xi \in [1, 2]\). However, when \(\xi \gt 2\), this matching with transfers is no longer stable, since the individual rationality condition \(u_i(j) + \tau _i\ge 0\) fails. (Intuitively, the larger \(\xi\) is, the further we are from stability.) But its utility difference remains at zero.

However, the Subset Instability of this matching with transfers is \(\xi - 2 \gt 0\) when \(\xi \gt 2\). In particular, Subset Instability increases with \(\xi\) in this regime, which is consistent with the intuition that outcomes with larger \(\xi\) should be more unstable.

B.2 Proof of Proposition 4.2

Proposition 4.2. Minimum stabilizing subsidy equals Subset Instability for any market outcome.

Proof of Proposition 4.2

We can take the dual of the linear program (5) to obtain:

By strong duality, the optimal values of (5) and (\(\ddagger\)) are equal. Thus, it suffices to show that Subset Instability is equal to (\(\ddagger\)). By Proposition 4.3, we know that Subset Instability is equal to the maximum unhappiness of any coalition. Thus, it suffices to show that (\(\ddagger\)) is equal to the maximum unhappiness of any coalition.

To interpret (\(\ddagger\)), observe that there exist optimal \(S^*\) and \(Z^*\) all of whose entries lie in \(\lbrace 0,1\rbrace\), because this linear program can be embedded into a maximum weight matching linear program. Take such a choice of optimal \(S^*\) and \(Z^*\). Then, \(S^*\) is an indicator vector corresponding to a (partial) matching on a subset of the agents such that all pairs in this matching are blocking with respect to \((X, \tau)\). Similarly, \(Z^*\) is an indicator vector of agents who would rather be unmatched than match according to \((X, \tau)\).

We first prove the claim that \(I(X, \tau ; u, \mathcal {A})\) is at least (\(\ddagger\)). Based on the above discussion, the optimal objective of (\(\ddagger\)) is obtained through \(S^*\) and \(Z^*\) that represent a matching and a subset of agents, respectively. Let S be the union of agents participating in \(S^*\) and \(Z^*\). We see that the objective of (\(\ddagger\)) is equal to the utility difference at S, i.e.: \(\begin{equation*} \left(\max _{X^{\prime }\in \mathscr{X}_{S}} \sum _{a\in S} u_a(\mu _{X^{\prime }}(a)) \right) - \left(\sum _{a \in S} u_a(\mu _{X}(a)) + \tau _a)\right). \end{equation*}\) This is no larger than Subset Instability by definition.

We next prove the claim that \(I(X, \tau ; u, \mathcal {A})\) is at most (\(\ddagger\)). Let us consider \(S^*\) that maximizes: \(\begin{equation*} \max _{S \subseteq \mathcal {A}} \left(\max _{X^{\prime }\in \mathscr{X}_{S}} \sum _{a\in S} u_a(\mu _{X^{\prime }}(a)) \right) - \left(\sum _{a \in S} u_a(\mu _{X}(a)) + \tau _a)\right). \end{equation*}\) Let us take the maximum weight matching of \(S^*\). Let S be given by the matched agents in this matching and let Z be given by the unmatched agents in this matching (using the interpretation of (\(\ddagger\)) described above). We see that the objective at (\(\ddagger\)) for \((S, Z)\) is equal to Subset Instability, which proves the desired statement.□

B.3 Proof of Proposition 4.3

We first formally define the unhappiness of a coalition , as follows. In particular, the unhappiness with respect to \((X, \tau)\) of a coalition \(\mathcal {S}\subseteq \mathcal {A}\) is defined to be:

(7)
with unhappiness being 0 if there are no feasible \(X^{\prime }\) and \(\tau ^{\prime }\). In the optimization program, \((X^{\prime }, \tau ^{\prime })\) represents a matching with transfers over \(\mathcal {S}\), with the constraint \(\tau ^{\prime }_a+ \tau ^{\prime }_{\mu _{X^{\prime }}(a)} = 0\) ensuring that it is zero-sum. The objective measures the difference between \((X, \tau)\) and \((X^{\prime }, \tau ^{\prime })\) of the total utility of the agents in \(\mathcal {S}\). The constraint \(u_a(\mu _{X^{\prime }}(a)) + \tau ^{\prime }_a\ge u_a(\mu _{X}(a)) + \tau _a\) encodes the requirement that all agents be at least as well off under \((X^{\prime }, \tau ^{\prime })\) as they were under \((X, \tau)\). This optimization program therefore captures the objective of \(\mathcal {S}\) to maximize their total payoff while ensuring that no member of the coalition is worse off than they were according to \((X, \tau)\).

Recall that, in terms of unhappiness, Proposition 4.3 is as follows:

Proposition 4.3. The maximum unhappiness of any coalition \(\mathcal {S}\subseteq \mathcal {A}\) with respect to \((X, \tau)\) equals the Subset Instability \(I(X, \tau ; u, \mathcal {A})\).

Proof of Proposition 4.3

By Proposition 4.2, we know that Subset Instability is equal to (5). Moreover, by strong duality, we know that Subset Instability is equal to (\(\ddagger\)) (the dual linear program of (5)). Thus, it suffices to prove that the maximum unhappiness of any coalition is equal to (\(\ddagger\)).

We first prove the claim that (\(\ddagger\)) is at most the maximum unhappiness of any coalition with respect to \((X, \tau)\). To do this, it suffices to construct a coalition \(\mathcal {S}\subseteq \mathcal {A}\) such that (\(\ddagger\)) is at most the unhappiness of \(\mathcal {S}\). We construct \(\mathcal {S}\) as follows: Recall that there exist optimal solutions \(S^*\) and \(Z^*\) to (\(\ddagger\)) such that \(S^*\) corresponds to a (partial) matching on \(\mathcal {I}\times \mathcal {J}\) and \(Z^*\) corresponds to a subset of \(\mathcal {A}\). We may take \(\mathcal {S}\) to be the union of the agents involved in \(S^*\) and in \(Z^*\). Now, we upper bound the unhappiness of \(\mathcal {S}\) by constructing \(X^{\prime }\) and \(\tau ^{\prime }\) that are feasible for (7). We can take \(X^{\prime }\) to be the matching that corresponds to the indicator vector \(S^*\). Because \((S^*, Z^*)\) is optimal for (\(\ddagger\)), \(\begin{equation*} u_i(j) + u_j(i)\ge (u_i(\mu _{X}(i)) + \tau _i) + (u_j(\mu _{X}(j)) + \tau _j) \end{equation*}\) for all \((i, j)\in X^{\prime }\). Thus, we can find a vector \(\tau ^{\prime }\) of transfers that is feasible for (7). Then, since \(\sum _{a\in \mathcal {S}} \tau ^{\prime }_a= 0\), the objective of (7) at \((X^{\prime }, \tau ^{\prime })\) is \(\begin{equation*} \sum _{a\in \mathcal {S}} \bigl (u_a(\mu _{X^{\prime }}(a)) - u_a(\mu _{X}(a)) - \tau _a\bigr). \end{equation*}\) This equals to the objective of (\(\ddagger\)) at \((S^*, Z^*)\), which equals (\(\ddagger\)), as desired.

We now show the inequality in the other direction, that (\(\ddagger\)) is at least the maximum unhappiness of any coalition with respect to \((X, \tau)\). It suffices to construct a feasible solution \((S, Z)\) to (\(\ddagger\)) that achieves at least the maximum unhappiness of any coalition. Let \(\mathcal {S}\) be a coalition with maximum unhappiness, and let \((X^{\prime }, \tau ^{\prime })\) be an optimal solution for (7). Moreover, let S be the indicator vector corresponding to agents who are matched in \(X^{\prime }\) and Z be the indicator vector corresponding to agents in \(\mathcal {S}\) who are unmatched. The objective of (7) at \((X^{\prime }, \tau ^{\prime })\) is \(\begin{equation*} \sum _{a\in \mathcal {S}} \bigl (u_a(\mu _{X^{\prime }}(a)) - u_a(\mu _{X}(a)) - \tau _a\bigr), \end{equation*}\) which equals the objective of (\(\ddagger\)) at the \((S, Z)\) that we constructed.□

B.4 Proof of Proposition 4.4

Proposition 4.4. Subset Instability satisfies the following properties:

(1)

Subset Instability is always nonnegative and is zero if and only if \((X, \tau)\) is stable.

(2)

Subset Instability is Lipschitz continuous with respect to agent utilities. That is, for any possible market outcome \((X, \tau)\) and any pair of utility functions u and \(\tilde{u}\) it holds that: \(\begin{equation*} |I(X, \tau ; u, \mathcal {A}) - I(X, \tau ; \tilde{u}, \mathcal {A})| \le 2\sum _{a\in \mathcal {A}} {\Vert }u_a- \tilde{u}_a{\Vert }_\infty . \end{equation*}\)

(3)

Subset Instability is always at least the utility difference.

Proof of Proposition 4.4

We first prove the third part of the Proposition statement, then the first part of the Proposition statement, and finally the second part.

Proof of part (c). Because \(\sum _{a\in \mathcal {A}} \tau _a= 0\), Subset Instability satisfies the following: \(\begin{align*} I(X, \tau ; u, \mathcal {A}) &\ge \left(\max _{X^{\prime }\in \mathscr{X}_\mathcal {A}} \sum _{a\in \mathcal {A}} u_a(\mu _{X^{\prime }}(a)) \right) - \left(\sum _{a\in \mathcal {A}} u_a(\mu _{X}(a)) + \tau _a\right)\\ &= \left(\max _{X^{\prime }\in \mathscr{X}_\mathcal {A}} \sum _{a\in \mathcal {A}} u_a(\mu _{X^{\prime }}(a)) \right) - \left(\sum _{a\in \mathcal {A}} u_a(\mu _{X}(a))\right). \end{align*}\) The second line is exactly the utility difference.

Proof of part (a). From above, we have that Subset Instability is lower bounded by the utility difference, which is always nonnegative. Hence, Subset Instability is also always nonnegative.

To see that Subset Instability is 0 if and only if \((X, \tau)\) is stable, first suppose \((X, \tau)\) is unstable. Then, there exists a blocking pair \((i, j)\), in which case \(\begin{equation*} I(X, \tau ; u, \mathcal {A}) \ge u_i(j) + u_j(i) - (u_{i}(\mu _{X}(i)) + u_{j}(\mu _{X}(j)) + \tau _i+ \tau _j) \gt 0 \end{equation*}\) by the definition of blocking. Now, suppose \(I(X, \tau ; u, \mathcal {A}) \gt 0\). Then, there exists a subset \(S\subseteq \mathcal {A}\) such that \(\begin{equation*} \left(\max _{X^{\prime }\in \mathscr{X}_S} \sum _{a\in S} u_a(\mu _{X^{\prime }}(a)) \right) - \left(\sum _{a\in S} u_a(\mu _{X}(a)) + \tau _a\right) \gt 0. \end{equation*}\) Let \(X^{\prime }\) be a maximum weight matching on S. We can rewrite the above as \(\begin{equation*} \sum _{(i,j)\in X^{\prime }} \bigl (u_i(j) + u_j(i) - (u_i(\mu _{X}(i)) + u_j(\mu _{X}(j)) + \tau _i+ \tau _j\bigr) \gt 0. \end{equation*}\) Some term in the sum on the left-hand side must be positive, so there exists a blocking pair \((i, j)\in X^{\prime }\). In particular, \((X, \tau)\) is not stable.

Proof of part (b). We prove that \(\begin{equation*} |I(X, \tau ; u, \mathcal {A}) - I(X, \tau ; \tilde{u}, \mathcal {A})| \le 2\sum _{a\in \mathcal {A}} {\Vert }u_a- \tilde{u}_a{\Vert }_\infty . \end{equation*}\) The supremum of L-Lipschitz functions is L-Lipschitz, so it suffices to show that \(\begin{equation*} \left(\max _{X^{\prime }\in \mathscr{X}_S} \sum _{a\in S} u_a(\mu _{X^{\prime }}(a)) \right) - \sum _{a\in S} (u_a(\mu _{X}(a)) + \tau _a) \end{equation*}\) satisfies the desired Lipschitz condition for any \(S \subseteq \mathcal {A}\). In particular, it suffices to show that (8) \(\begin{equation} \left| \sum _{a\in S} (u_a(\mu _{X}(a)) + \tau _a) - \sum _{a\in S}(\tilde{u}_a(\mu _{X}(a)) + \tau _a) \right| \le \sum _{a\in \mathcal {A}} {\Vert }u_a- \tilde{u}_a{\Vert }_\infty \end{equation}\) and (9) \(\begin{equation} \left| \left(\max _{X^{\prime }\in \mathscr{X}_S} \sum _{a\in S} u_a(\mu _{X^{\prime }}(a))\right) - \left(\max _{X^{\prime }\in \mathscr{X}_S} \sum _{a\in S} \tilde{u}_a(\mu _{X^{\prime }}(a))\right) \right| \le \sum _{a\in \mathcal {A}} {\Vert }u_a- \tilde{u}_a{\Vert }_\infty . \end{equation}\) For (8), we have \(\begin{equation*} \left| \sum _{a\in S} (u_a(\mu _{X}(a)) + \tau _a) - \sum _{a\in S} (\tilde{u}_a(\mu _{X}(a)) + \tau _a) \right| = \left| \sum _{a\in S} \bigl (u_a(\mu _{X}(a)) - \tilde{u}_a(\mu _{X}(a))\bigr) \right| \le \sum _{a\in \mathcal {A}} {\Vert }u_a- \tilde{u}_a{\Vert }_\infty . \end{equation*}\) For (9), this boils down to showing that the total utility of the maximum weight matching is Lipschitz. Using again the fact that the supremum of Lipschitz functions is Lipschitz, this follows from the total utility of any fixed matching being Lipschitz.□

Skip CProofs for Section 5 Section

C Proofs for Section 5

C.1 Proof of Theorem 5.1

Theorem 5.1. For preference class \(\hspace{1.111pt}\mathcal {U}_{\mathrm{unstructured}}\) (see Section 3), MatchUCB (defined in Section 5.3) incurs expected regret \(\mathbb {E}(R_T) = O(|\mathcal {A}| \sqrt {n T\log (|\mathcal {A}| T)})\), where \(n = \max _t|\mathcal {A}_t|\).

Proof of Theorem 5.1

The starting point for our proof of Theorem 5.1 is the typical approach in multi-armed bandits and combinatorial bandits [15, 22, 30] of bounding regret in terms of the sizes of the confidence interval of the chosen arms. However, rather than using the sizes of confidence intervals to bound the utility difference (as in the incentive-free maximum weight matching setting), we bound Subset Instability through Lemma 5.4. From here on, our approach composes cleanly with existing bandits analyses; in particular, we can follow the typical combinatorial bandits approach [15, 22] to get the desired upper bound.

For completeness, we present the full proof. We divide into two cases, based on the event E that all of the confidence sets contain their respective true utilities at every timestep \(t\le T\). That is, \(u_{i}(j)\in C_{i,j}\) and \(u_j(i)\in C_{j,i}\) for all \((i,j)\in \mathcal {I}\times \mathcal {J}\) at all t.

Case 1: E holds. By Lemma 5.4, we may bound

where \(n_{ij}^{t}\) is the number of times that the pair \((i,j)\) has been matched at the start of round t. Let \(w^t_{i, j} = \frac{1}{\sqrt {n_{ij}^{t}}}\) be the size of the confidence set (with the log factor scaled out) for \((i, j)\) at the start of round t.

At each timestep t, let us consider the list consisting of \(w^t_{i_t, j_t}\) for all \((i_t, j_t) \in X^t\). Let us now consider the overall list consisting of the concatenation of all of these lists over all rounds. Let us order this list in decreasing order to obtain a list \(\tilde{w}_1, \ldots , \tilde{w}_L\) where \(L = \sum _{t=1}^{T} |X^t| \le n T\). In this notation, we observe that: \(\begin{equation*} \sum _{t=1}^{T} I(X^t, \tau ^t; u, \mathcal {A}) \le \sum _{t=1}^{T} \sum _{a\in \mathcal {A}^t} (\max (C_{a, \mu _{X^t}(a)}) - \min (C_{a, \mu _{X^t}(a)})) = \log (|\mathcal {A}| T) \sum _{l=1}^L \tilde{w}_l. \end{equation*}\) We claim that \(\tilde{w}_l \le O({\min (1, \frac{1}{\sqrt {(l / |\mathcal {A}|^2) -1}})})\). The number of rounds that a pair of agents can have their confidence set have size at least \(\tilde{w}_l\) is upper bounded by \(1 + \frac{1}{\tilde{w}_l^2}\). Thus, the total number of times that any confidence set can have size at least \(\tilde{w}_l\) is upper bounded by \((|\mathcal {A}|^2)(1 + \frac{1}{\tilde{w}_l^2})\).

Putting this together, we see that:

Case 2: E does not hold. Since each \(n_{ij}(\hat{u}_i(j) -u_i(j))\) is mean-zero and 1-subgaussian, and we have \(O(|\mathcal {I}||\mathcal {J}|T)\) such random variables over T rounds, the probability that any of them exceeds \(\begin{equation*} 2\sqrt {\log (|\mathcal {I}||\mathcal {J}|T/\delta)}\le 2\sqrt {\log (|\mathcal {A}|^2T/\delta)} \end{equation*}\) is at most \(\delta\) by a standard tail bound for the maximum of subgaussian random variables. It follows that E fails to hold with probability at most \(|\mathcal {A}|^{-2}T^{-2}\). In the case that E fails to hold, our regret in any given round would be at most \(4|\mathcal {A}|\) by the Lipschitz property in Proposition 4.4. (Recall that our upper confidence bound for any utility is wrong by at most 2 due to clipping each confidence interval to lie in \([-1, 1]\).) Thus, the expected regret from this scenario is at most \(\begin{equation*} |\mathcal {A}|^{-2}T^{-2}\cdot 4|\mathcal {A}|T\le 4|\mathcal {A}|^{-1}T^{-1}, \end{equation*}\) which is negligible compared to the regret bound from when E does occur.□

C.2 Proof of Theorem 5.2

Theorem 5.2. For preference class \(\hspace{1.111pt}\mathcal {U}_{\mathrm{typed}}\) (see Section 3), MatchTypedUCB (defined in Section 5.3) incurs expected regret \(\mathbb {E}(R_T) = O({|\mathcal {C}|} \sqrt {n T\log (|\mathcal {A}|T)})\), where \(n = \max _t |\mathcal {A}_t|\).

Proof of Theorem 5.2

Like in the proof of Theorem 5.1, we divide into two cases, based on the event E that all of the confidence sets contain their respective true utilities at every timestep \(t\le T\). That is, \(u_{a}(a^{\prime })\in C_{a, a^{\prime }}\) for all pairs of agents at all t.

Case 1: E holds. By Lemma 5.4, we may bound

where \(n_{c_1c_2}^{t}\) is the number of times that an agent of type \(c_1\) has been matched with an agent of context \(c_2\) at the start of round t. (We define \(n_{c_1, c_2}^0 = 0\) by default.) Let \(w^t_{c_1, c_2} = \frac{1}{\sqrt {n^t_{c_1, c_2}}}\) be the size of the confidence set (with the log factor scaled out) for \((c_1, c_2)\) at the start of round t.

At each timestep t, let us consider the list consisting of \(w^t_{c_{i_t}, c_{j_t}}\) for all \((i_t, j_t) \in X^t\). Let us now consider the overall list consisting of the concatenation of all of these lists over all rounds. Let us order this list in decreasing order to obtain a list \(\tilde{w}_1, \ldots , \tilde{w}_L\) where \(L = \sum _{t=1}^{T} |X^t| \le n T\). In this notation, we observe that: \(\begin{equation*} \sum _{t=1}^{T} I(X^t, \tau ^t; u, \mathcal {A}^t) \le \sum _{t=1}^{T} \sum _{a\in \mathcal {A}^t} (\max (C_{c_a, c_{\mu _{X^t}(a)}}) - \min (C_{c_a, c_{\mu _{X^t}(a)}})) = \log (|\mathcal {A}| T) \sum _{l=1}^L \tilde{w}_l. \end{equation*}\) We claim that \(\tilde{w}_l \le O({\min (1, \frac{1}{\sqrt {(l / |\mathcal {C}|^2) -1}})})\). The number of instances that a pair of contexts can have their confidence set have size at least \(\tilde{w}_l\) is upper bounded by \(2n + \frac{1}{\tilde{w}_l^2}\). Thus, the total number of times that any confidence set can have size at least \(\tilde{w}_l\) is upper bounded by \((|\mathcal {C}|)(2n + \frac{1}{\tilde{w}_l^2})\).

Putting this together, we see that:

Case 2: E does not hold. Since each \(n_{ij}(\hat{u}_i(j) -u_i(j))\) is mean-zero and 1-subgaussian, and we have \(O(|\mathcal {I}||\mathcal {J}|T)\) such random variables over T rounds, the probability that any of them exceeds \(\begin{equation*} 2\sqrt {\log (|\mathcal {I}||\mathcal {J}|T/\delta)}\le 2\sqrt {\log (|\mathcal {A}|^2T/\delta)} \end{equation*}\) is at most \(\delta\) by a standard tail bound for the maximum of subgaussian random variables. It follows that E fails to hold with probability at most \(|\mathcal {A}|^{-2}T^{-2}\). In the case that E fails to hold, our regret in any given round would be at most \(4|\mathcal {A}|\) by the Lipschitz property in Proposition 4.4. (Recall that our upper confidence bound for any utility is wrong by at most two due to clipping each confidence interval to lie in \([-1, 1]\).) Thus, the expected regret from this scenario is at most \(\begin{equation*} |\mathcal {A}|^{-2}T^{-2}\cdot 4|\mathcal {A}|T\le 4|\mathcal {A}|^{-1}T^{-1}, \end{equation*}\) which is negligible compared to the regret bound from when E does occur.□

C.3 Proof of Theorem 5.3

Theorem 5.3. For preference class \(\hspace{1.111pt}\mathcal {U}_{\mathrm{linear}}\) (see Section 3), MatchLinUCB (defined in Section 5.3) incurs expected regret \(\mathbb {E}(R_T) = O (d \sqrt {|\mathcal {A}|}\sqrt {n T\log (|\mathcal {A}| T)})\), where \(n = \max _t |\mathcal {A}_t|\).

To prove Theorem 5.3, it suffices to (a) show that the confidence sets contain the true utilities with high probability and (b) bound the sum of the sizes of the confidence sets.

Part (a) follows from fact established in existing analysis of LinUCB in the classical linear contextual bandits setting [39].

Lemma C.1 ([39, Proposition 2]) Let the confidence sets be defined as above (and in MatchLinUCB). For each \(a\in \mathcal {A}\), it holds that: \(\begin{equation*} \mathbb {P}[\phi (a) \in C_{\phi (a)} \quad \forall 1 \le t\le T] \ge 1 - 1/(|\mathcal {A}|^3 T^2). \end{equation*}\)

Lemma C.2 Let the confidence sets be defined as above (and in MatchLinUCB). For each \(a\in \mathcal {A}\) and for any \(\varepsilon \gt 0\), it holds that: \(\begin{equation*} \sum _{t\mid a\in \mathcal {A}^t, \mu _{X^t}(a) \ne a}\mathbf {1}\left[\max \bigl (C_{a, \mu _{X^t} (a)}\bigr) - \min \bigl (C_{a, \mu _{X^t} (a)})\bigr) \gt \varepsilon \right] \le O\left(\left(\frac{4 \beta _T}{\varepsilon ^2} + 1\right) d \log (1/\varepsilon) \right). \end{equation*}\)

Proof.

We follow the same argument as the proof of Proposition 3 in Russo and Van Roy [39].

We first recall the definition of \(\varepsilon\)-dependence and \(\varepsilon\)-eluder dimension: We say that an agent \(a^{\prime }\) is \(\varepsilon\)-dependenton \(a_1^{\prime },\ldots ,a_s^{\prime }\) if for all \(\phi (a), \tilde{\phi }(a)\in \mathcal {B}^d\) such that \(\begin{equation*} \sum _{k=1}^s \langle c_{a_k^{\prime }}, \tilde{\phi }(a) - \phi (a)\rangle ^2 \le \varepsilon ^2, \end{equation*}\) we also have \(\langle c_{a^{\prime }}, \tilde{\phi }(a) - \phi (a)\rangle ^2 \le \varepsilon ^2\). The \(\varepsilon\)-eluder dimension \(d_{{\varepsilon} \textrm{-eluder}}\) of \(\mathcal {B}^d\) is the maximum length of a sequence \(a_1^{\prime }, \ldots , a_s^{\prime }\) such that no element is \(\varepsilon\)-dependent on a prefix.

Consider the subset \(S_a\) of \(\lbrace t\mid a\in \mathcal {A}^t, \mu _{X^t}(a) \ne a\rbrace\) such that \(\begin{equation*} \mathbf {1}\left[\max \bigl (C_{a, \mu _{X^t} (a)}\bigr) - \min \bigl (C_{a, \mu _{X^t} (a)})\bigr) \gt \varepsilon \right]. \end{equation*}\) Suppose for the sake of contradiction that \(\begin{equation*} |S_a| \gt \left(\frac{4 \beta _T}{\varepsilon ^2} + 1\right) d_{{\varepsilon}\textrm {-eluder}}. \end{equation*}\) Then, there exists an element \(t^*\) that is \(\varepsilon\)-dependent on \(\frac{4 \beta _T}{\varepsilon ^2} + 1\) disjoint subsets of \(S_a\): One can repeatedly remove sequences \(a_{\mu _{X^{t_1}}(a)}^{\prime }, \ldots , a_{\mu _{X^{t_s}}(a)}^{\prime }\) of maximal length such that no element is \(\varepsilon\)-dependent on a prefix; note that \(s\le d_{{\varepsilon}\textrm {-eluder}}\) always. Let the subsets be \(S_a^{(q)}\) for \(q = 1,\ldots ,\frac{4 \beta _T}{\varepsilon ^2} + 1\), and let \(\phi (a), \tilde{\phi }(a)\) be such that \(\langle c_{\mu _{X^{t^*}}(a)}, \tilde{\phi }(a) - \phi (a)\rangle \gt \varepsilon\). The above implies that \(\begin{equation*} \sum _{q=1}^{\frac{4 \beta _T}{\varepsilon ^2} + 1}\sum _{t\in S_a^{(q)}} \langle c{\mu _{X^t}(a)}, \tilde{\phi }(a) - \phi (a)\rangle ^2 \gt 4\beta _T \end{equation*}\) by the definition of \(\varepsilon\)-dependence. But this is impossible, since the left-hand side is upper bounded by \(\begin{equation*} \sum _{t=1}^T\langle c{\mu _{X^t}(a)}, \tilde{\phi }(a) - \phi (a)\rangle ^2\le 4\beta _T \end{equation*}\) by the definition of the confidence sets. Hence, it must hold that \(\begin{equation*} |S_a| \le \left(\frac{4 \beta _T}{\varepsilon ^2} + 1\right) d_{{\varepsilon}\textrm {-eluder}}. \end{equation*}\) Now, it follows from the bound on the eluder dimension for linear bandits (Proposition 6 in Reference [39]) that the bound of \(\tilde{O} ((\frac{4 \beta _T}{\varepsilon ^2} + 1) d \log (1/\varepsilon).)\) holds.□

Lemma C.3 Let the confidence sets be defined as above (and in MatchLinUCB). For any \(a\in \mathcal {A}\), it holds that: \(\begin{equation*} \sum _{t\mid a\in \mathcal {A}^t, \mu _{X^t}(a) \ne a}(\max (C_{a, \mu _{X^t} (a)}) - \min (C_{a, \mu _{X^t} (a)}))) \le O(d (\log (T |\mathcal {A}|)) \sqrt {T_a}) , \end{equation*}\) where \(T_a\) is the number of times that agent a has been matched.

Proof.

Let us consider the set of confidence set sizes \((\max (C_{a, \mu _{X^t} (a)}) - \min (C_{a, \mu _{X^t} (a)})))\) for t such that \(a\in \mathcal {A}^t, \mu _{X^t}\). Let us sort these confidence set sizes in decreasing order and label them \(w_1 \ge \cdots \ge w_{T_a}\). Restating C.2, we see that (10) \(\begin{equation} \sum _{t=1}^{T_a} w_t \mathbf {1}[w_t \gt \varepsilon ] \le O\left(\left(\frac{4 \beta _T}{\varepsilon ^2} + 1\right) d \log (1/\varepsilon) \right). \end{equation}\) for all \(\varepsilon \gt 0\).

We see that: \(\begin{align*} \sum _{t\mid a\in \mathcal {A}^t, \mu _{X^t}(a) \ne a}(\max (C_{a, \mu _{X^t} (a)}) - \min (C_{a, \mu _{X^t} (a)}))) &= \sum _{t=1}^{T_a} w_t \\ &\le \sum _{t=1}^{T_a} w_t \mathbf {1}[w_t \gt 1/T_a^2] + \sum _{t=1}^{T_a} w_t \mathbf {1}[w_t \le 1/T_a^2] \\ &\le \frac{1}{T_a} + \sum _{t=1}^{T_a} w_t \mathbf {1}[w_t \gt 1/T_a^2]. \end{align*}\)

We claim that \(w_i \le 2\) if \(i \ge d \log (T_a)\) and \(w_i \le \min (2, \frac{4 \beta _T (d \log T_a)}{i - d \log T_a})\) if \(i \gt d \log T_a\). The first part follows from the fact that we truncate the confidence sets to be within \([-1, 1]\). It thus suffices to show that \(w_i \le \frac{4 \beta _T (d \log T_a)}{i - d \log T_a}\) for \(t \le d \log T\). If \(w_i \ge \varepsilon \gt 1/T_a^2\), then we see that \(\sum _{t=1}^{T_a} \mathbf {1}[w_t \gt \varepsilon ] \ge i\), which means by (10) that \(i \le O((\frac{4 \beta _T}{\varepsilon ^2} + 1) d \log (1/\varepsilon)) \le O((\frac{4 \beta _T}{\varepsilon ^2} + 1) d \log (T_a))\), which means that \(\varepsilon \le \frac{4 \beta _T (d \log T_a)}{i - d \log T_a}\). This proves the desired statement.

Now, we can plug this into the above expression to obtain: \(\begin{align*} &{\sum _{t\mid a\in \mathcal {A}^t, \mu _{X^t}(a) \ne a}(\max (C_{a, \mu _{X^t} (a)}) - \min (C_{a, \mu _{X^t} (a)})))}\\ &\le \frac{1}{T_a} + \sum _{t=1}^{T_a} w_t \mathbf {1}[w_t \gt 1/T_a^2] \\ &\le \frac{1}{T_a} + 2 d \log (T_a) + \sum _{i \gt d \log T_a}^{T_a} \min \left(2, \frac{4 \beta _T (d \log T_a)}{i - d \log T_a}\right) \\ & \le \frac{1}{T_a} + 2 d \log (T_a) + 2 \sqrt {d \log T_a \beta _T} \int _{t=0}^{T_a} t^{-1/2} dt \\ & = \frac{1}{T_a} + 2 d \log (T_a) + 4 \sqrt {d T_a \log T_a \beta _T}. \end{align*}\) We now use the fact that: \(\begin{equation*} \beta _T = O\left(d \log T + \frac{1}{T} \sqrt {\log (T^2 |A|)}\right). \end{equation*}\) Plugging this into the above expression, we obtain the desired result.□

We are now ready to prove Theorem 5.3.

Proof of Theorem 5.3

Like in the proof of Theorem 5.1, we divide into two cases, based on the event E that all of the confidence sets contain their respective true utilities at every timestep \(t\le T\). That is, \(u_{c_1}(c_2)\in C_{c_1, c_2}\) for all \(c_1, c_2 \in \mathcal {C}\) at all t.

Case 1: E holds. By Lemma 5.4, we know that the cumulative regret is upper bounded by \(\begin{align*} R_T &\le \sum _{t=1}^{T} \sum _{a\in \mathcal {A}^t} (\max (C_{a, \mu _{X^t}(a)}) - \min (C_{a, \mu _{X^t}(a)})) \\ &= \sum _{a\in \mathcal {A}} \sum _{t\mid a\in \mathcal {A}^t, \mu _{X^t}(a) \ne a}(\max (C_{a, \mu _{X^t} (a)}) - \min (C_{a, \mu _{X^t} (a)}))) \\ &\le \sum _{a\in \mathcal {A}} O(d \log (T |\mathcal {A}|) \sqrt {T_a}), \end{align*}\) where the last inequality applies C.3 to the inner summand. We see that \(\sum _{a\in \mathcal {A}} T_a= \sum _t |\mathcal {A}_t| \le n T\) by definition, since at most n agents show up at every round. Let us now observe that: \(\begin{equation*} \sum _{a\in \mathcal {A}} \sqrt {T_a} \le \sqrt {|\mathcal {A}|} \sqrt {\sum _{a\in \mathcal {A}} T_a} \le \sqrt {|\mathcal {A}| n T}, \end{equation*}\) as desired.

Case 2: E does not hold. From C.1, it follows that: \(\begin{equation*} \mathbb {P}[\phi (a) \in C_{\phi (a)} \quad \forall 1 \le t\le T] \ge 1 - 1/(|\mathcal {A}|^3 T^2). \end{equation*}\) Union bounding, we see that \(\begin{equation*} \mathbb {P}[\phi (a) \in C_{\phi (a)} \quad \forall 1 \le t\le T\forall a\in \mathcal {A}] \ge 1 - 1/(|\mathcal {A}|^2 T^2). \end{equation*}\) By the definition of the confidence sets for the utilities, we see that: (11) \(\begin{equation} \mathbb {P}[u(a, a^{\prime }) \in C_{a, a^{\prime }} \quad \forall 1 \le t\le T, \forall a, a^{\prime } \in \mathcal {A}] \ge 1/(|\mathcal {A}|^2 T^2). \end{equation}\) Thus, the probability that event E does not hold is at most \(|\mathcal {A}|^{-2} T^{-2}\). In the case that E fails to hold, our regret in any given round would be at most \(4 |\mathcal {A}|\) by the Lipschitz property in Proposition 4.4. Thus, the expected regret is at most \(4 |\mathcal {A}|^{-1} T^{-1}\), which is negligible compared to the regret bound from when E does occur.□

C.4 Proof of Lemma 5.5

Lemma 5.5. For any algorithm that learns a stable matching with respect to unstructured preferences, there exists an instance on which it has expected regret \(\widetilde{\Omega }(|A|^{3/2} \sqrt {T})\) (where regret is given by Subset Instability).

Proof of Lemma 5.5

Recall that, by Proposition 4.4, the problem of learning a maximum weight matching with respect to utility difference is no harder than that of learning a stable matching with respect to Subset Instability. In the remainder of our proof, we reduce a standard “hard instance” for stochastic multi-armed bandits to our setting of learning a maximum weight matching.

Step 1: Constructing the hard instance for stochastic MAB. Consider the following family of stochastic multi-armed bandits instances: For a fixed K, let \(\mathcal {I}_\alpha\) for \(\alpha \in \lbrace 1,\ldots ,K\rbrace\) denote the stochastic multi-armed bandits problem where all arms have 0-1 rewards, and the k-th arm has mean reward \(\frac{1}{2} + \rho\) if \(k =\alpha\) and \(\frac{1}{2}\) otherwise, where \(\rho \gt 0\) will be set later. A classical lower bound for stochastic multi-armed bandits is the following:

Lemma C.4 ([6]) The expected regret of any stochastic multi-armed bandit algorithm on an instance \(\mathcal {I}_\alpha\) for \(\alpha\) selected uniformly at random from \(\lbrace 1,\ldots ,K\rbrace\) is \(\Omega (\sqrt {KT})\).□

Step 2: Constructing a (random) instance for the maximum weight matching problem. We will reduce solving the above distribution over stochastic multi-armed bandits problems to a distribution over instances of learning a maximum weight matching. Let us now construct this random instance of the maximum weight matching problem. Let \(|\mathcal {I}| = K\) and \(|\mathcal {J}| = 10 K\log (KT)\). Specifically, we sample inputs for learning a maximum weight matching as follows: For each man \(i\in \mathcal {I}\), select \(\alpha _i\in \lbrace 1,\ldots ,K\rbrace\) uniformly at random, and define \(u_i(j)\) to be \(\frac{1}{2} + \rho\) if \(\lfloor (j- 1) / \log K\rfloor = \alpha _i\) and \(\frac{1}{2}\) otherwise. Furthermore, let \(u_j(i) = 0\) for all \((i,j)\in \mathcal {I}\times \mathcal {J}\). Finally, suppose observations are always in \(\lbrace 0,1\rbrace\) (but are unbiased).

The key property of the above setup that we will exploit for our reduction is the fact that, due to the imbalance in the market, the maximum weight matching for these utilities has with high probability each i matched with some j whom they value at \(\frac{1}{2} + \rho\). Indeed, by a union bound, the probability that more than \(10\log (KT)\) different i have the same \(\alpha _i\) is at most

Thus, with probability \(1 - O(K^{-4}T^{-4})\), this event holds. The case where this event does not hold contributes negligibly to regret, so we do not consider it further.

Step 3: Establishing the reduction. Now, suppose for the sake of contradiction that some algorithm could solve our random instance of learning a maximum weight matching problem with expected regret \(o(K^{3/2}\sqrt {T})\). We can obtain a stochastic multi-armed bandits that solves the instances in C.4 as follows: Choose a random \(i^*\in \mathcal {I}\) and set \(\alpha _{i^*} = \alpha\). Simulate the remaining i by choosing \(\alpha _{i}\) for all \(i\ne i^*\) uniformly at random. Run the algorithm on this instance of learning a maximum weight matching, “forwarding” arm pulls to the true instance when matching \(i^*\).

To analyze the regret of this algorithm when faced with the distribution from C.4, we first note that with high probability, all the agents \(i\in \mathcal {I}\) can simultaneously be matched to a set of \(j\in \mathcal {J}\) such that each i is matched to some j whom they value at \(\frac{1}{2} + \rho\). Then, the regret of any matching is \(\rho\) times the number of \(i\in \mathcal {I}\) who are not matched to a j whom they value at \(\frac{1}{2} + \rho\). Thus, we can define the cumulative regret for an agent \(i\in \mathcal {I}\) as \(\rho\) times the number of rounds they were not matched to someone whom they value at \(\frac{1}{2} + \rho\). For \(i^*\), this regret is just the regret for the distribution from C.4. Since \(i^*\) was chosen uniformly at random, their expected cumulative regret is at most \(\begin{equation*} \frac{1}{K}\cdot o(K^{3/2}\sqrt T) = o(\sqrt {KT}), \end{equation*}\) in violation of C.4.

Step 4: Concluding the lower bound. This contradiction implies that no algorithm can hope to obtain \(o(K^{3/2}\sqrt {T})\) expected regret on this distribution over instances of learning a maximum weight matching. Since there are \(O(K\log (KT)) = \widetilde{O}(K)\) agents in the market total, the desired lower bound follows.

Skip DProof of Theorem 6.1 Section

D Proof of Theorem 6.1

Theorem 6.1 (Instance-Dependent Regret). Suppose that \(\mathcal {A}_t= \mathcal {A}\) for all t. Let \(u\in \hspace{1.111pt}\mathcal {U}_{\mathrm{unstructured}}\) be any utility function and put \(\begin{equation*} \Delta := \inf _{X\ne X^*} \Biggl \lbrace \sum _{a\in \mathcal {A}} u_a(\mu _{X^*}(a)) - \sum _{a\in \mathcal {A}} u_a(\mu _{X}(a)) \Biggr \rbrace . \end{equation*}\) Then, MatchUCB\(^{\prime }\) incurs expected regret \(\mathbb {E}(R_T) = O(|\mathcal {A}|^5 \cdot \log (|\mathcal {A}|T)/\Delta ^2)\).

D.1 MatchUCB'

MatchUCB\(^{\prime }\) is the same as MatchUCB, except we call ComputeMatch\(^{\prime }\) instead of ComputeMatch. The idea behind ComputeMatch\(^{\prime }\) is that we compute an optimal primal-dual solution for both the original confidence sets C as well as expanded confidence sets \(C^{\prime }\), which we define to be twice the width of the original confidence sets. More formally, we define \(\begin{equation*} C^{\prime }_{a, a^{\prime }} := \biggl [\min (C_{a, a^{\prime }}) - \frac{\max (C_{a, a^{\prime }}) - \min (C_{a, a^{\prime }})}{2}, \max (C_{a, a^{\prime }}) + \frac{\max (C_{a, a^{\prime }}) - \min (C_{a, a^{\prime }})}{2}\biggr ]. \end{equation*}\) We will adaptively explore (following UCB) according to both C and \(C^{\prime }\). Doing extra exploration according to the more pessimistic confidence sets \(C^{\prime }\) is necessary for us to be able to find “robust” dual solutions for setting transfers.

We define \((X^{*}, p^{*})\), which will be an optimal primal-dual solution for the upper confidence bounds of \({C}\) as follows: Let \(X^*\) be a maximum weight matching with respect to \(u^{\mathrm{UCB}}\). We next compute the gap \(\begin{equation*} \Delta ^{\text{UCB}} = \min _{X\ne X^{*}} \Biggl \lbrace \sum _{a\in \mathcal {A}} u^{\mathrm{UCB}}_a(\mu _{X^{*}}(a)) - \sum _{a\in \mathcal {A}} u^{\mathrm{UCB}}_a(\mu _{X}(a)) \Biggr \rbrace \end{equation*}\) with respect to \(u^{\mathrm{UCB}}\). We can compute this gap by computing the maximum weight matching and the second-best matching with respect to \(u^{\mathrm{UCB}}\).15 Next, define utility functions \(u^{\prime }_a\) such that \(\begin{equation*} u^{\prime }_a(a^{\prime }) = {\left\lbrace \begin{array}{ll}u^{\mathrm{UCB}}_a(a^{\prime }) - \frac{\Delta ^{\text{UCB}}}{|\mathcal {A}|} & \text{if $\mu _{X^{*}}(a) = a^{\prime }$ and $a\ne a^{\prime }$} \\ u^{\mathrm{UCB}}_a(a^{\prime }) & \text{otherwise} \end{array}\right.} \end{equation*}\) for all \(a\in \mathcal {A}\). (We show in D.1 that \(X^*\) is still a maximum weight matching for \(u^{\prime }\).) Now, compute an optimal dual solution \(p^{\prime }\) for utility function \(u^{\prime }\). To get \(p^*\), we add \(\Delta ^\text{UCB}/|\mathcal {A}|\) to \(p^{\prime }_a\) for each matched agent a in \(X^*\). (See D.2 for a proof that \((X^*, p^*)\) is an optimal primal-dual pair with respect to \(u^{\mathrm{UCB}}\).)

Finally, let \((X^{*,2}, p^{*,2})\) be any optimal primal-dual pair for the utility function \(u^{\mathrm{UCB},2}\) given by the upper confidence bounds \(\max (C^{\prime }_{a,a^{\prime }})\) of \({C}^{\prime }\).

With this setup, we define ComputeMatch\(^{\prime }\) as follows: If \(X^{*} \ne X^{*,2}\), then return \((X^{*,2}, \tau ^{*,2})\), where \(\tau ^{*,2}\) is given by \(\tau _a^{*,2} = p^{*,2}_a- u^{\mathrm{UCB},2}_a(\mu _{X^{*,2}}(a))\) if a is matched and \(\tau ^{*,2}_a= 0\) if a is unmatched. Otherwise, return \((X^{*}, \tau ^*)\), where \(\tau ^*\) is given by \(\tau ^*_a= p^{*}_a- u^{\mathrm{UCB}}_a(\mu _{X}(a))\) if a is matched and \(\tau ^*_a= 0\) if a is unmatched.

D.2 Proof of Theorem 6.1

We first verify (as claimed above) that \(X^*\) is a maximum weight matching with respect to \(u^{\prime }\).

Lemma D.1 Matching \(X^*\) is a maximum weight matching with respect to \(u^{\prime }\).

Proof.

Consider any matching \(X\ne X^*\). Since \(\begin{equation*} \sum _{a\in \mathcal {A}}u^{\mathrm{UCB}}_a(\mu _{X}(a))\le -\Delta ^{\mathrm{UCB}} + \sum _{a\in \mathcal {A}} u^{\mathrm{UCB}}_a(\mu _{X^*}(a)) \end{equation*}\) by the definition of \(\Delta ^\mathrm{UCB}\), we have

We now prove the main lemma for this analysis, restated below. Lemma 6.2 shows that if the confidence sets are small enough, then the selected matching will be stable with respect to the true utilities.

Lemma 6.2. Suppose ComputeMatch\(^{\prime }\) is run on a collection \(\mathscr{C}\) of confidence sets \(C_{i,j}\) and \(C_{j,i}\) over the agent utilities that satisfy \(\begin{equation*} \max (C_{i,j}) - \min (C_{i,j}) \le 0.05{} \frac{\Delta }{|\mathcal {A}|} \quad \text{and}\quad \max (C_{j, i}) - \min (C_{j, i}) \le 0.05{} \frac{\Delta }{|\mathcal {A}|} \end{equation*}\) for all \((i, j)\) in the matching returned by ComputeMatch\(^{\prime }\) . Suppose also that the confidence sets \(\mathscr{C}\) contain the true utilities for all pairs of agents. Then, the market outcome returned by ComputeMatch\(^{\prime }\) is stable with respect to the true utilities u.

Proof of Lemma 6.2

The proof proceeds in five steps, which we now outline. We first show the matching returned by ComputeMatch\(^{\prime }\) is the maximum weight matching \(X^{\text{opt}}\) with respect to u. We next show that \(X^*\) as defined in ComputeMatch\(^{\prime }\) also equals \(X^{\text{opt}}\). These facts let us conclude that ComputeMatch\(^{\prime }\) returns \((X^*, \tau ^*)\). We then show \(\Delta ^\mathrm{UCB}\) is at least \(0.1\Delta\). We then show that \((X^*, \tau ^*)\) is stable with respect to \(u^{\prime }\). We finish by showing that this implies \((X^*, \tau ^*)\) is stable with respect to u.

Throughout the proof, we will use the following observation about the expanded confidence sets: (12) \(\begin{equation} \max (C^{\prime }_{i,j}) - \min (C^{\prime }_{i,j}) \le 0.1{} \frac{\Delta }{|\mathcal {A}|} \quad \text{and}\quad \max (C^{\prime }_{j, i}) - \min (C^{\prime }_{j, i}) \le 0.1{} \frac{\Delta }{|\mathcal {A}|} \end{equation}\) for all \((i, j)\) in the matching returned by ComputeMatching\(^{\prime }\). This follows from the assumptions in the lemma statement.

Proving ComputeMatch\(^{\prime }\) returns \(X^{\mathrm{opt}}\) as the matching. ComputeMatch\(^{\prime }\) by definition returns \(X^{*,2}\) always, so it suffices to show that \(X^{*,2} = X^{\text{opt}}\). Note that \(X^{*,2}\) is a maximum weight matching with respect to \(u^{\text{UCB}, 2}\). This means that \(\begin{align*} \sum _{a\in \mathcal {A}} u_a(\mu _{X^{*, 2}}(a)) &\ge -\sum _{a\in \mathcal {A}} (\max (C^{\prime }_{a, \mu _{X^{*,2}}(a) }) - \min (C^{\prime }_{a, \mu _{X^{*,2}}(a) })) + \sum _{a\in \mathcal {A}} u^{\text{UCB}, 2}_a(\mu _{X^{*, 2}}(a)) \\ &\ge - 0.1{}\Delta + \sum _{a\in \mathcal {A}} u^{\text{UCB}, 2}_a(\mu _{X^{*, 2}}(a)) \\ &\ge -0.1{}\Delta + \sum _{a\in \mathcal {A}} u^{\text{UCB}, 2}_a(\mu _{X^{\text{opt}}}(a)) \\ &\ge -0.1{}\Delta + \sum _{a\in \mathcal {A}} u_a(\mu _{X^{\text{opt}}}(a)). \end{align*}\) By the definition of the gap \(\Delta\), we conclude that \(X^{*,2}= X^{\text{opt}}\).

Proving \(X^*= X^{\mathrm{opt}}\). Suppose for sake of contradiction that \(X^* \ne X^{\text{opt}}\). Then, \(\begin{equation*} \sum _{a\in \mathcal {A}} u^{\mathrm{UCB}}_a(\mu _{X^{*}}(a)) \ge \sum _{a\in \mathcal {A}} u^{\mathrm{UCB}}_a(\mu _{X^{\text{opt}}}(a)) \ge \sum _{a\in \mathcal {A}} u_a(\mu _{X^{\text{opt}}}(a)), \end{equation*}\) since \(X^*\) is a maximum weight matching with respect to \(u^{\mathrm{UCB}}\). Moreover, by the definition of the gap, we know that \(\sum _{a\in \mathcal {A}} u_a(\mu _{X^{*}}(a)) \le \sum _{a\in \mathcal {A}} u_a(\mu _{X^{\text{opt}}}(a)) - \Delta\). Putting this all together, we see that \(\begin{align*} \sum _{a\in \mathcal {A}} (\max (C_{a, \mu _{X^{*}}(a) }) - \min (C_{a, \mu _{X^{*}}(a) })) &\ge \sum _{a\in \mathcal {A}} u^{\mathrm{UCB}}_a(\mu _{X^{*}}(a)) - \sum _{a\in \mathcal {A}} u_a(\mu _{X^{*}}(a))\\ &\ge \Delta . \end{align*}\) We now use this to lower bound the utility of \(X^*\) on \(u^{\text{UCB}, 2}\). By the definition of the confidence sets, we see that \(\begin{align*} \sum _{a\in \mathcal {A}} u^{\text{UCB}, 2}_a(\mu _{X^{*}}(a)) &\ge \sum _{a\in \mathcal {A}} u^{\text{UCB}}_a(\mu _{X^{*}}(a)) + \frac{1}{2} \sum _{a\in \mathcal {A}} (\max (C_{a, \mu _{X^{*}}(a) }) - \min (C_{a, \mu _{X^{*}}(a) })) \\ &\ge \sum _{a\in \mathcal {A}} u^{\text{UCB}}_a(\mu _{X^{*}}(a)) + 0.5 \Delta . \end{align*}\) However, \(X^{\text{opt}}\) only achieves a utility of \(\begin{align*} \sum _{a\in \mathcal {A}} u^{\text{UCB}, 2}_a(\mu _{X^{\text{opt}}}(a)) &\le \sum _{a\in \mathcal {A}} u_a(\mu _{X^{\text{opt}}}(a)) + \sum _{a\in \mathcal {A}} (\max (C^{\prime }_{a, \mu _{X^{\text{opt}}}(a) }) - \min (C^{\prime }_{a, \mu _{X^{\text{opt}}}(a) })) \\ &\le \sum _{a\in \mathcal {A}} u_a(\mu _{X^{\text{opt}}}(a)) + 0.1{}\Delta . \end{align*}\) But this contradicts the fact (from above) that \(X^{\text{opt}} = X^{*,2}\) is a maximum weight matching with respect to \(u^{\mathrm{UCB}, 2}\). Therefore, it must be that \(X^*= X^{\text{opt}}\).

Putting the above two arguments together, we conclude ComputeMatch\(^{\prime }\) returns \((X^*, \tau ^*)\) in this case.

Bounding the gap \(\Delta ^{\mathrm{UCB}}\). We next show that \(\Delta ^{\text{UCB}} \ge 0.1{} \Delta\). We proceed by assuming (13) \(\begin{equation} \sum _{a\in \mathcal {A}} u^{\mathrm{UCB}}_a(\mu _{X}(a))\ge -0.1{}\Delta + \sum _{a\in \mathcal {A}} u^{\mathrm{UCB}}_a(\mu _{X^*}(a)) \end{equation}\) for some \(X\ne X^*\) and deriving a contradiction.

We first show that (13) implies a lower bound on \(\begin{equation*} S = \sum _{a\in \mathcal {A}} (\max (C_{a, \mu _{X}(a)}) - \min (C_{a, \mu _{X}(a)})) \end{equation*}\) in terms of \(\Delta\). Because the confidence sets contain the true utilities and \(u^{\mathrm{UCB}}_a\) upper bounds \(u_a\) pointwise, (13) implies \(\begin{equation*} S + \sum _{a\in \mathcal {A}}u_a(\mu _{X}(a)) \ge \sum _{a\in \mathcal {A}} u^{\mathrm{UCB}}_a(\mu _{X}(a))\ge -0.1{}\Delta + \sum _{a\in \mathcal {A}} u_a(\mu _{X^*}(a)). \end{equation*}\) Applying the definition of \(\Delta\), we obtain the lower bound \(\begin{equation*} S\ge - 0.1{}\Delta + \sum _{a\in \mathcal {A}} u_a(\mu _{X^*}(a)) - \sum _{a\in \mathcal {A}} u_a(\mu _{X}(a))\ge (1 - 0.1{})\Delta . \end{equation*}\)

Now, we apply the fact that \(X^{*} = X^{*,2} = X^{\text{opt}}\). We establish the following contradiction: \(\begin{align*} 0.1{}\Delta + \sum _{a\in \mathcal {A}} u^{\mathrm{UCB}}_a(\mu _{X^*}(a)) &\ge 0.1{}\Delta + \sum _{a\in \mathcal {A}} u_a(\mu _{X^*}(a)) \\ &=\sum _{a\in \mathcal {A}} (u_a(\mu _{X^*}(a)) + 0.1{}\Delta /|\mathcal {A}|) \\ &\stackrel{\text{(i)}}{\ge } \sum _{a\in \mathcal {A}} u^{\mathrm{UCB},2}_a(\mu _{X^*}(a)) \\ &\stackrel{\text{(ii)}}{\ge } \sum _{a\in \mathcal {A}} u^{\mathrm{UCB},2}_a(\mu _{X}(a)) \\ &\stackrel{\text{(iii)}}{\ge } \frac{S}{2} + \sum _{a\in \mathcal {A}} u^{\mathrm{UCB}}_a(\mu _{X}(a)) \\ &\stackrel{\text{(iv)}}{\ge } \left(\frac{1}{2}(1 - 0.1{})\right)\Delta + \sum _{a\in \mathcal {A}} u^{\mathrm{UCB}}_a(\mu _{X}(a)) \\ &\stackrel{\text{(v)}}{\ge } \left(\frac{1}{2}(1 - 0.1{}) - 0.1{}\right)\Delta + \sum _{a\in \mathcal {A}} u^{\mathrm{UCB}}_a(\mu _{X^*}(a)). \end{align*}\) Here, (i) comes from (12) in the lemma statement; (ii) holds because \(X^*= X^{*,2}\) is a maximum weight matching with respect to \(u^{\mathrm{UCB},2}\); (iii) is by the definition of \(u^{\mathrm{UCB},2}\); (iv) follows from our lower bound on S; and (v) follows from (13).

Proving that \((X^*, \tau ^*)\) is stable with respect to \(u^{\prime }\). By D.1, \((X^*, p^{\prime })\) is an optimal primal-dual pair with respect to \(u^{\prime }\). Now, it suffices to show that the primal-dual solution corresponds to the market outcome \((X^*, \tau ^*)\) for \(u^{\prime }\). To see this, notice that \(p^{\prime }_a= 0\) for unmatched agents and \(\begin{equation*} p^{\prime }_a= p^*_a- \frac{\Delta ^{\text{UCB}}}{2 |\mathcal {A}|} = \tau ^*_a+ u^{\prime }_a(\mu _{X^*}(a)) \end{equation*}\) for matched agents.

Proving that \((X^*, \tau ^*)\) is stable with respect to u. We show the stability \((X^*, \tau ^*)\) with respect to u by checking that individual rationality holds and that there are no blocking pairs.

The main fact that we will use is that \(\begin{equation*} u_a(\mu _{X^*}(a)) \ge u^{\prime }_a(\mu _{X^*}(a)). \end{equation*}\) To prove this, we split into two cases: (i) agent a is matched in \(X^*\) (i.e., \(\mu _{X^*}(a)\ne a\)), and (ii) agent a is not matched by \(X^*\). For (i), if a is matched by \(X^*\), then \(\begin{equation*} u_a(\mu _{X^*}(a)) \ge u^{\mathrm{UCB}}_a(\mu _{X^*}(a)) - 0.1{} \frac{\Delta }{|\mathcal {A}|} \ge u^{\mathrm{UCB}}_a(\mu _{X^*}(a)) - \frac{\Delta ^{\mathrm{UCB}}}{|\mathcal {A}|} = u^{\prime }_a(\mu _{X^*}(a)). \end{equation*}\) For (ii), if a is not matched by \(X^*\), then \(u_a(\mu _{X^*}(a))\ge u^{\prime }_a(\mu _{X^*}(a)),\) because both sides are 0.

For individual rationality, we thus have \(\begin{align*} u_a(\mu _{X^*}(a)) + \tau ^*_a&\ge u^{\prime }_a(\mu _{X^*}(a)) + \tau ^*_a\ge 0, \end{align*}\) where the second inequality comes from the individual rationality of \((X^*, \tau ^*)\) with respect to \(u^{\prime }\).

Let us next show that there are no blocking pairs. If \((i, j)\in X^*\), then we see that: \(\begin{equation*} u_i(\mu _{X^*}(i)) + \tau ^*_i+ u_{j}(\mu _{X^*}(j)) + \tau ^*_{j} = u_i(\mu _{X^*}(i)) +u_{j}(\mu _{X^*}(j)), \end{equation*}\) as desired. Next, consider any pair \((i, j)\not\in X^*\). Then, \(\begin{equation*} u_i(j) + u_j(i) \le u^{\mathrm{UCB}}_i(j) + u^{\mathrm{UCB}}_j(i) = u^{\prime }_i(j) + u^{\prime }_j(i). \end{equation*}\) It follows that \(\begin{align*} u_i(\mu _{X^*}(i)) + \tau ^*_i+ u_{j}(\mu _{X^*}(j)) + \tau ^*_{j} &\ge u^{\prime }_i(\mu _{X^*}(i)) + \tau ^*_i+ u_{j}(\mu _{X^*}(j)) + \tau ^*_{j} \\ &\ge u^{\prime }_i(j) + u^{\prime }_j(i) \\ &\ge u_i(j) + u_j(i), \end{align*}\) where the second inequality comes from the fact that \((X^*, \tau ^*)\) has no blocking pairs with respect to \(u^{\prime }\).

This completes our proof that \((X^*, \tau ^*)\) is stable with respect to u.□

Now, we are ready to prove Theorem 6.1.

Proof of Theorem 6.1

As in the proof of Theorem 5.1, the starting point for our proof is the typical approach in multi-armed bandits and combinatorial bandits [15, 22, 30] of bounding regret in terms of the sizes of the confidence interval of the chosen arms. Our approach does not quite compose cleanly with these proofs, since we need to handle the transfers in addition to the matching.

We divide in two cases, based on the event E that all of the confidence sets contain their respective true utilities at every timestep \(t\le T\). That is, \(u_{i}(j)\in C_{i,j}\) and \(u_j(i)\in C_{j,i}\) for all \((i,j)\in \mathcal {I}\times \mathcal {J}\) at all t.

Case 1: E holds. Let \(n_{ij}^{t}\) be the number of times that the pair \((i,j)\) has been matched by round t. For each pair \((i, j)\), we maintain a “blame” counter \(b^t_{ij}\). We will ultimately bound the total number of timesteps where the algorithm chooses a matching that is not stable by \(\sum _{(i,j)} b_{i,j}^T\).

We increment the blame counters as follows. First, suppose that \(\begin{equation*} \max (C_{a, \mu _{X^t}(a)}) - \min (C_{a, \mu _{X^t}(a)}) \le 0.1{}\frac{\Delta }{|\mathcal {A}|} \end{equation*}\) for every matched agent \(a\in \mathcal {A}\). By Lemma 6.2 and since the event E holds, we know the chosen matching is stable and thus incurs 0 regret. We do not increment any of the blame counters in this case. Now, suppose that \(\begin{equation*} \max (C_{a, \mu _{X^t}(a)}) - \min (C_{a, \mu _{X^t}(a)}) \gt 0.1{}\frac{\Delta }{ |\mathcal {A}|} \end{equation*}\) for some matched agent a. We increment the counter of the least-blamed pair \((i, j) \in X^t\).

We now bound the blame counter \(b^T_{ij}\). We use the fact that the blame counter is only incremented when the corresponding confidence set is sufficiently large, and that a new sample of the utilities is received whenever the blame counter is incremented. This means that: \(\begin{equation*} b^T_{ij} = O\left(\frac{|\mathcal {A}|^2 \log (|\mathcal {A}| T))}{\Delta ^2} \right). \end{equation*}\) The maximum regret incurred by any matching is at most \(12 |\mathcal {A}|\), which means that the regret incurred by this case is at most: \(\begin{equation*} 12 |\mathcal {A}| \sum _{(i, j)} b^T_{ij} \le 12 |\mathcal {A}| \sum _{(i, j)} O\left(\frac{|\mathcal {A}|^2 \log (|\mathcal {A}| T))}{\Delta ^2}\right) = O \left(\frac{|\mathcal {A}|^5 \log (|\mathcal {A}| T))}{\Delta ^2} \right). \end{equation*}\)

Case 2: E does not hold. Since each \(n_{ij}(\hat{u}_i(j) -u_i(j))\) is mean-zero and 1-subgaussian and we have \(O(|\mathcal {I}||\mathcal {J}|T)\) such random variables over T rounds, the probability that any of them exceeds \(\begin{equation*} 2\sqrt {\log (|\mathcal {I}||\mathcal {J}|T/\delta)}\le 2\sqrt {\log (|\mathcal {A}|^2T/\delta)} \end{equation*}\) is at most \(\delta\) by a standard tail bound for the maximum of subgaussian random variables. It follows that E fails to hold with probability at most \(|\mathcal {A}|^{-2}T^{-2}\). In the case that E fails to hold, our regret in any given round would be at most \(12|\mathcal {A}|\) by the Lipschitz property in Proposition 4.4. (Recall that our upper confidence bound is off by at most 6 due to clipping the confidence interval to lie in \([-1, 1]\), so the expanded confidence sets also necessarily lie in \([-3, 3]\).) Thus, the expected regret from this scenario is at most \(\begin{equation*} |\mathcal {A}|^{-2}T^{-2}\cdot 12|\mathcal {A}|T\le 12|\mathcal {A}|^{-1}T^{-1}, \end{equation*}\) which is negligible compared to the regret bound from when E does occur.□

D.3 Instance-independent Regret Bounds for MatchUCB\(^{\prime }\)

To establish instance-independent regret bounds for MatchUCB\(^{\prime }\), we show that \((X^{*}, p^{*})\) is indeed optimal with respect to \(u^{\mathrm{UCB}}\); the remainder then follows the same argument as Theorem 5.1.

Lemma D.2 The pair \((X^{*}, p^{*})\) is an optimal primal-dual pair with respect to \(u^{\mathrm{UCB}}\).

Proof.

It suffices to verify feasibility and, by weak duality, check that \(X^*\) and \(p^*\) achieve the same objective value. It is clear that \(X^*\) is primal feasible. For dual feasibility, if \((i, j)\not\in X^*\), then \(\begin{equation*} p^*_i+ p^*_j\ge p^{\prime }_i+ p^{\prime }_j\ge u^{\prime }_i(j) + u^{\prime }_j(i) = u^{\mathrm{UCB}}_i(j) + u^{\mathrm{UCB}}_j(i); \end{equation*}\) and if \((i, j)\in X^*\), then \(\begin{equation*} p^*_i+ p^*_j= p^{\prime }_i+ p^{\prime }_j+ 2\frac{\Delta ^{\text{UCB}}}{|\mathcal {A}|}\ge u^{\prime }_i(j) + u^{\prime }_j(i) + 2\frac{\Delta ^{\text{UCB}}}{|\mathcal {A}|} = u^{\mathrm{UCB}}_i(j) + u^{\mathrm{UCB}}_j(i). \end{equation*}\) Finally, we check that they achieve the same objective value with respect to \(u^{\mathrm{UCB}}\). By D.1 and strong duality, \(X^*\) achieves the same objective value as \(p^{\prime }\) with respect to \(u^{\prime }\). Hence,

Skip EProofs for Section 6.2 Section

E Proofs for Section 6.2

Theorem 6.3. For preference class \(\hspace{1.111pt}\mathcal {U}_{\mathrm{unstructured}}\) (see Section 3), there exists an algorithm giving the platform \(\begin{equation*} \varepsilon T \sum _{t=1}^T |\mathcal {A}_t| - O\Bigl (|\mathcal {A}| \sqrt {n T} \sqrt {\log (|\mathcal {A}| |T|)}\Bigr) \end{equation*}\) revenue in the presence of search frictions while maintaining stability with high probability.

Proof of Theorem 6.3

The algorithm is defined as follows: We set confidence sets according to MatchUCB and run essentially that algorithm, but with a modified ComputeMatch. Instead of ComputeMatch, we use the following algorithm. The platform first computes a matching with transfers \((X^*, \tau ^*)\) according to the UCB estimates \(u^{\mathrm{UCB}}\), like before. Then, the platform chooses \(X^*\) to be the selected matching and sets the transfers according to: \(\begin{equation*} \tau _a= \tau ^*_a- \varepsilon + \max (C_{a, \mu _{X}(a))}) - \min (C_{a, \mu _{X}(a)}). \end{equation*}\) This choice of transfers has a clean economic intuition: Agents should be compensated based on the platform’s uncertainty about their utilities with \(\varepsilon\) of their transfer shaved off as revenue for the platform.

First, we show that if the confidence sets contain the true utilities, then \((X^*, \tau)\) is \(\varepsilon\)-stable. It suffices to show that \((X^*, \tau ^{\prime })\) where: \(\begin{equation*} \tau ^{\prime }_a= \tau ^*_a+ \max (C_{a, \mu _{X}(a))}) - \min (C_{a, \mu _{X}(a)}) \end{equation*}\) is stable. First, we see that \(\begin{equation*} u_a(\mu _{X^\mathrm{UCB}}(a)) + \tau ^{\prime }_a= u^{\mathrm{UCB}}_a(\mu _{X^\mathrm{UCB}}(a)) + \tau ^*_a\ge 0, \end{equation*}\) since \((X, \tau ^*)\) is stable with respect to \(u^{\mathrm{UCB}}\). Furthermore, we see that: \(\begin{align*} \bigl (u_i(\mu _{X}(i)) + \tau ^{\prime }_i\bigr) + \bigl (u_j(\mu _{X}(j)) + \tau ^{\prime }_j\bigr) &\ge \bigl (u^{\mathrm{UCB}}_i(\mu _{X}(i)) + \tau ^*_i\bigr) + \bigl (u^{\mathrm{UCB}}_j(\mu _{X}(j)) + \tau ^*_j\bigr) \\ &\ge u^{\mathrm{UCB}}_i(j) + u^{\mathrm{UCB}}_j(i) \\ &\ge u_i(j) + u_j(i), \end{align*}\) where the second-to-last line follows from the fact that \((X, \tau ^*)\) is stable with respect to \(u^{\mathrm{UCB}}\).

We first show that s is a feasible solution to (\(\dagger\)): \(\begin{align*} &{\min \bigl (u_i(j) - u_i(\mu _{X^\mathrm{UCB}}(i))- s_i, \tilde{u}_j(i) - u_j(\mu _{X^\mathrm{UCB}}(j)) - s_j\bigr)} \\ &= \min \bigl (u_i(j) - u^{\mathrm{UCB}}_i(\mu _{X^\mathrm{UCB}}(i)), \tilde{u}_j(i) - u^{\mathrm{UCB}}_j(\mu _{X^\mathrm{UCB}}(j))\bigr) \\ &\le \min \bigl (u^{\mathrm{UCB}}_i(j) - u^{\mathrm{UCB}}_i(\mu _{X^\mathrm{UCB}}(i)), u^{\mathrm{UCB}}_j(i) - u^{\mathrm{UCB}}_j(\mu _{X^\mathrm{UCB}}(j))\bigr) \\ &\le 0, \end{align*}\) where the last step uses the fact that \(\mu _{X^\mathrm{UCB}}\) is stable with respect to \(u^{\mathrm{UCB}}\) by definition. Moreover, we see that \(\begin{equation*} u_a(\mu _{X^\mathrm{UCB}}(a)) + s_a= u^{\mathrm{UCB}}_a(\mu _{X^\mathrm{UCB}}(a)) \ge 0, \end{equation*}\) where the last inequality uses that \(\mu _{X^\mathrm{UCB}}\) is stable with respect to \(u^{\mathrm{UCB}}\) by definition. This implies that s is feasible.

We see that the platform’s revenue is equal to: \(\begin{align*} -\sum _{t=1}^T \sum _{a\in \mathcal {A}_t} \tau _a&= -\sum _{t=1}^T \sum _{a\in \mathcal {A}_t} \tau ^*_a+ \sum _{t=1}^T \sum _{a\in \mathcal {A}_t} \varepsilon + \sum _{t=1}^T \sum _{a\in \mathcal {A}_t} (\max (C_{a, \mu _{X}(a))}) - \min (C_{a, \mu _{X}(a)})) \\ &= \varepsilon \sum _{t=1}^T |\mathcal {A}_t| - \sum _{t=1}^T \sum _{a\in \mathcal {A}_t} (\max (C_{a, \mu _{X}(a))}) - \min (C_{a, \mu _{X}(a)})). \end{align*}\) Using the proof of Theorem 5.1, we see that \(\begin{equation*} \sum _{t=1}^T \sum _{a\in \mathcal {A}_t} (\max (C_{a, \mu _{X}(a))}) - \min (C_{a, \mu _{X}(a)})) \le O(|\mathcal {A}| \sqrt {n T} \log (|\mathcal {A}| T)), \end{equation*}\) as desired.□

Skip FProofs for Section 6.3 Section

F Proofs for Section 6.3

F.1 Proof of Proposition 6.5

Proof of Proposition 6.5

We first prove the first part of the statement, and then the second part of the statement.

Proof of part (a). We note that it follows immediately from Definition 6.4 that NTU Subset Instability is nonnegative. Let us now show that \(I(X; u, \mathcal {A})\) is zero if and only if \((X, \tau)\) is stable. It is not difficult to see that the infimum of (\(\dagger\)) is attained at some \(s^*\).

If \(I(X; u, \mathcal {A}) = 0\), then we know that \(s^*_a= 0\) for all \(a\in \mathcal {A}\). The constraints in the optimization problem imply that X has no blocking pairs and individual rationality is satisfied, as desired.

If X is stable, then we see that \(s= \vec{0}\) is a feasible solution to (\(\dagger\)), which means that the optimum of (\(\dagger\)) is at most zero. This, coupled with the fact that \(I(X; u, \mathcal {A})\) is always nonnegative, means that \(I(X; u, \mathcal {A}) = 0,\) as desired.

Proof of part (b). Consider two utility functions u and \(\tilde{u}\). To show Lipchitz continuity, it suffices to show that for any matching X: \(\begin{equation*} |I(X; u, \mathcal {A}) - I(X; \tilde{u}, \mathcal {A})| \le 2\sum _{a\in \mathcal {A}} {\Vert }u_a- \tilde{u}_a{\Vert }_\infty . \end{equation*}\) We show that: \(\begin{equation*} I(X; \tilde{u}, \mathcal {A}) \le I(X; u, \mathcal {A}) + 2\sum _{a\in \mathcal {A}} {\Vert }u_a- \tilde{u}_a{\Vert }_\infty , \end{equation*}\) noting that the other direction follows from an analogous argument. Let \(s^*\) be an optimal solution to (\(\dagger\)) for the utilities u. Consider the solution \(s_a= s^*_a+ 2 {\Vert }u_a- u_a{\Vert }_{\infty }\). We first verify that s is a feasible solution to (\(\dagger\)) for \(\tilde{u}\). We see that: \(\begin{align*} & \min \bigl (\tilde{u}_i(j) - \tilde{u}_i(\mu _{X}(i)) - s_i, \tilde{u}_j(i) - \tilde{u}_j(\mu _{X}(j)) - s_j\bigr) \\ &= \min \bigl (\tilde{u}_i(j) - \tilde{u}_i(\mu _{X}(i)) - s^*_i- 2 {\Vert }u_i- \tilde{u}_i{\Vert }_{\infty }, \tilde{u}_j(i) - \tilde{u}_j(\mu _{X}(j)) - s^*_j- 2 {\Vert }u_j- \tilde{u}_j{\Vert }_{\infty }\bigr) \\ &\le \min \bigl (u_i(j) - u_i(\mu _{X}(i)) - s^*_i, u_j(i) - u_j(\mu _{X}(j)) - s^*_j\bigr) \\ &\le 0, \end{align*}\) as desired. Moreover, we see that \(\begin{equation*} \tilde{u}_a(\mu _{X}(a)) + s_a= \tilde{u}_a(\mu _{X}(a)) + s^*_a+ 2 {\Vert }u_a- u_a{\Vert }_{\infty } \le u_a(\mu _{X}(a)) + s^*_a\ge 0. \end{equation*}\) Thus, we have demonstrated that s is feasible. This means that: \(\begin{equation*} I(X; \tilde{u}, \mathcal {A}) \le \sum _{a\in \mathcal {A}} s_a= \sum _{a\in \mathcal {A}} s^*_a+ 2\sum _{a\in \mathcal {A}} {\Vert }u_a- \tilde{u}_a{\Vert }_\infty = [I(X; u, \mathcal {A}) + 2\sum _{a\in \mathcal {A}} {\Vert }u_a- \tilde{u}_a{\Vert }_\infty , \end{equation*}\) as desired.□

F.2 Proof of Theorem 6.6

We show that the algorithmic approach from Section 5 can be adapted to the setting of matching with non-transferable utilities.

Drawing intuition from Section 5, at each round, we compute a stable matching for utilities given by the upper confidence bounds. More precisely, suppose we have a collection \(\mathscr{C}\) of confidence sets \(C_{i, j}, C_{j, i}\subseteq \mathbb {R}\) such that \(u_i(j)\in C_{i, j}\) and \(u_j(i)\in C_{j, i}\) for all \((i, j) \in \mathcal {I}\times \mathcal {J}\). Our algorithm uses \(\mathscr{C}\) to get an upper confidence bound for each agent’s utility function and then computes a stable matching with transfers as if these upper confidence bounds were the true utilities (see ComputeMatchNTU). This can be implemented efficiently if we use, e.g., the Gale-Shapley algorithm (either the customer-proposing algorithm or the provider-proposing algorithm will work).

The core property of ComputeMatchNTU is that we can upper bound NTU Subset Instability by the sum of the sizes of the relevant confidence sets, assuming that the confidence sets contain the true utilities.

Proposition F.1 Consider a collection confidence sets \(\mathscr{C}\) such that \(u_{i}(j) \in C_{i, j}\) and \(u_{j}(i) \in C_{j,i}\) for all \((i, j) \in \mathcal {I}\times \mathcal {J}\). The instability of the output \(X^\mathrm{UCB}\) of ComputeMatch satisfies (14) \(\begin{equation} I(X^\mathrm{UCB}; u, \mathcal {A}) \le \sum _{a\in \mathcal {A}^t} (\max (C_{a, \mu _{X^\mathrm{UCB}}(a)}) - \min (C_{a, \mu _{X^\mathrm{UCB}}(a)})). \end{equation}\)

Proof.

We construct subsidies for this setting to be: \(\begin{equation*} s_a= \max (C_{a, \mu _{X}(a)}) - u_a(\mu _{X}(a))\le \max (C_{a, \mu _{X}(a)}) - \min (C_{a, \mu _{X}(a)}). \end{equation*}\)

Step 1: Verifying feasibility. We first show that s is a feasible solution to (\(\dagger\)). \(\begin{align*} & \min \bigl (u_i(j) - u_i(\mu _{X^\mathrm{UCB}}(i))- s_i, \tilde{u}_j(i) - u_j(\mu _{X^\mathrm{UCB}}(j)) - s_j\bigr) \\ &= \min \bigl (u_i(j) - u^{\mathrm{UCB}}_i(\mu _{X^\mathrm{UCB}}(i)), \tilde{u}_j(i) - u^{\mathrm{UCB}}_j(\mu _{X^\mathrm{UCB}}(j))\bigr) \\ &\le \min \bigl (u^{\mathrm{UCB}}_i(j) - u^{\mathrm{UCB}}_i(\mu _{X^\mathrm{UCB}}(i)), u^{\mathrm{UCB}}_j(i) - u^{\mathrm{UCB}}_j(\mu _{X^\mathrm{UCB}}(j))\bigr) \\ &\le 0, \end{align*}\) where the last step uses the fact that \(\mu _{X^\mathrm{UCB}}\) is stable with respect to \(u^{\mathrm{UCB}}\) by definition. Moreover, we see that \(\begin{equation*} u_a(\mu _{X^\mathrm{UCB}}(a)) + s_a= u^{\mathrm{UCB}}_a(\mu _{X^\mathrm{UCB}}(a)) \ge 0, \end{equation*}\) where the last inequality uses that \(\mu _{X^\mathrm{UCB}}\) is stable with respect to \(u^{\mathrm{UCB}}\) by definition. This implies that s is feasible.

Step 2: Computing the objective. We next compute the objective of (\(\dagger\)) at s and use this to bound \(I(X^*; u, \mathcal {A})\). A simple calculation shows that: \(\begin{equation*} I(X^*; u, \mathcal {A}) \le \sum _as_a= \sum _{a\in \mathcal {A}} (\max (C_{a, \mu _{X^\mathrm{UCB}}(a)}) - \min (C_{a, \mu _{X^\mathrm{UCB}}(a)})), \end{equation*}\) as desired.□

F.2.1 Explicit Algorithm and Regret Bounds.

Using the same intuition as Section 5, the regret bound of F.1 hints at an algorithm: Each round, select the matching with transfers returned by ComputeMatchNTU and update confidence sets accordingly. To instantiate this approach, it remains to construct confidence intervals that contain the true utilities with high probability.

We showcase this algorithm in the simple setting of unstructured preferences. For this setting, we can construct our confidence intervals following the classical UCB approach. That is, for each utility value involving the pair \((i, j)\), we take a length \(O(\sqrt {\log (|\mathcal {A}|T)} / n_{ij})\) confidence interval centered around the empirical mean, where \(n_{ij}\) is the number of times the pair has been matched before. We describe this construction precisely in Algorithm 3 (MatchNTUUCB).

To analyze MatchNTUUCB, recall that Lemma 5.4 bounds the regret at each step by the lengths of the confidence intervals of each pair in the selected matching. Like in Section 5, this yields the following instance-independent regret bound:

Theorem F.2 MatchNTUUCB incurs expected regret \(\mathbb {E}(R_T) \le O\bigl ({|\mathcal {A}|}^{3/2} \sqrt {T} \sqrt {\log (|\mathcal {A}| T)}\bigr)\).

Proof.

This proof proceeds very similarly to the proof of Theorem 5.1. We consider the event E that all of the confidence sets contain their respective true utilities at every timestep \(t\le T\). That is, \(u_{i}(j)\in C_{i, j}\) and \(u_j(i)\in C_{j, i}\) for all \((i,j)\in \mathcal {I}\times \mathcal {J}\) at all t.

Case 1: E holds. By Lemma 5.4, we may bound

where \(n_{ij}^{t}\) is the number of times that the pair \((i,j)\) has been matched at the start of round t. Let \(w^t_{i, j} = \frac{1}{\sqrt {n_{ij}^{t}}}\) be the size of the confidence set (with the log factor scaled out) for \((i, j)\) at the start of round t.

At each timestep t, let us consider the list consisting of \(w^t_{i_t, j_t}\) for all \((i_t, j_t) \in X^t\). Let us now consider the overall list consisting of the concatenation of all of these lists over all rounds. Let us order this list in decreasing order to obtain a list \(\tilde{w}_1, \ldots , \tilde{w}_L\) where \(L = \sum _{t=1}^{T} |X^t| \le n T\). In this notation, we observe that: \(\begin{equation*} \sum _{t=1}^{T} I(X^t; u, \mathcal {A}^t) \le \sum _{t=1}^{T} \sum _{a\in \mathcal {A}^t} (\max (C_{a, \mu _{X^t}(a)}) - \min (C_{a, \mu _{X^t}(a)})) = \log (|\mathcal {A}| T) \sum _{l=1}^L \tilde{w}_l. \end{equation*}\) We claim that \(\tilde{w}_l \le O({\min (1, \frac{1}{\sqrt {(l / |\mathcal {A}|^2) -1}})})\). The number of rounds that a pair of agents can have their confidence set have size at least \(\tilde{w}_l\) is upper bounded by \(1 + \frac{1}{\tilde{w}_l^2}\). Thus, the total number of times that any confidence set can have size at least \(\tilde{w}_l\) is upper bounded by \((|\mathcal {A}|^2)(1 + \frac{1}{\tilde{w}_l^2})\).

Putting this together, we see that:

Case 2: E does not hold. Since each \(n_{ij}(\hat{u}_i(j) -u_i(j))\) is mean-zero and 1-subgaussian, and we have \(O(|\mathcal {I}||\mathcal {J}|T)\) such random variables over T rounds, the probability that any of them exceeds \(\begin{equation*} 2\sqrt {\log (|\mathcal {I}||\mathcal {J}|T/\delta)}\le 2\sqrt {\log (|\mathcal {A}|^2T/\delta)} \end{equation*}\) is at most \(\delta\) by a standard tail bound for the maximum of subgaussian random variables. It follows that E fails to hold with probability at most \(|\mathcal {A}|^{-2}T^{-2}\). In the case that E fails to hold, our regret in any given round would be at most \(4|\mathcal {A}|\) by the Lipschitz property in Proposition 6.5. (Recall that our upper confidence bound for any utility is wrong by at most two due to clipping each confidence interval to lie in \([-1, 1]\).) Thus, the expected regret from this scenario is at most \(\begin{equation*} |\mathcal {A}|^{-2}T^{-2}\cdot 4|\mathcal {A}|T\le 4|\mathcal {A}|^{-1}T^{-1}, \end{equation*}\) which is negligible compared to the regret bound from when E does occur.□

Footnotes

  1. 1 Feedback might arise from explicit sources (e.g., riders rating drivers after a Lyft ride) or implicit sources (e.g., engagement metrics on an app); in either case, feedback is likely to be sparse and noisy.

    Footnote
  2. 2 Previous work [17, 32] has investigated utility difference (i.e., the difference between the total utility achieved by the selected matching and the utility achieved by a stable matching) as a measure of regret. However, this does not capture distance from equilibrium in matching markets with monetary transfers (see Section 4) or without monetary transfers (see Section 6.3.1).

    Footnote
  3. 3 This formulation is inspired by the strong \(\varepsilon\)-core of Shapley and Shubik [41].

    Footnote
  4. 4 We observe that (2) corresponds to no pair of agents \((i, j)\) being able to agree on a transfer such that both would rather match with each other than abide by \((X, \tau)\). Notice that a pair \((i, j)\) violates (2) if and only if they can find a transfer \(\tau _i^{\prime } = -\tau _j^{\prime }\) such that \(u_i(j) + \tau _i^{\prime } \gt u_i(\mu _{X}(i)) + \tau _i\) and \(u_j(i) + \tau _j^{\prime } \gt u_j(\mu _{X}(j)) + \tau _j\).

    Footnote
  5. 5 Our feedback model corresponds to semi-bandit feedback, since the platform has (noisy) access to each agent’s utility within the matching rather than the overall utility of the matching.

    Footnote
  6. 6 Utility difference is standard as a measure of regret for learning a maximum weight matching in the combinatorial bandits literature (see, e.g., Gai et al. [22]). However, we show that for learning stable matchings, a fundamentally different measure of regret is needed.

    Footnote
  7. 7 The requirement that \(s_a \ge 0\) enforces that all subsidies are nonnegative; without it, (5) would reduce to the utility difference, which is not incentive-aware.

    Footnote
  8. 8 In this linear program, the first set of constraints ensures there are no blocking pairs, while the second set of constraints ensures individual rationality.

    Footnote
  9. 9 This bound can be achieved by adapting the explore-then-commit (ETC) approach, where the platform explores by choosing each pair of agents \(\widetilde{O}((T/N)^{2/3})\) times [30]. Thus, \(\widetilde{O}(N^{1/3} T^{2/3})\) rounds are spent exploring, and the Subset Instability of the matching selected in the commit phase is \(\widetilde{O}(N^{4/3} T^{2/3})\) with high probability. We omit further details, since this analysis is a straightforward adaptation of the typical ETC analysis.

    Footnote
  10. 10 For intuition, consider the classical stochastic multi-armed bandits setting and suppose that we could only guarantee that the loss incurred by an arm is bounded by the maximum of the sizes of the confidence sets over all arms. Then, we would only be able to obtain a weak bound on regret, since low-reward arms with large confidence sets may never be pulled.

    Footnote
  11. 11 Our bound is less fine-grained than the gap in Chen et al. [15], and in particular does not allow there to be multiple maximum weight matchings. We defer improving our definition of \(\Delta\) to future work.

    Footnote
  12. 12 The instance-independent regret bound can be shown using the same argument as the proof for Theorem 5.1.

    Footnote
  13. 13 The number of mistakes necessarily depends on the gap \(\Delta ,\) because there exist utility functions u and \(\tilde{u}\) where \({\Vert }u - \tilde{u}{\Vert }_{\infty }\) is arbitrarily small, but where the stable market outcomes with respect to u and \(\tilde{u}\) differ. To see this, consider a market where \(\mathcal {I}= \lbrace C\rbrace\) and \(\mathcal {J}= \lbrace P\rbrace\). Suppose that \(u_C(P) = \tilde{u}_C(P) = 1\), while \(u_{P}(C) = -1 + \varepsilon\) and \(\tilde{u}_P(C) = -1 - \varepsilon\). Then, the maximum weight matchings under these utility functions differ: \(\lbrace (C, P)\rbrace\) is the only maximum weight matching in the former, whereas \(\emptyset\) is the only maximum weight matching in the latter.

    Footnote
  14. 14 This definition corresponds to \((X, \tau)\) belonging to the weak \(\varepsilon\)-core of Shapley and Shubik [41]. We note that this definition also relaxes individual rationality. This formulation gives us the cleanest algorithmic results; while it can be extended to an analogue that does not relax individual rationality, it would involve bounds that (necessarily) depend on the specifics of agents’ utilities.

    Footnote
  15. 15 See Chegireddy and Hamacher [14] for efficient algorithms to compute the second-best matching.

REFERENCES

  1. [1] Abernethy Jacob D., Hazan Elad, and Rakhlin Alexander. 2008. Competing in the dark: An efficient algorithm for bandit linear optimization. In 21st Annual Conference on Learning Theory (COLT’08). Omnipress, 263274.Google ScholarGoogle Scholar
  2. [2] Alston Max. 2020. On the non-existence of stable matches with incomplete information. Games Econ. Behav. 120 (2020), 336344.Google ScholarGoogle ScholarCross RefCross Ref
  3. [3] Aridor Guy, Mansour Yishay, Slivkins Aleksandrs, and Wu Zhiwei Steven. 2020. Competing bandits: The perils of exploration under competition. CoRR abs/2007.10144 (2020).Google ScholarGoogle Scholar
  4. [4] Ashlagi Itai, Braverman Mark, Kanoria Yash, and Shi Peng. 2020. Clearing matching markets efficiently: Informative signals and match recommendations. Manag. Sci. 66, 5 (2020), 21632193.Google ScholarGoogle ScholarCross RefCross Ref
  5. [5] Auer Peter, Cesa-Bianchi Nicolò, and Fischer Paul. 2002. Finite-time analysis of the multiarmed bandit problem. Mach. Learn. 47, 2-3 (2002), 235256.Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. [6] Auer Peter, Cesa-Bianchi Nicolò, Freund Yoav, and Schapire Robert E.. 2002. The nonstochastic multiarmed bandit problem. SIAM J. Comput. 32, 1 (2002), 4877.Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. [7] Azevedo Eduardo M. and Hatfield John William. 2018. Existence of equilibrium in large matching markets with complementarities. Available at SSRN (2018). https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3268884.Google ScholarGoogle Scholar
  8. [8] Badanidiyuru Ashwinkumar, Kleinberg Robert, and Slivkins Aleksandrs. 2018. Bandits with knapsacks. J. ACM 65, 3 (2018), 13:1–13:55.Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. [9] Basu Soumya, Sankararaman Karthik Abinav, and Sankararaman Abishek. 2021. Beyond \(\log ^2(T)\) regret for decentralized bandits in matching markets. CoRR abs/2103.07501 (2021).Google ScholarGoogle Scholar
  10. [10] Bichler Martin, Fichtl Maximilian, and Schwarz Gregor. 2021. Walrasian equilibria from an optimization perspective: A guide to the literature. Naval Res. Logist. 68, 4 (2021), 496513.Google ScholarGoogle ScholarCross RefCross Ref
  11. [11] Bikhchandani Sushil. 2017. Stability with one-sided incomplete information. J. Econ. Theor. 168 (2017), 372399.Google ScholarGoogle ScholarCross RefCross Ref
  12. [12] Cen Sarah H. and Shah Devavrat. 2021. Regret, stability, and fairness in matching markets with bandit learners. CoRR abs/2102.06246 (2021).Google ScholarGoogle Scholar
  13. [13] Cesa-Bianchi Nicolò and Lugosi Gábor. 2012. Combinatorial bandits. J. Comput. Syst. Sci. 78, 5 (2012), 14041422.Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. [14] Chegireddy Chandra R. and Hamacher Horst W.. 1987. Algorithms for finding K-best perfect matchings. Discret. Appl. Math. 18, 2 (1987), 155165.Google ScholarGoogle ScholarCross RefCross Ref
  15. [15] Chen Wei, Wang Yajun, and Yuan Yang. 2013. Combinatorial multi-armed bandit: General framework and applications. In 30th International Conference on Machine Learning(JMLR Workshop and Conference Proceedings, Vol. 28). JMLR.org, 151159.Google ScholarGoogle Scholar
  16. [16] Combes Richard, Talebi Mohammad Sadegh, Proutière Alexandre, and Lelarge Marc. 2015. Combinatorial bandits revisited. In Annual Conference on Neural Information Processing Systems. 21162124.Google ScholarGoogle Scholar
  17. [17] Das Sanmay and Kamenica Emir. 2005. Two-sided bandits and the dating market. In 19th International Joint Conference on Artificial Intelligence. Professional Book Center, 947952.Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. [18] Debreu Gerard and Scarf Herbert. 1963. A limit theorem on the core of an economy. Int. Econ. Rev. 4, 3 (1963), 235246.Google ScholarGoogle ScholarCross RefCross Ref
  19. [19] Echenique Federico, Lee Sangmok, Shum Matthew, and Yenmez M. Bumin. 2013. The revealed preference theory of stable and extremal stable matchings. Econometrica 81, 1 (2013), 153171.Google ScholarGoogle ScholarCross RefCross Ref
  20. [20] Emamjomeh-Zadeh Ehsan, Gonczarowski Yannai A., and Kempe David. 2020. The complexity of interactively learning a stable matching by trial and error. In 21st ACM Conference on Economics and Computation. ACM.Google ScholarGoogle ScholarDigital LibraryDigital Library
  21. [21] Frazier Peter I., Kempe David, Kleinberg Jon M., and Kleinberg Robert. 2014. Incentivizing exploration. In ACM Conference on Economics and Computation (EC’14). ACM, 522.Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. [22] Gai Yi, Krishnamachari Bhaskar, and Jain Rahul. 2012. Combinatorial network optimization with unknown variables: Multi-armed bandits with linear rewards and individual observations. IEEE/ACM Trans. Netw. 20, 5 (2012), 14661478.Google ScholarGoogle ScholarDigital LibraryDigital Library
  23. [23] Gale D. and Shapley L. S.. 1962. College admissions and the stability of marriage. Amer. Math. Month. 69, 1 (1962), 915.Google ScholarGoogle ScholarCross RefCross Ref
  24. [24] Gonczarowski Yannai A., Nisan Noam, Ostrovsky Rafail, and Rosenbaum Will. 2019. A stable marriage requires communication. Games Econ. Behav. 118 (2019), 626647.Google ScholarGoogle ScholarCross RefCross Ref
  25. [25] Immorlica Nicole, Sankararaman Karthik Abinav, Schapire Robert E., and Slivkins Aleksandrs. 2019. Adversarial bandits with knapsacks. In 60th IEEE Annual Symposium on Foundations of Computer Science (FOCS’19). IEEE Computer Society, 202219.Google ScholarGoogle Scholar
  26. [26] Johari Ramesh, Kamble Vijay, and Kanoria Yash. 2021. Matching while learning. Oper. Res. 69, 2 (2021), 655681.Google ScholarGoogle ScholarCross RefCross Ref
  27. [27] Kleinberg Robert D. and Leighton Frank Thomson. 2003. The value of knowing a demand curve: Bounds on regret for online posted-price auctions. In 44th Symposium on Foundations of Computer Science (FOCS’03). IEEE Computer Society, 594605.Google ScholarGoogle Scholar
  28. [28] Kuhn H. W.. 1955. The Hungarian method for the assignment problem. Naval Res. Logist. Quart. 2, 1–2 (1955), 8397.Google ScholarGoogle ScholarCross RefCross Ref
  29. [29] Kveton Branislav, Wen Zheng, Ashkan Azin, and Szepesvári Csaba. 2015. Tight regret bounds for stochastic combinatorial semi-bandits. In 18th International Conference on Artificial Intelligence and Statistics(JMLR Workshop and Conference Proceedings, Vol. 38). JMLR.org.Google ScholarGoogle Scholar
  30. [30] Lattimore Tor and Szepesvári Csaba. 2020. Bandit Algorithms. Cambridge University Press.Google ScholarGoogle ScholarCross RefCross Ref
  31. [31] Li Xiaocheng, Sun Chunlin, and Ye Yinyu. 2021. The symmetry between arms and knapsacks: A primal-dual approach for bandits with knapsacks. CoRR abs/2102.06385 (2021).Google ScholarGoogle Scholar
  32. [32] Liu Lydia T., Mania Horia, and Jordan Michael I.. 2020. Competing bandits in matching markets. In 23rd International Conference on Artificial Intelligence and Statistics(Proceedings of Machine Learning Research, Vol. 108). PMLR, 16181628.Google ScholarGoogle Scholar
  33. [33] Liu Lydia T., Ruan Feng, Mania Horia, and Jordan Michael I.. 2020. Bandit learning in decentralized matching markets. CoRR abs/2012.07348 (2020).Google ScholarGoogle Scholar
  34. [34] Liu Qingmin. 2020. Stability and Bayesian consistency in two-sided markets. Amer. Econ. Rev. 110, 8 (Aug.2020).Google ScholarGoogle ScholarCross RefCross Ref
  35. [35] Liu Qingmin, Mailath George J., Postlewaite Andrew, and Samuelson Larry. 2014. Stable matching with incomplete information. Econometrica 82, 2 (2014), 541587.Google ScholarGoogle ScholarCross RefCross Ref
  36. [36] Mansour Yishay, Slivkins Aleksandrs, and Syrgkanis Vasilis. 2015. Bayesian incentive-compatible bandit exploration. In 16th ACM Conference on Economics and Computation (EC’15). ACM, 565582.Google ScholarGoogle ScholarDigital LibraryDigital Library
  37. [37] Martínez-Legaz J. E.. 1996. Dual representation of cooperative games based on Fenchel-Moreau conjugation. Optimization 36, 4 (1996), 291319.Google ScholarGoogle ScholarCross RefCross Ref
  38. [38] Rothschild Michael. 1974. A two-armed bandit theory of market pricing. J. Econ. Theor. 9, 2 (1974), 185202.Google ScholarGoogle ScholarCross RefCross Ref
  39. [39] Russo Daniel and Roy Benjamin Van. 2013. Eluder dimension and the sample complexity of optimistic exploration. In 27th Annual Conference on Neural Information Processing Systems. 22562264.Google ScholarGoogle Scholar
  40. [40] Sankararaman Abishek, Basu Soumya, and Sankararaman Karthik Abinav. 2021. Dominate or delete: Decentralized competing bandits in serial dictatorship. In 24th International Conference on Artificial Intelligence and Statistics(Proceedings of Machine Learning Research, Vol. 130). PMLR, 12521260.Google ScholarGoogle Scholar
  41. [41] Shapley L. S. and Shubik M.. 1966. Quasi-cores in a monetary economy with nonconvex preferences. Econometrica 34, 4 (1966), 805827.Google ScholarGoogle ScholarCross RefCross Ref
  42. [42] Shapley L. S. and Shubik M.. 1971. The assignment game I: The core. Int. J. Game Theor. 1, 1 (Dec.1971), 111130.Google ScholarGoogle ScholarDigital LibraryDigital Library
  43. [43] Shi Peng. 2020. Efficient matchmaking in assignment games with application to online platforms. In 21st ACM Conference on Economics and Computation. ACM, 601602.Google ScholarGoogle ScholarDigital LibraryDigital Library
  44. [44] Tirinzoni Andrea, Pirotta Matteo, Restelli Marcello, and Lazaric Alessandro. 2020. An asymptotically optimal primal-dual incremental algorithm for contextual linear bandits. In Annual Conference on Neural Information Processing Systems.Google ScholarGoogle Scholar

Index Terms

  1. Learning Equilibria in Matching Markets with Bandit Feedback

      Recommendations

      Comments

      Login options

      Check if you have access through your login credentials or your institution to get full access on this article.

      Sign in

      Full Access

      • Published in

        cover image Journal of the ACM
        Journal of the ACM  Volume 70, Issue 3
        June 2023
        284 pages
        ISSN:0004-5411
        EISSN:1557-735X
        DOI:10.1145/3599472
        Issue’s Table of Contents

        Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for third-party components of this work must be honored. For all other uses, contact the owner/author(s).

        Publisher

        Association for Computing Machinery

        New York, NY, United States

        Publication History

        • Published: 24 May 2023
        • Online AM: 16 February 2023
        • Accepted: 11 November 2022
        • Revised: 24 August 2022
        • Received: 31 December 2021
        Published in jacm Volume 70, Issue 3

        Permissions

        Request permissions about this article.

        Request Permissions

        Check for updates

        Qualifiers

        • research-article
      • Article Metrics

        • Downloads (Last 12 months)2,336
        • Downloads (Last 6 weeks)97

        Other Metrics

      PDF Format

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader